problem_id
stringlengths 18
22
| source
stringclasses 1
value | task_type
stringclasses 1
value | in_source_id
stringlengths 13
58
| prompt
stringlengths 1.71k
18.9k
| golden_diff
stringlengths 145
5.13k
| verification_info
stringlengths 465
23.6k
| num_tokens_prompt
int64 556
4.1k
| num_tokens_diff
int64 47
1.02k
|
---|---|---|---|---|---|---|---|---|
gh_patches_debug_7008
|
rasdani/github-patches
|
git_diff
|
sunpy__sunpy-1779
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Deprecation warnings in rescale.py
``` python
In [13]: dimensions = u.Quantity(np.array(aia.data.shape) / 2048*u.pixel)
In [14]: sp = aia.superpixel(dimensions)
d:\sunpy-master\sunpy\image\rescale.py:226: DeprecationWarning: using a non-integer number instead of an integer will result in an error in the future
offset[1]:offset[1] + nb*dimensions[1]]).reshape(na, dimensions[0], nb, dimensions[1])
d:\sunpy-master\sunpy\image\rescale.py:226: DeprecationWarning: using a non-integer number instead of an integer will result in an error in the future
offset[1]:offset[1] + nb*dimensions[1]]).reshape(na, dimensions[0], nb, dimensions[1])
d:\sunpy-master\sunpy\image\rescale.py:226: DeprecationWarning: using a non-integer number instead of an integer will result in an error in the future
offset[1]:offset[1] + nb*dimensions[1]]).reshape(na, dimensions[0], nb, dimensions[1])
d:\sunpy-master\sunpy\image\rescale.py:226: DeprecationWarning: using a non-integer number instead of an integer will result in an error in the future
offset[1]:offset[1] + nb*dimensions[1]]).reshape(na, dimensions[0], nb, dimensions[1])
d:\sunpy-master\sunpy\image\rescale.py:226: DeprecationWarning: using a non-integer number instead of an integer will result in an error in the future
offset[1]:offset[1] + nb*dimensions[1]]).reshape(na, dimensions[0], nb, dimensions[1])
d:\sunpy-master\sunpy\image\rescale.py:226: DeprecationWarning: using a non-integer number instead of an integer will result in an error in the future
offset[1]:offset[1] + nb*dimensions[1]]).reshape(na, dimensions[0], nb, dimensions[1])
```
</issue>
<code>
[start of sunpy/image/rescale.py]
1 """Image resampling methods"""
2 from __future__ import absolute_import, division, print_function
3
4 import numpy as np
5 import scipy.interpolate
6 import scipy.ndimage
7 from sunpy.extern.six.moves import range
8
9 __all__ = ['resample', 'reshape_image_to_4d_superpixel']
10
11 def resample(orig, dimensions, method='linear', center=False, minusone=False):
12 """Returns a new `numpy.ndarray` that has been resampled up or down.
13
14 Arbitrary resampling of source array to new dimension sizes.
15 Currently only supports maintaining the same number of dimensions.
16 To use 1-D arrays, first promote them to shape (x,1).
17
18 Uses the same parameters and creates the same co-ordinate lookup points
19 as IDL's congrid routine (which apparently originally came from a
20 VAX/VMS routine of the same name.)
21
22 Parameters
23 ----------
24 orig : `~numpy.ndarray`
25 Original inout array.
26 dimensions : tuple
27 Dimensions that new `~numpy.ndarray` should have.
28 method : {'neighbor' | 'nearest' | 'linear' | 'spline'}
29 Method to use for resampling interpolation.
30 * neighbor - Closest value from original data
31 * nearest and linear - Uses n x 1-D interpolations calculated by
32 `scipy.interpolate.interp1d`.
33 * spline - Uses ndimage.map_coordinates
34 center : bool
35 If True, interpolation points are at the centers of the bins,
36 otherwise points are at the front edge of the bin.
37 minusone : bool
38 For orig.shape = (i,j) & new dimensions = (x,y), if set to False
39 orig is resampled by factors of (i/x) * (j/y), otherwise orig
40 is resampled by(i-1)/(x-1) * (j-1)/(y-1)
41 This prevents extrapolation one element beyond bounds of input
42 array.
43
44 Returns
45 -------
46 out : `~numpy.ndarray`
47 A new `~numpy.ndarray` which has been resampled to the desired
48 dimensions.
49
50 References
51 ----------
52 | http://www.scipy.org/Cookbook/Rebinning (Original source, 2011/11/19)
53 """
54
55 # Verify that number dimensions requested matches original shape
56 if len(dimensions) != orig.ndim:
57 raise UnequalNumDimensions("Number of dimensions must remain the same "
58 "when calling resample.")
59
60 #@note: will this be okay for integer (e.g. JPEG 2000) data?
61 if orig.dtype not in [np.float64, np.float32]:
62 orig = orig.astype(np.float64)
63
64 dimensions = np.asarray(dimensions, dtype=np.float64)
65 m1 = np.array(minusone, dtype=np.int64) # array(0) or array(1)
66 offset = np.float64(center * 0.5) # float64(0.) or float64(0.5)
67
68 # Resample data
69 if method == 'neighbor':
70 data = _resample_neighbor(orig, dimensions, offset, m1)
71 elif method in ['nearest','linear']:
72 data = _resample_nearest_linear(orig, dimensions, method,
73 offset, m1)
74 elif method == 'spline':
75 data = _resample_spline(orig, dimensions, offset, m1)
76 else:
77 raise UnrecognizedInterpolationMethod("Unrecognized interpolation "
78 "method requested.")
79
80 return data
81
82
83 def _resample_nearest_linear(orig, dimensions, method, offset, m1):
84 """Resample Map using either linear or nearest interpolation."""
85
86 dimlist = []
87
88 # calculate new dims
89 for i in range(orig.ndim):
90 base = np.arange(dimensions[i])
91 dimlist.append((orig.shape[i] - m1) / (dimensions[i] - m1) *
92 (base + offset) - offset)
93
94 # specify old coordinates
95 old_coords = [np.arange(i, dtype=np.float) for i in orig.shape]
96
97 # first interpolation - for ndims = any
98 mint = scipy.interpolate.interp1d(old_coords[-1], orig, bounds_error=False,
99 fill_value=min(old_coords[-1]), kind=method)
100
101 new_data = mint(dimlist[-1])
102
103 trorder = [orig.ndim - 1] + list(range(orig.ndim - 1))
104 for i in range(orig.ndim - 2, -1, -1):
105 new_data = new_data.transpose(trorder)
106
107 mint = scipy.interpolate.interp1d(old_coords[i], new_data,
108 bounds_error=False, fill_value=min(old_coords[i]), kind=method)
109 new_data = mint(dimlist[i])
110
111 if orig.ndim > 1:
112 # need one more transpose to return to original dimensions
113 new_data = new_data.transpose(trorder)
114
115 return new_data
116
117
118 def _resample_neighbor(orig, dimensions, offset, m1):
119 """Resample Map using closest-value interpolation."""
120
121 dimlist = []
122
123 for i in range(orig.ndim):
124 base = np.indices(dimensions)[i]
125 dimlist.append((orig.shape[i] - m1) / (dimensions[i] - m1) *
126 (base + offset) - offset)
127 cd = np.array(dimlist).round().astype(int)
128
129 return orig[list(cd)]
130
131
132 def _resample_spline(orig, dimensions, offset, m1):
133 """Resample Map using spline-based interpolation."""
134
135 oslices = [slice(0, j) for j in orig.shape]
136 # FIXME: not used?!
137 old_coords = np.ogrid[oslices] #pylint: disable=W0612
138 nslices = [slice(0, j) for j in list(dimensions)]
139 newcoords = np.mgrid[nslices]
140
141 newcoords_dims = list(range(np.rank(newcoords)))
142
143 #make first index last
144 newcoords_dims.append(newcoords_dims.pop(0))
145 newcoords_tr = newcoords.transpose(newcoords_dims) #pylint: disable=W0612
146
147 # makes a view that affects newcoords
148 newcoords_tr += offset
149
150 deltas = (np.asarray(orig.shape) - m1) / (dimensions - m1)
151 newcoords_tr *= deltas
152
153 newcoords_tr -= offset
154
155 return scipy.ndimage.map_coordinates(orig, newcoords)
156
157
158 def reshape_image_to_4d_superpixel(img, dimensions, offset):
159 """Re-shape the two dimension input image into a a four dimensional
160 array whose first and third dimensions express the number of original
161 pixels in the x and y directions that form one superpixel. The reshaping
162 makes it very easy to perform operations on superpixels.
163
164 An application of this reshaping is the following. Let's say you have an
165 array
166
167 x = np.array([[0, 0, 0, 1, 1, 1],
168 [0, 0, 1, 1, 0, 0],
169 [1, 1, 0, 0, 1, 1],
170 [0, 0, 0, 0, 1, 1],
171 [1, 0, 1, 0, 1, 1],
172 [0, 0, 1, 0, 0, 0]])
173
174 and you want to sum over 2x2 non-overlapping sub-arrays. For example, you
175 could have a noisy image and you want to increase the signal-to-noise ratio.
176 Summing over all the non-overlapping 2x2 sub-arrays will create a
177 superpixel array of the original data. Every pixel in the superpixel array
178 is the sum of the values in a 2x2 sub-array of the original array,
179
180 This summing can be done by reshaping the array
181
182 y = x.reshape(3,2,3,2)
183
184 and then summing over the 1st and third directions
185
186 y2 = y.sum(axis=3).sum(axis=1)
187
188 which gives the expected array.
189
190 array([[0, 3, 2],
191 [2, 0, 4],
192 [1, 2, 2]])
193
194 Parameters
195 ----------
196 img : `numpy.ndarray`
197 A two-dimensional `~numpy.ndarray` of the form (y, x).
198
199 dimensions : array-like
200 A two element array-like object containing integers that describe the
201 superpixel summation in the (y, x) directions.
202
203 offset : array-like
204 A two element array-like object containing integers that describe
205 where in the input image the array reshaping begins in the (y, x)
206 directions.
207
208 Returns
209 -------
210 A four dimensional `~numpy.ndarray` that can be used to easily create
211 two-dimensional arrays of superpixels of the input image.
212
213 References
214 ----------
215 Taken from
216 http://mail.scipy.org/pipermail/numpy-discussion/2010-July/051760.html
217
218 """
219 # New dimensions of the final image
220 na = int(np.floor((img.shape[0] - offset[0]) / dimensions[0]))
221 nb = int(np.floor((img.shape[1] - offset[1]) / dimensions[1]))
222
223 # Reshape up to a higher dimensional array which is useful for higher
224 # level operations
225 return (img[offset[0]:offset[0] + na*dimensions[0],
226 offset[1]:offset[1] + nb*dimensions[1]]).reshape(na, dimensions[0], nb, dimensions[1])
227
228
229 class UnrecognizedInterpolationMethod(ValueError):
230 """Unrecognized interpolation method specified."""
231 pass
232
233
234 class UnequalNumDimensions(ValueError):
235 """Number of dimensions does not match input array"""
236 pass
237
[end of sunpy/image/rescale.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/sunpy/image/rescale.py b/sunpy/image/rescale.py
--- a/sunpy/image/rescale.py
+++ b/sunpy/image/rescale.py
@@ -216,6 +216,9 @@
http://mail.scipy.org/pipermail/numpy-discussion/2010-July/051760.html
"""
+ # make sure the input dimensions are integers
+ dimensions = [int(dim) for dim in dimensions]
+
# New dimensions of the final image
na = int(np.floor((img.shape[0] - offset[0]) / dimensions[0]))
nb = int(np.floor((img.shape[1] - offset[1]) / dimensions[1]))
|
{"golden_diff": "diff --git a/sunpy/image/rescale.py b/sunpy/image/rescale.py\n--- a/sunpy/image/rescale.py\n+++ b/sunpy/image/rescale.py\n@@ -216,6 +216,9 @@\n http://mail.scipy.org/pipermail/numpy-discussion/2010-July/051760.html\n \n \"\"\"\n+ # make sure the input dimensions are integers\n+ dimensions = [int(dim) for dim in dimensions]\n+\n # New dimensions of the final image\n na = int(np.floor((img.shape[0] - offset[0]) / dimensions[0]))\n nb = int(np.floor((img.shape[1] - offset[1]) / dimensions[1]))\n", "issue": "Deprecation warnings in rescale.py\n``` python\nIn [13]: dimensions = u.Quantity(np.array(aia.data.shape) / 2048*u.pixel)\n\nIn [14]: sp = aia.superpixel(dimensions)\nd:\\sunpy-master\\sunpy\\image\\rescale.py:226: DeprecationWarning: using a non-integer number instead of an integer will result in an error in the future\n offset[1]:offset[1] + nb*dimensions[1]]).reshape(na, dimensions[0], nb, dimensions[1])\nd:\\sunpy-master\\sunpy\\image\\rescale.py:226: DeprecationWarning: using a non-integer number instead of an integer will result in an error in the future\n offset[1]:offset[1] + nb*dimensions[1]]).reshape(na, dimensions[0], nb, dimensions[1])\nd:\\sunpy-master\\sunpy\\image\\rescale.py:226: DeprecationWarning: using a non-integer number instead of an integer will result in an error in the future\n offset[1]:offset[1] + nb*dimensions[1]]).reshape(na, dimensions[0], nb, dimensions[1])\nd:\\sunpy-master\\sunpy\\image\\rescale.py:226: DeprecationWarning: using a non-integer number instead of an integer will result in an error in the future\n offset[1]:offset[1] + nb*dimensions[1]]).reshape(na, dimensions[0], nb, dimensions[1])\nd:\\sunpy-master\\sunpy\\image\\rescale.py:226: DeprecationWarning: using a non-integer number instead of an integer will result in an error in the future\n offset[1]:offset[1] + nb*dimensions[1]]).reshape(na, dimensions[0], nb, dimensions[1])\nd:\\sunpy-master\\sunpy\\image\\rescale.py:226: DeprecationWarning: using a non-integer number instead of an integer will result in an error in the future\n offset[1]:offset[1] + nb*dimensions[1]]).reshape(na, dimensions[0], nb, dimensions[1])\n```\n\n", "before_files": [{"content": "\"\"\"Image resampling methods\"\"\"\nfrom __future__ import absolute_import, division, print_function\n\nimport numpy as np\nimport scipy.interpolate\nimport scipy.ndimage\nfrom sunpy.extern.six.moves import range\n\n__all__ = ['resample', 'reshape_image_to_4d_superpixel']\n\ndef resample(orig, dimensions, method='linear', center=False, minusone=False):\n \"\"\"Returns a new `numpy.ndarray` that has been resampled up or down.\n\n Arbitrary resampling of source array to new dimension sizes.\n Currently only supports maintaining the same number of dimensions.\n To use 1-D arrays, first promote them to shape (x,1).\n\n Uses the same parameters and creates the same co-ordinate lookup points\n as IDL's congrid routine (which apparently originally came from a\n VAX/VMS routine of the same name.)\n\n Parameters\n ----------\n orig : `~numpy.ndarray`\n Original inout array.\n dimensions : tuple\n Dimensions that new `~numpy.ndarray` should have.\n method : {'neighbor' | 'nearest' | 'linear' | 'spline'}\n Method to use for resampling interpolation.\n * neighbor - Closest value from original data\n * nearest and linear - Uses n x 1-D interpolations calculated by\n `scipy.interpolate.interp1d`.\n * spline - Uses ndimage.map_coordinates\n center : bool\n If True, interpolation points are at the centers of the bins,\n otherwise points are at the front edge of the bin.\n minusone : bool\n For orig.shape = (i,j) & new dimensions = (x,y), if set to False\n orig is resampled by factors of (i/x) * (j/y), otherwise orig\n is resampled by(i-1)/(x-1) * (j-1)/(y-1)\n This prevents extrapolation one element beyond bounds of input\n array.\n\n Returns\n -------\n out : `~numpy.ndarray`\n A new `~numpy.ndarray` which has been resampled to the desired\n dimensions.\n\n References\n ----------\n | http://www.scipy.org/Cookbook/Rebinning (Original source, 2011/11/19)\n \"\"\"\n\n # Verify that number dimensions requested matches original shape\n if len(dimensions) != orig.ndim:\n raise UnequalNumDimensions(\"Number of dimensions must remain the same \"\n \"when calling resample.\")\n\n #@note: will this be okay for integer (e.g. JPEG 2000) data?\n if orig.dtype not in [np.float64, np.float32]:\n orig = orig.astype(np.float64)\n\n dimensions = np.asarray(dimensions, dtype=np.float64)\n m1 = np.array(minusone, dtype=np.int64) # array(0) or array(1)\n offset = np.float64(center * 0.5) # float64(0.) or float64(0.5)\n\n # Resample data\n if method == 'neighbor':\n data = _resample_neighbor(orig, dimensions, offset, m1)\n elif method in ['nearest','linear']:\n data = _resample_nearest_linear(orig, dimensions, method,\n offset, m1)\n elif method == 'spline':\n data = _resample_spline(orig, dimensions, offset, m1)\n else:\n raise UnrecognizedInterpolationMethod(\"Unrecognized interpolation \"\n \"method requested.\")\n\n return data\n\n\ndef _resample_nearest_linear(orig, dimensions, method, offset, m1):\n \"\"\"Resample Map using either linear or nearest interpolation.\"\"\"\n\n dimlist = []\n\n # calculate new dims\n for i in range(orig.ndim):\n base = np.arange(dimensions[i])\n dimlist.append((orig.shape[i] - m1) / (dimensions[i] - m1) *\n (base + offset) - offset)\n\n # specify old coordinates\n old_coords = [np.arange(i, dtype=np.float) for i in orig.shape]\n\n # first interpolation - for ndims = any\n mint = scipy.interpolate.interp1d(old_coords[-1], orig, bounds_error=False,\n fill_value=min(old_coords[-1]), kind=method)\n\n new_data = mint(dimlist[-1])\n\n trorder = [orig.ndim - 1] + list(range(orig.ndim - 1))\n for i in range(orig.ndim - 2, -1, -1):\n new_data = new_data.transpose(trorder)\n\n mint = scipy.interpolate.interp1d(old_coords[i], new_data,\n bounds_error=False, fill_value=min(old_coords[i]), kind=method)\n new_data = mint(dimlist[i])\n\n if orig.ndim > 1:\n # need one more transpose to return to original dimensions\n new_data = new_data.transpose(trorder)\n\n return new_data\n\n\ndef _resample_neighbor(orig, dimensions, offset, m1):\n \"\"\"Resample Map using closest-value interpolation.\"\"\"\n\n dimlist = []\n\n for i in range(orig.ndim):\n base = np.indices(dimensions)[i]\n dimlist.append((orig.shape[i] - m1) / (dimensions[i] - m1) *\n (base + offset) - offset)\n cd = np.array(dimlist).round().astype(int)\n\n return orig[list(cd)]\n\n\ndef _resample_spline(orig, dimensions, offset, m1):\n \"\"\"Resample Map using spline-based interpolation.\"\"\"\n\n oslices = [slice(0, j) for j in orig.shape]\n # FIXME: not used?!\n old_coords = np.ogrid[oslices] #pylint: disable=W0612\n nslices = [slice(0, j) for j in list(dimensions)]\n newcoords = np.mgrid[nslices]\n\n newcoords_dims = list(range(np.rank(newcoords)))\n\n #make first index last\n newcoords_dims.append(newcoords_dims.pop(0))\n newcoords_tr = newcoords.transpose(newcoords_dims) #pylint: disable=W0612\n\n # makes a view that affects newcoords\n newcoords_tr += offset\n\n deltas = (np.asarray(orig.shape) - m1) / (dimensions - m1)\n newcoords_tr *= deltas\n\n newcoords_tr -= offset\n\n return scipy.ndimage.map_coordinates(orig, newcoords)\n\n\ndef reshape_image_to_4d_superpixel(img, dimensions, offset):\n \"\"\"Re-shape the two dimension input image into a a four dimensional\n array whose first and third dimensions express the number of original\n pixels in the x and y directions that form one superpixel. The reshaping\n makes it very easy to perform operations on superpixels.\n\n An application of this reshaping is the following. Let's say you have an\n array\n\n x = np.array([[0, 0, 0, 1, 1, 1],\n [0, 0, 1, 1, 0, 0],\n [1, 1, 0, 0, 1, 1],\n [0, 0, 0, 0, 1, 1],\n [1, 0, 1, 0, 1, 1],\n [0, 0, 1, 0, 0, 0]])\n\n and you want to sum over 2x2 non-overlapping sub-arrays. For example, you\n could have a noisy image and you want to increase the signal-to-noise ratio.\n Summing over all the non-overlapping 2x2 sub-arrays will create a\n superpixel array of the original data. Every pixel in the superpixel array\n is the sum of the values in a 2x2 sub-array of the original array,\n\n This summing can be done by reshaping the array\n\n y = x.reshape(3,2,3,2)\n\n and then summing over the 1st and third directions\n\n y2 = y.sum(axis=3).sum(axis=1)\n\n which gives the expected array.\n\n array([[0, 3, 2],\n [2, 0, 4],\n [1, 2, 2]])\n\n Parameters\n ----------\n img : `numpy.ndarray`\n A two-dimensional `~numpy.ndarray` of the form (y, x).\n\n dimensions : array-like\n A two element array-like object containing integers that describe the\n superpixel summation in the (y, x) directions.\n\n offset : array-like\n A two element array-like object containing integers that describe\n where in the input image the array reshaping begins in the (y, x)\n directions.\n\n Returns\n -------\n A four dimensional `~numpy.ndarray` that can be used to easily create\n two-dimensional arrays of superpixels of the input image.\n\n References\n ----------\n Taken from\n http://mail.scipy.org/pipermail/numpy-discussion/2010-July/051760.html\n\n \"\"\"\n # New dimensions of the final image\n na = int(np.floor((img.shape[0] - offset[0]) / dimensions[0]))\n nb = int(np.floor((img.shape[1] - offset[1]) / dimensions[1]))\n\n # Reshape up to a higher dimensional array which is useful for higher\n # level operations\n return (img[offset[0]:offset[0] + na*dimensions[0],\n offset[1]:offset[1] + nb*dimensions[1]]).reshape(na, dimensions[0], nb, dimensions[1])\n\n\nclass UnrecognizedInterpolationMethod(ValueError):\n \"\"\"Unrecognized interpolation method specified.\"\"\"\n pass\n\n\nclass UnequalNumDimensions(ValueError):\n \"\"\"Number of dimensions does not match input array\"\"\"\n pass\n", "path": "sunpy/image/rescale.py"}]}
| 3,805 | 164 |
gh_patches_debug_18423
|
rasdani/github-patches
|
git_diff
|
zulip__zulip-20366
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Use SHA-256 for BigBlueButton checksums
BigBlueButton's API uses SHA1 for its checksum; I marked is as dismissed for CodeQL since it's not something we can directly change without SHA256 support in BigBlueButton itself. I note that https://github.com/bigbluebutton/bigbluebutton/issues/17135 and related issues suggest it might be possible to do so today, though apparently the way to do that may not be documented. So probably not something we can change without testing with a running BigBlueButton instance.
Noticed by GitHub Actions in https://github.com/zulip/zulip/pull/19962; I disabled the CodeQL warning for this so CI remains happy.
@strifel FYI as the original author of our BigBlueButton plugin.
</issue>
<code>
[start of zerver/views/video_calls.py]
1 import hashlib
2 import json
3 import random
4 import secrets
5 from base64 import b32encode
6 from functools import partial
7 from typing import Dict
8 from urllib.parse import quote, urlencode, urljoin
9
10 import requests
11 from defusedxml import ElementTree
12 from django.conf import settings
13 from django.http import HttpRequest, HttpResponse
14 from django.middleware import csrf
15 from django.shortcuts import redirect, render
16 from django.utils.crypto import constant_time_compare, salted_hmac
17 from django.utils.translation import gettext as _
18 from django.views.decorators.cache import never_cache
19 from django.views.decorators.csrf import csrf_exempt
20 from django.views.decorators.http import require_POST
21 from oauthlib.oauth2 import OAuth2Error
22 from requests_oauthlib import OAuth2Session
23
24 from zerver.decorator import zulip_login_required
25 from zerver.lib.actions import do_set_zoom_token
26 from zerver.lib.exceptions import ErrorCode, JsonableError
27 from zerver.lib.outgoing_http import OutgoingSession
28 from zerver.lib.pysa import mark_sanitized
29 from zerver.lib.request import REQ, has_request_variables
30 from zerver.lib.response import json_success
31 from zerver.lib.subdomains import get_subdomain
32 from zerver.lib.url_encoding import append_url_query_string
33 from zerver.lib.validator import check_dict, check_string
34 from zerver.models import UserProfile, get_realm
35
36
37 class VideoCallSession(OutgoingSession):
38 def __init__(self) -> None:
39 super().__init__(role="video_calls", timeout=5)
40
41
42 class InvalidZoomTokenError(JsonableError):
43 code = ErrorCode.INVALID_ZOOM_TOKEN
44
45 def __init__(self) -> None:
46 super().__init__(_("Invalid Zoom access token"))
47
48
49 def get_zoom_session(user: UserProfile) -> OAuth2Session:
50 if settings.VIDEO_ZOOM_CLIENT_ID is None:
51 raise JsonableError(_("Zoom credentials have not been configured"))
52
53 client_id = settings.VIDEO_ZOOM_CLIENT_ID
54 client_secret = settings.VIDEO_ZOOM_CLIENT_SECRET
55 if user.realm.string_id in settings.VIDEO_ZOOM_TESTING_REALMS: # nocoverage
56 client_id = settings.VIDEO_ZOOM_TESTING_CLIENT_ID
57 client_secret = settings.VIDEO_ZOOM_TESTING_CLIENT_SECRET
58
59 return OAuth2Session(
60 client_id,
61 redirect_uri=urljoin(settings.ROOT_DOMAIN_URI, "/calls/zoom/complete"),
62 auto_refresh_url="https://zoom.us/oauth/token",
63 auto_refresh_kwargs={
64 "client_id": client_id,
65 "client_secret": client_secret,
66 },
67 token=user.zoom_token,
68 token_updater=partial(do_set_zoom_token, user),
69 )
70
71
72 def get_zoom_sid(request: HttpRequest) -> str:
73 # This is used to prevent CSRF attacks on the Zoom OAuth
74 # authentication flow. We want this value to be unpredictable and
75 # tied to the session, but we don’t want to expose the main CSRF
76 # token directly to the Zoom server.
77
78 csrf.get_token(request)
79 # Use 'mark_sanitized' to cause Pysa to ignore the flow of user controlled
80 # data out of this function. 'request.META' is indeed user controlled, but
81 # post-HMAC output is no longer meaningfully controllable.
82 return mark_sanitized(
83 ""
84 if getattr(request, "_dont_enforce_csrf_checks", False)
85 else salted_hmac("Zulip Zoom sid", request.META["CSRF_COOKIE"]).hexdigest()
86 )
87
88
89 @zulip_login_required
90 @never_cache
91 def register_zoom_user(request: HttpRequest) -> HttpResponse:
92 assert request.user.is_authenticated
93
94 oauth = get_zoom_session(request.user)
95 authorization_url, state = oauth.authorization_url(
96 "https://zoom.us/oauth/authorize",
97 state=json.dumps(
98 {"realm": get_subdomain(request), "sid": get_zoom_sid(request)},
99 ),
100 )
101 return redirect(authorization_url)
102
103
104 @never_cache
105 @has_request_variables
106 def complete_zoom_user(
107 request: HttpRequest,
108 state: Dict[str, str] = REQ(
109 json_validator=check_dict([("realm", check_string)], value_validator=check_string)
110 ),
111 ) -> HttpResponse:
112 if get_subdomain(request) != state["realm"]:
113 return redirect(urljoin(get_realm(state["realm"]).uri, request.get_full_path()))
114 return complete_zoom_user_in_realm(request)
115
116
117 @zulip_login_required
118 @has_request_variables
119 def complete_zoom_user_in_realm(
120 request: HttpRequest,
121 code: str = REQ(),
122 state: Dict[str, str] = REQ(
123 json_validator=check_dict([("sid", check_string)], value_validator=check_string)
124 ),
125 ) -> HttpResponse:
126 assert request.user.is_authenticated
127
128 if not constant_time_compare(state["sid"], get_zoom_sid(request)):
129 raise JsonableError(_("Invalid Zoom session identifier"))
130
131 client_secret = settings.VIDEO_ZOOM_CLIENT_SECRET
132 if request.user.realm.string_id in settings.VIDEO_ZOOM_TESTING_REALMS: # nocoverage
133 client_secret = settings.VIDEO_ZOOM_TESTING_CLIENT_SECRET
134
135 oauth = get_zoom_session(request.user)
136 try:
137 token = oauth.fetch_token(
138 "https://zoom.us/oauth/token",
139 code=code,
140 client_secret=client_secret,
141 )
142 except OAuth2Error:
143 raise JsonableError(_("Invalid Zoom credentials"))
144
145 do_set_zoom_token(request.user, token)
146 return render(request, "zerver/close_window.html")
147
148
149 def make_zoom_video_call(request: HttpRequest, user: UserProfile) -> HttpResponse:
150 oauth = get_zoom_session(user)
151 if not oauth.authorized:
152 raise InvalidZoomTokenError
153
154 try:
155 res = oauth.post("https://api.zoom.us/v2/users/me/meetings", json={})
156 except OAuth2Error:
157 do_set_zoom_token(user, None)
158 raise InvalidZoomTokenError
159
160 if res.status_code == 401:
161 do_set_zoom_token(user, None)
162 raise InvalidZoomTokenError
163 elif not res.ok:
164 raise JsonableError(_("Failed to create Zoom call"))
165
166 return json_success({"url": res.json()["join_url"]})
167
168
169 @csrf_exempt
170 @require_POST
171 @has_request_variables
172 def deauthorize_zoom_user(request: HttpRequest) -> HttpResponse:
173 return json_success()
174
175
176 def get_bigbluebutton_url(request: HttpRequest, user_profile: UserProfile) -> HttpResponse:
177 # https://docs.bigbluebutton.org/dev/api.html#create for reference on the API calls
178 # https://docs.bigbluebutton.org/dev/api.html#usage for reference for checksum
179 id = "zulip-" + str(random.randint(100000000000, 999999999999))
180 password = b32encode(secrets.token_bytes(7))[:10].decode()
181 checksum = hashlib.sha1(
182 (
183 "create"
184 + "meetingID="
185 + id
186 + "&moderatorPW="
187 + password
188 + "&attendeePW="
189 + password
190 + "a"
191 + settings.BIG_BLUE_BUTTON_SECRET
192 ).encode()
193 ).hexdigest()
194 url = append_url_query_string(
195 "/calls/bigbluebutton/join",
196 urlencode(
197 {
198 "meeting_id": id,
199 "password": password,
200 "checksum": checksum,
201 }
202 ),
203 )
204 return json_success({"url": url})
205
206
207 # We use zulip_login_required here mainly to get access to the user's
208 # full name from Zulip to prepopulate the user's name in the
209 # BigBlueButton meeting. Since the meeting's details are encoded in
210 # the link the user is clicking, there is no validation tying this
211 # meeting to the Zulip organization it was created in.
212 @zulip_login_required
213 @never_cache
214 @has_request_variables
215 def join_bigbluebutton(
216 request: HttpRequest,
217 meeting_id: str = REQ(),
218 password: str = REQ(),
219 checksum: str = REQ(),
220 ) -> HttpResponse:
221 assert request.user.is_authenticated
222
223 if settings.BIG_BLUE_BUTTON_URL is None or settings.BIG_BLUE_BUTTON_SECRET is None:
224 raise JsonableError(_("BigBlueButton is not configured."))
225 else:
226 try:
227 response = VideoCallSession().get(
228 append_url_query_string(
229 settings.BIG_BLUE_BUTTON_URL + "api/create",
230 urlencode(
231 {
232 "meetingID": meeting_id,
233 "moderatorPW": password,
234 "attendeePW": password + "a",
235 "checksum": checksum,
236 }
237 ),
238 )
239 )
240 response.raise_for_status()
241 except requests.RequestException:
242 raise JsonableError(_("Error connecting to the BigBlueButton server."))
243
244 payload = ElementTree.fromstring(response.text)
245 if payload.find("messageKey").text == "checksumError":
246 raise JsonableError(_("Error authenticating to the BigBlueButton server."))
247
248 if payload.find("returncode").text != "SUCCESS":
249 raise JsonableError(_("BigBlueButton server returned an unexpected error."))
250
251 join_params = urlencode( # type: ignore[type-var] # https://github.com/python/typeshed/issues/4234
252 {
253 "meetingID": meeting_id,
254 "password": password,
255 "fullName": request.user.full_name,
256 },
257 quote_via=quote,
258 )
259
260 checksum = hashlib.sha1(
261 ("join" + join_params + settings.BIG_BLUE_BUTTON_SECRET).encode()
262 ).hexdigest()
263 redirect_url_base = append_url_query_string(
264 settings.BIG_BLUE_BUTTON_URL + "api/join", join_params
265 )
266 return redirect(append_url_query_string(redirect_url_base, "checksum=" + checksum))
267
[end of zerver/views/video_calls.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/zerver/views/video_calls.py b/zerver/views/video_calls.py
--- a/zerver/views/video_calls.py
+++ b/zerver/views/video_calls.py
@@ -178,7 +178,7 @@
# https://docs.bigbluebutton.org/dev/api.html#usage for reference for checksum
id = "zulip-" + str(random.randint(100000000000, 999999999999))
password = b32encode(secrets.token_bytes(7))[:10].decode()
- checksum = hashlib.sha1(
+ checksum = hashlib.sha256(
(
"create"
+ "meetingID="
@@ -257,7 +257,7 @@
quote_via=quote,
)
- checksum = hashlib.sha1(
+ checksum = hashlib.sha256(
("join" + join_params + settings.BIG_BLUE_BUTTON_SECRET).encode()
).hexdigest()
redirect_url_base = append_url_query_string(
|
{"golden_diff": "diff --git a/zerver/views/video_calls.py b/zerver/views/video_calls.py\n--- a/zerver/views/video_calls.py\n+++ b/zerver/views/video_calls.py\n@@ -178,7 +178,7 @@\n # https://docs.bigbluebutton.org/dev/api.html#usage for reference for checksum\n id = \"zulip-\" + str(random.randint(100000000000, 999999999999))\n password = b32encode(secrets.token_bytes(7))[:10].decode()\n- checksum = hashlib.sha1(\n+ checksum = hashlib.sha256(\n (\n \"create\"\n + \"meetingID=\"\n@@ -257,7 +257,7 @@\n quote_via=quote,\n )\n \n- checksum = hashlib.sha1(\n+ checksum = hashlib.sha256(\n (\"join\" + join_params + settings.BIG_BLUE_BUTTON_SECRET).encode()\n ).hexdigest()\n redirect_url_base = append_url_query_string(\n", "issue": "Use SHA-256 for BigBlueButton checksums\nBigBlueButton's API uses SHA1 for its checksum; I marked is as dismissed for CodeQL since it's not something we can directly change without SHA256 support in BigBlueButton itself. I note that https://github.com/bigbluebutton/bigbluebutton/issues/17135 and related issues suggest it might be possible to do so today, though apparently the way to do that may not be documented. So probably not something we can change without testing with a running BigBlueButton instance.\r\n\r\nNoticed by GitHub Actions in https://github.com/zulip/zulip/pull/19962; I disabled the CodeQL warning for this so CI remains happy.\r\n\r\n@strifel FYI as the original author of our BigBlueButton plugin.\n", "before_files": [{"content": "import hashlib\nimport json\nimport random\nimport secrets\nfrom base64 import b32encode\nfrom functools import partial\nfrom typing import Dict\nfrom urllib.parse import quote, urlencode, urljoin\n\nimport requests\nfrom defusedxml import ElementTree\nfrom django.conf import settings\nfrom django.http import HttpRequest, HttpResponse\nfrom django.middleware import csrf\nfrom django.shortcuts import redirect, render\nfrom django.utils.crypto import constant_time_compare, salted_hmac\nfrom django.utils.translation import gettext as _\nfrom django.views.decorators.cache import never_cache\nfrom django.views.decorators.csrf import csrf_exempt\nfrom django.views.decorators.http import require_POST\nfrom oauthlib.oauth2 import OAuth2Error\nfrom requests_oauthlib import OAuth2Session\n\nfrom zerver.decorator import zulip_login_required\nfrom zerver.lib.actions import do_set_zoom_token\nfrom zerver.lib.exceptions import ErrorCode, JsonableError\nfrom zerver.lib.outgoing_http import OutgoingSession\nfrom zerver.lib.pysa import mark_sanitized\nfrom zerver.lib.request import REQ, has_request_variables\nfrom zerver.lib.response import json_success\nfrom zerver.lib.subdomains import get_subdomain\nfrom zerver.lib.url_encoding import append_url_query_string\nfrom zerver.lib.validator import check_dict, check_string\nfrom zerver.models import UserProfile, get_realm\n\n\nclass VideoCallSession(OutgoingSession):\n def __init__(self) -> None:\n super().__init__(role=\"video_calls\", timeout=5)\n\n\nclass InvalidZoomTokenError(JsonableError):\n code = ErrorCode.INVALID_ZOOM_TOKEN\n\n def __init__(self) -> None:\n super().__init__(_(\"Invalid Zoom access token\"))\n\n\ndef get_zoom_session(user: UserProfile) -> OAuth2Session:\n if settings.VIDEO_ZOOM_CLIENT_ID is None:\n raise JsonableError(_(\"Zoom credentials have not been configured\"))\n\n client_id = settings.VIDEO_ZOOM_CLIENT_ID\n client_secret = settings.VIDEO_ZOOM_CLIENT_SECRET\n if user.realm.string_id in settings.VIDEO_ZOOM_TESTING_REALMS: # nocoverage\n client_id = settings.VIDEO_ZOOM_TESTING_CLIENT_ID\n client_secret = settings.VIDEO_ZOOM_TESTING_CLIENT_SECRET\n\n return OAuth2Session(\n client_id,\n redirect_uri=urljoin(settings.ROOT_DOMAIN_URI, \"/calls/zoom/complete\"),\n auto_refresh_url=\"https://zoom.us/oauth/token\",\n auto_refresh_kwargs={\n \"client_id\": client_id,\n \"client_secret\": client_secret,\n },\n token=user.zoom_token,\n token_updater=partial(do_set_zoom_token, user),\n )\n\n\ndef get_zoom_sid(request: HttpRequest) -> str:\n # This is used to prevent CSRF attacks on the Zoom OAuth\n # authentication flow. We want this value to be unpredictable and\n # tied to the session, but we don\u2019t want to expose the main CSRF\n # token directly to the Zoom server.\n\n csrf.get_token(request)\n # Use 'mark_sanitized' to cause Pysa to ignore the flow of user controlled\n # data out of this function. 'request.META' is indeed user controlled, but\n # post-HMAC output is no longer meaningfully controllable.\n return mark_sanitized(\n \"\"\n if getattr(request, \"_dont_enforce_csrf_checks\", False)\n else salted_hmac(\"Zulip Zoom sid\", request.META[\"CSRF_COOKIE\"]).hexdigest()\n )\n\n\n@zulip_login_required\n@never_cache\ndef register_zoom_user(request: HttpRequest) -> HttpResponse:\n assert request.user.is_authenticated\n\n oauth = get_zoom_session(request.user)\n authorization_url, state = oauth.authorization_url(\n \"https://zoom.us/oauth/authorize\",\n state=json.dumps(\n {\"realm\": get_subdomain(request), \"sid\": get_zoom_sid(request)},\n ),\n )\n return redirect(authorization_url)\n\n\n@never_cache\n@has_request_variables\ndef complete_zoom_user(\n request: HttpRequest,\n state: Dict[str, str] = REQ(\n json_validator=check_dict([(\"realm\", check_string)], value_validator=check_string)\n ),\n) -> HttpResponse:\n if get_subdomain(request) != state[\"realm\"]:\n return redirect(urljoin(get_realm(state[\"realm\"]).uri, request.get_full_path()))\n return complete_zoom_user_in_realm(request)\n\n\n@zulip_login_required\n@has_request_variables\ndef complete_zoom_user_in_realm(\n request: HttpRequest,\n code: str = REQ(),\n state: Dict[str, str] = REQ(\n json_validator=check_dict([(\"sid\", check_string)], value_validator=check_string)\n ),\n) -> HttpResponse:\n assert request.user.is_authenticated\n\n if not constant_time_compare(state[\"sid\"], get_zoom_sid(request)):\n raise JsonableError(_(\"Invalid Zoom session identifier\"))\n\n client_secret = settings.VIDEO_ZOOM_CLIENT_SECRET\n if request.user.realm.string_id in settings.VIDEO_ZOOM_TESTING_REALMS: # nocoverage\n client_secret = settings.VIDEO_ZOOM_TESTING_CLIENT_SECRET\n\n oauth = get_zoom_session(request.user)\n try:\n token = oauth.fetch_token(\n \"https://zoom.us/oauth/token\",\n code=code,\n client_secret=client_secret,\n )\n except OAuth2Error:\n raise JsonableError(_(\"Invalid Zoom credentials\"))\n\n do_set_zoom_token(request.user, token)\n return render(request, \"zerver/close_window.html\")\n\n\ndef make_zoom_video_call(request: HttpRequest, user: UserProfile) -> HttpResponse:\n oauth = get_zoom_session(user)\n if not oauth.authorized:\n raise InvalidZoomTokenError\n\n try:\n res = oauth.post(\"https://api.zoom.us/v2/users/me/meetings\", json={})\n except OAuth2Error:\n do_set_zoom_token(user, None)\n raise InvalidZoomTokenError\n\n if res.status_code == 401:\n do_set_zoom_token(user, None)\n raise InvalidZoomTokenError\n elif not res.ok:\n raise JsonableError(_(\"Failed to create Zoom call\"))\n\n return json_success({\"url\": res.json()[\"join_url\"]})\n\n\n@csrf_exempt\n@require_POST\n@has_request_variables\ndef deauthorize_zoom_user(request: HttpRequest) -> HttpResponse:\n return json_success()\n\n\ndef get_bigbluebutton_url(request: HttpRequest, user_profile: UserProfile) -> HttpResponse:\n # https://docs.bigbluebutton.org/dev/api.html#create for reference on the API calls\n # https://docs.bigbluebutton.org/dev/api.html#usage for reference for checksum\n id = \"zulip-\" + str(random.randint(100000000000, 999999999999))\n password = b32encode(secrets.token_bytes(7))[:10].decode()\n checksum = hashlib.sha1(\n (\n \"create\"\n + \"meetingID=\"\n + id\n + \"&moderatorPW=\"\n + password\n + \"&attendeePW=\"\n + password\n + \"a\"\n + settings.BIG_BLUE_BUTTON_SECRET\n ).encode()\n ).hexdigest()\n url = append_url_query_string(\n \"/calls/bigbluebutton/join\",\n urlencode(\n {\n \"meeting_id\": id,\n \"password\": password,\n \"checksum\": checksum,\n }\n ),\n )\n return json_success({\"url\": url})\n\n\n# We use zulip_login_required here mainly to get access to the user's\n# full name from Zulip to prepopulate the user's name in the\n# BigBlueButton meeting. Since the meeting's details are encoded in\n# the link the user is clicking, there is no validation tying this\n# meeting to the Zulip organization it was created in.\n@zulip_login_required\n@never_cache\n@has_request_variables\ndef join_bigbluebutton(\n request: HttpRequest,\n meeting_id: str = REQ(),\n password: str = REQ(),\n checksum: str = REQ(),\n) -> HttpResponse:\n assert request.user.is_authenticated\n\n if settings.BIG_BLUE_BUTTON_URL is None or settings.BIG_BLUE_BUTTON_SECRET is None:\n raise JsonableError(_(\"BigBlueButton is not configured.\"))\n else:\n try:\n response = VideoCallSession().get(\n append_url_query_string(\n settings.BIG_BLUE_BUTTON_URL + \"api/create\",\n urlencode(\n {\n \"meetingID\": meeting_id,\n \"moderatorPW\": password,\n \"attendeePW\": password + \"a\",\n \"checksum\": checksum,\n }\n ),\n )\n )\n response.raise_for_status()\n except requests.RequestException:\n raise JsonableError(_(\"Error connecting to the BigBlueButton server.\"))\n\n payload = ElementTree.fromstring(response.text)\n if payload.find(\"messageKey\").text == \"checksumError\":\n raise JsonableError(_(\"Error authenticating to the BigBlueButton server.\"))\n\n if payload.find(\"returncode\").text != \"SUCCESS\":\n raise JsonableError(_(\"BigBlueButton server returned an unexpected error.\"))\n\n join_params = urlencode( # type: ignore[type-var] # https://github.com/python/typeshed/issues/4234\n {\n \"meetingID\": meeting_id,\n \"password\": password,\n \"fullName\": request.user.full_name,\n },\n quote_via=quote,\n )\n\n checksum = hashlib.sha1(\n (\"join\" + join_params + settings.BIG_BLUE_BUTTON_SECRET).encode()\n ).hexdigest()\n redirect_url_base = append_url_query_string(\n settings.BIG_BLUE_BUTTON_URL + \"api/join\", join_params\n )\n return redirect(append_url_query_string(redirect_url_base, \"checksum=\" + checksum))\n", "path": "zerver/views/video_calls.py"}]}
| 3,486 | 231 |
gh_patches_debug_11468
|
rasdani/github-patches
|
git_diff
|
getredash__redash-4582
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
TreasureData getSchema fails when setting non-default region
<!--
#####################################################################
#
# Need support? USE THE FORUM! https://discuss.redash.io/c/support.
#
# Don't have steps to reproduce and actually not sure it's a bug?
# Use the forum! https://discuss.redash.io/c/support.
#
#####################################################################
**Got an idea for a new feature?** Check if it isn't on the roadmap already: https://bit.ly/redash-roadmap and start a new discussion in the features category: https://discuss.redash.io/c/feature-requests 🌟.
Found a bug? Please fill out the sections below... thank you 👍
Found a security vulnerability? Please email [email protected] to report any security vulnerabilities. We will acknowledge receipt of your vulnerability and strive to send you regular updates about our progress. If you're curious about the status of your disclosure please feel free to email us again. If you want to encrypt your disclosure email, you can use this PGP key.
-->
### Issue Summary
There are some regions in Treasure Data, but getSchema alsways fails when setting non-default region.
### Steps to Reproduce
1. Set datasource using non-default region (e.g. Tokyo region)
2. Push schema refresh then "Schema refresh failed" error occurs
### Technical details:
* Redash Version: confirmed v5.0.2
* Browser/OS: any Browsers/OSs
* How did you install Redash: from Amazon AMI
### Details
When accessing Treasure Data to get schema, always default region will be set because the parameter is not prepared.
https://github.com/getredash/redash/blob/6c364369bb0eb98e2191c2e502fed72abe5a74c7/redash/query_runner/treasuredata.py#L82
</issue>
<code>
[start of redash/query_runner/treasuredata.py]
1 import logging
2
3 from redash.query_runner import *
4 from redash.utils import json_dumps
5
6 logger = logging.getLogger(__name__)
7
8 try:
9 import tdclient
10 from tdclient import errors
11
12 enabled = True
13
14 except ImportError:
15 enabled = False
16
17 TD_TYPES_MAPPING = {
18 "bigint": TYPE_INTEGER,
19 "tinyint": TYPE_INTEGER,
20 "smallint": TYPE_INTEGER,
21 "int": TYPE_INTEGER,
22 "integer": TYPE_INTEGER,
23 "long": TYPE_INTEGER,
24 "double": TYPE_FLOAT,
25 "decimal": TYPE_FLOAT,
26 "float": TYPE_FLOAT,
27 "real": TYPE_FLOAT,
28 "boolean": TYPE_BOOLEAN,
29 "timestamp": TYPE_DATETIME,
30 "date": TYPE_DATETIME,
31 "char": TYPE_STRING,
32 "string": TYPE_STRING,
33 "varchar": TYPE_STRING,
34 }
35
36
37 class TreasureData(BaseQueryRunner):
38 should_annotate_query = False
39 noop_query = "SELECT 1"
40
41 @classmethod
42 def configuration_schema(cls):
43 return {
44 "type": "object",
45 "properties": {
46 "endpoint": {"type": "string"},
47 "apikey": {"type": "string"},
48 "type": {"type": "string"},
49 "db": {"type": "string", "title": "Database Name"},
50 "get_schema": {
51 "type": "boolean",
52 "title": "Auto Schema Retrieval",
53 "default": False,
54 },
55 },
56 "required": ["apikey", "db"],
57 }
58
59 @classmethod
60 def enabled(cls):
61 return enabled
62
63 @classmethod
64 def type(cls):
65 return "treasuredata"
66
67 def get_schema(self, get_stats=False):
68 schema = {}
69 if self.configuration.get("get_schema", False):
70 try:
71 with tdclient.Client(self.configuration.get("apikey")) as client:
72 for table in client.tables(self.configuration.get("db")):
73 table_name = "{}.{}".format(
74 self.configuration.get("db"), table.name
75 )
76 for table_schema in table.schema:
77 schema[table_name] = {
78 "name": table_name,
79 "columns": [column[0] for column in table.schema],
80 }
81 except Exception as ex:
82 raise Exception("Failed getting schema")
83 return list(schema.values())
84
85 def run_query(self, query, user):
86 connection = tdclient.connect(
87 endpoint=self.configuration.get("endpoint", "https://api.treasuredata.com"),
88 apikey=self.configuration.get("apikey"),
89 type=self.configuration.get("type", "hive").lower(),
90 db=self.configuration.get("db"),
91 )
92
93 cursor = connection.cursor()
94 try:
95 cursor.execute(query)
96 columns_tuples = [
97 (i[0], TD_TYPES_MAPPING.get(i[1], None))
98 for i in cursor.show_job()["hive_result_schema"]
99 ]
100 columns = self.fetch_columns(columns_tuples)
101
102 if cursor.rowcount == 0:
103 rows = []
104 else:
105 rows = [
106 dict(zip(([column["name"] for column in columns]), r))
107 for r in cursor.fetchall()
108 ]
109 data = {"columns": columns, "rows": rows}
110 json_data = json_dumps(data)
111 error = None
112 except errors.InternalError as e:
113 json_data = None
114 error = "%s: %s" % (
115 str(e),
116 cursor.show_job()
117 .get("debug", {})
118 .get("stderr", "No stderr message in the response"),
119 )
120 return json_data, error
121
122
123 register(TreasureData)
124
[end of redash/query_runner/treasuredata.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/redash/query_runner/treasuredata.py b/redash/query_runner/treasuredata.py
--- a/redash/query_runner/treasuredata.py
+++ b/redash/query_runner/treasuredata.py
@@ -68,7 +68,7 @@
schema = {}
if self.configuration.get("get_schema", False):
try:
- with tdclient.Client(self.configuration.get("apikey")) as client:
+ with tdclient.Client(self.configuration.get("apikey"),endpoint=self.configuration.get("endpoint")) as client:
for table in client.tables(self.configuration.get("db")):
table_name = "{}.{}".format(
self.configuration.get("db"), table.name
|
{"golden_diff": "diff --git a/redash/query_runner/treasuredata.py b/redash/query_runner/treasuredata.py\n--- a/redash/query_runner/treasuredata.py\n+++ b/redash/query_runner/treasuredata.py\n@@ -68,7 +68,7 @@\n schema = {}\n if self.configuration.get(\"get_schema\", False):\n try:\n- with tdclient.Client(self.configuration.get(\"apikey\")) as client:\n+ with tdclient.Client(self.configuration.get(\"apikey\"),endpoint=self.configuration.get(\"endpoint\")) as client:\n for table in client.tables(self.configuration.get(\"db\")):\n table_name = \"{}.{}\".format(\n self.configuration.get(\"db\"), table.name\n", "issue": "TreasureData getSchema fails when setting non-default region\n<!--\r\n#####################################################################\r\n#\r\n# Need support? USE THE FORUM! https://discuss.redash.io/c/support.\r\n#\r\n# Don't have steps to reproduce and actually not sure it's a bug?\r\n# Use the forum! https://discuss.redash.io/c/support.\r\n#\r\n#####################################################################\r\n\r\n**Got an idea for a new feature?** Check if it isn't on the roadmap already: https://bit.ly/redash-roadmap and start a new discussion in the features category: https://discuss.redash.io/c/feature-requests \ud83c\udf1f.\r\n\r\nFound a bug? Please fill out the sections below... thank you \ud83d\udc4d\r\n\r\nFound a security vulnerability? Please email [email protected] to report any security vulnerabilities. We will acknowledge receipt of your vulnerability and strive to send you regular updates about our progress. If you're curious about the status of your disclosure please feel free to email us again. If you want to encrypt your disclosure email, you can use this PGP key.\r\n\r\n-->\r\n\r\n### Issue Summary\r\n\r\nThere are some regions in Treasure Data, but getSchema alsways fails when setting non-default region.\r\n\r\n### Steps to Reproduce\r\n\r\n1. Set datasource using non-default region (e.g. Tokyo region)\r\n2. Push schema refresh then \"Schema refresh failed\" error occurs\r\n\r\n### Technical details:\r\n\r\n* Redash Version: confirmed v5.0.2\r\n* Browser/OS: any Browsers/OSs\r\n* How did you install Redash: from Amazon AMI\r\n\r\n### Details\r\n\r\nWhen accessing Treasure Data to get schema, always default region will be set because the parameter is not prepared.\r\nhttps://github.com/getredash/redash/blob/6c364369bb0eb98e2191c2e502fed72abe5a74c7/redash/query_runner/treasuredata.py#L82\n", "before_files": [{"content": "import logging\n\nfrom redash.query_runner import *\nfrom redash.utils import json_dumps\n\nlogger = logging.getLogger(__name__)\n\ntry:\n import tdclient\n from tdclient import errors\n\n enabled = True\n\nexcept ImportError:\n enabled = False\n\nTD_TYPES_MAPPING = {\n \"bigint\": TYPE_INTEGER,\n \"tinyint\": TYPE_INTEGER,\n \"smallint\": TYPE_INTEGER,\n \"int\": TYPE_INTEGER,\n \"integer\": TYPE_INTEGER,\n \"long\": TYPE_INTEGER,\n \"double\": TYPE_FLOAT,\n \"decimal\": TYPE_FLOAT,\n \"float\": TYPE_FLOAT,\n \"real\": TYPE_FLOAT,\n \"boolean\": TYPE_BOOLEAN,\n \"timestamp\": TYPE_DATETIME,\n \"date\": TYPE_DATETIME,\n \"char\": TYPE_STRING,\n \"string\": TYPE_STRING,\n \"varchar\": TYPE_STRING,\n}\n\n\nclass TreasureData(BaseQueryRunner):\n should_annotate_query = False\n noop_query = \"SELECT 1\"\n\n @classmethod\n def configuration_schema(cls):\n return {\n \"type\": \"object\",\n \"properties\": {\n \"endpoint\": {\"type\": \"string\"},\n \"apikey\": {\"type\": \"string\"},\n \"type\": {\"type\": \"string\"},\n \"db\": {\"type\": \"string\", \"title\": \"Database Name\"},\n \"get_schema\": {\n \"type\": \"boolean\",\n \"title\": \"Auto Schema Retrieval\",\n \"default\": False,\n },\n },\n \"required\": [\"apikey\", \"db\"],\n }\n\n @classmethod\n def enabled(cls):\n return enabled\n\n @classmethod\n def type(cls):\n return \"treasuredata\"\n\n def get_schema(self, get_stats=False):\n schema = {}\n if self.configuration.get(\"get_schema\", False):\n try:\n with tdclient.Client(self.configuration.get(\"apikey\")) as client:\n for table in client.tables(self.configuration.get(\"db\")):\n table_name = \"{}.{}\".format(\n self.configuration.get(\"db\"), table.name\n )\n for table_schema in table.schema:\n schema[table_name] = {\n \"name\": table_name,\n \"columns\": [column[0] for column in table.schema],\n }\n except Exception as ex:\n raise Exception(\"Failed getting schema\")\n return list(schema.values())\n\n def run_query(self, query, user):\n connection = tdclient.connect(\n endpoint=self.configuration.get(\"endpoint\", \"https://api.treasuredata.com\"),\n apikey=self.configuration.get(\"apikey\"),\n type=self.configuration.get(\"type\", \"hive\").lower(),\n db=self.configuration.get(\"db\"),\n )\n\n cursor = connection.cursor()\n try:\n cursor.execute(query)\n columns_tuples = [\n (i[0], TD_TYPES_MAPPING.get(i[1], None))\n for i in cursor.show_job()[\"hive_result_schema\"]\n ]\n columns = self.fetch_columns(columns_tuples)\n\n if cursor.rowcount == 0:\n rows = []\n else:\n rows = [\n dict(zip(([column[\"name\"] for column in columns]), r))\n for r in cursor.fetchall()\n ]\n data = {\"columns\": columns, \"rows\": rows}\n json_data = json_dumps(data)\n error = None\n except errors.InternalError as e:\n json_data = None\n error = \"%s: %s\" % (\n str(e),\n cursor.show_job()\n .get(\"debug\", {})\n .get(\"stderr\", \"No stderr message in the response\"),\n )\n return json_data, error\n\n\nregister(TreasureData)\n", "path": "redash/query_runner/treasuredata.py"}]}
| 1,971 | 147 |
gh_patches_debug_34692
|
rasdani/github-patches
|
git_diff
|
bridgecrewio__checkov-3747
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Checkov failing to perform a check against a Terraform Plan when skipping a specific AWS check
**Describe the issue**
Checkov failing to perform a check against a Terraform Plan when skipping a specific AWS check
**Examples**
```
resource "aws_iam_role" "backend" {
#checkov:skip=CKV_AWS_274:TODO Generate a policy from CloudTrail later
name = "${var.repo}-foo"
assume_role_policy = data.aws_iam_policy_document.backend-assume.json
managed_policy_arns = ["arn:aws:iam::aws:policy/AdministratorAccess"]
inline_policy {
name = "foo"
policy = data.aws_iam_policy_document.backend-permissions.json
}
tags = {
component = var.foo
}
}
```
**Exception Trace**
```
Traceback (most recent call last):
File "/usr/local/bin/checkov", line 9, in <module>
sys.exit(run())
File "/usr/local/lib/python3.9/site-packages/checkov/main.py", line 355, in run
scan_reports = runner_registry.run(external_checks_dir=external_checks_dir, files=config.file,
File "/usr/local/lib/python3.9/site-packages/checkov/common/runners/runner_registry.py", line 79, in run
self.runners[0].run(root_folder, external_checks_dir=external_checks_dir, files=files,
File "/usr/local/lib/python3.9/site-packages/checkov/terraform/plan_runner.py", line 81, in run
self.check_tf_definition(report, root_folder, runner_filter)
File "/usr/local/lib/python3.9/site-packages/checkov/terraform/plan_runner.py", line 99, in check_tf_definition
self.run_block(definition[block_type], None, full_file_path, root_folder, report, scanned_file,
File "/usr/local/lib/python3.9/site-packages/checkov/terraform/plan_runner.py", line 119, in run_block
results = registry.scan(scanned_file, entity, [], runner_filter, report_type=CheckType.TERRAFORM_PLAN)
File "/usr/local/lib/python3.9/site-packages/checkov/common/checks/base_check_registry.py", line 127, in scan
result = self.run_check(check, entity_configuration, entity_name, entity_type, scanned_file, skip_info)
File "/usr/local/lib/python3.9/site-packages/checkov/common/checks/base_check_registry.py", line 141, in run_check
result = check.run(
File "/usr/local/lib/python3.9/site-packages/checkov/common/checks/base_check.py", line 70, in run
check_result["result"] = self.scan_entity_conf(entity_configuration, entity_type)
File "/usr/local/lib/python3.9/site-packages/checkov/terraform/checks/resource/base_resource_check.py", line 43, in scan_entity_conf
return self.scan_resource_conf(conf)
File "/usr/local/lib/python3.9/site-packages/checkov/terraform/checks/resource/aws/IAMManagedAdminPolicy.py", line 42, in scan_resource_conf
if conf.get("policy_arn")[0] == ADMIN_POLICY_ARN:
TypeError: 'NoneType' object is not subscriptable
```
**Desktop (please complete the following information):**
- MacOS 11.7
- Checkov Version 2.2.0
**Additional context**
This fails as of whateverv version CKV_AWS_274 was added. Last time a build didn't crash I was using 2.1.294 and it worked.
Also if I skip it with a command-line switch then this crash does not happen (which is going to be my temp workaround)
</issue>
<code>
[start of checkov/terraform/checks/resource/aws/IAMManagedAdminPolicy.py]
1 from checkov.common.models.enums import CheckResult, CheckCategories
2 from checkov.terraform.checks.resource.base_resource_check import BaseResourceCheck
3
4
5 ADMIN_POLICY_NAME = "AdministratorAccess"
6 ADMIN_POLICY_ARN = f"arn:aws:iam::aws:policy/{ADMIN_POLICY_NAME}"
7
8
9 class IAMManagedAdminPolicy(BaseResourceCheck):
10 def __init__(self):
11 # This is the full description of your check
12 description = "Disallow IAM roles, users, and groups from using the AWS AdministratorAccess policy"
13
14 # This is the Unique ID for your check
15 id = "CKV_AWS_274"
16
17 # These are the terraform objects supported by this check (ex: aws_iam_policy_document)
18 supported_resources = (
19 "aws_iam_role",
20 "aws_iam_policy_attachment",
21 "aws_iam_role_policy_attachment",
22 "aws_iam_user_policy_attachment",
23 "aws_iam_group_policy_attachment",
24 )
25
26 # Valid CheckCategories are defined in checkov/common/models/enums.py
27 categories = (CheckCategories.IAM,)
28 super().__init__(name=description, id=id, categories=categories, supported_resources=supported_resources)
29
30 def scan_resource_conf(self, conf):
31 if self.entity_type == "aws_iam_role":
32 if "managed_policy_arns" in conf.keys():
33 if ADMIN_POLICY_ARN in conf.get("managed_policy_arns")[0]:
34 return CheckResult.FAILED
35
36 elif self.entity_type in (
37 "aws_iam_policy_attachment",
38 "aws_iam_role_policy_attachment",
39 "aws_iam_user_policy_attachment",
40 "aws_iam_group_policy_attachment",
41 ):
42 if conf.get("policy_arn")[0] == ADMIN_POLICY_ARN:
43 return CheckResult.FAILED
44
45 return CheckResult.PASSED
46
47
48 check = IAMManagedAdminPolicy()
[end of checkov/terraform/checks/resource/aws/IAMManagedAdminPolicy.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/checkov/terraform/checks/resource/aws/IAMManagedAdminPolicy.py b/checkov/terraform/checks/resource/aws/IAMManagedAdminPolicy.py
--- a/checkov/terraform/checks/resource/aws/IAMManagedAdminPolicy.py
+++ b/checkov/terraform/checks/resource/aws/IAMManagedAdminPolicy.py
@@ -1,3 +1,7 @@
+from __future__ import annotations
+
+from typing import Any
+
from checkov.common.models.enums import CheckResult, CheckCategories
from checkov.terraform.checks.resource.base_resource_check import BaseResourceCheck
@@ -7,7 +11,7 @@
class IAMManagedAdminPolicy(BaseResourceCheck):
- def __init__(self):
+ def __init__(self) -> None:
# This is the full description of your check
description = "Disallow IAM roles, users, and groups from using the AWS AdministratorAccess policy"
@@ -27,10 +31,10 @@
categories = (CheckCategories.IAM,)
super().__init__(name=description, id=id, categories=categories, supported_resources=supported_resources)
- def scan_resource_conf(self, conf):
+ def scan_resource_conf(self, conf: dict[str, list[Any]]) -> CheckResult:
if self.entity_type == "aws_iam_role":
if "managed_policy_arns" in conf.keys():
- if ADMIN_POLICY_ARN in conf.get("managed_policy_arns")[0]:
+ if ADMIN_POLICY_ARN in conf["managed_policy_arns"][0]:
return CheckResult.FAILED
elif self.entity_type in (
@@ -39,10 +43,11 @@
"aws_iam_user_policy_attachment",
"aws_iam_group_policy_attachment",
):
- if conf.get("policy_arn")[0] == ADMIN_POLICY_ARN:
+ policy_arn = conf.get("policy_arn")
+ if policy_arn and policy_arn[0] == ADMIN_POLICY_ARN:
return CheckResult.FAILED
return CheckResult.PASSED
-check = IAMManagedAdminPolicy()
\ No newline at end of file
+check = IAMManagedAdminPolicy()
|
{"golden_diff": "diff --git a/checkov/terraform/checks/resource/aws/IAMManagedAdminPolicy.py b/checkov/terraform/checks/resource/aws/IAMManagedAdminPolicy.py\n--- a/checkov/terraform/checks/resource/aws/IAMManagedAdminPolicy.py\n+++ b/checkov/terraform/checks/resource/aws/IAMManagedAdminPolicy.py\n@@ -1,3 +1,7 @@\n+from __future__ import annotations\n+\n+from typing import Any\n+\n from checkov.common.models.enums import CheckResult, CheckCategories\n from checkov.terraform.checks.resource.base_resource_check import BaseResourceCheck\n \n@@ -7,7 +11,7 @@\n \n \n class IAMManagedAdminPolicy(BaseResourceCheck):\n- def __init__(self):\n+ def __init__(self) -> None:\n # This is the full description of your check\n description = \"Disallow IAM roles, users, and groups from using the AWS AdministratorAccess policy\"\n \n@@ -27,10 +31,10 @@\n categories = (CheckCategories.IAM,)\n super().__init__(name=description, id=id, categories=categories, supported_resources=supported_resources)\n \n- def scan_resource_conf(self, conf):\n+ def scan_resource_conf(self, conf: dict[str, list[Any]]) -> CheckResult:\n if self.entity_type == \"aws_iam_role\":\n if \"managed_policy_arns\" in conf.keys():\n- if ADMIN_POLICY_ARN in conf.get(\"managed_policy_arns\")[0]:\n+ if ADMIN_POLICY_ARN in conf[\"managed_policy_arns\"][0]:\n return CheckResult.FAILED\n \n elif self.entity_type in (\n@@ -39,10 +43,11 @@\n \"aws_iam_user_policy_attachment\",\n \"aws_iam_group_policy_attachment\",\n ):\n- if conf.get(\"policy_arn\")[0] == ADMIN_POLICY_ARN:\n+ policy_arn = conf.get(\"policy_arn\")\n+ if policy_arn and policy_arn[0] == ADMIN_POLICY_ARN:\n return CheckResult.FAILED\n \n return CheckResult.PASSED\n \n \n-check = IAMManagedAdminPolicy()\n\\ No newline at end of file\n+check = IAMManagedAdminPolicy()\n", "issue": "Checkov failing to perform a check against a Terraform Plan when skipping a specific AWS check\n**Describe the issue**\r\nCheckov failing to perform a check against a Terraform Plan when skipping a specific AWS check\r\n\r\n**Examples**\r\n```\r\nresource \"aws_iam_role\" \"backend\" {\r\n #checkov:skip=CKV_AWS_274:TODO Generate a policy from CloudTrail later\r\n name = \"${var.repo}-foo\"\r\n assume_role_policy = data.aws_iam_policy_document.backend-assume.json\r\n managed_policy_arns = [\"arn:aws:iam::aws:policy/AdministratorAccess\"]\r\n inline_policy {\r\n name = \"foo\"\r\n policy = data.aws_iam_policy_document.backend-permissions.json\r\n }\r\n tags = {\r\n component = var.foo\r\n }\r\n}\r\n```\r\n\r\n**Exception Trace**\r\n```\r\nTraceback (most recent call last):\r\n File \"/usr/local/bin/checkov\", line 9, in <module>\r\n sys.exit(run())\r\n File \"/usr/local/lib/python3.9/site-packages/checkov/main.py\", line 355, in run\r\n scan_reports = runner_registry.run(external_checks_dir=external_checks_dir, files=config.file,\r\n File \"/usr/local/lib/python3.9/site-packages/checkov/common/runners/runner_registry.py\", line 79, in run\r\n self.runners[0].run(root_folder, external_checks_dir=external_checks_dir, files=files,\r\n File \"/usr/local/lib/python3.9/site-packages/checkov/terraform/plan_runner.py\", line 81, in run\r\n self.check_tf_definition(report, root_folder, runner_filter)\r\n File \"/usr/local/lib/python3.9/site-packages/checkov/terraform/plan_runner.py\", line 99, in check_tf_definition\r\n self.run_block(definition[block_type], None, full_file_path, root_folder, report, scanned_file,\r\n File \"/usr/local/lib/python3.9/site-packages/checkov/terraform/plan_runner.py\", line 119, in run_block\r\n results = registry.scan(scanned_file, entity, [], runner_filter, report_type=CheckType.TERRAFORM_PLAN)\r\n File \"/usr/local/lib/python3.9/site-packages/checkov/common/checks/base_check_registry.py\", line 127, in scan\r\n result = self.run_check(check, entity_configuration, entity_name, entity_type, scanned_file, skip_info)\r\n File \"/usr/local/lib/python3.9/site-packages/checkov/common/checks/base_check_registry.py\", line 141, in run_check\r\n result = check.run(\r\n File \"/usr/local/lib/python3.9/site-packages/checkov/common/checks/base_check.py\", line 70, in run\r\n check_result[\"result\"] = self.scan_entity_conf(entity_configuration, entity_type)\r\n File \"/usr/local/lib/python3.9/site-packages/checkov/terraform/checks/resource/base_resource_check.py\", line 43, in scan_entity_conf\r\n return self.scan_resource_conf(conf)\r\n File \"/usr/local/lib/python3.9/site-packages/checkov/terraform/checks/resource/aws/IAMManagedAdminPolicy.py\", line 42, in scan_resource_conf\r\n if conf.get(\"policy_arn\")[0] == ADMIN_POLICY_ARN:\r\nTypeError: 'NoneType' object is not subscriptable\r\n```\r\n\r\n**Desktop (please complete the following information):**\r\n - MacOS 11.7\r\n - Checkov Version 2.2.0\r\n\r\n**Additional context**\r\nThis fails as of whateverv version CKV_AWS_274 was added. Last time a build didn't crash I was using 2.1.294 and it worked.\r\nAlso if I skip it with a command-line switch then this crash does not happen (which is going to be my temp workaround)\r\n\n", "before_files": [{"content": "from checkov.common.models.enums import CheckResult, CheckCategories\nfrom checkov.terraform.checks.resource.base_resource_check import BaseResourceCheck\n\n\nADMIN_POLICY_NAME = \"AdministratorAccess\"\nADMIN_POLICY_ARN = f\"arn:aws:iam::aws:policy/{ADMIN_POLICY_NAME}\"\n\n\nclass IAMManagedAdminPolicy(BaseResourceCheck):\n def __init__(self):\n # This is the full description of your check\n description = \"Disallow IAM roles, users, and groups from using the AWS AdministratorAccess policy\"\n\n # This is the Unique ID for your check\n id = \"CKV_AWS_274\"\n\n # These are the terraform objects supported by this check (ex: aws_iam_policy_document)\n supported_resources = (\n \"aws_iam_role\",\n \"aws_iam_policy_attachment\",\n \"aws_iam_role_policy_attachment\",\n \"aws_iam_user_policy_attachment\",\n \"aws_iam_group_policy_attachment\",\n )\n\n # Valid CheckCategories are defined in checkov/common/models/enums.py\n categories = (CheckCategories.IAM,)\n super().__init__(name=description, id=id, categories=categories, supported_resources=supported_resources)\n\n def scan_resource_conf(self, conf):\n if self.entity_type == \"aws_iam_role\":\n if \"managed_policy_arns\" in conf.keys():\n if ADMIN_POLICY_ARN in conf.get(\"managed_policy_arns\")[0]:\n return CheckResult.FAILED\n\n elif self.entity_type in (\n \"aws_iam_policy_attachment\",\n \"aws_iam_role_policy_attachment\",\n \"aws_iam_user_policy_attachment\",\n \"aws_iam_group_policy_attachment\",\n ):\n if conf.get(\"policy_arn\")[0] == ADMIN_POLICY_ARN:\n return CheckResult.FAILED\n\n return CheckResult.PASSED\n\n\ncheck = IAMManagedAdminPolicy()", "path": "checkov/terraform/checks/resource/aws/IAMManagedAdminPolicy.py"}]}
| 1,851 | 472 |
gh_patches_debug_18921
|
rasdani/github-patches
|
git_diff
|
beeware__toga-756
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Added textInput placeholder in winforms
Added placeholder implementation to textInput in winforms. The code could be extracted to winforms.libs so that it could be reused on multilineTextInput, but I couldn't get it to work.
Also changed winforms specific handler name to winforms_ prefixed name.
<!--- Describe your changes in detail -->
<!--- What problem does this change solve? -->
<!--- If this PR relates to an issue, include Refs #XXX or Fixes #XXX -->
## PR Checklist:
<!--- Go over all the following points, and put an `x` in all the boxes that apply. -->
- [x] All new features have been tested
- [ ] All new features have been documented
- [x] I have read the **CONTRIBUTING.md** file
- [x] I will abide by the code of conduct
</issue>
<code>
[start of docs/conf.py]
1 # -*- coding: utf-8 -*-
2 #
3 # Toga documentation build configuration file, created by
4 # sphinx-quickstart on Sat Jul 27 14:58:42 2013.
5 #
6 # This file is execfile()d with the current directory set to its containing dir.
7 #
8 # Note that not all possible configuration values are present in this
9 # autogenerated file.
10 #
11 # All configuration values have a default; values that are commented out
12 # serve to show the default.
13
14 # If extensions (or modules to document with autodoc) are in another directory,
15 # add these directories to sys.path here. If the directory is relative to the
16 # documentation root, use os.path.abspath to make it absolute, like shown here.
17 import sys
18 import os
19 sys.path.insert(0, os.path.abspath('../src/core/'))
20
21 # -- General configuration -----------------------------------------------------
22
23 # If your documentation needs a minimal Sphinx version, state it here.
24 #needs_sphinx = '1.0'
25
26 # Add any Sphinx extension module names here, as strings. They can be extensions
27 # coming with Sphinx (named 'sphinx.ext.*') or your custom ones.
28 extensions = ['sphinx.ext.autodoc', 'sphinx.ext.napoleon', 'sphinx.ext.todo', 'sphinx_tabs.tabs']
29
30 # Add any paths that contain templates here, relative to this directory.
31 templates_path = ['_templates']
32
33 # The suffix of source filenames.
34 source_suffix = '.rst'
35
36 # The encoding of source files.
37 #source_encoding = 'utf-8-sig'
38
39 # The master toctree document.
40 master_doc = 'index'
41
42 # General information about the project.
43 project = u'Toga'
44 copyright = u'2013, Russell Keith-Magee'
45
46 # The version info for the project you're documenting, acts as replacement for
47 # |version| and |release|, also used in various other places throughout the
48 # built documents.
49 #
50 # The full version, including alpha/beta/rc tags.
51 import io, re
52 with io.open('../src/core/toga/__init__.py', encoding='utf8') as version_file:
53 version_match = re.search(r"^__version__ = ['\"]([^'\"]*)['\"]", version_file.read(), re.M)
54 if version_match:
55 release = version_match.group(1)
56 else:
57 raise RuntimeError("Unable to find version string.")
58
59 # The short X.Y version.
60 version = '.'.join(release.split('.')[:2])
61
62 # Fix the autodoc import issues
63 os.environ['TOGA_PLATFORM'] = 'dummy'
64
65 autoclass_content = 'both'
66
67 # The language for content autogenerated by Sphinx. Refer to documentation
68 # for a list of supported languages.
69 #language = None
70
71 # There are two options for replacing |today|: either, you set today to some
72 # non-false value, then it is used:
73 #today = ''
74 # Else, today_fmt is used as the format for a strftime call.
75 #today_fmt = '%B %d, %Y'
76
77 # List of patterns, relative to source directory, that match files and
78 # directories to ignore when looking for source files.
79 exclude_patterns = ['_build']
80
81 # The reST default role (used for this markup: `text`) to use for all documents.
82 #default_role = None
83
84 # If true, '()' will be appended to :func: etc. cross-reference text.
85 #add_function_parentheses = True
86
87 # If true, the current module name will be prepended to all description
88 # unit titles (such as .. function::).
89 #add_module_names = True
90
91 # If true, sectionauthor and moduleauthor directives will be shown in the
92 # output. They are ignored by default.
93 #show_authors = False
94
95 # The name of the Pygments (syntax highlighting) style to use.
96 pygments_style = 'sphinx'
97
98 # A list of ignored prefixes for module index sorting.
99 #modindex_common_prefix = []
100
101
102 # -- Options for HTML output ---------------------------------------------------
103
104 # on_rtd: whether we are on readthedocs.org
105 on_rtd = os.environ.get('READTHEDOCS', None) == 'True'
106
107 if not on_rtd: # only import and set the theme if we're building docs locally
108 try:
109 import sphinx_rtd_theme
110 except ModuleNotFoundError:
111 html_theme = 'default'
112 else:
113 html_theme = 'sphinx_rtd_theme'
114 html_theme_path = [sphinx_rtd_theme.get_html_theme_path()]
115
116 # Theme options are theme-specific and customize the look and feel of a theme
117 # further. For a list of options available for each theme, see the
118 # documentation.
119 #html_theme_options = {}
120
121 # Add any paths that contain custom themes here, relative to this directory.
122 #html_theme_path = []
123
124 # The name for this set of Sphinx documents. If None, it defaults to
125 # "<project> v<release> documentation".
126 #html_title = None
127
128 # A shorter title for the navigation bar. Default is the same as html_title.
129 #html_short_title = None
130
131 # The name of an image file (relative to this directory) to place at the top
132 # of the sidebar.
133 html_logo = "images/toga.png"
134
135 # The name of an image file (within the static path) to use as favicon of the
136 # docs. This file should be a Windows icon file (.ico) being 16x16 or 32x32
137 # pixels large.
138 #html_favicon = None
139
140 # Add any paths that contain custom static files (such as style sheets) here,
141 # relative to this directory. They are copied after the builtin static files,
142 # so a file named "default.css" will overwrite the builtin "default.css".
143 html_static_path = ['_static']
144
145 # If not '', a 'Last updated on:' timestamp is inserted at every page bottom,
146 # using the given strftime format.
147 #html_last_updated_fmt = '%b %d, %Y'
148
149 # If true, SmartyPants will be used to convert quotes and dashes to
150 # typographically correct entities.
151 #html_use_smartypants = True
152
153 # Custom sidebar templates, maps document names to template names.
154 #html_sidebars = {}
155
156 # Additional templates that should be rendered to pages, maps page names to
157 # template names.
158 #html_additional_pages = {}
159
160 # If false, no module index is generated.
161 #html_domain_indices = True
162
163 # If false, no index is generated.
164 #html_use_index = True
165
166 # If true, the index is split into individual pages for each letter.
167 #html_split_index = False
168
169 # If true, links to the reST sources are added to the pages.
170 #html_show_sourcelink = True
171
172 # If true, "Created using Sphinx" is shown in the HTML footer. Default is True.
173 #html_show_sphinx = True
174
175 # If true, "(C) Copyright ..." is shown in the HTML footer. Default is True.
176 #html_show_copyright = True
177
178 # If true, an OpenSearch description file will be output, and all pages will
179 # contain a <link> tag referring to it. The value of this option must be the
180 # base URL from which the finished HTML is served.
181 #html_use_opensearch = ''
182
183 # This is the file name suffix for HTML files (e.g. ".xhtml").
184 #html_file_suffix = None
185
186 # Output file base name for HTML help builder.
187 htmlhelp_basename = 'togadoc'
188
189 try:
190 import sphinx_rtd_theme
191 html_theme = "sphinx_rtd_theme"
192 html_theme_path = [sphinx_rtd_theme.get_html_theme_path()]
193 except ImportError:
194 # The sphinx-rtd-theme package is not installed, so to the default
195 pass
196
197 # -- Options for LaTeX output --------------------------------------------------
198
199 latex_elements = {
200 # The paper size ('letterpaper' or 'a4paper').
201 #'papersize': 'letterpaper',
202
203 # The font size ('10pt', '11pt' or '12pt').
204 #'pointsize': '10pt',
205
206 # Additional stuff for the LaTeX preamble.
207 #'preamble': '',
208 }
209
210 # Grouping the document tree into LaTeX files. List of tuples
211 # (source start file, target name, title, author, documentclass [howto/manual]).
212 latex_documents = [
213 ('index', 'toga.tex', u'Toga Documentation',
214 u'Russell Keith-Magee', 'manual'),
215 ]
216
217 # The name of an image file (relative to this directory) to place at the top of
218 # the title page.
219 #latex_logo = None
220
221 # For "manual" documents, if this is true, then toplevel headings are parts,
222 # not chapters.
223 #latex_use_parts = False
224
225 # If true, show page references after internal links.
226 #latex_show_pagerefs = False
227
228 # If true, show URL addresses after external links.
229 #latex_show_urls = False
230
231 # Documents to append as an appendix to all manuals.
232 #latex_appendices = []
233
234 # If false, no module index is generated.
235 #latex_domain_indices = True
236
237
238 # -- Options for manual page output --------------------------------------------
239
240 # One entry per manual page. List of tuples
241 # (source start file, name, description, authors, manual section).
242 man_pages = [
243 ('index', 'toga', u'Toga Documentation',
244 [u'Russell Keith-Magee'], 1)
245 ]
246
247 # If true, show URL addresses after external links.
248 #man_show_urls = False
249
250
251 # -- Options for Texinfo output ------------------------------------------------
252
253 # Grouping the document tree into Texinfo files. List of tuples
254 # (source start file, target name, title, author,
255 # dir menu entry, description, category)
256 texinfo_documents = [
257 ('index', 'toga', u'Toga Documentation',
258 u'Russell Keith-Magee', 'Toga', 'A Python native, OS native GUI toolkit.',
259 'Miscellaneous'),
260 ]
261
262 # Documents to append as an appendix to all manuals.
263 #texinfo_appendices = []
264
265 # If false, no module index is generated.
266 #texinfo_domain_indices = True
267
268 # How to display URL addresses: 'footnote', 'no', or 'inline'.
269 #texinfo_show_urls = 'footnote'
270
271 # -- Options for spelling -------------------------------------------
272
273 # Spelling check needs an additional module that is not installed by default.
274 # Add it only if spelling check is requested so docs can be generated without it.
275 if 'spelling' in sys.argv:
276 extensions.append("sphinxcontrib.spelling")
277
278 # Spelling language.
279 spelling_lang = 'en_US'
280
281 # Location of word list.
282 spelling_word_list_filename = 'spelling_wordlist'
283
284 spelling_ignore_pypi_package_names = True
285
286 # -- Options for Todos -------------------------------------------
287
288 # If this is True, todo and todolist produce output, else they produce nothing. The default is False.
289 todo_include_todos = True
290
291 # If this is True, todo emits a warning for each TODO entries. The default is False.
292 # todo_emit_warnings = False
293
294 # If this is True, todolist produce output without file path and line, The default is False.
295 # todo_link_only = False
296
[end of docs/conf.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/docs/conf.py b/docs/conf.py
--- a/docs/conf.py
+++ b/docs/conf.py
@@ -25,8 +25,7 @@
# Add any Sphinx extension module names here, as strings. They can be extensions
# coming with Sphinx (named 'sphinx.ext.*') or your custom ones.
-extensions = ['sphinx.ext.autodoc', 'sphinx.ext.napoleon', 'sphinx.ext.todo', 'sphinx_tabs.tabs']
-
+extensions = ['sphinx.ext.autodoc', 'sphinx.ext.napoleon', 'sphinx.ext.todo', 'sphinx_tabs.tabs', 'crate.sphinx.csv']
# Add any paths that contain templates here, relative to this directory.
templates_path = ['_templates']
@@ -194,6 +193,10 @@
# The sphinx-rtd-theme package is not installed, so to the default
pass
+html_css_files = [
+ 'custom.css',
+]
+
# -- Options for LaTeX output --------------------------------------------------
latex_elements = {
|
{"golden_diff": "diff --git a/docs/conf.py b/docs/conf.py\n--- a/docs/conf.py\n+++ b/docs/conf.py\n@@ -25,8 +25,7 @@\n \n # Add any Sphinx extension module names here, as strings. They can be extensions\n # coming with Sphinx (named 'sphinx.ext.*') or your custom ones.\n-extensions = ['sphinx.ext.autodoc', 'sphinx.ext.napoleon', 'sphinx.ext.todo', 'sphinx_tabs.tabs']\n-\n+extensions = ['sphinx.ext.autodoc', 'sphinx.ext.napoleon', 'sphinx.ext.todo', 'sphinx_tabs.tabs', 'crate.sphinx.csv']\n # Add any paths that contain templates here, relative to this directory.\n templates_path = ['_templates']\n \n@@ -194,6 +193,10 @@\n # The sphinx-rtd-theme package is not installed, so to the default\n pass\n \n+html_css_files = [\n+ 'custom.css',\n+]\n+\n # -- Options for LaTeX output --------------------------------------------------\n \n latex_elements = {\n", "issue": "Added textInput placeholder in winforms\nAdded placeholder implementation to textInput in winforms. The code could be extracted to winforms.libs so that it could be reused on multilineTextInput, but I couldn't get it to work.\r\n\r\nAlso changed winforms specific handler name to winforms_ prefixed name.\r\n\r\n<!--- Describe your changes in detail -->\r\n<!--- What problem does this change solve? -->\r\n<!--- If this PR relates to an issue, include Refs #XXX or Fixes #XXX -->\r\n\r\n## PR Checklist:\r\n<!--- Go over all the following points, and put an `x` in all the boxes that apply. -->\r\n- [x] All new features have been tested\r\n- [ ] All new features have been documented\r\n- [x] I have read the **CONTRIBUTING.md** file\r\n- [x] I will abide by the code of conduct\r\n\n", "before_files": [{"content": "# -*- coding: utf-8 -*-\n#\n# Toga documentation build configuration file, created by\n# sphinx-quickstart on Sat Jul 27 14:58:42 2013.\n#\n# This file is execfile()d with the current directory set to its containing dir.\n#\n# Note that not all possible configuration values are present in this\n# autogenerated file.\n#\n# All configuration values have a default; values that are commented out\n# serve to show the default.\n\n# If extensions (or modules to document with autodoc) are in another directory,\n# add these directories to sys.path here. If the directory is relative to the\n# documentation root, use os.path.abspath to make it absolute, like shown here.\nimport sys\nimport os\nsys.path.insert(0, os.path.abspath('../src/core/'))\n\n# -- General configuration -----------------------------------------------------\n\n# If your documentation needs a minimal Sphinx version, state it here.\n#needs_sphinx = '1.0'\n\n# Add any Sphinx extension module names here, as strings. They can be extensions\n# coming with Sphinx (named 'sphinx.ext.*') or your custom ones.\nextensions = ['sphinx.ext.autodoc', 'sphinx.ext.napoleon', 'sphinx.ext.todo', 'sphinx_tabs.tabs']\n\n# Add any paths that contain templates here, relative to this directory.\ntemplates_path = ['_templates']\n\n# The suffix of source filenames.\nsource_suffix = '.rst'\n\n# The encoding of source files.\n#source_encoding = 'utf-8-sig'\n\n# The master toctree document.\nmaster_doc = 'index'\n\n# General information about the project.\nproject = u'Toga'\ncopyright = u'2013, Russell Keith-Magee'\n\n# The version info for the project you're documenting, acts as replacement for\n# |version| and |release|, also used in various other places throughout the\n# built documents.\n#\n# The full version, including alpha/beta/rc tags.\nimport io, re\nwith io.open('../src/core/toga/__init__.py', encoding='utf8') as version_file:\n version_match = re.search(r\"^__version__ = ['\\\"]([^'\\\"]*)['\\\"]\", version_file.read(), re.M)\n if version_match:\n release = version_match.group(1)\n else:\n raise RuntimeError(\"Unable to find version string.\")\n\n# The short X.Y version.\nversion = '.'.join(release.split('.')[:2])\n\n# Fix the autodoc import issues\nos.environ['TOGA_PLATFORM'] = 'dummy'\n\nautoclass_content = 'both'\n\n# The language for content autogenerated by Sphinx. Refer to documentation\n# for a list of supported languages.\n#language = None\n\n# There are two options for replacing |today|: either, you set today to some\n# non-false value, then it is used:\n#today = ''\n# Else, today_fmt is used as the format for a strftime call.\n#today_fmt = '%B %d, %Y'\n\n# List of patterns, relative to source directory, that match files and\n# directories to ignore when looking for source files.\nexclude_patterns = ['_build']\n\n# The reST default role (used for this markup: `text`) to use for all documents.\n#default_role = None\n\n# If true, '()' will be appended to :func: etc. cross-reference text.\n#add_function_parentheses = True\n\n# If true, the current module name will be prepended to all description\n# unit titles (such as .. function::).\n#add_module_names = True\n\n# If true, sectionauthor and moduleauthor directives will be shown in the\n# output. They are ignored by default.\n#show_authors = False\n\n# The name of the Pygments (syntax highlighting) style to use.\npygments_style = 'sphinx'\n\n# A list of ignored prefixes for module index sorting.\n#modindex_common_prefix = []\n\n\n# -- Options for HTML output ---------------------------------------------------\n\n# on_rtd: whether we are on readthedocs.org\non_rtd = os.environ.get('READTHEDOCS', None) == 'True'\n\nif not on_rtd: # only import and set the theme if we're building docs locally\n try:\n import sphinx_rtd_theme\n except ModuleNotFoundError:\n html_theme = 'default'\n else:\n html_theme = 'sphinx_rtd_theme'\n html_theme_path = [sphinx_rtd_theme.get_html_theme_path()]\n\n# Theme options are theme-specific and customize the look and feel of a theme\n# further. For a list of options available for each theme, see the\n# documentation.\n#html_theme_options = {}\n\n# Add any paths that contain custom themes here, relative to this directory.\n#html_theme_path = []\n\n# The name for this set of Sphinx documents. If None, it defaults to\n# \"<project> v<release> documentation\".\n#html_title = None\n\n# A shorter title for the navigation bar. Default is the same as html_title.\n#html_short_title = None\n\n# The name of an image file (relative to this directory) to place at the top\n# of the sidebar.\nhtml_logo = \"images/toga.png\"\n\n# The name of an image file (within the static path) to use as favicon of the\n# docs. This file should be a Windows icon file (.ico) being 16x16 or 32x32\n# pixels large.\n#html_favicon = None\n\n# Add any paths that contain custom static files (such as style sheets) here,\n# relative to this directory. They are copied after the builtin static files,\n# so a file named \"default.css\" will overwrite the builtin \"default.css\".\nhtml_static_path = ['_static']\n\n# If not '', a 'Last updated on:' timestamp is inserted at every page bottom,\n# using the given strftime format.\n#html_last_updated_fmt = '%b %d, %Y'\n\n# If true, SmartyPants will be used to convert quotes and dashes to\n# typographically correct entities.\n#html_use_smartypants = True\n\n# Custom sidebar templates, maps document names to template names.\n#html_sidebars = {}\n\n# Additional templates that should be rendered to pages, maps page names to\n# template names.\n#html_additional_pages = {}\n\n# If false, no module index is generated.\n#html_domain_indices = True\n\n# If false, no index is generated.\n#html_use_index = True\n\n# If true, the index is split into individual pages for each letter.\n#html_split_index = False\n\n# If true, links to the reST sources are added to the pages.\n#html_show_sourcelink = True\n\n# If true, \"Created using Sphinx\" is shown in the HTML footer. Default is True.\n#html_show_sphinx = True\n\n# If true, \"(C) Copyright ...\" is shown in the HTML footer. Default is True.\n#html_show_copyright = True\n\n# If true, an OpenSearch description file will be output, and all pages will\n# contain a <link> tag referring to it. The value of this option must be the\n# base URL from which the finished HTML is served.\n#html_use_opensearch = ''\n\n# This is the file name suffix for HTML files (e.g. \".xhtml\").\n#html_file_suffix = None\n\n# Output file base name for HTML help builder.\nhtmlhelp_basename = 'togadoc'\n\ntry:\n import sphinx_rtd_theme\n html_theme = \"sphinx_rtd_theme\"\n html_theme_path = [sphinx_rtd_theme.get_html_theme_path()]\nexcept ImportError:\n # The sphinx-rtd-theme package is not installed, so to the default\n pass\n\n# -- Options for LaTeX output --------------------------------------------------\n\nlatex_elements = {\n# The paper size ('letterpaper' or 'a4paper').\n#'papersize': 'letterpaper',\n\n# The font size ('10pt', '11pt' or '12pt').\n#'pointsize': '10pt',\n\n# Additional stuff for the LaTeX preamble.\n#'preamble': '',\n}\n\n# Grouping the document tree into LaTeX files. List of tuples\n# (source start file, target name, title, author, documentclass [howto/manual]).\nlatex_documents = [\n ('index', 'toga.tex', u'Toga Documentation',\n u'Russell Keith-Magee', 'manual'),\n]\n\n# The name of an image file (relative to this directory) to place at the top of\n# the title page.\n#latex_logo = None\n\n# For \"manual\" documents, if this is true, then toplevel headings are parts,\n# not chapters.\n#latex_use_parts = False\n\n# If true, show page references after internal links.\n#latex_show_pagerefs = False\n\n# If true, show URL addresses after external links.\n#latex_show_urls = False\n\n# Documents to append as an appendix to all manuals.\n#latex_appendices = []\n\n# If false, no module index is generated.\n#latex_domain_indices = True\n\n\n# -- Options for manual page output --------------------------------------------\n\n# One entry per manual page. List of tuples\n# (source start file, name, description, authors, manual section).\nman_pages = [\n ('index', 'toga', u'Toga Documentation',\n [u'Russell Keith-Magee'], 1)\n]\n\n# If true, show URL addresses after external links.\n#man_show_urls = False\n\n\n# -- Options for Texinfo output ------------------------------------------------\n\n# Grouping the document tree into Texinfo files. List of tuples\n# (source start file, target name, title, author,\n# dir menu entry, description, category)\ntexinfo_documents = [\n ('index', 'toga', u'Toga Documentation',\n u'Russell Keith-Magee', 'Toga', 'A Python native, OS native GUI toolkit.',\n 'Miscellaneous'),\n]\n\n# Documents to append as an appendix to all manuals.\n#texinfo_appendices = []\n\n# If false, no module index is generated.\n#texinfo_domain_indices = True\n\n# How to display URL addresses: 'footnote', 'no', or 'inline'.\n#texinfo_show_urls = 'footnote'\n\n# -- Options for spelling -------------------------------------------\n\n# Spelling check needs an additional module that is not installed by default.\n# Add it only if spelling check is requested so docs can be generated without it.\nif 'spelling' in sys.argv:\n extensions.append(\"sphinxcontrib.spelling\")\n\n# Spelling language.\nspelling_lang = 'en_US'\n\n# Location of word list.\nspelling_word_list_filename = 'spelling_wordlist'\n\nspelling_ignore_pypi_package_names = True\n\n# -- Options for Todos -------------------------------------------\n\n# If this is True, todo and todolist produce output, else they produce nothing. The default is False.\ntodo_include_todos = True\n\n# If this is True, todo emits a warning for each TODO entries. The default is False.\n# todo_emit_warnings = False\n\n# If this is True, todolist produce output without file path and line, The default is False.\n# todo_link_only = False\n", "path": "docs/conf.py"}]}
| 3,901 | 226 |
gh_patches_debug_40754
|
rasdani/github-patches
|
git_diff
|
qtile__qtile-4716
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Clock and tzupdate
### The issue:
qtile version:
0.21.0
These days I need to travel with my system and have the need to update my timezone in order to work with my calendar. I'm using `tzupdate` which easily updates my timezone with one command.
I'm using the Clock widget in the qtile bar as so:
``` python
widget.Clock(format="%A %d %b %Y %H:%M:%S %z"),
```
Updating the timezone with `tzupdate` however does not change the timezone on the Clock widget. It requires restarting qtile in order to get this done. I would expect qtile to poll the current system timezone at each tik.
### Required:
- [X] I have searched past issues to see if this bug has already been reported.
</issue>
<code>
[start of libqtile/widget/clock.py]
1 # Copyright (c) 2010 Aldo Cortesi
2 # Copyright (c) 2012 Andrew Grigorev
3 # Copyright (c) 2014 Sean Vig
4 # Copyright (c) 2014 Tycho Andersen
5 #
6 # Permission is hereby granted, free of charge, to any person obtaining a copy
7 # of this software and associated documentation files (the "Software"), to deal
8 # in the Software without restriction, including without limitation the rights
9 # to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
10 # copies of the Software, and to permit persons to whom the Software is
11 # furnished to do so, subject to the following conditions:
12 #
13 # The above copyright notice and this permission notice shall be included in
14 # all copies or substantial portions of the Software.
15 #
16 # THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
17 # IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
18 # FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
19 # AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
20 # LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
21 # OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
22 # SOFTWARE.
23
24 import sys
25 import time
26 from datetime import datetime, timedelta, timezone
27
28 from libqtile.log_utils import logger
29 from libqtile.widget import base
30
31 try:
32 import pytz
33 except ImportError:
34 pass
35
36 try:
37 import dateutil.tz
38 except ImportError:
39 pass
40
41
42 class Clock(base.InLoopPollText):
43 """A simple but flexible text-based clock"""
44
45 defaults = [
46 ("format", "%H:%M", "A Python datetime format string"),
47 ("update_interval", 1.0, "Update interval for the clock"),
48 (
49 "timezone",
50 None,
51 "The timezone to use for this clock, either as"
52 ' string if pytz or dateutil is installed (e.g. "US/Central" or'
53 " anything in /usr/share/zoneinfo), or as tzinfo (e.g."
54 " datetime.timezone.utc). None means the system local timezone and is"
55 " the default.",
56 ),
57 ]
58 DELTA = timedelta(seconds=0.5)
59
60 def __init__(self, **config):
61 base.InLoopPollText.__init__(self, **config)
62 self.add_defaults(Clock.defaults)
63 if isinstance(self.timezone, str):
64 if "pytz" in sys.modules:
65 self.timezone = pytz.timezone(self.timezone)
66 elif "dateutil" in sys.modules:
67 self.timezone = dateutil.tz.gettz(self.timezone)
68 else:
69 logger.warning(
70 "Clock widget can not infer its timezone from a"
71 " string without pytz or dateutil. Install one"
72 " of these libraries, or give it a"
73 " datetime.tzinfo instance."
74 )
75 if self.timezone is None:
76 logger.debug("Defaulting to the system local timezone.")
77
78 def tick(self):
79 self.update(self.poll())
80 return self.update_interval - time.time() % self.update_interval
81
82 # adding .5 to get a proper seconds value because glib could
83 # theoreticaly call our method too early and we could get something
84 # like (x-1).999 instead of x.000
85 def poll(self):
86 if self.timezone:
87 now = datetime.now(timezone.utc).astimezone(self.timezone)
88 else:
89 now = datetime.now(timezone.utc).astimezone()
90 return (now + self.DELTA).strftime(self.format)
91
[end of libqtile/widget/clock.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/libqtile/widget/clock.py b/libqtile/widget/clock.py
--- a/libqtile/widget/clock.py
+++ b/libqtile/widget/clock.py
@@ -20,11 +20,13 @@
# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
# SOFTWARE.
+from __future__ import annotations
import sys
import time
-from datetime import datetime, timedelta, timezone
+from datetime import datetime, timedelta, timezone, tzinfo
+from libqtile.command.base import expose_command
from libqtile.log_utils import logger
from libqtile.widget import base
@@ -60,11 +62,24 @@
def __init__(self, **config):
base.InLoopPollText.__init__(self, **config)
self.add_defaults(Clock.defaults)
- if isinstance(self.timezone, str):
+ self.timezone = self._lift_timezone(self.timezone)
+
+ if self.timezone is None:
+ logger.debug("Defaulting to the system local timezone.")
+
+ def _lift_timezone(self, timezone):
+ if isinstance(timezone, tzinfo):
+ return timezone
+ elif isinstance(timezone, str):
+ # Empty string can be used to force use of system time
+ if not timezone:
+ return None
+
+ # A string timezone needs to be converted to a tzinfo object
if "pytz" in sys.modules:
- self.timezone = pytz.timezone(self.timezone)
+ return pytz.timezone(timezone)
elif "dateutil" in sys.modules:
- self.timezone = dateutil.tz.gettz(self.timezone)
+ return dateutil.tz.gettz(timezone)
else:
logger.warning(
"Clock widget can not infer its timezone from a"
@@ -72,8 +87,12 @@
" of these libraries, or give it a"
" datetime.tzinfo instance."
)
- if self.timezone is None:
- logger.debug("Defaulting to the system local timezone.")
+ elif timezone is None:
+ pass
+ else:
+ logger.warning("Invalid timezone value %s.", timezone)
+
+ return None
def tick(self):
self.update(self.poll())
@@ -88,3 +107,27 @@
else:
now = datetime.now(timezone.utc).astimezone()
return (now + self.DELTA).strftime(self.format)
+
+ @expose_command
+ def update_timezone(self, timezone: str | tzinfo | None = None):
+ """
+ Force the clock to update timezone information.
+
+ If the method is called with no arguments then the widget will reload
+ the timzeone set on the computer (e.g. via ``timedatectl set-timezone ..``).
+ This will have no effect if you have previously set a ``timezone`` value.
+
+ Alternatively, you can pass a timezone string (e.g. ``"Europe/Lisbon"``) to change
+ the specified timezone. Setting this to an empty string will cause the clock
+ to rely on the system timezone.
+ """
+ self.timezone = self._lift_timezone(timezone)
+
+ # Force python to update timezone info (e.g. if system time has changed)
+ time.tzset()
+ self.update(self.poll())
+
+ @expose_command
+ def use_system_timezone(self):
+ """Force clock to use system timezone."""
+ self.update_timezone("")
|
{"golden_diff": "diff --git a/libqtile/widget/clock.py b/libqtile/widget/clock.py\n--- a/libqtile/widget/clock.py\n+++ b/libqtile/widget/clock.py\n@@ -20,11 +20,13 @@\n # LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,\n # OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE\n # SOFTWARE.\n+from __future__ import annotations\n \n import sys\n import time\n-from datetime import datetime, timedelta, timezone\n+from datetime import datetime, timedelta, timezone, tzinfo\n \n+from libqtile.command.base import expose_command\n from libqtile.log_utils import logger\n from libqtile.widget import base\n \n@@ -60,11 +62,24 @@\n def __init__(self, **config):\n base.InLoopPollText.__init__(self, **config)\n self.add_defaults(Clock.defaults)\n- if isinstance(self.timezone, str):\n+ self.timezone = self._lift_timezone(self.timezone)\n+\n+ if self.timezone is None:\n+ logger.debug(\"Defaulting to the system local timezone.\")\n+\n+ def _lift_timezone(self, timezone):\n+ if isinstance(timezone, tzinfo):\n+ return timezone\n+ elif isinstance(timezone, str):\n+ # Empty string can be used to force use of system time\n+ if not timezone:\n+ return None\n+\n+ # A string timezone needs to be converted to a tzinfo object\n if \"pytz\" in sys.modules:\n- self.timezone = pytz.timezone(self.timezone)\n+ return pytz.timezone(timezone)\n elif \"dateutil\" in sys.modules:\n- self.timezone = dateutil.tz.gettz(self.timezone)\n+ return dateutil.tz.gettz(timezone)\n else:\n logger.warning(\n \"Clock widget can not infer its timezone from a\"\n@@ -72,8 +87,12 @@\n \" of these libraries, or give it a\"\n \" datetime.tzinfo instance.\"\n )\n- if self.timezone is None:\n- logger.debug(\"Defaulting to the system local timezone.\")\n+ elif timezone is None:\n+ pass\n+ else:\n+ logger.warning(\"Invalid timezone value %s.\", timezone)\n+\n+ return None\n \n def tick(self):\n self.update(self.poll())\n@@ -88,3 +107,27 @@\n else:\n now = datetime.now(timezone.utc).astimezone()\n return (now + self.DELTA).strftime(self.format)\n+\n+ @expose_command\n+ def update_timezone(self, timezone: str | tzinfo | None = None):\n+ \"\"\"\n+ Force the clock to update timezone information.\n+\n+ If the method is called with no arguments then the widget will reload\n+ the timzeone set on the computer (e.g. via ``timedatectl set-timezone ..``).\n+ This will have no effect if you have previously set a ``timezone`` value.\n+\n+ Alternatively, you can pass a timezone string (e.g. ``\"Europe/Lisbon\"``) to change\n+ the specified timezone. Setting this to an empty string will cause the clock\n+ to rely on the system timezone.\n+ \"\"\"\n+ self.timezone = self._lift_timezone(timezone)\n+\n+ # Force python to update timezone info (e.g. if system time has changed)\n+ time.tzset()\n+ self.update(self.poll())\n+\n+ @expose_command\n+ def use_system_timezone(self):\n+ \"\"\"Force clock to use system timezone.\"\"\"\n+ self.update_timezone(\"\")\n", "issue": "Clock and tzupdate \n### The issue:\n\nqtile version:\r\n0.21.0\r\n\r\nThese days I need to travel with my system and have the need to update my timezone in order to work with my calendar. I'm using `tzupdate` which easily updates my timezone with one command. \r\n\r\nI'm using the Clock widget in the qtile bar as so:\r\n``` python \r\n widget.Clock(format=\"%A %d %b %Y %H:%M:%S %z\"),\r\n```\r\n\r\nUpdating the timezone with `tzupdate` however does not change the timezone on the Clock widget. It requires restarting qtile in order to get this done. I would expect qtile to poll the current system timezone at each tik.\n\n### Required:\n\n- [X] I have searched past issues to see if this bug has already been reported.\n", "before_files": [{"content": "# Copyright (c) 2010 Aldo Cortesi\n# Copyright (c) 2012 Andrew Grigorev\n# Copyright (c) 2014 Sean Vig\n# Copyright (c) 2014 Tycho Andersen\n#\n# Permission is hereby granted, free of charge, to any person obtaining a copy\n# of this software and associated documentation files (the \"Software\"), to deal\n# in the Software without restriction, including without limitation the rights\n# to use, copy, modify, merge, publish, distribute, sublicense, and/or sell\n# copies of the Software, and to permit persons to whom the Software is\n# furnished to do so, subject to the following conditions:\n#\n# The above copyright notice and this permission notice shall be included in\n# all copies or substantial portions of the Software.\n#\n# THE SOFTWARE IS PROVIDED \"AS IS\", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR\n# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,\n# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE\n# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER\n# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,\n# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE\n# SOFTWARE.\n\nimport sys\nimport time\nfrom datetime import datetime, timedelta, timezone\n\nfrom libqtile.log_utils import logger\nfrom libqtile.widget import base\n\ntry:\n import pytz\nexcept ImportError:\n pass\n\ntry:\n import dateutil.tz\nexcept ImportError:\n pass\n\n\nclass Clock(base.InLoopPollText):\n \"\"\"A simple but flexible text-based clock\"\"\"\n\n defaults = [\n (\"format\", \"%H:%M\", \"A Python datetime format string\"),\n (\"update_interval\", 1.0, \"Update interval for the clock\"),\n (\n \"timezone\",\n None,\n \"The timezone to use for this clock, either as\"\n ' string if pytz or dateutil is installed (e.g. \"US/Central\" or'\n \" anything in /usr/share/zoneinfo), or as tzinfo (e.g.\"\n \" datetime.timezone.utc). None means the system local timezone and is\"\n \" the default.\",\n ),\n ]\n DELTA = timedelta(seconds=0.5)\n\n def __init__(self, **config):\n base.InLoopPollText.__init__(self, **config)\n self.add_defaults(Clock.defaults)\n if isinstance(self.timezone, str):\n if \"pytz\" in sys.modules:\n self.timezone = pytz.timezone(self.timezone)\n elif \"dateutil\" in sys.modules:\n self.timezone = dateutil.tz.gettz(self.timezone)\n else:\n logger.warning(\n \"Clock widget can not infer its timezone from a\"\n \" string without pytz or dateutil. Install one\"\n \" of these libraries, or give it a\"\n \" datetime.tzinfo instance.\"\n )\n if self.timezone is None:\n logger.debug(\"Defaulting to the system local timezone.\")\n\n def tick(self):\n self.update(self.poll())\n return self.update_interval - time.time() % self.update_interval\n\n # adding .5 to get a proper seconds value because glib could\n # theoreticaly call our method too early and we could get something\n # like (x-1).999 instead of x.000\n def poll(self):\n if self.timezone:\n now = datetime.now(timezone.utc).astimezone(self.timezone)\n else:\n now = datetime.now(timezone.utc).astimezone()\n return (now + self.DELTA).strftime(self.format)\n", "path": "libqtile/widget/clock.py"}]}
| 1,668 | 786 |
gh_patches_debug_21242
|
rasdani/github-patches
|
git_diff
|
chainer__chainer-415
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
generateNormal and generateNormalDouble raise errors when size is odd
When `curand.generateNormal` is called with argument `size` odd (as long as I investigated).
For example, the following code fails.
```
In [84]: generator = curand.createGenerator(curand.CURAND_RNG_PSEUDO_DEFAULT)
In [85]: curand.generateNormal(generator, out.data.ptr, 1, 0.0, 1.0)
```
The error message is as follows:
```
CURANDError Traceback (most recent call last)
<ipython-input-85-e3c65b1a19eb> in <module>()
----> 1 curand.generateNormal(generator, out.data.ptr, 1, 0.0, 1.0)
/home/delta/dev/chainer/cupy/cuda/curand.py in generateNormal(generator, outputPtr, n, mean, stddev)
185 status = _curand.curandGenerateNormal(generator, outputPtr, n, mean,
186 stddev)
--> 187 check_status(status)
188
189
/home/delta/dev/chainer/cupy/cuda/curand.py in check_status(status)
77 def check_status(status):
78 if status != 0:
---> 79 raise CURANDError(status)
80
81 ###############################################################################
CURANDError: CURAND_STATUS_LENGTH_NOT_MULTIPLE
```
But when `size` is even, no error is raised.
```
In [86]: curand.generateNormal(generator, out.data.ptr, 2, 0.0, 1.0)
```
The situation is true for `curand.generateNormalDouble`.
</issue>
<code>
[start of cupy/cuda/curand.py]
1 """Thin wrapper of cuRAND."""
2 import ctypes
3 import sys
4
5 from cupy.cuda import internal
6 from cupy.cuda import runtime
7
8 if 'win32' == sys.platform:
9 _curand = internal.load_library(
10 internal.get_windows_cuda_library_names('curand'))
11 else:
12 _curand = internal.load_library('curand')
13
14 _I = ctypes.c_int
15 _U = ctypes.c_uint
16 _S = ctypes.c_size_t
17 _ULL = ctypes.c_ulonglong
18 _P = ctypes.c_void_p
19 _IP = ctypes.POINTER(_I)
20 _UP = ctypes.POINTER(_U)
21 _ULLP = ctypes.POINTER(_ULL)
22 _F = ctypes.c_float
23 _D = ctypes.c_double
24 _FP = ctypes.POINTER(_F)
25 _DP = ctypes.POINTER(_D)
26
27 CURAND_RNG_PSEUDO_DEFAULT = 100
28 CURAND_RNG_PSEUDO_XORWOW = 101
29 CURAND_RNG_PSEUDO_MRG32K3A = 121
30 CURAND_RNG_PSEUDO_MTGP32 = 141
31 CURAND_RNG_PSEUDO_MT19937 = 142
32 CURAND_RNG_PSEUDO_PHILOX4_32_10 = 161
33 CURAND_RNG_QUASI_DEFAULT = 200
34 CURAND_RNG_QUASI_SOBOL32 = 201
35 CURAND_RNG_QUASI_SCRAMBLED_SOBOL32 = 202
36 CURAND_RNG_QUASI_SOBOL64 = 203
37 CURAND_RNG_QUASI_SCRAMBLED_SOBOL64 = 204
38
39 CURAND_ORDERING_PSEUDO_BEST = 100
40 CURAND_ORDERING_PSEUDO_DEFAULT = 101
41 CURAND_ORDERING_PSEUDO_SEEDED = 102
42 CURAND_ORDERING_QUASI_DEFAULT = 201
43
44 Generator = _P
45 Distribution = _DP
46
47 DiscreteDistribution = _P
48
49 ###############################################################################
50 # Error handling
51 ###############################################################################
52
53 STATUS = {
54 0: 'CURAND_STATUS_SUCCESS',
55 100: 'CURAND_STATUS_VERSION_MISMATCH',
56 101: 'CURAND_STATUS_NOT_INITIALIZED',
57 102: 'CURAND_STATUS_ALLOCATION_FAILED',
58 103: 'CURAND_STATUS_TYPE_ERROR',
59 104: 'CURAND_STATUS_OUT_OF_RANGE',
60 105: 'CURAND_STATUS_LENGTH_NOT_MULTIPLE',
61 106: 'CURAND_STATUS_DOUBLE_PRECISION_REQUIRED',
62 201: 'CURAND_STATUS_LAUNCH_FAILURE',
63 202: 'CURAND_STATUS_PREEXISTING_FAILURE',
64 203: 'CURAND_STATUS_INITIALIZATION_FAILED',
65 204: 'CURAND_STATUS_ARCH_MISMATCH',
66 999: 'CURAND_STATUS_INTERNAL_ERROR',
67 }
68
69
70 class CURANDError(RuntimeError):
71
72 def __init__(self, status):
73 self.status = status
74 super(CURANDError, self).__init__(STATUS[status])
75
76
77 def check_status(status):
78 if status != 0:
79 raise CURANDError(status)
80
81 ###############################################################################
82 # Generator
83 ###############################################################################
84
85 _curand.curandCreateGenerator.argtypes = (_P, _I)
86
87
88 def createGenerator(rng_type):
89 generator = Generator()
90 status = _curand.curandCreateGenerator(ctypes.byref(generator), rng_type)
91 check_status(status)
92 return generator
93
94
95 _curand.curandDestroyGenerator.argtypes = (Generator,)
96
97
98 def destroyGenerator(generator):
99 status = _curand.curandDestroyGenerator(generator)
100 check_status(status)
101
102
103 _curand.curandGetVersion.argtypes = (_IP,)
104
105
106 def getVersion():
107 version = _I()
108 status = _curand.curandGetVersion(ctypes.byref(version))
109 check_status(status)
110 return version
111
112
113 _curand.curandSetStream.argtypes = (Generator, runtime.Stream)
114
115
116 def setStream(generator, stream):
117 status = _curand.curandSetStream(generator, stream)
118 check_status(status)
119
120
121 _curand.curandSetPseudoRandomGeneratorSeed.argtypes = (Generator, _ULL)
122
123
124 def setPseudoRandomGeneratorSeed(generator, seed):
125 status = _curand.curandSetPseudoRandomGeneratorSeed(generator, seed)
126 check_status(status)
127
128
129 _curand.curandSetGeneratorOffset.argtypes = (Generator, _ULL)
130
131
132 def setGeneratorOffset(generator, offset):
133 status = _curand.curandSetGeneratorOffset(generator, offset)
134 check_status(status)
135
136
137 _curand.curandSetGeneratorOrdering.argtypes = (Generator, _I)
138
139
140 def setGeneratorOrdering(generator, order):
141 status = _curand.curandSetGeneratorOrdering(generator, order)
142 check_status(status)
143
144
145 ###############################################################################
146 # Generation functions
147 ###############################################################################
148
149 _curand.curandGenerate.argtypes = (Generator, _P, _S)
150
151
152 def generate(generator, outputPtr, num):
153 status = _curand.curandGenerate(generator, outputPtr, num)
154 check_status(status)
155
156
157 _curand.curandGenerateLongLong.argtypes = (Generator, _P, _S)
158
159
160 def generateLongLong(generator, outputPtr, num):
161 status = _curand.curandGenerateLongLong(generator, outputPtr, num)
162 check_status(status)
163
164
165 _curand.curandGenerateUniform.argtypes = (Generator, _P, _S)
166
167
168 def generateUniform(generator, outputPtr, num):
169 status = _curand.curandGenerateUniform(generator, outputPtr, num)
170 check_status(status)
171
172
173 _curand.curandGenerateUniformDouble.argtypes = (Generator, _P, _S)
174
175
176 def generateUniformDouble(generator, outputPtr, num):
177 status = _curand.curandGenerateUniformDouble(generator, outputPtr, num)
178 check_status(status)
179
180
181 _curand.curandGenerateNormal.argtypes = (Generator, _P, _S, _F, _F)
182
183
184 def generateNormal(generator, outputPtr, n, mean, stddev):
185 status = _curand.curandGenerateNormal(generator, outputPtr, n, mean,
186 stddev)
187 check_status(status)
188
189
190 _curand.curandGenerateNormalDouble.argtypes = (Generator, _P, _S, _D, _D)
191
192
193 def generateNormalDouble(generator, outputPtr, n, mean, stddev):
194 status = _curand.curandGenerateNormalDouble(generator, outputPtr, n, mean,
195 stddev)
196 check_status(status)
197
198
199 _curand.curandGenerateLogNormal.argtypes = (Generator, _P, _S, _F, _F)
200
201
202 def generateLogNormal(generator, outputPtr, n, mean, stddev):
203 status = _curand.curandGenerateLogNormal(generator, outputPtr, n,
204 mean, stddev)
205 check_status(status)
206
207
208 _curand.curandGenerateLogNormalDouble.argtypes = (Generator, _P, _S, _D, _D)
209
210
211 def generateLogNormalDouble(generator, outputPtr, n, mean, stddev):
212 status = _curand.curandGenerateLogNormalDouble(generator, outputPtr, n,
213 mean, stddev)
214 check_status(status)
215
216
217 _curand.curandGeneratePoisson.argtypes = (Generator, _P, _S, _D)
218
219
220 def generatePoisson(generator, outputPtr, n, lam):
221 status = _curand.curandGeneratePoisson(generator, outputPtr, n, lam)
222 check_status(status)
223
[end of cupy/cuda/curand.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/cupy/cuda/curand.py b/cupy/cuda/curand.py
--- a/cupy/cuda/curand.py
+++ b/cupy/cuda/curand.py
@@ -182,6 +182,10 @@
def generateNormal(generator, outputPtr, n, mean, stddev):
+ if n % 2 == 1:
+ msg = 'curandGenerateNormal can only generate even number of '\
+ 'random variables simultaneously. See issue #390 for detail.'
+ raise ValueError(msg)
status = _curand.curandGenerateNormal(generator, outputPtr, n, mean,
stddev)
check_status(status)
@@ -191,6 +195,10 @@
def generateNormalDouble(generator, outputPtr, n, mean, stddev):
+ if n % 2 == 1:
+ msg = 'curandGenerateNormalDouble can only generate even number of '\
+ 'random variables simultaneously. See issue #390 for detail.'
+ raise ValueError(msg)
status = _curand.curandGenerateNormalDouble(generator, outputPtr, n, mean,
stddev)
check_status(status)
|
{"golden_diff": "diff --git a/cupy/cuda/curand.py b/cupy/cuda/curand.py\n--- a/cupy/cuda/curand.py\n+++ b/cupy/cuda/curand.py\n@@ -182,6 +182,10 @@\n \n \n def generateNormal(generator, outputPtr, n, mean, stddev):\n+ if n % 2 == 1:\n+ msg = 'curandGenerateNormal can only generate even number of '\\\n+ 'random variables simultaneously. See issue #390 for detail.'\n+ raise ValueError(msg)\n status = _curand.curandGenerateNormal(generator, outputPtr, n, mean,\n stddev)\n check_status(status)\n@@ -191,6 +195,10 @@\n \n \n def generateNormalDouble(generator, outputPtr, n, mean, stddev):\n+ if n % 2 == 1:\n+ msg = 'curandGenerateNormalDouble can only generate even number of '\\\n+ 'random variables simultaneously. See issue #390 for detail.'\n+ raise ValueError(msg)\n status = _curand.curandGenerateNormalDouble(generator, outputPtr, n, mean,\n stddev)\n check_status(status)\n", "issue": "generateNormal and generateNormalDouble raise errors when size is odd\nWhen `curand.generateNormal` is called with argument `size` odd (as long as I investigated).\nFor example, the following code fails.\n\n```\nIn [84]: generator = curand.createGenerator(curand.CURAND_RNG_PSEUDO_DEFAULT)\nIn [85]: curand.generateNormal(generator, out.data.ptr, 1, 0.0, 1.0)\n```\n\nThe error message is as follows:\n\n```\nCURANDError Traceback (most recent call last)\n<ipython-input-85-e3c65b1a19eb> in <module>()\n----> 1 curand.generateNormal(generator, out.data.ptr, 1, 0.0, 1.0)\n\n/home/delta/dev/chainer/cupy/cuda/curand.py in generateNormal(generator, outputPtr, n, mean, stddev)\n 185 status = _curand.curandGenerateNormal(generator, outputPtr, n, mean,\n 186 stddev)\n--> 187 check_status(status)\n 188 \n 189 \n\n/home/delta/dev/chainer/cupy/cuda/curand.py in check_status(status)\n 77 def check_status(status):\n 78 if status != 0:\n---> 79 raise CURANDError(status)\n 80 \n 81 ###############################################################################\n\nCURANDError: CURAND_STATUS_LENGTH_NOT_MULTIPLE\n```\n\nBut when `size` is even, no error is raised.\n\n```\nIn [86]: curand.generateNormal(generator, out.data.ptr, 2, 0.0, 1.0)\n```\n\nThe situation is true for `curand.generateNormalDouble`.\n\n", "before_files": [{"content": "\"\"\"Thin wrapper of cuRAND.\"\"\"\nimport ctypes\nimport sys\n\nfrom cupy.cuda import internal\nfrom cupy.cuda import runtime\n\nif 'win32' == sys.platform:\n _curand = internal.load_library(\n internal.get_windows_cuda_library_names('curand'))\nelse:\n _curand = internal.load_library('curand')\n\n_I = ctypes.c_int\n_U = ctypes.c_uint\n_S = ctypes.c_size_t\n_ULL = ctypes.c_ulonglong\n_P = ctypes.c_void_p\n_IP = ctypes.POINTER(_I)\n_UP = ctypes.POINTER(_U)\n_ULLP = ctypes.POINTER(_ULL)\n_F = ctypes.c_float\n_D = ctypes.c_double\n_FP = ctypes.POINTER(_F)\n_DP = ctypes.POINTER(_D)\n\nCURAND_RNG_PSEUDO_DEFAULT = 100\nCURAND_RNG_PSEUDO_XORWOW = 101\nCURAND_RNG_PSEUDO_MRG32K3A = 121\nCURAND_RNG_PSEUDO_MTGP32 = 141\nCURAND_RNG_PSEUDO_MT19937 = 142\nCURAND_RNG_PSEUDO_PHILOX4_32_10 = 161\nCURAND_RNG_QUASI_DEFAULT = 200\nCURAND_RNG_QUASI_SOBOL32 = 201\nCURAND_RNG_QUASI_SCRAMBLED_SOBOL32 = 202\nCURAND_RNG_QUASI_SOBOL64 = 203\nCURAND_RNG_QUASI_SCRAMBLED_SOBOL64 = 204\n\nCURAND_ORDERING_PSEUDO_BEST = 100\nCURAND_ORDERING_PSEUDO_DEFAULT = 101\nCURAND_ORDERING_PSEUDO_SEEDED = 102\nCURAND_ORDERING_QUASI_DEFAULT = 201\n\nGenerator = _P\nDistribution = _DP\n\nDiscreteDistribution = _P\n\n###############################################################################\n# Error handling\n###############################################################################\n\nSTATUS = {\n 0: 'CURAND_STATUS_SUCCESS',\n 100: 'CURAND_STATUS_VERSION_MISMATCH',\n 101: 'CURAND_STATUS_NOT_INITIALIZED',\n 102: 'CURAND_STATUS_ALLOCATION_FAILED',\n 103: 'CURAND_STATUS_TYPE_ERROR',\n 104: 'CURAND_STATUS_OUT_OF_RANGE',\n 105: 'CURAND_STATUS_LENGTH_NOT_MULTIPLE',\n 106: 'CURAND_STATUS_DOUBLE_PRECISION_REQUIRED',\n 201: 'CURAND_STATUS_LAUNCH_FAILURE',\n 202: 'CURAND_STATUS_PREEXISTING_FAILURE',\n 203: 'CURAND_STATUS_INITIALIZATION_FAILED',\n 204: 'CURAND_STATUS_ARCH_MISMATCH',\n 999: 'CURAND_STATUS_INTERNAL_ERROR',\n}\n\n\nclass CURANDError(RuntimeError):\n\n def __init__(self, status):\n self.status = status\n super(CURANDError, self).__init__(STATUS[status])\n\n\ndef check_status(status):\n if status != 0:\n raise CURANDError(status)\n\n###############################################################################\n# Generator\n###############################################################################\n\n_curand.curandCreateGenerator.argtypes = (_P, _I)\n\n\ndef createGenerator(rng_type):\n generator = Generator()\n status = _curand.curandCreateGenerator(ctypes.byref(generator), rng_type)\n check_status(status)\n return generator\n\n\n_curand.curandDestroyGenerator.argtypes = (Generator,)\n\n\ndef destroyGenerator(generator):\n status = _curand.curandDestroyGenerator(generator)\n check_status(status)\n\n\n_curand.curandGetVersion.argtypes = (_IP,)\n\n\ndef getVersion():\n version = _I()\n status = _curand.curandGetVersion(ctypes.byref(version))\n check_status(status)\n return version\n\n\n_curand.curandSetStream.argtypes = (Generator, runtime.Stream)\n\n\ndef setStream(generator, stream):\n status = _curand.curandSetStream(generator, stream)\n check_status(status)\n\n\n_curand.curandSetPseudoRandomGeneratorSeed.argtypes = (Generator, _ULL)\n\n\ndef setPseudoRandomGeneratorSeed(generator, seed):\n status = _curand.curandSetPseudoRandomGeneratorSeed(generator, seed)\n check_status(status)\n\n\n_curand.curandSetGeneratorOffset.argtypes = (Generator, _ULL)\n\n\ndef setGeneratorOffset(generator, offset):\n status = _curand.curandSetGeneratorOffset(generator, offset)\n check_status(status)\n\n\n_curand.curandSetGeneratorOrdering.argtypes = (Generator, _I)\n\n\ndef setGeneratorOrdering(generator, order):\n status = _curand.curandSetGeneratorOrdering(generator, order)\n check_status(status)\n\n\n###############################################################################\n# Generation functions\n###############################################################################\n\n_curand.curandGenerate.argtypes = (Generator, _P, _S)\n\n\ndef generate(generator, outputPtr, num):\n status = _curand.curandGenerate(generator, outputPtr, num)\n check_status(status)\n\n\n_curand.curandGenerateLongLong.argtypes = (Generator, _P, _S)\n\n\ndef generateLongLong(generator, outputPtr, num):\n status = _curand.curandGenerateLongLong(generator, outputPtr, num)\n check_status(status)\n\n\n_curand.curandGenerateUniform.argtypes = (Generator, _P, _S)\n\n\ndef generateUniform(generator, outputPtr, num):\n status = _curand.curandGenerateUniform(generator, outputPtr, num)\n check_status(status)\n\n\n_curand.curandGenerateUniformDouble.argtypes = (Generator, _P, _S)\n\n\ndef generateUniformDouble(generator, outputPtr, num):\n status = _curand.curandGenerateUniformDouble(generator, outputPtr, num)\n check_status(status)\n\n\n_curand.curandGenerateNormal.argtypes = (Generator, _P, _S, _F, _F)\n\n\ndef generateNormal(generator, outputPtr, n, mean, stddev):\n status = _curand.curandGenerateNormal(generator, outputPtr, n, mean,\n stddev)\n check_status(status)\n\n\n_curand.curandGenerateNormalDouble.argtypes = (Generator, _P, _S, _D, _D)\n\n\ndef generateNormalDouble(generator, outputPtr, n, mean, stddev):\n status = _curand.curandGenerateNormalDouble(generator, outputPtr, n, mean,\n stddev)\n check_status(status)\n\n\n_curand.curandGenerateLogNormal.argtypes = (Generator, _P, _S, _F, _F)\n\n\ndef generateLogNormal(generator, outputPtr, n, mean, stddev):\n status = _curand.curandGenerateLogNormal(generator, outputPtr, n,\n mean, stddev)\n check_status(status)\n\n\n_curand.curandGenerateLogNormalDouble.argtypes = (Generator, _P, _S, _D, _D)\n\n\ndef generateLogNormalDouble(generator, outputPtr, n, mean, stddev):\n status = _curand.curandGenerateLogNormalDouble(generator, outputPtr, n,\n mean, stddev)\n check_status(status)\n\n\n_curand.curandGeneratePoisson.argtypes = (Generator, _P, _S, _D)\n\n\ndef generatePoisson(generator, outputPtr, n, lam):\n status = _curand.curandGeneratePoisson(generator, outputPtr, n, lam)\n check_status(status)\n", "path": "cupy/cuda/curand.py"}]}
| 3,149 | 261 |
gh_patches_debug_4445
|
rasdani/github-patches
|
git_diff
|
kubeflow__pipelines-4134
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
allow output artifact store configuration (vs hard coded)
it seems like the output artifacts are always stored in a specific minio service, port, namespace, bucket, secrets, etc (`minio-service.kubeflow:9000`).
see: https://github.com/kubeflow/pipelines/blob/f40a22a3f4a8e06d20cf3e3f425b5058d5c87e0b/sdk/python/kfp/compiler/_op_to_template.py#L148
it would be great to make it flexible, e.g. allow using S3, or change namespace or bucket names.
i suggest making it configurable, i can do such PR if we agree its needed.
flexible pipeline service (host) path in client SDK
when creating an SDK `Client()` the path to `ml-pipeline` API service is loaded from a hard coded value (`ml-pipeline.kubeflow.svc.cluster.local:8888`) which indicate a specific k8s namespace. it can be valuable to load that default value from an env variable, i.e. changing the line in `_client.py` from:
`config.host = host if host else Client.IN_CLUSTER_DNS_NAME`
to:
`config.host = host or os.environ.get('ML_PIPELINE_DNS_NAME',Client.IN_CLUSTER_DNS_NAME)`
also note that when a user provide the `host` parameter, the ipython output points to the API server and not to the UI service (see the logic in `_get_url_prefix()`), it seems like a potential bug
if its acceptable i can submit a PR for the line change above
</issue>
<code>
[start of sdk/python/kfp/gcp.py]
1 # Copyright 2018 Google LLC
2 #
3 # Licensed under the Apache License, Version 2.0 (the "License");
4 # you may not use this file except in compliance with the License.
5 # You may obtain a copy of the License at
6 #
7 # http://www.apache.org/licenses/LICENSE-2.0
8 #
9 # Unless required by applicable law or agreed to in writing, software
10 # distributed under the License is distributed on an "AS IS" BASIS,
11 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12 # See the License for the specific language governing permissions and
13 # limitations under the License.
14
15 from kubernetes.client import V1Toleration, V1Affinity, V1NodeAffinity, \
16 V1NodeSelector, V1NodeSelectorTerm, V1NodeSelectorRequirement, V1PreferredSchedulingTerm
17
18 def use_gcp_secret(secret_name='user-gcp-sa', secret_file_path_in_volume=None, volume_name=None, secret_volume_mount_path='/secret/gcp-credentials'):
19 """An operator that configures the container to use GCP service account by service account key
20 stored in a Kubernetes secret.
21
22 For cluster setup and alternatives to using service account key, check https://www.kubeflow.org/docs/gke/authentication-pipelines/.
23 """
24
25 # permitted values for secret_name = ['admin-gcp-sa', 'user-gcp-sa']
26 if secret_file_path_in_volume is None:
27 secret_file_path_in_volume = '/' + secret_name + '.json'
28
29 if volume_name is None:
30 volume_name = 'gcp-credentials-' + secret_name
31
32 else:
33 import warnings
34 warnings.warn('The volume_name parameter is deprecated and will be removed in next release. The volume names are now generated automatically.', DeprecationWarning)
35
36 def _use_gcp_secret(task):
37 from kubernetes import client as k8s_client
38 task = task.add_volume(
39 k8s_client.V1Volume(
40 name=volume_name,
41 secret=k8s_client.V1SecretVolumeSource(
42 secret_name=secret_name,
43 )
44 )
45 )
46 task.container \
47 .add_volume_mount(
48 k8s_client.V1VolumeMount(
49 name=volume_name,
50 mount_path=secret_volume_mount_path,
51 )
52 ) \
53 .add_env_variable(
54 k8s_client.V1EnvVar(
55 name='GOOGLE_APPLICATION_CREDENTIALS',
56 value=secret_volume_mount_path + secret_file_path_in_volume,
57 )
58 ) \
59 .add_env_variable(
60 k8s_client.V1EnvVar(
61 name='CLOUDSDK_AUTH_CREDENTIAL_FILE_OVERRIDE',
62 value=secret_volume_mount_path + secret_file_path_in_volume,
63 )
64 ) # Set GCloud Credentials by using the env var override.
65 # TODO: Is there a better way for GCloud to pick up the credential?
66 return task
67
68 return _use_gcp_secret
69
70 def use_tpu(tpu_cores: int, tpu_resource: str, tf_version: str):
71 """An operator that configures GCP TPU spec in a container op.
72
73 Args:
74 tpu_cores: Required. The number of cores of TPU resource.
75 For example, the value can be '8', '32', '128', etc.
76 Check more details at: https://cloud.google.com/tpu/docs/kubernetes-engine-setup#pod-spec.
77 tpu_resource: Required. The resource name of the TPU resource.
78 For example, the value can be 'v2', 'preemptible-v1', 'v3' or 'preemptible-v3'.
79 Check more details at: https://cloud.google.com/tpu/docs/kubernetes-engine-setup#pod-spec.
80 tf_version: Required. The TensorFlow version that the TPU nodes use.
81 For example, the value can be '1.12', '1.11', '1.9' or '1.8'.
82 Check more details at: https://cloud.google.com/tpu/docs/supported-versions.
83 """
84
85 def _set_tpu_spec(task):
86 task.add_pod_annotation('tf-version.cloud-tpus.google.com', tf_version)
87 task.add_resource_limit('cloud-tpus.google.com/{}'.format(tpu_resource), str(tpu_cores))
88 return task
89
90 return _set_tpu_spec
91
92 def use_preemptible_nodepool(toleration: V1Toleration = V1Toleration(effect='NoSchedule',
93 key='preemptible',
94 operator='Equal',
95 value='true'),
96 hard_constraint: bool = False):
97 """An operator that configures the GKE preemptible in a container op.
98 Args:
99 toleration (V1Toleration): toleration to pods, default is the preemptible label.
100 hard_constraint (bool): the constraint of scheduling the pods on preemptible
101 nodepools is hard. (Default: False)
102 """
103
104 def _set_preemptible(task):
105 task.add_toleration(toleration)
106 node_selector_term = V1NodeSelectorTerm(match_expressions=[
107 V1NodeSelectorRequirement(key='cloud.google.com/gke-preemptible',
108 operator='In',
109 values=['true'])]
110 )
111 if hard_constraint:
112 node_affinity = V1NodeAffinity(required_during_scheduling_ignored_during_execution=
113 V1NodeSelector(node_selector_terms=[node_selector_term]))
114 else:
115 node_affinity = V1NodeAffinity(preferred_during_scheduling_ignored_during_execution=[
116 V1PreferredSchedulingTerm(preference=node_selector_term,
117 weight=50)
118 ])
119 affinity = V1Affinity(node_affinity=node_affinity)
120 task.add_affinity(affinity=affinity)
121 return task
122
123 return _set_preemptible
124
125 def add_gpu_toleration(toleration: V1Toleration = V1Toleration(
126 effect='NoSchedule', key='nvidia.com/gpu', operator='Equal', value='true')):
127 """An operator that configures the GKE GPU nodes in a container op.
128
129 Args:
130 toleration {V1Toleration} -- toleration to pods, default is the nvidia.com/gpu label.
131 """
132
133 def _set_toleration(task):
134 task.add_toleration(toleration)
135
136 return _set_toleration
137
[end of sdk/python/kfp/gcp.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/sdk/python/kfp/gcp.py b/sdk/python/kfp/gcp.py
--- a/sdk/python/kfp/gcp.py
+++ b/sdk/python/kfp/gcp.py
@@ -84,7 +84,7 @@
def _set_tpu_spec(task):
task.add_pod_annotation('tf-version.cloud-tpus.google.com', tf_version)
- task.add_resource_limit('cloud-tpus.google.com/{}'.format(tpu_resource), str(tpu_cores))
+ task.container.add_resource_limit('cloud-tpus.google.com/{}'.format(tpu_resource), str(tpu_cores))
return task
return _set_tpu_spec
|
{"golden_diff": "diff --git a/sdk/python/kfp/gcp.py b/sdk/python/kfp/gcp.py\n--- a/sdk/python/kfp/gcp.py\n+++ b/sdk/python/kfp/gcp.py\n@@ -84,7 +84,7 @@\n \n def _set_tpu_spec(task):\n task.add_pod_annotation('tf-version.cloud-tpus.google.com', tf_version)\n- task.add_resource_limit('cloud-tpus.google.com/{}'.format(tpu_resource), str(tpu_cores))\n+ task.container.add_resource_limit('cloud-tpus.google.com/{}'.format(tpu_resource), str(tpu_cores))\n return task\n \n return _set_tpu_spec\n", "issue": "allow output artifact store configuration (vs hard coded)\nit seems like the output artifacts are always stored in a specific minio service, port, namespace, bucket, secrets, etc (`minio-service.kubeflow:9000`). \r\n\r\nsee: https://github.com/kubeflow/pipelines/blob/f40a22a3f4a8e06d20cf3e3f425b5058d5c87e0b/sdk/python/kfp/compiler/_op_to_template.py#L148\r\n\r\nit would be great to make it flexible, e.g. allow using S3, or change namespace or bucket names.\r\ni suggest making it configurable, i can do such PR if we agree its needed. \nflexible pipeline service (host) path in client SDK \nwhen creating an SDK `Client()` the path to `ml-pipeline` API service is loaded from a hard coded value (`ml-pipeline.kubeflow.svc.cluster.local:8888`) which indicate a specific k8s namespace. it can be valuable to load that default value from an env variable, i.e. changing the line in `_client.py` from:\r\n\r\n`config.host = host if host else Client.IN_CLUSTER_DNS_NAME`\r\n\r\nto:\r\n\r\n`config.host = host or os.environ.get('ML_PIPELINE_DNS_NAME',Client.IN_CLUSTER_DNS_NAME)`\r\n\r\nalso note that when a user provide the `host` parameter, the ipython output points to the API server and not to the UI service (see the logic in `_get_url_prefix()`), it seems like a potential bug\r\n\r\nif its acceptable i can submit a PR for the line change above\r\n \n", "before_files": [{"content": "# Copyright 2018 Google LLC\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom kubernetes.client import V1Toleration, V1Affinity, V1NodeAffinity, \\\n V1NodeSelector, V1NodeSelectorTerm, V1NodeSelectorRequirement, V1PreferredSchedulingTerm\n\ndef use_gcp_secret(secret_name='user-gcp-sa', secret_file_path_in_volume=None, volume_name=None, secret_volume_mount_path='/secret/gcp-credentials'):\n \"\"\"An operator that configures the container to use GCP service account by service account key\n stored in a Kubernetes secret.\n\n For cluster setup and alternatives to using service account key, check https://www.kubeflow.org/docs/gke/authentication-pipelines/.\n \"\"\"\n\n # permitted values for secret_name = ['admin-gcp-sa', 'user-gcp-sa']\n if secret_file_path_in_volume is None:\n secret_file_path_in_volume = '/' + secret_name + '.json'\n\n if volume_name is None:\n volume_name = 'gcp-credentials-' + secret_name\n\n else:\n import warnings\n warnings.warn('The volume_name parameter is deprecated and will be removed in next release. The volume names are now generated automatically.', DeprecationWarning)\n \n def _use_gcp_secret(task):\n from kubernetes import client as k8s_client\n task = task.add_volume(\n k8s_client.V1Volume(\n name=volume_name,\n secret=k8s_client.V1SecretVolumeSource(\n secret_name=secret_name,\n )\n )\n )\n task.container \\\n .add_volume_mount(\n k8s_client.V1VolumeMount(\n name=volume_name,\n mount_path=secret_volume_mount_path,\n )\n ) \\\n .add_env_variable(\n k8s_client.V1EnvVar(\n name='GOOGLE_APPLICATION_CREDENTIALS',\n value=secret_volume_mount_path + secret_file_path_in_volume,\n )\n ) \\\n .add_env_variable(\n k8s_client.V1EnvVar(\n name='CLOUDSDK_AUTH_CREDENTIAL_FILE_OVERRIDE',\n value=secret_volume_mount_path + secret_file_path_in_volume,\n )\n ) # Set GCloud Credentials by using the env var override.\n # TODO: Is there a better way for GCloud to pick up the credential?\n return task\n \n return _use_gcp_secret\n\ndef use_tpu(tpu_cores: int, tpu_resource: str, tf_version: str):\n \"\"\"An operator that configures GCP TPU spec in a container op.\n\n Args:\n tpu_cores: Required. The number of cores of TPU resource. \n For example, the value can be '8', '32', '128', etc.\n Check more details at: https://cloud.google.com/tpu/docs/kubernetes-engine-setup#pod-spec.\n tpu_resource: Required. The resource name of the TPU resource. \n For example, the value can be 'v2', 'preemptible-v1', 'v3' or 'preemptible-v3'.\n Check more details at: https://cloud.google.com/tpu/docs/kubernetes-engine-setup#pod-spec.\n tf_version: Required. The TensorFlow version that the TPU nodes use.\n For example, the value can be '1.12', '1.11', '1.9' or '1.8'.\n Check more details at: https://cloud.google.com/tpu/docs/supported-versions.\n \"\"\"\n\n def _set_tpu_spec(task):\n task.add_pod_annotation('tf-version.cloud-tpus.google.com', tf_version)\n task.add_resource_limit('cloud-tpus.google.com/{}'.format(tpu_resource), str(tpu_cores))\n return task\n\n return _set_tpu_spec\n\ndef use_preemptible_nodepool(toleration: V1Toleration = V1Toleration(effect='NoSchedule',\n key='preemptible',\n operator='Equal',\n value='true'),\n hard_constraint: bool = False):\n \"\"\"An operator that configures the GKE preemptible in a container op.\n Args:\n toleration (V1Toleration): toleration to pods, default is the preemptible label.\n hard_constraint (bool): the constraint of scheduling the pods on preemptible\n nodepools is hard. (Default: False)\n \"\"\"\n\n def _set_preemptible(task):\n task.add_toleration(toleration)\n node_selector_term = V1NodeSelectorTerm(match_expressions=[\n V1NodeSelectorRequirement(key='cloud.google.com/gke-preemptible',\n operator='In',\n values=['true'])]\n )\n if hard_constraint:\n node_affinity = V1NodeAffinity(required_during_scheduling_ignored_during_execution=\n V1NodeSelector(node_selector_terms=[node_selector_term]))\n else:\n node_affinity = V1NodeAffinity(preferred_during_scheduling_ignored_during_execution=[\n V1PreferredSchedulingTerm(preference=node_selector_term,\n weight=50)\n ])\n affinity = V1Affinity(node_affinity=node_affinity)\n task.add_affinity(affinity=affinity)\n return task\n\n return _set_preemptible\n\ndef add_gpu_toleration(toleration: V1Toleration = V1Toleration(\n effect='NoSchedule', key='nvidia.com/gpu', operator='Equal', value='true')):\n \"\"\"An operator that configures the GKE GPU nodes in a container op.\n\n Args:\n toleration {V1Toleration} -- toleration to pods, default is the nvidia.com/gpu label.\n \"\"\"\n\n def _set_toleration(task):\n task.add_toleration(toleration)\n\n return _set_toleration\n", "path": "sdk/python/kfp/gcp.py"}]}
| 2,556 | 143 |
gh_patches_debug_11504
|
rasdani/github-patches
|
git_diff
|
alltheplaces__alltheplaces-3347
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Spider subway is broken
During the global build at 2021-09-15-14-42-44, spider **subway** failed with **31396 features** and **22 errors**.
Here's [the log](https://data.alltheplaces.xyz/runs/2021-09-15-14-42-44/logs/subway.txt) and [the output](https://data.alltheplaces.xyz/runs/2021-09-15-14-42-44/output/subway.geojson) ([on a map](https://data.alltheplaces.xyz/map.html?show=https://data.alltheplaces.xyz/runs/2021-09-15-14-42-44/output/subway.geojson))
</issue>
<code>
[start of locations/spiders/subway.py]
1 # -*- coding: utf-8 -*-
2 import scrapy
3 from locations.items import GeojsonPointItem
4 from locations.hours import OpeningHours
5
6 from urllib.parse import urlparse
7 import json
8
9
10 DAY_MAPPING = {
11 "MONDAY": "Mo",
12 "TUESDAY": "Tu",
13 "WEDNESDAY": "We",
14 "THURSDAY": "Th",
15 "FRIDAY": "Fr",
16 "SATURDAY": "Sa",
17 "SUNDAY": "Su",
18 }
19
20
21 class SubwaySpider(scrapy.Spider):
22 name = "subway"
23 item_attributes = {"name": "Subway", "brand": "Subway", "brand_wikidata": "Q244457"}
24 allowed_domains = ["restaurants.subway.com"]
25 start_urls = ["https://restaurants.subway.com/"]
26
27 link_extractor = scrapy.linkextractors.LinkExtractor(
28 restrict_css=".Directory-listLinks, .Directory-listTeasers"
29 )
30
31 def parse(self, response):
32 for link in self.link_extractor.extract_links(response):
33 yield scrapy.Request(link.url)
34
35 js = response.xpath('//script[@class="js-hours-config"]/text()').get()
36 if js:
37 yield from self.parse_restaurant(json.loads(js))
38
39 def parse_restaurant(self, js):
40 # Note: Of the five different coordinate fields, this is the one that always exists
41 lat_long = js["profile"]["yextDisplayCoordinate"]
42 website = urlparse(js["profile"]["websiteUrl"])._replace(query="").geturl()
43 properties = {
44 "lat": lat_long["lat"],
45 "lon": lat_long["long"],
46 "ref": js["profile"]["meta"]["id"],
47 "addr_full": js["profile"]["address"]["line1"],
48 "extras": {
49 "addr:unit": js["profile"]["address"]["line2"],
50 # Note: line3 is always null
51 "loc_name": js["profile"]["address"]["extraDescription"],
52 },
53 "city": js["profile"]["address"]["city"],
54 "state": js["profile"]["address"]["region"],
55 "postcode": js["profile"]["address"]["postalCode"],
56 "country": js["profile"]["address"]["countryCode"],
57 "phone": js["profile"].get("mainPhone", {}).get("number"),
58 "opening_hours": self.parse_hours(js["profile"]["hours"]["normalHours"]),
59 "website": website,
60 }
61 yield GeojsonPointItem(**properties)
62
63 def parse_hours(self, hours_json):
64 opening_hours = OpeningHours()
65 for date in hours_json:
66 day = DAY_MAPPING[date["day"]]
67 for interval in date["intervals"]:
68 start_hr, start_min = divmod(interval["start"], 100)
69 end_hr, end_min = divmod(interval["end"], 100)
70 opening_hours.add_range(
71 day, f"{start_hr}:{start_min}", f"{end_hr}:{end_min}"
72 )
73 return opening_hours.as_opening_hours()
74
[end of locations/spiders/subway.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/locations/spiders/subway.py b/locations/spiders/subway.py
--- a/locations/spiders/subway.py
+++ b/locations/spiders/subway.py
@@ -39,7 +39,9 @@
def parse_restaurant(self, js):
# Note: Of the five different coordinate fields, this is the one that always exists
lat_long = js["profile"]["yextDisplayCoordinate"]
- website = urlparse(js["profile"]["websiteUrl"])._replace(query="").geturl()
+ website = ""
+ if 'websiteUrl' in js["profile"]:
+ website = urlparse(js["profile"]["websiteUrl"])._replace(query="").geturl()
properties = {
"lat": lat_long["lat"],
"lon": lat_long["long"],
|
{"golden_diff": "diff --git a/locations/spiders/subway.py b/locations/spiders/subway.py\n--- a/locations/spiders/subway.py\n+++ b/locations/spiders/subway.py\n@@ -39,7 +39,9 @@\n def parse_restaurant(self, js):\n # Note: Of the five different coordinate fields, this is the one that always exists\n lat_long = js[\"profile\"][\"yextDisplayCoordinate\"]\n- website = urlparse(js[\"profile\"][\"websiteUrl\"])._replace(query=\"\").geturl()\n+ website = \"\"\n+ if 'websiteUrl' in js[\"profile\"]:\n+ website = urlparse(js[\"profile\"][\"websiteUrl\"])._replace(query=\"\").geturl()\n properties = {\n \"lat\": lat_long[\"lat\"],\n \"lon\": lat_long[\"long\"],\n", "issue": "Spider subway is broken\nDuring the global build at 2021-09-15-14-42-44, spider **subway** failed with **31396 features** and **22 errors**.\n\nHere's [the log](https://data.alltheplaces.xyz/runs/2021-09-15-14-42-44/logs/subway.txt) and [the output](https://data.alltheplaces.xyz/runs/2021-09-15-14-42-44/output/subway.geojson) ([on a map](https://data.alltheplaces.xyz/map.html?show=https://data.alltheplaces.xyz/runs/2021-09-15-14-42-44/output/subway.geojson))\n", "before_files": [{"content": "# -*- coding: utf-8 -*-\nimport scrapy\nfrom locations.items import GeojsonPointItem\nfrom locations.hours import OpeningHours\n\nfrom urllib.parse import urlparse\nimport json\n\n\nDAY_MAPPING = {\n \"MONDAY\": \"Mo\",\n \"TUESDAY\": \"Tu\",\n \"WEDNESDAY\": \"We\",\n \"THURSDAY\": \"Th\",\n \"FRIDAY\": \"Fr\",\n \"SATURDAY\": \"Sa\",\n \"SUNDAY\": \"Su\",\n}\n\n\nclass SubwaySpider(scrapy.Spider):\n name = \"subway\"\n item_attributes = {\"name\": \"Subway\", \"brand\": \"Subway\", \"brand_wikidata\": \"Q244457\"}\n allowed_domains = [\"restaurants.subway.com\"]\n start_urls = [\"https://restaurants.subway.com/\"]\n\n link_extractor = scrapy.linkextractors.LinkExtractor(\n restrict_css=\".Directory-listLinks, .Directory-listTeasers\"\n )\n\n def parse(self, response):\n for link in self.link_extractor.extract_links(response):\n yield scrapy.Request(link.url)\n\n js = response.xpath('//script[@class=\"js-hours-config\"]/text()').get()\n if js:\n yield from self.parse_restaurant(json.loads(js))\n\n def parse_restaurant(self, js):\n # Note: Of the five different coordinate fields, this is the one that always exists\n lat_long = js[\"profile\"][\"yextDisplayCoordinate\"]\n website = urlparse(js[\"profile\"][\"websiteUrl\"])._replace(query=\"\").geturl()\n properties = {\n \"lat\": lat_long[\"lat\"],\n \"lon\": lat_long[\"long\"],\n \"ref\": js[\"profile\"][\"meta\"][\"id\"],\n \"addr_full\": js[\"profile\"][\"address\"][\"line1\"],\n \"extras\": {\n \"addr:unit\": js[\"profile\"][\"address\"][\"line2\"],\n # Note: line3 is always null\n \"loc_name\": js[\"profile\"][\"address\"][\"extraDescription\"],\n },\n \"city\": js[\"profile\"][\"address\"][\"city\"],\n \"state\": js[\"profile\"][\"address\"][\"region\"],\n \"postcode\": js[\"profile\"][\"address\"][\"postalCode\"],\n \"country\": js[\"profile\"][\"address\"][\"countryCode\"],\n \"phone\": js[\"profile\"].get(\"mainPhone\", {}).get(\"number\"),\n \"opening_hours\": self.parse_hours(js[\"profile\"][\"hours\"][\"normalHours\"]),\n \"website\": website,\n }\n yield GeojsonPointItem(**properties)\n\n def parse_hours(self, hours_json):\n opening_hours = OpeningHours()\n for date in hours_json:\n day = DAY_MAPPING[date[\"day\"]]\n for interval in date[\"intervals\"]:\n start_hr, start_min = divmod(interval[\"start\"], 100)\n end_hr, end_min = divmod(interval[\"end\"], 100)\n opening_hours.add_range(\n day, f\"{start_hr}:{start_min}\", f\"{end_hr}:{end_min}\"\n )\n return opening_hours.as_opening_hours()\n", "path": "locations/spiders/subway.py"}]}
| 1,501 | 172 |
gh_patches_debug_29135
|
rasdani/github-patches
|
git_diff
|
googleapis__google-auth-library-python-628
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Use extra for asyncio dependencies
Hello! The latest release for this library pulls in aiohttp and its dependencies unconditionally, which adds non-trivial burden to projects that don’t need it. Would you consider using a packaging extra so that people can opt-in?
</issue>
<code>
[start of noxfile.py]
1 # Copyright 2019 Google LLC
2 #
3 # Licensed under the Apache License, Version 2.0 (the "License");
4 # you may not use this file except in compliance with the License.
5 # You may obtain a copy of the License at
6 #
7 # http://www.apache.org/licenses/LICENSE-2.0
8 #
9 # Unless required by applicable law or agreed to in writing, software
10 # distributed under the License is distributed on an "AS IS" BASIS,
11 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12 # See the License for the specific language governing permissions and
13 # limitations under the License.
14
15 import nox
16
17 TEST_DEPENDENCIES = [
18 "flask",
19 "freezegun",
20 "mock",
21 "oauth2client",
22 "pyopenssl",
23 "pytest",
24 "pytest-cov",
25 "pytest-localserver",
26 "requests",
27 "urllib3",
28 "cryptography",
29 "responses",
30 "grpcio",
31 ]
32
33 ASYNC_DEPENDENCIES = ["pytest-asyncio", "aioresponses"]
34
35 BLACK_VERSION = "black==19.3b0"
36 BLACK_PATHS = [
37 "google",
38 "tests",
39 "tests_async",
40 "noxfile.py",
41 "setup.py",
42 "docs/conf.py",
43 ]
44
45
46 @nox.session(python="3.7")
47 def lint(session):
48 session.install("flake8", "flake8-import-order", "docutils", BLACK_VERSION)
49 session.install(".")
50 session.run("black", "--check", *BLACK_PATHS)
51 session.run(
52 "flake8",
53 "--import-order-style=google",
54 "--application-import-names=google,tests,system_tests",
55 "google",
56 "tests",
57 "tests_async",
58 )
59 session.run(
60 "python", "setup.py", "check", "--metadata", "--restructuredtext", "--strict"
61 )
62
63
64 @nox.session(python="3.6")
65 def blacken(session):
66 """Run black.
67
68 Format code to uniform standard.
69
70 This currently uses Python 3.6 due to the automated Kokoro run of synthtool.
71 That run uses an image that doesn't have 3.6 installed. Before updating this
72 check the state of the `gcp_ubuntu_config` we use for that Kokoro run.
73 """
74 session.install(BLACK_VERSION)
75 session.run("black", *BLACK_PATHS)
76
77
78 @nox.session(python=["3.6", "3.7", "3.8"])
79 def unit(session):
80 session.install(*TEST_DEPENDENCIES)
81 session.install(*(ASYNC_DEPENDENCIES))
82 session.install(".")
83 session.run(
84 "pytest",
85 "--cov=google.auth",
86 "--cov=google.oauth2",
87 "--cov=tests",
88 "tests",
89 "tests_async",
90 )
91
92
93 @nox.session(python=["2.7", "3.5"])
94 def unit_prev_versions(session):
95 session.install(*TEST_DEPENDENCIES)
96 session.install(".")
97 session.run(
98 "pytest", "--cov=google.auth", "--cov=google.oauth2", "--cov=tests", "tests"
99 )
100
101
102 @nox.session(python="3.7")
103 def cover(session):
104 session.install(*TEST_DEPENDENCIES)
105 session.install(*(ASYNC_DEPENDENCIES))
106 session.install(".")
107 session.run(
108 "pytest",
109 "--cov=google.auth",
110 "--cov=google.oauth2",
111 "--cov=tests",
112 "--cov=tests_async",
113 "--cov-report=",
114 "tests",
115 "tests_async",
116 )
117 session.run("coverage", "report", "--show-missing", "--fail-under=100")
118
119
120 @nox.session(python="3.7")
121 def docgen(session):
122 session.env["SPHINX_APIDOC_OPTIONS"] = "members,inherited-members,show-inheritance"
123 session.install(*TEST_DEPENDENCIES)
124 session.install("sphinx")
125 session.install(".")
126 session.run("rm", "-r", "docs/reference")
127 session.run(
128 "sphinx-apidoc",
129 "--output-dir",
130 "docs/reference",
131 "--separate",
132 "--module-first",
133 "google",
134 )
135
136
137 @nox.session(python="3.7")
138 def docs(session):
139 session.install("sphinx", "-r", "docs/requirements-docs.txt")
140 session.install(".")
141 session.run("make", "-C", "docs", "html")
142
143
144 @nox.session(python="pypy")
145 def pypy(session):
146 session.install(*TEST_DEPENDENCIES)
147 session.install(".")
148 session.run(
149 "pytest",
150 "--cov=google.auth",
151 "--cov=google.oauth2",
152 "--cov=tests",
153 "tests",
154 "tests_async",
155 )
156
[end of noxfile.py]
[start of setup.py]
1 # Copyright 2014 Google Inc.
2 #
3 # Licensed under the Apache License, Version 2.0 (the "License");
4 # you may not use this file except in compliance with the License.
5 # You may obtain a copy of the License at
6 #
7 # http://www.apache.org/licenses/LICENSE-2.0
8 #
9 # Unless required by applicable law or agreed to in writing, software
10 # distributed under the License is distributed on an "AS IS" BASIS,
11 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12 # See the License for the specific language governing permissions and
13 # limitations under the License.
14
15 import io
16
17 from setuptools import find_packages
18 from setuptools import setup
19
20
21 DEPENDENCIES = (
22 "cachetools>=2.0.0,<5.0",
23 'enum34; python_version < "3.4"',
24 "pyasn1-modules>=0.2.1",
25 # rsa==4.5 is the last version to support 2.7
26 # https://github.com/sybrenstuvel/python-rsa/issues/152#issuecomment-643470233
27 'rsa<4.6; python_version < "3.5"',
28 'rsa>=3.1.4,<5; python_version >= "3.5"',
29 "setuptools>=40.3.0",
30 "six>=1.9.0",
31 'aiohttp >= 3.6.2, < 4.0.0dev; python_version>="3.6"',
32 )
33
34 extras = {"aiohttp": "aiohttp >= 3.6.2, < 4.0.0dev; python_version>='3.6'"}
35
36 with io.open("README.rst", "r") as fh:
37 long_description = fh.read()
38
39 version = "1.22.1"
40
41 setup(
42 name="google-auth",
43 version=version,
44 author="Google Cloud Platform",
45 author_email="[email protected]",
46 description="Google Authentication Library",
47 long_description=long_description,
48 url="https://github.com/googleapis/google-auth-library-python",
49 packages=find_packages(exclude=("tests*", "system_tests*")),
50 namespace_packages=("google",),
51 install_requires=DEPENDENCIES,
52 extras_require=extras,
53 python_requires=">=2.7,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*",
54 license="Apache 2.0",
55 keywords="google auth oauth client",
56 classifiers=[
57 "Programming Language :: Python :: 2",
58 "Programming Language :: Python :: 2.7",
59 "Programming Language :: Python :: 3",
60 "Programming Language :: Python :: 3.5",
61 "Programming Language :: Python :: 3.6",
62 "Programming Language :: Python :: 3.7",
63 "Programming Language :: Python :: 3.8",
64 "Development Status :: 5 - Production/Stable",
65 "Intended Audience :: Developers",
66 "License :: OSI Approved :: Apache Software License",
67 "Operating System :: POSIX",
68 "Operating System :: Microsoft :: Windows",
69 "Operating System :: MacOS :: MacOS X",
70 "Operating System :: OS Independent",
71 "Topic :: Internet :: WWW/HTTP",
72 ],
73 )
74
[end of setup.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/noxfile.py b/noxfile.py
--- a/noxfile.py
+++ b/noxfile.py
@@ -30,7 +30,7 @@
"grpcio",
]
-ASYNC_DEPENDENCIES = ["pytest-asyncio", "aioresponses"]
+ASYNC_DEPENDENCIES = ["pytest-asyncio", "aioresponses", "asynctest"]
BLACK_VERSION = "black==19.3b0"
BLACK_PATHS = [
@@ -64,9 +64,7 @@
@nox.session(python="3.6")
def blacken(session):
"""Run black.
-
Format code to uniform standard.
-
This currently uses Python 3.6 due to the automated Kokoro run of synthtool.
That run uses an image that doesn't have 3.6 installed. Before updating this
check the state of the `gcp_ubuntu_config` we use for that Kokoro run.
diff --git a/setup.py b/setup.py
--- a/setup.py
+++ b/setup.py
@@ -20,7 +20,6 @@
DEPENDENCIES = (
"cachetools>=2.0.0,<5.0",
- 'enum34; python_version < "3.4"',
"pyasn1-modules>=0.2.1",
# rsa==4.5 is the last version to support 2.7
# https://github.com/sybrenstuvel/python-rsa/issues/152#issuecomment-643470233
@@ -28,7 +27,6 @@
'rsa>=3.1.4,<5; python_version >= "3.5"',
"setuptools>=40.3.0",
"six>=1.9.0",
- 'aiohttp >= 3.6.2, < 4.0.0dev; python_version>="3.6"',
)
extras = {"aiohttp": "aiohttp >= 3.6.2, < 4.0.0dev; python_version>='3.6'"}
|
{"golden_diff": "diff --git a/noxfile.py b/noxfile.py\n--- a/noxfile.py\n+++ b/noxfile.py\n@@ -30,7 +30,7 @@\n \"grpcio\",\n ]\n \n-ASYNC_DEPENDENCIES = [\"pytest-asyncio\", \"aioresponses\"]\n+ASYNC_DEPENDENCIES = [\"pytest-asyncio\", \"aioresponses\", \"asynctest\"]\n \n BLACK_VERSION = \"black==19.3b0\"\n BLACK_PATHS = [\n@@ -64,9 +64,7 @@\n @nox.session(python=\"3.6\")\n def blacken(session):\n \"\"\"Run black.\n-\n Format code to uniform standard.\n-\n This currently uses Python 3.6 due to the automated Kokoro run of synthtool.\n That run uses an image that doesn't have 3.6 installed. Before updating this\n check the state of the `gcp_ubuntu_config` we use for that Kokoro run.\ndiff --git a/setup.py b/setup.py\n--- a/setup.py\n+++ b/setup.py\n@@ -20,7 +20,6 @@\n \n DEPENDENCIES = (\n \"cachetools>=2.0.0,<5.0\",\n- 'enum34; python_version < \"3.4\"',\n \"pyasn1-modules>=0.2.1\",\n # rsa==4.5 is the last version to support 2.7\n # https://github.com/sybrenstuvel/python-rsa/issues/152#issuecomment-643470233\n@@ -28,7 +27,6 @@\n 'rsa>=3.1.4,<5; python_version >= \"3.5\"',\n \"setuptools>=40.3.0\",\n \"six>=1.9.0\",\n- 'aiohttp >= 3.6.2, < 4.0.0dev; python_version>=\"3.6\"',\n )\n \n extras = {\"aiohttp\": \"aiohttp >= 3.6.2, < 4.0.0dev; python_version>='3.6'\"}\n", "issue": "Use extra for asyncio dependencies\nHello! The latest release for this library pulls in aiohttp and its dependencies unconditionally, which adds non-trivial burden to projects that don\u2019t need it. Would you consider using a packaging extra so that people can opt-in?\n", "before_files": [{"content": "# Copyright 2019 Google LLC\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nimport nox\n\nTEST_DEPENDENCIES = [\n \"flask\",\n \"freezegun\",\n \"mock\",\n \"oauth2client\",\n \"pyopenssl\",\n \"pytest\",\n \"pytest-cov\",\n \"pytest-localserver\",\n \"requests\",\n \"urllib3\",\n \"cryptography\",\n \"responses\",\n \"grpcio\",\n]\n\nASYNC_DEPENDENCIES = [\"pytest-asyncio\", \"aioresponses\"]\n\nBLACK_VERSION = \"black==19.3b0\"\nBLACK_PATHS = [\n \"google\",\n \"tests\",\n \"tests_async\",\n \"noxfile.py\",\n \"setup.py\",\n \"docs/conf.py\",\n]\n\n\[email protected](python=\"3.7\")\ndef lint(session):\n session.install(\"flake8\", \"flake8-import-order\", \"docutils\", BLACK_VERSION)\n session.install(\".\")\n session.run(\"black\", \"--check\", *BLACK_PATHS)\n session.run(\n \"flake8\",\n \"--import-order-style=google\",\n \"--application-import-names=google,tests,system_tests\",\n \"google\",\n \"tests\",\n \"tests_async\",\n )\n session.run(\n \"python\", \"setup.py\", \"check\", \"--metadata\", \"--restructuredtext\", \"--strict\"\n )\n\n\[email protected](python=\"3.6\")\ndef blacken(session):\n \"\"\"Run black.\n\n Format code to uniform standard.\n\n This currently uses Python 3.6 due to the automated Kokoro run of synthtool.\n That run uses an image that doesn't have 3.6 installed. Before updating this\n check the state of the `gcp_ubuntu_config` we use for that Kokoro run.\n \"\"\"\n session.install(BLACK_VERSION)\n session.run(\"black\", *BLACK_PATHS)\n\n\[email protected](python=[\"3.6\", \"3.7\", \"3.8\"])\ndef unit(session):\n session.install(*TEST_DEPENDENCIES)\n session.install(*(ASYNC_DEPENDENCIES))\n session.install(\".\")\n session.run(\n \"pytest\",\n \"--cov=google.auth\",\n \"--cov=google.oauth2\",\n \"--cov=tests\",\n \"tests\",\n \"tests_async\",\n )\n\n\[email protected](python=[\"2.7\", \"3.5\"])\ndef unit_prev_versions(session):\n session.install(*TEST_DEPENDENCIES)\n session.install(\".\")\n session.run(\n \"pytest\", \"--cov=google.auth\", \"--cov=google.oauth2\", \"--cov=tests\", \"tests\"\n )\n\n\[email protected](python=\"3.7\")\ndef cover(session):\n session.install(*TEST_DEPENDENCIES)\n session.install(*(ASYNC_DEPENDENCIES))\n session.install(\".\")\n session.run(\n \"pytest\",\n \"--cov=google.auth\",\n \"--cov=google.oauth2\",\n \"--cov=tests\",\n \"--cov=tests_async\",\n \"--cov-report=\",\n \"tests\",\n \"tests_async\",\n )\n session.run(\"coverage\", \"report\", \"--show-missing\", \"--fail-under=100\")\n\n\[email protected](python=\"3.7\")\ndef docgen(session):\n session.env[\"SPHINX_APIDOC_OPTIONS\"] = \"members,inherited-members,show-inheritance\"\n session.install(*TEST_DEPENDENCIES)\n session.install(\"sphinx\")\n session.install(\".\")\n session.run(\"rm\", \"-r\", \"docs/reference\")\n session.run(\n \"sphinx-apidoc\",\n \"--output-dir\",\n \"docs/reference\",\n \"--separate\",\n \"--module-first\",\n \"google\",\n )\n\n\[email protected](python=\"3.7\")\ndef docs(session):\n session.install(\"sphinx\", \"-r\", \"docs/requirements-docs.txt\")\n session.install(\".\")\n session.run(\"make\", \"-C\", \"docs\", \"html\")\n\n\[email protected](python=\"pypy\")\ndef pypy(session):\n session.install(*TEST_DEPENDENCIES)\n session.install(\".\")\n session.run(\n \"pytest\",\n \"--cov=google.auth\",\n \"--cov=google.oauth2\",\n \"--cov=tests\",\n \"tests\",\n \"tests_async\",\n )\n", "path": "noxfile.py"}, {"content": "# Copyright 2014 Google Inc.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nimport io\n\nfrom setuptools import find_packages\nfrom setuptools import setup\n\n\nDEPENDENCIES = (\n \"cachetools>=2.0.0,<5.0\",\n 'enum34; python_version < \"3.4\"',\n \"pyasn1-modules>=0.2.1\",\n # rsa==4.5 is the last version to support 2.7\n # https://github.com/sybrenstuvel/python-rsa/issues/152#issuecomment-643470233\n 'rsa<4.6; python_version < \"3.5\"',\n 'rsa>=3.1.4,<5; python_version >= \"3.5\"',\n \"setuptools>=40.3.0\",\n \"six>=1.9.0\",\n 'aiohttp >= 3.6.2, < 4.0.0dev; python_version>=\"3.6\"',\n)\n\nextras = {\"aiohttp\": \"aiohttp >= 3.6.2, < 4.0.0dev; python_version>='3.6'\"}\n\nwith io.open(\"README.rst\", \"r\") as fh:\n long_description = fh.read()\n\nversion = \"1.22.1\"\n\nsetup(\n name=\"google-auth\",\n version=version,\n author=\"Google Cloud Platform\",\n author_email=\"[email protected]\",\n description=\"Google Authentication Library\",\n long_description=long_description,\n url=\"https://github.com/googleapis/google-auth-library-python\",\n packages=find_packages(exclude=(\"tests*\", \"system_tests*\")),\n namespace_packages=(\"google\",),\n install_requires=DEPENDENCIES,\n extras_require=extras,\n python_requires=\">=2.7,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*\",\n license=\"Apache 2.0\",\n keywords=\"google auth oauth client\",\n classifiers=[\n \"Programming Language :: Python :: 2\",\n \"Programming Language :: Python :: 2.7\",\n \"Programming Language :: Python :: 3\",\n \"Programming Language :: Python :: 3.5\",\n \"Programming Language :: Python :: 3.6\",\n \"Programming Language :: Python :: 3.7\",\n \"Programming Language :: Python :: 3.8\",\n \"Development Status :: 5 - Production/Stable\",\n \"Intended Audience :: Developers\",\n \"License :: OSI Approved :: Apache Software License\",\n \"Operating System :: POSIX\",\n \"Operating System :: Microsoft :: Windows\",\n \"Operating System :: MacOS :: MacOS X\",\n \"Operating System :: OS Independent\",\n \"Topic :: Internet :: WWW/HTTP\",\n ],\n)\n", "path": "setup.py"}]}
| 2,827 | 465 |
gh_patches_debug_23194
|
rasdani/github-patches
|
git_diff
|
microsoft__botbuilder-python-1221
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
[PORT] Update telemetry logger to include attachments
> Port this change from botbuilder-dotnet/master branch:
https://github.com/microsoft/botbuilder-dotnet/pull/3737
Addresses #1574 in JS repo (https://github.com/microsoft/botbuilder-js/issues/1574)
Telemetry logger will now include the collection of any attachments sent with the activity, if the LogPersonalInformation flag is set to true.
# Changed projects
* Microsoft.Bot.Builder
</issue>
<code>
[start of libraries/botbuilder-core/botbuilder/core/telemetry_constants.py]
1 # Copyright (c) Microsoft Corporation. All rights reserved.
2 # Licensed under the MIT License
3
4
5 class TelemetryConstants:
6 """Telemetry logger property names."""
7
8 CHANNEL_ID_PROPERTY: str = "channelId"
9 CONVERSATION_ID_PROPERTY: str = "conversationId"
10 CONVERSATION_NAME_PROPERTY: str = "conversationName"
11 DIALOG_ID_PROPERTY: str = "dialogId"
12 FROM_ID_PROPERTY: str = "fromId"
13 FROM_NAME_PROPERTY: str = "fromName"
14 LOCALE_PROPERTY: str = "locale"
15 RECIPIENT_ID_PROPERTY: str = "recipientId"
16 RECIPIENT_NAME_PROPERTY: str = "recipientName"
17 REPLY_ACTIVITY_ID_PROPERTY: str = "replyActivityId"
18 TEXT_PROPERTY: str = "text"
19 SPEAK_PROPERTY: str = "speak"
20 USER_ID_PROPERTY: str = "userId"
21
[end of libraries/botbuilder-core/botbuilder/core/telemetry_constants.py]
[start of libraries/botbuilder-core/botbuilder/core/telemetry_logger_middleware.py]
1 # Copyright (c) Microsoft Corporation. All rights reserved.
2 # Licensed under the MIT License.
3 """Middleware Component for logging Activity messages."""
4
5 from typing import Awaitable, Callable, List, Dict
6 from botbuilder.schema import Activity, ConversationReference, ActivityTypes
7 from .bot_telemetry_client import BotTelemetryClient
8 from .bot_assert import BotAssert
9 from .middleware_set import Middleware
10 from .null_telemetry_client import NullTelemetryClient
11 from .turn_context import TurnContext
12 from .telemetry_constants import TelemetryConstants
13 from .telemetry_logger_constants import TelemetryLoggerConstants
14
15
16 # pylint: disable=line-too-long
17 class TelemetryLoggerMiddleware(Middleware):
18 """Middleware for logging incoming, outgoing, updated or deleted Activity messages."""
19
20 def __init__(
21 self, telemetry_client: BotTelemetryClient, log_personal_information: bool
22 ) -> None:
23 super(TelemetryLoggerMiddleware, self).__init__()
24 self._telemetry_client = telemetry_client or NullTelemetryClient()
25 self._log_personal_information = log_personal_information
26
27 @property
28 def telemetry_client(self) -> BotTelemetryClient:
29 """Gets the currently configured BotTelemetryClient."""
30 return self._telemetry_client
31
32 @property
33 def log_personal_information(self) -> bool:
34 """ Gets a value indicating whether determines whether to log personal
35 information that came from the user."""
36 return self._log_personal_information
37
38 # pylint: disable=arguments-differ
39 async def on_turn(
40 self, context: TurnContext, logic_fn: Callable[[TurnContext], Awaitable]
41 ) -> None:
42 """Logs events based on incoming and outgoing activities using
43 BotTelemetryClient base class
44
45 :param turn_context: The context object for this turn.
46 :param logic: Callable to continue the bot middleware pipeline
47
48 :return: None
49 """
50 BotAssert.context_not_none(context)
51
52 # Log incoming activity at beginning of turn
53 if context.activity:
54 activity = context.activity
55 # Log Bot Message Received
56 await self.on_receive_activity(activity)
57
58 # hook up onSend pipeline
59 # pylint: disable=unused-argument
60 async def send_activities_handler(
61 ctx: TurnContext,
62 activities: List[Activity],
63 next_send: Callable[[], Awaitable[None]],
64 ):
65 # Run full pipeline
66 responses = await next_send()
67 for activity in activities:
68 await self.on_send_activity(activity)
69 return responses
70
71 context.on_send_activities(send_activities_handler)
72
73 # hook up update activity pipeline
74 async def update_activity_handler(
75 ctx: TurnContext, activity: Activity, next_update: Callable[[], Awaitable]
76 ):
77 # Run full pipeline
78 response = await next_update()
79 await self.on_update_activity(activity)
80 return response
81
82 context.on_update_activity(update_activity_handler)
83
84 # hook up delete activity pipeline
85 async def delete_activity_handler(
86 ctx: TurnContext,
87 reference: ConversationReference,
88 next_delete: Callable[[], Awaitable],
89 ):
90 # Run full pipeline
91 await next_delete()
92
93 delete_msg = Activity(
94 type=ActivityTypes.message_delete, id=reference.activity_id
95 )
96 deleted_activity: Activity = TurnContext.apply_conversation_reference(
97 delete_msg, reference, False
98 )
99 await self.on_delete_activity(deleted_activity)
100
101 context.on_delete_activity(delete_activity_handler)
102
103 if logic_fn:
104 await logic_fn()
105
106 async def on_receive_activity(self, activity: Activity) -> None:
107 """Invoked when a message is received from the user.
108 Performs logging of telemetry data using the BotTelemetryClient.track_event() method.
109 This event name used is "BotMessageReceived".
110 :param activity: Current activity sent from user.
111 """
112 self.telemetry_client.track_event(
113 TelemetryLoggerConstants.BOT_MSG_RECEIVE_EVENT,
114 await self.fill_receive_event_properties(activity),
115 )
116
117 async def on_send_activity(self, activity: Activity) -> None:
118 """Invoked when the bot sends a message to the user.
119 Performs logging of telemetry data using the BotTelemetryClient.track_event() method.
120 This event name used is "BotMessageSend".
121 :param activity: Current activity sent from bot.
122 """
123 self.telemetry_client.track_event(
124 TelemetryLoggerConstants.BOT_MSG_SEND_EVENT,
125 await self.fill_send_event_properties(activity),
126 )
127
128 async def on_update_activity(self, activity: Activity) -> None:
129 """Invoked when the bot updates a message.
130 Performs logging of telemetry data using the BotTelemetryClient.track_event() method.
131 This event name used is "BotMessageUpdate".
132 :param activity: Current activity sent from user.
133 """
134 self.telemetry_client.track_event(
135 TelemetryLoggerConstants.BOT_MSG_UPDATE_EVENT,
136 await self.fill_update_event_properties(activity),
137 )
138
139 async def on_delete_activity(self, activity: Activity) -> None:
140 """Invoked when the bot deletes a message.
141 Performs logging of telemetry data using the BotTelemetryClient.track_event() method.
142 This event name used is "BotMessageDelete".
143 :param activity: Current activity sent from user.
144 """
145 self.telemetry_client.track_event(
146 TelemetryLoggerConstants.BOT_MSG_DELETE_EVENT,
147 await self.fill_delete_event_properties(activity),
148 )
149
150 async def fill_receive_event_properties(
151 self, activity: Activity, additional_properties: Dict[str, str] = None
152 ) -> Dict[str, str]:
153 """Fills the event properties for the BotMessageReceived.
154 Adheres to the LogPersonalInformation flag to filter Name, Text and Speak properties.
155 :param activity: activity sent from user.
156 :param additional_properties: Additional properties to add to the event.
157 Additional properties can override "stock" properties.
158
159 :return: A dictionary that is sent as "Properties" to
160 BotTelemetryClient.track_event method for the BotMessageReceived event.
161 """
162 properties = {
163 TelemetryConstants.FROM_ID_PROPERTY: activity.from_property.id
164 if activity.from_property
165 else None,
166 TelemetryConstants.CONVERSATION_NAME_PROPERTY: activity.conversation.name,
167 TelemetryConstants.LOCALE_PROPERTY: activity.locale,
168 TelemetryConstants.RECIPIENT_ID_PROPERTY: activity.recipient.id,
169 TelemetryConstants.RECIPIENT_NAME_PROPERTY: activity.recipient.name,
170 }
171
172 if self.log_personal_information:
173 if (
174 activity.from_property
175 and activity.from_property.name
176 and activity.from_property.name.strip()
177 ):
178 properties[
179 TelemetryConstants.FROM_NAME_PROPERTY
180 ] = activity.from_property.name
181 if activity.text and activity.text.strip():
182 properties[TelemetryConstants.TEXT_PROPERTY] = activity.text
183 if activity.speak and activity.speak.strip():
184 properties[TelemetryConstants.SPEAK_PROPERTY] = activity.speak
185
186 # Additional properties can override "stock" properties
187 if additional_properties:
188 for prop in additional_properties:
189 properties[prop.key] = prop.value
190
191 return properties
192
193 async def fill_send_event_properties(
194 self, activity: Activity, additional_properties: Dict[str, str] = None
195 ) -> Dict[str, str]:
196 """Fills the event properties for the BotMessageSend.
197 These properties are logged when an activity message is sent by the Bot to the user.
198 :param activity: activity sent from user.
199 :param additional_properties: Additional properties to add to the event.
200 Additional properties can override "stock" properties.
201
202 :return: A dictionary that is sent as "Properties" to the
203 BotTelemetryClient.track_event method for the BotMessageSend event.
204 """
205 properties = {
206 TelemetryConstants.REPLY_ACTIVITY_ID_PROPERTY: activity.reply_to_id,
207 TelemetryConstants.RECIPIENT_ID_PROPERTY: activity.recipient.id,
208 TelemetryConstants.CONVERSATION_NAME_PROPERTY: activity.conversation.name,
209 TelemetryConstants.LOCALE_PROPERTY: activity.locale,
210 }
211
212 # Use the LogPersonalInformation flag to toggle logging PII data, text and user name are common examples
213 if self.log_personal_information:
214 if activity.from_property.name and activity.from_property.name.strip():
215 properties[
216 TelemetryConstants.FROM_NAME_PROPERTY
217 ] = activity.from_property.name
218 if activity.text and activity.text.strip():
219 properties[TelemetryConstants.TEXT_PROPERTY] = activity.text
220 if activity.speak and activity.speak.strip():
221 properties[TelemetryConstants.SPEAK_PROPERTY] = activity.speak
222
223 # Additional properties can override "stock" properties
224 if additional_properties:
225 for prop in additional_properties:
226 properties[prop.key] = prop.value
227
228 return properties
229
230 async def fill_update_event_properties(
231 self, activity: Activity, additional_properties: Dict[str, str] = None
232 ) -> Dict[str, str]:
233 """Fills the event properties for the BotMessageUpdate.
234 These properties are logged when an activity message is updated by the Bot.
235 For example, if a card is interacted with by the use, and the card needs
236 to be updated to reflect some interaction.
237 :param activity: activity sent from user.
238 :param additional_properties: Additional properties to add to the event.
239 Additional properties can override "stock" properties.
240
241 :return: A dictionary that is sent as "Properties" to the
242 BotTelemetryClient.track_event method for the BotMessageUpdate event.
243 """
244 properties = {
245 TelemetryConstants.RECIPIENT_ID_PROPERTY: activity.recipient.id,
246 TelemetryConstants.CONVERSATION_ID_PROPERTY: activity.conversation.id,
247 TelemetryConstants.CONVERSATION_NAME_PROPERTY: activity.conversation.name,
248 TelemetryConstants.LOCALE_PROPERTY: activity.locale,
249 }
250
251 # Use the LogPersonalInformation flag to toggle logging PII data, text is a common examples
252 if self.log_personal_information:
253 if activity.text and activity.text.strip():
254 properties[TelemetryConstants.TEXT_PROPERTY] = activity.text
255
256 # Additional properties can override "stock" properties
257 if additional_properties:
258 for prop in additional_properties:
259 properties[prop.key] = prop.value
260
261 return properties
262
263 async def fill_delete_event_properties(
264 self, activity: Activity, additional_properties: Dict[str, str] = None
265 ) -> Dict[str, str]:
266 """Fills the event properties for the BotMessageDelete.
267 These properties are logged when an activity message is deleted by the Bot.
268 :param activity: activity sent from user.
269 :param additional_properties: Additional properties to add to the event.
270 Additional properties can override "stock" properties.
271
272 :return: A dictionary that is sent as "Properties" to the
273 BotTelemetryClient.track_event method for the BotMessageUpdate event.
274 """
275 properties = {
276 TelemetryConstants.RECIPIENT_ID_PROPERTY: activity.recipient.id,
277 TelemetryConstants.CONVERSATION_ID_PROPERTY: activity.conversation.id,
278 TelemetryConstants.CONVERSATION_NAME_PROPERTY: activity.conversation.name,
279 }
280
281 # Additional properties can override "stock" properties
282 if additional_properties:
283 for prop in additional_properties:
284 properties[prop.key] = prop.value
285
286 return properties
287
[end of libraries/botbuilder-core/botbuilder/core/telemetry_logger_middleware.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/libraries/botbuilder-core/botbuilder/core/telemetry_constants.py b/libraries/botbuilder-core/botbuilder/core/telemetry_constants.py
--- a/libraries/botbuilder-core/botbuilder/core/telemetry_constants.py
+++ b/libraries/botbuilder-core/botbuilder/core/telemetry_constants.py
@@ -5,6 +5,7 @@
class TelemetryConstants:
"""Telemetry logger property names."""
+ ATTACHMENTS_PROPERTY: str = "attachments"
CHANNEL_ID_PROPERTY: str = "channelId"
CONVERSATION_ID_PROPERTY: str = "conversationId"
CONVERSATION_NAME_PROPERTY: str = "conversationName"
diff --git a/libraries/botbuilder-core/botbuilder/core/telemetry_logger_middleware.py b/libraries/botbuilder-core/botbuilder/core/telemetry_logger_middleware.py
--- a/libraries/botbuilder-core/botbuilder/core/telemetry_logger_middleware.py
+++ b/libraries/botbuilder-core/botbuilder/core/telemetry_logger_middleware.py
@@ -211,6 +211,10 @@
# Use the LogPersonalInformation flag to toggle logging PII data, text and user name are common examples
if self.log_personal_information:
+ if activity.attachments and activity.attachments.strip():
+ properties[
+ TelemetryConstants.ATTACHMENTS_PROPERTY
+ ] = activity.attachments
if activity.from_property.name and activity.from_property.name.strip():
properties[
TelemetryConstants.FROM_NAME_PROPERTY
|
{"golden_diff": "diff --git a/libraries/botbuilder-core/botbuilder/core/telemetry_constants.py b/libraries/botbuilder-core/botbuilder/core/telemetry_constants.py\n--- a/libraries/botbuilder-core/botbuilder/core/telemetry_constants.py\n+++ b/libraries/botbuilder-core/botbuilder/core/telemetry_constants.py\n@@ -5,6 +5,7 @@\n class TelemetryConstants:\n \"\"\"Telemetry logger property names.\"\"\"\n \n+ ATTACHMENTS_PROPERTY: str = \"attachments\"\n CHANNEL_ID_PROPERTY: str = \"channelId\"\n CONVERSATION_ID_PROPERTY: str = \"conversationId\"\n CONVERSATION_NAME_PROPERTY: str = \"conversationName\"\ndiff --git a/libraries/botbuilder-core/botbuilder/core/telemetry_logger_middleware.py b/libraries/botbuilder-core/botbuilder/core/telemetry_logger_middleware.py\n--- a/libraries/botbuilder-core/botbuilder/core/telemetry_logger_middleware.py\n+++ b/libraries/botbuilder-core/botbuilder/core/telemetry_logger_middleware.py\n@@ -211,6 +211,10 @@\n \n # Use the LogPersonalInformation flag to toggle logging PII data, text and user name are common examples\n if self.log_personal_information:\n+ if activity.attachments and activity.attachments.strip():\n+ properties[\n+ TelemetryConstants.ATTACHMENTS_PROPERTY\n+ ] = activity.attachments\n if activity.from_property.name and activity.from_property.name.strip():\n properties[\n TelemetryConstants.FROM_NAME_PROPERTY\n", "issue": "[PORT] Update telemetry logger to include attachments\n> Port this change from botbuilder-dotnet/master branch:\nhttps://github.com/microsoft/botbuilder-dotnet/pull/3737\n\nAddresses #1574 in JS repo (https://github.com/microsoft/botbuilder-js/issues/1574)\r\n\r\nTelemetry logger will now include the collection of any attachments sent with the activity, if the LogPersonalInformation flag is set to true.\n\n\r\n# Changed projects\r\n* Microsoft.Bot.Builder\r\n\r\n\r\n\n\n", "before_files": [{"content": "# Copyright (c) Microsoft Corporation. All rights reserved.\n# Licensed under the MIT License\n\n\nclass TelemetryConstants:\n \"\"\"Telemetry logger property names.\"\"\"\n\n CHANNEL_ID_PROPERTY: str = \"channelId\"\n CONVERSATION_ID_PROPERTY: str = \"conversationId\"\n CONVERSATION_NAME_PROPERTY: str = \"conversationName\"\n DIALOG_ID_PROPERTY: str = \"dialogId\"\n FROM_ID_PROPERTY: str = \"fromId\"\n FROM_NAME_PROPERTY: str = \"fromName\"\n LOCALE_PROPERTY: str = \"locale\"\n RECIPIENT_ID_PROPERTY: str = \"recipientId\"\n RECIPIENT_NAME_PROPERTY: str = \"recipientName\"\n REPLY_ACTIVITY_ID_PROPERTY: str = \"replyActivityId\"\n TEXT_PROPERTY: str = \"text\"\n SPEAK_PROPERTY: str = \"speak\"\n USER_ID_PROPERTY: str = \"userId\"\n", "path": "libraries/botbuilder-core/botbuilder/core/telemetry_constants.py"}, {"content": "# Copyright (c) Microsoft Corporation. All rights reserved.\n# Licensed under the MIT License.\n\"\"\"Middleware Component for logging Activity messages.\"\"\"\n\nfrom typing import Awaitable, Callable, List, Dict\nfrom botbuilder.schema import Activity, ConversationReference, ActivityTypes\nfrom .bot_telemetry_client import BotTelemetryClient\nfrom .bot_assert import BotAssert\nfrom .middleware_set import Middleware\nfrom .null_telemetry_client import NullTelemetryClient\nfrom .turn_context import TurnContext\nfrom .telemetry_constants import TelemetryConstants\nfrom .telemetry_logger_constants import TelemetryLoggerConstants\n\n\n# pylint: disable=line-too-long\nclass TelemetryLoggerMiddleware(Middleware):\n \"\"\"Middleware for logging incoming, outgoing, updated or deleted Activity messages.\"\"\"\n\n def __init__(\n self, telemetry_client: BotTelemetryClient, log_personal_information: bool\n ) -> None:\n super(TelemetryLoggerMiddleware, self).__init__()\n self._telemetry_client = telemetry_client or NullTelemetryClient()\n self._log_personal_information = log_personal_information\n\n @property\n def telemetry_client(self) -> BotTelemetryClient:\n \"\"\"Gets the currently configured BotTelemetryClient.\"\"\"\n return self._telemetry_client\n\n @property\n def log_personal_information(self) -> bool:\n \"\"\" Gets a value indicating whether determines whether to log personal\n information that came from the user.\"\"\"\n return self._log_personal_information\n\n # pylint: disable=arguments-differ\n async def on_turn(\n self, context: TurnContext, logic_fn: Callable[[TurnContext], Awaitable]\n ) -> None:\n \"\"\"Logs events based on incoming and outgoing activities using\n BotTelemetryClient base class\n\n :param turn_context: The context object for this turn.\n :param logic: Callable to continue the bot middleware pipeline\n\n :return: None\n \"\"\"\n BotAssert.context_not_none(context)\n\n # Log incoming activity at beginning of turn\n if context.activity:\n activity = context.activity\n # Log Bot Message Received\n await self.on_receive_activity(activity)\n\n # hook up onSend pipeline\n # pylint: disable=unused-argument\n async def send_activities_handler(\n ctx: TurnContext,\n activities: List[Activity],\n next_send: Callable[[], Awaitable[None]],\n ):\n # Run full pipeline\n responses = await next_send()\n for activity in activities:\n await self.on_send_activity(activity)\n return responses\n\n context.on_send_activities(send_activities_handler)\n\n # hook up update activity pipeline\n async def update_activity_handler(\n ctx: TurnContext, activity: Activity, next_update: Callable[[], Awaitable]\n ):\n # Run full pipeline\n response = await next_update()\n await self.on_update_activity(activity)\n return response\n\n context.on_update_activity(update_activity_handler)\n\n # hook up delete activity pipeline\n async def delete_activity_handler(\n ctx: TurnContext,\n reference: ConversationReference,\n next_delete: Callable[[], Awaitable],\n ):\n # Run full pipeline\n await next_delete()\n\n delete_msg = Activity(\n type=ActivityTypes.message_delete, id=reference.activity_id\n )\n deleted_activity: Activity = TurnContext.apply_conversation_reference(\n delete_msg, reference, False\n )\n await self.on_delete_activity(deleted_activity)\n\n context.on_delete_activity(delete_activity_handler)\n\n if logic_fn:\n await logic_fn()\n\n async def on_receive_activity(self, activity: Activity) -> None:\n \"\"\"Invoked when a message is received from the user.\n Performs logging of telemetry data using the BotTelemetryClient.track_event() method.\n This event name used is \"BotMessageReceived\".\n :param activity: Current activity sent from user.\n \"\"\"\n self.telemetry_client.track_event(\n TelemetryLoggerConstants.BOT_MSG_RECEIVE_EVENT,\n await self.fill_receive_event_properties(activity),\n )\n\n async def on_send_activity(self, activity: Activity) -> None:\n \"\"\"Invoked when the bot sends a message to the user.\n Performs logging of telemetry data using the BotTelemetryClient.track_event() method.\n This event name used is \"BotMessageSend\".\n :param activity: Current activity sent from bot.\n \"\"\"\n self.telemetry_client.track_event(\n TelemetryLoggerConstants.BOT_MSG_SEND_EVENT,\n await self.fill_send_event_properties(activity),\n )\n\n async def on_update_activity(self, activity: Activity) -> None:\n \"\"\"Invoked when the bot updates a message.\n Performs logging of telemetry data using the BotTelemetryClient.track_event() method.\n This event name used is \"BotMessageUpdate\".\n :param activity: Current activity sent from user.\n \"\"\"\n self.telemetry_client.track_event(\n TelemetryLoggerConstants.BOT_MSG_UPDATE_EVENT,\n await self.fill_update_event_properties(activity),\n )\n\n async def on_delete_activity(self, activity: Activity) -> None:\n \"\"\"Invoked when the bot deletes a message.\n Performs logging of telemetry data using the BotTelemetryClient.track_event() method.\n This event name used is \"BotMessageDelete\".\n :param activity: Current activity sent from user.\n \"\"\"\n self.telemetry_client.track_event(\n TelemetryLoggerConstants.BOT_MSG_DELETE_EVENT,\n await self.fill_delete_event_properties(activity),\n )\n\n async def fill_receive_event_properties(\n self, activity: Activity, additional_properties: Dict[str, str] = None\n ) -> Dict[str, str]:\n \"\"\"Fills the event properties for the BotMessageReceived.\n Adheres to the LogPersonalInformation flag to filter Name, Text and Speak properties.\n :param activity: activity sent from user.\n :param additional_properties: Additional properties to add to the event.\n Additional properties can override \"stock\" properties.\n\n :return: A dictionary that is sent as \"Properties\" to\n BotTelemetryClient.track_event method for the BotMessageReceived event.\n \"\"\"\n properties = {\n TelemetryConstants.FROM_ID_PROPERTY: activity.from_property.id\n if activity.from_property\n else None,\n TelemetryConstants.CONVERSATION_NAME_PROPERTY: activity.conversation.name,\n TelemetryConstants.LOCALE_PROPERTY: activity.locale,\n TelemetryConstants.RECIPIENT_ID_PROPERTY: activity.recipient.id,\n TelemetryConstants.RECIPIENT_NAME_PROPERTY: activity.recipient.name,\n }\n\n if self.log_personal_information:\n if (\n activity.from_property\n and activity.from_property.name\n and activity.from_property.name.strip()\n ):\n properties[\n TelemetryConstants.FROM_NAME_PROPERTY\n ] = activity.from_property.name\n if activity.text and activity.text.strip():\n properties[TelemetryConstants.TEXT_PROPERTY] = activity.text\n if activity.speak and activity.speak.strip():\n properties[TelemetryConstants.SPEAK_PROPERTY] = activity.speak\n\n # Additional properties can override \"stock\" properties\n if additional_properties:\n for prop in additional_properties:\n properties[prop.key] = prop.value\n\n return properties\n\n async def fill_send_event_properties(\n self, activity: Activity, additional_properties: Dict[str, str] = None\n ) -> Dict[str, str]:\n \"\"\"Fills the event properties for the BotMessageSend.\n These properties are logged when an activity message is sent by the Bot to the user.\n :param activity: activity sent from user.\n :param additional_properties: Additional properties to add to the event.\n Additional properties can override \"stock\" properties.\n\n :return: A dictionary that is sent as \"Properties\" to the\n BotTelemetryClient.track_event method for the BotMessageSend event.\n \"\"\"\n properties = {\n TelemetryConstants.REPLY_ACTIVITY_ID_PROPERTY: activity.reply_to_id,\n TelemetryConstants.RECIPIENT_ID_PROPERTY: activity.recipient.id,\n TelemetryConstants.CONVERSATION_NAME_PROPERTY: activity.conversation.name,\n TelemetryConstants.LOCALE_PROPERTY: activity.locale,\n }\n\n # Use the LogPersonalInformation flag to toggle logging PII data, text and user name are common examples\n if self.log_personal_information:\n if activity.from_property.name and activity.from_property.name.strip():\n properties[\n TelemetryConstants.FROM_NAME_PROPERTY\n ] = activity.from_property.name\n if activity.text and activity.text.strip():\n properties[TelemetryConstants.TEXT_PROPERTY] = activity.text\n if activity.speak and activity.speak.strip():\n properties[TelemetryConstants.SPEAK_PROPERTY] = activity.speak\n\n # Additional properties can override \"stock\" properties\n if additional_properties:\n for prop in additional_properties:\n properties[prop.key] = prop.value\n\n return properties\n\n async def fill_update_event_properties(\n self, activity: Activity, additional_properties: Dict[str, str] = None\n ) -> Dict[str, str]:\n \"\"\"Fills the event properties for the BotMessageUpdate.\n These properties are logged when an activity message is updated by the Bot.\n For example, if a card is interacted with by the use, and the card needs\n to be updated to reflect some interaction.\n :param activity: activity sent from user.\n :param additional_properties: Additional properties to add to the event.\n Additional properties can override \"stock\" properties.\n\n :return: A dictionary that is sent as \"Properties\" to the\n BotTelemetryClient.track_event method for the BotMessageUpdate event.\n \"\"\"\n properties = {\n TelemetryConstants.RECIPIENT_ID_PROPERTY: activity.recipient.id,\n TelemetryConstants.CONVERSATION_ID_PROPERTY: activity.conversation.id,\n TelemetryConstants.CONVERSATION_NAME_PROPERTY: activity.conversation.name,\n TelemetryConstants.LOCALE_PROPERTY: activity.locale,\n }\n\n # Use the LogPersonalInformation flag to toggle logging PII data, text is a common examples\n if self.log_personal_information:\n if activity.text and activity.text.strip():\n properties[TelemetryConstants.TEXT_PROPERTY] = activity.text\n\n # Additional properties can override \"stock\" properties\n if additional_properties:\n for prop in additional_properties:\n properties[prop.key] = prop.value\n\n return properties\n\n async def fill_delete_event_properties(\n self, activity: Activity, additional_properties: Dict[str, str] = None\n ) -> Dict[str, str]:\n \"\"\"Fills the event properties for the BotMessageDelete.\n These properties are logged when an activity message is deleted by the Bot.\n :param activity: activity sent from user.\n :param additional_properties: Additional properties to add to the event.\n Additional properties can override \"stock\" properties.\n\n :return: A dictionary that is sent as \"Properties\" to the\n BotTelemetryClient.track_event method for the BotMessageUpdate event.\n \"\"\"\n properties = {\n TelemetryConstants.RECIPIENT_ID_PROPERTY: activity.recipient.id,\n TelemetryConstants.CONVERSATION_ID_PROPERTY: activity.conversation.id,\n TelemetryConstants.CONVERSATION_NAME_PROPERTY: activity.conversation.name,\n }\n\n # Additional properties can override \"stock\" properties\n if additional_properties:\n for prop in additional_properties:\n properties[prop.key] = prop.value\n\n return properties\n", "path": "libraries/botbuilder-core/botbuilder/core/telemetry_logger_middleware.py"}]}
| 4,057 | 331 |
gh_patches_debug_3893
|
rasdani/github-patches
|
git_diff
|
frappe__frappe-5292
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
[Blog] Image Meta Tags Not Generated Properly
#### Expected Behaviour
When meta tags are created for a blog, the image tags should contain link to the image.
#### Actual Behaviour
When the image meta tags are generated, it assigns "name" to it.
eg: domain.name/image instead of domain.name/filename.png
Frappé version:
v11.x.x-develop (ebde145) (develop)
</issue>
<code>
[start of frappe/website/context.py]
1 # Copyright (c) 2015, Frappe Technologies Pvt. Ltd. and Contributors
2 # MIT License. See license.txt
3
4 from __future__ import unicode_literals
5 import frappe, os, json
6
7 from frappe.website.doctype.website_settings.website_settings import get_website_settings
8 from frappe.website.router import get_page_context
9 from frappe.model.document import Document
10
11 def get_context(path, args=None):
12 if args and args.source:
13 context = args
14 else:
15 context = get_page_context(path)
16 if args:
17 context.update(args)
18
19 if hasattr(frappe.local, 'request'):
20 # for <body data-path=""> (remove leading slash)
21 # path could be overriden in render.resolve_from_map
22 context["path"] = frappe.local.request.path.strip('/ ')
23 else:
24 context["path"] = path
25
26 context.route = context.path
27
28 context = build_context(context)
29
30 # set using frappe.respond_as_web_page
31 if hasattr(frappe.local, 'response') and frappe.local.response.get('context'):
32 context.update(frappe.local.response.context)
33
34 return context
35
36 def update_controller_context(context, controller):
37 module = frappe.get_module(controller)
38
39 if module:
40 # get config fields
41 for prop in ("base_template_path", "template", "no_cache", "no_sitemap",
42 "condition_field"):
43 if hasattr(module, prop):
44 context[prop] = getattr(module, prop)
45
46 if hasattr(module, "get_context"):
47 try:
48 ret = module.get_context(context)
49 if ret:
50 context.update(ret)
51 except frappe.Redirect:
52 raise
53 except (frappe.PermissionError, frappe.DoesNotExistError):
54 raise
55 except:
56 if not frappe.flags.in_migrate:
57 frappe.errprint(frappe.utils.get_traceback())
58
59 if hasattr(module, "get_children"):
60 context.children = module.get_children(context)
61
62
63 def build_context(context):
64 """get_context method of doc or module is supposed to render
65 content templates and push it into context"""
66 context = frappe._dict(context)
67
68 if not "url_prefix" in context:
69 context.url_prefix = ""
70
71 if context.url_prefix and context.url_prefix[-1]!='/':
72 context.url_prefix += '/'
73
74 # for backward compatibility
75 context.docs_base_url = '/docs'
76
77 context.update(get_website_settings())
78 context.update(frappe.local.conf.get("website_context") or {})
79
80 # provide doc
81 if context.doc:
82 context.update(context.doc.as_dict())
83 context.update(context.doc.get_website_properties())
84
85 if not context.template:
86 context.template = context.doc.meta.get_web_template()
87
88 if hasattr(context.doc, "get_context"):
89 ret = context.doc.get_context(context)
90
91 if ret:
92 context.update(ret)
93
94 for prop in ("no_cache", "no_sitemap"):
95 if not prop in context:
96 context[prop] = getattr(context.doc, prop, False)
97
98 elif context.controller:
99 # controller based context
100 update_controller_context(context, context.controller)
101
102 # controller context extensions
103 context_controller_hooks = frappe.get_hooks("extend_website_page_controller_context") or {}
104 for controller, extension in context_controller_hooks.items():
105 if isinstance(extension, list):
106 for ext in extension:
107 if controller == context.controller:
108 update_controller_context(context, ext)
109 else:
110 update_controller_context(context, extension)
111
112 add_metatags(context)
113 add_sidebar_and_breadcrumbs(context)
114
115 # determine templates to be used
116 if not context.base_template_path:
117 app_base = frappe.get_hooks("base_template")
118 context.base_template_path = app_base[0] if app_base else "templates/base.html"
119
120 if context.title_prefix and context.title and not context.title.startswith(context.title_prefix):
121 context.title = '{0} - {1}'.format(context.title_prefix, context.title)
122
123 return context
124
125 def add_sidebar_and_breadcrumbs(context):
126 '''Add sidebar and breadcrumbs to context'''
127 from frappe.website.router import get_page_info_from_template
128 if context.show_sidebar:
129 context.no_cache = 1
130 add_sidebar_data(context)
131 else:
132 if context.basepath:
133 sidebar_json_path = os.path.join(context.basepath, '_sidebar.json')
134 if os.path.exists(sidebar_json_path):
135 with open(sidebar_json_path, 'r') as sidebarfile:
136 context.sidebar_items = json.loads(sidebarfile.read())
137 context.show_sidebar = 1
138
139 if context.add_breadcrumbs and not context.parents:
140 if context.basepath:
141 parent_path = os.path.dirname(context.path).rstrip('/')
142 page_info = get_page_info_from_template(parent_path)
143 if page_info:
144 context.parents = [dict(route=parent_path, title=page_info.title)]
145
146 def add_sidebar_data(context):
147 from frappe.utils.user import get_fullname_and_avatar
148 import frappe.www.list
149
150 if context.show_sidebar and context.website_sidebar:
151 context.sidebar_items = frappe.get_all('Website Sidebar Item',
152 filters=dict(parent=context.website_sidebar), fields=['title', 'route', '`group`'],
153 order_by='idx asc')
154
155 if not context.sidebar_items:
156 sidebar_items = frappe.cache().hget('portal_menu_items', frappe.session.user)
157 if sidebar_items == None:
158 sidebar_items = []
159 roles = frappe.get_roles()
160 portal_settings = frappe.get_doc('Portal Settings', 'Portal Settings')
161
162 def add_items(sidebar_items, items):
163 for d in items:
164 if d.get('enabled') and ((not d.get('role')) or d.get('role') in roles):
165 sidebar_items.append(d.as_dict() if isinstance(d, Document) else d)
166
167 if not portal_settings.hide_standard_menu:
168 add_items(sidebar_items, portal_settings.get('menu'))
169
170 if portal_settings.custom_menu:
171 add_items(sidebar_items, portal_settings.get('custom_menu'))
172
173 items_via_hooks = frappe.get_hooks('portal_menu_items')
174 if items_via_hooks:
175 for i in items_via_hooks: i['enabled'] = 1
176 add_items(sidebar_items, items_via_hooks)
177
178 frappe.cache().hset('portal_menu_items', frappe.session.user, sidebar_items)
179
180 context.sidebar_items = sidebar_items
181
182 info = get_fullname_and_avatar(frappe.session.user)
183 context["fullname"] = info.fullname
184 context["user_image"] = info.avatar
185 context["user"] = info.name
186
187
188 def add_metatags(context):
189 tags = context.get("metatags")
190 if tags:
191 if not "twitter:card" in tags:
192 tags["twitter:card"] = "summary_large_image"
193 if not "og:type" in tags:
194 tags["og:type"] = "article"
195 if tags.get("name"):
196 tags["og:title"] = tags["twitter:title"] = tags["name"]
197 if tags.get("description"):
198 tags["og:description"] = tags["twitter:description"] = tags["description"]
199 if tags.get("image"):
200 tags["og:image"] = tags["twitter:image:src"] = tags["image"] = frappe.utils.get_url("image")
201
202
[end of frappe/website/context.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/frappe/website/context.py b/frappe/website/context.py
--- a/frappe/website/context.py
+++ b/frappe/website/context.py
@@ -197,5 +197,4 @@
if tags.get("description"):
tags["og:description"] = tags["twitter:description"] = tags["description"]
if tags.get("image"):
- tags["og:image"] = tags["twitter:image:src"] = tags["image"] = frappe.utils.get_url("image")
-
+ tags["og:image"] = tags["twitter:image:src"] = tags["image"] = frappe.utils.get_url(tags.get("image"))
|
{"golden_diff": "diff --git a/frappe/website/context.py b/frappe/website/context.py\n--- a/frappe/website/context.py\n+++ b/frappe/website/context.py\n@@ -197,5 +197,4 @@\n \t\tif tags.get(\"description\"):\n \t\t\ttags[\"og:description\"] = tags[\"twitter:description\"] = tags[\"description\"]\n \t\tif tags.get(\"image\"):\n-\t\t\ttags[\"og:image\"] = tags[\"twitter:image:src\"] = tags[\"image\"] = frappe.utils.get_url(\"image\")\n- \n+\t\t\ttags[\"og:image\"] = tags[\"twitter:image:src\"] = tags[\"image\"] = frappe.utils.get_url(tags.get(\"image\"))\n", "issue": "[Blog] Image Meta Tags Not Generated Properly\n#### Expected Behaviour\r\nWhen meta tags are created for a blog, the image tags should contain link to the image.\r\n\r\n#### Actual Behaviour\r\nWhen the image meta tags are generated, it assigns \"name\" to it.\r\neg: domain.name/image instead of domain.name/filename.png\r\n\r\nFrapp\u00e9 version: \r\n v11.x.x-develop (ebde145) (develop)\r\n\n", "before_files": [{"content": "# Copyright (c) 2015, Frappe Technologies Pvt. Ltd. and Contributors\n# MIT License. See license.txt\n\nfrom __future__ import unicode_literals\nimport frappe, os, json\n\nfrom frappe.website.doctype.website_settings.website_settings import get_website_settings\nfrom frappe.website.router import get_page_context\nfrom frappe.model.document import Document\n\ndef get_context(path, args=None):\n\tif args and args.source:\n\t\tcontext = args\n\telse:\n\t\tcontext = get_page_context(path)\n\t\tif args:\n\t\t\tcontext.update(args)\n\n\tif hasattr(frappe.local, 'request'):\n\t\t# for <body data-path=\"\"> (remove leading slash)\n\t\t# path could be overriden in render.resolve_from_map\n\t\tcontext[\"path\"] = frappe.local.request.path.strip('/ ')\n\telse:\n\t\tcontext[\"path\"] = path\n\n\tcontext.route = context.path\n\n\tcontext = build_context(context)\n\n\t# set using frappe.respond_as_web_page\n\tif hasattr(frappe.local, 'response') and frappe.local.response.get('context'):\n\t\tcontext.update(frappe.local.response.context)\n\n\treturn context\n\ndef update_controller_context(context, controller):\n\tmodule = frappe.get_module(controller)\n\n\tif module:\n\t\t# get config fields\n\t\tfor prop in (\"base_template_path\", \"template\", \"no_cache\", \"no_sitemap\",\n\t\t\t\"condition_field\"):\n\t\t\tif hasattr(module, prop):\n\t\t\t\tcontext[prop] = getattr(module, prop)\n\n\t\tif hasattr(module, \"get_context\"):\n\t\t\ttry:\n\t\t\t\tret = module.get_context(context)\n\t\t\t\tif ret:\n\t\t\t\t\tcontext.update(ret)\n\t\t\texcept frappe.Redirect:\n\t\t\t\traise\n\t\t\texcept (frappe.PermissionError, frappe.DoesNotExistError):\n\t\t\t\traise\n\t\t\texcept:\n\t\t\t\tif not frappe.flags.in_migrate:\n\t\t\t\t\tfrappe.errprint(frappe.utils.get_traceback())\n\n\t\tif hasattr(module, \"get_children\"):\n\t\t\tcontext.children = module.get_children(context)\n\n\ndef build_context(context):\n\t\"\"\"get_context method of doc or module is supposed to render\n\t\tcontent templates and push it into context\"\"\"\n\tcontext = frappe._dict(context)\n\n\tif not \"url_prefix\" in context:\n\t\tcontext.url_prefix = \"\"\n\n\tif context.url_prefix and context.url_prefix[-1]!='/':\n\t\tcontext.url_prefix += '/'\n\n\t# for backward compatibility\n\tcontext.docs_base_url = '/docs'\n\n\tcontext.update(get_website_settings())\n\tcontext.update(frappe.local.conf.get(\"website_context\") or {})\n\n\t# provide doc\n\tif context.doc:\n\t\tcontext.update(context.doc.as_dict())\n\t\tcontext.update(context.doc.get_website_properties())\n\n\t\tif not context.template:\n\t\t\tcontext.template = context.doc.meta.get_web_template()\n\n\t\tif hasattr(context.doc, \"get_context\"):\n\t\t\tret = context.doc.get_context(context)\n\n\t\t\tif ret:\n\t\t\t\tcontext.update(ret)\n\n\t\tfor prop in (\"no_cache\", \"no_sitemap\"):\n\t\t\tif not prop in context:\n\t\t\t\tcontext[prop] = getattr(context.doc, prop, False)\n\n\telif context.controller:\n\t\t# controller based context\n\t\tupdate_controller_context(context, context.controller)\n\n\t\t# controller context extensions\n\t\tcontext_controller_hooks = frappe.get_hooks(\"extend_website_page_controller_context\") or {}\n\t\tfor controller, extension in context_controller_hooks.items():\n\t\t\tif isinstance(extension, list):\n\t\t\t\tfor ext in extension:\n\t\t\t\t\tif controller == context.controller:\n\t\t\t\t\t\tupdate_controller_context(context, ext)\n\t\t\telse:\n\t\t\t\tupdate_controller_context(context, extension)\n\n\tadd_metatags(context)\n\tadd_sidebar_and_breadcrumbs(context)\n\n\t# determine templates to be used\n\tif not context.base_template_path:\n\t\tapp_base = frappe.get_hooks(\"base_template\")\n\t\tcontext.base_template_path = app_base[0] if app_base else \"templates/base.html\"\n\n\tif context.title_prefix and context.title and not context.title.startswith(context.title_prefix):\n\t\tcontext.title = '{0} - {1}'.format(context.title_prefix, context.title)\n\n\treturn context\n\ndef add_sidebar_and_breadcrumbs(context):\n\t'''Add sidebar and breadcrumbs to context'''\n\tfrom frappe.website.router import get_page_info_from_template\n\tif context.show_sidebar:\n\t\tcontext.no_cache = 1\n\t\tadd_sidebar_data(context)\n\telse:\n\t\tif context.basepath:\n\t\t\tsidebar_json_path = os.path.join(context.basepath, '_sidebar.json')\n\t\t\tif os.path.exists(sidebar_json_path):\n\t\t\t\twith open(sidebar_json_path, 'r') as sidebarfile:\n\t\t\t\t\tcontext.sidebar_items = json.loads(sidebarfile.read())\n\t\t\t\t\tcontext.show_sidebar = 1\n\n\tif context.add_breadcrumbs and not context.parents:\n\t\tif context.basepath:\n\t\t\tparent_path = os.path.dirname(context.path).rstrip('/')\n\t\t\tpage_info = get_page_info_from_template(parent_path)\n\t\t\tif page_info:\n\t\t\t\tcontext.parents = [dict(route=parent_path, title=page_info.title)]\n\ndef add_sidebar_data(context):\n\tfrom frappe.utils.user import get_fullname_and_avatar\n\timport frappe.www.list\n\n\tif context.show_sidebar and context.website_sidebar:\n\t\tcontext.sidebar_items = frappe.get_all('Website Sidebar Item',\n\t\t\tfilters=dict(parent=context.website_sidebar), fields=['title', 'route', '`group`'],\n\t\t\torder_by='idx asc')\n\n\tif not context.sidebar_items:\n\t\tsidebar_items = frappe.cache().hget('portal_menu_items', frappe.session.user)\n\t\tif sidebar_items == None:\n\t\t\tsidebar_items = []\n\t\t\troles = frappe.get_roles()\n\t\t\tportal_settings = frappe.get_doc('Portal Settings', 'Portal Settings')\n\n\t\t\tdef add_items(sidebar_items, items):\n\t\t\t\tfor d in items:\n\t\t\t\t\tif d.get('enabled') and ((not d.get('role')) or d.get('role') in roles):\n\t\t\t\t\t\tsidebar_items.append(d.as_dict() if isinstance(d, Document) else d)\n\n\t\t\tif not portal_settings.hide_standard_menu:\n\t\t\t\tadd_items(sidebar_items, portal_settings.get('menu'))\n\n\t\t\tif portal_settings.custom_menu:\n\t\t\t\tadd_items(sidebar_items, portal_settings.get('custom_menu'))\n\n\t\t\titems_via_hooks = frappe.get_hooks('portal_menu_items')\n\t\t\tif items_via_hooks:\n\t\t\t\tfor i in items_via_hooks: i['enabled'] = 1\n\t\t\t\tadd_items(sidebar_items, items_via_hooks)\n\n\t\t\tfrappe.cache().hset('portal_menu_items', frappe.session.user, sidebar_items)\n\n\t\tcontext.sidebar_items = sidebar_items\n\n\tinfo = get_fullname_and_avatar(frappe.session.user)\n\tcontext[\"fullname\"] = info.fullname\n\tcontext[\"user_image\"] = info.avatar\n\tcontext[\"user\"] = info.name\n\n\ndef add_metatags(context):\n\ttags = context.get(\"metatags\")\n\tif tags:\n\t\tif not \"twitter:card\" in tags:\n\t\t\ttags[\"twitter:card\"] = \"summary_large_image\"\n\t\tif not \"og:type\" in tags:\n\t\t\ttags[\"og:type\"] = \"article\"\n\t\tif tags.get(\"name\"):\n\t\t\ttags[\"og:title\"] = tags[\"twitter:title\"] = tags[\"name\"]\n\t\tif tags.get(\"description\"):\n\t\t\ttags[\"og:description\"] = tags[\"twitter:description\"] = tags[\"description\"]\n\t\tif tags.get(\"image\"):\n\t\t\ttags[\"og:image\"] = tags[\"twitter:image:src\"] = tags[\"image\"] = frappe.utils.get_url(\"image\")\n \n", "path": "frappe/website/context.py"}]}
| 2,706 | 146 |
gh_patches_debug_1207
|
rasdani/github-patches
|
git_diff
|
pytorch__vision-2933
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Change default value of eps in FrozenBatchNorm to match BatchNorm
## ❓ Questions and Help
Hello
Loss is nan error occurs when I learn fast rcnn with resnext101 backbone
My code is as follows
```python
backbone = resnet_fpn_backbone('resnext101_32x8d', pretrained=True)
model = FasterRCNN(backbone, num_classes)
in_features = model.roi_heads.box_predictor.cls_score.in_features
model.roi_heads.box_predictor = FastRCNNPredictor(in_features, num_classes)
```
error message
```
Epoch: [0] [ 0/7208] eta: 1:27:42 lr: 0.000040 loss: 40613806080.0000 (40613806080.0000) loss_box_reg: 7979147264.0000 (7979147264.0000) loss_classifier: 11993160704.0000 (11993160704.0000) loss_objectness: 9486380032.0000 (9486380032.0000) loss_rpn_box_reg: 11155118080.0000 (11155118080.0000) time: 0.7301 data: 0.4106 max mem: 1241
Loss is nan, stopping training
```
When i change the backbone to resnet50 and resnet152, no error occrus.
### Please note that this issue tracker is not a help form and this issue will be closed.
We have a set of [listed resources available on the website](https://pytorch.org/resources). Our primary means of support is our discussion forum:
- [Discussion Forum](https://discuss.pytorch.org/)
</issue>
<code>
[start of torchvision/ops/misc.py]
1 """
2 helper class that supports empty tensors on some nn functions.
3
4 Ideally, add support directly in PyTorch to empty tensors in
5 those functions.
6
7 This can be removed once https://github.com/pytorch/pytorch/issues/12013
8 is implemented
9 """
10
11 import warnings
12 import torch
13 from torch import Tensor, Size
14 from torch.jit.annotations import List, Optional, Tuple
15
16
17 class Conv2d(torch.nn.Conv2d):
18 def __init__(self, *args, **kwargs):
19 super().__init__(*args, **kwargs)
20 warnings.warn(
21 "torchvision.ops.misc.Conv2d is deprecated and will be "
22 "removed in future versions, use torch.nn.Conv2d instead.", FutureWarning)
23
24
25 class ConvTranspose2d(torch.nn.ConvTranspose2d):
26 def __init__(self, *args, **kwargs):
27 super().__init__(*args, **kwargs)
28 warnings.warn(
29 "torchvision.ops.misc.ConvTranspose2d is deprecated and will be "
30 "removed in future versions, use torch.nn.ConvTranspose2d instead.", FutureWarning)
31
32
33 class BatchNorm2d(torch.nn.BatchNorm2d):
34 def __init__(self, *args, **kwargs):
35 super().__init__(*args, **kwargs)
36 warnings.warn(
37 "torchvision.ops.misc.BatchNorm2d is deprecated and will be "
38 "removed in future versions, use torch.nn.BatchNorm2d instead.", FutureWarning)
39
40
41 interpolate = torch.nn.functional.interpolate
42
43
44 # This is not in nn
45 class FrozenBatchNorm2d(torch.nn.Module):
46 """
47 BatchNorm2d where the batch statistics and the affine parameters
48 are fixed
49 """
50
51 def __init__(
52 self,
53 num_features: int,
54 eps: float = 0.,
55 n: Optional[int] = None,
56 ):
57 # n=None for backward-compatibility
58 if n is not None:
59 warnings.warn("`n` argument is deprecated and has been renamed `num_features`",
60 DeprecationWarning)
61 num_features = n
62 super(FrozenBatchNorm2d, self).__init__()
63 self.eps = eps
64 self.register_buffer("weight", torch.ones(num_features))
65 self.register_buffer("bias", torch.zeros(num_features))
66 self.register_buffer("running_mean", torch.zeros(num_features))
67 self.register_buffer("running_var", torch.ones(num_features))
68
69 def _load_from_state_dict(
70 self,
71 state_dict: dict,
72 prefix: str,
73 local_metadata: dict,
74 strict: bool,
75 missing_keys: List[str],
76 unexpected_keys: List[str],
77 error_msgs: List[str],
78 ):
79 num_batches_tracked_key = prefix + 'num_batches_tracked'
80 if num_batches_tracked_key in state_dict:
81 del state_dict[num_batches_tracked_key]
82
83 super(FrozenBatchNorm2d, self)._load_from_state_dict(
84 state_dict, prefix, local_metadata, strict,
85 missing_keys, unexpected_keys, error_msgs)
86
87 def forward(self, x: Tensor) -> Tensor:
88 # move reshapes to the beginning
89 # to make it fuser-friendly
90 w = self.weight.reshape(1, -1, 1, 1)
91 b = self.bias.reshape(1, -1, 1, 1)
92 rv = self.running_var.reshape(1, -1, 1, 1)
93 rm = self.running_mean.reshape(1, -1, 1, 1)
94 scale = w * (rv + self.eps).rsqrt()
95 bias = b - rm * scale
96 return x * scale + bias
97
98 def __repr__(self) -> str:
99 return f"{self.__class__.__name__}({self.weight.shape[0]}, eps={self.eps})"
100
[end of torchvision/ops/misc.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/torchvision/ops/misc.py b/torchvision/ops/misc.py
--- a/torchvision/ops/misc.py
+++ b/torchvision/ops/misc.py
@@ -51,7 +51,7 @@
def __init__(
self,
num_features: int,
- eps: float = 0.,
+ eps: float = 1e-5,
n: Optional[int] = None,
):
# n=None for backward-compatibility
|
{"golden_diff": "diff --git a/torchvision/ops/misc.py b/torchvision/ops/misc.py\n--- a/torchvision/ops/misc.py\n+++ b/torchvision/ops/misc.py\n@@ -51,7 +51,7 @@\n def __init__(\n self,\n num_features: int,\n- eps: float = 0.,\n+ eps: float = 1e-5,\n n: Optional[int] = None,\n ):\n # n=None for backward-compatibility\n", "issue": "Change default value of eps in FrozenBatchNorm to match BatchNorm\n## \u2753 Questions and Help\r\nHello\r\nLoss is nan error occurs when I learn fast rcnn with resnext101 backbone\r\nMy code is as follows\r\n```python\r\nbackbone = resnet_fpn_backbone('resnext101_32x8d', pretrained=True)\r\nmodel = FasterRCNN(backbone, num_classes)\r\nin_features = model.roi_heads.box_predictor.cls_score.in_features\r\nmodel.roi_heads.box_predictor = FastRCNNPredictor(in_features, num_classes)\r\n```\r\n\r\nerror message\r\n```\r\nEpoch: [0] [ 0/7208] eta: 1:27:42 lr: 0.000040 loss: 40613806080.0000 (40613806080.0000) loss_box_reg: 7979147264.0000 (7979147264.0000) loss_classifier: 11993160704.0000 (11993160704.0000) loss_objectness: 9486380032.0000 (9486380032.0000) loss_rpn_box_reg: 11155118080.0000 (11155118080.0000) time: 0.7301 data: 0.4106 max mem: 1241\r\nLoss is nan, stopping training\r\n```\r\n\r\nWhen i change the backbone to resnet50 and resnet152, no error occrus.\r\n### Please note that this issue tracker is not a help form and this issue will be closed.\r\n\r\nWe have a set of [listed resources available on the website](https://pytorch.org/resources). Our primary means of support is our discussion forum:\r\n\r\n- [Discussion Forum](https://discuss.pytorch.org/)\r\n\n", "before_files": [{"content": "\"\"\"\nhelper class that supports empty tensors on some nn functions.\n\nIdeally, add support directly in PyTorch to empty tensors in\nthose functions.\n\nThis can be removed once https://github.com/pytorch/pytorch/issues/12013\nis implemented\n\"\"\"\n\nimport warnings\nimport torch\nfrom torch import Tensor, Size\nfrom torch.jit.annotations import List, Optional, Tuple\n\n\nclass Conv2d(torch.nn.Conv2d):\n def __init__(self, *args, **kwargs):\n super().__init__(*args, **kwargs)\n warnings.warn(\n \"torchvision.ops.misc.Conv2d is deprecated and will be \"\n \"removed in future versions, use torch.nn.Conv2d instead.\", FutureWarning)\n\n\nclass ConvTranspose2d(torch.nn.ConvTranspose2d):\n def __init__(self, *args, **kwargs):\n super().__init__(*args, **kwargs)\n warnings.warn(\n \"torchvision.ops.misc.ConvTranspose2d is deprecated and will be \"\n \"removed in future versions, use torch.nn.ConvTranspose2d instead.\", FutureWarning)\n\n\nclass BatchNorm2d(torch.nn.BatchNorm2d):\n def __init__(self, *args, **kwargs):\n super().__init__(*args, **kwargs)\n warnings.warn(\n \"torchvision.ops.misc.BatchNorm2d is deprecated and will be \"\n \"removed in future versions, use torch.nn.BatchNorm2d instead.\", FutureWarning)\n\n\ninterpolate = torch.nn.functional.interpolate\n\n\n# This is not in nn\nclass FrozenBatchNorm2d(torch.nn.Module):\n \"\"\"\n BatchNorm2d where the batch statistics and the affine parameters\n are fixed\n \"\"\"\n\n def __init__(\n self,\n num_features: int,\n eps: float = 0.,\n n: Optional[int] = None,\n ):\n # n=None for backward-compatibility\n if n is not None:\n warnings.warn(\"`n` argument is deprecated and has been renamed `num_features`\",\n DeprecationWarning)\n num_features = n\n super(FrozenBatchNorm2d, self).__init__()\n self.eps = eps\n self.register_buffer(\"weight\", torch.ones(num_features))\n self.register_buffer(\"bias\", torch.zeros(num_features))\n self.register_buffer(\"running_mean\", torch.zeros(num_features))\n self.register_buffer(\"running_var\", torch.ones(num_features))\n\n def _load_from_state_dict(\n self,\n state_dict: dict,\n prefix: str,\n local_metadata: dict,\n strict: bool,\n missing_keys: List[str],\n unexpected_keys: List[str],\n error_msgs: List[str],\n ):\n num_batches_tracked_key = prefix + 'num_batches_tracked'\n if num_batches_tracked_key in state_dict:\n del state_dict[num_batches_tracked_key]\n\n super(FrozenBatchNorm2d, self)._load_from_state_dict(\n state_dict, prefix, local_metadata, strict,\n missing_keys, unexpected_keys, error_msgs)\n\n def forward(self, x: Tensor) -> Tensor:\n # move reshapes to the beginning\n # to make it fuser-friendly\n w = self.weight.reshape(1, -1, 1, 1)\n b = self.bias.reshape(1, -1, 1, 1)\n rv = self.running_var.reshape(1, -1, 1, 1)\n rm = self.running_mean.reshape(1, -1, 1, 1)\n scale = w * (rv + self.eps).rsqrt()\n bias = b - rm * scale\n return x * scale + bias\n\n def __repr__(self) -> str:\n return f\"{self.__class__.__name__}({self.weight.shape[0]}, eps={self.eps})\"\n", "path": "torchvision/ops/misc.py"}]}
| 2,017 | 107 |
gh_patches_debug_17358
|
rasdani/github-patches
|
git_diff
|
weecology__retriever-548
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
download_only w/path fails to use path argument when checking for file
When `download_only` checks to see if the file already exists before copying it, it ignores the path argument. This means that:
```
retriever download MoM2003 -p testdata
```
will keep overwriting the file in `testdata` if it exists, and it will not copy the file to `testdata` if the file exists in `.`.
Fixes this is probably just a little logic improvement in the `final_cleanup` function of `download_only`.
</issue>
<code>
[start of engines/download_only.py]
1 from __future__ import print_function
2 from builtins import object
3 import os
4 import platform
5 import shutil
6 import inspect
7
8 from retriever.lib.engine import filename_from_url
9 from retriever.lib.models import Engine, no_cleanup
10 from retriever import DATA_DIR, HOME_DIR
11
12
13 class DummyConnection(object):
14
15 def cursor(self):
16 pass
17
18 def commit(self):
19 pass
20
21 def rollback(self):
22 pass
23
24 def close(self):
25 pass
26
27
28 class DummyCursor(DummyConnection):
29 pass
30
31
32 class engine(Engine):
33 """Engine instance for writing data to a CSV file."""
34 name = "Download Only"
35 abbreviation = "download"
36 required_opts = [("path",
37 "File path to copy data files",
38 "./"),
39 ("subdir",
40 "Keep the subdirectories for archived files",
41 False)
42 ]
43
44 def table_exists(self, dbname, tablename):
45 """Checks if the file to be downloaded already exists"""
46 try:
47 tablename = self.table_name(name=tablename, dbname=dbname)
48 return os.path.exists(tablename)
49 except:
50 return False
51
52 def get_connection(self):
53 """Gets the db connection."""
54 self.get_input()
55 return DummyConnection()
56
57 def final_cleanup(self):
58 """Copies downloaded files to desired directory
59
60 Copies the downloaded files into the chosen directory unless files with the same
61 name already exist in the directory.
62
63 """
64 if hasattr(self, "all_files"):
65 for file_name in self.all_files:
66 file_path, file_name_nopath = os.path.split(file_name)
67 subdir = os.path.split(file_path)[1] if self.opts['subdir'] else ''
68 dest_path = os.path.join(self.opts['path'], subdir)
69 if os.path.abspath(file_path) == os.path.abspath(os.path.join(DATA_DIR, subdir)):
70 print ("%s is already in the working directory" %
71 file_name_nopath)
72 print("Keeping existing copy.")
73 else:
74 print("Copying %s from %s" % (file_name_nopath, file_path))
75 if os.path.isdir(dest_path):
76 try:
77 shutil.copy(file_name, dest_path)
78 except:
79 print("Couldn't copy file to %s" % dest_path)
80 else:
81 try:
82 print("Creating directory %s" % dest_path)
83 os.makedirs(dest_path)
84 shutil.copy(file_name, dest_path)
85 except:
86 print("Couldn't create directory %s" % dest_path)
87 self.all_files = set()
88
89 def auto_create_table(self, table, url=None, filename=None, pk=None):
90 """Download the file if it doesn't exist"""
91 if url and not filename:
92 filename = filename_from_url(url)
93
94 if url and not self.find_file(filename):
95 # If the file doesn't exist, download it
96 self.download_file(url, filename)
97
98 def insert_data_from_url(self, url):
99 """Insert data from a web resource"""
100 filename = filename_from_url(url)
101 find = self.find_file(filename)
102 if not find:
103 self.create_raw_data_dir()
104 self.download_file(url, filename)
105
106 def find_file(self, filename):
107 """Checks for the given file and adds it to the list of all files"""
108 result = Engine.find_file(self, filename)
109 if not hasattr(self, "all_files"):
110 self.all_files = set()
111 if result:
112 self.all_files.add(result)
113 return result
114
115 def register_files(self, filenames):
116 """Identify a list of files to be moved by the download
117
118 When downloading archives with multiple files the engine needs to be
119 informed of all of the file names so that it can move them.
120
121 """
122 full_filenames = {self.find_file(filename) for filename in filenames
123 if self.find_file(filename)}
124 self.all_files = self.all_files.union(full_filenames)
125
126
127 # replace all other methods with a function that does nothing
128 def dummy_method(self, *args, **kwargs):
129 pass
130
131
132 methods = inspect.getmembers(engine, predicate=inspect.ismethod)
133 keep_methods = {'table_exists',
134 'get_connection',
135 'final_cleanup',
136 'auto_create_table',
137 'insert_data_from_url',
138 }
139 remove_methods = ['insert_data_from_file']
140 for name, method in methods:
141 if (name not in keep_methods and
142 'download' not in name and
143 'file' not in name and
144 'dir' not in name):
145 setattr(engine, name, dummy_method)
146 for name in remove_methods:
147 setattr(engine, name, dummy_method)
148
[end of engines/download_only.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/engines/download_only.py b/engines/download_only.py
--- a/engines/download_only.py
+++ b/engines/download_only.py
@@ -66,7 +66,9 @@
file_path, file_name_nopath = os.path.split(file_name)
subdir = os.path.split(file_path)[1] if self.opts['subdir'] else ''
dest_path = os.path.join(self.opts['path'], subdir)
- if os.path.abspath(file_path) == os.path.abspath(os.path.join(DATA_DIR, subdir)):
+ if os.path.isfile(os.path.join(dest_path, file_name_nopath)):
+ print ("File already exists at specified location")
+ elif os.path.abspath(file_path) == os.path.abspath(os.path.join(DATA_DIR, subdir)):
print ("%s is already in the working directory" %
file_name_nopath)
print("Keeping existing copy.")
|
{"golden_diff": "diff --git a/engines/download_only.py b/engines/download_only.py\n--- a/engines/download_only.py\n+++ b/engines/download_only.py\n@@ -66,7 +66,9 @@\n file_path, file_name_nopath = os.path.split(file_name)\n subdir = os.path.split(file_path)[1] if self.opts['subdir'] else ''\n dest_path = os.path.join(self.opts['path'], subdir)\n- if os.path.abspath(file_path) == os.path.abspath(os.path.join(DATA_DIR, subdir)):\n+ if os.path.isfile(os.path.join(dest_path, file_name_nopath)):\n+ print (\"File already exists at specified location\")\n+ elif os.path.abspath(file_path) == os.path.abspath(os.path.join(DATA_DIR, subdir)):\n print (\"%s is already in the working directory\" %\n file_name_nopath)\n print(\"Keeping existing copy.\")\n", "issue": "download_only w/path fails to use path argument when checking for file\nWhen `download_only` checks to see if the file already exists before copying it, it ignores the path argument. This means that:\n\n```\nretriever download MoM2003 -p testdata\n```\n\nwill keep overwriting the file in `testdata` if it exists, and it will not copy the file to `testdata` if the file exists in `.`.\n\nFixes this is probably just a little logic improvement in the `final_cleanup` function of `download_only`.\n\n", "before_files": [{"content": "from __future__ import print_function\nfrom builtins import object\nimport os\nimport platform\nimport shutil\nimport inspect\n\nfrom retriever.lib.engine import filename_from_url\nfrom retriever.lib.models import Engine, no_cleanup\nfrom retriever import DATA_DIR, HOME_DIR\n\n\nclass DummyConnection(object):\n\n def cursor(self):\n pass\n\n def commit(self):\n pass\n\n def rollback(self):\n pass\n\n def close(self):\n pass\n\n\nclass DummyCursor(DummyConnection):\n pass\n\n\nclass engine(Engine):\n \"\"\"Engine instance for writing data to a CSV file.\"\"\"\n name = \"Download Only\"\n abbreviation = \"download\"\n required_opts = [(\"path\",\n \"File path to copy data files\",\n \"./\"),\n (\"subdir\",\n \"Keep the subdirectories for archived files\",\n False)\n ]\n\n def table_exists(self, dbname, tablename):\n \"\"\"Checks if the file to be downloaded already exists\"\"\"\n try:\n tablename = self.table_name(name=tablename, dbname=dbname)\n return os.path.exists(tablename)\n except:\n return False\n\n def get_connection(self):\n \"\"\"Gets the db connection.\"\"\"\n self.get_input()\n return DummyConnection()\n\n def final_cleanup(self):\n \"\"\"Copies downloaded files to desired directory\n\n Copies the downloaded files into the chosen directory unless files with the same\n name already exist in the directory.\n\n \"\"\"\n if hasattr(self, \"all_files\"):\n for file_name in self.all_files:\n file_path, file_name_nopath = os.path.split(file_name)\n subdir = os.path.split(file_path)[1] if self.opts['subdir'] else ''\n dest_path = os.path.join(self.opts['path'], subdir)\n if os.path.abspath(file_path) == os.path.abspath(os.path.join(DATA_DIR, subdir)):\n print (\"%s is already in the working directory\" %\n file_name_nopath)\n print(\"Keeping existing copy.\")\n else:\n print(\"Copying %s from %s\" % (file_name_nopath, file_path))\n if os.path.isdir(dest_path):\n try:\n shutil.copy(file_name, dest_path)\n except:\n print(\"Couldn't copy file to %s\" % dest_path)\n else:\n try:\n print(\"Creating directory %s\" % dest_path)\n os.makedirs(dest_path)\n shutil.copy(file_name, dest_path)\n except:\n print(\"Couldn't create directory %s\" % dest_path)\n self.all_files = set()\n\n def auto_create_table(self, table, url=None, filename=None, pk=None):\n \"\"\"Download the file if it doesn't exist\"\"\"\n if url and not filename:\n filename = filename_from_url(url)\n\n if url and not self.find_file(filename):\n # If the file doesn't exist, download it\n self.download_file(url, filename)\n\n def insert_data_from_url(self, url):\n \"\"\"Insert data from a web resource\"\"\"\n filename = filename_from_url(url)\n find = self.find_file(filename)\n if not find:\n self.create_raw_data_dir()\n self.download_file(url, filename)\n\n def find_file(self, filename):\n \"\"\"Checks for the given file and adds it to the list of all files\"\"\"\n result = Engine.find_file(self, filename)\n if not hasattr(self, \"all_files\"):\n self.all_files = set()\n if result:\n self.all_files.add(result)\n return result\n\n def register_files(self, filenames):\n \"\"\"Identify a list of files to be moved by the download\n\n When downloading archives with multiple files the engine needs to be\n informed of all of the file names so that it can move them.\n\n \"\"\"\n full_filenames = {self.find_file(filename) for filename in filenames\n if self.find_file(filename)}\n self.all_files = self.all_files.union(full_filenames)\n\n\n# replace all other methods with a function that does nothing\ndef dummy_method(self, *args, **kwargs):\n pass\n\n\nmethods = inspect.getmembers(engine, predicate=inspect.ismethod)\nkeep_methods = {'table_exists',\n 'get_connection',\n 'final_cleanup',\n 'auto_create_table',\n 'insert_data_from_url',\n }\nremove_methods = ['insert_data_from_file']\nfor name, method in methods:\n if (name not in keep_methods and\n 'download' not in name and\n 'file' not in name and\n 'dir' not in name):\n setattr(engine, name, dummy_method)\nfor name in remove_methods:\n setattr(engine, name, dummy_method)\n", "path": "engines/download_only.py"}]}
| 1,963 | 192 |
gh_patches_debug_20458
|
rasdani/github-patches
|
git_diff
|
Zeroto521__my-data-toolkit-444
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
BUG: fix jupyer notebook can’t render
<!--
Thanks for contributing a pull request!
Please follow these standard acronyms to start the commit message:
- ENH: enhancement
- BUG: bug fix
- DOC: documentation
- TYP: type annotations
- TST: addition or modification of tests
- MAINT: maintenance commit (refactoring, typos, etc.)
- BLD: change related to building
- REL: related to releasing
- API: an (incompatible) API change
- DEP: deprecate something, or remove a deprecated object
- DEV: development tool or utility
- REV: revert an earlier commit
- PERF: performance improvement
- BOT: always commit via a bot
- CI: related to CI or CD
- CLN: Code cleanup
-->
- [x] closes #436 and https://my-data-toolkit.readthedocs.io/en/v0.0.10/guide/tips_about_getattr.html
- [ ] whatsnew entry
</issue>
<code>
[start of doc/source/conf.py]
1 # Configuration file for the Sphinx documentation builder.
2 #
3 # This file only contains a selection of the most common options.
4 # For a full list see the documentation:
5 # https://www.sphinx-doc.org/en/master/usage/configuration.html
6
7
8 # -- Project information -----------------------------------------------------
9
10 # The version info for the project you're documenting, acts as replacement for
11 # |version| and |release|, also used in various other places throughout the
12 # built documents.
13 from __future__ import annotations
14
15 import inspect
16 import os
17 import sys
18 from datetime import datetime
19
20 import dtoolkit
21
22 version = release = dtoolkit.__version__
23 version = version.replace(".dev0", "")
24 version = version.replace(".post0", "")
25
26
27 project = "DToolKit"
28 author = "Zero <@Zeroto521>"
29 copyright = f"2021-{datetime.now().year}, {author}" # pylint: disable=redefined-builtin
30 github_url = "https://github.com/Zeroto521/my-data-toolkit"
31
32
33 # -- General configuration ---------------------------------------------------
34
35 # Add any Sphinx extension module names here, as strings. They can be
36 # extensions coming with Sphinx (named 'sphinx.ext.*') or your custom
37 # ones.
38 extensions = [
39 "myst_parser",
40 "numpydoc",
41 "sphinx.ext.autodoc",
42 "sphinx.ext.autosummary",
43 "sphinx.ext.autosectionlabel",
44 "sphinx.ext.extlinks",
45 "sphinx.ext.intersphinx",
46 "sphinx.ext.linkcode",
47 "sphinx_toggleprompt",
48 "IPython.sphinxext.ipython_console_highlighting",
49 "IPython.sphinxext.ipython_directive",
50 "nbsphinx",
51 ]
52
53 # The suffix of source filenames.
54 source_suffix = [".rst", ".md"]
55
56 # Add any paths that contain templates here, relative to this directory.
57 templates_path = ["_templates"]
58
59 # List of patterns, relative to source directory, that match files and
60 # directories to ignore when looking for source files.
61 # This pattern also affects html_static_path and html_extra_path.
62 exclude_patterns = []
63
64
65 # -- Options for HTML output -------------------------------------------------
66
67 # The theme to use for HTML and HTML Help pages. See the documentation for
68 # a list of builtin themes.
69 html_theme = "pydata_sphinx_theme"
70
71 # Theme options are theme-specific and customize the look and feel of a theme
72 # further. For a list of options available for each theme, see the
73 # documentation.
74 html_theme_options = {
75 "search_bar_position": "sidebar",
76 "github_url": github_url,
77 "icon_links": [
78 {
79 "name": "PyPI",
80 "url": "https://pypi.org/project/my-data-toolkit",
81 "icon": "fas fa-box",
82 },
83 ],
84 }
85
86 # Add any paths that contain custom static files (such as style sheets) here,
87 # relative to this directory. They are copied after the builtin static files,
88 # so a file named "default.css" will overwrite the builtin "default.css".
89 html_static_path = ["_static"]
90
91
92 # Add redirect for previously existing pages, each item is like `(from_old, to_new)`
93
94 moved_pages = [
95 ("py-modindex", "reference"),
96 ]
97
98 html_additional_pages = {page[0]: "redirect.html" for page in moved_pages}
99
100 html_context = {"redirects": dict(moved_pages)}
101
102
103 nbsphinx_prolog = r"""
104 {% set docname = env.doc2path(env.docname, base=None) %}
105
106 .. tip::
107
108 This page was generated from `{{ docname }}`__.
109
110 __ :ghlink:`my-data-toolkit/blob/main/doc/source/{{ docname }}`
111 """
112
113
114 # --Options for sphinx extensions -----------------------------------------------
115
116 # connect docs in other projects
117 intersphinx_mapping = {
118 "python": ("http://docs.python.org/", None),
119 "pandas": ("https://pandas.pydata.org/pandas-docs/stable/", None),
120 "numpy": ("https://numpy.org/doc/stable/", None),
121 "sklearn": ("https://scikit-learn.org/stable/", None),
122 "geopandas": ("https://geopandas.readthedocs.io/en/stable/", None),
123 "shapely": ("https://shapely.readthedocs.io/en/stable/", None),
124 "pyproj": ("https://pyproj4.github.io/pyproj/stable/", None),
125 "pygeos": ("https://pygeos.readthedocs.io/en/stable/", None),
126 }
127
128 # extlinks alias
129 extlinks = {
130 "issue": (f"{github_url}/issues/%s", "issue#"),
131 "pr": (f"{github_url}/issues/%s", "pr#"),
132 "user": (r"https://github.com/%s", "@"),
133 # github's link
134 "ghlink": (r"https://github.com/%s", None),
135 }
136
137 myst_enable_extensions = [
138 "colon_fence",
139 ]
140
141 autosummary_generate = True
142
143 nbsphinx_execute = "always"
144 nbsphinx_allow_errors = True
145
146
147 # based on pandas doc/source/conf.py
148 def linkcode_resolve(domain: str, info: dict[str, str]) -> str | None:
149 """
150 Determine the URL corresponding to Python object
151 """
152
153 if domain != "py":
154 return None
155
156 modname = info["module"]
157 fullname = info["fullname"]
158
159 submod = sys.modules.get(modname)
160 if submod is None:
161 return None
162
163 obj = submod
164 for part in fullname.split("."):
165 try:
166 obj = getattr(obj, part)
167 except AttributeError:
168 return None
169
170 try:
171 fn = inspect.getsourcefile(inspect.unwrap(obj))
172 except TypeError:
173 fn = None
174
175 if not fn:
176 return None
177
178 # to fix these doc doesn't exist in dtoolkit
179 if project.lower() not in fn:
180 return None
181
182 try:
183 source, lineno = inspect.getsourcelines(obj)
184 except OSError:
185 lineno = None
186
187 linespec = f"#L{lineno}-L{lineno + len(source) - 1}" if lineno else ""
188 fn = os.path.relpath(fn, start=os.path.dirname(dtoolkit.__file__))
189
190 base_link = f"{github_url}/blob/" + "{branch}" + f"/dtoolkit/{fn}{linespec}"
191 if "post" in version:
192 return base_link.format(branch="main")
193
194 return base_link.format(branch=f"v{version}")
195
[end of doc/source/conf.py]
[start of dtoolkit/geoaccessor/accessor.py]
1 import geopandas as gpd
2 from pandas.core.accessor import _register_accessor
3 from pandas.util._decorators import doc
4
5
6 @doc(klass=":class:`~geopandas.GeoSeries`")
7 def register_geoseries_accessor(name: str):
8 """
9 Register a custom accessor on {klass} objects.
10
11 This is a temparatory solution to hook method into
12 :class:`~geopandas.GeoSeries` or :class:`~geopandas.GeoDataFrame`.
13 If `geopandas#1952`_ done, it would be removed from
14 :mod:`dtoolkit.geoaccessor`.
15
16 .. _geopandas#1952: :ghlink:`geopandas/geopandas/pull/1952`
17
18 Parameters
19 ----------
20 name : str
21 Name under which the accessor should be registered. A warning is issued
22 if this name conflicts with a preexisting attribute.
23
24 Returns
25 -------
26 callable
27 A class decorator.
28
29 See Also
30 --------
31 register_geoseries_accessor
32 register_geodataframe_accessor
33 dtoolkit.geoaccessor.register_geoseries_method
34 dtoolkit.geoaccessor.register_geodataframe_method
35
36 Notes
37 -----
38 When accessed, your accessor will be initialized with the geopandas object
39 the user is interacting with. So the signature must be::
40
41 def __init__(self, geopandas_object): # noqa: E999
42 ...
43
44 For consistency with geopandas methods, you should raise an
45 ``AttributeError`` if the data passed to your accessor has an incorrect
46 dtype.
47
48 >>> import geopandas as gpd
49 >>> gpd.GeoSeries().dt
50 Traceback (most recent call last):
51 ...
52 AttributeError: Can only use .dt accessor with datetimelike values
53
54 Examples
55 --------
56 In your library code::
57
58 from pygeos import count_coordinates, from_shapely
59
60 @register_geodataframe_accessor("coords")
61 @register_geoseries_accessor("coords")
62 class CoordinateAccessor:
63 def __init__(self, gpd_obj):
64 self._obj = gpd_obj
65 self.geometry = gpd_obj.geometry
66
67 @property
68 def count_coordinates(self):
69 # Counts the number of coordinate pairs in geometry
70
71 return self.geometry.apply(count_coordinates(from_shapely(x)))
72
73 Back in an interactive IPython session:
74
75 .. code-block:: ipython
76
77 In [1]: import geopandas as gpd
78
79 In [2]: s = gpd.GeoSeries.from_wkt(["POINT (0 0)", "POINT (1 1)", None])
80
81 In [3]: s
82 Out[3]:
83 0 POINT (0.00000 0.00000)
84 1 POINT (1.00000 1.00000)
85 2 None
86 dtype: geometry
87
88 In [4]: s.coords.count_coordinates
89 Out[4]:
90 0 1
91 1 1
92 2 0
93 dtype: int64
94
95 In [5]: d = s.to_frame("geometry")
96 Out[5]:
97 geometry
98 0 POINT (0.00000 0.00000)
99 1 POINT (1.00000 1.00000)
100 2 None
101
102 In [6]: d.coords.count_coordinates
103 Out[6]:
104 0 1
105 1 1
106 2 0
107 Name: geometry, dtype: int64
108 """
109
110 return _register_accessor(name, gpd.GeoSeries)
111
112
113 @doc(register_geoseries_accessor, klass=":class:`~geopandas.GeoDataFrame`")
114 def register_geodataframe_accessor(name: str):
115
116 return _register_accessor(name, gpd.GeoDataFrame)
117
[end of dtoolkit/geoaccessor/accessor.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/doc/source/conf.py b/doc/source/conf.py
--- a/doc/source/conf.py
+++ b/doc/source/conf.py
@@ -107,7 +107,7 @@
This page was generated from `{{ docname }}`__.
- __ :ghlink:`my-data-toolkit/blob/main/doc/source/{{ docname }}`
+ __ https://github.com/zeroto521/my-data-toolkit/blob/main/doc/source/{{ docname }}
"""
@@ -130,8 +130,6 @@
"issue": (f"{github_url}/issues/%s", "issue#"),
"pr": (f"{github_url}/issues/%s", "pr#"),
"user": (r"https://github.com/%s", "@"),
- # github's link
- "ghlink": (r"https://github.com/%s", None),
}
myst_enable_extensions = [
diff --git a/dtoolkit/geoaccessor/accessor.py b/dtoolkit/geoaccessor/accessor.py
--- a/dtoolkit/geoaccessor/accessor.py
+++ b/dtoolkit/geoaccessor/accessor.py
@@ -13,7 +13,7 @@
If `geopandas#1952`_ done, it would be removed from
:mod:`dtoolkit.geoaccessor`.
- .. _geopandas#1952: :ghlink:`geopandas/geopandas/pull/1952`
+ .. _geopandas#1952: https://github.com/geopandas/geopandas/pull/1952
Parameters
----------
|
{"golden_diff": "diff --git a/doc/source/conf.py b/doc/source/conf.py\n--- a/doc/source/conf.py\n+++ b/doc/source/conf.py\n@@ -107,7 +107,7 @@\n \n This page was generated from `{{ docname }}`__.\n \n- __ :ghlink:`my-data-toolkit/blob/main/doc/source/{{ docname }}`\n+ __ https://github.com/zeroto521/my-data-toolkit/blob/main/doc/source/{{ docname }}\n \"\"\"\n \n \n@@ -130,8 +130,6 @@\n \"issue\": (f\"{github_url}/issues/%s\", \"issue#\"),\n \"pr\": (f\"{github_url}/issues/%s\", \"pr#\"),\n \"user\": (r\"https://github.com/%s\", \"@\"),\n- # github's link\n- \"ghlink\": (r\"https://github.com/%s\", None),\n }\n \n myst_enable_extensions = [\ndiff --git a/dtoolkit/geoaccessor/accessor.py b/dtoolkit/geoaccessor/accessor.py\n--- a/dtoolkit/geoaccessor/accessor.py\n+++ b/dtoolkit/geoaccessor/accessor.py\n@@ -13,7 +13,7 @@\n If `geopandas#1952`_ done, it would be removed from\n :mod:`dtoolkit.geoaccessor`.\n \n- .. _geopandas#1952: :ghlink:`geopandas/geopandas/pull/1952`\n+ .. _geopandas#1952: https://github.com/geopandas/geopandas/pull/1952\n \n Parameters\n ----------\n", "issue": "BUG: fix jupyer notebook can\u2019t render\n<!--\r\nThanks for contributing a pull request!\r\n\r\nPlease follow these standard acronyms to start the commit message:\r\n\r\n- ENH: enhancement\r\n- BUG: bug fix\r\n- DOC: documentation\r\n- TYP: type annotations\r\n- TST: addition or modification of tests\r\n- MAINT: maintenance commit (refactoring, typos, etc.)\r\n- BLD: change related to building\r\n- REL: related to releasing\r\n- API: an (incompatible) API change\r\n- DEP: deprecate something, or remove a deprecated object\r\n- DEV: development tool or utility\r\n- REV: revert an earlier commit\r\n- PERF: performance improvement\r\n- BOT: always commit via a bot\r\n- CI: related to CI or CD\r\n- CLN: Code cleanup\r\n-->\r\n\r\n- [x] closes #436 and https://my-data-toolkit.readthedocs.io/en/v0.0.10/guide/tips_about_getattr.html\r\n- [ ] whatsnew entry\r\n\n", "before_files": [{"content": "# Configuration file for the Sphinx documentation builder.\n#\n# This file only contains a selection of the most common options.\n# For a full list see the documentation:\n# https://www.sphinx-doc.org/en/master/usage/configuration.html\n\n\n# -- Project information -----------------------------------------------------\n\n# The version info for the project you're documenting, acts as replacement for\n# |version| and |release|, also used in various other places throughout the\n# built documents.\nfrom __future__ import annotations\n\nimport inspect\nimport os\nimport sys\nfrom datetime import datetime\n\nimport dtoolkit\n\nversion = release = dtoolkit.__version__\nversion = version.replace(\".dev0\", \"\")\nversion = version.replace(\".post0\", \"\")\n\n\nproject = \"DToolKit\"\nauthor = \"Zero <@Zeroto521>\"\ncopyright = f\"2021-{datetime.now().year}, {author}\" # pylint: disable=redefined-builtin\ngithub_url = \"https://github.com/Zeroto521/my-data-toolkit\"\n\n\n# -- General configuration ---------------------------------------------------\n\n# Add any Sphinx extension module names here, as strings. They can be\n# extensions coming with Sphinx (named 'sphinx.ext.*') or your custom\n# ones.\nextensions = [\n \"myst_parser\",\n \"numpydoc\",\n \"sphinx.ext.autodoc\",\n \"sphinx.ext.autosummary\",\n \"sphinx.ext.autosectionlabel\",\n \"sphinx.ext.extlinks\",\n \"sphinx.ext.intersphinx\",\n \"sphinx.ext.linkcode\",\n \"sphinx_toggleprompt\",\n \"IPython.sphinxext.ipython_console_highlighting\",\n \"IPython.sphinxext.ipython_directive\",\n \"nbsphinx\",\n]\n\n# The suffix of source filenames.\nsource_suffix = [\".rst\", \".md\"]\n\n# Add any paths that contain templates here, relative to this directory.\ntemplates_path = [\"_templates\"]\n\n# List of patterns, relative to source directory, that match files and\n# directories to ignore when looking for source files.\n# This pattern also affects html_static_path and html_extra_path.\nexclude_patterns = []\n\n\n# -- Options for HTML output -------------------------------------------------\n\n# The theme to use for HTML and HTML Help pages. See the documentation for\n# a list of builtin themes.\nhtml_theme = \"pydata_sphinx_theme\"\n\n# Theme options are theme-specific and customize the look and feel of a theme\n# further. For a list of options available for each theme, see the\n# documentation.\nhtml_theme_options = {\n \"search_bar_position\": \"sidebar\",\n \"github_url\": github_url,\n \"icon_links\": [\n {\n \"name\": \"PyPI\",\n \"url\": \"https://pypi.org/project/my-data-toolkit\",\n \"icon\": \"fas fa-box\",\n },\n ],\n}\n\n# Add any paths that contain custom static files (such as style sheets) here,\n# relative to this directory. They are copied after the builtin static files,\n# so a file named \"default.css\" will overwrite the builtin \"default.css\".\nhtml_static_path = [\"_static\"]\n\n\n# Add redirect for previously existing pages, each item is like `(from_old, to_new)`\n\nmoved_pages = [\n (\"py-modindex\", \"reference\"),\n]\n\nhtml_additional_pages = {page[0]: \"redirect.html\" for page in moved_pages}\n\nhtml_context = {\"redirects\": dict(moved_pages)}\n\n\nnbsphinx_prolog = r\"\"\"\n{% set docname = env.doc2path(env.docname, base=None) %}\n\n.. tip::\n\n This page was generated from `{{ docname }}`__.\n\n __ :ghlink:`my-data-toolkit/blob/main/doc/source/{{ docname }}`\n\"\"\"\n\n\n# --Options for sphinx extensions -----------------------------------------------\n\n# connect docs in other projects\nintersphinx_mapping = {\n \"python\": (\"http://docs.python.org/\", None),\n \"pandas\": (\"https://pandas.pydata.org/pandas-docs/stable/\", None),\n \"numpy\": (\"https://numpy.org/doc/stable/\", None),\n \"sklearn\": (\"https://scikit-learn.org/stable/\", None),\n \"geopandas\": (\"https://geopandas.readthedocs.io/en/stable/\", None),\n \"shapely\": (\"https://shapely.readthedocs.io/en/stable/\", None),\n \"pyproj\": (\"https://pyproj4.github.io/pyproj/stable/\", None),\n \"pygeos\": (\"https://pygeos.readthedocs.io/en/stable/\", None),\n}\n\n# extlinks alias\nextlinks = {\n \"issue\": (f\"{github_url}/issues/%s\", \"issue#\"),\n \"pr\": (f\"{github_url}/issues/%s\", \"pr#\"),\n \"user\": (r\"https://github.com/%s\", \"@\"),\n # github's link\n \"ghlink\": (r\"https://github.com/%s\", None),\n}\n\nmyst_enable_extensions = [\n \"colon_fence\",\n]\n\nautosummary_generate = True\n\nnbsphinx_execute = \"always\"\nnbsphinx_allow_errors = True\n\n\n# based on pandas doc/source/conf.py\ndef linkcode_resolve(domain: str, info: dict[str, str]) -> str | None:\n \"\"\"\n Determine the URL corresponding to Python object\n \"\"\"\n\n if domain != \"py\":\n return None\n\n modname = info[\"module\"]\n fullname = info[\"fullname\"]\n\n submod = sys.modules.get(modname)\n if submod is None:\n return None\n\n obj = submod\n for part in fullname.split(\".\"):\n try:\n obj = getattr(obj, part)\n except AttributeError:\n return None\n\n try:\n fn = inspect.getsourcefile(inspect.unwrap(obj))\n except TypeError:\n fn = None\n\n if not fn:\n return None\n\n # to fix these doc doesn't exist in dtoolkit\n if project.lower() not in fn:\n return None\n\n try:\n source, lineno = inspect.getsourcelines(obj)\n except OSError:\n lineno = None\n\n linespec = f\"#L{lineno}-L{lineno + len(source) - 1}\" if lineno else \"\"\n fn = os.path.relpath(fn, start=os.path.dirname(dtoolkit.__file__))\n\n base_link = f\"{github_url}/blob/\" + \"{branch}\" + f\"/dtoolkit/{fn}{linespec}\"\n if \"post\" in version:\n return base_link.format(branch=\"main\")\n\n return base_link.format(branch=f\"v{version}\")\n", "path": "doc/source/conf.py"}, {"content": "import geopandas as gpd\nfrom pandas.core.accessor import _register_accessor\nfrom pandas.util._decorators import doc\n\n\n@doc(klass=\":class:`~geopandas.GeoSeries`\")\ndef register_geoseries_accessor(name: str):\n \"\"\"\n Register a custom accessor on {klass} objects.\n\n This is a temparatory solution to hook method into\n :class:`~geopandas.GeoSeries` or :class:`~geopandas.GeoDataFrame`.\n If `geopandas#1952`_ done, it would be removed from\n :mod:`dtoolkit.geoaccessor`.\n\n .. _geopandas#1952: :ghlink:`geopandas/geopandas/pull/1952`\n\n Parameters\n ----------\n name : str\n Name under which the accessor should be registered. A warning is issued\n if this name conflicts with a preexisting attribute.\n\n Returns\n -------\n callable\n A class decorator.\n\n See Also\n --------\n register_geoseries_accessor\n register_geodataframe_accessor\n dtoolkit.geoaccessor.register_geoseries_method\n dtoolkit.geoaccessor.register_geodataframe_method\n\n Notes\n -----\n When accessed, your accessor will be initialized with the geopandas object\n the user is interacting with. So the signature must be::\n\n def __init__(self, geopandas_object): # noqa: E999\n ...\n\n For consistency with geopandas methods, you should raise an\n ``AttributeError`` if the data passed to your accessor has an incorrect\n dtype.\n\n >>> import geopandas as gpd\n >>> gpd.GeoSeries().dt\n Traceback (most recent call last):\n ...\n AttributeError: Can only use .dt accessor with datetimelike values\n\n Examples\n --------\n In your library code::\n\n from pygeos import count_coordinates, from_shapely\n\n @register_geodataframe_accessor(\"coords\")\n @register_geoseries_accessor(\"coords\")\n class CoordinateAccessor:\n def __init__(self, gpd_obj):\n self._obj = gpd_obj\n self.geometry = gpd_obj.geometry\n\n @property\n def count_coordinates(self):\n # Counts the number of coordinate pairs in geometry\n\n return self.geometry.apply(count_coordinates(from_shapely(x)))\n\n Back in an interactive IPython session:\n\n .. code-block:: ipython\n\n In [1]: import geopandas as gpd\n\n In [2]: s = gpd.GeoSeries.from_wkt([\"POINT (0 0)\", \"POINT (1 1)\", None])\n\n In [3]: s\n Out[3]:\n 0 POINT (0.00000 0.00000)\n 1 POINT (1.00000 1.00000)\n 2 None\n dtype: geometry\n\n In [4]: s.coords.count_coordinates\n Out[4]:\n 0 1\n 1 1\n 2 0\n dtype: int64\n\n In [5]: d = s.to_frame(\"geometry\")\n Out[5]:\n geometry\n 0 POINT (0.00000 0.00000)\n 1 POINT (1.00000 1.00000)\n 2 None\n\n In [6]: d.coords.count_coordinates\n Out[6]:\n 0 1\n 1 1\n 2 0\n Name: geometry, dtype: int64\n \"\"\"\n\n return _register_accessor(name, gpd.GeoSeries)\n\n\n@doc(register_geoseries_accessor, klass=\":class:`~geopandas.GeoDataFrame`\")\ndef register_geodataframe_accessor(name: str):\n\n return _register_accessor(name, gpd.GeoDataFrame)\n", "path": "dtoolkit/geoaccessor/accessor.py"}]}
| 3,805 | 365 |
gh_patches_debug_6830
|
rasdani/github-patches
|
git_diff
|
platformsh__platformsh-docs-2432
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
🐛 Allow searching for code strings with separators
### Where on docs.platform.sh should be changed?
The search
### What exactly should be updated?
Searching for strings with separators like `X-Frame-Options` and `memory_ratio` don't show the results for pages that have those strings directly. Putting quotes around the strings doesn't help.
We'd like people to be able to get info on specific properties and strings, so the search should return these results.
### Additional context
_No response_
</issue>
<code>
[start of search/main.py]
1 import os
2 import glob
3 import json
4 import meilisearch
5 from platformshconfig import Config
6
7 class Search:
8 def __init__(self):
9 self.default = {
10 "host": "http://127.0.0.1",
11 "key": None,
12 "port": 7700
13 }
14
15 self.scrape_dir = "output"
16 self.scrape_config = "config/scrape.json"
17 self.docs_index = "docs"
18 self.primaryKey = "documentId"
19 self.index_name = "Docs"
20
21 # Below are Platform.sh custom settings for how the search engine functions.
22
23 # Data available to the dropdown React app in docs, used to fill out autocomplete results.
24 self.displayed_attributes = ['keywords', 'title', 'text', 'url', 'site', 'section']
25 # Data actually searchable by our queries.
26 self.searchable_attributes = ['keywords', 'title', 'pageUrl', 'section', 'text', 'url']
27
28 # Show results for one query with the listed pages, when they by default would not show up as best results.
29 # Note: these aren't automatically two-way, which is why they're all defined twice.
30 self.synonyms = {
31 "cron": ["crons"],
32 "crons": ["cron tasks", "cron jobs"],
33 "e-mail": ["email"],
34 "routes.yaml": ["routes"],
35 "routes": ["routes.yaml"],
36 "services": ["services.yaml"],
37 "services.yaml": ["services"],
38 "application": [".platform.app.yaml", "app.yaml", "applications.yaml"],
39 ".platform.app.yaml": ["application"],
40 "app.yaml": ["application"],
41 "applications.yaml": ["application", "multi-app"],
42 "multi-app": ["applications.yaml"],
43 "regions": ["public ip addresses"],
44 "public ip addresses": ["regions"],
45 "ssl": ["https", "tls"],
46 "https": ["ssl"],
47 "auth": ["authentication", "access control"], # Only needs to be one way since we don't use "auth" in the docs
48 }
49
50 # Ranking rules:
51 #
52 # - Default order: ["words", "typo", "proximity", "attribute", "sort", "exactness"]
53 #
54 # - words: number of times query is in document (greater number gets priority)
55 # - typo: fewer typos > more typos
56 # - proximity: smaller distance between multiple occurences of query in same document > larger distances
57 # - attribute: sorted according to order of importance of attributes (searchable_attributes). terms in
58 # more important attributes first.
59 # - sort: queries are sorted at query time
60 # - exactness: similarity of matched words in document with query
61
62 self.ranking_rules = ["rank:asc", "attribute", "typo", "words", "proximity", "exactness"]
63
64 self.updated_settings = {
65 "rankingRules": self.ranking_rules,
66 "searchableAttributes": self.searchable_attributes,
67 "displayedAttributes": self.displayed_attributes
68 }
69
70 # Group results by page
71 self.distinct_attribute = "pageUrl"
72
73 def getConnectionString(self):
74 """
75 Sets the Meilisearch host string, depending on the environment.
76
77 Returns:
78 string: Meilisearch host string.
79 """
80 if os.environ.get('PORT'):
81 return "{}:{}".format(self.default["host"], os.environ['PORT'])
82 else:
83 return "{}:{}".format(self.default["host"], self.default["port"])
84
85 def getMasterKey(self):
86 """
87 Retrieves the Meilisearch master key, either from the Platform.sh environment or locally.
88 """
89 config = Config()
90 if config.is_valid_platform():
91 return config.projectEntropy
92 elif os.environ.get("MEILI_MASTER_KEY"):
93 return os.environ["MEILI_MASTER_KEY"]
94 else:
95 return self.default["key"]
96
97 def add_documents(self, index):
98 """
99 Cycle through the individual site indexes in /outputs so their individual documents can be added to Meilisearch.
100 """
101 documents = [f for f in glob.glob("{}/*.json".format(self.scrape_dir))]
102 for doc in documents:
103 self.add(doc, index)
104
105 def add(self, doc, index):
106 """
107 Add an individual site's index to the Meilisearch service.
108 """
109 with open(doc) as scraped_index:
110 data = json.load(scraped_index)
111 index.add_documents(data)
112
113 def update(self):
114 """
115 Updates the Meilisearch index.
116 """
117 # Create a Meilisearch client.
118 client = meilisearch.Client(self.getConnectionString(), self.getMasterKey())
119
120 # Delete previous index
121 if len(client.get_indexes()):
122 client.index(self.docs_index).delete()
123
124 # Create a new index
125 create_index_task = client.create_index(uid=self.docs_index, options={'primaryKey': self.primaryKey, 'uid': self.index_name})
126
127 client.wait_for_task(create_index_task['uid'])
128
129 index = client.get_index(create_index_task['indexUid'])
130
131 # Add synonyms for the index
132 index.update_synonyms(self.synonyms)
133
134 # Update its settings: what can be searched, what's displayable, and how results should be ranked.
135 index.update_settings(self.updated_settings)
136
137 # Update distinct attribute.
138 index.update_distinct_attribute(self.distinct_attribute)
139
140 # Add documents to the index
141 self.add_documents(index)
142
143 if __name__ == "__main__":
144 meili = Search()
145 meili.update()
146
[end of search/main.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/search/main.py b/search/main.py
--- a/search/main.py
+++ b/search/main.py
@@ -59,7 +59,7 @@
# - sort: queries are sorted at query time
# - exactness: similarity of matched words in document with query
- self.ranking_rules = ["rank:asc", "attribute", "typo", "words", "proximity", "exactness"]
+ self.ranking_rules = ["rank:asc", "exactness", "attribute", "proximity", "typo", "words"]
self.updated_settings = {
"rankingRules": self.ranking_rules,
|
{"golden_diff": "diff --git a/search/main.py b/search/main.py\n--- a/search/main.py\n+++ b/search/main.py\n@@ -59,7 +59,7 @@\n # - sort: queries are sorted at query time\n # - exactness: similarity of matched words in document with query\n \n- self.ranking_rules = [\"rank:asc\", \"attribute\", \"typo\", \"words\", \"proximity\", \"exactness\"]\n+ self.ranking_rules = [\"rank:asc\", \"exactness\", \"attribute\", \"proximity\", \"typo\", \"words\"]\n \n self.updated_settings = {\n \"rankingRules\": self.ranking_rules,\n", "issue": "\ud83d\udc1b Allow searching for code strings with separators\n### Where on docs.platform.sh should be changed?\r\n\r\nThe search\r\n\r\n### What exactly should be updated?\r\n\r\nSearching for strings with separators like `X-Frame-Options` and `memory_ratio` don't show the results for pages that have those strings directly. Putting quotes around the strings doesn't help.\r\n\r\nWe'd like people to be able to get info on specific properties and strings, so the search should return these results.\r\n\r\n### Additional context\r\n\r\n_No response_\n", "before_files": [{"content": "import os\nimport glob\nimport json\nimport meilisearch\nfrom platformshconfig import Config\n\nclass Search:\n def __init__(self):\n self.default = {\n \"host\": \"http://127.0.0.1\",\n \"key\": None,\n \"port\": 7700\n }\n\n self.scrape_dir = \"output\"\n self.scrape_config = \"config/scrape.json\"\n self.docs_index = \"docs\"\n self.primaryKey = \"documentId\"\n self.index_name = \"Docs\"\n\n # Below are Platform.sh custom settings for how the search engine functions.\n\n # Data available to the dropdown React app in docs, used to fill out autocomplete results.\n self.displayed_attributes = ['keywords', 'title', 'text', 'url', 'site', 'section']\n # Data actually searchable by our queries.\n self.searchable_attributes = ['keywords', 'title', 'pageUrl', 'section', 'text', 'url']\n\n # Show results for one query with the listed pages, when they by default would not show up as best results.\n # Note: these aren't automatically two-way, which is why they're all defined twice.\n self.synonyms = {\n \"cron\": [\"crons\"],\n \"crons\": [\"cron tasks\", \"cron jobs\"],\n \"e-mail\": [\"email\"],\n \"routes.yaml\": [\"routes\"],\n \"routes\": [\"routes.yaml\"],\n \"services\": [\"services.yaml\"],\n \"services.yaml\": [\"services\"],\n \"application\": [\".platform.app.yaml\", \"app.yaml\", \"applications.yaml\"],\n \".platform.app.yaml\": [\"application\"],\n \"app.yaml\": [\"application\"],\n \"applications.yaml\": [\"application\", \"multi-app\"],\n \"multi-app\": [\"applications.yaml\"],\n \"regions\": [\"public ip addresses\"],\n \"public ip addresses\": [\"regions\"],\n \"ssl\": [\"https\", \"tls\"],\n \"https\": [\"ssl\"],\n \"auth\": [\"authentication\", \"access control\"], # Only needs to be one way since we don't use \"auth\" in the docs\n }\n\n # Ranking rules:\n #\n # - Default order: [\"words\", \"typo\", \"proximity\", \"attribute\", \"sort\", \"exactness\"]\n #\n # - words: number of times query is in document (greater number gets priority)\n # - typo: fewer typos > more typos\n # - proximity: smaller distance between multiple occurences of query in same document > larger distances\n # - attribute: sorted according to order of importance of attributes (searchable_attributes). terms in\n # more important attributes first.\n # - sort: queries are sorted at query time\n # - exactness: similarity of matched words in document with query\n\n self.ranking_rules = [\"rank:asc\", \"attribute\", \"typo\", \"words\", \"proximity\", \"exactness\"]\n\n self.updated_settings = {\n \"rankingRules\": self.ranking_rules,\n \"searchableAttributes\": self.searchable_attributes,\n \"displayedAttributes\": self.displayed_attributes\n }\n\n # Group results by page\n self.distinct_attribute = \"pageUrl\"\n\n def getConnectionString(self):\n \"\"\"\n Sets the Meilisearch host string, depending on the environment.\n\n Returns:\n string: Meilisearch host string.\n \"\"\"\n if os.environ.get('PORT'):\n return \"{}:{}\".format(self.default[\"host\"], os.environ['PORT'])\n else:\n return \"{}:{}\".format(self.default[\"host\"], self.default[\"port\"])\n\n def getMasterKey(self):\n \"\"\"\n Retrieves the Meilisearch master key, either from the Platform.sh environment or locally.\n \"\"\"\n config = Config()\n if config.is_valid_platform():\n return config.projectEntropy\n elif os.environ.get(\"MEILI_MASTER_KEY\"):\n return os.environ[\"MEILI_MASTER_KEY\"]\n else:\n return self.default[\"key\"]\n\n def add_documents(self, index):\n \"\"\"\n Cycle through the individual site indexes in /outputs so their individual documents can be added to Meilisearch.\n \"\"\"\n documents = [f for f in glob.glob(\"{}/*.json\".format(self.scrape_dir))]\n for doc in documents:\n self.add(doc, index)\n\n def add(self, doc, index):\n \"\"\"\n Add an individual site's index to the Meilisearch service.\n \"\"\"\n with open(doc) as scraped_index:\n data = json.load(scraped_index)\n index.add_documents(data)\n\n def update(self):\n \"\"\"\n Updates the Meilisearch index.\n \"\"\"\n # Create a Meilisearch client.\n client = meilisearch.Client(self.getConnectionString(), self.getMasterKey())\n\n # Delete previous index\n if len(client.get_indexes()):\n client.index(self.docs_index).delete()\n\n # Create a new index\n create_index_task = client.create_index(uid=self.docs_index, options={'primaryKey': self.primaryKey, 'uid': self.index_name})\n\n client.wait_for_task(create_index_task['uid'])\n\n index = client.get_index(create_index_task['indexUid'])\n\n # Add synonyms for the index\n index.update_synonyms(self.synonyms)\n\n # Update its settings: what can be searched, what's displayable, and how results should be ranked.\n index.update_settings(self.updated_settings)\n\n # Update distinct attribute.\n index.update_distinct_attribute(self.distinct_attribute)\n\n # Add documents to the index\n self.add_documents(index)\n\nif __name__ == \"__main__\":\n meili = Search()\n meili.update()\n", "path": "search/main.py"}]}
| 2,173 | 142 |
gh_patches_debug_7495
|
rasdani/github-patches
|
git_diff
|
kymatio__kymatio-890
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Make NumPy the default frontend
Since we promised earlier:
```
/home/jenkins/workspace/kymatio_dev/kymatio/frontend/entry.py:20: DeprecationWarning: Torch frontend is currently the default, but NumPy will become the default in the next version.
warnings.warn("Torch frontend is currently the default, but NumPy will become the default in the next"
```
</issue>
<code>
[start of kymatio/frontend/entry.py]
1 import logging
2 import warnings
3 import importlib
4
5
6 class ScatteringEntry(object):
7 def __init__(self, *args, **kwargs):
8 self.name = kwargs['name']
9 self.class_name = kwargs['class_name']
10 kwargs.pop('name')
11 kwargs.pop('class_name')
12
13 frontend_suffixes = {'torch' : 'Torch',
14 'numpy' : 'NumPy',
15 'tensorflow' : 'TensorFlow',
16 'keras': 'Keras',
17 'sklearn': 'Transformer'}
18
19 if 'frontend' not in kwargs:
20 warnings.warn("Torch frontend is currently the default, but NumPy will become the default in the next"
21 " version.", DeprecationWarning)
22 frontend = 'torch'
23 else:
24 frontend = kwargs['frontend'].lower()
25 kwargs.pop('frontend')
26
27 frontends = list(frontend_suffixes.keys())
28
29 if frontend not in frontends:
30 raise RuntimeError('The frontend \'%s\" is not valid. Must be '
31 'one of \'%s\', or \'%s\'.' %
32 (frontend, '\', \''.join(frontends[:-1]),
33 frontends[-1]))
34
35 try:
36 module = importlib.import_module('kymatio.' + self.class_name + '.frontend.' + frontend + '_frontend')
37
38 # Create frontend-specific class name by inserting frontend name
39 # after `Scattering`.
40 frontend = frontend_suffixes[frontend]
41
42 class_name = self.__class__.__name__
43
44 base_name = class_name[:-len('Entry*D')]
45 dim_suffix = class_name[-len('*D'):]
46
47 class_name = base_name + frontend + dim_suffix
48
49 self.__class__ = getattr(module, class_name)
50 self.__init__(*args, **kwargs)
51 except Exception as e:
52 raise e from RuntimeError('\nThe frontend \'' + frontend + '\' could not be correctly imported.')
53
54 logging.info('The ' + self.name + ' frontend ' + frontend + ' was imported.')
55
56
57 __all__ = ['ScatteringEntry']
58
[end of kymatio/frontend/entry.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/kymatio/frontend/entry.py b/kymatio/frontend/entry.py
--- a/kymatio/frontend/entry.py
+++ b/kymatio/frontend/entry.py
@@ -17,9 +17,7 @@
'sklearn': 'Transformer'}
if 'frontend' not in kwargs:
- warnings.warn("Torch frontend is currently the default, but NumPy will become the default in the next"
- " version.", DeprecationWarning)
- frontend = 'torch'
+ frontend = 'numpy'
else:
frontend = kwargs['frontend'].lower()
kwargs.pop('frontend')
|
{"golden_diff": "diff --git a/kymatio/frontend/entry.py b/kymatio/frontend/entry.py\n--- a/kymatio/frontend/entry.py\n+++ b/kymatio/frontend/entry.py\n@@ -17,9 +17,7 @@\n 'sklearn': 'Transformer'}\n \n if 'frontend' not in kwargs:\n- warnings.warn(\"Torch frontend is currently the default, but NumPy will become the default in the next\"\n- \" version.\", DeprecationWarning)\n- frontend = 'torch'\n+ frontend = 'numpy'\n else:\n frontend = kwargs['frontend'].lower()\n kwargs.pop('frontend')\n", "issue": "Make NumPy the default frontend\nSince we promised earlier:\r\n\r\n```\r\n /home/jenkins/workspace/kymatio_dev/kymatio/frontend/entry.py:20: DeprecationWarning: Torch frontend is currently the default, but NumPy will become the default in the next version.\r\n warnings.warn(\"Torch frontend is currently the default, but NumPy will become the default in the next\"\r\n```\n", "before_files": [{"content": "import logging\nimport warnings\nimport importlib\n\n\nclass ScatteringEntry(object):\n def __init__(self, *args, **kwargs):\n self.name = kwargs['name']\n self.class_name = kwargs['class_name']\n kwargs.pop('name')\n kwargs.pop('class_name')\n\n frontend_suffixes = {'torch' : 'Torch',\n 'numpy' : 'NumPy',\n 'tensorflow' : 'TensorFlow',\n 'keras': 'Keras',\n 'sklearn': 'Transformer'}\n\n if 'frontend' not in kwargs:\n warnings.warn(\"Torch frontend is currently the default, but NumPy will become the default in the next\"\n \" version.\", DeprecationWarning)\n frontend = 'torch'\n else:\n frontend = kwargs['frontend'].lower()\n kwargs.pop('frontend')\n\n frontends = list(frontend_suffixes.keys())\n\n if frontend not in frontends:\n raise RuntimeError('The frontend \\'%s\\\" is not valid. Must be '\n 'one of \\'%s\\', or \\'%s\\'.' %\n (frontend, '\\', \\''.join(frontends[:-1]),\n frontends[-1]))\n\n try:\n module = importlib.import_module('kymatio.' + self.class_name + '.frontend.' + frontend + '_frontend')\n\n # Create frontend-specific class name by inserting frontend name\n # after `Scattering`.\n frontend = frontend_suffixes[frontend]\n\n class_name = self.__class__.__name__\n\n base_name = class_name[:-len('Entry*D')]\n dim_suffix = class_name[-len('*D'):]\n\n class_name = base_name + frontend + dim_suffix\n\n self.__class__ = getattr(module, class_name)\n self.__init__(*args, **kwargs)\n except Exception as e:\n raise e from RuntimeError('\\nThe frontend \\'' + frontend + '\\' could not be correctly imported.')\n\n logging.info('The ' + self.name + ' frontend ' + frontend + ' was imported.')\n\n\n__all__ = ['ScatteringEntry']\n", "path": "kymatio/frontend/entry.py"}]}
| 1,168 | 136 |
gh_patches_debug_14010
|
rasdani/github-patches
|
git_diff
|
vaexio__vaex-404
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Failing to open arrow file format & misleading error message
Vaex fails to open arrow file format. It confusingly expects a file written not in the arrow file-format but in arrow stream-format.
If this is a non-negotiable vaex constraint, a less misleading error message might help very confused novices...
```
import pandas as pd
import pyarrow as pa
import numpy as np
import vaex
df = pd.DataFrame(
{
'col1': range(5)
}
)
table = pa.Table.from_pandas(df)
with pa.OSFile('test2.arrow', 'wb') as sink:
with pa.RecordBatchFileWriter(sink, table.schema) as writer:
writer.write_table(table)
with pa.OSFile('test2.arrow', 'rb') as source:
df = pa.ipc.open_file(source).read_pandas()
df = vaex.open('test2.arrow')
```
Error messages:
```
ERROR:MainThread:vaex:error opening 'test2.arrow'
---------------------------------------------------------------------------
ArrowInvalid Traceback (most recent call last)
in
16 with pa.OSFile('test2.arrow', 'rb') as source:
17 df = pa.ipc.open_file(source).read_pandas()
---> 18 df = vaex.open('test2.arrow')
Z:\Systemdateien\Miniconda3\envs\finance\lib\site-packages\vaex\__init__.py in open(path, convert, shuffle, copy_index, *args, **kwargs)
189 ds = from_csv(path, copy_index=copy_index, **kwargs)
190 else:
--> 191 ds = vaex.file.open(path, *args, **kwargs)
192 if convert and ds:
193 ds.export_hdf5(filename_hdf5, shuffle=shuffle)
Z:\Systemdateien\Miniconda3\envs\finance\lib\site-packages\vaex\file\__init__.py in open(path, *args, **kwargs)
28 for opener in opener_classes:
29 if opener.can_open(path, *args, **kwargs):
---> 30 return opener.open(path, *args, **kwargs)
31 if hdf5:
32 openers.extend(hdf5.dataset.dataset_type_map.items())
Z:\Systemdateien\Miniconda3\envs\finance\lib\site-packages\vaex_arrow\opener.py in open(path, *args, **kwargs)
9 def open(path, *args, **kwargs):
10 from .dataset import DatasetArrow
---> 11 return DatasetArrow(path, *args, **kwargs)
12
13 class ParquetOpener:
Z:\Systemdateien\Miniconda3\envs\finance\lib\site-packages\vaex_arrow\dataset.py in __init__(self, filename, table, write)
18 self._write = write
19 if table is None:
---> 20 self._load()
21 else:
22 self._load_table(table)
Z:\Systemdateien\Miniconda3\envs\finance\lib\site-packages\vaex_arrow\dataset.py in _load(self)
24 def _load(self):
25 source = pa.memory_map(self.path)
---> 26 reader = pa.ipc.open_stream(source)
27 table = pa.Table.from_batches([b for b in reader])
28 self._load_table(table)
Z:\Systemdateien\Miniconda3\envs\finance\lib\site-packages\pyarrow\ipc.py in open_stream(source)
123 reader : RecordBatchStreamReader
124 """
--> 125 return RecordBatchStreamReader(source)
126
127
Z:\Systemdateien\Miniconda3\envs\finance\lib\site-packages\pyarrow\ipc.py in __init__(self, source)
58 """
59 def __init__(self, source):
---> 60 self._open(source)
61
62
Z:\Systemdateien\Miniconda3\envs\finance\lib\site-packages\pyarrow\ipc.pxi in pyarrow.lib._RecordBatchStreamReader._open()
Z:\Systemdateien\Miniconda3\envs\finance\lib\site-packages\pyarrow\error.pxi in pyarrow.lib.check_status()
ArrowInvalid: Expected to read 1330795073 metadata bytes, but only read 1474
```
</issue>
<code>
[start of packages/vaex-arrow/vaex_arrow/dataset.py]
1 __author__ = 'maartenbreddels'
2 import logging
3
4 import pyarrow as pa
5 import pyarrow.parquet as pq
6
7 import vaex.dataset
8 import vaex.file.other
9 from .convert import column_from_arrow_array
10 logger = logging.getLogger("vaex_arrow")
11
12
13 class DatasetArrow(vaex.dataset.DatasetLocal):
14 """Implements storage using arrow"""
15
16 def __init__(self, filename=None, table=None, write=False):
17 super(DatasetArrow, self).__init__(name=filename, path=filename, column_names=[])
18 self._write = write
19 if table is None:
20 self._load()
21 else:
22 self._load_table(table)
23
24 def _load(self):
25 source = pa.memory_map(self.path)
26 reader = pa.ipc.open_stream(source)
27 table = pa.Table.from_batches([b for b in reader])
28 self._load_table(table)
29
30 def _load_table(self, table):
31 self._length_unfiltered = self._length_original = table.num_rows
32 self._index_end = self._length_original = table.num_rows
33 for col in table.columns:
34 name = col.name
35 # TODO: keep the arrow columns, and support and test chunks
36 arrow_array = col.data.chunks[0]
37 column = column_from_arrow_array(arrow_array)
38
39 self.columns[name] = column
40 self.column_names.append(name)
41 self._save_assign_expression(name, vaex.expression.Expression(self, name))
42
43
44 @classmethod
45 def can_open(cls, path, *args, **kwargs):
46 return path.rpartition('.')[2] == 'arrow'
47
48 @classmethod
49 def get_options(cls, path):
50 return []
51
52 @classmethod
53 def option_to_args(cls, option):
54 return []
55
56 class DatasetParquet(DatasetArrow):
57 def _load(self):
58 # might not be optimal, but it works, we can always see if we can
59 # do mmapping later on
60 table = pq.read_table(self.path)
61 self._load_table(table)
62
63 vaex.file.other.dataset_type_map["arrow"] = DatasetArrow
64 vaex.file.other.dataset_type_map["parquet"] = DatasetParquet
65
66
[end of packages/vaex-arrow/vaex_arrow/dataset.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/packages/vaex-arrow/vaex_arrow/dataset.py b/packages/vaex-arrow/vaex_arrow/dataset.py
--- a/packages/vaex-arrow/vaex_arrow/dataset.py
+++ b/packages/vaex-arrow/vaex_arrow/dataset.py
@@ -23,8 +23,18 @@
def _load(self):
source = pa.memory_map(self.path)
- reader = pa.ipc.open_stream(source)
- table = pa.Table.from_batches([b for b in reader])
+ try:
+ # first we try if it opens as stream
+ reader = pa.ipc.open_stream(source)
+ except pa.lib.ArrowInvalid:
+ # if not, we open as file
+ reader = pa.ipc.open_file(source)
+ # for some reason this reader is not iterable
+ batches = [reader.get_batch(i) for i in range(reader.num_record_batches)]
+ else:
+ # if a stream, we're good
+ batches = reader # this reader is iterable
+ table = pa.Table.from_batches(batches)
self._load_table(table)
def _load_table(self, table):
|
{"golden_diff": "diff --git a/packages/vaex-arrow/vaex_arrow/dataset.py b/packages/vaex-arrow/vaex_arrow/dataset.py\n--- a/packages/vaex-arrow/vaex_arrow/dataset.py\n+++ b/packages/vaex-arrow/vaex_arrow/dataset.py\n@@ -23,8 +23,18 @@\n \n def _load(self):\n source = pa.memory_map(self.path)\n- reader = pa.ipc.open_stream(source)\n- table = pa.Table.from_batches([b for b in reader])\n+ try:\n+ # first we try if it opens as stream\n+ reader = pa.ipc.open_stream(source)\n+ except pa.lib.ArrowInvalid:\n+ # if not, we open as file\n+ reader = pa.ipc.open_file(source)\n+ # for some reason this reader is not iterable\n+ batches = [reader.get_batch(i) for i in range(reader.num_record_batches)]\n+ else:\n+ # if a stream, we're good\n+ batches = reader # this reader is iterable\n+ table = pa.Table.from_batches(batches)\n self._load_table(table)\n \n def _load_table(self, table):\n", "issue": "Failing to open arrow file format & misleading error message\nVaex fails to open arrow file format. It confusingly expects a file written not in the arrow file-format but in arrow stream-format.\r\n\r\nIf this is a non-negotiable vaex constraint, a less misleading error message might help very confused novices...\r\n\r\n```\r\n\r\nimport pandas as pd\r\nimport pyarrow as pa\r\nimport numpy as np\r\nimport vaex\r\ndf = pd.DataFrame(\r\n {\r\n 'col1': range(5)\r\n }\r\n)\r\ntable = pa.Table.from_pandas(df)\r\n\r\nwith pa.OSFile('test2.arrow', 'wb') as sink:\r\n with pa.RecordBatchFileWriter(sink, table.schema) as writer:\r\n writer.write_table(table)\r\n\r\nwith pa.OSFile('test2.arrow', 'rb') as source:\r\n df = pa.ipc.open_file(source).read_pandas()\r\ndf = vaex.open('test2.arrow')\r\n```\r\n\r\nError messages:\r\n```\r\nERROR:MainThread:vaex:error opening 'test2.arrow'\r\n---------------------------------------------------------------------------\r\nArrowInvalid Traceback (most recent call last)\r\n in \r\n 16 with pa.OSFile('test2.arrow', 'rb') as source:\r\n 17 df = pa.ipc.open_file(source).read_pandas()\r\n---> 18 df = vaex.open('test2.arrow')\r\n\r\nZ:\\Systemdateien\\Miniconda3\\envs\\finance\\lib\\site-packages\\vaex\\__init__.py in open(path, convert, shuffle, copy_index, *args, **kwargs)\r\n 189 ds = from_csv(path, copy_index=copy_index, **kwargs)\r\n 190 else:\r\n--> 191 ds = vaex.file.open(path, *args, **kwargs)\r\n 192 if convert and ds:\r\n 193 ds.export_hdf5(filename_hdf5, shuffle=shuffle)\r\n\r\nZ:\\Systemdateien\\Miniconda3\\envs\\finance\\lib\\site-packages\\vaex\\file\\__init__.py in open(path, *args, **kwargs)\r\n 28 for opener in opener_classes:\r\n 29 if opener.can_open(path, *args, **kwargs):\r\n---> 30 return opener.open(path, *args, **kwargs)\r\n 31 if hdf5:\r\n 32 openers.extend(hdf5.dataset.dataset_type_map.items())\r\n\r\nZ:\\Systemdateien\\Miniconda3\\envs\\finance\\lib\\site-packages\\vaex_arrow\\opener.py in open(path, *args, **kwargs)\r\n 9 def open(path, *args, **kwargs):\r\n 10 from .dataset import DatasetArrow\r\n---> 11 return DatasetArrow(path, *args, **kwargs)\r\n 12 \r\n 13 class ParquetOpener:\r\n\r\nZ:\\Systemdateien\\Miniconda3\\envs\\finance\\lib\\site-packages\\vaex_arrow\\dataset.py in __init__(self, filename, table, write)\r\n 18 self._write = write\r\n 19 if table is None:\r\n---> 20 self._load()\r\n 21 else:\r\n 22 self._load_table(table)\r\n\r\nZ:\\Systemdateien\\Miniconda3\\envs\\finance\\lib\\site-packages\\vaex_arrow\\dataset.py in _load(self)\r\n 24 def _load(self):\r\n 25 source = pa.memory_map(self.path)\r\n---> 26 reader = pa.ipc.open_stream(source)\r\n 27 table = pa.Table.from_batches([b for b in reader])\r\n 28 self._load_table(table)\r\n\r\nZ:\\Systemdateien\\Miniconda3\\envs\\finance\\lib\\site-packages\\pyarrow\\ipc.py in open_stream(source)\r\n 123 reader : RecordBatchStreamReader\r\n 124 \"\"\"\r\n--> 125 return RecordBatchStreamReader(source)\r\n 126 \r\n 127 \r\n\r\nZ:\\Systemdateien\\Miniconda3\\envs\\finance\\lib\\site-packages\\pyarrow\\ipc.py in __init__(self, source)\r\n 58 \"\"\"\r\n 59 def __init__(self, source):\r\n---> 60 self._open(source)\r\n 61 \r\n 62 \r\n\r\nZ:\\Systemdateien\\Miniconda3\\envs\\finance\\lib\\site-packages\\pyarrow\\ipc.pxi in pyarrow.lib._RecordBatchStreamReader._open()\r\n\r\nZ:\\Systemdateien\\Miniconda3\\envs\\finance\\lib\\site-packages\\pyarrow\\error.pxi in pyarrow.lib.check_status()\r\n\r\nArrowInvalid: Expected to read 1330795073 metadata bytes, but only read 1474\r\n```\n", "before_files": [{"content": "__author__ = 'maartenbreddels'\nimport logging\n\nimport pyarrow as pa\nimport pyarrow.parquet as pq\n\nimport vaex.dataset\nimport vaex.file.other\nfrom .convert import column_from_arrow_array\nlogger = logging.getLogger(\"vaex_arrow\")\n\n\nclass DatasetArrow(vaex.dataset.DatasetLocal):\n \"\"\"Implements storage using arrow\"\"\"\n\n def __init__(self, filename=None, table=None, write=False):\n super(DatasetArrow, self).__init__(name=filename, path=filename, column_names=[])\n self._write = write\n if table is None:\n self._load()\n else:\n self._load_table(table)\n\n def _load(self):\n source = pa.memory_map(self.path)\n reader = pa.ipc.open_stream(source)\n table = pa.Table.from_batches([b for b in reader])\n self._load_table(table)\n \n def _load_table(self, table):\n self._length_unfiltered = self._length_original = table.num_rows\n self._index_end = self._length_original = table.num_rows\n for col in table.columns:\n name = col.name\n # TODO: keep the arrow columns, and support and test chunks\n arrow_array = col.data.chunks[0]\n column = column_from_arrow_array(arrow_array)\n\n self.columns[name] = column\n self.column_names.append(name)\n self._save_assign_expression(name, vaex.expression.Expression(self, name))\n\n\n @classmethod\n def can_open(cls, path, *args, **kwargs):\n return path.rpartition('.')[2] == 'arrow'\n\n @classmethod\n def get_options(cls, path):\n return []\n\n @classmethod\n def option_to_args(cls, option):\n return []\n\nclass DatasetParquet(DatasetArrow):\n def _load(self):\n # might not be optimal, but it works, we can always see if we can\n # do mmapping later on\n table = pq.read_table(self.path)\n self._load_table(table)\n\nvaex.file.other.dataset_type_map[\"arrow\"] = DatasetArrow\nvaex.file.other.dataset_type_map[\"parquet\"] = DatasetParquet\n\n", "path": "packages/vaex-arrow/vaex_arrow/dataset.py"}]}
| 2,192 | 259 |
gh_patches_debug_42783
|
rasdani/github-patches
|
git_diff
|
conda-forge__conda-smithy-1372
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
search for teams by handle/slug and not name
I think this is a simple as using `team.slug` instead of `team.name` in the relevant code.
cc @isuruf
</issue>
<code>
[start of conda_smithy/github.py]
1 import os
2 from random import choice
3
4 from git import Repo
5
6 from github import Github
7 from github.GithubException import GithubException
8 from github.Organization import Organization
9 from github.Team import Team
10 import github
11
12 import conda_build.api
13 from conda_smithy.utils import get_feedstock_name_from_meta
14
15
16 def gh_token():
17 try:
18 with open(
19 os.path.expanduser("~/.conda-smithy/github.token"), "r"
20 ) as fh:
21 token = fh.read().strip()
22 if not token:
23 raise ValueError()
24 except (IOError, ValueError):
25 msg = (
26 "No github token. Go to https://github.com/settings/tokens/new and generate\n"
27 "a token with repo access. Put it in ~/.conda-smithy/github.token"
28 )
29 raise RuntimeError(msg)
30 return token
31
32
33 def create_team(org, name, description, repo_names=[]):
34 # PyGithub creates secret teams, and has no way of turning that off! :(
35 post_parameters = {
36 "name": name,
37 "description": description,
38 "privacy": "closed",
39 "permission": "push",
40 "repo_names": repo_names,
41 }
42 headers, data = org._requester.requestJsonAndCheck(
43 "POST", org.url + "/teams", input=post_parameters
44 )
45 return Team(org._requester, headers, data, completed=True)
46
47
48 def add_membership(team, member):
49 headers, data = team._requester.requestJsonAndCheck(
50 "PUT", team.url + "/memberships/" + member
51 )
52 return (headers, data)
53
54
55 def remove_membership(team, member):
56 headers, data = team._requester.requestJsonAndCheck(
57 "DELETE", team.url + "/memberships/" + member
58 )
59 return (headers, data)
60
61
62 def has_in_members(team, member):
63 status, headers, data = team._requester.requestJson(
64 "GET", team.url + "/members/" + member
65 )
66 return status == 204
67
68
69 def get_cached_team(org, team_name, description=""):
70 cached_file = os.path.expanduser(
71 "~/.conda-smithy/{}-{}-team".format(org.login, team_name)
72 )
73 try:
74 with open(cached_file, "r") as fh:
75 team_id = int(fh.read().strip())
76 return org.get_team(team_id)
77 except IOError:
78 pass
79
80 try:
81 repo = org.get_repo("{}-feedstock".format(team_name))
82 team = next(
83 (team for team in repo.get_teams() if team.name == team_name), None
84 )
85 if team:
86 return team
87 except GithubException:
88 pass
89
90 team = next(
91 (team for team in org.get_teams() if team.name == team_name), None
92 )
93 if not team:
94 if description:
95 team = create_team(org, team_name, description, [])
96 else:
97 raise RuntimeError("Couldn't find team {}".format(team_name))
98
99 with open(cached_file, "w") as fh:
100 fh.write(str(team.id))
101
102 return team
103
104
105 def create_github_repo(args):
106 token = gh_token()
107 meta = conda_build.api.render(
108 args.feedstock_directory,
109 permit_undefined_jinja=True,
110 finalize=False,
111 bypass_env_check=True,
112 trim_skip=False,
113 )[0][0]
114
115 feedstock_name = get_feedstock_name_from_meta(meta)
116
117 gh = Github(token)
118 user_or_org = None
119 if args.user is not None:
120 pass
121 # User has been defined, and organization has not.
122 user_or_org = gh.get_user()
123 else:
124 # Use the organization provided.
125 user_or_org = gh.get_organization(args.organization)
126
127 repo_name = "{}-feedstock".format(feedstock_name)
128 try:
129 gh_repo = user_or_org.create_repo(
130 repo_name,
131 has_wiki=False,
132 private=args.private,
133 description="A conda-smithy repository for {}.".format(
134 feedstock_name
135 ),
136 )
137 print("Created {} on github".format(gh_repo.full_name))
138 except GithubException as gh_except:
139 if (
140 gh_except.data.get("errors", [{}])[0].get("message", "")
141 != u"name already exists on this account"
142 ):
143 raise
144 gh_repo = user_or_org.get_repo(repo_name)
145 print("Github repository already exists.")
146
147 # Now add this new repo as a remote on the local clone.
148 repo = Repo(args.feedstock_directory)
149 remote_name = args.remote_name.strip()
150 if remote_name:
151 if remote_name in [remote.name for remote in repo.remotes]:
152 existing_remote = repo.remotes[remote_name]
153 if existing_remote.url != gh_repo.ssh_url:
154 print(
155 "Remote {} already exists, and doesn't point to {} "
156 "(it points to {}).".format(
157 remote_name, gh_repo.ssh_url, existing_remote.url
158 )
159 )
160 else:
161 repo.create_remote(remote_name, gh_repo.ssh_url)
162
163 if args.extra_admin_users is not None:
164 for user in args.extra_admin_users:
165 gh_repo.add_to_collaborators(user, "admin")
166
167 if args.add_teams:
168 if isinstance(user_or_org, Organization):
169 configure_github_team(meta, gh_repo, user_or_org, feedstock_name)
170
171
172 def accept_all_repository_invitations(gh):
173 user = gh.get_user()
174 invitations = github.PaginatedList.PaginatedList(
175 github.Invitation.Invitation,
176 user._requester,
177 user.url + "/repository_invitations",
178 None,
179 )
180 for invite in invitations:
181 invite._requester.requestJsonAndCheck("PATCH", invite.url)
182
183
184 def remove_from_project(gh, org, project):
185 user = gh.get_user()
186 repo = gh.get_repo("{}/{}".format(org, project))
187 repo.remove_from_collaborators(user.login)
188
189
190 def configure_github_team(meta, gh_repo, org, feedstock_name):
191
192 # Add a team for this repo and add the maintainers to it.
193 superlative = [
194 "awesome",
195 "slick",
196 "formidable",
197 "awe-inspiring",
198 "breathtaking",
199 "magnificent",
200 "wonderous",
201 "stunning",
202 "astonishing",
203 "superb",
204 "splendid",
205 "impressive",
206 "unbeatable",
207 "excellent",
208 "top",
209 "outstanding",
210 "exalted",
211 "standout",
212 "smashing",
213 ]
214
215 maintainers = set(meta.meta.get("extra", {}).get("recipe-maintainers", []))
216 maintainers = set(maintainer.lower() for maintainer in maintainers)
217 maintainer_teams = set(m for m in maintainers if "/" in m)
218 maintainers = set(m for m in maintainers if "/" not in m)
219
220 # Try to get team or create it if it doesn't exist.
221 team_name = feedstock_name
222 current_maintainer_teams = list(gh_repo.get_teams())
223 team = next(
224 (team for team in current_maintainer_teams if team.name == team_name),
225 None,
226 )
227 current_maintainers = set()
228 if not team:
229 team = create_team(
230 org,
231 team_name,
232 "The {} {} contributors!".format(choice(superlative), team_name),
233 )
234 team.add_to_repos(gh_repo)
235 else:
236 current_maintainers = set(
237 [e.login.lower() for e in team.get_members()]
238 )
239
240 # Get the all-members team
241 description = "All of the awesome {} contributors!".format(org.login)
242 all_members_team = get_cached_team(org, "all-members", description)
243 new_org_members = set()
244
245 # Add only the new maintainers to the team.
246 # Also add the new maintainers to all-members if not already included.
247 for new_maintainer in maintainers - current_maintainers:
248 add_membership(team, new_maintainer)
249
250 if not has_in_members(all_members_team, new_maintainer):
251 print(
252 "Adding a new member ({}) to {}. Welcome! :)".format(
253 new_maintainer, org.login
254 )
255 )
256 add_membership(all_members_team, new_maintainer)
257 new_org_members.add(new_maintainer)
258
259 # Mention any maintainers that need to be removed (unlikely here).
260 for old_maintainer in current_maintainers - maintainers:
261 print(
262 "AN OLD MEMBER ({}) NEEDS TO BE REMOVED FROM {}".format(
263 old_maintainer, gh_repo
264 )
265 )
266
267 # Add any new maintainer team
268 maintainer_teams = set(
269 m.split("/")[1]
270 for m in maintainer_teams
271 if m.startswith(str(org.login))
272 )
273 current_maintainer_teams = [team.name for team in current_maintainer_teams]
274 for maintainer_team in maintainer_teams - set(current_maintainer_teams):
275 print(
276 "Adding a new team ({}) to {}. Welcome! :)".format(
277 maintainer_team, org.login
278 )
279 )
280
281 team = get_cached_team(org, maintainer_team)
282 team.add_to_repos(gh_repo)
283
284 return maintainers, current_maintainers, new_org_members
285
[end of conda_smithy/github.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/conda_smithy/github.py b/conda_smithy/github.py
--- a/conda_smithy/github.py
+++ b/conda_smithy/github.py
@@ -187,7 +187,7 @@
repo.remove_from_collaborators(user.login)
-def configure_github_team(meta, gh_repo, org, feedstock_name):
+def configure_github_team(meta, gh_repo, org, feedstock_name, remove=True):
# Add a team for this repo and add the maintainers to it.
superlative = [
@@ -220,21 +220,21 @@
# Try to get team or create it if it doesn't exist.
team_name = feedstock_name
current_maintainer_teams = list(gh_repo.get_teams())
- team = next(
+ fs_team = next(
(team for team in current_maintainer_teams if team.name == team_name),
None,
)
current_maintainers = set()
- if not team:
- team = create_team(
+ if not fs_team:
+ fs_team = create_team(
org,
team_name,
"The {} {} contributors!".format(choice(superlative), team_name),
)
- team.add_to_repos(gh_repo)
+ fs_team.add_to_repos(gh_repo)
else:
current_maintainers = set(
- [e.login.lower() for e in team.get_members()]
+ [e.login.lower() for e in fs_team.get_members()]
)
# Get the all-members team
@@ -245,40 +245,41 @@
# Add only the new maintainers to the team.
# Also add the new maintainers to all-members if not already included.
for new_maintainer in maintainers - current_maintainers:
- add_membership(team, new_maintainer)
+ add_membership(fs_team, new_maintainer)
if not has_in_members(all_members_team, new_maintainer):
- print(
- "Adding a new member ({}) to {}. Welcome! :)".format(
- new_maintainer, org.login
- )
- )
add_membership(all_members_team, new_maintainer)
new_org_members.add(new_maintainer)
- # Mention any maintainers that need to be removed (unlikely here).
- for old_maintainer in current_maintainers - maintainers:
- print(
- "AN OLD MEMBER ({}) NEEDS TO BE REMOVED FROM {}".format(
- old_maintainer, gh_repo
- )
- )
+ # Remove any maintainers that need to be removed (unlikely here).
+ if remove:
+ for old_maintainer in current_maintainers - maintainers:
+ remove_membership(fs_team, old_maintainer)
- # Add any new maintainer team
+ # Add any new maintainer teams
maintainer_teams = set(
m.split("/")[1]
for m in maintainer_teams
if m.startswith(str(org.login))
)
- current_maintainer_teams = [team.name for team in current_maintainer_teams]
- for maintainer_team in maintainer_teams - set(current_maintainer_teams):
- print(
- "Adding a new team ({}) to {}. Welcome! :)".format(
- maintainer_team, org.login
- )
- )
-
- team = get_cached_team(org, maintainer_team)
+ current_maintainer_team_objs = {
+ team.slug: team for team in current_maintainer_teams
+ }
+ current_maintainer_teams = set(
+ [team.slug for team in current_maintainer_teams]
+ )
+ for new_team in maintainer_teams - current_maintainer_teams:
+ team = org.get_team_by_slug(new_team)
team.add_to_repos(gh_repo)
+ # remove any old teams
+ if remove:
+ for old_team in current_maintainer_teams - maintainer_teams:
+ team = current_maintainer_team_objs.get(
+ old_team, org.get_team_by_slug(old_team)
+ )
+ if team.name == fs_team.name:
+ continue
+ team.remove_from_repos(gh_repo)
+
return maintainers, current_maintainers, new_org_members
|
{"golden_diff": "diff --git a/conda_smithy/github.py b/conda_smithy/github.py\n--- a/conda_smithy/github.py\n+++ b/conda_smithy/github.py\n@@ -187,7 +187,7 @@\n repo.remove_from_collaborators(user.login)\n \n \n-def configure_github_team(meta, gh_repo, org, feedstock_name):\n+def configure_github_team(meta, gh_repo, org, feedstock_name, remove=True):\n \n # Add a team for this repo and add the maintainers to it.\n superlative = [\n@@ -220,21 +220,21 @@\n # Try to get team or create it if it doesn't exist.\n team_name = feedstock_name\n current_maintainer_teams = list(gh_repo.get_teams())\n- team = next(\n+ fs_team = next(\n (team for team in current_maintainer_teams if team.name == team_name),\n None,\n )\n current_maintainers = set()\n- if not team:\n- team = create_team(\n+ if not fs_team:\n+ fs_team = create_team(\n org,\n team_name,\n \"The {} {} contributors!\".format(choice(superlative), team_name),\n )\n- team.add_to_repos(gh_repo)\n+ fs_team.add_to_repos(gh_repo)\n else:\n current_maintainers = set(\n- [e.login.lower() for e in team.get_members()]\n+ [e.login.lower() for e in fs_team.get_members()]\n )\n \n # Get the all-members team\n@@ -245,40 +245,41 @@\n # Add only the new maintainers to the team.\n # Also add the new maintainers to all-members if not already included.\n for new_maintainer in maintainers - current_maintainers:\n- add_membership(team, new_maintainer)\n+ add_membership(fs_team, new_maintainer)\n \n if not has_in_members(all_members_team, new_maintainer):\n- print(\n- \"Adding a new member ({}) to {}. Welcome! :)\".format(\n- new_maintainer, org.login\n- )\n- )\n add_membership(all_members_team, new_maintainer)\n new_org_members.add(new_maintainer)\n \n- # Mention any maintainers that need to be removed (unlikely here).\n- for old_maintainer in current_maintainers - maintainers:\n- print(\n- \"AN OLD MEMBER ({}) NEEDS TO BE REMOVED FROM {}\".format(\n- old_maintainer, gh_repo\n- )\n- )\n+ # Remove any maintainers that need to be removed (unlikely here).\n+ if remove:\n+ for old_maintainer in current_maintainers - maintainers:\n+ remove_membership(fs_team, old_maintainer)\n \n- # Add any new maintainer team\n+ # Add any new maintainer teams\n maintainer_teams = set(\n m.split(\"/\")[1]\n for m in maintainer_teams\n if m.startswith(str(org.login))\n )\n- current_maintainer_teams = [team.name for team in current_maintainer_teams]\n- for maintainer_team in maintainer_teams - set(current_maintainer_teams):\n- print(\n- \"Adding a new team ({}) to {}. Welcome! :)\".format(\n- maintainer_team, org.login\n- )\n- )\n-\n- team = get_cached_team(org, maintainer_team)\n+ current_maintainer_team_objs = {\n+ team.slug: team for team in current_maintainer_teams\n+ }\n+ current_maintainer_teams = set(\n+ [team.slug for team in current_maintainer_teams]\n+ )\n+ for new_team in maintainer_teams - current_maintainer_teams:\n+ team = org.get_team_by_slug(new_team)\n team.add_to_repos(gh_repo)\n \n+ # remove any old teams\n+ if remove:\n+ for old_team in current_maintainer_teams - maintainer_teams:\n+ team = current_maintainer_team_objs.get(\n+ old_team, org.get_team_by_slug(old_team)\n+ )\n+ if team.name == fs_team.name:\n+ continue\n+ team.remove_from_repos(gh_repo)\n+\n return maintainers, current_maintainers, new_org_members\n", "issue": "search for teams by handle/slug and not name\nI think this is a simple as using `team.slug` instead of `team.name` in the relevant code.\r\n\r\ncc @isuruf \n", "before_files": [{"content": "import os\nfrom random import choice\n\nfrom git import Repo\n\nfrom github import Github\nfrom github.GithubException import GithubException\nfrom github.Organization import Organization\nfrom github.Team import Team\nimport github\n\nimport conda_build.api\nfrom conda_smithy.utils import get_feedstock_name_from_meta\n\n\ndef gh_token():\n try:\n with open(\n os.path.expanduser(\"~/.conda-smithy/github.token\"), \"r\"\n ) as fh:\n token = fh.read().strip()\n if not token:\n raise ValueError()\n except (IOError, ValueError):\n msg = (\n \"No github token. Go to https://github.com/settings/tokens/new and generate\\n\"\n \"a token with repo access. Put it in ~/.conda-smithy/github.token\"\n )\n raise RuntimeError(msg)\n return token\n\n\ndef create_team(org, name, description, repo_names=[]):\n # PyGithub creates secret teams, and has no way of turning that off! :(\n post_parameters = {\n \"name\": name,\n \"description\": description,\n \"privacy\": \"closed\",\n \"permission\": \"push\",\n \"repo_names\": repo_names,\n }\n headers, data = org._requester.requestJsonAndCheck(\n \"POST\", org.url + \"/teams\", input=post_parameters\n )\n return Team(org._requester, headers, data, completed=True)\n\n\ndef add_membership(team, member):\n headers, data = team._requester.requestJsonAndCheck(\n \"PUT\", team.url + \"/memberships/\" + member\n )\n return (headers, data)\n\n\ndef remove_membership(team, member):\n headers, data = team._requester.requestJsonAndCheck(\n \"DELETE\", team.url + \"/memberships/\" + member\n )\n return (headers, data)\n\n\ndef has_in_members(team, member):\n status, headers, data = team._requester.requestJson(\n \"GET\", team.url + \"/members/\" + member\n )\n return status == 204\n\n\ndef get_cached_team(org, team_name, description=\"\"):\n cached_file = os.path.expanduser(\n \"~/.conda-smithy/{}-{}-team\".format(org.login, team_name)\n )\n try:\n with open(cached_file, \"r\") as fh:\n team_id = int(fh.read().strip())\n return org.get_team(team_id)\n except IOError:\n pass\n\n try:\n repo = org.get_repo(\"{}-feedstock\".format(team_name))\n team = next(\n (team for team in repo.get_teams() if team.name == team_name), None\n )\n if team:\n return team\n except GithubException:\n pass\n\n team = next(\n (team for team in org.get_teams() if team.name == team_name), None\n )\n if not team:\n if description:\n team = create_team(org, team_name, description, [])\n else:\n raise RuntimeError(\"Couldn't find team {}\".format(team_name))\n\n with open(cached_file, \"w\") as fh:\n fh.write(str(team.id))\n\n return team\n\n\ndef create_github_repo(args):\n token = gh_token()\n meta = conda_build.api.render(\n args.feedstock_directory,\n permit_undefined_jinja=True,\n finalize=False,\n bypass_env_check=True,\n trim_skip=False,\n )[0][0]\n\n feedstock_name = get_feedstock_name_from_meta(meta)\n\n gh = Github(token)\n user_or_org = None\n if args.user is not None:\n pass\n # User has been defined, and organization has not.\n user_or_org = gh.get_user()\n else:\n # Use the organization provided.\n user_or_org = gh.get_organization(args.organization)\n\n repo_name = \"{}-feedstock\".format(feedstock_name)\n try:\n gh_repo = user_or_org.create_repo(\n repo_name,\n has_wiki=False,\n private=args.private,\n description=\"A conda-smithy repository for {}.\".format(\n feedstock_name\n ),\n )\n print(\"Created {} on github\".format(gh_repo.full_name))\n except GithubException as gh_except:\n if (\n gh_except.data.get(\"errors\", [{}])[0].get(\"message\", \"\")\n != u\"name already exists on this account\"\n ):\n raise\n gh_repo = user_or_org.get_repo(repo_name)\n print(\"Github repository already exists.\")\n\n # Now add this new repo as a remote on the local clone.\n repo = Repo(args.feedstock_directory)\n remote_name = args.remote_name.strip()\n if remote_name:\n if remote_name in [remote.name for remote in repo.remotes]:\n existing_remote = repo.remotes[remote_name]\n if existing_remote.url != gh_repo.ssh_url:\n print(\n \"Remote {} already exists, and doesn't point to {} \"\n \"(it points to {}).\".format(\n remote_name, gh_repo.ssh_url, existing_remote.url\n )\n )\n else:\n repo.create_remote(remote_name, gh_repo.ssh_url)\n\n if args.extra_admin_users is not None:\n for user in args.extra_admin_users:\n gh_repo.add_to_collaborators(user, \"admin\")\n\n if args.add_teams:\n if isinstance(user_or_org, Organization):\n configure_github_team(meta, gh_repo, user_or_org, feedstock_name)\n\n\ndef accept_all_repository_invitations(gh):\n user = gh.get_user()\n invitations = github.PaginatedList.PaginatedList(\n github.Invitation.Invitation,\n user._requester,\n user.url + \"/repository_invitations\",\n None,\n )\n for invite in invitations:\n invite._requester.requestJsonAndCheck(\"PATCH\", invite.url)\n\n\ndef remove_from_project(gh, org, project):\n user = gh.get_user()\n repo = gh.get_repo(\"{}/{}\".format(org, project))\n repo.remove_from_collaborators(user.login)\n\n\ndef configure_github_team(meta, gh_repo, org, feedstock_name):\n\n # Add a team for this repo and add the maintainers to it.\n superlative = [\n \"awesome\",\n \"slick\",\n \"formidable\",\n \"awe-inspiring\",\n \"breathtaking\",\n \"magnificent\",\n \"wonderous\",\n \"stunning\",\n \"astonishing\",\n \"superb\",\n \"splendid\",\n \"impressive\",\n \"unbeatable\",\n \"excellent\",\n \"top\",\n \"outstanding\",\n \"exalted\",\n \"standout\",\n \"smashing\",\n ]\n\n maintainers = set(meta.meta.get(\"extra\", {}).get(\"recipe-maintainers\", []))\n maintainers = set(maintainer.lower() for maintainer in maintainers)\n maintainer_teams = set(m for m in maintainers if \"/\" in m)\n maintainers = set(m for m in maintainers if \"/\" not in m)\n\n # Try to get team or create it if it doesn't exist.\n team_name = feedstock_name\n current_maintainer_teams = list(gh_repo.get_teams())\n team = next(\n (team for team in current_maintainer_teams if team.name == team_name),\n None,\n )\n current_maintainers = set()\n if not team:\n team = create_team(\n org,\n team_name,\n \"The {} {} contributors!\".format(choice(superlative), team_name),\n )\n team.add_to_repos(gh_repo)\n else:\n current_maintainers = set(\n [e.login.lower() for e in team.get_members()]\n )\n\n # Get the all-members team\n description = \"All of the awesome {} contributors!\".format(org.login)\n all_members_team = get_cached_team(org, \"all-members\", description)\n new_org_members = set()\n\n # Add only the new maintainers to the team.\n # Also add the new maintainers to all-members if not already included.\n for new_maintainer in maintainers - current_maintainers:\n add_membership(team, new_maintainer)\n\n if not has_in_members(all_members_team, new_maintainer):\n print(\n \"Adding a new member ({}) to {}. Welcome! :)\".format(\n new_maintainer, org.login\n )\n )\n add_membership(all_members_team, new_maintainer)\n new_org_members.add(new_maintainer)\n\n # Mention any maintainers that need to be removed (unlikely here).\n for old_maintainer in current_maintainers - maintainers:\n print(\n \"AN OLD MEMBER ({}) NEEDS TO BE REMOVED FROM {}\".format(\n old_maintainer, gh_repo\n )\n )\n\n # Add any new maintainer team\n maintainer_teams = set(\n m.split(\"/\")[1]\n for m in maintainer_teams\n if m.startswith(str(org.login))\n )\n current_maintainer_teams = [team.name for team in current_maintainer_teams]\n for maintainer_team in maintainer_teams - set(current_maintainer_teams):\n print(\n \"Adding a new team ({}) to {}. Welcome! :)\".format(\n maintainer_team, org.login\n )\n )\n\n team = get_cached_team(org, maintainer_team)\n team.add_to_repos(gh_repo)\n\n return maintainers, current_maintainers, new_org_members\n", "path": "conda_smithy/github.py"}]}
| 3,360 | 958 |
gh_patches_debug_8010
|
rasdani/github-patches
|
git_diff
|
wagtail__wagtail-7538
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Preview error when using StreamField with EditHandler on per-request basis
### Issue Summary
We're using `EditHandler`s, which show/hide fields depending on the user's permission. This is based on the example in wagtail's tests (see https://github.com/wagtail/wagtail/blob/main/wagtail/tests/testapp/models.py#L1418).
**Problem:**
Unfortunately, there are multiple places in wagtail's codebase, where `BaseFormEditHandler.get_form_class()` is called, without binding the current request (and also instance etc). One example is the preview-mode (see [preview.py](https://github.com/wagtail/wagtail/blob/main/wagtail/admin/views/pages/preview.py#L51)).
As a result, a user which can only view a subset of the full admin form of a page for example, actually validates his POST-data (which only contains a subset of the fields) with the full form, since the logic that would hide some fields based on the request isn't called.
This probably works for simple cases like [wagtail's test](https://github.com/wagtail/wagtail/blob/main/wagtail/tests/testapp/models.py#L1442) and described in multiple tickets (e.g. https://github.com/wagtail/wagtail/issues/5546), but in our case it results in crashes when showing/hiding `StreamField`s.
**Solution:**
At least bind `instance` and `request` before `get_form_class()` is called, where possible, similar to how the normal edit-view does it: https://github.com/wagtail/wagtail/blob/main/wagtail/admin/views/pages/edit.py#L311
I will try to add those calls to all the places I can find and let the tests run and I hope this can be fixed quickly.
### Steps to Reproduce
I don't have a simple example available yet, but basically we have a custom `BaseCompositeEditHandler` which may skip some of its children, if the user is not a superuser. This is very similar to the [sample](https://github.com/wagtail/wagtail/blob/main/wagtail/tests/testapp/models.py#L1393) in wagtail's tests, but instead of switching between two panel-lists, we have one default list, and based on the field-keys we may exclude some of those fields.
So this all works for the superuser, since he can edit all of the fields. But if one of those excluded fields is a `StreamField` and the non-superuser clicks on the preview button, the server crashes with something like that:
2021-09-24 11:43:13 ERROR django.request: Internal Server Error: /admin/content/pages/add/base/testpage/1337/preview/ (log.py:224)
Traceback (most recent call last):
File "/****/lib/python3.8/site-packages/django/utils/datastructures.py", line 76, in __getitem__
list_ = super().__getitem__(key)
KeyError: 'form_element_blocks-count'
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
...****...
return view_func(request, *args, **kwargs)
File "/****/lib/python3.8/site-packages/django/views/decorators/cache.py", line 44, in _wrapped_view_func
response = view_func(request, *args, **kwargs)
File "/****/lib/python3.8/site-packages/wagtail/admin/urls/__init__.py", line 121, in wrapper
return view_func(request, *args, **kwargs)
File "/****/lib/python3.8/site-packages/wagtail/admin/auth.py", line 193, in decorated_view
return view_func(request, *args, **kwargs)
File "/****/lib/python3.8/site-packages/django/views/generic/base.py", line 70, in view
return self.dispatch(request, *args, **kwargs)
File "/****/lib/python3.8/site-packages/django/views/generic/base.py", line 98, in dispatch
return handler(request, *args, **kwargs)
File "/****/lib/python3.8/site-packages/wagtail/admin/views/pages/preview.py", line 64, in post
form = self.get_form(self.get_page(), request.POST)
File "/****/lib/python3.8/site-packages/wagtail/admin/views/pages/preview.py", line 129, in get_form
if form.is_valid():
File "/****/lib/python3.8/site-packages/wagtail/admin/forms/pages.py", line 150, in is_valid
return super().is_valid()
File "/****/lib/python3.8/site-packages/modelcluster/forms.py", line 316, in is_valid
form_is_valid = super(ClusterForm, self).is_valid()
File "/****/lib/python3.8/site-packages/django/forms/forms.py", line 175, in is_valid
return self.is_bound and not self.errors
File "/****/lib/python3.8/site-packages/django/forms/forms.py", line 170, in errors
self.full_clean()
File "/****/lib/python3.8/site-packages/django/forms/forms.py", line 372, in full_clean
self._clean_fields()
File "/****/lib/python3.8/site-packages/django/forms/forms.py", line 384, in _clean_fields
value = field.widget.value_from_datadict(self.data, self.files, self.add_prefix(name))
File "/****/lib/python3.8/site-packages/wagtail/core/blocks/base.py", line 535, in value_from_datadict
return self.block_def.value_from_datadict(data, files, name)
File "/****/lib/python3.8/site-packages/wagtail/core/blocks/stream_block.py", line 97, in value_from_datadict
count = int(data['%s-count' % prefix])
File "/****/lib/python3.8/site-packages/django/utils/datastructures.py", line 78, in __getitem__
raise MultiValueDictKeyError(key)
django.utils.datastructures.MultiValueDictKeyError: 'form_element_blocks-count'
2021-09-24 11:43:13 ERROR django.channels.server: HTTP POST /admin/content/pages/add/base/testpage/1337/preview/ 500 [0.15, 127.0.0.1:44074] (runserver.py:160)
In the above stacktrace, the `form_element_blocks` (which is a `StreamField`) are not visible to the non-superuser user in the frontend. But the preview-request actually creates a form with the `StreamField` (because our code in `on_request_bound()` isn't called), which fails when validating against the POST-data, which doesn't contain any of the `StreamField`'s POST-data.
### Technical details
* Python version: 3.8.10
* Django version: 3.2.7
* Wagtail version: 2.14.1
</issue>
<code>
[start of wagtail/admin/views/pages/preview.py]
1 from time import time
2
3 from django.contrib.contenttypes.models import ContentType
4 from django.core.exceptions import PermissionDenied
5 from django.http import Http404, JsonResponse
6 from django.http.request import QueryDict
7 from django.shortcuts import get_object_or_404
8 from django.template.response import TemplateResponse
9 from django.views.generic import View
10
11 from wagtail.core.models import Page
12
13
14 def view_draft(request, page_id):
15 page = get_object_or_404(Page, id=page_id).get_latest_revision_as_page()
16 perms = page.permissions_for_user(request.user)
17 if not (perms.can_publish() or perms.can_edit()):
18 raise PermissionDenied
19
20 try:
21 preview_mode = page.default_preview_mode
22 except IndexError:
23 raise PermissionDenied
24
25 return page.make_preview_request(request, preview_mode)
26
27
28 class PreviewOnEdit(View):
29 http_method_names = ('post', 'get')
30 preview_expiration_timeout = 60 * 60 * 24 # seconds
31 session_key_prefix = 'wagtail-preview-'
32
33 def remove_old_preview_data(self):
34 expiration = time() - self.preview_expiration_timeout
35 expired_keys = [
36 k for k, v in self.request.session.items()
37 if k.startswith(self.session_key_prefix) and v[1] < expiration]
38 # Removes the session key gracefully
39 for k in expired_keys:
40 self.request.session.pop(k)
41
42 @property
43 def session_key(self):
44 return '{}{}'.format(self.session_key_prefix, self.kwargs['page_id'])
45
46 def get_page(self):
47 return get_object_or_404(Page,
48 id=self.kwargs["page_id"]).get_latest_revision_as_page()
49
50 def get_form(self, page, query_dict):
51 form_class = page.get_edit_handler().get_form_class()
52 parent_page = page.get_parent().specific
53
54 if self.session_key not in self.request.session:
55 # Session key not in session, returning null form
56 return form_class(instance=page, parent_page=parent_page)
57
58 return form_class(query_dict, instance=page, parent_page=parent_page)
59
60 def post(self, request, *args, **kwargs):
61 # TODO: Handle request.FILES.
62 request.session[self.session_key] = request.POST.urlencode(), time()
63 self.remove_old_preview_data()
64 form = self.get_form(self.get_page(), request.POST)
65 return JsonResponse({'is_valid': form.is_valid()})
66
67 def error_response(self, page):
68 return TemplateResponse(
69 self.request, 'wagtailadmin/pages/preview_error.html',
70 {'page': page}
71 )
72
73 def get(self, request, *args, **kwargs):
74 page = self.get_page()
75
76 post_data, timestamp = self.request.session.get(self.session_key,
77 (None, None))
78 if not isinstance(post_data, str):
79 post_data = ''
80 form = self.get_form(page, QueryDict(post_data))
81
82 if not form.is_valid():
83 return self.error_response(page)
84
85 form.save(commit=False)
86
87 try:
88 preview_mode = request.GET.get('mode', page.default_preview_mode)
89 except IndexError:
90 raise PermissionDenied
91
92 return page.make_preview_request(request, preview_mode)
93
94
95 class PreviewOnCreate(PreviewOnEdit):
96 @property
97 def session_key(self):
98 return '{}{}-{}-{}'.format(
99 self.session_key_prefix,
100 self.kwargs['content_type_app_name'],
101 self.kwargs['content_type_model_name'],
102 self.kwargs['parent_page_id'],
103 )
104
105 def get_page(self):
106 content_type_app_name = self.kwargs["content_type_app_name"]
107 content_type_model_name = self.kwargs["content_type_model_name"]
108 parent_page_id = self.kwargs["parent_page_id"]
109 try:
110 content_type = ContentType.objects.get_by_natural_key(
111 content_type_app_name, content_type_model_name)
112 except ContentType.DoesNotExist:
113 raise Http404
114
115 page = content_type.model_class()()
116 parent_page = get_object_or_404(Page, id=parent_page_id).specific
117 # We need to populate treebeard's path / depth fields in order to
118 # pass validation. We can't make these 100% consistent with the rest
119 # of the tree without making actual database changes (such as
120 # incrementing the parent's numchild field), but by calling treebeard's
121 # internal _get_path method, we can set a 'realistic' value that will
122 # hopefully enable tree traversal operations
123 # to at least partially work.
124 page.depth = parent_page.depth + 1
125 # Puts the page at the next available path
126 # for a child of `parent_page`.
127 if parent_page.is_leaf():
128 # set the path as the first child of parent_page
129 page.path = page._get_path(parent_page.path, page.depth, 1)
130 else:
131 # add the new page after the last child of parent_page
132 page.path = parent_page.get_last_child()._inc_path()
133
134 return page
135
136 def get_form(self, page, query_dict):
137 form = super().get_form(page, query_dict)
138 if form.is_valid():
139 # Ensures our unsaved page has a suitable url.
140 form.instance.set_url_path(form.parent_page)
141
142 form.instance.full_clean()
143 return form
144
[end of wagtail/admin/views/pages/preview.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/wagtail/admin/views/pages/preview.py b/wagtail/admin/views/pages/preview.py
--- a/wagtail/admin/views/pages/preview.py
+++ b/wagtail/admin/views/pages/preview.py
@@ -48,7 +48,7 @@
id=self.kwargs["page_id"]).get_latest_revision_as_page()
def get_form(self, page, query_dict):
- form_class = page.get_edit_handler().get_form_class()
+ form_class = page.get_edit_handler().bind_to(instance=page, request=self.request).get_form_class()
parent_page = page.get_parent().specific
if self.session_key not in self.request.session:
|
{"golden_diff": "diff --git a/wagtail/admin/views/pages/preview.py b/wagtail/admin/views/pages/preview.py\n--- a/wagtail/admin/views/pages/preview.py\n+++ b/wagtail/admin/views/pages/preview.py\n@@ -48,7 +48,7 @@\n id=self.kwargs[\"page_id\"]).get_latest_revision_as_page()\n \n def get_form(self, page, query_dict):\n- form_class = page.get_edit_handler().get_form_class()\n+ form_class = page.get_edit_handler().bind_to(instance=page, request=self.request).get_form_class()\n parent_page = page.get_parent().specific\n \n if self.session_key not in self.request.session:\n", "issue": "Preview error when using StreamField with EditHandler on per-request basis\n### Issue Summary\r\n\r\nWe're using `EditHandler`s, which show/hide fields depending on the user's permission. This is based on the example in wagtail's tests (see https://github.com/wagtail/wagtail/blob/main/wagtail/tests/testapp/models.py#L1418).\r\n\r\n**Problem:**\r\nUnfortunately, there are multiple places in wagtail's codebase, where `BaseFormEditHandler.get_form_class()` is called, without binding the current request (and also instance etc). One example is the preview-mode (see [preview.py](https://github.com/wagtail/wagtail/blob/main/wagtail/admin/views/pages/preview.py#L51)).\r\n\r\nAs a result, a user which can only view a subset of the full admin form of a page for example, actually validates his POST-data (which only contains a subset of the fields) with the full form, since the logic that would hide some fields based on the request isn't called.\r\n\r\nThis probably works for simple cases like [wagtail's test](https://github.com/wagtail/wagtail/blob/main/wagtail/tests/testapp/models.py#L1442) and described in multiple tickets (e.g. https://github.com/wagtail/wagtail/issues/5546), but in our case it results in crashes when showing/hiding `StreamField`s.\r\n\r\n**Solution:**\r\nAt least bind `instance` and `request` before `get_form_class()` is called, where possible, similar to how the normal edit-view does it: https://github.com/wagtail/wagtail/blob/main/wagtail/admin/views/pages/edit.py#L311\r\n\r\nI will try to add those calls to all the places I can find and let the tests run and I hope this can be fixed quickly.\r\n\r\n\r\n### Steps to Reproduce\r\n\r\nI don't have a simple example available yet, but basically we have a custom `BaseCompositeEditHandler` which may skip some of its children, if the user is not a superuser. This is very similar to the [sample](https://github.com/wagtail/wagtail/blob/main/wagtail/tests/testapp/models.py#L1393) in wagtail's tests, but instead of switching between two panel-lists, we have one default list, and based on the field-keys we may exclude some of those fields.\r\n\r\nSo this all works for the superuser, since he can edit all of the fields. But if one of those excluded fields is a `StreamField` and the non-superuser clicks on the preview button, the server crashes with something like that:\r\n\r\n 2021-09-24 11:43:13 ERROR django.request: Internal Server Error: /admin/content/pages/add/base/testpage/1337/preview/ (log.py:224)\r\n Traceback (most recent call last):\r\n File \"/****/lib/python3.8/site-packages/django/utils/datastructures.py\", line 76, in __getitem__\r\n list_ = super().__getitem__(key)\r\n KeyError: 'form_element_blocks-count'\r\n\r\n During handling of the above exception, another exception occurred:\r\n\r\n Traceback (most recent call last):\r\n ...****...\r\n return view_func(request, *args, **kwargs)\r\n File \"/****/lib/python3.8/site-packages/django/views/decorators/cache.py\", line 44, in _wrapped_view_func\r\n response = view_func(request, *args, **kwargs)\r\n File \"/****/lib/python3.8/site-packages/wagtail/admin/urls/__init__.py\", line 121, in wrapper\r\n return view_func(request, *args, **kwargs)\r\n File \"/****/lib/python3.8/site-packages/wagtail/admin/auth.py\", line 193, in decorated_view\r\n return view_func(request, *args, **kwargs)\r\n File \"/****/lib/python3.8/site-packages/django/views/generic/base.py\", line 70, in view\r\n return self.dispatch(request, *args, **kwargs)\r\n File \"/****/lib/python3.8/site-packages/django/views/generic/base.py\", line 98, in dispatch\r\n return handler(request, *args, **kwargs)\r\n File \"/****/lib/python3.8/site-packages/wagtail/admin/views/pages/preview.py\", line 64, in post\r\n form = self.get_form(self.get_page(), request.POST)\r\n File \"/****/lib/python3.8/site-packages/wagtail/admin/views/pages/preview.py\", line 129, in get_form\r\n if form.is_valid():\r\n File \"/****/lib/python3.8/site-packages/wagtail/admin/forms/pages.py\", line 150, in is_valid\r\n return super().is_valid()\r\n File \"/****/lib/python3.8/site-packages/modelcluster/forms.py\", line 316, in is_valid\r\n form_is_valid = super(ClusterForm, self).is_valid()\r\n File \"/****/lib/python3.8/site-packages/django/forms/forms.py\", line 175, in is_valid\r\n return self.is_bound and not self.errors\r\n File \"/****/lib/python3.8/site-packages/django/forms/forms.py\", line 170, in errors\r\n self.full_clean()\r\n File \"/****/lib/python3.8/site-packages/django/forms/forms.py\", line 372, in full_clean\r\n self._clean_fields()\r\n File \"/****/lib/python3.8/site-packages/django/forms/forms.py\", line 384, in _clean_fields\r\n value = field.widget.value_from_datadict(self.data, self.files, self.add_prefix(name))\r\n File \"/****/lib/python3.8/site-packages/wagtail/core/blocks/base.py\", line 535, in value_from_datadict\r\n return self.block_def.value_from_datadict(data, files, name)\r\n File \"/****/lib/python3.8/site-packages/wagtail/core/blocks/stream_block.py\", line 97, in value_from_datadict\r\n count = int(data['%s-count' % prefix])\r\n File \"/****/lib/python3.8/site-packages/django/utils/datastructures.py\", line 78, in __getitem__\r\n raise MultiValueDictKeyError(key)\r\n django.utils.datastructures.MultiValueDictKeyError: 'form_element_blocks-count'\r\n 2021-09-24 11:43:13 ERROR django.channels.server: HTTP POST /admin/content/pages/add/base/testpage/1337/preview/ 500 [0.15, 127.0.0.1:44074] (runserver.py:160)\r\n\r\nIn the above stacktrace, the `form_element_blocks` (which is a `StreamField`) are not visible to the non-superuser user in the frontend. But the preview-request actually creates a form with the `StreamField` (because our code in `on_request_bound()` isn't called), which fails when validating against the POST-data, which doesn't contain any of the `StreamField`'s POST-data.\r\n\r\n\r\n### Technical details\r\n\r\n* Python version: 3.8.10\r\n* Django version: 3.2.7\r\n* Wagtail version: 2.14.1\r\n\n", "before_files": [{"content": "from time import time\n\nfrom django.contrib.contenttypes.models import ContentType\nfrom django.core.exceptions import PermissionDenied\nfrom django.http import Http404, JsonResponse\nfrom django.http.request import QueryDict\nfrom django.shortcuts import get_object_or_404\nfrom django.template.response import TemplateResponse\nfrom django.views.generic import View\n\nfrom wagtail.core.models import Page\n\n\ndef view_draft(request, page_id):\n page = get_object_or_404(Page, id=page_id).get_latest_revision_as_page()\n perms = page.permissions_for_user(request.user)\n if not (perms.can_publish() or perms.can_edit()):\n raise PermissionDenied\n\n try:\n preview_mode = page.default_preview_mode\n except IndexError:\n raise PermissionDenied\n\n return page.make_preview_request(request, preview_mode)\n\n\nclass PreviewOnEdit(View):\n http_method_names = ('post', 'get')\n preview_expiration_timeout = 60 * 60 * 24 # seconds\n session_key_prefix = 'wagtail-preview-'\n\n def remove_old_preview_data(self):\n expiration = time() - self.preview_expiration_timeout\n expired_keys = [\n k for k, v in self.request.session.items()\n if k.startswith(self.session_key_prefix) and v[1] < expiration]\n # Removes the session key gracefully\n for k in expired_keys:\n self.request.session.pop(k)\n\n @property\n def session_key(self):\n return '{}{}'.format(self.session_key_prefix, self.kwargs['page_id'])\n\n def get_page(self):\n return get_object_or_404(Page,\n id=self.kwargs[\"page_id\"]).get_latest_revision_as_page()\n\n def get_form(self, page, query_dict):\n form_class = page.get_edit_handler().get_form_class()\n parent_page = page.get_parent().specific\n\n if self.session_key not in self.request.session:\n # Session key not in session, returning null form\n return form_class(instance=page, parent_page=parent_page)\n\n return form_class(query_dict, instance=page, parent_page=parent_page)\n\n def post(self, request, *args, **kwargs):\n # TODO: Handle request.FILES.\n request.session[self.session_key] = request.POST.urlencode(), time()\n self.remove_old_preview_data()\n form = self.get_form(self.get_page(), request.POST)\n return JsonResponse({'is_valid': form.is_valid()})\n\n def error_response(self, page):\n return TemplateResponse(\n self.request, 'wagtailadmin/pages/preview_error.html',\n {'page': page}\n )\n\n def get(self, request, *args, **kwargs):\n page = self.get_page()\n\n post_data, timestamp = self.request.session.get(self.session_key,\n (None, None))\n if not isinstance(post_data, str):\n post_data = ''\n form = self.get_form(page, QueryDict(post_data))\n\n if not form.is_valid():\n return self.error_response(page)\n\n form.save(commit=False)\n\n try:\n preview_mode = request.GET.get('mode', page.default_preview_mode)\n except IndexError:\n raise PermissionDenied\n\n return page.make_preview_request(request, preview_mode)\n\n\nclass PreviewOnCreate(PreviewOnEdit):\n @property\n def session_key(self):\n return '{}{}-{}-{}'.format(\n self.session_key_prefix,\n self.kwargs['content_type_app_name'],\n self.kwargs['content_type_model_name'],\n self.kwargs['parent_page_id'],\n )\n\n def get_page(self):\n content_type_app_name = self.kwargs[\"content_type_app_name\"]\n content_type_model_name = self.kwargs[\"content_type_model_name\"]\n parent_page_id = self.kwargs[\"parent_page_id\"]\n try:\n content_type = ContentType.objects.get_by_natural_key(\n content_type_app_name, content_type_model_name)\n except ContentType.DoesNotExist:\n raise Http404\n\n page = content_type.model_class()()\n parent_page = get_object_or_404(Page, id=parent_page_id).specific\n # We need to populate treebeard's path / depth fields in order to\n # pass validation. We can't make these 100% consistent with the rest\n # of the tree without making actual database changes (such as\n # incrementing the parent's numchild field), but by calling treebeard's\n # internal _get_path method, we can set a 'realistic' value that will\n # hopefully enable tree traversal operations\n # to at least partially work.\n page.depth = parent_page.depth + 1\n # Puts the page at the next available path\n # for a child of `parent_page`.\n if parent_page.is_leaf():\n # set the path as the first child of parent_page\n page.path = page._get_path(parent_page.path, page.depth, 1)\n else:\n # add the new page after the last child of parent_page\n page.path = parent_page.get_last_child()._inc_path()\n\n return page\n\n def get_form(self, page, query_dict):\n form = super().get_form(page, query_dict)\n if form.is_valid():\n # Ensures our unsaved page has a suitable url.\n form.instance.set_url_path(form.parent_page)\n\n form.instance.full_clean()\n return form\n", "path": "wagtail/admin/views/pages/preview.py"}]}
| 3,601 | 145 |
gh_patches_debug_30712
|
rasdani/github-patches
|
git_diff
|
nvaccess__nvda-11841
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Visual Studio: IntelliSense tooltips reported twice
### Steps to reproduce:
1. Open Visual Studio 2019
2. Open a C# project
3. Enable reporting of Tooltips
4. Trigger an IntelliSense autocomplete suggestion by typing something.
5. Arrow through the suggestions
### Actual behavior:
The selected item is announced, followed by twice the tooltip.
### Expected behavior:
The selected item is announced, followed by once the tooltip.
### System configuration
#### NVDA installed/portable/running from source:
Installed
#### NVDA version:
alpha-20957
#### Windows version:
Windows 10 2004
#### Name and version of other software in use when reproducing the issue:
Visual Studio 2019 16.7.3 Enterprise
### Other questions
#### Does the issue still occur after restarting your computer?
Yes
#### Have you tried any other versions of NVDA? If so, please report their behaviors.
No
#### If addons are disabled, is your problem still occuring?
Yes
#### Did you try to run the COM registry fixing tool in NVDA menu / tools?
Yes
</issue>
<code>
[start of source/NVDAObjects/UIA/VisualStudio.py]
1 # This file is covered by the GNU General Public License.
2 # See the file COPYING for more details.
3 # Copyright (C) 2020 NV Access Limited, Leonard de Ruijter
4
5 """
6 Object overlay classes for Visual Studio components
7 available in Visual Studio and SQL Server Management Studio.
8 """
9
10 from . import UIA
11 import speech
12 import braille
13 import api
14
15
16 class IntelliSenseItem(UIA):
17
18 def _get_name(self):
19 return self.UIAElement.cachedAutomationID
20
21 def event_UIA_elementSelected(self):
22 # Cancel speech to have speech announce the selection as soon as possible.
23 # This is needed because L{reportFocus} does not cancel speech.
24 # Therefore, if speech wouldn't be cancelled,
25 # selection announcements would queue up when changing selection rapidly.
26 speech.cancelSpeech()
27 api.setNavigatorObject(self, isFocus=True)
28 self.reportFocus()
29 # Display results as flash messages.
30 braille.handler.message(braille.getPropertiesBraille(
31 name=self.name, role=self.role, positionInfo=self.positionInfo, description=self.description
32 ))
33
34
35 class IntelliSenseList(UIA):
36 ...
37
38
39 class IntelliSenseLiveRegion(UIA):
40 """
41 Visual Studio uses both Intellisense menu item objects and a live region
42 to communicate Intellisense selections.
43 NVDA uses the menu item approach and therefore the live region provides doubled information
44 and is disabled.
45 """
46
47 _shouldAllowUIALiveRegionChangeEvent = False
48
49
50 _INTELLISENSE_LIST_AUTOMATION_IDS = {
51 "listBoxCompletions",
52 "CompletionList"
53 }
54
55
56 def findExtraOverlayClasses(obj, clsList):
57 if obj.UIAAutomationId in _INTELLISENSE_LIST_AUTOMATION_IDS:
58 clsList.insert(0, IntelliSenseList)
59 elif isinstance(obj.parent, IntelliSenseList) and obj.UIAElement.cachedClassName == "IntellisenseMenuItem":
60 clsList.insert(0, IntelliSenseItem)
61 elif (
62 obj.UIAElement.cachedClassName == "LiveTextBlock"
63 and obj.previous
64 and isinstance(obj.previous.previous, IntelliSenseList)
65 ):
66 clsList.insert(0, IntelliSenseLiveRegion)
67
[end of source/NVDAObjects/UIA/VisualStudio.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/source/NVDAObjects/UIA/VisualStudio.py b/source/NVDAObjects/UIA/VisualStudio.py
--- a/source/NVDAObjects/UIA/VisualStudio.py
+++ b/source/NVDAObjects/UIA/VisualStudio.py
@@ -7,10 +7,11 @@
available in Visual Studio and SQL Server Management Studio.
"""
-from . import UIA
+from . import UIA, ToolTip
import speech
import braille
import api
+import time
class IntelliSenseItem(UIA):
@@ -53,6 +54,34 @@
}
+class CompletionToolTip(ToolTip):
+ """ A tool tip for which duplicate open events can be fired.
+ """
+
+ #: Keeps track of the last ToolTipOpened event (text, time)
+ _lastToolTipOpenedInfo = (None, None)
+ #: The duplicate tooltip events will be dropped within this time window
+ _preventDuplicateToolTipSeconds = 0.2
+
+ def event_UIA_toolTipOpened(self):
+ oldText, oldTime = self._lastToolTipOpenedInfo
+ newText = self.name
+ newTime = time.time()
+ self.__class__._lastToolTipOpenedInfo = (newText, newTime)
+ withinPossibleDupToolTipTimeWindow = (
+ oldTime is not None
+ and (newTime - oldTime) < self._preventDuplicateToolTipSeconds
+ )
+ if newText == oldText and withinPossibleDupToolTipTimeWindow:
+ # Tool-tip event suspected to be a duplicate, drop the event.
+ # - Users attempting to rapidly re-announce tool-tips may
+ # have the announcement erroneously suppressed
+ # - Users on slower systems (or systems under load) may still
+ # receive duplicate announcements
+ return
+ super().event_UIA_toolTipOpened()
+
+
def findExtraOverlayClasses(obj, clsList):
if obj.UIAAutomationId in _INTELLISENSE_LIST_AUTOMATION_IDS:
clsList.insert(0, IntelliSenseList)
@@ -64,3 +93,5 @@
and isinstance(obj.previous.previous, IntelliSenseList)
):
clsList.insert(0, IntelliSenseLiveRegion)
+ elif obj.UIAAutomationId == "completion tooltip":
+ clsList.insert(0, CompletionToolTip)
|
{"golden_diff": "diff --git a/source/NVDAObjects/UIA/VisualStudio.py b/source/NVDAObjects/UIA/VisualStudio.py\n--- a/source/NVDAObjects/UIA/VisualStudio.py\n+++ b/source/NVDAObjects/UIA/VisualStudio.py\n@@ -7,10 +7,11 @@\n available in Visual Studio and SQL Server Management Studio.\n \"\"\"\n \n-from . import UIA\n+from . import UIA, ToolTip\n import speech\n import braille\n import api\n+import time\n \n \n class IntelliSenseItem(UIA):\n@@ -53,6 +54,34 @@\n }\n \n \n+class CompletionToolTip(ToolTip):\n+\t\"\"\" A tool tip for which duplicate open events can be fired.\n+\t\"\"\"\n+\n+\t#: Keeps track of the last ToolTipOpened event (text, time)\n+\t_lastToolTipOpenedInfo = (None, None)\n+\t#: The duplicate tooltip events will be dropped within this time window\n+\t_preventDuplicateToolTipSeconds = 0.2\n+\n+\tdef event_UIA_toolTipOpened(self):\n+\t\toldText, oldTime = self._lastToolTipOpenedInfo\n+\t\tnewText = self.name\n+\t\tnewTime = time.time()\n+\t\tself.__class__._lastToolTipOpenedInfo = (newText, newTime)\n+\t\twithinPossibleDupToolTipTimeWindow = (\n+\t\t\toldTime is not None\n+\t\t\tand (newTime - oldTime) < self._preventDuplicateToolTipSeconds\n+\t\t)\n+\t\tif newText == oldText and withinPossibleDupToolTipTimeWindow:\n+\t\t\t# Tool-tip event suspected to be a duplicate, drop the event.\n+\t\t\t# - Users attempting to rapidly re-announce tool-tips may\n+\t\t\t# have the announcement erroneously suppressed\n+\t\t\t# - Users on slower systems (or systems under load) may still\n+\t\t\t# receive duplicate announcements\n+\t\t\treturn\n+\t\tsuper().event_UIA_toolTipOpened()\n+\n+\n def findExtraOverlayClasses(obj, clsList):\n \tif obj.UIAAutomationId in _INTELLISENSE_LIST_AUTOMATION_IDS:\n \t\tclsList.insert(0, IntelliSenseList)\n@@ -64,3 +93,5 @@\n \t\tand isinstance(obj.previous.previous, IntelliSenseList)\n \t):\n \t\tclsList.insert(0, IntelliSenseLiveRegion)\n+\telif obj.UIAAutomationId == \"completion tooltip\":\n+\t\tclsList.insert(0, CompletionToolTip)\n", "issue": "Visual Studio: IntelliSense tooltips reported twice\n### Steps to reproduce:\r\n1. Open Visual Studio 2019\r\n2. Open a C# project\r\n3. Enable reporting of Tooltips\r\n4. Trigger an IntelliSense autocomplete suggestion by typing something.\r\n5. Arrow through the suggestions\r\n\r\n### Actual behavior:\r\nThe selected item is announced, followed by twice the tooltip.\r\n\r\n### Expected behavior:\r\nThe selected item is announced, followed by once the tooltip.\r\n\r\n### System configuration\r\n#### NVDA installed/portable/running from source:\r\nInstalled\r\n\r\n#### NVDA version:\r\nalpha-20957\r\n\r\n#### Windows version:\r\nWindows 10 2004\r\n\r\n#### Name and version of other software in use when reproducing the issue:\r\nVisual Studio 2019 16.7.3 Enterprise\r\n\r\n### Other questions\r\n#### Does the issue still occur after restarting your computer?\r\nYes\r\n\r\n#### Have you tried any other versions of NVDA? If so, please report their behaviors.\r\nNo\r\n\r\n#### If addons are disabled, is your problem still occuring?\r\nYes\r\n\r\n#### Did you try to run the COM registry fixing tool in NVDA menu / tools?\r\nYes\n", "before_files": [{"content": "# This file is covered by the GNU General Public License.\n# See the file COPYING for more details.\n# Copyright (C) 2020 NV Access Limited, Leonard de Ruijter\n\n\"\"\"\nObject overlay classes for Visual Studio components\navailable in Visual Studio and SQL Server Management Studio.\n\"\"\"\n\nfrom . import UIA\nimport speech\nimport braille\nimport api\n\n\nclass IntelliSenseItem(UIA):\n\n\tdef _get_name(self):\n\t\treturn self.UIAElement.cachedAutomationID\n\n\tdef event_UIA_elementSelected(self):\n\t\t# Cancel speech to have speech announce the selection as soon as possible.\n\t\t# This is needed because L{reportFocus} does not cancel speech.\n\t\t# Therefore, if speech wouldn't be cancelled,\n\t\t# selection announcements would queue up when changing selection rapidly.\n\t\tspeech.cancelSpeech()\n\t\tapi.setNavigatorObject(self, isFocus=True)\n\t\tself.reportFocus()\n\t\t# Display results as flash messages.\n\t\tbraille.handler.message(braille.getPropertiesBraille(\n\t\t\tname=self.name, role=self.role, positionInfo=self.positionInfo, description=self.description\n\t\t))\n\n\nclass IntelliSenseList(UIA):\n\t...\n\n\nclass IntelliSenseLiveRegion(UIA):\n\t\"\"\"\n\tVisual Studio uses both Intellisense menu item objects and a live region\n\tto communicate Intellisense selections.\n\tNVDA uses the menu item approach and therefore the live region provides doubled information\n\tand is disabled.\n\t\"\"\"\n\n\t_shouldAllowUIALiveRegionChangeEvent = False\n\n\n_INTELLISENSE_LIST_AUTOMATION_IDS = {\n\t\"listBoxCompletions\",\n\t\"CompletionList\"\n}\n\n\ndef findExtraOverlayClasses(obj, clsList):\n\tif obj.UIAAutomationId in _INTELLISENSE_LIST_AUTOMATION_IDS:\n\t\tclsList.insert(0, IntelliSenseList)\n\telif isinstance(obj.parent, IntelliSenseList) and obj.UIAElement.cachedClassName == \"IntellisenseMenuItem\":\n\t\tclsList.insert(0, IntelliSenseItem)\n\telif (\n\t\tobj.UIAElement.cachedClassName == \"LiveTextBlock\"\n\t\tand obj.previous\n\t\tand isinstance(obj.previous.previous, IntelliSenseList)\n\t):\n\t\tclsList.insert(0, IntelliSenseLiveRegion)\n", "path": "source/NVDAObjects/UIA/VisualStudio.py"}]}
| 1,405 | 529 |
gh_patches_debug_35620
|
rasdani/github-patches
|
git_diff
|
mitmproxy__mitmproxy-4741
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
GZIP request body encoding falling back to raw
#### Problem Description
GZIP requests are not decoded correctly.
This only occurs in the request and not in the response. In the response I see the body as uncompromised gzip but not in the request.
I created a own content view to test the gzip decompression:
```python
from typing import Optional
from mitmproxy import contentviews, flow
from mitmproxy import http
import gzip
class ViewGZIPCase(contentviews.View):
name = "GZIP"
def __call__(self, data, **metadata) -> contentviews.TViewResult:
return "gzip", contentviews.format_text(gzip.decompress(data))
def render_priority(
self,
data: bytes,
*,
content_type: Optional[str] = None,
flow: Optional[flow.Flow] = None,
http_message: Optional[http.Message] = None,
**unknown_metadata,
) -> float:
if http_message.headers["content-encoding"] == "GZIP":
return 1
else:
return 0
view = ViewGZIPCase()
def load(l):
contentviews.add(view)
def done():
contentviews.remove(view)
```
##### request
with selected self-made view
<img width="1637" alt="Screenshot 2021-08-05 at 16 42 57" src="https://user-images.githubusercontent.com/1338331/128370647-68c5af49-dfca-4862-a3ff-b3413c4226fd.png">
without any special view
<img width="1637" alt="Screenshot 2021-08-05 at 16 42 42" src="https://user-images.githubusercontent.com/1338331/128370652-e2001fd6-0909-491a-8c77-19cfa226b7d6.png">
##### response
Mitmproxy standalone decompression and json parsing in the response but why not in the request?
<img width="1637" alt="Screenshot 2021-08-05 at 16 47 55" src="https://user-images.githubusercontent.com/1338331/128370639-75287a4f-5c54-47e9-a126-3ff95a1dc90f.png">
#### Steps to reproduce the behavior:
1. start Mitmproxy
2. wait for requests ( iPhone 12 as client is connected)
3. open the request
#### System Information
Mitmproxy: 7.0.2
Python: 3.9.6
OpenSSL: OpenSSL 1.1.1k 25 Mar 2021
Platform: macOS-11.5.1-x86_64-i386-64bit
GZIP request body encoding falling back to raw
#### Problem Description
GZIP requests are not decoded correctly.
This only occurs in the request and not in the response. In the response I see the body as uncompromised gzip but not in the request.
I created a own content view to test the gzip decompression:
```python
from typing import Optional
from mitmproxy import contentviews, flow
from mitmproxy import http
import gzip
class ViewGZIPCase(contentviews.View):
name = "GZIP"
def __call__(self, data, **metadata) -> contentviews.TViewResult:
return "gzip", contentviews.format_text(gzip.decompress(data))
def render_priority(
self,
data: bytes,
*,
content_type: Optional[str] = None,
flow: Optional[flow.Flow] = None,
http_message: Optional[http.Message] = None,
**unknown_metadata,
) -> float:
if http_message.headers["content-encoding"] == "GZIP":
return 1
else:
return 0
view = ViewGZIPCase()
def load(l):
contentviews.add(view)
def done():
contentviews.remove(view)
```
##### request
with selected self-made view
<img width="1637" alt="Screenshot 2021-08-05 at 16 42 57" src="https://user-images.githubusercontent.com/1338331/128370647-68c5af49-dfca-4862-a3ff-b3413c4226fd.png">
without any special view
<img width="1637" alt="Screenshot 2021-08-05 at 16 42 42" src="https://user-images.githubusercontent.com/1338331/128370652-e2001fd6-0909-491a-8c77-19cfa226b7d6.png">
##### response
Mitmproxy standalone decompression and json parsing in the response but why not in the request?
<img width="1637" alt="Screenshot 2021-08-05 at 16 47 55" src="https://user-images.githubusercontent.com/1338331/128370639-75287a4f-5c54-47e9-a126-3ff95a1dc90f.png">
#### Steps to reproduce the behavior:
1. start Mitmproxy
2. wait for requests ( iPhone 12 as client is connected)
3. open the request
#### System Information
Mitmproxy: 7.0.2
Python: 3.9.6
OpenSSL: OpenSSL 1.1.1k 25 Mar 2021
Platform: macOS-11.5.1-x86_64-i386-64bit
</issue>
<code>
[start of mitmproxy/net/encoding.py]
1 """
2 Utility functions for decoding response bodies.
3 """
4
5 import codecs
6 import collections
7 from io import BytesIO
8
9 import gzip
10 import zlib
11 import brotli
12 import zstandard as zstd
13
14 from typing import Union, Optional, AnyStr, overload # noqa
15
16 # We have a shared single-element cache for encoding and decoding.
17 # This is quite useful in practice, e.g.
18 # flow.request.content = flow.request.content.replace(b"foo", b"bar")
19 # does not require an .encode() call if content does not contain b"foo"
20 CachedDecode = collections.namedtuple(
21 "CachedDecode", "encoded encoding errors decoded"
22 )
23 _cache = CachedDecode(None, None, None, None)
24
25
26 @overload
27 def decode(encoded: None, encoding: str, errors: str = 'strict') -> None:
28 ...
29
30
31 @overload
32 def decode(encoded: str, encoding: str, errors: str = 'strict') -> str:
33 ...
34
35
36 @overload
37 def decode(encoded: bytes, encoding: str, errors: str = 'strict') -> Union[str, bytes]:
38 ...
39
40
41 def decode(
42 encoded: Union[None, str, bytes], encoding: str, errors: str = 'strict'
43 ) -> Union[None, str, bytes]:
44 """
45 Decode the given input object
46
47 Returns:
48 The decoded value
49
50 Raises:
51 ValueError, if decoding fails.
52 """
53 if encoded is None:
54 return None
55
56 global _cache
57 cached = (
58 isinstance(encoded, bytes) and
59 _cache.encoded == encoded and
60 _cache.encoding == encoding and
61 _cache.errors == errors
62 )
63 if cached:
64 return _cache.decoded
65 try:
66 try:
67 decoded = custom_decode[encoding](encoded)
68 except KeyError:
69 decoded = codecs.decode(encoded, encoding, errors) # type: ignore
70 if encoding in ("gzip", "deflate", "br", "zstd"):
71 _cache = CachedDecode(encoded, encoding, errors, decoded)
72 return decoded
73 except TypeError:
74 raise
75 except Exception as e:
76 raise ValueError("{} when decoding {} with {}: {}".format(
77 type(e).__name__,
78 repr(encoded)[:10],
79 repr(encoding),
80 repr(e),
81 ))
82
83
84 @overload
85 def encode(decoded: None, encoding: str, errors: str = 'strict') -> None:
86 ...
87
88
89 @overload
90 def encode(decoded: str, encoding: str, errors: str = 'strict') -> Union[str, bytes]:
91 ...
92
93
94 @overload
95 def encode(decoded: bytes, encoding: str, errors: str = 'strict') -> bytes:
96 ...
97
98
99 def encode(decoded: Union[None, str, bytes], encoding, errors='strict') -> Union[None, str, bytes]:
100 """
101 Encode the given input object
102
103 Returns:
104 The encoded value
105
106 Raises:
107 ValueError, if encoding fails.
108 """
109 if decoded is None:
110 return None
111
112 global _cache
113 cached = (
114 isinstance(decoded, bytes) and
115 _cache.decoded == decoded and
116 _cache.encoding == encoding and
117 _cache.errors == errors
118 )
119 if cached:
120 return _cache.encoded
121 try:
122 try:
123 encoded = custom_encode[encoding](decoded)
124 except KeyError:
125 encoded = codecs.encode(decoded, encoding, errors) # type: ignore
126 if encoding in ("gzip", "deflate", "br", "zstd"):
127 _cache = CachedDecode(encoded, encoding, errors, decoded)
128 return encoded
129 except TypeError:
130 raise
131 except Exception as e:
132 raise ValueError("{} when encoding {} with {}: {}".format(
133 type(e).__name__,
134 repr(decoded)[:10],
135 repr(encoding),
136 repr(e),
137 ))
138
139
140 def identity(content):
141 """
142 Returns content unchanged. Identity is the default value of
143 Accept-Encoding headers.
144 """
145 return content
146
147
148 def decode_gzip(content: bytes) -> bytes:
149 if not content:
150 return b""
151 gfile = gzip.GzipFile(fileobj=BytesIO(content))
152 return gfile.read()
153
154
155 def encode_gzip(content: bytes) -> bytes:
156 s = BytesIO()
157 gf = gzip.GzipFile(fileobj=s, mode='wb')
158 gf.write(content)
159 gf.close()
160 return s.getvalue()
161
162
163 def decode_brotli(content: bytes) -> bytes:
164 if not content:
165 return b""
166 return brotli.decompress(content)
167
168
169 def encode_brotli(content: bytes) -> bytes:
170 return brotli.compress(content)
171
172
173 def decode_zstd(content: bytes) -> bytes:
174 if not content:
175 return b""
176 zstd_ctx = zstd.ZstdDecompressor()
177 try:
178 return zstd_ctx.decompress(content)
179 except zstd.ZstdError:
180 # If the zstd stream is streamed without a size header,
181 # try decoding with a 10MiB output buffer
182 return zstd_ctx.decompress(content, max_output_size=10 * 2 ** 20)
183
184
185 def encode_zstd(content: bytes) -> bytes:
186 zstd_ctx = zstd.ZstdCompressor()
187 return zstd_ctx.compress(content)
188
189
190 def decode_deflate(content: bytes) -> bytes:
191 """
192 Returns decompressed data for DEFLATE. Some servers may respond with
193 compressed data without a zlib header or checksum. An undocumented
194 feature of zlib permits the lenient decompression of data missing both
195 values.
196
197 http://bugs.python.org/issue5784
198 """
199 if not content:
200 return b""
201 try:
202 return zlib.decompress(content)
203 except zlib.error:
204 return zlib.decompress(content, -15)
205
206
207 def encode_deflate(content: bytes) -> bytes:
208 """
209 Returns compressed content, always including zlib header and checksum.
210 """
211 return zlib.compress(content)
212
213
214 custom_decode = {
215 "none": identity,
216 "identity": identity,
217 "gzip": decode_gzip,
218 "deflate": decode_deflate,
219 "deflateRaw": decode_deflate,
220 "br": decode_brotli,
221 "zstd": decode_zstd,
222 }
223 custom_encode = {
224 "none": identity,
225 "identity": identity,
226 "gzip": encode_gzip,
227 "deflate": encode_deflate,
228 "deflateRaw": encode_deflate,
229 "br": encode_brotli,
230 "zstd": encode_zstd,
231 }
232
233 __all__ = ["encode", "decode"]
234
[end of mitmproxy/net/encoding.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/mitmproxy/net/encoding.py b/mitmproxy/net/encoding.py
--- a/mitmproxy/net/encoding.py
+++ b/mitmproxy/net/encoding.py
@@ -52,6 +52,7 @@
"""
if encoded is None:
return None
+ encoding = encoding.lower()
global _cache
cached = (
@@ -67,7 +68,7 @@
decoded = custom_decode[encoding](encoded)
except KeyError:
decoded = codecs.decode(encoded, encoding, errors) # type: ignore
- if encoding in ("gzip", "deflate", "br", "zstd"):
+ if encoding in ("gzip", "deflate", "deflateraw", "br", "zstd"):
_cache = CachedDecode(encoded, encoding, errors, decoded)
return decoded
except TypeError:
@@ -108,6 +109,7 @@
"""
if decoded is None:
return None
+ encoding = encoding.lower()
global _cache
cached = (
@@ -123,7 +125,7 @@
encoded = custom_encode[encoding](decoded)
except KeyError:
encoded = codecs.encode(decoded, encoding, errors) # type: ignore
- if encoding in ("gzip", "deflate", "br", "zstd"):
+ if encoding in ("gzip", "deflate", "deflateraw", "br", "zstd"):
_cache = CachedDecode(encoded, encoding, errors, decoded)
return encoded
except TypeError:
@@ -216,7 +218,7 @@
"identity": identity,
"gzip": decode_gzip,
"deflate": decode_deflate,
- "deflateRaw": decode_deflate,
+ "deflateraw": decode_deflate,
"br": decode_brotli,
"zstd": decode_zstd,
}
@@ -225,7 +227,7 @@
"identity": identity,
"gzip": encode_gzip,
"deflate": encode_deflate,
- "deflateRaw": encode_deflate,
+ "deflateraw": encode_deflate,
"br": encode_brotli,
"zstd": encode_zstd,
}
|
{"golden_diff": "diff --git a/mitmproxy/net/encoding.py b/mitmproxy/net/encoding.py\n--- a/mitmproxy/net/encoding.py\n+++ b/mitmproxy/net/encoding.py\n@@ -52,6 +52,7 @@\n \"\"\"\n if encoded is None:\n return None\n+ encoding = encoding.lower()\n \n global _cache\n cached = (\n@@ -67,7 +68,7 @@\n decoded = custom_decode[encoding](encoded)\n except KeyError:\n decoded = codecs.decode(encoded, encoding, errors) # type: ignore\n- if encoding in (\"gzip\", \"deflate\", \"br\", \"zstd\"):\n+ if encoding in (\"gzip\", \"deflate\", \"deflateraw\", \"br\", \"zstd\"):\n _cache = CachedDecode(encoded, encoding, errors, decoded)\n return decoded\n except TypeError:\n@@ -108,6 +109,7 @@\n \"\"\"\n if decoded is None:\n return None\n+ encoding = encoding.lower()\n \n global _cache\n cached = (\n@@ -123,7 +125,7 @@\n encoded = custom_encode[encoding](decoded)\n except KeyError:\n encoded = codecs.encode(decoded, encoding, errors) # type: ignore\n- if encoding in (\"gzip\", \"deflate\", \"br\", \"zstd\"):\n+ if encoding in (\"gzip\", \"deflate\", \"deflateraw\", \"br\", \"zstd\"):\n _cache = CachedDecode(encoded, encoding, errors, decoded)\n return encoded\n except TypeError:\n@@ -216,7 +218,7 @@\n \"identity\": identity,\n \"gzip\": decode_gzip,\n \"deflate\": decode_deflate,\n- \"deflateRaw\": decode_deflate,\n+ \"deflateraw\": decode_deflate,\n \"br\": decode_brotli,\n \"zstd\": decode_zstd,\n }\n@@ -225,7 +227,7 @@\n \"identity\": identity,\n \"gzip\": encode_gzip,\n \"deflate\": encode_deflate,\n- \"deflateRaw\": encode_deflate,\n+ \"deflateraw\": encode_deflate,\n \"br\": encode_brotli,\n \"zstd\": encode_zstd,\n }\n", "issue": "GZIP request body encoding falling back to raw\n#### Problem Description\r\nGZIP requests are not decoded correctly.\r\nThis only occurs in the request and not in the response. In the response I see the body as uncompromised gzip but not in the request. \r\n\r\nI created a own content view to test the gzip decompression:\r\n\r\n```python\r\nfrom typing import Optional\r\nfrom mitmproxy import contentviews, flow\r\nfrom mitmproxy import http\r\nimport gzip\r\n\r\n\r\nclass ViewGZIPCase(contentviews.View):\r\n name = \"GZIP\"\r\n\r\n def __call__(self, data, **metadata) -> contentviews.TViewResult:\r\n return \"gzip\", contentviews.format_text(gzip.decompress(data))\r\n\r\n def render_priority(\r\n self,\r\n data: bytes,\r\n *,\r\n content_type: Optional[str] = None,\r\n flow: Optional[flow.Flow] = None,\r\n http_message: Optional[http.Message] = None,\r\n **unknown_metadata,\r\n ) -> float:\r\n\r\n if http_message.headers[\"content-encoding\"] == \"GZIP\":\r\n return 1\r\n\r\n else:\r\n return 0\r\n\r\n\r\nview = ViewGZIPCase()\r\n\r\n\r\ndef load(l):\r\n contentviews.add(view)\r\n\r\n\r\ndef done():\r\n contentviews.remove(view)\r\n```\r\n\r\n##### request\r\n\r\nwith selected self-made view\r\n<img width=\"1637\" alt=\"Screenshot 2021-08-05 at 16 42 57\" src=\"https://user-images.githubusercontent.com/1338331/128370647-68c5af49-dfca-4862-a3ff-b3413c4226fd.png\">\r\n\r\nwithout any special view\r\n<img width=\"1637\" alt=\"Screenshot 2021-08-05 at 16 42 42\" src=\"https://user-images.githubusercontent.com/1338331/128370652-e2001fd6-0909-491a-8c77-19cfa226b7d6.png\">\r\n\r\n\r\n##### response\r\n\r\nMitmproxy standalone decompression and json parsing in the response but why not in the request?\r\n<img width=\"1637\" alt=\"Screenshot 2021-08-05 at 16 47 55\" src=\"https://user-images.githubusercontent.com/1338331/128370639-75287a4f-5c54-47e9-a126-3ff95a1dc90f.png\">\r\n\r\n\r\n#### Steps to reproduce the behavior:\r\n1. start Mitmproxy \r\n2. wait for requests ( iPhone 12 as client is connected)\r\n3. open the request\r\n\r\n#### System Information\r\nMitmproxy: 7.0.2\r\nPython: 3.9.6\r\nOpenSSL: OpenSSL 1.1.1k 25 Mar 2021\r\nPlatform: macOS-11.5.1-x86_64-i386-64bit\r\n\r\n\r\n\nGZIP request body encoding falling back to raw\n#### Problem Description\r\nGZIP requests are not decoded correctly.\r\nThis only occurs in the request and not in the response. In the response I see the body as uncompromised gzip but not in the request. \r\n\r\nI created a own content view to test the gzip decompression:\r\n\r\n```python\r\nfrom typing import Optional\r\nfrom mitmproxy import contentviews, flow\r\nfrom mitmproxy import http\r\nimport gzip\r\n\r\n\r\nclass ViewGZIPCase(contentviews.View):\r\n name = \"GZIP\"\r\n\r\n def __call__(self, data, **metadata) -> contentviews.TViewResult:\r\n return \"gzip\", contentviews.format_text(gzip.decompress(data))\r\n\r\n def render_priority(\r\n self,\r\n data: bytes,\r\n *,\r\n content_type: Optional[str] = None,\r\n flow: Optional[flow.Flow] = None,\r\n http_message: Optional[http.Message] = None,\r\n **unknown_metadata,\r\n ) -> float:\r\n\r\n if http_message.headers[\"content-encoding\"] == \"GZIP\":\r\n return 1\r\n\r\n else:\r\n return 0\r\n\r\n\r\nview = ViewGZIPCase()\r\n\r\n\r\ndef load(l):\r\n contentviews.add(view)\r\n\r\n\r\ndef done():\r\n contentviews.remove(view)\r\n```\r\n\r\n##### request\r\n\r\nwith selected self-made view\r\n<img width=\"1637\" alt=\"Screenshot 2021-08-05 at 16 42 57\" src=\"https://user-images.githubusercontent.com/1338331/128370647-68c5af49-dfca-4862-a3ff-b3413c4226fd.png\">\r\n\r\nwithout any special view\r\n<img width=\"1637\" alt=\"Screenshot 2021-08-05 at 16 42 42\" src=\"https://user-images.githubusercontent.com/1338331/128370652-e2001fd6-0909-491a-8c77-19cfa226b7d6.png\">\r\n\r\n\r\n##### response\r\n\r\nMitmproxy standalone decompression and json parsing in the response but why not in the request?\r\n<img width=\"1637\" alt=\"Screenshot 2021-08-05 at 16 47 55\" src=\"https://user-images.githubusercontent.com/1338331/128370639-75287a4f-5c54-47e9-a126-3ff95a1dc90f.png\">\r\n\r\n\r\n#### Steps to reproduce the behavior:\r\n1. start Mitmproxy \r\n2. wait for requests ( iPhone 12 as client is connected)\r\n3. open the request\r\n\r\n#### System Information\r\nMitmproxy: 7.0.2\r\nPython: 3.9.6\r\nOpenSSL: OpenSSL 1.1.1k 25 Mar 2021\r\nPlatform: macOS-11.5.1-x86_64-i386-64bit\r\n\r\n\r\n\n", "before_files": [{"content": "\"\"\"\nUtility functions for decoding response bodies.\n\"\"\"\n\nimport codecs\nimport collections\nfrom io import BytesIO\n\nimport gzip\nimport zlib\nimport brotli\nimport zstandard as zstd\n\nfrom typing import Union, Optional, AnyStr, overload # noqa\n\n# We have a shared single-element cache for encoding and decoding.\n# This is quite useful in practice, e.g.\n# flow.request.content = flow.request.content.replace(b\"foo\", b\"bar\")\n# does not require an .encode() call if content does not contain b\"foo\"\nCachedDecode = collections.namedtuple(\n \"CachedDecode\", \"encoded encoding errors decoded\"\n)\n_cache = CachedDecode(None, None, None, None)\n\n\n@overload\ndef decode(encoded: None, encoding: str, errors: str = 'strict') -> None:\n ...\n\n\n@overload\ndef decode(encoded: str, encoding: str, errors: str = 'strict') -> str:\n ...\n\n\n@overload\ndef decode(encoded: bytes, encoding: str, errors: str = 'strict') -> Union[str, bytes]:\n ...\n\n\ndef decode(\n encoded: Union[None, str, bytes], encoding: str, errors: str = 'strict'\n) -> Union[None, str, bytes]:\n \"\"\"\n Decode the given input object\n\n Returns:\n The decoded value\n\n Raises:\n ValueError, if decoding fails.\n \"\"\"\n if encoded is None:\n return None\n\n global _cache\n cached = (\n isinstance(encoded, bytes) and\n _cache.encoded == encoded and\n _cache.encoding == encoding and\n _cache.errors == errors\n )\n if cached:\n return _cache.decoded\n try:\n try:\n decoded = custom_decode[encoding](encoded)\n except KeyError:\n decoded = codecs.decode(encoded, encoding, errors) # type: ignore\n if encoding in (\"gzip\", \"deflate\", \"br\", \"zstd\"):\n _cache = CachedDecode(encoded, encoding, errors, decoded)\n return decoded\n except TypeError:\n raise\n except Exception as e:\n raise ValueError(\"{} when decoding {} with {}: {}\".format(\n type(e).__name__,\n repr(encoded)[:10],\n repr(encoding),\n repr(e),\n ))\n\n\n@overload\ndef encode(decoded: None, encoding: str, errors: str = 'strict') -> None:\n ...\n\n\n@overload\ndef encode(decoded: str, encoding: str, errors: str = 'strict') -> Union[str, bytes]:\n ...\n\n\n@overload\ndef encode(decoded: bytes, encoding: str, errors: str = 'strict') -> bytes:\n ...\n\n\ndef encode(decoded: Union[None, str, bytes], encoding, errors='strict') -> Union[None, str, bytes]:\n \"\"\"\n Encode the given input object\n\n Returns:\n The encoded value\n\n Raises:\n ValueError, if encoding fails.\n \"\"\"\n if decoded is None:\n return None\n\n global _cache\n cached = (\n isinstance(decoded, bytes) and\n _cache.decoded == decoded and\n _cache.encoding == encoding and\n _cache.errors == errors\n )\n if cached:\n return _cache.encoded\n try:\n try:\n encoded = custom_encode[encoding](decoded)\n except KeyError:\n encoded = codecs.encode(decoded, encoding, errors) # type: ignore\n if encoding in (\"gzip\", \"deflate\", \"br\", \"zstd\"):\n _cache = CachedDecode(encoded, encoding, errors, decoded)\n return encoded\n except TypeError:\n raise\n except Exception as e:\n raise ValueError(\"{} when encoding {} with {}: {}\".format(\n type(e).__name__,\n repr(decoded)[:10],\n repr(encoding),\n repr(e),\n ))\n\n\ndef identity(content):\n \"\"\"\n Returns content unchanged. Identity is the default value of\n Accept-Encoding headers.\n \"\"\"\n return content\n\n\ndef decode_gzip(content: bytes) -> bytes:\n if not content:\n return b\"\"\n gfile = gzip.GzipFile(fileobj=BytesIO(content))\n return gfile.read()\n\n\ndef encode_gzip(content: bytes) -> bytes:\n s = BytesIO()\n gf = gzip.GzipFile(fileobj=s, mode='wb')\n gf.write(content)\n gf.close()\n return s.getvalue()\n\n\ndef decode_brotli(content: bytes) -> bytes:\n if not content:\n return b\"\"\n return brotli.decompress(content)\n\n\ndef encode_brotli(content: bytes) -> bytes:\n return brotli.compress(content)\n\n\ndef decode_zstd(content: bytes) -> bytes:\n if not content:\n return b\"\"\n zstd_ctx = zstd.ZstdDecompressor()\n try:\n return zstd_ctx.decompress(content)\n except zstd.ZstdError:\n # If the zstd stream is streamed without a size header,\n # try decoding with a 10MiB output buffer\n return zstd_ctx.decompress(content, max_output_size=10 * 2 ** 20)\n\n\ndef encode_zstd(content: bytes) -> bytes:\n zstd_ctx = zstd.ZstdCompressor()\n return zstd_ctx.compress(content)\n\n\ndef decode_deflate(content: bytes) -> bytes:\n \"\"\"\n Returns decompressed data for DEFLATE. Some servers may respond with\n compressed data without a zlib header or checksum. An undocumented\n feature of zlib permits the lenient decompression of data missing both\n values.\n\n http://bugs.python.org/issue5784\n \"\"\"\n if not content:\n return b\"\"\n try:\n return zlib.decompress(content)\n except zlib.error:\n return zlib.decompress(content, -15)\n\n\ndef encode_deflate(content: bytes) -> bytes:\n \"\"\"\n Returns compressed content, always including zlib header and checksum.\n \"\"\"\n return zlib.compress(content)\n\n\ncustom_decode = {\n \"none\": identity,\n \"identity\": identity,\n \"gzip\": decode_gzip,\n \"deflate\": decode_deflate,\n \"deflateRaw\": decode_deflate,\n \"br\": decode_brotli,\n \"zstd\": decode_zstd,\n}\ncustom_encode = {\n \"none\": identity,\n \"identity\": identity,\n \"gzip\": encode_gzip,\n \"deflate\": encode_deflate,\n \"deflateRaw\": encode_deflate,\n \"br\": encode_brotli,\n \"zstd\": encode_zstd,\n}\n\n__all__ = [\"encode\", \"decode\"]\n", "path": "mitmproxy/net/encoding.py"}]}
| 3,933 | 498 |
gh_patches_debug_17450
|
rasdani/github-patches
|
git_diff
|
DataBiosphere__toil-2931
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Toil fails on python 3.7 due to dependency
Toil has an indirect dependency on http_parser package which is no longer maintained on pypi (last update 2013). This package does not compile with python 3.7. It is required by pymesos
However, it does not appear that pymesos has released a version that corrects this.
┆Issue is synchronized with this [Jira Task](https://ucsc-cgl.atlassian.net/browse/TOIL-489)
┆Issue Number: TOIL-489
</issue>
<code>
[start of setup.py]
1 # Copyright (C) 2015-2016 Regents of the University of California
2 #
3 # Licensed under the Apache License, Version 2.0 (the "License");
4 # you may not use this file except in compliance with the License.
5 # You may obtain a copy of the License at
6 #
7 # http://www.apache.org/licenses/LICENSE-2.0
8 #
9 # Unless required by applicable law or agreed to in writing, software
10 # distributed under the License is distributed on an "AS IS" BASIS,
11 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12 # See the License for the specific language governing permissions and
13 # limitations under the License.
14
15 from setuptools import find_packages, setup
16 import sys
17
18 def runSetup():
19 """
20 Calls setup(). This function exists so the setup() invocation preceded more internal
21 functionality. The `version` module is imported dynamically by importVersion() below.
22 """
23 boto = 'boto==2.48.0'
24 boto3 = 'boto3>=1.7.50, <2.0'
25 futures = 'futures==3.1.1'
26 pycryptodome = 'pycryptodome==3.5.1'
27 pymesos = 'pymesos==0.3.7'
28 psutil = 'psutil >= 3.0.1, <6'
29 pynacl = 'pynacl==1.1.2'
30 gcs = 'google-cloud-storage==1.6.0'
31 gcs_oauth2_boto_plugin = 'gcs_oauth2_boto_plugin==1.14'
32 apacheLibcloud = 'apache-libcloud==2.2.1'
33 cwltool = 'cwltool==1.0.20190906054215'
34 schemaSalad = 'schema-salad<5,>=4.5.20190815125611'
35 galaxyLib = 'galaxy-lib==18.9.2'
36 htcondor = 'htcondor>=8.6.0'
37 kubernetes = 'kubernetes>=10, <11'
38 pytx = 'pytz>=2012'
39 dill = 'dill==0.2.7.1'
40 six = 'six>=1.10.0'
41 future = 'future'
42 requests = 'requests>=2, <3'
43 docker = 'docker==2.5.1'
44 subprocess32 = 'subprocess32<=3.5.2'
45 dateutil = 'python-dateutil'
46 addict = 'addict<=2.2.0'
47 sphinx = 'sphinx==1.7.5'
48 pathlib2 = 'pathlib2==2.3.2'
49
50 core_reqs = [
51 dill,
52 six,
53 future,
54 requests,
55 docker,
56 dateutil,
57 psutil,
58 subprocess32,
59 addict,
60 sphinx,
61 pathlib2]
62
63 aws_reqs = [
64 boto,
65 boto3,
66 futures,
67 pycryptodome]
68 cwl_reqs = [
69 cwltool,
70 schemaSalad,
71 galaxyLib]
72 encryption_reqs = [
73 pynacl]
74 google_reqs = [
75 gcs_oauth2_boto_plugin, # is this being used??
76 apacheLibcloud,
77 gcs]
78 htcondor_reqs = [
79 htcondor]
80 kubernetes_reqs = [
81 kubernetes]
82 mesos_reqs = [
83 pymesos,
84 psutil]
85 wdl_reqs = []
86
87
88 # htcondor is not supported by apple
89 # this is tricky to conditionally support in 'all' due
90 # to how wheels work, so it is not included in all and
91 # must be explicitly installed as an extra
92 all_reqs = \
93 aws_reqs + \
94 cwl_reqs + \
95 encryption_reqs + \
96 google_reqs + \
97 kubernetes_reqs + \
98 mesos_reqs
99
100 # remove the subprocess32 backport if not python2
101 if not sys.version_info[0] == 2:
102 core_reqs.remove(subprocess32)
103
104 setup(
105 name='toil',
106 version=version.distVersion,
107 description='Pipeline management software for clusters.',
108 author='Benedict Paten',
109 author_email='[email protected]',
110 url="https://github.com/BD2KGenomics/toil",
111 classifiers=[
112 'Development Status :: 5 - Production/Stable',
113 'Environment :: Console',
114 'Intended Audience :: Developers',
115 'Intended Audience :: Science/Research',
116 'Intended Audience :: Healthcare Industry',
117 'License :: OSI Approved :: Apache Software License',
118 'Natural Language :: English',
119 'Operating System :: MacOS :: MacOS X',
120 'Operating System :: POSIX',
121 'Operating System :: POSIX :: Linux',
122 'Programming Language :: Python :: 2.7',
123 'Programming Language :: Python :: 3.5',
124 'Programming Language :: Python :: 3.6',
125 'Topic :: Scientific/Engineering',
126 'Topic :: Scientific/Engineering :: Bio-Informatics',
127 'Topic :: Scientific/Engineering :: Astronomy',
128 'Topic :: Scientific/Engineering :: Atmospheric Science',
129 'Topic :: Scientific/Engineering :: Information Analysis',
130 'Topic :: Scientific/Engineering :: Medical Science Apps.',
131 'Topic :: System :: Distributed Computing',
132 'Topic :: Utilities'],
133 license="Apache License v2.0",
134 install_requires=core_reqs,
135 extras_require={
136 'aws': aws_reqs,
137 'cwl': cwl_reqs,
138 'encryption': encryption_reqs,
139 'google': google_reqs,
140 'htcondor:sys_platform!="darwin"': htcondor_reqs,
141 'kubernetes': kubernetes_reqs,
142 'mesos': mesos_reqs,
143 'wdl': wdl_reqs,
144 'all': all_reqs},
145 package_dir={'': 'src'},
146 packages=find_packages(where='src',
147 # Note that we intentionally include the top-level `test` package for
148 # functionality like the @experimental and @integrative decoratorss:
149 exclude=['*.test.*']),
150 package_data = {
151 '': ['*.yml', 'cloud-config'],
152 },
153 # Unfortunately, the names of the entry points are hard-coded elsewhere in the code base so
154 # you can't just change them here. Luckily, most of them are pretty unique strings, and thus
155 # easy to search for.
156 entry_points={
157 'console_scripts': [
158 'toil = toil.utils.toilMain:main',
159 '_toil_worker = toil.worker:main',
160 'cwltoil = toil.cwl.cwltoil:main [cwl]',
161 'toil-cwl-runner = toil.cwl.cwltoil:main [cwl]',
162 'toil-wdl-runner = toil.wdl.toilwdl:main',
163 '_toil_mesos_executor = toil.batchSystems.mesos.executor:main [mesos]',
164 '_toil_kubernetes_executor = toil.batchSystems.kubernetes:executor [kubernetes]']})
165
166
167 def importVersion():
168 """
169 Load and return the module object for src/toil/version.py, generating it from the template if
170 required.
171 """
172 import imp
173 try:
174 # Attempt to load the template first. It only exists in a working copy cloned via git.
175 import version_template
176 except ImportError:
177 # If loading the template fails we must be in a unpacked source distribution and
178 # src/toil/version.py will already exist.
179 pass
180 else:
181 # Use the template to generate src/toil/version.py
182 import os
183 import errno
184 from tempfile import NamedTemporaryFile
185
186 new = version_template.expand_()
187 try:
188 with open('src/toil/version.py') as f:
189 old = f.read()
190 except IOError as e:
191 if e.errno == errno.ENOENT:
192 old = None
193 else:
194 raise
195
196 if old != new:
197 with NamedTemporaryFile(mode='w',dir='src/toil', prefix='version.py.', delete=False) as f:
198 f.write(new)
199 os.rename(f.name, 'src/toil/version.py')
200 # Unfortunately, we can't use a straight import here because that would also load the stuff
201 # defined in src/toil/__init__.py which imports modules from external dependencies that may
202 # yet to be installed when setup.py is invoked.
203 return imp.load_source('toil.version', 'src/toil/version.py')
204
205
206 version = importVersion()
207 runSetup()
208
[end of setup.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/setup.py b/setup.py
--- a/setup.py
+++ b/setup.py
@@ -25,6 +25,11 @@
futures = 'futures==3.1.1'
pycryptodome = 'pycryptodome==3.5.1'
pymesos = 'pymesos==0.3.7'
+ # We need this specific http-parser that still claims to be version 0.8.3
+ # but which builds on Python 3.7+, to satisfy pymesos
+ http_parser = 'http-parser' + \
+ '@https://github.com/adamnovak/http-parser/archive/190a17839ba229c635b59d960579451a81fe73f3.zip' + \
+ '#sha256=3d30c84a426627e468657c44de199daee9d3210a48e392d4ad2e7497c5010949'
psutil = 'psutil >= 3.0.1, <6'
pynacl = 'pynacl==1.1.2'
gcs = 'google-cloud-storage==1.6.0'
@@ -80,6 +85,7 @@
kubernetes_reqs = [
kubernetes]
mesos_reqs = [
+ http_parser,
pymesos,
psutil]
wdl_reqs = []
|
{"golden_diff": "diff --git a/setup.py b/setup.py\n--- a/setup.py\n+++ b/setup.py\n@@ -25,6 +25,11 @@\n futures = 'futures==3.1.1'\n pycryptodome = 'pycryptodome==3.5.1'\n pymesos = 'pymesos==0.3.7'\n+ # We need this specific http-parser that still claims to be version 0.8.3\n+ # but which builds on Python 3.7+, to satisfy pymesos\n+ http_parser = 'http-parser' + \\\n+ '@https://github.com/adamnovak/http-parser/archive/190a17839ba229c635b59d960579451a81fe73f3.zip' + \\\n+ '#sha256=3d30c84a426627e468657c44de199daee9d3210a48e392d4ad2e7497c5010949'\n psutil = 'psutil >= 3.0.1, <6'\n pynacl = 'pynacl==1.1.2'\n gcs = 'google-cloud-storage==1.6.0'\n@@ -80,6 +85,7 @@\n kubernetes_reqs = [\n kubernetes]\n mesos_reqs = [\n+ http_parser,\n pymesos,\n psutil]\n wdl_reqs = []\n", "issue": "Toil fails on python 3.7 due to dependency\nToil has an indirect dependency on http_parser package which is no longer maintained on pypi (last update 2013). This package does not compile with python 3.7. It is required by pymesos\n\nHowever, it does not appear that pymesos has released a version that corrects this.\n\n\u2506Issue is synchronized with this [Jira Task](https://ucsc-cgl.atlassian.net/browse/TOIL-489)\n\u2506Issue Number: TOIL-489\n\n", "before_files": [{"content": "# Copyright (C) 2015-2016 Regents of the University of California\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom setuptools import find_packages, setup\nimport sys\n\ndef runSetup():\n \"\"\"\n Calls setup(). This function exists so the setup() invocation preceded more internal\n functionality. The `version` module is imported dynamically by importVersion() below.\n \"\"\"\n boto = 'boto==2.48.0'\n boto3 = 'boto3>=1.7.50, <2.0'\n futures = 'futures==3.1.1'\n pycryptodome = 'pycryptodome==3.5.1'\n pymesos = 'pymesos==0.3.7'\n psutil = 'psutil >= 3.0.1, <6'\n pynacl = 'pynacl==1.1.2'\n gcs = 'google-cloud-storage==1.6.0'\n gcs_oauth2_boto_plugin = 'gcs_oauth2_boto_plugin==1.14'\n apacheLibcloud = 'apache-libcloud==2.2.1'\n cwltool = 'cwltool==1.0.20190906054215'\n schemaSalad = 'schema-salad<5,>=4.5.20190815125611'\n galaxyLib = 'galaxy-lib==18.9.2'\n htcondor = 'htcondor>=8.6.0'\n kubernetes = 'kubernetes>=10, <11'\n pytx = 'pytz>=2012'\n dill = 'dill==0.2.7.1'\n six = 'six>=1.10.0'\n future = 'future'\n requests = 'requests>=2, <3'\n docker = 'docker==2.5.1'\n subprocess32 = 'subprocess32<=3.5.2'\n dateutil = 'python-dateutil'\n addict = 'addict<=2.2.0'\n sphinx = 'sphinx==1.7.5'\n pathlib2 = 'pathlib2==2.3.2'\n\n core_reqs = [\n dill,\n six,\n future,\n requests,\n docker,\n dateutil,\n psutil,\n subprocess32,\n addict,\n sphinx,\n pathlib2]\n\n aws_reqs = [\n boto,\n boto3,\n futures,\n pycryptodome]\n cwl_reqs = [\n cwltool,\n schemaSalad,\n galaxyLib]\n encryption_reqs = [\n pynacl]\n google_reqs = [\n gcs_oauth2_boto_plugin, # is this being used??\n apacheLibcloud,\n gcs]\n htcondor_reqs = [\n htcondor]\n kubernetes_reqs = [\n kubernetes]\n mesos_reqs = [\n pymesos,\n psutil]\n wdl_reqs = []\n \n\n # htcondor is not supported by apple\n # this is tricky to conditionally support in 'all' due\n # to how wheels work, so it is not included in all and\n # must be explicitly installed as an extra\n all_reqs = \\\n aws_reqs + \\\n cwl_reqs + \\\n encryption_reqs + \\\n google_reqs + \\\n kubernetes_reqs + \\\n mesos_reqs\n\n # remove the subprocess32 backport if not python2\n if not sys.version_info[0] == 2:\n core_reqs.remove(subprocess32)\n\n setup(\n name='toil',\n version=version.distVersion,\n description='Pipeline management software for clusters.',\n author='Benedict Paten',\n author_email='[email protected]',\n url=\"https://github.com/BD2KGenomics/toil\",\n classifiers=[\n 'Development Status :: 5 - Production/Stable',\n 'Environment :: Console',\n 'Intended Audience :: Developers',\n 'Intended Audience :: Science/Research',\n 'Intended Audience :: Healthcare Industry',\n 'License :: OSI Approved :: Apache Software License',\n 'Natural Language :: English',\n 'Operating System :: MacOS :: MacOS X',\n 'Operating System :: POSIX',\n 'Operating System :: POSIX :: Linux',\n 'Programming Language :: Python :: 2.7',\n 'Programming Language :: Python :: 3.5',\n 'Programming Language :: Python :: 3.6',\n 'Topic :: Scientific/Engineering',\n 'Topic :: Scientific/Engineering :: Bio-Informatics',\n 'Topic :: Scientific/Engineering :: Astronomy',\n 'Topic :: Scientific/Engineering :: Atmospheric Science',\n 'Topic :: Scientific/Engineering :: Information Analysis',\n 'Topic :: Scientific/Engineering :: Medical Science Apps.',\n 'Topic :: System :: Distributed Computing',\n 'Topic :: Utilities'],\n license=\"Apache License v2.0\",\n install_requires=core_reqs,\n extras_require={\n 'aws': aws_reqs,\n 'cwl': cwl_reqs,\n 'encryption': encryption_reqs,\n 'google': google_reqs,\n 'htcondor:sys_platform!=\"darwin\"': htcondor_reqs,\n 'kubernetes': kubernetes_reqs,\n 'mesos': mesos_reqs,\n 'wdl': wdl_reqs,\n 'all': all_reqs},\n package_dir={'': 'src'},\n packages=find_packages(where='src',\n # Note that we intentionally include the top-level `test` package for\n # functionality like the @experimental and @integrative decoratorss:\n exclude=['*.test.*']),\n package_data = {\n '': ['*.yml', 'cloud-config'],\n },\n # Unfortunately, the names of the entry points are hard-coded elsewhere in the code base so\n # you can't just change them here. Luckily, most of them are pretty unique strings, and thus\n # easy to search for.\n entry_points={\n 'console_scripts': [\n 'toil = toil.utils.toilMain:main',\n '_toil_worker = toil.worker:main',\n 'cwltoil = toil.cwl.cwltoil:main [cwl]',\n 'toil-cwl-runner = toil.cwl.cwltoil:main [cwl]',\n 'toil-wdl-runner = toil.wdl.toilwdl:main',\n '_toil_mesos_executor = toil.batchSystems.mesos.executor:main [mesos]',\n '_toil_kubernetes_executor = toil.batchSystems.kubernetes:executor [kubernetes]']})\n\n\ndef importVersion():\n \"\"\"\n Load and return the module object for src/toil/version.py, generating it from the template if\n required.\n \"\"\"\n import imp\n try:\n # Attempt to load the template first. It only exists in a working copy cloned via git.\n import version_template\n except ImportError:\n # If loading the template fails we must be in a unpacked source distribution and\n # src/toil/version.py will already exist.\n pass\n else:\n # Use the template to generate src/toil/version.py\n import os\n import errno\n from tempfile import NamedTemporaryFile\n\n new = version_template.expand_()\n try:\n with open('src/toil/version.py') as f:\n old = f.read()\n except IOError as e:\n if e.errno == errno.ENOENT:\n old = None\n else:\n raise\n\n if old != new:\n with NamedTemporaryFile(mode='w',dir='src/toil', prefix='version.py.', delete=False) as f:\n f.write(new)\n os.rename(f.name, 'src/toil/version.py')\n # Unfortunately, we can't use a straight import here because that would also load the stuff\n # defined in src/toil/__init__.py which imports modules from external dependencies that may\n # yet to be installed when setup.py is invoked.\n return imp.load_source('toil.version', 'src/toil/version.py')\n\n\nversion = importVersion()\nrunSetup()\n", "path": "setup.py"}]}
| 3,073 | 359 |
gh_patches_debug_34356
|
rasdani/github-patches
|
git_diff
|
ansible-collections__community.general-6369
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
idrac_redfish_command module - requesting job_id on the output
### Summary
idrac_redfish_command CreateBiosConfigJob command doesn't return job id now. I'm trying to check if job is completed before moving to next steps. This will ensure system is consistent state before next steps are executed.
### Issue Type
Feature Idea
### Component Name
idrac_redfish_command
### Additional Information
<!--- Paste example playbooks or commands between quotes below -->
```yaml (paste below)
- name: Create BIOS Config job
community.general.idrac_redfish_command:
baseuri: "{{ baseuri }}"
username: "{{ username }}"
password: "{{ password }}"
category: Systems
command: CreateBiosConfigJob
register: bios_job
- debug:
msg: "{{ bios_job.job_id }}
```
### Code of Conduct
- [X] I agree to follow the Ansible Code of Conduct
</issue>
<code>
[start of plugins/modules/idrac_redfish_command.py]
1 #!/usr/bin/python
2 # -*- coding: utf-8 -*-
3
4 # Copyright (c) 2018 Dell EMC Inc.
5 # GNU General Public License v3.0+ (see LICENSES/GPL-3.0-or-later.txt or https://www.gnu.org/licenses/gpl-3.0.txt)
6 # SPDX-License-Identifier: GPL-3.0-or-later
7
8 from __future__ import absolute_import, division, print_function
9 __metaclass__ = type
10
11 DOCUMENTATION = '''
12 ---
13 module: idrac_redfish_command
14 short_description: Manages Out-Of-Band controllers using iDRAC OEM Redfish APIs
15 description:
16 - Builds Redfish URIs locally and sends them to remote OOB controllers to
17 perform an action.
18 - For use with Dell iDRAC operations that require Redfish OEM extensions.
19 extends_documentation_fragment:
20 - community.general.attributes
21 attributes:
22 check_mode:
23 support: none
24 diff_mode:
25 support: none
26 options:
27 category:
28 required: true
29 description:
30 - Category to execute on iDRAC.
31 type: str
32 command:
33 required: true
34 description:
35 - List of commands to execute on iDRAC.
36 type: list
37 elements: str
38 baseuri:
39 required: true
40 description:
41 - Base URI of iDRAC.
42 type: str
43 username:
44 description:
45 - Username for authenticating to iDRAC.
46 type: str
47 password:
48 description:
49 - Password for authenticating to iDRAC.
50 type: str
51 auth_token:
52 description:
53 - Security token for authenticating to iDRAC.
54 type: str
55 version_added: 2.3.0
56 timeout:
57 description:
58 - Timeout in seconds for HTTP requests to iDRAC.
59 default: 10
60 type: int
61 resource_id:
62 required: false
63 description:
64 - ID of the System, Manager or Chassis to modify.
65 type: str
66 version_added: '0.2.0'
67
68 author: "Jose Delarosa (@jose-delarosa)"
69 '''
70
71 EXAMPLES = '''
72 - name: Create BIOS configuration job (schedule BIOS setting update)
73 community.general.idrac_redfish_command:
74 category: Systems
75 command: CreateBiosConfigJob
76 resource_id: System.Embedded.1
77 baseuri: "{{ baseuri }}"
78 username: "{{ username }}"
79 password: "{{ password }}"
80 '''
81
82 RETURN = '''
83 msg:
84 description: Message with action result or error description
85 returned: always
86 type: str
87 sample: "Action was successful"
88 '''
89
90 import re
91 from ansible.module_utils.basic import AnsibleModule
92 from ansible_collections.community.general.plugins.module_utils.redfish_utils import RedfishUtils
93 from ansible.module_utils.common.text.converters import to_native
94
95
96 class IdracRedfishUtils(RedfishUtils):
97
98 def create_bios_config_job(self):
99 result = {}
100 key = "Bios"
101 jobs = "Jobs"
102
103 # Search for 'key' entry and extract URI from it
104 response = self.get_request(self.root_uri + self.systems_uris[0])
105 if response['ret'] is False:
106 return response
107 result['ret'] = True
108 data = response['data']
109
110 if key not in data:
111 return {'ret': False, 'msg': "Key %s not found" % key}
112
113 bios_uri = data[key]["@odata.id"]
114
115 # Extract proper URI
116 response = self.get_request(self.root_uri + bios_uri)
117 if response['ret'] is False:
118 return response
119 result['ret'] = True
120 data = response['data']
121 set_bios_attr_uri = data["@Redfish.Settings"]["SettingsObject"][
122 "@odata.id"]
123
124 payload = {"TargetSettingsURI": set_bios_attr_uri}
125 response = self.post_request(
126 self.root_uri + self.manager_uri + "/" + jobs, payload)
127 if response['ret'] is False:
128 return response
129
130 response_output = response['resp'].__dict__
131 job_id = response_output["headers"]["Location"]
132 job_id = re.search("JID_.+", job_id).group()
133 # Currently not passing job_id back to user but patch is coming
134 return {'ret': True, 'msg': "Config job %s created" % job_id}
135
136
137 CATEGORY_COMMANDS_ALL = {
138 "Systems": ["CreateBiosConfigJob"],
139 "Accounts": [],
140 "Manager": []
141 }
142
143
144 def main():
145 result = {}
146 module = AnsibleModule(
147 argument_spec=dict(
148 category=dict(required=True),
149 command=dict(required=True, type='list', elements='str'),
150 baseuri=dict(required=True),
151 username=dict(),
152 password=dict(no_log=True),
153 auth_token=dict(no_log=True),
154 timeout=dict(type='int', default=10),
155 resource_id=dict()
156 ),
157 required_together=[
158 ('username', 'password'),
159 ],
160 required_one_of=[
161 ('username', 'auth_token'),
162 ],
163 mutually_exclusive=[
164 ('username', 'auth_token'),
165 ],
166 supports_check_mode=False
167 )
168
169 category = module.params['category']
170 command_list = module.params['command']
171
172 # admin credentials used for authentication
173 creds = {'user': module.params['username'],
174 'pswd': module.params['password'],
175 'token': module.params['auth_token']}
176
177 # timeout
178 timeout = module.params['timeout']
179
180 # System, Manager or Chassis ID to modify
181 resource_id = module.params['resource_id']
182
183 # Build root URI
184 root_uri = "https://" + module.params['baseuri']
185 rf_utils = IdracRedfishUtils(creds, root_uri, timeout, module,
186 resource_id=resource_id, data_modification=True)
187
188 # Check that Category is valid
189 if category not in CATEGORY_COMMANDS_ALL:
190 module.fail_json(msg=to_native("Invalid Category '%s'. Valid Categories = %s" % (category, list(CATEGORY_COMMANDS_ALL.keys()))))
191
192 # Check that all commands are valid
193 for cmd in command_list:
194 # Fail if even one command given is invalid
195 if cmd not in CATEGORY_COMMANDS_ALL[category]:
196 module.fail_json(msg=to_native("Invalid Command '%s'. Valid Commands = %s" % (cmd, CATEGORY_COMMANDS_ALL[category])))
197
198 # Organize by Categories / Commands
199
200 if category == "Systems":
201 # execute only if we find a System resource
202 result = rf_utils._find_systems_resource()
203 if result['ret'] is False:
204 module.fail_json(msg=to_native(result['msg']))
205
206 for command in command_list:
207 if command == "CreateBiosConfigJob":
208 # execute only if we find a Managers resource
209 result = rf_utils._find_managers_resource()
210 if result['ret'] is False:
211 module.fail_json(msg=to_native(result['msg']))
212 result = rf_utils.create_bios_config_job()
213
214 # Return data back or fail with proper message
215 if result['ret'] is True:
216 del result['ret']
217 module.exit_json(changed=True, msg='Action was successful')
218 else:
219 module.fail_json(msg=to_native(result['msg']))
220
221
222 if __name__ == '__main__':
223 main()
224
[end of plugins/modules/idrac_redfish_command.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/plugins/modules/idrac_redfish_command.py b/plugins/modules/idrac_redfish_command.py
--- a/plugins/modules/idrac_redfish_command.py
+++ b/plugins/modules/idrac_redfish_command.py
@@ -85,6 +85,14 @@
returned: always
type: str
sample: "Action was successful"
+return_values:
+ description: Dictionary containing command-specific response data from the action.
+ returned: on success
+ type: dict
+ version_added: 6.6.0
+ sample: {
+ "job_id": "/redfish/v1/Managers/iDRAC.Embedded.1/Jobs/JID_471269252011"
+ }
'''
import re
@@ -128,10 +136,9 @@
return response
response_output = response['resp'].__dict__
- job_id = response_output["headers"]["Location"]
- job_id = re.search("JID_.+", job_id).group()
- # Currently not passing job_id back to user but patch is coming
- return {'ret': True, 'msg': "Config job %s created" % job_id}
+ job_id_full = response_output["headers"]["Location"]
+ job_id = re.search("JID_.+", job_id_full).group()
+ return {'ret': True, 'msg': "Config job %s created" % job_id, 'job_id': job_id_full}
CATEGORY_COMMANDS_ALL = {
@@ -143,6 +150,7 @@
def main():
result = {}
+ return_values = {}
module = AnsibleModule(
argument_spec=dict(
category=dict(required=True),
@@ -210,11 +218,13 @@
if result['ret'] is False:
module.fail_json(msg=to_native(result['msg']))
result = rf_utils.create_bios_config_job()
+ if 'job_id' in result:
+ return_values['job_id'] = result['job_id']
# Return data back or fail with proper message
if result['ret'] is True:
del result['ret']
- module.exit_json(changed=True, msg='Action was successful')
+ module.exit_json(changed=True, msg='Action was successful', return_values=return_values)
else:
module.fail_json(msg=to_native(result['msg']))
|
{"golden_diff": "diff --git a/plugins/modules/idrac_redfish_command.py b/plugins/modules/idrac_redfish_command.py\n--- a/plugins/modules/idrac_redfish_command.py\n+++ b/plugins/modules/idrac_redfish_command.py\n@@ -85,6 +85,14 @@\n returned: always\n type: str\n sample: \"Action was successful\"\n+return_values:\n+ description: Dictionary containing command-specific response data from the action.\n+ returned: on success\n+ type: dict\n+ version_added: 6.6.0\n+ sample: {\n+ \"job_id\": \"/redfish/v1/Managers/iDRAC.Embedded.1/Jobs/JID_471269252011\"\n+ }\n '''\n \n import re\n@@ -128,10 +136,9 @@\n return response\n \n response_output = response['resp'].__dict__\n- job_id = response_output[\"headers\"][\"Location\"]\n- job_id = re.search(\"JID_.+\", job_id).group()\n- # Currently not passing job_id back to user but patch is coming\n- return {'ret': True, 'msg': \"Config job %s created\" % job_id}\n+ job_id_full = response_output[\"headers\"][\"Location\"]\n+ job_id = re.search(\"JID_.+\", job_id_full).group()\n+ return {'ret': True, 'msg': \"Config job %s created\" % job_id, 'job_id': job_id_full}\n \n \n CATEGORY_COMMANDS_ALL = {\n@@ -143,6 +150,7 @@\n \n def main():\n result = {}\n+ return_values = {}\n module = AnsibleModule(\n argument_spec=dict(\n category=dict(required=True),\n@@ -210,11 +218,13 @@\n if result['ret'] is False:\n module.fail_json(msg=to_native(result['msg']))\n result = rf_utils.create_bios_config_job()\n+ if 'job_id' in result:\n+ return_values['job_id'] = result['job_id']\n \n # Return data back or fail with proper message\n if result['ret'] is True:\n del result['ret']\n- module.exit_json(changed=True, msg='Action was successful')\n+ module.exit_json(changed=True, msg='Action was successful', return_values=return_values)\n else:\n module.fail_json(msg=to_native(result['msg']))\n", "issue": "idrac_redfish_command module - requesting job_id on the output\n### Summary\n\nidrac_redfish_command CreateBiosConfigJob command doesn't return job id now. I'm trying to check if job is completed before moving to next steps. This will ensure system is consistent state before next steps are executed.\n\n### Issue Type\n\nFeature Idea\n\n### Component Name\n\nidrac_redfish_command\n\n### Additional Information\n\n<!--- Paste example playbooks or commands between quotes below -->\r\n```yaml (paste below)\r\n- name: Create BIOS Config job\r\n community.general.idrac_redfish_command:\r\n baseuri: \"{{ baseuri }}\"\r\n username: \"{{ username }}\"\r\n password: \"{{ password }}\"\r\n category: Systems\r\n command: CreateBiosConfigJob\r\n register: bios_job\r\n\r\n- debug:\r\n msg: \"{{ bios_job.job_id }}\r\n\r\n\r\n```\r\n\n\n### Code of Conduct\n\n- [X] I agree to follow the Ansible Code of Conduct\n", "before_files": [{"content": "#!/usr/bin/python\n# -*- coding: utf-8 -*-\n\n# Copyright (c) 2018 Dell EMC Inc.\n# GNU General Public License v3.0+ (see LICENSES/GPL-3.0-or-later.txt or https://www.gnu.org/licenses/gpl-3.0.txt)\n# SPDX-License-Identifier: GPL-3.0-or-later\n\nfrom __future__ import absolute_import, division, print_function\n__metaclass__ = type\n\nDOCUMENTATION = '''\n---\nmodule: idrac_redfish_command\nshort_description: Manages Out-Of-Band controllers using iDRAC OEM Redfish APIs\ndescription:\n - Builds Redfish URIs locally and sends them to remote OOB controllers to\n perform an action.\n - For use with Dell iDRAC operations that require Redfish OEM extensions.\nextends_documentation_fragment:\n - community.general.attributes\nattributes:\n check_mode:\n support: none\n diff_mode:\n support: none\noptions:\n category:\n required: true\n description:\n - Category to execute on iDRAC.\n type: str\n command:\n required: true\n description:\n - List of commands to execute on iDRAC.\n type: list\n elements: str\n baseuri:\n required: true\n description:\n - Base URI of iDRAC.\n type: str\n username:\n description:\n - Username for authenticating to iDRAC.\n type: str\n password:\n description:\n - Password for authenticating to iDRAC.\n type: str\n auth_token:\n description:\n - Security token for authenticating to iDRAC.\n type: str\n version_added: 2.3.0\n timeout:\n description:\n - Timeout in seconds for HTTP requests to iDRAC.\n default: 10\n type: int\n resource_id:\n required: false\n description:\n - ID of the System, Manager or Chassis to modify.\n type: str\n version_added: '0.2.0'\n\nauthor: \"Jose Delarosa (@jose-delarosa)\"\n'''\n\nEXAMPLES = '''\n - name: Create BIOS configuration job (schedule BIOS setting update)\n community.general.idrac_redfish_command:\n category: Systems\n command: CreateBiosConfigJob\n resource_id: System.Embedded.1\n baseuri: \"{{ baseuri }}\"\n username: \"{{ username }}\"\n password: \"{{ password }}\"\n'''\n\nRETURN = '''\nmsg:\n description: Message with action result or error description\n returned: always\n type: str\n sample: \"Action was successful\"\n'''\n\nimport re\nfrom ansible.module_utils.basic import AnsibleModule\nfrom ansible_collections.community.general.plugins.module_utils.redfish_utils import RedfishUtils\nfrom ansible.module_utils.common.text.converters import to_native\n\n\nclass IdracRedfishUtils(RedfishUtils):\n\n def create_bios_config_job(self):\n result = {}\n key = \"Bios\"\n jobs = \"Jobs\"\n\n # Search for 'key' entry and extract URI from it\n response = self.get_request(self.root_uri + self.systems_uris[0])\n if response['ret'] is False:\n return response\n result['ret'] = True\n data = response['data']\n\n if key not in data:\n return {'ret': False, 'msg': \"Key %s not found\" % key}\n\n bios_uri = data[key][\"@odata.id\"]\n\n # Extract proper URI\n response = self.get_request(self.root_uri + bios_uri)\n if response['ret'] is False:\n return response\n result['ret'] = True\n data = response['data']\n set_bios_attr_uri = data[\"@Redfish.Settings\"][\"SettingsObject\"][\n \"@odata.id\"]\n\n payload = {\"TargetSettingsURI\": set_bios_attr_uri}\n response = self.post_request(\n self.root_uri + self.manager_uri + \"/\" + jobs, payload)\n if response['ret'] is False:\n return response\n\n response_output = response['resp'].__dict__\n job_id = response_output[\"headers\"][\"Location\"]\n job_id = re.search(\"JID_.+\", job_id).group()\n # Currently not passing job_id back to user but patch is coming\n return {'ret': True, 'msg': \"Config job %s created\" % job_id}\n\n\nCATEGORY_COMMANDS_ALL = {\n \"Systems\": [\"CreateBiosConfigJob\"],\n \"Accounts\": [],\n \"Manager\": []\n}\n\n\ndef main():\n result = {}\n module = AnsibleModule(\n argument_spec=dict(\n category=dict(required=True),\n command=dict(required=True, type='list', elements='str'),\n baseuri=dict(required=True),\n username=dict(),\n password=dict(no_log=True),\n auth_token=dict(no_log=True),\n timeout=dict(type='int', default=10),\n resource_id=dict()\n ),\n required_together=[\n ('username', 'password'),\n ],\n required_one_of=[\n ('username', 'auth_token'),\n ],\n mutually_exclusive=[\n ('username', 'auth_token'),\n ],\n supports_check_mode=False\n )\n\n category = module.params['category']\n command_list = module.params['command']\n\n # admin credentials used for authentication\n creds = {'user': module.params['username'],\n 'pswd': module.params['password'],\n 'token': module.params['auth_token']}\n\n # timeout\n timeout = module.params['timeout']\n\n # System, Manager or Chassis ID to modify\n resource_id = module.params['resource_id']\n\n # Build root URI\n root_uri = \"https://\" + module.params['baseuri']\n rf_utils = IdracRedfishUtils(creds, root_uri, timeout, module,\n resource_id=resource_id, data_modification=True)\n\n # Check that Category is valid\n if category not in CATEGORY_COMMANDS_ALL:\n module.fail_json(msg=to_native(\"Invalid Category '%s'. Valid Categories = %s\" % (category, list(CATEGORY_COMMANDS_ALL.keys()))))\n\n # Check that all commands are valid\n for cmd in command_list:\n # Fail if even one command given is invalid\n if cmd not in CATEGORY_COMMANDS_ALL[category]:\n module.fail_json(msg=to_native(\"Invalid Command '%s'. Valid Commands = %s\" % (cmd, CATEGORY_COMMANDS_ALL[category])))\n\n # Organize by Categories / Commands\n\n if category == \"Systems\":\n # execute only if we find a System resource\n result = rf_utils._find_systems_resource()\n if result['ret'] is False:\n module.fail_json(msg=to_native(result['msg']))\n\n for command in command_list:\n if command == \"CreateBiosConfigJob\":\n # execute only if we find a Managers resource\n result = rf_utils._find_managers_resource()\n if result['ret'] is False:\n module.fail_json(msg=to_native(result['msg']))\n result = rf_utils.create_bios_config_job()\n\n # Return data back or fail with proper message\n if result['ret'] is True:\n del result['ret']\n module.exit_json(changed=True, msg='Action was successful')\n else:\n module.fail_json(msg=to_native(result['msg']))\n\n\nif __name__ == '__main__':\n main()\n", "path": "plugins/modules/idrac_redfish_command.py"}]}
| 2,896 | 531 |
gh_patches_debug_1740
|
rasdani/github-patches
|
git_diff
|
flairNLP__flair-239
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Bug in tokenizer?
Here's a minimum viable code to reproduce:
```
from flair.data import Sentence
from flair.models import SequenceTagger
model = SequenceTagger.load("ner-ontonotes-fast")
full_text = "\"In the 1960s and 1970s...\" Then came Thierry Mugler and Gianni Versace."
sentence = Sentence(full_text, use_tokenizer=True)
model.predict(sentence)
print(f"full text : {full_text}")
print(f"text length: {len(full_text)}")
print("tag\tstart\tend\tto_original_text()")
for entity in sentence.get_spans('ner'):
print(f"{entity.tag}\t{entity.start_pos}\t{entity.end_pos}\t{entity.to_original_text()}")
```
Output:
``` $ python predict.py
full text : "In the 1960s and 1970s..." Then came Thierry Mugler and Gianni Versace.
text length: 72
tag start end to_original_text()
DATE 8 13 1960s
DATE 18 23 1970s
PERSON 81 94 ThierryMugler
PERSON 97 110 GianniVersace
```
Seems the resulting tokens have start_pos and end_pos indexes larger than the real text length. Note also that the method to_original_text() is eating the spaces, so I suppose it is related.
Any ideas about what is causing the trouble?
</issue>
<code>
[start of setup.py]
1 from setuptools import setup, find_packages
2
3 setup(
4 name='flair',
5 version='0.3.2',
6 description='A very simple framework for state-of-the-art NLP',
7 long_description=open("README.md", encoding='utf-8').read(),
8 long_description_content_type="text/markdown",
9 author='Alan Akbik',
10 author_email='[email protected]',
11 url='https://github.com/zalandoresearch/flair',
12 packages=find_packages(exclude='test'), # same as name
13 license='MIT',
14 install_requires=[
15 'torch==0.4.1',
16 'gensim==3.4.0',
17 'typing==3.6.4',
18 'tqdm==4.23.4',
19 'segtok==1.5.6',
20 'matplotlib==3.0.0',
21 'mpld3==0.3',
22 'sklearn',
23 'sqlitedict==1.6.0',
24 'deprecated==1.2.4',
25 ],
26 include_package_data=True,
27 python_requires='>=3.6',
28 )
29
[end of setup.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/setup.py b/setup.py
--- a/setup.py
+++ b/setup.py
@@ -15,8 +15,8 @@
'torch==0.4.1',
'gensim==3.4.0',
'typing==3.6.4',
- 'tqdm==4.23.4',
- 'segtok==1.5.6',
+ 'tqdm==4.26.0',
+ 'segtok==1.5.7',
'matplotlib==3.0.0',
'mpld3==0.3',
'sklearn',
|
{"golden_diff": "diff --git a/setup.py b/setup.py\n--- a/setup.py\n+++ b/setup.py\n@@ -15,8 +15,8 @@\n 'torch==0.4.1',\n 'gensim==3.4.0',\n 'typing==3.6.4',\n- 'tqdm==4.23.4',\n- 'segtok==1.5.6',\n+ 'tqdm==4.26.0',\n+ 'segtok==1.5.7',\n 'matplotlib==3.0.0',\n 'mpld3==0.3',\n 'sklearn',\n", "issue": "Bug in tokenizer?\nHere's a minimum viable code to reproduce:\r\n\r\n```\r\nfrom flair.data import Sentence\r\nfrom flair.models import SequenceTagger\r\n\r\nmodel = SequenceTagger.load(\"ner-ontonotes-fast\")\r\nfull_text = \"\\\"In the 1960s and 1970s...\\\" Then came Thierry Mugler and Gianni Versace.\"\r\nsentence = Sentence(full_text, use_tokenizer=True)\r\nmodel.predict(sentence)\r\nprint(f\"full text : {full_text}\")\r\nprint(f\"text length: {len(full_text)}\")\r\nprint(\"tag\\tstart\\tend\\tto_original_text()\")\r\nfor entity in sentence.get_spans('ner'):\r\n print(f\"{entity.tag}\\t{entity.start_pos}\\t{entity.end_pos}\\t{entity.to_original_text()}\")\r\n```\r\n\r\nOutput:\r\n\r\n``` $ python predict.py \r\nfull text : \"In the 1960s and 1970s...\" Then came Thierry Mugler and Gianni Versace.\r\ntext length: 72\r\ntag\tstart\tend\tto_original_text()\r\nDATE\t8\t13\t1960s\r\nDATE\t18\t23\t1970s\r\nPERSON\t81\t94\tThierryMugler\r\nPERSON\t97\t110\tGianniVersace\r\n```\r\nSeems the resulting tokens have start_pos and end_pos indexes larger than the real text length. Note also that the method to_original_text() is eating the spaces, so I suppose it is related.\r\n\r\nAny ideas about what is causing the trouble?\n", "before_files": [{"content": "from setuptools import setup, find_packages\n\nsetup(\n name='flair',\n version='0.3.2',\n description='A very simple framework for state-of-the-art NLP',\n long_description=open(\"README.md\", encoding='utf-8').read(),\n long_description_content_type=\"text/markdown\",\n author='Alan Akbik',\n author_email='[email protected]',\n url='https://github.com/zalandoresearch/flair',\n packages=find_packages(exclude='test'), # same as name\n license='MIT',\n install_requires=[\n 'torch==0.4.1',\n 'gensim==3.4.0',\n 'typing==3.6.4',\n 'tqdm==4.23.4',\n 'segtok==1.5.6',\n 'matplotlib==3.0.0',\n 'mpld3==0.3',\n 'sklearn',\n 'sqlitedict==1.6.0',\n 'deprecated==1.2.4',\n ],\n include_package_data=True,\n python_requires='>=3.6',\n)\n", "path": "setup.py"}]}
| 1,154 | 142 |
gh_patches_debug_21620
|
rasdani/github-patches
|
git_diff
|
openfun__marsha-98
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Integrate sentry
## Purpose
[Sentry](https://sentry.io) is a great tool to track and report application errors. Looks like a requirement for any Django project.
## Proposal
Add raven as a project requirement and configure sentry.
Reference documentation: https://docs.sentry.io/platforms/python/django/
</issue>
<code>
[start of marsha/settings.py]
1 """Django settings for marsha project.
2
3 Uses django-configurations to manage environments inheritance and the loading of some
4 config from the environment
5
6 """
7
8 from datetime import timedelta
9 import os
10
11 from django.utils.translation import gettext_lazy as _
12
13 from configurations import Configuration, values
14
15
16 class Base(Configuration):
17 """Base configuration every configuration (aka environment) should inherit from.
18
19 It depends on an environment variable that SHOULD be defined:
20 - DJANGO_SECRET_KEY
21
22 You may also want to override default configuration by setting the following
23 environment variables:
24 - DJANGO_DEBUG
25 - DATABASE_URL
26 """
27
28 BASE_DIR = os.path.dirname(__file__)
29 DATA_DIR = values.Value(os.path.join("/", "data"))
30
31 # Static files (CSS, JavaScript, Images)
32 STATICFILES_DIRS = (os.path.join(BASE_DIR, "static"),)
33 STATIC_URL = "/static/"
34 MEDIA_URL = "/media/"
35 # Allow to configure location of static/media files for non-Docker installation
36 MEDIA_ROOT = values.Value(os.path.join(str(DATA_DIR), "media"))
37 STATIC_ROOT = values.Value(os.path.join(str(DATA_DIR), "static"))
38
39 SECRET_KEY = values.SecretValue()
40
41 DEBUG = values.BooleanValue(False)
42
43 DATABASES = {
44 "default": {
45 "ENGINE": values.Value(
46 "django.db.backends.postgresql_psycopg2",
47 environ_name="DATABASE_ENGINE",
48 environ_prefix=None,
49 ),
50 "NAME": values.Value(
51 "marsha", environ_name="POSTGRES_DB", environ_prefix=None
52 ),
53 "USER": values.Value(
54 "marsha_user", environ_name="POSTGRES_USER", environ_prefix=None
55 ),
56 "PASSWORD": values.Value(
57 "pass", environ_name="POSTGRES_PASSWORD", environ_prefix=None
58 ),
59 "HOST": values.Value(
60 "localhost", environ_name="POSTGRES_HOST", environ_prefix=None
61 ),
62 "PORT": values.Value(
63 5432, environ_name="POSTGRES_PORT", environ_prefix=None
64 ),
65 }
66 }
67
68 ALLOWED_HOSTS = []
69
70 # Application definition
71
72 INSTALLED_APPS = [
73 "django.contrib.admin.apps.SimpleAdminConfig",
74 "django.contrib.auth",
75 "django.contrib.contenttypes",
76 "django.contrib.sessions",
77 "django.contrib.messages",
78 "django.contrib.staticfiles",
79 "django_extensions",
80 "rest_framework",
81 "marsha.core.apps.CoreConfig",
82 ]
83
84 MIDDLEWARE = [
85 "django.middleware.security.SecurityMiddleware",
86 "django.contrib.sessions.middleware.SessionMiddleware",
87 "django.middleware.common.CommonMiddleware",
88 "django.middleware.csrf.CsrfViewMiddleware",
89 "django.contrib.auth.middleware.AuthenticationMiddleware",
90 "django.contrib.messages.middleware.MessageMiddleware",
91 "django.middleware.clickjacking.XFrameOptionsMiddleware",
92 ]
93
94 ROOT_URLCONF = "marsha.urls"
95
96 TEMPLATES = [
97 {
98 "BACKEND": "django.template.backends.django.DjangoTemplates",
99 "DIRS": [],
100 "APP_DIRS": True,
101 "OPTIONS": {
102 "context_processors": [
103 "django.template.context_processors.debug",
104 "django.template.context_processors.request",
105 "django.contrib.auth.context_processors.auth",
106 "django.contrib.messages.context_processors.messages",
107 ]
108 },
109 }
110 ]
111
112 AUTH_USER_MODEL = "core.User"
113
114 WSGI_APPLICATION = "marsha.wsgi.application"
115
116 REST_FRAMEWORK = {
117 "DEFAULT_AUTHENTICATION_CLASSES": (
118 "rest_framework_simplejwt.authentication.JWTTokenUserAuthentication",
119 )
120 }
121
122 # Password validation
123 # https://docs.djangoproject.com/en/2.0/ref/settings/#auth-password-validators
124 AUTH_PASSWORD_VALIDATORS = [
125 {
126 "NAME": "django.contrib.auth.password_validation.UserAttributeSimilarityValidator"
127 },
128 {"NAME": "django.contrib.auth.password_validation.MinimumLengthValidator"},
129 {"NAME": "django.contrib.auth.password_validation.CommonPasswordValidator"},
130 {"NAME": "django.contrib.auth.password_validation.NumericPasswordValidator"},
131 ]
132
133 JWT_SIGNING_KEY = values.SecretValue()
134
135 # Internationalization
136 # https://docs.djangoproject.com/en/2.0/topics/i18n/
137
138 LANGUAGE_CODE = "en-us"
139
140 LANGUAGES = [("en", _("english")), ("fr", _("french"))]
141
142 # Internationalization
143 TIME_ZONE = "UTC"
144 USE_I18N = True
145 USE_L10N = True
146 USE_TZ = True
147
148 VIDEO_RESOLUTIONS = [144, 240, 480, 720, 1080]
149
150 # AWS
151 AWS_ACCESS_KEY_ID = values.SecretValue()
152 AWS_SECRET_ACCESS_KEY = values.SecretValue()
153 AWS_DEFAULT_REGION = values.Value("eu-west-1")
154 UPDATE_STATE_SHARED_SECRETS = values.ListValue()
155
156 # Cloud Front key pair for signed urls
157 CLOUDFRONT_URL = values.SecretValue()
158 CLOUDFRONT_ACCESS_KEY_ID = values.Value(None)
159 CLOUDFRONT_PRIVATE_KEY_PATH = values.Value(
160 os.path.join(BASE_DIR, "..", ".ssh", "cloudfront_private_key")
161 )
162 CLOUDFRONT_SIGNED_URLS_ACTIVE = True
163 CLOUDFRONT_SIGNED_URLS_VALIDITY = 2 * 60 * 60 # 2 hours
164
165 # pylint: disable=invalid-name
166 @property
167 def SIMPLE_JWT(self):
168 """Define settings for `djangorestframework_simplejwt`.
169
170 The JWT_SIGNING_KEY must be evaluated late as the jwt library check for string type.
171 """
172 return {
173 "ACCESS_TOKEN_LIFETIME": timedelta(days=1),
174 "ALGORITHM": "HS256",
175 "SIGNING_KEY": str(self.JWT_SIGNING_KEY),
176 "USER_ID_CLAIM": "video_id",
177 "AUTH_TOKEN_CLASSES": ("rest_framework_simplejwt.tokens.AccessToken",),
178 }
179
180
181 class Development(Base):
182 """Development environment settings.
183
184 We set ``DEBUG`` to ``True`` by default, configure the server to respond to all hosts,
185 and use a local sqlite database by default.
186 """
187
188 DEBUG = values.BooleanValue(True)
189 ALLOWED_HOSTS = ["*"]
190
191 AWS_SOURCE_BUCKET_NAME = "development-marsha-source"
192
193
194 class Test(Base):
195 """Test environment settings."""
196
197 AWS_SOURCE_BUCKET_NAME = "test-marsha-source"
198
199 CLOUDFRONT_SIGNED_URLS_ACTIVE = False
200
[end of marsha/settings.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/marsha/settings.py b/marsha/settings.py
--- a/marsha/settings.py
+++ b/marsha/settings.py
@@ -11,6 +11,8 @@
from django.utils.translation import gettext_lazy as _
from configurations import Configuration, values
+import sentry_sdk
+from sentry_sdk.integrations.django import DjangoIntegration
class Base(Configuration):
@@ -177,6 +179,25 @@
"AUTH_TOKEN_CLASSES": ("rest_framework_simplejwt.tokens.AccessToken",),
}
+ @classmethod
+ def post_setup(cls):
+ """Post setup configuration.
+
+ This is the place where you can configure settings that require other
+ settings to be loaded.
+ """
+ super().post_setup()
+
+ # The DJANGO_SENTRY_DSN environment variable should be set to activate
+ # sentry for an environment
+ sentry_dsn = values.Value(None, environ_name="SENTRY_DSN")
+ if sentry_dsn is not None:
+ sentry_sdk.init(
+ dsn=sentry_dsn,
+ environment=cls.__name__.lower(),
+ integrations=[DjangoIntegration()],
+ )
+
class Development(Base):
"""Development environment settings.
|
{"golden_diff": "diff --git a/marsha/settings.py b/marsha/settings.py\n--- a/marsha/settings.py\n+++ b/marsha/settings.py\n@@ -11,6 +11,8 @@\n from django.utils.translation import gettext_lazy as _\n \n from configurations import Configuration, values\n+import sentry_sdk\n+from sentry_sdk.integrations.django import DjangoIntegration\n \n \n class Base(Configuration):\n@@ -177,6 +179,25 @@\n \"AUTH_TOKEN_CLASSES\": (\"rest_framework_simplejwt.tokens.AccessToken\",),\n }\n \n+ @classmethod\n+ def post_setup(cls):\n+ \"\"\"Post setup configuration.\n+\n+ This is the place where you can configure settings that require other\n+ settings to be loaded.\n+ \"\"\"\n+ super().post_setup()\n+\n+ # The DJANGO_SENTRY_DSN environment variable should be set to activate\n+ # sentry for an environment\n+ sentry_dsn = values.Value(None, environ_name=\"SENTRY_DSN\")\n+ if sentry_dsn is not None:\n+ sentry_sdk.init(\n+ dsn=sentry_dsn,\n+ environment=cls.__name__.lower(),\n+ integrations=[DjangoIntegration()],\n+ )\n+\n \n class Development(Base):\n \"\"\"Development environment settings.\n", "issue": "Integrate sentry\n## Purpose\r\n\r\n[Sentry](https://sentry.io) is a great tool to track and report application errors. Looks like a requirement for any Django project.\r\n\r\n## Proposal\r\n\r\nAdd raven as a project requirement and configure sentry.\r\n\r\nReference documentation: https://docs.sentry.io/platforms/python/django/\n", "before_files": [{"content": "\"\"\"Django settings for marsha project.\n\nUses django-configurations to manage environments inheritance and the loading of some\nconfig from the environment\n\n\"\"\"\n\nfrom datetime import timedelta\nimport os\n\nfrom django.utils.translation import gettext_lazy as _\n\nfrom configurations import Configuration, values\n\n\nclass Base(Configuration):\n \"\"\"Base configuration every configuration (aka environment) should inherit from.\n\n It depends on an environment variable that SHOULD be defined:\n - DJANGO_SECRET_KEY\n\n You may also want to override default configuration by setting the following\n environment variables:\n - DJANGO_DEBUG\n - DATABASE_URL\n \"\"\"\n\n BASE_DIR = os.path.dirname(__file__)\n DATA_DIR = values.Value(os.path.join(\"/\", \"data\"))\n\n # Static files (CSS, JavaScript, Images)\n STATICFILES_DIRS = (os.path.join(BASE_DIR, \"static\"),)\n STATIC_URL = \"/static/\"\n MEDIA_URL = \"/media/\"\n # Allow to configure location of static/media files for non-Docker installation\n MEDIA_ROOT = values.Value(os.path.join(str(DATA_DIR), \"media\"))\n STATIC_ROOT = values.Value(os.path.join(str(DATA_DIR), \"static\"))\n\n SECRET_KEY = values.SecretValue()\n\n DEBUG = values.BooleanValue(False)\n\n DATABASES = {\n \"default\": {\n \"ENGINE\": values.Value(\n \"django.db.backends.postgresql_psycopg2\",\n environ_name=\"DATABASE_ENGINE\",\n environ_prefix=None,\n ),\n \"NAME\": values.Value(\n \"marsha\", environ_name=\"POSTGRES_DB\", environ_prefix=None\n ),\n \"USER\": values.Value(\n \"marsha_user\", environ_name=\"POSTGRES_USER\", environ_prefix=None\n ),\n \"PASSWORD\": values.Value(\n \"pass\", environ_name=\"POSTGRES_PASSWORD\", environ_prefix=None\n ),\n \"HOST\": values.Value(\n \"localhost\", environ_name=\"POSTGRES_HOST\", environ_prefix=None\n ),\n \"PORT\": values.Value(\n 5432, environ_name=\"POSTGRES_PORT\", environ_prefix=None\n ),\n }\n }\n\n ALLOWED_HOSTS = []\n\n # Application definition\n\n INSTALLED_APPS = [\n \"django.contrib.admin.apps.SimpleAdminConfig\",\n \"django.contrib.auth\",\n \"django.contrib.contenttypes\",\n \"django.contrib.sessions\",\n \"django.contrib.messages\",\n \"django.contrib.staticfiles\",\n \"django_extensions\",\n \"rest_framework\",\n \"marsha.core.apps.CoreConfig\",\n ]\n\n MIDDLEWARE = [\n \"django.middleware.security.SecurityMiddleware\",\n \"django.contrib.sessions.middleware.SessionMiddleware\",\n \"django.middleware.common.CommonMiddleware\",\n \"django.middleware.csrf.CsrfViewMiddleware\",\n \"django.contrib.auth.middleware.AuthenticationMiddleware\",\n \"django.contrib.messages.middleware.MessageMiddleware\",\n \"django.middleware.clickjacking.XFrameOptionsMiddleware\",\n ]\n\n ROOT_URLCONF = \"marsha.urls\"\n\n TEMPLATES = [\n {\n \"BACKEND\": \"django.template.backends.django.DjangoTemplates\",\n \"DIRS\": [],\n \"APP_DIRS\": True,\n \"OPTIONS\": {\n \"context_processors\": [\n \"django.template.context_processors.debug\",\n \"django.template.context_processors.request\",\n \"django.contrib.auth.context_processors.auth\",\n \"django.contrib.messages.context_processors.messages\",\n ]\n },\n }\n ]\n\n AUTH_USER_MODEL = \"core.User\"\n\n WSGI_APPLICATION = \"marsha.wsgi.application\"\n\n REST_FRAMEWORK = {\n \"DEFAULT_AUTHENTICATION_CLASSES\": (\n \"rest_framework_simplejwt.authentication.JWTTokenUserAuthentication\",\n )\n }\n\n # Password validation\n # https://docs.djangoproject.com/en/2.0/ref/settings/#auth-password-validators\n AUTH_PASSWORD_VALIDATORS = [\n {\n \"NAME\": \"django.contrib.auth.password_validation.UserAttributeSimilarityValidator\"\n },\n {\"NAME\": \"django.contrib.auth.password_validation.MinimumLengthValidator\"},\n {\"NAME\": \"django.contrib.auth.password_validation.CommonPasswordValidator\"},\n {\"NAME\": \"django.contrib.auth.password_validation.NumericPasswordValidator\"},\n ]\n\n JWT_SIGNING_KEY = values.SecretValue()\n\n # Internationalization\n # https://docs.djangoproject.com/en/2.0/topics/i18n/\n\n LANGUAGE_CODE = \"en-us\"\n\n LANGUAGES = [(\"en\", _(\"english\")), (\"fr\", _(\"french\"))]\n\n # Internationalization\n TIME_ZONE = \"UTC\"\n USE_I18N = True\n USE_L10N = True\n USE_TZ = True\n\n VIDEO_RESOLUTIONS = [144, 240, 480, 720, 1080]\n\n # AWS\n AWS_ACCESS_KEY_ID = values.SecretValue()\n AWS_SECRET_ACCESS_KEY = values.SecretValue()\n AWS_DEFAULT_REGION = values.Value(\"eu-west-1\")\n UPDATE_STATE_SHARED_SECRETS = values.ListValue()\n\n # Cloud Front key pair for signed urls\n CLOUDFRONT_URL = values.SecretValue()\n CLOUDFRONT_ACCESS_KEY_ID = values.Value(None)\n CLOUDFRONT_PRIVATE_KEY_PATH = values.Value(\n os.path.join(BASE_DIR, \"..\", \".ssh\", \"cloudfront_private_key\")\n )\n CLOUDFRONT_SIGNED_URLS_ACTIVE = True\n CLOUDFRONT_SIGNED_URLS_VALIDITY = 2 * 60 * 60 # 2 hours\n\n # pylint: disable=invalid-name\n @property\n def SIMPLE_JWT(self):\n \"\"\"Define settings for `djangorestframework_simplejwt`.\n\n The JWT_SIGNING_KEY must be evaluated late as the jwt library check for string type.\n \"\"\"\n return {\n \"ACCESS_TOKEN_LIFETIME\": timedelta(days=1),\n \"ALGORITHM\": \"HS256\",\n \"SIGNING_KEY\": str(self.JWT_SIGNING_KEY),\n \"USER_ID_CLAIM\": \"video_id\",\n \"AUTH_TOKEN_CLASSES\": (\"rest_framework_simplejwt.tokens.AccessToken\",),\n }\n\n\nclass Development(Base):\n \"\"\"Development environment settings.\n\n We set ``DEBUG`` to ``True`` by default, configure the server to respond to all hosts,\n and use a local sqlite database by default.\n \"\"\"\n\n DEBUG = values.BooleanValue(True)\n ALLOWED_HOSTS = [\"*\"]\n\n AWS_SOURCE_BUCKET_NAME = \"development-marsha-source\"\n\n\nclass Test(Base):\n \"\"\"Test environment settings.\"\"\"\n\n AWS_SOURCE_BUCKET_NAME = \"test-marsha-source\"\n\n CLOUDFRONT_SIGNED_URLS_ACTIVE = False\n", "path": "marsha/settings.py"}]}
| 2,474 | 281 |
gh_patches_debug_18422
|
rasdani/github-patches
|
git_diff
|
ansible__awx-7280
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Mattermost Notification fails on latest release
##### ISSUE TYPE
- Bug Report
##### SUMMARY
Trying to send a (test) notification to a Mattermost Channel fails with
```
mattermostinfo: Notification failed.
Error sending notification mattermost: {"id":"api.webhook.incoming.error","message":"Could not decode the multipart payload of incoming webhook.","detailed_error":"","request_id":"<request ID>","status_code":400}
```
##### ENVIRONMENT
* AWX version: 11.2.0
* AWX install method: docker on linux
* Ansible version: 2.9.7
* Operating System: CentOS 7.8
* Web Browser: Chrome,Chromium,Firefox
* Mattermost Server Version: 5.22.1
##### STEPS TO REPRODUCE
- Create an incomming webhook
- Create a mattermost notification
- Send a test notification
##### EXPECTED RESULTS
Having a notification in the Channel
##### ACTUAL RESULTS
Sending failed with above error message
##### ADDITIONAL INFORMATION
The error message in the mattermost log shows
```
{"level":"error","ts":1591342011.6592789,"caller":"mlog/log.go:175","msg":"Could not decode the multipart payload of incoming webhook.","path":"/
hooks/<hook ID>","request_id":"<request ID>","ip_addr":"<IP Address>","user_id":"","method":"POST","err_where":"
incomingWebhook","http_code":400,"err_details":"mime: no media type"}
```
---
edit: some ID removed in the log sample, mattermost server version added
</issue>
<code>
[start of awx/main/notifications/mattermost_backend.py]
1 # Copyright (c) 2016 Ansible, Inc.
2 # All Rights Reserved.
3
4 import logging
5 import requests
6 import json
7
8 from django.utils.encoding import smart_text
9 from django.utils.translation import ugettext_lazy as _
10
11 from awx.main.notifications.base import AWXBaseEmailBackend
12 from awx.main.notifications.custom_notification_base import CustomNotificationBase
13
14 logger = logging.getLogger('awx.main.notifications.mattermost_backend')
15
16
17 class MattermostBackend(AWXBaseEmailBackend, CustomNotificationBase):
18
19 init_parameters = {"mattermost_url": {"label": "Target URL", "type": "string"},
20 "mattermost_no_verify_ssl": {"label": "Verify SSL", "type": "bool"}}
21 recipient_parameter = "mattermost_url"
22 sender_parameter = None
23
24 def __init__(self, mattermost_no_verify_ssl=False, mattermost_channel=None, mattermost_username=None,
25 mattermost_icon_url=None, fail_silently=False, **kwargs):
26 super(MattermostBackend, self).__init__(fail_silently=fail_silently)
27 self.mattermost_channel = mattermost_channel
28 self.mattermost_username = mattermost_username
29 self.mattermost_icon_url = mattermost_icon_url
30 self.mattermost_no_verify_ssl = mattermost_no_verify_ssl
31
32 def format_body(self, body):
33 return body
34
35 def send_messages(self, messages):
36 sent_messages = 0
37 for m in messages:
38 payload = {}
39 for opt, optval in {'mattermost_icon_url':'icon_url',
40 'mattermost_channel': 'channel', 'mattermost_username': 'username'}.items():
41 optvalue = getattr(self, opt)
42 if optvalue is not None:
43 payload[optval] = optvalue.strip()
44
45 payload['text'] = m.subject
46
47 r = requests.post("{}".format(m.recipients()[0]),
48 data=json.dumps(payload), verify=(not self.mattermost_no_verify_ssl))
49 if r.status_code >= 400:
50 logger.error(smart_text(_("Error sending notification mattermost: {}").format(r.text)))
51 if not self.fail_silently:
52 raise Exception(smart_text(_("Error sending notification mattermost: {}").format(r.text)))
53 sent_messages += 1
54 return sent_messages
55
[end of awx/main/notifications/mattermost_backend.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/awx/main/notifications/mattermost_backend.py b/awx/main/notifications/mattermost_backend.py
--- a/awx/main/notifications/mattermost_backend.py
+++ b/awx/main/notifications/mattermost_backend.py
@@ -3,7 +3,6 @@
import logging
import requests
-import json
from django.utils.encoding import smart_text
from django.utils.translation import ugettext_lazy as _
@@ -45,7 +44,7 @@
payload['text'] = m.subject
r = requests.post("{}".format(m.recipients()[0]),
- data=json.dumps(payload), verify=(not self.mattermost_no_verify_ssl))
+ json=payload, verify=(not self.mattermost_no_verify_ssl))
if r.status_code >= 400:
logger.error(smart_text(_("Error sending notification mattermost: {}").format(r.text)))
if not self.fail_silently:
|
{"golden_diff": "diff --git a/awx/main/notifications/mattermost_backend.py b/awx/main/notifications/mattermost_backend.py\n--- a/awx/main/notifications/mattermost_backend.py\n+++ b/awx/main/notifications/mattermost_backend.py\n@@ -3,7 +3,6 @@\n \n import logging\n import requests\n-import json\n \n from django.utils.encoding import smart_text\n from django.utils.translation import ugettext_lazy as _\n@@ -45,7 +44,7 @@\n payload['text'] = m.subject\n \n r = requests.post(\"{}\".format(m.recipients()[0]),\n- data=json.dumps(payload), verify=(not self.mattermost_no_verify_ssl))\n+ json=payload, verify=(not self.mattermost_no_verify_ssl))\n if r.status_code >= 400:\n logger.error(smart_text(_(\"Error sending notification mattermost: {}\").format(r.text)))\n if not self.fail_silently:\n", "issue": "Mattermost Notification fails on latest release\n##### ISSUE TYPE\r\n - Bug Report\r\n\r\n##### SUMMARY\r\nTrying to send a (test) notification to a Mattermost Channel fails with\r\n```\r\n mattermostinfo: Notification failed.\r\nError sending notification mattermost: {\"id\":\"api.webhook.incoming.error\",\"message\":\"Could not decode the multipart payload of incoming webhook.\",\"detailed_error\":\"\",\"request_id\":\"<request ID>\",\"status_code\":400}\r\n```\r\n##### ENVIRONMENT\r\n* AWX version: 11.2.0\r\n* AWX install method: docker on linux\r\n* Ansible version: 2.9.7\r\n* Operating System: CentOS 7.8\r\n* Web Browser: Chrome,Chromium,Firefox\r\n* Mattermost Server Version: 5.22.1\r\n\r\n##### STEPS TO REPRODUCE\r\n- Create an incomming webhook\r\n- Create a mattermost notification\r\n- Send a test notification\r\n\r\n\r\n##### EXPECTED RESULTS\r\nHaving a notification in the Channel\r\n\r\n\r\n##### ACTUAL RESULTS\r\n\r\nSending failed with above error message\r\n\r\n##### ADDITIONAL INFORMATION\r\n\r\nThe error message in the mattermost log shows\r\n```\r\n{\"level\":\"error\",\"ts\":1591342011.6592789,\"caller\":\"mlog/log.go:175\",\"msg\":\"Could not decode the multipart payload of incoming webhook.\",\"path\":\"/\r\nhooks/<hook ID>\",\"request_id\":\"<request ID>\",\"ip_addr\":\"<IP Address>\",\"user_id\":\"\",\"method\":\"POST\",\"err_where\":\"\r\nincomingWebhook\",\"http_code\":400,\"err_details\":\"mime: no media type\"}\r\n```\r\n---\r\nedit: some ID removed in the log sample, mattermost server version added\n", "before_files": [{"content": "# Copyright (c) 2016 Ansible, Inc.\n# All Rights Reserved.\n\nimport logging\nimport requests\nimport json\n\nfrom django.utils.encoding import smart_text\nfrom django.utils.translation import ugettext_lazy as _\n\nfrom awx.main.notifications.base import AWXBaseEmailBackend\nfrom awx.main.notifications.custom_notification_base import CustomNotificationBase\n\nlogger = logging.getLogger('awx.main.notifications.mattermost_backend')\n\n\nclass MattermostBackend(AWXBaseEmailBackend, CustomNotificationBase):\n\n init_parameters = {\"mattermost_url\": {\"label\": \"Target URL\", \"type\": \"string\"},\n \"mattermost_no_verify_ssl\": {\"label\": \"Verify SSL\", \"type\": \"bool\"}}\n recipient_parameter = \"mattermost_url\"\n sender_parameter = None\n\n def __init__(self, mattermost_no_verify_ssl=False, mattermost_channel=None, mattermost_username=None,\n mattermost_icon_url=None, fail_silently=False, **kwargs):\n super(MattermostBackend, self).__init__(fail_silently=fail_silently)\n self.mattermost_channel = mattermost_channel\n self.mattermost_username = mattermost_username\n self.mattermost_icon_url = mattermost_icon_url\n self.mattermost_no_verify_ssl = mattermost_no_verify_ssl\n\n def format_body(self, body):\n return body\n\n def send_messages(self, messages):\n sent_messages = 0\n for m in messages:\n payload = {}\n for opt, optval in {'mattermost_icon_url':'icon_url',\n 'mattermost_channel': 'channel', 'mattermost_username': 'username'}.items():\n optvalue = getattr(self, opt)\n if optvalue is not None:\n payload[optval] = optvalue.strip()\n\n payload['text'] = m.subject\n\n r = requests.post(\"{}\".format(m.recipients()[0]),\n data=json.dumps(payload), verify=(not self.mattermost_no_verify_ssl))\n if r.status_code >= 400:\n logger.error(smart_text(_(\"Error sending notification mattermost: {}\").format(r.text)))\n if not self.fail_silently:\n raise Exception(smart_text(_(\"Error sending notification mattermost: {}\").format(r.text)))\n sent_messages += 1\n return sent_messages\n", "path": "awx/main/notifications/mattermost_backend.py"}]}
| 1,503 | 203 |
gh_patches_debug_8442
|
rasdani/github-patches
|
git_diff
|
PaddlePaddle__PaddleSeg-980
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
DecoupledSegNet错误
1.环境:(首先安装cpu版本,测试成功)

2.安装GPU版本
命令:python -m pip install paddlepaddle-gpu==2.0.0 -i https://mirror.baidu.com/pypi/simple
3.出现问题

4.解决方案
(1)env | grep cuda
(2)port LD_LIBRARY_PATH=$LD_LIBRARY_PATH:/usr/local/cuda/lib64
(3) /usr/local/cuda/lib64
(4)ln -s libcusolver.so.10 libcusolver.so
5. 运行
Python./train.py -config ./configs/decoupled_segnet/decoupledsegnet_resnet50_os8_cityscapes_832x832_80k.yml
6.新问题

</issue>
<code>
[start of paddleseg/models/decoupled_segnet.py]
1 # Copyright (c) 2020 PaddlePaddle Authors. All Rights Reserved.
2 #
3 # Licensed under the Apache License, Version 2.0 (the "License");
4 # you may not use this file except in compliance with the License.
5 # You may obtain a copy of the License at
6 #
7 # http://www.apache.org/licenses/LICENSE-2.0
8 #
9 # Unless required by applicable law or agreed to in writing, software
10 # distributed under the License is distributed on an "AS IS" BASIS,
11 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12 # See the License for the specific language governing permissions and
13 # limitations under the License.
14
15 import cv2
16 import numpy as np
17 import paddle
18 import paddle.nn as nn
19 import paddle.nn.functional as F
20
21 from paddleseg.cvlibs import manager
22 from paddleseg.models import layers
23 from paddleseg.models.backbones import resnet_vd
24 from paddleseg.models import deeplab
25 from paddleseg.utils import utils
26
27
28 @manager.MODELS.add_component
29 class DecoupledSegNet(nn.Layer):
30 """
31 The DecoupledSegNet implementation based on PaddlePaddle.
32
33 The original article refers to
34 Xiangtai Li, et, al. "Improving Semantic Segmentation via Decoupled Body and Edge Supervision"
35 (https://arxiv.org/pdf/2007.10035.pdf)
36
37 Args:
38 num_classes (int): The unique number of target classes.
39 backbone (paddle.nn.Layer): Backbone network, currently support Resnet50_vd/Resnet101_vd.
40 backbone_indices (tuple, optional): Two values in the tuple indicate the indices of output of backbone.
41 Default: (0, 3).
42 aspp_ratios (tuple, optional): The dilation rate using in ASSP module.
43 If output_stride=16, aspp_ratios should be set as (1, 6, 12, 18).
44 If output_stride=8, aspp_ratios is (1, 12, 24, 36).
45 Default: (1, 6, 12, 18).
46 aspp_out_channels (int, optional): The output channels of ASPP module. Default: 256.
47 align_corners (bool, optional): An argument of F.interpolate. It should be set to False when the feature size is even,
48 e.g. 1024x512, otherwise it is True, e.g. 769x769. Default: False.
49 pretrained (str, optional): The path or url of pretrained model. Default: None.
50 """
51
52 def __init__(self,
53 num_classes,
54 backbone,
55 backbone_indices=(0, 3),
56 aspp_ratios=(1, 6, 12, 18),
57 aspp_out_channels=256,
58 align_corners=False,
59 pretrained=None):
60 super().__init__()
61 self.backbone = backbone
62 backbone_channels = self.backbone.feat_channels
63 self.head = DecoupledSegNetHead(num_classes, backbone_indices,
64 backbone_channels, aspp_ratios,
65 aspp_out_channels, align_corners)
66 self.align_corners = align_corners
67 self.pretrained = pretrained
68 self.init_weight()
69
70 def forward(self, x):
71 feat_list = self.backbone(x)
72 logit_list = self.head(feat_list)
73
74 seg_logit, body_logit, edge_logit = [
75 F.interpolate(
76 logit,
77 paddle.shape(x)[2:],
78 mode='bilinear',
79 align_corners=self.align_corners) for logit in logit_list
80 ]
81
82 return [seg_logit, body_logit, edge_logit, (seg_logit, edge_logit)]
83
84 def init_weight(self):
85 if self.pretrained is not None:
86 utils.load_entire_model(self, self.pretrained)
87
88
89 class DecoupledSegNetHead(nn.Layer):
90 """
91 The DecoupledSegNetHead implementation based on PaddlePaddle.
92
93 Args:
94 num_classes (int): The unique number of target classes.
95 backbone_indices (tuple): Two values in the tuple indicate the indices of output of backbone.
96 the first index will be taken as a low-level feature in Edge presevation component;
97 the second one will be taken as input of ASPP component.
98 backbone_channels (tuple): The channels of output of backbone.
99 aspp_ratios (tuple): The dilation rates using in ASSP module.
100 aspp_out_channels (int): The output channels of ASPP module.
101 align_corners (bool): An argument of F.interpolate. It should be set to False when the output size of feature
102 is even, e.g. 1024x512, otherwise it is True, e.g. 769x769.
103 """
104
105 def __init__(self, num_classes, backbone_indices, backbone_channels,
106 aspp_ratios, aspp_out_channels, align_corners):
107 super().__init__()
108 self.backbone_indices = backbone_indices
109 self.align_corners = align_corners
110 self.aspp = layers.ASPPModule(
111 aspp_ratios=aspp_ratios,
112 in_channels=backbone_channels[backbone_indices[1]],
113 out_channels=aspp_out_channels,
114 align_corners=align_corners,
115 image_pooling=True)
116
117 self.bot_fine = nn.Conv2D(
118 backbone_channels[backbone_indices[0]], 48, 1, bias_attr=False)
119 # decoupled
120 self.squeeze_body_edge = SqueezeBodyEdge(
121 256, align_corners=self.align_corners)
122 self.edge_fusion = nn.Conv2D(256 + 48, 256, 1, bias_attr=False)
123 self.sigmoid_edge = nn.Sigmoid()
124 self.edge_out = nn.Sequential(
125 layers.ConvBNReLU(
126 in_channels=256,
127 out_channels=48,
128 kernel_size=3,
129 bias_attr=False), nn.Conv2D(48, 1, 1, bias_attr=False))
130 self.dsn_seg_body = nn.Sequential(
131 layers.ConvBNReLU(
132 in_channels=256,
133 out_channels=256,
134 kernel_size=3,
135 bias_attr=False), nn.Conv2D(
136 256, num_classes, 1, bias_attr=False))
137
138 self.final_seg = nn.Sequential(
139 layers.ConvBNReLU(
140 in_channels=512,
141 out_channels=256,
142 kernel_size=3,
143 bias_attr=False),
144 layers.ConvBNReLU(
145 in_channels=256,
146 out_channels=256,
147 kernel_size=3,
148 bias_attr=False),
149 nn.Conv2D(256, num_classes, kernel_size=1, bias_attr=False))
150
151 def forward(self, feat_list):
152 fine_fea = feat_list[self.backbone_indices[0]]
153 fine_size = paddle.shape(fine_fea)
154 x = feat_list[self.backbone_indices[1]]
155 aspp = self.aspp(x)
156
157 # decoupled
158 seg_body, seg_edge = self.squeeze_body_edge(aspp)
159 # Edge presevation and edge out
160 fine_fea = self.bot_fine(fine_fea)
161 seg_edge = F.interpolate(
162 seg_edge,
163 fine_size[2:],
164 mode='bilinear',
165 align_corners=self.align_corners)
166 seg_edge = self.edge_fusion(paddle.concat([seg_edge, fine_fea], axis=1))
167 seg_edge_out = self.edge_out(seg_edge)
168 seg_edge_out = self.sigmoid_edge(seg_edge_out) # seg_edge output
169 seg_body_out = self.dsn_seg_body(seg_body) # body out
170
171 # seg_final out
172 seg_out = seg_edge + F.interpolate(
173 seg_body,
174 fine_size[2:],
175 mode='bilinear',
176 align_corners=self.align_corners)
177 aspp = F.interpolate(
178 aspp,
179 fine_size[2:],
180 mode='bilinear',
181 align_corners=self.align_corners)
182 seg_out = paddle.concat([aspp, seg_out], axis=1)
183 seg_final_out = self.final_seg(seg_out)
184
185 return [seg_final_out, seg_body_out, seg_edge_out]
186
187
188 class SqueezeBodyEdge(nn.Layer):
189 def __init__(self, inplane, align_corners=False):
190 super().__init__()
191 self.align_corners = align_corners
192 self.down = nn.Sequential(
193 layers.ConvBNReLU(
194 inplane, inplane, kernel_size=3, groups=inplane, stride=2),
195 layers.ConvBNReLU(
196 inplane, inplane, kernel_size=3, groups=inplane, stride=2))
197 self.flow_make = nn.Conv2D(
198 inplane * 2, 2, kernel_size=3, padding='same', bias_attr=False)
199
200 def forward(self, x):
201 size = paddle.shape(x)[2:]
202 seg_down = self.down(x)
203 seg_down = F.interpolate(
204 seg_down,
205 size=size,
206 mode='bilinear',
207 align_corners=self.align_corners)
208 flow = self.flow_make(paddle.concat([x, seg_down], axis=1))
209 seg_flow_warp = self.flow_warp(x, flow, size)
210 seg_edge = x - seg_flow_warp
211 return seg_flow_warp, seg_edge
212
213 def flow_warp(self, input, flow, size):
214 input_shape = paddle.shape(input)
215 norm = size[::-1].reshape([1, 1, 1, -1])
216 h_grid = paddle.linspace(-1.0, 1.0, size[0]).reshape([-1, 1])
217 h_grid = h_grid.tile([size[1]])
218 w_grid = paddle.linspace(-1.0, 1.0, size[1]).reshape([-1, 1])
219 w_grid = w_grid.tile([size[0]]).transpose([1, 0])
220 grid = paddle.concat([w_grid.unsqueeze(2), h_grid.unsqueeze(2)], axis=2)
221 grid.unsqueeze(0).tile([input_shape[0], 1, 1, 1])
222 grid = grid + paddle.transpose(flow, (0, 2, 3, 1)) / norm
223
224 output = F.grid_sample(input, grid)
225 return output
226
[end of paddleseg/models/decoupled_segnet.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/paddleseg/models/decoupled_segnet.py b/paddleseg/models/decoupled_segnet.py
--- a/paddleseg/models/decoupled_segnet.py
+++ b/paddleseg/models/decoupled_segnet.py
@@ -213,6 +213,7 @@
def flow_warp(self, input, flow, size):
input_shape = paddle.shape(input)
norm = size[::-1].reshape([1, 1, 1, -1])
+ norm.stop_gradient = True
h_grid = paddle.linspace(-1.0, 1.0, size[0]).reshape([-1, 1])
h_grid = h_grid.tile([size[1]])
w_grid = paddle.linspace(-1.0, 1.0, size[1]).reshape([-1, 1])
|
{"golden_diff": "diff --git a/paddleseg/models/decoupled_segnet.py b/paddleseg/models/decoupled_segnet.py\n--- a/paddleseg/models/decoupled_segnet.py\n+++ b/paddleseg/models/decoupled_segnet.py\n@@ -213,6 +213,7 @@\n def flow_warp(self, input, flow, size):\n input_shape = paddle.shape(input)\n norm = size[::-1].reshape([1, 1, 1, -1])\n+ norm.stop_gradient = True\n h_grid = paddle.linspace(-1.0, 1.0, size[0]).reshape([-1, 1])\n h_grid = h_grid.tile([size[1]])\n w_grid = paddle.linspace(-1.0, 1.0, size[1]).reshape([-1, 1])\n", "issue": "DecoupledSegNet\u9519\u8bef\n1.\u73af\u5883\uff1a\uff08\u9996\u5148\u5b89\u88c5cpu\u7248\u672c\uff0c\u6d4b\u8bd5\u6210\u529f\uff09\r\n \r\n\r\n\r\n\r\n2.\u5b89\u88c5GPU\u7248\u672c\r\n\u547d\u4ee4\uff1apython -m pip install paddlepaddle-gpu==2.0.0 -i https://mirror.baidu.com/pypi/simple\r\n\r\n3.\u51fa\u73b0\u95ee\u9898\r\n \r\n\r\n\r\n4.\u89e3\u51b3\u65b9\u6848\r\n\uff081\uff09env | grep cuda\r\n\uff082\uff09port LD_LIBRARY_PATH=$LD_LIBRARY_PATH:/usr/local/cuda/lib64\r\n\uff083\uff09 /usr/local/cuda/lib64\r\n\uff084\uff09ln -s libcusolver.so.10 libcusolver.so\r\n\r\n5. \u8fd0\u884c \r\nPython./train.py -config ./configs/decoupled_segnet/decoupledsegnet_resnet50_os8_cityscapes_832x832_80k.yml\r\n\r\n6.\u65b0\u95ee\u9898\r\n\r\n\r\n\r\n\n", "before_files": [{"content": "# Copyright (c) 2020 PaddlePaddle Authors. All Rights Reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nimport cv2\nimport numpy as np\nimport paddle\nimport paddle.nn as nn\nimport paddle.nn.functional as F\n\nfrom paddleseg.cvlibs import manager\nfrom paddleseg.models import layers\nfrom paddleseg.models.backbones import resnet_vd\nfrom paddleseg.models import deeplab\nfrom paddleseg.utils import utils\n\n\[email protected]_component\nclass DecoupledSegNet(nn.Layer):\n \"\"\"\n The DecoupledSegNet implementation based on PaddlePaddle.\n\n The original article refers to\n Xiangtai Li, et, al. \"Improving Semantic Segmentation via Decoupled Body and Edge Supervision\"\n (https://arxiv.org/pdf/2007.10035.pdf)\n\n Args:\n num_classes (int): The unique number of target classes.\n backbone (paddle.nn.Layer): Backbone network, currently support Resnet50_vd/Resnet101_vd.\n backbone_indices (tuple, optional): Two values in the tuple indicate the indices of output of backbone.\n Default: (0, 3).\n aspp_ratios (tuple, optional): The dilation rate using in ASSP module.\n If output_stride=16, aspp_ratios should be set as (1, 6, 12, 18).\n If output_stride=8, aspp_ratios is (1, 12, 24, 36).\n Default: (1, 6, 12, 18).\n aspp_out_channels (int, optional): The output channels of ASPP module. Default: 256.\n align_corners (bool, optional): An argument of F.interpolate. It should be set to False when the feature size is even,\n e.g. 1024x512, otherwise it is True, e.g. 769x769. Default: False.\n pretrained (str, optional): The path or url of pretrained model. Default: None.\n \"\"\"\n\n def __init__(self,\n num_classes,\n backbone,\n backbone_indices=(0, 3),\n aspp_ratios=(1, 6, 12, 18),\n aspp_out_channels=256,\n align_corners=False,\n pretrained=None):\n super().__init__()\n self.backbone = backbone\n backbone_channels = self.backbone.feat_channels\n self.head = DecoupledSegNetHead(num_classes, backbone_indices,\n backbone_channels, aspp_ratios,\n aspp_out_channels, align_corners)\n self.align_corners = align_corners\n self.pretrained = pretrained\n self.init_weight()\n\n def forward(self, x):\n feat_list = self.backbone(x)\n logit_list = self.head(feat_list)\n\n seg_logit, body_logit, edge_logit = [\n F.interpolate(\n logit,\n paddle.shape(x)[2:],\n mode='bilinear',\n align_corners=self.align_corners) for logit in logit_list\n ]\n\n return [seg_logit, body_logit, edge_logit, (seg_logit, edge_logit)]\n\n def init_weight(self):\n if self.pretrained is not None:\n utils.load_entire_model(self, self.pretrained)\n\n\nclass DecoupledSegNetHead(nn.Layer):\n \"\"\"\n The DecoupledSegNetHead implementation based on PaddlePaddle.\n\n Args:\n num_classes (int): The unique number of target classes.\n backbone_indices (tuple): Two values in the tuple indicate the indices of output of backbone.\n the first index will be taken as a low-level feature in Edge presevation component;\n the second one will be taken as input of ASPP component.\n backbone_channels (tuple): The channels of output of backbone.\n aspp_ratios (tuple): The dilation rates using in ASSP module.\n aspp_out_channels (int): The output channels of ASPP module.\n align_corners (bool): An argument of F.interpolate. It should be set to False when the output size of feature\n is even, e.g. 1024x512, otherwise it is True, e.g. 769x769.\n \"\"\"\n\n def __init__(self, num_classes, backbone_indices, backbone_channels,\n aspp_ratios, aspp_out_channels, align_corners):\n super().__init__()\n self.backbone_indices = backbone_indices\n self.align_corners = align_corners\n self.aspp = layers.ASPPModule(\n aspp_ratios=aspp_ratios,\n in_channels=backbone_channels[backbone_indices[1]],\n out_channels=aspp_out_channels,\n align_corners=align_corners,\n image_pooling=True)\n\n self.bot_fine = nn.Conv2D(\n backbone_channels[backbone_indices[0]], 48, 1, bias_attr=False)\n # decoupled\n self.squeeze_body_edge = SqueezeBodyEdge(\n 256, align_corners=self.align_corners)\n self.edge_fusion = nn.Conv2D(256 + 48, 256, 1, bias_attr=False)\n self.sigmoid_edge = nn.Sigmoid()\n self.edge_out = nn.Sequential(\n layers.ConvBNReLU(\n in_channels=256,\n out_channels=48,\n kernel_size=3,\n bias_attr=False), nn.Conv2D(48, 1, 1, bias_attr=False))\n self.dsn_seg_body = nn.Sequential(\n layers.ConvBNReLU(\n in_channels=256,\n out_channels=256,\n kernel_size=3,\n bias_attr=False), nn.Conv2D(\n 256, num_classes, 1, bias_attr=False))\n\n self.final_seg = nn.Sequential(\n layers.ConvBNReLU(\n in_channels=512,\n out_channels=256,\n kernel_size=3,\n bias_attr=False),\n layers.ConvBNReLU(\n in_channels=256,\n out_channels=256,\n kernel_size=3,\n bias_attr=False),\n nn.Conv2D(256, num_classes, kernel_size=1, bias_attr=False))\n\n def forward(self, feat_list):\n fine_fea = feat_list[self.backbone_indices[0]]\n fine_size = paddle.shape(fine_fea)\n x = feat_list[self.backbone_indices[1]]\n aspp = self.aspp(x)\n\n # decoupled\n seg_body, seg_edge = self.squeeze_body_edge(aspp)\n # Edge presevation and edge out\n fine_fea = self.bot_fine(fine_fea)\n seg_edge = F.interpolate(\n seg_edge,\n fine_size[2:],\n mode='bilinear',\n align_corners=self.align_corners)\n seg_edge = self.edge_fusion(paddle.concat([seg_edge, fine_fea], axis=1))\n seg_edge_out = self.edge_out(seg_edge)\n seg_edge_out = self.sigmoid_edge(seg_edge_out) # seg_edge output\n seg_body_out = self.dsn_seg_body(seg_body) # body out\n\n # seg_final out\n seg_out = seg_edge + F.interpolate(\n seg_body,\n fine_size[2:],\n mode='bilinear',\n align_corners=self.align_corners)\n aspp = F.interpolate(\n aspp,\n fine_size[2:],\n mode='bilinear',\n align_corners=self.align_corners)\n seg_out = paddle.concat([aspp, seg_out], axis=1)\n seg_final_out = self.final_seg(seg_out)\n\n return [seg_final_out, seg_body_out, seg_edge_out]\n\n\nclass SqueezeBodyEdge(nn.Layer):\n def __init__(self, inplane, align_corners=False):\n super().__init__()\n self.align_corners = align_corners\n self.down = nn.Sequential(\n layers.ConvBNReLU(\n inplane, inplane, kernel_size=3, groups=inplane, stride=2),\n layers.ConvBNReLU(\n inplane, inplane, kernel_size=3, groups=inplane, stride=2))\n self.flow_make = nn.Conv2D(\n inplane * 2, 2, kernel_size=3, padding='same', bias_attr=False)\n\n def forward(self, x):\n size = paddle.shape(x)[2:]\n seg_down = self.down(x)\n seg_down = F.interpolate(\n seg_down,\n size=size,\n mode='bilinear',\n align_corners=self.align_corners)\n flow = self.flow_make(paddle.concat([x, seg_down], axis=1))\n seg_flow_warp = self.flow_warp(x, flow, size)\n seg_edge = x - seg_flow_warp\n return seg_flow_warp, seg_edge\n\n def flow_warp(self, input, flow, size):\n input_shape = paddle.shape(input)\n norm = size[::-1].reshape([1, 1, 1, -1])\n h_grid = paddle.linspace(-1.0, 1.0, size[0]).reshape([-1, 1])\n h_grid = h_grid.tile([size[1]])\n w_grid = paddle.linspace(-1.0, 1.0, size[1]).reshape([-1, 1])\n w_grid = w_grid.tile([size[0]]).transpose([1, 0])\n grid = paddle.concat([w_grid.unsqueeze(2), h_grid.unsqueeze(2)], axis=2)\n grid.unsqueeze(0).tile([input_shape[0], 1, 1, 1])\n grid = grid + paddle.transpose(flow, (0, 2, 3, 1)) / norm\n\n output = F.grid_sample(input, grid)\n return output\n", "path": "paddleseg/models/decoupled_segnet.py"}]}
| 3,734 | 185 |
gh_patches_debug_10731
|
rasdani/github-patches
|
git_diff
|
litestar-org__litestar-2982
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Bug: openapi schema generation fails for Union of/in msgspec.Struct models
### Description
Hello!
In the latest versions(s) (I think this originates from the changes regarding nested models in openapi generation) we cannot use Unions of `msgspec.Struct`s anymore. Neither as direct return types for routes nor nested within return types.
The result is a 500 Error. The MCVE below raises `'types.UnionType' object has no attribute '__qualname__'` internally. In our production app I get `typing.Union is not a module, class, method, or function.` instead.
Cheers
### URL to code causing the issue
_No response_
### MCVE
```python
import msgspec
import uvicorn
from litestar import Litestar, get
class SubStructA(msgspec.Struct):
a: int
class SubStructB(msgspec.Struct):
a: int
class StructyStruct(msgspec.Struct):
sub: SubStructA | SubStructB
@get("/subunion")
async def testSubUnion() -> StructyStruct:
return StructyStruct(SubStructA(0))
@get("/union")
async def testUnion() -> SubStructA | SubStructB:
return SubStructA(0)
app = Litestar(route_handlers=[test2]) # or test
uvicorn.run(app)
```
### Steps to reproduce
```bash
Run the example and browse to `localhost:8000/schema`
```
### Screenshots
_No response_
### Logs
_No response_
### Litestar Version
2.5.0
### Platform
- [X] Linux
- [ ] Mac
- [ ] Windows
- [ ] Other (Please specify in the description above)
<!-- POLAR PLEDGE BADGE START -->
---
> [!NOTE]
> While we are open for sponsoring on [GitHub Sponsors](https://github.com/sponsors/litestar-org/) and
> [OpenCollective](https://opencollective.com/litestar), we also utilize [Polar.sh](https://polar.sh/) to engage in pledge-based sponsorship.
>
> Check out all issues funded or available for funding [on our Polar.sh dashboard](https://polar.sh/litestar-org)
> * If you would like to see an issue prioritized, make a pledge towards it!
> * We receive the pledge once the issue is completed & verified
> * This, along with engagement in the community, helps us know which features are a priority to our users.
<a href="https://polar.sh/litestar-org/litestar/issues/2971">
<picture>
<source media="(prefers-color-scheme: dark)" srcset="https://polar.sh/api/github/litestar-org/litestar/issues/2971/pledge.svg?darkmode=1">
<img alt="Fund with Polar" src="https://polar.sh/api/github/litestar-org/litestar/issues/2971/pledge.svg">
</picture>
</a>
<!-- POLAR PLEDGE BADGE END -->
</issue>
<code>
[start of litestar/_openapi/schema_generation/plugins/struct.py]
1 from __future__ import annotations
2
3 from typing import TYPE_CHECKING
4
5 from msgspec import Struct
6 from msgspec.structs import fields
7
8 from litestar.plugins import OpenAPISchemaPlugin
9 from litestar.types.empty import Empty
10 from litestar.typing import FieldDefinition
11 from litestar.utils.predicates import is_optional_union
12
13 if TYPE_CHECKING:
14 from msgspec.structs import FieldInfo
15
16 from litestar._openapi.schema_generation import SchemaCreator
17 from litestar.openapi.spec import Schema
18
19
20 class StructSchemaPlugin(OpenAPISchemaPlugin):
21 def is_plugin_supported_field(self, field_definition: FieldDefinition) -> bool:
22 return field_definition.is_subclass_of(Struct)
23
24 def to_openapi_schema(self, field_definition: FieldDefinition, schema_creator: SchemaCreator) -> Schema:
25 def is_field_required(field: FieldInfo) -> bool:
26 return field.required or field.default_factory is Empty
27
28 type_hints = field_definition.get_type_hints(include_extras=True, resolve_generics=True)
29 struct_fields = fields(field_definition.type_)
30
31 return schema_creator.create_component_schema(
32 field_definition,
33 required=sorted(
34 [
35 field.encode_name
36 for field in struct_fields
37 if is_field_required(field=field) and not is_optional_union(type_hints[field.name])
38 ]
39 ),
40 property_fields={
41 field.encode_name: FieldDefinition.from_kwarg(type_hints[field.name], field.encode_name)
42 for field in struct_fields
43 },
44 )
45
[end of litestar/_openapi/schema_generation/plugins/struct.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/litestar/_openapi/schema_generation/plugins/struct.py b/litestar/_openapi/schema_generation/plugins/struct.py
--- a/litestar/_openapi/schema_generation/plugins/struct.py
+++ b/litestar/_openapi/schema_generation/plugins/struct.py
@@ -19,7 +19,7 @@
class StructSchemaPlugin(OpenAPISchemaPlugin):
def is_plugin_supported_field(self, field_definition: FieldDefinition) -> bool:
- return field_definition.is_subclass_of(Struct)
+ return not field_definition.is_union and field_definition.is_subclass_of(Struct)
def to_openapi_schema(self, field_definition: FieldDefinition, schema_creator: SchemaCreator) -> Schema:
def is_field_required(field: FieldInfo) -> bool:
|
{"golden_diff": "diff --git a/litestar/_openapi/schema_generation/plugins/struct.py b/litestar/_openapi/schema_generation/plugins/struct.py\n--- a/litestar/_openapi/schema_generation/plugins/struct.py\n+++ b/litestar/_openapi/schema_generation/plugins/struct.py\n@@ -19,7 +19,7 @@\n \n class StructSchemaPlugin(OpenAPISchemaPlugin):\n def is_plugin_supported_field(self, field_definition: FieldDefinition) -> bool:\n- return field_definition.is_subclass_of(Struct)\n+ return not field_definition.is_union and field_definition.is_subclass_of(Struct)\n \n def to_openapi_schema(self, field_definition: FieldDefinition, schema_creator: SchemaCreator) -> Schema:\n def is_field_required(field: FieldInfo) -> bool:\n", "issue": "Bug: openapi schema generation fails for Union of/in msgspec.Struct models\n### Description\r\n\r\nHello!\r\n\r\nIn the latest versions(s) (I think this originates from the changes regarding nested models in openapi generation) we cannot use Unions of `msgspec.Struct`s anymore. Neither as direct return types for routes nor nested within return types. \r\n\r\nThe result is a 500 Error. The MCVE below raises `'types.UnionType' object has no attribute '__qualname__'` internally. In our production app I get `typing.Union is not a module, class, method, or function.` instead.\r\n\r\nCheers\r\n\r\n### URL to code causing the issue\r\n\r\n_No response_\r\n\r\n### MCVE\r\n\r\n```python\r\nimport msgspec\r\nimport uvicorn\r\nfrom litestar import Litestar, get\r\n\r\n\r\nclass SubStructA(msgspec.Struct):\r\n a: int\r\n\r\n\r\nclass SubStructB(msgspec.Struct):\r\n a: int\r\n\r\n\r\nclass StructyStruct(msgspec.Struct):\r\n sub: SubStructA | SubStructB\r\n\r\n\r\n@get(\"/subunion\")\r\nasync def testSubUnion() -> StructyStruct:\r\n return StructyStruct(SubStructA(0))\r\n\r\n\r\n@get(\"/union\")\r\nasync def testUnion() -> SubStructA | SubStructB:\r\n return SubStructA(0)\r\n\r\n\r\napp = Litestar(route_handlers=[test2]) # or test\r\nuvicorn.run(app)\r\n```\r\n\r\n\r\n### Steps to reproduce\r\n\r\n```bash\r\nRun the example and browse to `localhost:8000/schema`\r\n```\r\n\r\n\r\n### Screenshots\r\n\r\n_No response_\r\n\r\n### Logs\r\n\r\n_No response_\r\n\r\n### Litestar Version\r\n\r\n2.5.0\r\n\r\n### Platform\r\n\r\n- [X] Linux\r\n- [ ] Mac\r\n- [ ] Windows\r\n- [ ] Other (Please specify in the description above)\r\n\r\n<!-- POLAR PLEDGE BADGE START -->\r\n---\r\n> [!NOTE] \r\n> While we are open for sponsoring on [GitHub Sponsors](https://github.com/sponsors/litestar-org/) and \r\n> [OpenCollective](https://opencollective.com/litestar), we also utilize [Polar.sh](https://polar.sh/) to engage in pledge-based sponsorship.\r\n>\r\n> Check out all issues funded or available for funding [on our Polar.sh dashboard](https://polar.sh/litestar-org)\r\n> * If you would like to see an issue prioritized, make a pledge towards it!\r\n> * We receive the pledge once the issue is completed & verified\r\n> * This, along with engagement in the community, helps us know which features are a priority to our users.\r\n\r\n<a href=\"https://polar.sh/litestar-org/litestar/issues/2971\">\r\n<picture>\r\n <source media=\"(prefers-color-scheme: dark)\" srcset=\"https://polar.sh/api/github/litestar-org/litestar/issues/2971/pledge.svg?darkmode=1\">\r\n <img alt=\"Fund with Polar\" src=\"https://polar.sh/api/github/litestar-org/litestar/issues/2971/pledge.svg\">\r\n</picture>\r\n</a>\r\n<!-- POLAR PLEDGE BADGE END -->\r\n\n", "before_files": [{"content": "from __future__ import annotations\n\nfrom typing import TYPE_CHECKING\n\nfrom msgspec import Struct\nfrom msgspec.structs import fields\n\nfrom litestar.plugins import OpenAPISchemaPlugin\nfrom litestar.types.empty import Empty\nfrom litestar.typing import FieldDefinition\nfrom litestar.utils.predicates import is_optional_union\n\nif TYPE_CHECKING:\n from msgspec.structs import FieldInfo\n\n from litestar._openapi.schema_generation import SchemaCreator\n from litestar.openapi.spec import Schema\n\n\nclass StructSchemaPlugin(OpenAPISchemaPlugin):\n def is_plugin_supported_field(self, field_definition: FieldDefinition) -> bool:\n return field_definition.is_subclass_of(Struct)\n\n def to_openapi_schema(self, field_definition: FieldDefinition, schema_creator: SchemaCreator) -> Schema:\n def is_field_required(field: FieldInfo) -> bool:\n return field.required or field.default_factory is Empty\n\n type_hints = field_definition.get_type_hints(include_extras=True, resolve_generics=True)\n struct_fields = fields(field_definition.type_)\n\n return schema_creator.create_component_schema(\n field_definition,\n required=sorted(\n [\n field.encode_name\n for field in struct_fields\n if is_field_required(field=field) and not is_optional_union(type_hints[field.name])\n ]\n ),\n property_fields={\n field.encode_name: FieldDefinition.from_kwarg(type_hints[field.name], field.encode_name)\n for field in struct_fields\n },\n )\n", "path": "litestar/_openapi/schema_generation/plugins/struct.py"}]}
| 1,605 | 168 |
gh_patches_debug_6714
|
rasdani/github-patches
|
git_diff
|
open-mmlab__mmocr-570
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Write image name to pickle file
Hi MMOCR team,
Thank you for this awesome framework. I have a task to get coordinate of bounding box from Textsnake model, so I use --out argument in test.py to export to a pickle file. But when I load this pickle, I just got ‘boundary_result’ and don't know this ‘boundary_result’ belongs to which image. How can I get the image to write to the pickle file? Thank you.
</issue>
<code>
[start of mmocr/models/textdet/dense_heads/head_mixin.py]
1 # Copyright (c) OpenMMLab. All rights reserved.
2 import numpy as np
3
4 from mmocr.models.builder import HEADS
5 from mmocr.models.textdet.postprocess import decode
6 from mmocr.utils import check_argument
7
8
9 @HEADS.register_module()
10 class HeadMixin:
11 """The head minxin for dbnet and pannet heads."""
12
13 def resize_boundary(self, boundaries, scale_factor):
14 """Rescale boundaries via scale_factor.
15
16 Args:
17 boundaries (list[list[float]]): The boundary list. Each boundary
18 with size 2k+1 with k>=4.
19 scale_factor(ndarray): The scale factor of size (4,).
20
21 Returns:
22 boundaries (list[list[float]]): The scaled boundaries.
23 """
24 assert check_argument.is_2dlist(boundaries)
25 assert isinstance(scale_factor, np.ndarray)
26 assert scale_factor.shape[0] == 4
27
28 for b in boundaries:
29 sz = len(b)
30 check_argument.valid_boundary(b, True)
31 b[:sz -
32 1] = (np.array(b[:sz - 1]) *
33 (np.tile(scale_factor[:2], int(
34 (sz - 1) / 2)).reshape(1, sz - 1))).flatten().tolist()
35 return boundaries
36
37 def get_boundary(self, score_maps, img_metas, rescale):
38 """Compute text boundaries via post processing.
39
40 Args:
41 score_maps (Tensor): The text score map.
42 img_metas (dict): The image meta info.
43 rescale (bool): Rescale boundaries to the original image resolution
44 if true, and keep the score_maps resolution if false.
45
46 Returns:
47 results (dict): The result dict.
48 """
49
50 assert check_argument.is_type_list(img_metas, dict)
51 assert isinstance(rescale, bool)
52
53 score_maps = score_maps.squeeze()
54 boundaries = decode(
55 decoding_type=self.decoding_type,
56 preds=score_maps,
57 text_repr_type=self.text_repr_type)
58 if rescale:
59 boundaries = self.resize_boundary(
60 boundaries,
61 1.0 / self.downsample_ratio / img_metas[0]['scale_factor'])
62 results = dict(boundary_result=boundaries)
63 return results
64
65 def loss(self, pred_maps, **kwargs):
66 """Compute the loss for text detection.
67
68 Args:
69 pred_maps (tensor): The input score maps of NxCxHxW.
70
71 Returns:
72 losses (dict): The dict for losses.
73 """
74 losses = self.loss_module(pred_maps, self.downsample_ratio, **kwargs)
75 return losses
76
[end of mmocr/models/textdet/dense_heads/head_mixin.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/mmocr/models/textdet/dense_heads/head_mixin.py b/mmocr/models/textdet/dense_heads/head_mixin.py
--- a/mmocr/models/textdet/dense_heads/head_mixin.py
+++ b/mmocr/models/textdet/dense_heads/head_mixin.py
@@ -59,7 +59,9 @@
boundaries = self.resize_boundary(
boundaries,
1.0 / self.downsample_ratio / img_metas[0]['scale_factor'])
- results = dict(boundary_result=boundaries)
+ results = dict(
+ boundary_result=boundaries, filename=img_metas[0]['filename'])
+
return results
def loss(self, pred_maps, **kwargs):
|
{"golden_diff": "diff --git a/mmocr/models/textdet/dense_heads/head_mixin.py b/mmocr/models/textdet/dense_heads/head_mixin.py\n--- a/mmocr/models/textdet/dense_heads/head_mixin.py\n+++ b/mmocr/models/textdet/dense_heads/head_mixin.py\n@@ -59,7 +59,9 @@\n boundaries = self.resize_boundary(\n boundaries,\n 1.0 / self.downsample_ratio / img_metas[0]['scale_factor'])\n- results = dict(boundary_result=boundaries)\n+ results = dict(\n+ boundary_result=boundaries, filename=img_metas[0]['filename'])\n+\n return results\n \n def loss(self, pred_maps, **kwargs):\n", "issue": "Write image name to pickle file\nHi MMOCR team,\nThank you for this awesome framework. I have a task to get coordinate of bounding box from Textsnake model, so I use --out argument in test.py to export to a pickle file. But when I load this pickle, I just got \u2018boundary_result\u2019 and don't know this \u2018boundary_result\u2019 belongs to which image. How can I get the image to write to the pickle file? Thank you.\n", "before_files": [{"content": "# Copyright (c) OpenMMLab. All rights reserved.\nimport numpy as np\n\nfrom mmocr.models.builder import HEADS\nfrom mmocr.models.textdet.postprocess import decode\nfrom mmocr.utils import check_argument\n\n\[email protected]_module()\nclass HeadMixin:\n \"\"\"The head minxin for dbnet and pannet heads.\"\"\"\n\n def resize_boundary(self, boundaries, scale_factor):\n \"\"\"Rescale boundaries via scale_factor.\n\n Args:\n boundaries (list[list[float]]): The boundary list. Each boundary\n with size 2k+1 with k>=4.\n scale_factor(ndarray): The scale factor of size (4,).\n\n Returns:\n boundaries (list[list[float]]): The scaled boundaries.\n \"\"\"\n assert check_argument.is_2dlist(boundaries)\n assert isinstance(scale_factor, np.ndarray)\n assert scale_factor.shape[0] == 4\n\n for b in boundaries:\n sz = len(b)\n check_argument.valid_boundary(b, True)\n b[:sz -\n 1] = (np.array(b[:sz - 1]) *\n (np.tile(scale_factor[:2], int(\n (sz - 1) / 2)).reshape(1, sz - 1))).flatten().tolist()\n return boundaries\n\n def get_boundary(self, score_maps, img_metas, rescale):\n \"\"\"Compute text boundaries via post processing.\n\n Args:\n score_maps (Tensor): The text score map.\n img_metas (dict): The image meta info.\n rescale (bool): Rescale boundaries to the original image resolution\n if true, and keep the score_maps resolution if false.\n\n Returns:\n results (dict): The result dict.\n \"\"\"\n\n assert check_argument.is_type_list(img_metas, dict)\n assert isinstance(rescale, bool)\n\n score_maps = score_maps.squeeze()\n boundaries = decode(\n decoding_type=self.decoding_type,\n preds=score_maps,\n text_repr_type=self.text_repr_type)\n if rescale:\n boundaries = self.resize_boundary(\n boundaries,\n 1.0 / self.downsample_ratio / img_metas[0]['scale_factor'])\n results = dict(boundary_result=boundaries)\n return results\n\n def loss(self, pred_maps, **kwargs):\n \"\"\"Compute the loss for text detection.\n\n Args:\n pred_maps (tensor): The input score maps of NxCxHxW.\n\n Returns:\n losses (dict): The dict for losses.\n \"\"\"\n losses = self.loss_module(pred_maps, self.downsample_ratio, **kwargs)\n return losses\n", "path": "mmocr/models/textdet/dense_heads/head_mixin.py"}]}
| 1,348 | 153 |
gh_patches_debug_12224
|
rasdani/github-patches
|
git_diff
|
Zeroto521__my-data-toolkit-413
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
DEV: rename master to main
https://github.com/github/renaming
</issue>
<code>
[start of doc/source/conf.py]
1 # Configuration file for the Sphinx documentation builder.
2 #
3 # This file only contains a selection of the most common options.
4 # For a full list see the documentation:
5 # https://www.sphinx-doc.org/en/master/usage/configuration.html
6
7
8 # -- Project information -----------------------------------------------------
9
10 # The version info for the project you're documenting, acts as replacement for
11 # |version| and |release|, also used in various other places throughout the
12 # built documents.
13 from __future__ import annotations
14
15 import inspect
16 import os
17 import sys
18 from datetime import datetime
19
20 import dtoolkit
21
22 version = release = dtoolkit.__version__
23 version = version.replace(".dev0", "")
24 version = version.replace(".post0", "")
25
26
27 project = "DToolKit"
28 author = "Zero <@Zeroto521>"
29 copyright = f"2021-{datetime.now().year}, {author}" # pylint: disable=redefined-builtin
30 github_url = "https://github.com/Zeroto521/my-data-toolkit"
31
32
33 # -- General configuration ---------------------------------------------------
34
35 # Add any Sphinx extension module names here, as strings. They can be
36 # extensions coming with Sphinx (named 'sphinx.ext.*') or your custom
37 # ones.
38 extensions = [
39 "myst_parser",
40 "numpydoc",
41 "sphinx.ext.autodoc",
42 "sphinx.ext.autosummary",
43 "sphinx.ext.autosectionlabel",
44 "sphinx.ext.extlinks",
45 "sphinx.ext.intersphinx",
46 "sphinx.ext.linkcode",
47 "sphinx_toggleprompt",
48 "IPython.sphinxext.ipython_console_highlighting",
49 "IPython.sphinxext.ipython_directive",
50 "nbsphinx",
51 ]
52
53 # The suffix of source filenames.
54 source_suffix = [".rst", ".md"]
55
56 # Add any paths that contain templates here, relative to this directory.
57 templates_path = ["_templates"]
58
59 # List of patterns, relative to source directory, that match files and
60 # directories to ignore when looking for source files.
61 # This pattern also affects html_static_path and html_extra_path.
62 exclude_patterns = []
63
64
65 # -- Options for HTML output -------------------------------------------------
66
67 # The theme to use for HTML and HTML Help pages. See the documentation for
68 # a list of builtin themes.
69 html_theme = "pydata_sphinx_theme"
70
71 # Theme options are theme-specific and customize the look and feel of a theme
72 # further. For a list of options available for each theme, see the
73 # documentation.
74 html_theme_options = {
75 "search_bar_position": "sidebar",
76 "github_url": github_url,
77 "icon_links": [
78 {
79 "name": "PyPI",
80 "url": "https://pypi.org/project/my-data-toolkit",
81 "icon": "fas fa-box",
82 },
83 ],
84 }
85
86 # Add any paths that contain custom static files (such as style sheets) here,
87 # relative to this directory. They are copied after the builtin static files,
88 # so a file named "default.css" will overwrite the builtin "default.css".
89 html_static_path = ["_static"]
90
91
92 # Add redirect for previously existing pages, each item is like `(from_old, to_new)`
93
94 moved_pages = [
95 ("py-modindex", "reference"),
96 ]
97
98 html_additional_pages = {page[0]: "redirect.html" for page in moved_pages}
99
100 html_context = {"redirects": dict(moved_pages)}
101
102
103 nbsphinx_prolog = r"""
104 {% set docname = env.doc2path(env.docname, base=None) %}
105
106 .. tip::
107
108 This page was generated from `{{ docname }}`__.
109
110 __ https://github.com/zeroto521/my-data-toolkit/blob/master/doc/source/{{ docname }}
111 """
112
113
114 # --Options for sphinx extensions -----------------------------------------------
115
116 # connect docs in other projects
117 intersphinx_mapping = {
118 "python": ("http://docs.python.org/", None),
119 "pandas": ("https://pandas.pydata.org/pandas-docs/stable/", None),
120 "numpy": ("https://numpy.org/doc/stable/", None),
121 "sklearn": ("https://scikit-learn.org/stable/", None),
122 "geopandas": ("https://geopandas.readthedocs.io/en/stable/", None),
123 "shapely": ("https://shapely.readthedocs.io/en/stable/", None),
124 "pyproj": ("https://pyproj4.github.io/pyproj/stable/", None),
125 "pygeos": ("https://pygeos.readthedocs.io/en/stable/", None),
126 }
127
128 # extlinks alias
129 extlinks = {
130 "issue": (f"{github_url}/issues/%s", "issue#"),
131 "pr": (f"{github_url}/issues/%s", "pr#"),
132 "user": ("https://github.com/%s", "@"),
133 }
134
135 myst_enable_extensions = [
136 "colon_fence",
137 ]
138
139 autosummary_generate = True
140
141 nbsphinx_execute = "always"
142 nbsphinx_allow_errors = True
143
144
145 # based on pandas doc/source/conf.py
146 def linkcode_resolve(domain: str, info: dict[str, str]) -> str | None:
147 """
148 Determine the URL corresponding to Python object
149 """
150
151 if domain != "py":
152 return None
153
154 modname = info["module"]
155 fullname = info["fullname"]
156
157 submod = sys.modules.get(modname)
158 if submod is None:
159 return None
160
161 obj = submod
162 for part in fullname.split("."):
163 try:
164 obj = getattr(obj, part)
165 except AttributeError:
166 return None
167
168 try:
169 fn = inspect.getsourcefile(inspect.unwrap(obj))
170 except TypeError:
171 fn = None
172
173 if not fn:
174 return None
175
176 # to fix these doc doesn't exist in dtoolkit
177 if project.lower() not in fn:
178 return None
179
180 try:
181 source, lineno = inspect.getsourcelines(obj)
182 except OSError:
183 lineno = None
184
185 linespec = f"#L{lineno}-L{lineno + len(source) - 1}" if lineno else ""
186 fn = os.path.relpath(fn, start=os.path.dirname(dtoolkit.__file__))
187
188 base_link = f"{github_url}/blob/" + "{branch}" + f"/dtoolkit/{fn}{linespec}"
189 if "post" in version:
190 return base_link.format(branch="master")
191
192 return base_link.format(branch=f"v{version}")
193
[end of doc/source/conf.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/doc/source/conf.py b/doc/source/conf.py
--- a/doc/source/conf.py
+++ b/doc/source/conf.py
@@ -107,7 +107,7 @@
This page was generated from `{{ docname }}`__.
- __ https://github.com/zeroto521/my-data-toolkit/blob/master/doc/source/{{ docname }}
+ __ https://github.com/zeroto521/my-data-toolkit/blob/main/doc/source/{{ docname }}
"""
@@ -187,6 +187,6 @@
base_link = f"{github_url}/blob/" + "{branch}" + f"/dtoolkit/{fn}{linespec}"
if "post" in version:
- return base_link.format(branch="master")
+ return base_link.format(branch="main")
return base_link.format(branch=f"v{version}")
|
{"golden_diff": "diff --git a/doc/source/conf.py b/doc/source/conf.py\n--- a/doc/source/conf.py\n+++ b/doc/source/conf.py\n@@ -107,7 +107,7 @@\n \n This page was generated from `{{ docname }}`__.\n \n- __ https://github.com/zeroto521/my-data-toolkit/blob/master/doc/source/{{ docname }}\n+ __ https://github.com/zeroto521/my-data-toolkit/blob/main/doc/source/{{ docname }}\n \"\"\"\n \n \n@@ -187,6 +187,6 @@\n \n base_link = f\"{github_url}/blob/\" + \"{branch}\" + f\"/dtoolkit/{fn}{linespec}\"\n if \"post\" in version:\n- return base_link.format(branch=\"master\")\n+ return base_link.format(branch=\"main\")\n \n return base_link.format(branch=f\"v{version}\")\n", "issue": "DEV: rename master to main\nhttps://github.com/github/renaming\n", "before_files": [{"content": "# Configuration file for the Sphinx documentation builder.\n#\n# This file only contains a selection of the most common options.\n# For a full list see the documentation:\n# https://www.sphinx-doc.org/en/master/usage/configuration.html\n\n\n# -- Project information -----------------------------------------------------\n\n# The version info for the project you're documenting, acts as replacement for\n# |version| and |release|, also used in various other places throughout the\n# built documents.\nfrom __future__ import annotations\n\nimport inspect\nimport os\nimport sys\nfrom datetime import datetime\n\nimport dtoolkit\n\nversion = release = dtoolkit.__version__\nversion = version.replace(\".dev0\", \"\")\nversion = version.replace(\".post0\", \"\")\n\n\nproject = \"DToolKit\"\nauthor = \"Zero <@Zeroto521>\"\ncopyright = f\"2021-{datetime.now().year}, {author}\" # pylint: disable=redefined-builtin\ngithub_url = \"https://github.com/Zeroto521/my-data-toolkit\"\n\n\n# -- General configuration ---------------------------------------------------\n\n# Add any Sphinx extension module names here, as strings. They can be\n# extensions coming with Sphinx (named 'sphinx.ext.*') or your custom\n# ones.\nextensions = [\n \"myst_parser\",\n \"numpydoc\",\n \"sphinx.ext.autodoc\",\n \"sphinx.ext.autosummary\",\n \"sphinx.ext.autosectionlabel\",\n \"sphinx.ext.extlinks\",\n \"sphinx.ext.intersphinx\",\n \"sphinx.ext.linkcode\",\n \"sphinx_toggleprompt\",\n \"IPython.sphinxext.ipython_console_highlighting\",\n \"IPython.sphinxext.ipython_directive\",\n \"nbsphinx\",\n]\n\n# The suffix of source filenames.\nsource_suffix = [\".rst\", \".md\"]\n\n# Add any paths that contain templates here, relative to this directory.\ntemplates_path = [\"_templates\"]\n\n# List of patterns, relative to source directory, that match files and\n# directories to ignore when looking for source files.\n# This pattern also affects html_static_path and html_extra_path.\nexclude_patterns = []\n\n\n# -- Options for HTML output -------------------------------------------------\n\n# The theme to use for HTML and HTML Help pages. See the documentation for\n# a list of builtin themes.\nhtml_theme = \"pydata_sphinx_theme\"\n\n# Theme options are theme-specific and customize the look and feel of a theme\n# further. For a list of options available for each theme, see the\n# documentation.\nhtml_theme_options = {\n \"search_bar_position\": \"sidebar\",\n \"github_url\": github_url,\n \"icon_links\": [\n {\n \"name\": \"PyPI\",\n \"url\": \"https://pypi.org/project/my-data-toolkit\",\n \"icon\": \"fas fa-box\",\n },\n ],\n}\n\n# Add any paths that contain custom static files (such as style sheets) here,\n# relative to this directory. They are copied after the builtin static files,\n# so a file named \"default.css\" will overwrite the builtin \"default.css\".\nhtml_static_path = [\"_static\"]\n\n\n# Add redirect for previously existing pages, each item is like `(from_old, to_new)`\n\nmoved_pages = [\n (\"py-modindex\", \"reference\"),\n]\n\nhtml_additional_pages = {page[0]: \"redirect.html\" for page in moved_pages}\n\nhtml_context = {\"redirects\": dict(moved_pages)}\n\n\nnbsphinx_prolog = r\"\"\"\n{% set docname = env.doc2path(env.docname, base=None) %}\n\n.. tip::\n\n This page was generated from `{{ docname }}`__.\n\n __ https://github.com/zeroto521/my-data-toolkit/blob/master/doc/source/{{ docname }}\n\"\"\"\n\n\n# --Options for sphinx extensions -----------------------------------------------\n\n# connect docs in other projects\nintersphinx_mapping = {\n \"python\": (\"http://docs.python.org/\", None),\n \"pandas\": (\"https://pandas.pydata.org/pandas-docs/stable/\", None),\n \"numpy\": (\"https://numpy.org/doc/stable/\", None),\n \"sklearn\": (\"https://scikit-learn.org/stable/\", None),\n \"geopandas\": (\"https://geopandas.readthedocs.io/en/stable/\", None),\n \"shapely\": (\"https://shapely.readthedocs.io/en/stable/\", None),\n \"pyproj\": (\"https://pyproj4.github.io/pyproj/stable/\", None),\n \"pygeos\": (\"https://pygeos.readthedocs.io/en/stable/\", None),\n}\n\n# extlinks alias\nextlinks = {\n \"issue\": (f\"{github_url}/issues/%s\", \"issue#\"),\n \"pr\": (f\"{github_url}/issues/%s\", \"pr#\"),\n \"user\": (\"https://github.com/%s\", \"@\"),\n}\n\nmyst_enable_extensions = [\n \"colon_fence\",\n]\n\nautosummary_generate = True\n\nnbsphinx_execute = \"always\"\nnbsphinx_allow_errors = True\n\n\n# based on pandas doc/source/conf.py\ndef linkcode_resolve(domain: str, info: dict[str, str]) -> str | None:\n \"\"\"\n Determine the URL corresponding to Python object\n \"\"\"\n\n if domain != \"py\":\n return None\n\n modname = info[\"module\"]\n fullname = info[\"fullname\"]\n\n submod = sys.modules.get(modname)\n if submod is None:\n return None\n\n obj = submod\n for part in fullname.split(\".\"):\n try:\n obj = getattr(obj, part)\n except AttributeError:\n return None\n\n try:\n fn = inspect.getsourcefile(inspect.unwrap(obj))\n except TypeError:\n fn = None\n\n if not fn:\n return None\n\n # to fix these doc doesn't exist in dtoolkit\n if project.lower() not in fn:\n return None\n\n try:\n source, lineno = inspect.getsourcelines(obj)\n except OSError:\n lineno = None\n\n linespec = f\"#L{lineno}-L{lineno + len(source) - 1}\" if lineno else \"\"\n fn = os.path.relpath(fn, start=os.path.dirname(dtoolkit.__file__))\n\n base_link = f\"{github_url}/blob/\" + \"{branch}\" + f\"/dtoolkit/{fn}{linespec}\"\n if \"post\" in version:\n return base_link.format(branch=\"master\")\n\n return base_link.format(branch=f\"v{version}\")\n", "path": "doc/source/conf.py"}]}
| 2,424 | 193 |
gh_patches_debug_29188
|
rasdani/github-patches
|
git_diff
|
nilearn__nilearn-2670
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
check_paradigm should check for invalid keys in passed dict
Using the old `nipy` user logic, I passed `amplitude=somethx` instead of `modulation=somethx` in the `make_design_matrix`. I didn't crash but the values where ignored (e.g Error: unknown param, etc.). A default value of 1 was forced...
</issue>
<code>
[start of nilearn/glm/first_level/experimental_paradigm.py]
1 """
2 An experimental protocol is handled as a pandas DataFrame
3 that includes an 'onset' field.
4
5 This yields the onset time of the events in the experimental paradigm.
6 It can also contain:
7
8 * a 'trial_type' field that yields the condition identifier.
9 * a 'duration' field that yields event duration (for so-called block
10 paradigms).
11 * a 'modulation' field that associated a scalar value to each event.
12
13 Author: Bertrand Thirion, 2015
14
15 """
16 import warnings
17
18 import numpy as np
19
20
21 def check_events(events):
22 """Test that the events data describes a valid experimental paradigm
23
24 It is valid if the events data has an 'onset' key.
25
26 Parameters
27 ----------
28 events : pandas DataFrame
29 Events data that describes a functional experimental paradigm.
30
31 Returns
32 -------
33 trial_type : array of shape (n_events,), dtype='s'
34 Per-event experimental conditions identifier.
35 Defaults to np.repeat('dummy', len(onsets)).
36
37 onset : array of shape (n_events,), dtype='f'
38 Per-event onset time (in seconds)
39
40 duration : array of shape (n_events,), dtype='f'
41 Per-event durantion, (in seconds)
42 defaults to zeros(n_events) when no duration is provided
43
44 modulation : array of shape (n_events,), dtype='f'
45 Per-event modulation, (in seconds)
46 defaults to ones(n_events) when no duration is provided.
47
48 """
49 if 'onset' not in events.keys():
50 raise ValueError('The provided events data has no onset column.')
51 if 'duration' not in events.keys():
52 raise ValueError('The provided events data has no duration column.')
53
54 onset = np.array(events['onset'])
55 duration = np.array(events['duration']).astype(np.float)
56 n_events = len(onset)
57 trial_type = np.array(events['trial_type'])
58 modulation = np.ones(n_events)
59 if 'trial_type' not in events.keys():
60 warnings.warn("'trial_type' column not found "
61 "in the given events data.")
62 trial_type = np.repeat('dummy', n_events)
63 if 'modulation' in events.keys():
64 warnings.warn("'modulation' column found in the given events data.")
65 modulation = np.array(events['modulation']).astype(np.float)
66 return trial_type, onset, duration, modulation
67
[end of nilearn/glm/first_level/experimental_paradigm.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/nilearn/glm/first_level/experimental_paradigm.py b/nilearn/glm/first_level/experimental_paradigm.py
--- a/nilearn/glm/first_level/experimental_paradigm.py
+++ b/nilearn/glm/first_level/experimental_paradigm.py
@@ -17,6 +17,11 @@
import numpy as np
+VALID_FIELDS = set(["onset",
+ "duration",
+ "trial_type",
+ "modulation",
+ ])
def check_events(events):
"""Test that the events data describes a valid experimental paradigm
@@ -54,13 +59,19 @@
onset = np.array(events['onset'])
duration = np.array(events['duration']).astype(np.float)
n_events = len(onset)
- trial_type = np.array(events['trial_type'])
modulation = np.ones(n_events)
if 'trial_type' not in events.keys():
warnings.warn("'trial_type' column not found "
"in the given events data.")
trial_type = np.repeat('dummy', n_events)
+ else:
+ trial_type = np.array(events['trial_type'])
if 'modulation' in events.keys():
warnings.warn("'modulation' column found in the given events data.")
modulation = np.array(events['modulation']).astype(np.float)
+ for event,_ in events.items():
+ if event not in VALID_FIELDS:
+ warnings.warn("Unexpected key `{}` in events "
+ "will be ignored.".format(
+ event))
return trial_type, onset, duration, modulation
|
{"golden_diff": "diff --git a/nilearn/glm/first_level/experimental_paradigm.py b/nilearn/glm/first_level/experimental_paradigm.py\n--- a/nilearn/glm/first_level/experimental_paradigm.py\n+++ b/nilearn/glm/first_level/experimental_paradigm.py\n@@ -17,6 +17,11 @@\n \n import numpy as np\n \n+VALID_FIELDS = set([\"onset\",\n+ \"duration\",\n+ \"trial_type\",\n+ \"modulation\",\n+ ])\n \n def check_events(events):\n \"\"\"Test that the events data describes a valid experimental paradigm\n@@ -54,13 +59,19 @@\n onset = np.array(events['onset'])\n duration = np.array(events['duration']).astype(np.float)\n n_events = len(onset)\n- trial_type = np.array(events['trial_type'])\n modulation = np.ones(n_events)\n if 'trial_type' not in events.keys():\n warnings.warn(\"'trial_type' column not found \"\n \"in the given events data.\")\n trial_type = np.repeat('dummy', n_events)\n+ else:\n+ trial_type = np.array(events['trial_type'])\n if 'modulation' in events.keys():\n warnings.warn(\"'modulation' column found in the given events data.\")\n modulation = np.array(events['modulation']).astype(np.float)\n+ for event,_ in events.items():\n+ if event not in VALID_FIELDS:\n+ warnings.warn(\"Unexpected key `{}` in events \"\n+ \"will be ignored.\".format(\n+ event))\n return trial_type, onset, duration, modulation\n", "issue": "check_paradigm should check for invalid keys in passed dict\nUsing the old `nipy` user logic, I passed `amplitude=somethx` instead of `modulation=somethx` in the `make_design_matrix`. I didn't crash but the values where ignored (e.g Error: unknown param, etc.). A default value of 1 was forced...\n\n", "before_files": [{"content": "\"\"\"\nAn experimental protocol is handled as a pandas DataFrame\nthat includes an 'onset' field.\n\nThis yields the onset time of the events in the experimental paradigm.\nIt can also contain:\n\n * a 'trial_type' field that yields the condition identifier.\n * a 'duration' field that yields event duration (for so-called block\n paradigms).\n * a 'modulation' field that associated a scalar value to each event.\n\nAuthor: Bertrand Thirion, 2015\n\n\"\"\"\nimport warnings\n\nimport numpy as np\n\n\ndef check_events(events):\n \"\"\"Test that the events data describes a valid experimental paradigm\n\n It is valid if the events data has an 'onset' key.\n\n Parameters\n ----------\n events : pandas DataFrame\n Events data that describes a functional experimental paradigm.\n\n Returns\n -------\n trial_type : array of shape (n_events,), dtype='s'\n Per-event experimental conditions identifier.\n Defaults to np.repeat('dummy', len(onsets)).\n\n onset : array of shape (n_events,), dtype='f'\n Per-event onset time (in seconds)\n\n duration : array of shape (n_events,), dtype='f'\n Per-event durantion, (in seconds)\n defaults to zeros(n_events) when no duration is provided\n\n modulation : array of shape (n_events,), dtype='f'\n Per-event modulation, (in seconds)\n defaults to ones(n_events) when no duration is provided.\n\n \"\"\"\n if 'onset' not in events.keys():\n raise ValueError('The provided events data has no onset column.')\n if 'duration' not in events.keys():\n raise ValueError('The provided events data has no duration column.')\n\n onset = np.array(events['onset'])\n duration = np.array(events['duration']).astype(np.float)\n n_events = len(onset)\n trial_type = np.array(events['trial_type'])\n modulation = np.ones(n_events)\n if 'trial_type' not in events.keys():\n warnings.warn(\"'trial_type' column not found \"\n \"in the given events data.\")\n trial_type = np.repeat('dummy', n_events)\n if 'modulation' in events.keys():\n warnings.warn(\"'modulation' column found in the given events data.\")\n modulation = np.array(events['modulation']).astype(np.float)\n return trial_type, onset, duration, modulation\n", "path": "nilearn/glm/first_level/experimental_paradigm.py"}]}
| 1,268 | 351 |
gh_patches_debug_40978
|
rasdani/github-patches
|
git_diff
|
interlegis__sapl-1093
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Migração de Documentos: separar documentos por ano
Nas models de MateriaLegislativa, DocumentoAcessorio e NormaJurídica, seria bom alterar a estrutura de pastas para organizar os documentos migrados de acordo com o ano que foi criado.
Como exemplo, a mudança passaria de:
`media/sapl/{model_name}/{id}/{id}_nome_documento.pdf`
Para:
`media/sapl/{model_name}/{ano}/{id}/{id}_nome_documento.pdf`
</issue>
<code>
[start of sapl/legacy/migracao_documentos.py]
1 import mimetypes
2 import os
3 import re
4
5 import magic
6
7 from django.db.models.signals import post_delete, post_save
8 from sapl.base.models import CasaLegislativa
9 from sapl.materia.models import (DocumentoAcessorio, MateriaLegislativa,
10 Proposicao)
11 from sapl.norma.models import NormaJuridica
12 from sapl.parlamentares.models import Parlamentar
13 from sapl.protocoloadm.models import DocumentoAdministrativo
14 from sapl.protocoloadm.models import DocumentoAcessorioAdministrativo
15 from sapl.sessao.models import SessaoPlenaria
16 from sapl.settings import MEDIA_ROOT
17 from sapl.utils import delete_texto, save_texto
18
19
20 # MIGRAÇÃO DE DOCUMENTOS ###################################################
21 EXTENSOES = {
22 'application/msword': '.doc',
23 'application/pdf': '.pdf',
24 'application/vnd.oasis.opendocument.text': '.odt',
25 'application/vnd.openxmlformats-officedocument.wordprocessingml.document': '.docx', # noqa
26 'application/xml': '.xml',
27 'application/zip': '.zip',
28 'image/jpeg': '.jpeg',
29 'image/png': '.png',
30 'text/html': '.html',
31 'text/rtf': '.rtf',
32 'text/x-python': '.py',
33 'text/plain': '.ksh',
34 'text/plain': '.c',
35 'text/plain': '.h',
36 'text/plain': '.txt',
37 'text/plain': '.bat',
38 'text/plain': '.pl',
39 'text/plain': '.asc',
40 'text/plain': '.text',
41 'text/plain': '.pot',
42 'text/plain': '.brf',
43 'text/plain': '.srt',
44
45 # sem extensao
46 'application/octet-stream': '', # binário
47 'inode/x-empty': '', # vazio
48 }
49
50 DOCS = {
51 CasaLegislativa: [(
52 'logotipo',
53 'props_sapl/logo_casa.gif',
54 'casa/logotipo/logo_casa.gif')],
55 Parlamentar: [(
56 'fotografia',
57 'parlamentar/fotos/{}_foto_parlamentar',
58 'parlamentar/{0}/{0}_foto_parlamentar{1}')],
59 MateriaLegislativa: [(
60 'texto_original',
61 'materia/{}_texto_integral',
62 'materialegislativa/{0}/{0}_texto_integral{1}')],
63 DocumentoAcessorio: [(
64 'arquivo',
65 'materia/{}',
66 'documentoacessorio/{0}/{0}{1}')],
67 NormaJuridica: [(
68 'texto_integral',
69 'norma_juridica/{}_texto_integral',
70 'normajuridica/{0}/{0}_texto_integral{1}')],
71 SessaoPlenaria: [
72 ('upload_ata',
73 'ata_sessao/{}_ata_sessao',
74 'sessaoplenaria/{0}/ata/{0}_ata_sessao{1}'),
75 ('upload_anexo',
76 'anexo_sessao/{}_texto_anexado',
77 'sessaoplenaria/{0}/anexo/{0}_texto_anexado{1}')
78 ],
79 Proposicao: [(
80 'texto_original',
81 'proposicao/{}',
82 'proposicao/{0}/{0}{1}')],
83 DocumentoAdministrativo: [(
84 'texto_integral',
85 'administrativo/{}_texto_integral',
86 'documentoadministrativo/{0}/{0}_texto_integral{1}')
87 ],
88 DocumentoAcessorioAdministrativo: [(
89 'arquivo',
90 'administrativo/{}',
91 'documentoacessorioadministrativo/{0}/{0}_acessorio_administrativo{1}')
92 ],
93 }
94
95 DOCS = {tipo: [(campo,
96 os.path.join('sapl_documentos', origem),
97 os.path.join('sapl', destino))
98 for campo, origem, destino in campos]
99 for tipo, campos in DOCS.items()}
100
101
102 def em_media(caminho):
103 return os.path.join(MEDIA_ROOT, caminho)
104
105
106 def mover_documento(origem, destino):
107 origem, destino = [em_media(c) if not os.path.isabs(c) else c
108 for c in (origem, destino)]
109 os.makedirs(os.path.dirname(destino), exist_ok=True)
110 os.rename(origem, destino)
111
112
113 def get_casa_legislativa():
114 casa = CasaLegislativa.objects.first()
115 if not casa:
116 casa = CasaLegislativa.objects.create(**{k: 'PREENCHER...' for k in [
117 'codigo', 'nome', 'sigla', 'endereco', 'cep', 'municipio', 'uf',
118 ]})
119 return casa
120
121
122 def migrar_docs_logo():
123 print('#### Migrando logotipo da casa ####')
124 [(_, origem, destino)] = DOCS[CasaLegislativa]
125 props_sapl = os.path.dirname(origem)
126
127 # a pasta props_sapl deve conter apenas o origem e metadatas!
128 # Edit: Aparentemente há diretório que contém properties ao invés de
129 # metadata. O assert foi modificado para essa situação.
130 assert set(os.listdir(em_media(props_sapl))) < {
131 'logo_casa.gif', '.metadata', 'logo_casa.gif.metadata',
132 '.properties', 'logo_casa.gif.properties', '.objects'}
133
134 mover_documento(origem, destino)
135 casa = get_casa_legislativa()
136 casa.logotipo = destino
137 casa.save()
138
139
140 def get_extensao(caminho):
141 mime = magic.from_file(caminho, mime=True)
142 try:
143 return EXTENSOES[mime]
144 except KeyError as e:
145 raise Exception('\n'.join([
146 'Extensão não conhecida para o arquivo:',
147 caminho,
148 'E mimetype:',
149 mime,
150 ' Algumas possibilidades são:', ] +
151 [" '{}': '{}',".format(mime, ext)
152 for ext in mimetypes.guess_all_extensions(mime)] +
153 ['Atualize o código do dicionário EXTENSOES!']
154 )) from e
155
156
157 def migrar_docs_por_ids(tipo):
158 for campo, base_origem, base_destino in DOCS[tipo]:
159 print('#### Migrando {} de {} ####'.format(campo, tipo.__name__))
160
161 dir_origem, nome_origem = os.path.split(em_media(base_origem))
162 pat = re.compile('^{}$'.format(nome_origem.format('(\d+)')))
163
164 if not os.path.isdir(dir_origem):
165 print(' >>> O diretório {} não existe! Abortado.'.format(
166 dir_origem))
167 continue
168
169 for arq in os.listdir(dir_origem):
170 match = pat.match(arq)
171 if match:
172 origem = os.path.join(dir_origem, match.group(0))
173 id = match.group(1)
174 extensao = get_extensao(origem)
175 destino = base_destino.format(id, extensao)
176 mover_documento(origem, destino)
177
178 # associa documento ao objeto
179 try:
180 obj = tipo.objects.get(pk=id)
181 setattr(obj, campo, destino)
182 obj.save()
183 except tipo.DoesNotExist:
184 msg = ' {} (pk={}) não encontrado para documento em [{}]'
185 print(msg.format(
186 tipo.__name__, id, destino))
187
188
189 def desconecta_sinais_indexacao():
190 post_save.disconnect(save_texto, NormaJuridica)
191 post_save.disconnect(save_texto, DocumentoAcessorio)
192 post_save.disconnect(save_texto, MateriaLegislativa)
193 post_delete.disconnect(delete_texto, NormaJuridica)
194 post_delete.disconnect(delete_texto, DocumentoAcessorio)
195 post_delete.disconnect(delete_texto, MateriaLegislativa)
196
197
198 def conecta_sinais_indexacao():
199 post_save.connect(save_texto, NormaJuridica)
200 post_save.connect(save_texto, DocumentoAcessorio)
201 post_save.connect(save_texto, MateriaLegislativa)
202 post_delete.connect(delete_texto, NormaJuridica)
203 post_delete.connect(delete_texto, DocumentoAcessorio)
204 post_delete.connect(delete_texto, MateriaLegislativa)
205
206
207 def migrar_documentos():
208 # precisamos excluir os sinais de post_save e post_delete para não que o
209 # computador não trave com a criação de threads desnecessárias
210 desconecta_sinais_indexacao()
211
212 # aqui supomos que uma pasta chamada sapl_documentos está em MEDIA_ROOT
213 # com o conteúdo da pasta de mesmo nome do zope
214 # Os arquivos da pasta serão movidos para a nova estrutura e a pasta será
215 # apagada
216 migrar_docs_logo()
217 for tipo in [
218 Parlamentar,
219 MateriaLegislativa,
220 DocumentoAcessorio,
221 NormaJuridica,
222 SessaoPlenaria,
223 Proposicao,
224 DocumentoAdministrativo,
225 DocumentoAcessorioAdministrativo,
226 ]:
227 migrar_docs_por_ids(tipo)
228
229 sobrando = [os.path.join(dir, file)
230 for (dir, _, files) in os.walk(em_media('sapl_documentos'))
231 for file in files]
232 if sobrando:
233 print('\n#### Encerrado ####\n\n'
234 '{} documentos sobraram sem ser migrados!!!'.format(
235 len(sobrando)))
236 for doc in sobrando:
237 print(' {}'. format(doc))
238 #
239 # reconexão dos sinais desligados no inicio da migração de documentos
240 conecta_sinais_indexacao()
241
[end of sapl/legacy/migracao_documentos.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/sapl/legacy/migracao_documentos.py b/sapl/legacy/migracao_documentos.py
--- a/sapl/legacy/migracao_documentos.py
+++ b/sapl/legacy/migracao_documentos.py
@@ -55,40 +55,40 @@
Parlamentar: [(
'fotografia',
'parlamentar/fotos/{}_foto_parlamentar',
- 'parlamentar/{0}/{0}_foto_parlamentar{1}')],
+ 'public/parlamentar/{0}/{0}_foto_parlamentar{1}')],
MateriaLegislativa: [(
'texto_original',
'materia/{}_texto_integral',
- 'materialegislativa/{0}/{0}_texto_integral{1}')],
+ 'public/materialegislativa/{2}/{0}/{0}_texto_integral{1}')],
DocumentoAcessorio: [(
'arquivo',
'materia/{}',
- 'documentoacessorio/{0}/{0}{1}')],
+ 'public/documentoacessorio/{2}/{0}/{0}{1}')],
NormaJuridica: [(
'texto_integral',
'norma_juridica/{}_texto_integral',
- 'normajuridica/{0}/{0}_texto_integral{1}')],
+ 'public/normajuridica/{2}/{0}/{0}_texto_integral{1}')],
SessaoPlenaria: [
('upload_ata',
'ata_sessao/{}_ata_sessao',
- 'sessaoplenaria/{0}/ata/{0}_ata_sessao{1}'),
+ 'public/sessaoplenaria/{0}/ata/{0}_ata_sessao{1}'),
('upload_anexo',
'anexo_sessao/{}_texto_anexado',
- 'sessaoplenaria/{0}/anexo/{0}_texto_anexado{1}')
+ 'public/sessaoplenaria/{0}/anexo/{0}_texto_anexado{1}')
],
Proposicao: [(
'texto_original',
'proposicao/{}',
- 'proposicao/{0}/{0}{1}')],
+ 'private/proposicao/{0}/{0}{1}')],
DocumentoAdministrativo: [(
'texto_integral',
'administrativo/{}_texto_integral',
- 'documentoadministrativo/{0}/{0}_texto_integral{1}')
+ 'private/documentoadministrativo/{0}/{0}_texto_integral{1}')
],
DocumentoAcessorioAdministrativo: [(
- 'arquivo',
+ 'arquivo',
'administrativo/{}',
- 'documentoacessorioadministrativo/{0}/{0}_acessorio_administrativo{1}')
+ 'private/documentoacessorioadministrativo/{0}/{0}_acessorio_administrativo{1}')
],
}
@@ -169,15 +169,22 @@
for arq in os.listdir(dir_origem):
match = pat.match(arq)
if match:
- origem = os.path.join(dir_origem, match.group(0))
- id = match.group(1)
- extensao = get_extensao(origem)
- destino = base_destino.format(id, extensao)
- mover_documento(origem, destino)
-
# associa documento ao objeto
try:
+ origem = os.path.join(dir_origem, match.group(0))
+ id = match.group(1)
obj = tipo.objects.get(pk=id)
+
+ extensao = get_extensao(origem)
+ if hasattr(obj, "ano"):
+ destino = base_destino.format(id, extensao, obj.ano)
+ elif isinstance(obj, DocumentoAcessorio):
+ destino = base_destino.format(
+ id, extensao, obj.materia.ano)
+ else:
+ destino = base_destino.format(id, extensao)
+ mover_documento(origem, destino)
+
setattr(obj, campo, destino)
obj.save()
except tipo.DoesNotExist:
|
{"golden_diff": "diff --git a/sapl/legacy/migracao_documentos.py b/sapl/legacy/migracao_documentos.py\n--- a/sapl/legacy/migracao_documentos.py\n+++ b/sapl/legacy/migracao_documentos.py\n@@ -55,40 +55,40 @@\n Parlamentar: [(\n 'fotografia',\n 'parlamentar/fotos/{}_foto_parlamentar',\n- 'parlamentar/{0}/{0}_foto_parlamentar{1}')],\n+ 'public/parlamentar/{0}/{0}_foto_parlamentar{1}')],\n MateriaLegislativa: [(\n 'texto_original',\n 'materia/{}_texto_integral',\n- 'materialegislativa/{0}/{0}_texto_integral{1}')],\n+ 'public/materialegislativa/{2}/{0}/{0}_texto_integral{1}')],\n DocumentoAcessorio: [(\n 'arquivo',\n 'materia/{}',\n- 'documentoacessorio/{0}/{0}{1}')],\n+ 'public/documentoacessorio/{2}/{0}/{0}{1}')],\n NormaJuridica: [(\n 'texto_integral',\n 'norma_juridica/{}_texto_integral',\n- 'normajuridica/{0}/{0}_texto_integral{1}')],\n+ 'public/normajuridica/{2}/{0}/{0}_texto_integral{1}')],\n SessaoPlenaria: [\n ('upload_ata',\n 'ata_sessao/{}_ata_sessao',\n- 'sessaoplenaria/{0}/ata/{0}_ata_sessao{1}'),\n+ 'public/sessaoplenaria/{0}/ata/{0}_ata_sessao{1}'),\n ('upload_anexo',\n 'anexo_sessao/{}_texto_anexado',\n- 'sessaoplenaria/{0}/anexo/{0}_texto_anexado{1}')\n+ 'public/sessaoplenaria/{0}/anexo/{0}_texto_anexado{1}')\n ],\n Proposicao: [(\n 'texto_original',\n 'proposicao/{}',\n- 'proposicao/{0}/{0}{1}')],\n+ 'private/proposicao/{0}/{0}{1}')],\n DocumentoAdministrativo: [(\n 'texto_integral',\n 'administrativo/{}_texto_integral',\n- 'documentoadministrativo/{0}/{0}_texto_integral{1}') \n+ 'private/documentoadministrativo/{0}/{0}_texto_integral{1}')\n ],\n DocumentoAcessorioAdministrativo: [(\n- \t 'arquivo',\n+ 'arquivo',\n 'administrativo/{}',\n- 'documentoacessorioadministrativo/{0}/{0}_acessorio_administrativo{1}')\n+ 'private/documentoacessorioadministrativo/{0}/{0}_acessorio_administrativo{1}')\n ],\n }\n \n@@ -169,15 +169,22 @@\n for arq in os.listdir(dir_origem):\n match = pat.match(arq)\n if match:\n- origem = os.path.join(dir_origem, match.group(0))\n- id = match.group(1)\n- extensao = get_extensao(origem)\n- destino = base_destino.format(id, extensao)\n- mover_documento(origem, destino)\n-\n # associa documento ao objeto\n try:\n+ origem = os.path.join(dir_origem, match.group(0))\n+ id = match.group(1)\n obj = tipo.objects.get(pk=id)\n+\n+ extensao = get_extensao(origem)\n+ if hasattr(obj, \"ano\"):\n+ destino = base_destino.format(id, extensao, obj.ano)\n+ elif isinstance(obj, DocumentoAcessorio):\n+ destino = base_destino.format(\n+ id, extensao, obj.materia.ano)\n+ else:\n+ destino = base_destino.format(id, extensao)\n+ mover_documento(origem, destino)\n+\n setattr(obj, campo, destino)\n obj.save()\n except tipo.DoesNotExist:\n", "issue": "Migra\u00e7\u00e3o de Documentos: separar documentos por ano\nNas models de MateriaLegislativa, DocumentoAcessorio e NormaJur\u00eddica, seria bom alterar a estrutura de pastas para organizar os documentos migrados de acordo com o ano que foi criado.\r\n\r\nComo exemplo, a mudan\u00e7a passaria de:\r\n`media/sapl/{model_name}/{id}/{id}_nome_documento.pdf`\r\nPara:\r\n`media/sapl/{model_name}/{ano}/{id}/{id}_nome_documento.pdf`\n", "before_files": [{"content": "import mimetypes\nimport os\nimport re\n\nimport magic\n\nfrom django.db.models.signals import post_delete, post_save\nfrom sapl.base.models import CasaLegislativa\nfrom sapl.materia.models import (DocumentoAcessorio, MateriaLegislativa,\n Proposicao)\nfrom sapl.norma.models import NormaJuridica\nfrom sapl.parlamentares.models import Parlamentar\nfrom sapl.protocoloadm.models import DocumentoAdministrativo\nfrom sapl.protocoloadm.models import DocumentoAcessorioAdministrativo\nfrom sapl.sessao.models import SessaoPlenaria\nfrom sapl.settings import MEDIA_ROOT\nfrom sapl.utils import delete_texto, save_texto\n\n\n# MIGRA\u00c7\u00c3O DE DOCUMENTOS ###################################################\nEXTENSOES = {\n 'application/msword': '.doc',\n 'application/pdf': '.pdf',\n 'application/vnd.oasis.opendocument.text': '.odt',\n 'application/vnd.openxmlformats-officedocument.wordprocessingml.document': '.docx', # noqa\n 'application/xml': '.xml',\n 'application/zip': '.zip',\n 'image/jpeg': '.jpeg',\n 'image/png': '.png',\n 'text/html': '.html',\n 'text/rtf': '.rtf',\n 'text/x-python': '.py',\n 'text/plain': '.ksh',\n 'text/plain': '.c',\n 'text/plain': '.h',\n 'text/plain': '.txt',\n 'text/plain': '.bat',\n 'text/plain': '.pl',\n 'text/plain': '.asc',\n 'text/plain': '.text',\n 'text/plain': '.pot',\n 'text/plain': '.brf',\n 'text/plain': '.srt',\n\n # sem extensao\n 'application/octet-stream': '', # bin\u00e1rio\n 'inode/x-empty': '', # vazio\n}\n\nDOCS = {\n CasaLegislativa: [(\n 'logotipo',\n 'props_sapl/logo_casa.gif',\n 'casa/logotipo/logo_casa.gif')],\n Parlamentar: [(\n 'fotografia',\n 'parlamentar/fotos/{}_foto_parlamentar',\n 'parlamentar/{0}/{0}_foto_parlamentar{1}')],\n MateriaLegislativa: [(\n 'texto_original',\n 'materia/{}_texto_integral',\n 'materialegislativa/{0}/{0}_texto_integral{1}')],\n DocumentoAcessorio: [(\n 'arquivo',\n 'materia/{}',\n 'documentoacessorio/{0}/{0}{1}')],\n NormaJuridica: [(\n 'texto_integral',\n 'norma_juridica/{}_texto_integral',\n 'normajuridica/{0}/{0}_texto_integral{1}')],\n SessaoPlenaria: [\n ('upload_ata',\n 'ata_sessao/{}_ata_sessao',\n 'sessaoplenaria/{0}/ata/{0}_ata_sessao{1}'),\n ('upload_anexo',\n 'anexo_sessao/{}_texto_anexado',\n 'sessaoplenaria/{0}/anexo/{0}_texto_anexado{1}')\n ],\n Proposicao: [(\n 'texto_original',\n 'proposicao/{}',\n 'proposicao/{0}/{0}{1}')],\n DocumentoAdministrativo: [(\n 'texto_integral',\n 'administrativo/{}_texto_integral',\n 'documentoadministrativo/{0}/{0}_texto_integral{1}') \n ],\n DocumentoAcessorioAdministrativo: [(\n \t 'arquivo',\n 'administrativo/{}',\n 'documentoacessorioadministrativo/{0}/{0}_acessorio_administrativo{1}')\n ],\n}\n\nDOCS = {tipo: [(campo,\n os.path.join('sapl_documentos', origem),\n os.path.join('sapl', destino))\n for campo, origem, destino in campos]\n for tipo, campos in DOCS.items()}\n\n\ndef em_media(caminho):\n return os.path.join(MEDIA_ROOT, caminho)\n\n\ndef mover_documento(origem, destino):\n origem, destino = [em_media(c) if not os.path.isabs(c) else c\n for c in (origem, destino)]\n os.makedirs(os.path.dirname(destino), exist_ok=True)\n os.rename(origem, destino)\n\n\ndef get_casa_legislativa():\n casa = CasaLegislativa.objects.first()\n if not casa:\n casa = CasaLegislativa.objects.create(**{k: 'PREENCHER...' for k in [\n 'codigo', 'nome', 'sigla', 'endereco', 'cep', 'municipio', 'uf',\n ]})\n return casa\n\n\ndef migrar_docs_logo():\n print('#### Migrando logotipo da casa ####')\n [(_, origem, destino)] = DOCS[CasaLegislativa]\n props_sapl = os.path.dirname(origem)\n\n # a pasta props_sapl deve conter apenas o origem e metadatas!\n # Edit: Aparentemente h\u00e1 diret\u00f3rio que cont\u00e9m properties ao inv\u00e9s de\n # metadata. O assert foi modificado para essa situa\u00e7\u00e3o.\n assert set(os.listdir(em_media(props_sapl))) < {\n 'logo_casa.gif', '.metadata', 'logo_casa.gif.metadata',\n '.properties', 'logo_casa.gif.properties', '.objects'}\n\n mover_documento(origem, destino)\n casa = get_casa_legislativa()\n casa.logotipo = destino\n casa.save()\n\n\ndef get_extensao(caminho):\n mime = magic.from_file(caminho, mime=True)\n try:\n return EXTENSOES[mime]\n except KeyError as e:\n raise Exception('\\n'.join([\n 'Extens\u00e3o n\u00e3o conhecida para o arquivo:',\n caminho,\n 'E mimetype:',\n mime,\n ' Algumas possibilidades s\u00e3o:', ] +\n [\" '{}': '{}',\".format(mime, ext)\n for ext in mimetypes.guess_all_extensions(mime)] +\n ['Atualize o c\u00f3digo do dicion\u00e1rio EXTENSOES!']\n )) from e\n\n\ndef migrar_docs_por_ids(tipo):\n for campo, base_origem, base_destino in DOCS[tipo]:\n print('#### Migrando {} de {} ####'.format(campo, tipo.__name__))\n\n dir_origem, nome_origem = os.path.split(em_media(base_origem))\n pat = re.compile('^{}$'.format(nome_origem.format('(\\d+)')))\n\n if not os.path.isdir(dir_origem):\n print(' >>> O diret\u00f3rio {} n\u00e3o existe! Abortado.'.format(\n dir_origem))\n continue\n\n for arq in os.listdir(dir_origem):\n match = pat.match(arq)\n if match:\n origem = os.path.join(dir_origem, match.group(0))\n id = match.group(1)\n extensao = get_extensao(origem)\n destino = base_destino.format(id, extensao)\n mover_documento(origem, destino)\n\n # associa documento ao objeto\n try:\n obj = tipo.objects.get(pk=id)\n setattr(obj, campo, destino)\n obj.save()\n except tipo.DoesNotExist:\n msg = ' {} (pk={}) n\u00e3o encontrado para documento em [{}]'\n print(msg.format(\n tipo.__name__, id, destino))\n\n\ndef desconecta_sinais_indexacao():\n post_save.disconnect(save_texto, NormaJuridica)\n post_save.disconnect(save_texto, DocumentoAcessorio)\n post_save.disconnect(save_texto, MateriaLegislativa)\n post_delete.disconnect(delete_texto, NormaJuridica)\n post_delete.disconnect(delete_texto, DocumentoAcessorio)\n post_delete.disconnect(delete_texto, MateriaLegislativa)\n\n\ndef conecta_sinais_indexacao():\n post_save.connect(save_texto, NormaJuridica)\n post_save.connect(save_texto, DocumentoAcessorio)\n post_save.connect(save_texto, MateriaLegislativa)\n post_delete.connect(delete_texto, NormaJuridica)\n post_delete.connect(delete_texto, DocumentoAcessorio)\n post_delete.connect(delete_texto, MateriaLegislativa)\n\n\ndef migrar_documentos():\n # precisamos excluir os sinais de post_save e post_delete para n\u00e3o que o\n # computador n\u00e3o trave com a cria\u00e7\u00e3o de threads desnecess\u00e1rias\n desconecta_sinais_indexacao()\n\n # aqui supomos que uma pasta chamada sapl_documentos est\u00e1 em MEDIA_ROOT\n # com o conte\u00fado da pasta de mesmo nome do zope\n # Os arquivos da pasta ser\u00e3o movidos para a nova estrutura e a pasta ser\u00e1\n # apagada\n migrar_docs_logo()\n for tipo in [\n Parlamentar,\n MateriaLegislativa,\n DocumentoAcessorio,\n NormaJuridica,\n SessaoPlenaria,\n Proposicao,\n DocumentoAdministrativo,\n \tDocumentoAcessorioAdministrativo,\n ]:\n migrar_docs_por_ids(tipo)\n\n sobrando = [os.path.join(dir, file)\n for (dir, _, files) in os.walk(em_media('sapl_documentos'))\n for file in files]\n if sobrando:\n print('\\n#### Encerrado ####\\n\\n'\n '{} documentos sobraram sem ser migrados!!!'.format(\n len(sobrando)))\n for doc in sobrando:\n print(' {}'. format(doc))\n #\n # reconex\u00e3o dos sinais desligados no inicio da migra\u00e7\u00e3o de documentos\n conecta_sinais_indexacao()\n", "path": "sapl/legacy/migracao_documentos.py"}]}
| 3,430 | 931 |
gh_patches_debug_25318
|
rasdani/github-patches
|
git_diff
|
getsentry__sentry-python-484
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Celery - Queue object has no attribute 'all_tasks_done'
Hi all,
I'm integrating Sentry on a project in python that uses Celery. I'm getting this error when shutting down the worker:
```
Error in atexit._run_exitfuncs:
Traceback (most recent call last):
File "/Users/jibanez/API/.conda/envs/cimrender/lib/python3.6/site-packages/sentry_sdk/worker.py", line 84, in flush
self._wait_flush(timeout, callback)
File "/Users/jibanez/API/.conda/envs/cimrender/lib/python3.6/site-packages/sentry_sdk/worker.py", line 90, in _wait_flush
if not self._timed_queue_join(initial_timeout):
File "/Users/jibanez/API/.conda/envs/cimrender/lib/python3.6/site-packages/sentry_sdk/worker.py", line 48, in _timed_queue_join
queue.all_tasks_done.acquire() # type: ignore
AttributeError: 'Queue' object has no attribute 'all_tasks_done'
```
I'm using:
- Python 3.6
- Celery 4.3.0
- OSX Mojave
Any thoughts?
</issue>
<code>
[start of sentry_sdk/worker.py]
1 import os
2
3 from threading import Thread, Lock
4 from time import sleep, time
5 from sentry_sdk._compat import queue, check_thread_support
6 from sentry_sdk.utils import logger
7
8
9 from sentry_sdk._types import MYPY
10
11 if MYPY:
12 from queue import Queue
13 from typing import Any
14 from typing import Optional
15 from typing import Callable
16
17
18 _TERMINATOR = object()
19
20
21 class BackgroundWorker(object):
22 def __init__(self):
23 # type: () -> None
24 check_thread_support()
25 self._queue = queue.Queue(-1) # type: Queue[Any]
26 self._lock = Lock()
27 self._thread = None # type: Optional[Thread]
28 self._thread_for_pid = None # type: Optional[int]
29
30 @property
31 def is_alive(self):
32 # type: () -> bool
33 if self._thread_for_pid != os.getpid():
34 return False
35 if not self._thread:
36 return False
37 return self._thread.is_alive()
38
39 def _ensure_thread(self):
40 # type: () -> None
41 if not self.is_alive:
42 self.start()
43
44 def _timed_queue_join(self, timeout):
45 # type: (float) -> bool
46 deadline = time() + timeout
47 queue = self._queue
48 queue.all_tasks_done.acquire() # type: ignore
49 try:
50 while queue.unfinished_tasks: # type: ignore
51 delay = deadline - time()
52 if delay <= 0:
53 return False
54 queue.all_tasks_done.wait(timeout=delay) # type: ignore
55 return True
56 finally:
57 queue.all_tasks_done.release() # type: ignore
58
59 def start(self):
60 # type: () -> None
61 with self._lock:
62 if not self.is_alive:
63 self._thread = Thread(
64 target=self._target, name="raven-sentry.BackgroundWorker"
65 )
66 self._thread.setDaemon(True)
67 self._thread.start()
68 self._thread_for_pid = os.getpid()
69
70 def kill(self):
71 # type: () -> None
72 logger.debug("background worker got kill request")
73 with self._lock:
74 if self._thread:
75 self._queue.put_nowait(_TERMINATOR)
76 self._thread = None
77 self._thread_for_pid = None
78
79 def flush(self, timeout, callback=None):
80 # type: (float, Optional[Any]) -> None
81 logger.debug("background worker got flush request")
82 with self._lock:
83 if self.is_alive and timeout > 0.0:
84 self._wait_flush(timeout, callback)
85 logger.debug("background worker flushed")
86
87 def _wait_flush(self, timeout, callback):
88 # type: (float, Optional[Any]) -> None
89 initial_timeout = min(0.1, timeout)
90 if not self._timed_queue_join(initial_timeout):
91 pending = self._queue.qsize()
92 logger.debug("%d event(s) pending on flush", pending)
93 if callback is not None:
94 callback(pending, timeout)
95 self._timed_queue_join(timeout - initial_timeout)
96
97 def submit(self, callback):
98 # type: (Callable[[], None]) -> None
99 self._ensure_thread()
100 self._queue.put_nowait(callback)
101
102 def _target(self):
103 # type: () -> None
104 while True:
105 callback = self._queue.get()
106 try:
107 if callback is _TERMINATOR:
108 break
109 try:
110 callback()
111 except Exception:
112 logger.error("Failed processing job", exc_info=True)
113 finally:
114 self._queue.task_done()
115 sleep(0)
116
[end of sentry_sdk/worker.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/sentry_sdk/worker.py b/sentry_sdk/worker.py
--- a/sentry_sdk/worker.py
+++ b/sentry_sdk/worker.py
@@ -45,16 +45,33 @@
# type: (float) -> bool
deadline = time() + timeout
queue = self._queue
- queue.all_tasks_done.acquire() # type: ignore
+
+ real_all_tasks_done = getattr(
+ queue, "all_tasks_done", None
+ ) # type: Optional[Any]
+ if real_all_tasks_done is not None:
+ real_all_tasks_done.acquire()
+ all_tasks_done = real_all_tasks_done # type: Optional[Any]
+ elif queue.__module__.startswith("eventlet."):
+ all_tasks_done = getattr(queue, "_cond", None)
+ else:
+ all_tasks_done = None
+
try:
while queue.unfinished_tasks: # type: ignore
delay = deadline - time()
if delay <= 0:
return False
- queue.all_tasks_done.wait(timeout=delay) # type: ignore
+ if all_tasks_done is not None:
+ all_tasks_done.wait(timeout=delay)
+ else:
+ # worst case, we just poll the number of remaining tasks
+ sleep(0.1)
+
return True
finally:
- queue.all_tasks_done.release() # type: ignore
+ if real_all_tasks_done is not None:
+ real_all_tasks_done.release() # type: ignore
def start(self):
# type: () -> None
|
{"golden_diff": "diff --git a/sentry_sdk/worker.py b/sentry_sdk/worker.py\n--- a/sentry_sdk/worker.py\n+++ b/sentry_sdk/worker.py\n@@ -45,16 +45,33 @@\n # type: (float) -> bool\n deadline = time() + timeout\n queue = self._queue\n- queue.all_tasks_done.acquire() # type: ignore\n+\n+ real_all_tasks_done = getattr(\n+ queue, \"all_tasks_done\", None\n+ ) # type: Optional[Any]\n+ if real_all_tasks_done is not None:\n+ real_all_tasks_done.acquire()\n+ all_tasks_done = real_all_tasks_done # type: Optional[Any]\n+ elif queue.__module__.startswith(\"eventlet.\"):\n+ all_tasks_done = getattr(queue, \"_cond\", None)\n+ else:\n+ all_tasks_done = None\n+\n try:\n while queue.unfinished_tasks: # type: ignore\n delay = deadline - time()\n if delay <= 0:\n return False\n- queue.all_tasks_done.wait(timeout=delay) # type: ignore\n+ if all_tasks_done is not None:\n+ all_tasks_done.wait(timeout=delay)\n+ else:\n+ # worst case, we just poll the number of remaining tasks\n+ sleep(0.1)\n+\n return True\n finally:\n- queue.all_tasks_done.release() # type: ignore\n+ if real_all_tasks_done is not None:\n+ real_all_tasks_done.release() # type: ignore\n \n def start(self):\n # type: () -> None\n", "issue": "Celery - Queue object has no attribute 'all_tasks_done'\nHi all, \r\n\r\nI'm integrating Sentry on a project in python that uses Celery. I'm getting this error when shutting down the worker: \r\n\r\n```\r\nError in atexit._run_exitfuncs:\r\nTraceback (most recent call last):\r\n File \"/Users/jibanez/API/.conda/envs/cimrender/lib/python3.6/site-packages/sentry_sdk/worker.py\", line 84, in flush\r\n self._wait_flush(timeout, callback)\r\n File \"/Users/jibanez/API/.conda/envs/cimrender/lib/python3.6/site-packages/sentry_sdk/worker.py\", line 90, in _wait_flush\r\n if not self._timed_queue_join(initial_timeout):\r\n File \"/Users/jibanez/API/.conda/envs/cimrender/lib/python3.6/site-packages/sentry_sdk/worker.py\", line 48, in _timed_queue_join\r\n queue.all_tasks_done.acquire() # type: ignore\r\nAttributeError: 'Queue' object has no attribute 'all_tasks_done'\r\n```\r\n\r\nI'm using: \r\n- Python 3.6\r\n- Celery 4.3.0\r\n- OSX Mojave\r\n\r\nAny thoughts? \n", "before_files": [{"content": "import os\n\nfrom threading import Thread, Lock\nfrom time import sleep, time\nfrom sentry_sdk._compat import queue, check_thread_support\nfrom sentry_sdk.utils import logger\n\n\nfrom sentry_sdk._types import MYPY\n\nif MYPY:\n from queue import Queue\n from typing import Any\n from typing import Optional\n from typing import Callable\n\n\n_TERMINATOR = object()\n\n\nclass BackgroundWorker(object):\n def __init__(self):\n # type: () -> None\n check_thread_support()\n self._queue = queue.Queue(-1) # type: Queue[Any]\n self._lock = Lock()\n self._thread = None # type: Optional[Thread]\n self._thread_for_pid = None # type: Optional[int]\n\n @property\n def is_alive(self):\n # type: () -> bool\n if self._thread_for_pid != os.getpid():\n return False\n if not self._thread:\n return False\n return self._thread.is_alive()\n\n def _ensure_thread(self):\n # type: () -> None\n if not self.is_alive:\n self.start()\n\n def _timed_queue_join(self, timeout):\n # type: (float) -> bool\n deadline = time() + timeout\n queue = self._queue\n queue.all_tasks_done.acquire() # type: ignore\n try:\n while queue.unfinished_tasks: # type: ignore\n delay = deadline - time()\n if delay <= 0:\n return False\n queue.all_tasks_done.wait(timeout=delay) # type: ignore\n return True\n finally:\n queue.all_tasks_done.release() # type: ignore\n\n def start(self):\n # type: () -> None\n with self._lock:\n if not self.is_alive:\n self._thread = Thread(\n target=self._target, name=\"raven-sentry.BackgroundWorker\"\n )\n self._thread.setDaemon(True)\n self._thread.start()\n self._thread_for_pid = os.getpid()\n\n def kill(self):\n # type: () -> None\n logger.debug(\"background worker got kill request\")\n with self._lock:\n if self._thread:\n self._queue.put_nowait(_TERMINATOR)\n self._thread = None\n self._thread_for_pid = None\n\n def flush(self, timeout, callback=None):\n # type: (float, Optional[Any]) -> None\n logger.debug(\"background worker got flush request\")\n with self._lock:\n if self.is_alive and timeout > 0.0:\n self._wait_flush(timeout, callback)\n logger.debug(\"background worker flushed\")\n\n def _wait_flush(self, timeout, callback):\n # type: (float, Optional[Any]) -> None\n initial_timeout = min(0.1, timeout)\n if not self._timed_queue_join(initial_timeout):\n pending = self._queue.qsize()\n logger.debug(\"%d event(s) pending on flush\", pending)\n if callback is not None:\n callback(pending, timeout)\n self._timed_queue_join(timeout - initial_timeout)\n\n def submit(self, callback):\n # type: (Callable[[], None]) -> None\n self._ensure_thread()\n self._queue.put_nowait(callback)\n\n def _target(self):\n # type: () -> None\n while True:\n callback = self._queue.get()\n try:\n if callback is _TERMINATOR:\n break\n try:\n callback()\n except Exception:\n logger.error(\"Failed processing job\", exc_info=True)\n finally:\n self._queue.task_done()\n sleep(0)\n", "path": "sentry_sdk/worker.py"}]}
| 1,837 | 354 |
gh_patches_debug_9960
|
rasdani/github-patches
|
git_diff
|
RocketMap__RocketMap-243
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Error with JS
Recently I clone the repository and I get this errors:
> map.built.js:227 Uncaught ReferenceError: center_lat is not defined
>
> map.built.js:1082 Uncaught TypeError: Cannot read property 'setMapTypeId' of undefined
>
> 19 map.built.js:677 Uncaught TypeError: Cannot read property 'getBounds' of undefined
Error with JS
Recently I clone the repository and I get this errors:
> map.built.js:227 Uncaught ReferenceError: center_lat is not defined
>
> map.built.js:1082 Uncaught TypeError: Cannot read property 'setMapTypeId' of undefined
>
> 19 map.built.js:677 Uncaught TypeError: Cannot read property 'getBounds' of undefined
</issue>
<code>
[start of runserver.py]
1 #!/usr/bin/python
2 # -*- coding: utf-8 -*-
3
4 import os
5 import sys
6 import shutil
7 import logging
8 import time
9 import re
10 import requests
11
12 # Currently supported pgoapi
13 pgoapi_version = "1.1.6"
14
15 # Moved here so logger is configured at load time
16 logging.basicConfig(format='%(asctime)s [%(threadName)16s][%(module)14s][%(levelname)8s] %(message)s')
17 log = logging.getLogger()
18
19 # Make sure pogom/pgoapi is actually removed if it is an empty directory
20 # This is a leftover directory from the time pgoapi was embedded in PokemonGo-Map
21 # The empty directory will cause problems with `import pgoapi` so it needs to go
22 oldpgoapiPath = os.path.join(os.path.dirname(__file__), "pogom/pgoapi")
23 if os.path.isdir(oldpgoapiPath):
24 log.info("I found %s, but its no longer used. Going to remove it...", oldpgoapiPath)
25 shutil.rmtree(oldpgoapiPath)
26 log.info("Done!")
27
28 # Assert pgoapi is installed
29 try:
30 import pgoapi
31 except ImportError:
32 log.critical("It seems `pgoapi` is not installed. You must run pip install -r requirements.txt again")
33 sys.exit(1)
34
35 # Assert pgoapi >= 1.1.6 is installed
36 from distutils.version import StrictVersion
37 if not hasattr(pgoapi, "__version__") or StrictVersion(pgoapi.__version__) < StrictVersion(pgoapi_version):
38 log.critical("It seems `pgoapi` is not up-to-date. You must run pip install -r requirements.txt again")
39 sys.exit(1)
40
41 from threading import Thread, Event
42 from queue import Queue
43 from flask_cors import CORS
44
45 from pogom import config
46 from pogom.app import Pogom
47 from pogom.utils import get_args, insert_mock_data, get_encryption_lib_path
48
49 from pogom.search import search_overseer_thread, fake_search_loop
50 from pogom.models import init_database, create_tables, drop_tables, Pokemon, Pokestop, Gym
51
52 from pgoapi import utilities as util
53
54 if __name__ == '__main__':
55 # Check if we have the proper encryption library file and get its path
56 encryption_lib_path = get_encryption_lib_path()
57 if encryption_lib_path is "":
58 sys.exit(1)
59
60 args = get_args()
61
62 if args.debug:
63 log.setLevel(logging.DEBUG);
64 else:
65 log.setLevel(logging.INFO);
66
67 # Let's not forget to run Grunt / Only needed when running with webserver
68 if not args.no_server:
69 if not os.path.exists(os.path.join(os.path.dirname(__file__), 'static/dist')):
70 log.critical('Missing front-end assets (static/dist) -- please run "npm install && npm run build" before starting the server');
71 sys.exit();
72
73 # These are very noisey, let's shush them up a bit
74 logging.getLogger('peewee').setLevel(logging.INFO)
75 logging.getLogger('requests').setLevel(logging.WARNING)
76 logging.getLogger('pgoapi.pgoapi').setLevel(logging.WARNING)
77 logging.getLogger('pgoapi.rpc_api').setLevel(logging.INFO)
78 logging.getLogger('werkzeug').setLevel(logging.ERROR)
79
80 config['parse_pokemon'] = not args.no_pokemon
81 config['parse_pokestops'] = not args.no_pokestops
82 config['parse_gyms'] = not args.no_gyms
83
84 # Turn these back up if debugging
85 if args.debug:
86 logging.getLogger('requests').setLevel(logging.DEBUG)
87 logging.getLogger('pgoapi').setLevel(logging.DEBUG)
88 logging.getLogger('rpc_api').setLevel(logging.DEBUG)
89
90
91 # use lat/lng directly if matches such a pattern
92 prog = re.compile("^(\-?\d+\.\d+),?\s?(\-?\d+\.\d+)$")
93 res = prog.match(args.location)
94 if res:
95 log.debug('Using coordinates from CLI directly')
96 position = (float(res.group(1)), float(res.group(2)), 0)
97 else:
98 log.debug('Looking up coordinates in API')
99 position = util.get_pos_by_name(args.location)
100
101 # Use the latitude and longitude to get the local altitude from Google
102 try:
103 url = 'https://maps.googleapis.com/maps/api/elevation/json?locations={},{}'.format(
104 str(position[0]), str(position[1]))
105 altitude = requests.get(url).json()[u'results'][0][u'elevation']
106 log.debug('Local altitude is: %sm', altitude)
107 position = (position[0], position[1], altitude)
108 except requests.exceptions.RequestException:
109 log.error('Unable to retrieve altitude from Google APIs; setting to 0')
110
111 if not any(position):
112 log.error('Could not get a position by name, aborting')
113 sys.exit()
114
115 log.info('Parsed location is: %.4f/%.4f/%.4f (lat/lng/alt)',
116 position[0], position[1], position[2])
117
118 if args.no_pokemon:
119 log.info('Parsing of Pokemon disabled')
120 if args.no_pokestops:
121 log.info('Parsing of Pokestops disabled')
122 if args.no_gyms:
123 log.info('Parsing of Gyms disabled')
124
125 config['LOCALE'] = args.locale
126 config['CHINA'] = args.china
127
128 app = Pogom(__name__)
129 db = init_database(app)
130 if args.clear_db:
131 log.info('Clearing database')
132 if args.db_type == 'mysql':
133 drop_tables(db)
134 elif os.path.isfile(args.db):
135 os.remove(args.db)
136 create_tables(db)
137
138 app.set_current_location(position);
139
140 # Control the search status (running or not) across threads
141 pause_bit = Event()
142 pause_bit.clear()
143
144 # Setup the location tracking queue and push the first location on
145 new_location_queue = Queue()
146 new_location_queue.put(position)
147
148 if not args.only_server:
149 # Gather the pokemons!
150 if not args.mock:
151 log.debug('Starting a real search thread')
152 search_thread = Thread(target=search_overseer_thread, args=(args, new_location_queue, pause_bit, encryption_lib_path))
153 else:
154 log.debug('Starting a fake search thread')
155 insert_mock_data(position)
156 search_thread = Thread(target=fake_search_loop)
157
158 search_thread.daemon = True
159 search_thread.name = 'search_thread'
160 search_thread.start()
161
162 if args.cors:
163 CORS(app);
164
165 app.set_search_control(pause_bit)
166 app.set_location_queue(new_location_queue)
167
168 config['ROOT_PATH'] = app.root_path
169 config['GMAPS_KEY'] = args.gmaps_key
170
171 if args.no_server:
172 # This loop allows for ctrl-c interupts to work since flask won't be holding the program open
173 while search_thread.is_alive():
174 time.sleep(60)
175 else:
176 app.run(threaded=True, use_reloader=False, debug=args.debug, host=args.host, port=args.port)
177
[end of runserver.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/runserver.py b/runserver.py
--- a/runserver.py
+++ b/runserver.py
@@ -41,6 +41,7 @@
from threading import Thread, Event
from queue import Queue
from flask_cors import CORS
+from flask.ext import cache_bust
from pogom import config
from pogom.app import Pogom
@@ -162,6 +163,9 @@
if args.cors:
CORS(app);
+ # No more stale JS
+ cache_bust.init_cache_busting(app)
+
app.set_search_control(pause_bit)
app.set_location_queue(new_location_queue)
|
{"golden_diff": "diff --git a/runserver.py b/runserver.py\n--- a/runserver.py\n+++ b/runserver.py\n@@ -41,6 +41,7 @@\n from threading import Thread, Event\n from queue import Queue\n from flask_cors import CORS\n+from flask.ext import cache_bust\n \n from pogom import config\n from pogom.app import Pogom\n@@ -162,6 +163,9 @@\n if args.cors:\n CORS(app);\n \n+ # No more stale JS\n+ cache_bust.init_cache_busting(app)\n+\n app.set_search_control(pause_bit)\n app.set_location_queue(new_location_queue)\n", "issue": "Error with JS\nRecently I clone the repository and I get this errors:\n\n> map.built.js:227 Uncaught ReferenceError: center_lat is not defined\n> \n> map.built.js:1082 Uncaught TypeError: Cannot read property 'setMapTypeId' of undefined\n> \n> 19 map.built.js:677 Uncaught TypeError: Cannot read property 'getBounds' of undefined\n\nError with JS\nRecently I clone the repository and I get this errors:\n\n> map.built.js:227 Uncaught ReferenceError: center_lat is not defined\n> \n> map.built.js:1082 Uncaught TypeError: Cannot read property 'setMapTypeId' of undefined\n> \n> 19 map.built.js:677 Uncaught TypeError: Cannot read property 'getBounds' of undefined\n\n", "before_files": [{"content": "#!/usr/bin/python\n# -*- coding: utf-8 -*-\n\nimport os\nimport sys\nimport shutil\nimport logging\nimport time\nimport re\nimport requests\n\n# Currently supported pgoapi\npgoapi_version = \"1.1.6\"\n\n# Moved here so logger is configured at load time\nlogging.basicConfig(format='%(asctime)s [%(threadName)16s][%(module)14s][%(levelname)8s] %(message)s')\nlog = logging.getLogger()\n\n# Make sure pogom/pgoapi is actually removed if it is an empty directory\n# This is a leftover directory from the time pgoapi was embedded in PokemonGo-Map\n# The empty directory will cause problems with `import pgoapi` so it needs to go\noldpgoapiPath = os.path.join(os.path.dirname(__file__), \"pogom/pgoapi\")\nif os.path.isdir(oldpgoapiPath):\n log.info(\"I found %s, but its no longer used. Going to remove it...\", oldpgoapiPath)\n shutil.rmtree(oldpgoapiPath)\n log.info(\"Done!\")\n\n# Assert pgoapi is installed\ntry:\n import pgoapi\nexcept ImportError:\n log.critical(\"It seems `pgoapi` is not installed. You must run pip install -r requirements.txt again\")\n sys.exit(1)\n\n# Assert pgoapi >= 1.1.6 is installed\nfrom distutils.version import StrictVersion\nif not hasattr(pgoapi, \"__version__\") or StrictVersion(pgoapi.__version__) < StrictVersion(pgoapi_version):\n log.critical(\"It seems `pgoapi` is not up-to-date. You must run pip install -r requirements.txt again\")\n sys.exit(1)\n\nfrom threading import Thread, Event\nfrom queue import Queue\nfrom flask_cors import CORS\n\nfrom pogom import config\nfrom pogom.app import Pogom\nfrom pogom.utils import get_args, insert_mock_data, get_encryption_lib_path\n\nfrom pogom.search import search_overseer_thread, fake_search_loop\nfrom pogom.models import init_database, create_tables, drop_tables, Pokemon, Pokestop, Gym\n\nfrom pgoapi import utilities as util\n\nif __name__ == '__main__':\n # Check if we have the proper encryption library file and get its path\n encryption_lib_path = get_encryption_lib_path()\n if encryption_lib_path is \"\":\n sys.exit(1)\n\n args = get_args()\n\n if args.debug:\n log.setLevel(logging.DEBUG);\n else:\n log.setLevel(logging.INFO);\n\n # Let's not forget to run Grunt / Only needed when running with webserver\n if not args.no_server:\n if not os.path.exists(os.path.join(os.path.dirname(__file__), 'static/dist')):\n log.critical('Missing front-end assets (static/dist) -- please run \"npm install && npm run build\" before starting the server');\n sys.exit();\n\n # These are very noisey, let's shush them up a bit\n logging.getLogger('peewee').setLevel(logging.INFO)\n logging.getLogger('requests').setLevel(logging.WARNING)\n logging.getLogger('pgoapi.pgoapi').setLevel(logging.WARNING)\n logging.getLogger('pgoapi.rpc_api').setLevel(logging.INFO)\n logging.getLogger('werkzeug').setLevel(logging.ERROR)\n\n config['parse_pokemon'] = not args.no_pokemon\n config['parse_pokestops'] = not args.no_pokestops\n config['parse_gyms'] = not args.no_gyms\n\n # Turn these back up if debugging\n if args.debug:\n logging.getLogger('requests').setLevel(logging.DEBUG)\n logging.getLogger('pgoapi').setLevel(logging.DEBUG)\n logging.getLogger('rpc_api').setLevel(logging.DEBUG)\n\n\n # use lat/lng directly if matches such a pattern\n prog = re.compile(\"^(\\-?\\d+\\.\\d+),?\\s?(\\-?\\d+\\.\\d+)$\")\n res = prog.match(args.location)\n if res:\n log.debug('Using coordinates from CLI directly')\n position = (float(res.group(1)), float(res.group(2)), 0)\n else:\n log.debug('Looking up coordinates in API')\n position = util.get_pos_by_name(args.location)\n\n # Use the latitude and longitude to get the local altitude from Google\n try:\n url = 'https://maps.googleapis.com/maps/api/elevation/json?locations={},{}'.format(\n str(position[0]), str(position[1]))\n altitude = requests.get(url).json()[u'results'][0][u'elevation']\n log.debug('Local altitude is: %sm', altitude)\n position = (position[0], position[1], altitude)\n except requests.exceptions.RequestException:\n log.error('Unable to retrieve altitude from Google APIs; setting to 0')\n\n if not any(position):\n log.error('Could not get a position by name, aborting')\n sys.exit()\n\n log.info('Parsed location is: %.4f/%.4f/%.4f (lat/lng/alt)',\n position[0], position[1], position[2])\n\n if args.no_pokemon:\n log.info('Parsing of Pokemon disabled')\n if args.no_pokestops:\n log.info('Parsing of Pokestops disabled')\n if args.no_gyms:\n log.info('Parsing of Gyms disabled')\n\n config['LOCALE'] = args.locale\n config['CHINA'] = args.china\n\n app = Pogom(__name__)\n db = init_database(app)\n if args.clear_db:\n log.info('Clearing database')\n if args.db_type == 'mysql':\n drop_tables(db)\n elif os.path.isfile(args.db):\n os.remove(args.db)\n create_tables(db)\n\n app.set_current_location(position);\n\n # Control the search status (running or not) across threads\n pause_bit = Event()\n pause_bit.clear()\n\n # Setup the location tracking queue and push the first location on\n new_location_queue = Queue()\n new_location_queue.put(position)\n\n if not args.only_server:\n # Gather the pokemons!\n if not args.mock:\n log.debug('Starting a real search thread')\n search_thread = Thread(target=search_overseer_thread, args=(args, new_location_queue, pause_bit, encryption_lib_path))\n else:\n log.debug('Starting a fake search thread')\n insert_mock_data(position)\n search_thread = Thread(target=fake_search_loop)\n\n search_thread.daemon = True\n search_thread.name = 'search_thread'\n search_thread.start()\n\n if args.cors:\n CORS(app);\n\n app.set_search_control(pause_bit)\n app.set_location_queue(new_location_queue)\n\n config['ROOT_PATH'] = app.root_path\n config['GMAPS_KEY'] = args.gmaps_key\n\n if args.no_server:\n # This loop allows for ctrl-c interupts to work since flask won't be holding the program open\n while search_thread.is_alive():\n time.sleep(60)\n else:\n app.run(threaded=True, use_reloader=False, debug=args.debug, host=args.host, port=args.port)\n", "path": "runserver.py"}]}
| 2,686 | 142 |
gh_patches_debug_37737
|
rasdani/github-patches
|
git_diff
|
optuna__optuna-972
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Hyperband Usage Documentation
First many thanks for the new hyperband pruner.
I would like to request some more information on how to use it. I guess that it is necessary to tune the default parameters. Maybe you could provide some more information then just link to the papers. When should I set which parameter to what value?
The paper talks about an infinite horizon setting. Since I often do not set the number of iterations on optuna but stop it with KeyboardInterrupt is that relevant for me?
Many thanks
Philip
</issue>
<code>
[start of optuna/pruners/hyperband.py]
1 import math
2 from typing import List
3 from typing import Optional
4 import warnings
5
6 import optuna
7 from optuna._experimental import experimental
8 from optuna import logging
9 from optuna.pruners.base import BasePruner
10 from optuna.pruners.successive_halving import SuccessiveHalvingPruner
11 from optuna.trial import FrozenTrial
12
13 _logger = logging.get_logger(__name__)
14
15
16 @experimental("1.1.0")
17 class HyperbandPruner(BasePruner):
18 """Pruner using Hyperband.
19
20 As SuccessiveHalving (SHA) requires the number of configurations
21 :math:`n` as its hyperparameter. For a given finite budget :math:`B`,
22 all the configurations have the resources of :math:`B \\over n` on average.
23 As you can see, there will be a trade-off of :math:`B` and :math:`B \\over n`.
24 `Hyperband <http://www.jmlr.org/papers/volume18/16-558/16-558.pdf>`_ attacks this trade-off
25 by trying different :math:`n` values for a fixed budget.
26
27 .. note::
28 * In the Hyperband paper, the counterpart of :class:`~optuna.samplers.RandomSampler`
29 is used.
30 * Optuna uses :class:`~optuna.samplers.TPESampler` by default.
31 * `The benchmark result
32 <https://github.com/optuna/optuna/pull/828#issuecomment-575457360>`_
33 shows that :class:`optuna.pruners.HyperbandPruner` supports both samplers.
34
35 .. note::
36 If you use ``HyperbandPruner`` with :class:`~optuna.samplers.TPESampler`,
37 it's recommended to consider to set larger ``n_trials`` or ``timeout`` to make full use of
38 the characteristics of :class:`~optuna.samplers.TPESampler`
39 because :class:`~optuna.samplers.TPESampler` uses some (by default, :math:`10`)
40 :class:`~optuna.trial.Trial`\\ s for its startup.
41
42 As Hyperband runs multiple :class:`~optuna.pruners.SuccessiveHalvingPruner` and collect
43 trials based on the current :class:`~optuna.trial.Trial`\\ 's bracket ID, each bracket
44 needs to observe more than :math:`10` :class:`~optuna.trial.Trial`\\ s
45 for :class:`~optuna.samplers.TPESampler` to adapt its search space.
46
47 Thus, for example, if ``HyperbandPruner`` has :math:`4` pruners in it,
48 at least :math:`4 \\times 10` trials are consumed for startup.
49
50 .. note::
51 Hyperband has several :class:`~optuna.pruners.SuccessiveHalvingPruner`. Each
52 :class:`~optuna.pruners.SuccessiveHalvingPruner` is referred as "bracket" in the original
53 paper. The number of brackets is an important factor to control the early stopping behavior
54 of Hyperband and is automatically determined by ``max_resource`` and ``reduction_factor``
55 as `The number of brackets = floor(log(max_resource) / log(reduction_factor)) + 1`. Please
56 set ``reduction_factor`` so that the number of brackets is not too large (about 4 ~ 6 in
57 most use cases). Please see Section 3.6 of the `original paper
58 <http://www.jmlr.org/papers/volume18/16-558/16-558.pdf>`_ for the detail.
59
60 Args:
61 min_resource:
62 A parameter for specifying the minimum resource allocated to a trial noted as :math:`r`
63 in the paper.
64 See the details for :class:`~optuna.pruners.SuccessiveHalvingPruner`.
65 max_resource:
66 A parameter for specifying the maximum resource allocated to a trial noted as :math:`R`
67 in the paper. This value represents and should match the maximum iteration steps (e.g.,
68 the number of epochs for neural networks).
69 reduction_factor:
70 A parameter for specifying reduction factor of promotable trials noted as
71 :math:`\\eta` in the paper. See the details for
72 :class:`~optuna.pruners.SuccessiveHalvingPruner`.
73 n_brackets:
74
75 .. deprecated:: 1.4.0
76 This argument will be removed from :class:`~optuna.pruners.HyperbandPruner`. The
77 number of brackets are automatically determined based on ``max_resource`` and
78 ``reduction_factor``.
79
80 The number of :class:`~optuna.pruners.SuccessiveHalvingPruner`\\ s (brackets).
81 Defaults to :math:`4`.
82 min_early_stopping_rate_low:
83
84 .. deprecated:: 1.4.0
85 This argument will be removed from :class:`~optuna.pruners.HyperbandPruner`.
86
87 A parameter for specifying the minimum early-stopping rate.
88 This parameter is related to a parameter that is referred to as :math:`s` and used in
89 `Asynchronous SuccessiveHalving paper <http://arxiv.org/abs/1810.05934>`_.
90 The minimum early stopping rate for :math:`i` th bracket is :math:`i + s`.
91 """
92
93 def __init__(
94 self,
95 min_resource: int = 1,
96 max_resource: int = 80,
97 reduction_factor: int = 3,
98 n_brackets: Optional[int] = None,
99 min_early_stopping_rate_low: Optional[int] = None,
100 ) -> None:
101
102 self._pruners = [] # type: List[SuccessiveHalvingPruner]
103 self._reduction_factor = reduction_factor
104 self._total_trial_allocation_budget = 0
105
106 if n_brackets is None:
107 # In the original paper http://www.jmlr.org/papers/volume18/16-558/16-558.pdf, the
108 # inputs of Hyperband are `R`: max resource and `\eta`: reduction factor. The
109 # number of brackets (this is referred as `s_{max} + 1` in the paper) is calculated
110 # by s_{max} + 1 = \floor{\log_{\eta} (R)} + 1 in Algorithm 1 of the original paper.
111 self._n_brackets = (
112 math.floor(math.log2(max_resource) / math.log2(reduction_factor)) + 1
113 )
114 else:
115 message = (
116 "The argument of `n_brackets` is deprecated. "
117 "The number of brackets is automatically determined by `max_resource` and "
118 "`reduction_factor` as "
119 "`n_brackets = floor(log(max_resource) / log(reduction_factor)) + 1`. "
120 "Please specify `reduction_factor` appropriately."
121 )
122 warnings.warn(message, DeprecationWarning)
123 _logger.warning(message)
124 self._n_brackets = n_brackets
125
126 self._trial_allocation_budgets = [] # type: List[int]
127
128 _logger.debug("Hyperband has {} brackets".format(self._n_brackets))
129
130 for i in range(self._n_brackets):
131 trial_allocation_budget = self._calculate_trial_allocation_budget(i)
132 self._total_trial_allocation_budget += trial_allocation_budget
133 self._trial_allocation_budgets.append(trial_allocation_budget)
134
135 # N.B. (crcrpar): `min_early_stopping_rate` has the information of `bracket_index`.
136 if min_early_stopping_rate_low is None:
137 min_early_stopping_rate = i
138 else:
139 message = (
140 "The argument of `min_early_stopping_rate_low` is deprecated. "
141 "Please specify `min_resource` appropriately."
142 )
143 warnings.warn(message, DeprecationWarning)
144 _logger.warning(message)
145 min_early_stopping_rate = min_early_stopping_rate_low + i
146
147 _logger.debug(
148 "{}th bracket has minimum early stopping rate of {}".format(
149 i, min_early_stopping_rate
150 )
151 )
152
153 pruner = SuccessiveHalvingPruner(
154 min_resource=min_resource,
155 reduction_factor=reduction_factor,
156 min_early_stopping_rate=min_early_stopping_rate,
157 )
158 self._pruners.append(pruner)
159
160 def prune(self, study: "optuna.study.Study", trial: FrozenTrial) -> bool:
161 i = self._get_bracket_id(study, trial)
162 _logger.debug("{}th bracket is selected".format(i))
163 bracket_study = self._create_bracket_study(study, i)
164 return self._pruners[i].prune(bracket_study, trial)
165
166 def _calculate_trial_allocation_budget(self, pruner_index: int) -> int:
167 """Compute the trial allocated budget for a bracket of ``pruner_index``.
168
169 In the `original paper <http://www.jmlr.org/papers/volume18/16-558/16-558.pdf>`, the
170 number of trials per one bracket is referred as ``n`` in Algorithm 1. Since we do not know
171 the total number of trials in the leaning scheme of Optuna, we calculate the ratio of the
172 number of trials here instead.
173 """
174
175 s = self._n_brackets - 1 - pruner_index
176 return math.ceil(self._n_brackets * (self._reduction_factor ** s) / (s + 1))
177
178 def _get_bracket_id(self, study: "optuna.study.Study", trial: FrozenTrial) -> int:
179 """Compute the index of bracket for a trial of ``trial_number``.
180
181 The index of a bracket is noted as :math:`s` in
182 `Hyperband paper <http://www.jmlr.org/papers/volume18/16-558/16-558.pdf>`_.
183 """
184
185 n = (
186 hash("{}_{}".format(study.study_name, trial.number))
187 % self._total_trial_allocation_budget
188 )
189 for i in range(self._n_brackets):
190 n -= self._trial_allocation_budgets[i]
191 if n < 0:
192 return i
193
194 assert False, "This line should be unreachable."
195
196 def _create_bracket_study(
197 self, study: "optuna.study.Study", bracket_index: int
198 ) -> "optuna.study.Study":
199 # This class is assumed to be passed to
200 # `SuccessiveHalvingPruner.prune` in which `get_trials`,
201 # `direction`, and `storage` are used.
202 # But for safety, prohibit the other attributes explicitly.
203 class _BracketStudy(optuna.study.Study):
204
205 _VALID_ATTRS = (
206 "get_trials",
207 "direction",
208 "_storage",
209 "_study_id",
210 "pruner",
211 "study_name",
212 "_bracket_id",
213 "sampler",
214 )
215
216 def __init__(self, study: "optuna.study.Study", bracket_id: int) -> None:
217 super().__init__(
218 study_name=study.study_name,
219 storage=study._storage,
220 sampler=study.sampler,
221 pruner=study.pruner,
222 )
223 self._bracket_id = bracket_id
224
225 def get_trials(self, deepcopy: bool = True) -> List[FrozenTrial]:
226 trials = super().get_trials(deepcopy=deepcopy)
227 pruner = self.pruner
228 assert isinstance(pruner, HyperbandPruner)
229 return [t for t in trials if pruner._get_bracket_id(self, t) == self._bracket_id]
230
231 def __getattribute__(self, attr_name): # type: ignore
232 if attr_name not in _BracketStudy._VALID_ATTRS:
233 raise AttributeError(
234 "_BracketStudy does not have attribute of '{}'".format(attr_name)
235 )
236 else:
237 return object.__getattribute__(self, attr_name)
238
239 return _BracketStudy(study, bracket_index)
240
[end of optuna/pruners/hyperband.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/optuna/pruners/hyperband.py b/optuna/pruners/hyperband.py
--- a/optuna/pruners/hyperband.py
+++ b/optuna/pruners/hyperband.py
@@ -57,10 +57,58 @@
most use cases). Please see Section 3.6 of the `original paper
<http://www.jmlr.org/papers/volume18/16-558/16-558.pdf>`_ for the detail.
+ Example:
+
+ We minimize an objective function with Hyperband pruning algorithm.
+
+ .. testsetup::
+
+ import numpy as np
+ from sklearn.model_selection import train_test_split
+
+ np.random.seed(seed=0)
+ X = np.random.randn(200).reshape(-1, 1)
+ y = np.where(X[:, 0] < 0.5, 0, 1)
+ X_train, X_test, y_train, y_test = train_test_split(X, y, random_state=0)
+ classes = np.unique(y)
+
+ .. testcode::
+
+ import optuna
+ from sklearn.linear_model import SGDClassifier
+
+ n_train_iter = 100
+
+ def objective(trial):
+ alpha = trial.suggest_uniform('alpha', 0.0, 1.0)
+ clf = SGDClassifier(alpha=alpha)
+
+ for step in range(n_train_iter):
+ clf.partial_fit(X_train, y_train, classes=classes)
+
+ intermediate_value = clf.score(X_test, y_test)
+ trial.report(intermediate_value, step)
+
+ if trial.should_prune():
+ raise optuna.exceptions.TrialPruned()
+
+ return clf.score(X_test, y_test)
+
+ study = optuna.create_study(
+ direction='maximize',
+ pruner=optuna.pruners.HyperbandPruner(
+ min_resource=1,
+ max_resource=n_train_iter,
+ reduction_factor=3
+ )
+ )
+ study.optimize(objective, n_trials=20)
+
Args:
min_resource:
A parameter for specifying the minimum resource allocated to a trial noted as :math:`r`
- in the paper.
+ in the paper. A smaller :math:`r` will give a result faster, but a larger
+ :math:`r` will give a better guarantee of successful judging between configurations.
See the details for :class:`~optuna.pruners.SuccessiveHalvingPruner`.
max_resource:
A parameter for specifying the maximum resource allocated to a trial noted as :math:`R`
@@ -68,8 +116,8 @@
the number of epochs for neural networks).
reduction_factor:
A parameter for specifying reduction factor of promotable trials noted as
- :math:`\\eta` in the paper. See the details for
- :class:`~optuna.pruners.SuccessiveHalvingPruner`.
+ :math:`\\eta` in the paper.
+ See the details for :class:`~optuna.pruners.SuccessiveHalvingPruner`.
n_brackets:
.. deprecated:: 1.4.0
|
{"golden_diff": "diff --git a/optuna/pruners/hyperband.py b/optuna/pruners/hyperband.py\n--- a/optuna/pruners/hyperband.py\n+++ b/optuna/pruners/hyperband.py\n@@ -57,10 +57,58 @@\n most use cases).\u3000Please see Section 3.6 of the `original paper\n <http://www.jmlr.org/papers/volume18/16-558/16-558.pdf>`_ for the detail.\n \n+ Example:\n+\n+ We minimize an objective function with Hyperband pruning algorithm.\n+\n+ .. testsetup::\n+\n+ import numpy as np\n+ from sklearn.model_selection import train_test_split\n+\n+ np.random.seed(seed=0)\n+ X = np.random.randn(200).reshape(-1, 1)\n+ y = np.where(X[:, 0] < 0.5, 0, 1)\n+ X_train, X_test, y_train, y_test = train_test_split(X, y, random_state=0)\n+ classes = np.unique(y)\n+\n+ .. testcode::\n+\n+ import optuna\n+ from sklearn.linear_model import SGDClassifier\n+\n+ n_train_iter = 100\n+\n+ def objective(trial):\n+ alpha = trial.suggest_uniform('alpha', 0.0, 1.0)\n+ clf = SGDClassifier(alpha=alpha)\n+\n+ for step in range(n_train_iter):\n+ clf.partial_fit(X_train, y_train, classes=classes)\n+\n+ intermediate_value = clf.score(X_test, y_test)\n+ trial.report(intermediate_value, step)\n+\n+ if trial.should_prune():\n+ raise optuna.exceptions.TrialPruned()\n+\n+ return clf.score(X_test, y_test)\n+\n+ study = optuna.create_study(\n+ direction='maximize',\n+ pruner=optuna.pruners.HyperbandPruner(\n+ min_resource=1,\n+ max_resource=n_train_iter,\n+ reduction_factor=3\n+ )\n+ )\n+ study.optimize(objective, n_trials=20)\n+\n Args:\n min_resource:\n A parameter for specifying the minimum resource allocated to a trial noted as :math:`r`\n- in the paper.\n+ in the paper. A smaller :math:`r` will give a result faster, but a larger\n+ :math:`r` will give a better guarantee of successful judging between configurations.\n See the details for :class:`~optuna.pruners.SuccessiveHalvingPruner`.\n max_resource:\n A parameter for specifying the maximum resource allocated to a trial noted as :math:`R`\n@@ -68,8 +116,8 @@\n the number of epochs for neural networks).\n reduction_factor:\n A parameter for specifying reduction factor of promotable trials noted as\n- :math:`\\\\eta` in the paper. See the details for\n- :class:`~optuna.pruners.SuccessiveHalvingPruner`.\n+ :math:`\\\\eta` in the paper.\n+ See the details for :class:`~optuna.pruners.SuccessiveHalvingPruner`.\n n_brackets:\n \n .. deprecated:: 1.4.0\n", "issue": "Hyperband Usage Documentation\nFirst many thanks for the new hyperband pruner.\r\n\r\nI would like to request some more information on how to use it. I guess that it is necessary to tune the default parameters. Maybe you could provide some more information then just link to the papers. When should I set which parameter to what value?\r\n\r\nThe paper talks about an infinite horizon setting. Since I often do not set the number of iterations on optuna but stop it with KeyboardInterrupt is that relevant for me?\r\n\r\nMany thanks\r\nPhilip\n", "before_files": [{"content": "import math\nfrom typing import List\nfrom typing import Optional\nimport warnings\n\nimport optuna\nfrom optuna._experimental import experimental\nfrom optuna import logging\nfrom optuna.pruners.base import BasePruner\nfrom optuna.pruners.successive_halving import SuccessiveHalvingPruner\nfrom optuna.trial import FrozenTrial\n\n_logger = logging.get_logger(__name__)\n\n\n@experimental(\"1.1.0\")\nclass HyperbandPruner(BasePruner):\n \"\"\"Pruner using Hyperband.\n\n As SuccessiveHalving (SHA) requires the number of configurations\n :math:`n` as its hyperparameter. For a given finite budget :math:`B`,\n all the configurations have the resources of :math:`B \\\\over n` on average.\n As you can see, there will be a trade-off of :math:`B` and :math:`B \\\\over n`.\n `Hyperband <http://www.jmlr.org/papers/volume18/16-558/16-558.pdf>`_ attacks this trade-off\n by trying different :math:`n` values for a fixed budget.\n\n .. note::\n * In the Hyperband paper, the counterpart of :class:`~optuna.samplers.RandomSampler`\n is used.\n * Optuna uses :class:`~optuna.samplers.TPESampler` by default.\n * `The benchmark result\n <https://github.com/optuna/optuna/pull/828#issuecomment-575457360>`_\n shows that :class:`optuna.pruners.HyperbandPruner` supports both samplers.\n\n .. note::\n If you use ``HyperbandPruner`` with :class:`~optuna.samplers.TPESampler`,\n it's recommended to consider to set larger ``n_trials`` or ``timeout`` to make full use of\n the characteristics of :class:`~optuna.samplers.TPESampler`\n because :class:`~optuna.samplers.TPESampler` uses some (by default, :math:`10`)\n :class:`~optuna.trial.Trial`\\\\ s for its startup.\n\n As Hyperband runs multiple :class:`~optuna.pruners.SuccessiveHalvingPruner` and collect\n trials based on the current :class:`~optuna.trial.Trial`\\\\ 's bracket ID, each bracket\n needs to observe more than :math:`10` :class:`~optuna.trial.Trial`\\\\ s\n for :class:`~optuna.samplers.TPESampler` to adapt its search space.\n\n Thus, for example, if ``HyperbandPruner`` has :math:`4` pruners in it,\n at least :math:`4 \\\\times 10` trials are consumed for startup.\n\n .. note::\n Hyperband has several :class:`~optuna.pruners.SuccessiveHalvingPruner`. Each\n :class:`~optuna.pruners.SuccessiveHalvingPruner` is referred as \"bracket\" in the original\n paper. The number of brackets is an important factor to control the early stopping behavior\n of Hyperband and is automatically determined by ``max_resource`` and ``reduction_factor``\n as `The number of brackets = floor(log(max_resource) / log(reduction_factor)) + 1`. Please\n set ``reduction_factor`` so that the number of brackets is not too large\u3000(about 4 ~ 6 in\n most use cases).\u3000Please see Section 3.6 of the `original paper\n <http://www.jmlr.org/papers/volume18/16-558/16-558.pdf>`_ for the detail.\n\n Args:\n min_resource:\n A parameter for specifying the minimum resource allocated to a trial noted as :math:`r`\n in the paper.\n See the details for :class:`~optuna.pruners.SuccessiveHalvingPruner`.\n max_resource:\n A parameter for specifying the maximum resource allocated to a trial noted as :math:`R`\n in the paper. This value represents and should match the maximum iteration steps (e.g.,\n the number of epochs for neural networks).\n reduction_factor:\n A parameter for specifying reduction factor of promotable trials noted as\n :math:`\\\\eta` in the paper. See the details for\n :class:`~optuna.pruners.SuccessiveHalvingPruner`.\n n_brackets:\n\n .. deprecated:: 1.4.0\n This argument will be removed from :class:`~optuna.pruners.HyperbandPruner`. The\n number of brackets are automatically determined based on ``max_resource`` and\n ``reduction_factor``.\n\n The number of :class:`~optuna.pruners.SuccessiveHalvingPruner`\\\\ s (brackets).\n Defaults to :math:`4`.\n min_early_stopping_rate_low:\n\n .. deprecated:: 1.4.0\n This argument will be removed from :class:`~optuna.pruners.HyperbandPruner`.\n\n A parameter for specifying the minimum early-stopping rate.\n This parameter is related to a parameter that is referred to as :math:`s` and used in\n `Asynchronous SuccessiveHalving paper <http://arxiv.org/abs/1810.05934>`_.\n The minimum early stopping rate for :math:`i` th bracket is :math:`i + s`.\n \"\"\"\n\n def __init__(\n self,\n min_resource: int = 1,\n max_resource: int = 80,\n reduction_factor: int = 3,\n n_brackets: Optional[int] = None,\n min_early_stopping_rate_low: Optional[int] = None,\n ) -> None:\n\n self._pruners = [] # type: List[SuccessiveHalvingPruner]\n self._reduction_factor = reduction_factor\n self._total_trial_allocation_budget = 0\n\n if n_brackets is None:\n # In the original paper http://www.jmlr.org/papers/volume18/16-558/16-558.pdf, the\n # inputs of Hyperband are `R`: max resource and `\\eta`: reduction factor. The\n # number of brackets (this is referred as `s_{max} + 1` in the paper) is calculated\n # by s_{max} + 1 = \\floor{\\log_{\\eta} (R)} + 1 in Algorithm 1 of the original paper.\n self._n_brackets = (\n math.floor(math.log2(max_resource) / math.log2(reduction_factor)) + 1\n )\n else:\n message = (\n \"The argument of `n_brackets` is deprecated. \"\n \"The number of brackets is automatically determined by `max_resource` and \"\n \"`reduction_factor` as \"\n \"`n_brackets = floor(log(max_resource) / log(reduction_factor)) + 1`. \"\n \"Please specify `reduction_factor` appropriately.\"\n )\n warnings.warn(message, DeprecationWarning)\n _logger.warning(message)\n self._n_brackets = n_brackets\n\n self._trial_allocation_budgets = [] # type: List[int]\n\n _logger.debug(\"Hyperband has {} brackets\".format(self._n_brackets))\n\n for i in range(self._n_brackets):\n trial_allocation_budget = self._calculate_trial_allocation_budget(i)\n self._total_trial_allocation_budget += trial_allocation_budget\n self._trial_allocation_budgets.append(trial_allocation_budget)\n\n # N.B. (crcrpar): `min_early_stopping_rate` has the information of `bracket_index`.\n if min_early_stopping_rate_low is None:\n min_early_stopping_rate = i\n else:\n message = (\n \"The argument of `min_early_stopping_rate_low` is deprecated. \"\n \"Please specify `min_resource` appropriately.\"\n )\n warnings.warn(message, DeprecationWarning)\n _logger.warning(message)\n min_early_stopping_rate = min_early_stopping_rate_low + i\n\n _logger.debug(\n \"{}th bracket has minimum early stopping rate of {}\".format(\n i, min_early_stopping_rate\n )\n )\n\n pruner = SuccessiveHalvingPruner(\n min_resource=min_resource,\n reduction_factor=reduction_factor,\n min_early_stopping_rate=min_early_stopping_rate,\n )\n self._pruners.append(pruner)\n\n def prune(self, study: \"optuna.study.Study\", trial: FrozenTrial) -> bool:\n i = self._get_bracket_id(study, trial)\n _logger.debug(\"{}th bracket is selected\".format(i))\n bracket_study = self._create_bracket_study(study, i)\n return self._pruners[i].prune(bracket_study, trial)\n\n def _calculate_trial_allocation_budget(self, pruner_index: int) -> int:\n \"\"\"Compute the trial allocated budget for a bracket of ``pruner_index``.\n\n In the `original paper <http://www.jmlr.org/papers/volume18/16-558/16-558.pdf>`, the\n number of trials per one bracket is referred as ``n`` in Algorithm 1. Since we do not know\n the total number of trials in the leaning scheme of Optuna, we calculate the ratio of the\n number of trials here instead.\n \"\"\"\n\n s = self._n_brackets - 1 - pruner_index\n return math.ceil(self._n_brackets * (self._reduction_factor ** s) / (s + 1))\n\n def _get_bracket_id(self, study: \"optuna.study.Study\", trial: FrozenTrial) -> int:\n \"\"\"Compute the index of bracket for a trial of ``trial_number``.\n\n The index of a bracket is noted as :math:`s` in\n `Hyperband paper <http://www.jmlr.org/papers/volume18/16-558/16-558.pdf>`_.\n \"\"\"\n\n n = (\n hash(\"{}_{}\".format(study.study_name, trial.number))\n % self._total_trial_allocation_budget\n )\n for i in range(self._n_brackets):\n n -= self._trial_allocation_budgets[i]\n if n < 0:\n return i\n\n assert False, \"This line should be unreachable.\"\n\n def _create_bracket_study(\n self, study: \"optuna.study.Study\", bracket_index: int\n ) -> \"optuna.study.Study\":\n # This class is assumed to be passed to\n # `SuccessiveHalvingPruner.prune` in which `get_trials`,\n # `direction`, and `storage` are used.\n # But for safety, prohibit the other attributes explicitly.\n class _BracketStudy(optuna.study.Study):\n\n _VALID_ATTRS = (\n \"get_trials\",\n \"direction\",\n \"_storage\",\n \"_study_id\",\n \"pruner\",\n \"study_name\",\n \"_bracket_id\",\n \"sampler\",\n )\n\n def __init__(self, study: \"optuna.study.Study\", bracket_id: int) -> None:\n super().__init__(\n study_name=study.study_name,\n storage=study._storage,\n sampler=study.sampler,\n pruner=study.pruner,\n )\n self._bracket_id = bracket_id\n\n def get_trials(self, deepcopy: bool = True) -> List[FrozenTrial]:\n trials = super().get_trials(deepcopy=deepcopy)\n pruner = self.pruner\n assert isinstance(pruner, HyperbandPruner)\n return [t for t in trials if pruner._get_bracket_id(self, t) == self._bracket_id]\n\n def __getattribute__(self, attr_name): # type: ignore\n if attr_name not in _BracketStudy._VALID_ATTRS:\n raise AttributeError(\n \"_BracketStudy does not have attribute of '{}'\".format(attr_name)\n )\n else:\n return object.__getattribute__(self, attr_name)\n\n return _BracketStudy(study, bracket_index)\n", "path": "optuna/pruners/hyperband.py"}]}
| 3,929 | 720 |
gh_patches_debug_12533
|
rasdani/github-patches
|
git_diff
|
getnikola__nikola-2108
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
output/assets/css/code.css is orphaned?
```
~/blog$ nikola build
Scanning posts................done!
copy_assets:output/assets/css/base.css
Scanning posts................done!
~/blog$
~/blog$ nikola build
Scanning posts................done!
~/blog$ nikola check -f
Scanning posts................done!
WARNING: check: Files from unknown origins (orphans):
WARNING: check: output/assets/css/code.css
~/blog$ nikola build
Scanning posts................done!
copy_assets:output/assets/css/base.css
~/blog$ nikola check -f
Scanning posts................done!
WARNING: check: Files from unknown origins (orphans):
WARNING: check: output/assets/css/code.css
```
</issue>
<code>
[start of nikola/plugins/task/copy_assets.py]
1 # -*- coding: utf-8 -*-
2
3 # Copyright © 2012-2015 Roberto Alsina and others.
4
5 # Permission is hereby granted, free of charge, to any
6 # person obtaining a copy of this software and associated
7 # documentation files (the "Software"), to deal in the
8 # Software without restriction, including without limitation
9 # the rights to use, copy, modify, merge, publish,
10 # distribute, sublicense, and/or sell copies of the
11 # Software, and to permit persons to whom the Software is
12 # furnished to do so, subject to the following conditions:
13 #
14 # The above copyright notice and this permission notice
15 # shall be included in all copies or substantial portions of
16 # the Software.
17 #
18 # THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY
19 # KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE
20 # WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR
21 # PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS
22 # OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR
23 # OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR
24 # OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE
25 # SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
26
27 """Copy theme assets into output."""
28
29 from __future__ import unicode_literals
30
31 import io
32 import os
33
34 from nikola.plugin_categories import Task
35 from nikola import utils
36
37
38 class CopyAssets(Task):
39
40 """Copy theme assets into output."""
41
42 name = "copy_assets"
43
44 def gen_tasks(self):
45 """Create tasks to copy the assets of the whole theme chain.
46
47 If a file is present on two themes, use the version
48 from the "youngest" theme.
49 """
50 kw = {
51 "themes": self.site.THEMES,
52 "files_folders": self.site.config['FILES_FOLDERS'],
53 "output_folder": self.site.config['OUTPUT_FOLDER'],
54 "filters": self.site.config['FILTERS'],
55 "code_color_scheme": self.site.config['CODE_COLOR_SCHEME'],
56 "code.css_selectors": 'pre.code',
57 "code.css_head": '/* code.css file generated by Nikola */\n',
58 "code.css_close": "\ntable.codetable { width: 100%;} td.linenos {text-align: right; width: 4em;}\n",
59 }
60 tasks = {}
61 code_css_path = os.path.join(kw['output_folder'], 'assets', 'css', 'code.css')
62 code_css_input = utils.get_asset_path('assets/css/code.css',
63 themes=kw['themes'],
64 files_folders=kw['files_folders'])
65
66 kw["code.css_input"] = code_css_input
67
68 yield self.group_task()
69
70 for theme_name in kw['themes']:
71 src = os.path.join(utils.get_theme_path(theme_name), 'assets')
72 dst = os.path.join(kw['output_folder'], 'assets')
73 for task in utils.copy_tree(src, dst):
74 if task['name'] in tasks:
75 continue
76 tasks[task['name']] = task
77 task['uptodate'] = [utils.config_changed(kw, 'nikola.plugins.task.copy_assets')]
78 task['basename'] = self.name
79 if code_css_input:
80 if 'file_dep' not in task:
81 task['file_dep'] = []
82 task['file_dep'].append(code_css_input)
83 yield utils.apply_filters(task, kw['filters'])
84
85 # Check whether or not there is a code.css file around.
86 if not code_css_input:
87 def create_code_css():
88 from pygments.formatters import get_formatter_by_name
89 formatter = get_formatter_by_name('html', style=kw["code_color_scheme"])
90 utils.makedirs(os.path.dirname(code_css_path))
91 with io.open(code_css_path, 'w+', encoding='utf8') as outf:
92 outf.write(kw["code.css_head"])
93 outf.write(formatter.get_style_defs(kw["code.css_selectors"]))
94 outf.write(kw["code.css_close"])
95
96 if os.path.exists(code_css_path):
97 with io.open(code_css_path, 'r', encoding='utf-8') as fh:
98 testcontents = fh.read(len(kw["code.css_head"])) == kw["code.css_head"]
99 else:
100 testcontents = False
101
102 task = {
103 'basename': self.name,
104 'name': code_css_path,
105 'targets': [code_css_path],
106 'uptodate': [utils.config_changed(kw, 'nikola.plugins.task.copy_assets'), testcontents],
107 'actions': [(create_code_css, [])],
108 'clean': True,
109 }
110 yield utils.apply_filters(task, kw['filters'])
111
[end of nikola/plugins/task/copy_assets.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/nikola/plugins/task/copy_assets.py b/nikola/plugins/task/copy_assets.py
--- a/nikola/plugins/task/copy_assets.py
+++ b/nikola/plugins/task/copy_assets.py
@@ -61,10 +61,7 @@
code_css_path = os.path.join(kw['output_folder'], 'assets', 'css', 'code.css')
code_css_input = utils.get_asset_path('assets/css/code.css',
themes=kw['themes'],
- files_folders=kw['files_folders'])
-
- kw["code.css_input"] = code_css_input
-
+ files_folders=kw['files_folders'], output_dir=None)
yield self.group_task()
for theme_name in kw['themes']:
|
{"golden_diff": "diff --git a/nikola/plugins/task/copy_assets.py b/nikola/plugins/task/copy_assets.py\n--- a/nikola/plugins/task/copy_assets.py\n+++ b/nikola/plugins/task/copy_assets.py\n@@ -61,10 +61,7 @@\n code_css_path = os.path.join(kw['output_folder'], 'assets', 'css', 'code.css')\n code_css_input = utils.get_asset_path('assets/css/code.css',\n themes=kw['themes'],\n- files_folders=kw['files_folders'])\n-\n- kw[\"code.css_input\"] = code_css_input\n-\n+ files_folders=kw['files_folders'], output_dir=None)\n yield self.group_task()\n \n for theme_name in kw['themes']:\n", "issue": "output/assets/css/code.css is orphaned?\n```\n~/blog$ nikola build\nScanning posts................done!\ncopy_assets:output/assets/css/base.css\nScanning posts................done!\n~/blog$ \n~/blog$ nikola build\nScanning posts................done!\n~/blog$ nikola check -f\nScanning posts................done!\nWARNING: check: Files from unknown origins (orphans):\nWARNING: check: output/assets/css/code.css\n~/blog$ nikola build\nScanning posts................done!\ncopy_assets:output/assets/css/base.css\n~/blog$ nikola check -f\nScanning posts................done!\nWARNING: check: Files from unknown origins (orphans):\nWARNING: check: output/assets/css/code.css\n```\n\n", "before_files": [{"content": "# -*- coding: utf-8 -*-\n\n# Copyright \u00a9 2012-2015 Roberto Alsina and others.\n\n# Permission is hereby granted, free of charge, to any\n# person obtaining a copy of this software and associated\n# documentation files (the \"Software\"), to deal in the\n# Software without restriction, including without limitation\n# the rights to use, copy, modify, merge, publish,\n# distribute, sublicense, and/or sell copies of the\n# Software, and to permit persons to whom the Software is\n# furnished to do so, subject to the following conditions:\n#\n# The above copyright notice and this permission notice\n# shall be included in all copies or substantial portions of\n# the Software.\n#\n# THE SOFTWARE IS PROVIDED \"AS IS\", WITHOUT WARRANTY OF ANY\n# KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE\n# WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR\n# PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS\n# OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR\n# OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR\n# OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE\n# SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.\n\n\"\"\"Copy theme assets into output.\"\"\"\n\nfrom __future__ import unicode_literals\n\nimport io\nimport os\n\nfrom nikola.plugin_categories import Task\nfrom nikola import utils\n\n\nclass CopyAssets(Task):\n\n \"\"\"Copy theme assets into output.\"\"\"\n\n name = \"copy_assets\"\n\n def gen_tasks(self):\n \"\"\"Create tasks to copy the assets of the whole theme chain.\n\n If a file is present on two themes, use the version\n from the \"youngest\" theme.\n \"\"\"\n kw = {\n \"themes\": self.site.THEMES,\n \"files_folders\": self.site.config['FILES_FOLDERS'],\n \"output_folder\": self.site.config['OUTPUT_FOLDER'],\n \"filters\": self.site.config['FILTERS'],\n \"code_color_scheme\": self.site.config['CODE_COLOR_SCHEME'],\n \"code.css_selectors\": 'pre.code',\n \"code.css_head\": '/* code.css file generated by Nikola */\\n',\n \"code.css_close\": \"\\ntable.codetable { width: 100%;} td.linenos {text-align: right; width: 4em;}\\n\",\n }\n tasks = {}\n code_css_path = os.path.join(kw['output_folder'], 'assets', 'css', 'code.css')\n code_css_input = utils.get_asset_path('assets/css/code.css',\n themes=kw['themes'],\n files_folders=kw['files_folders'])\n\n kw[\"code.css_input\"] = code_css_input\n\n yield self.group_task()\n\n for theme_name in kw['themes']:\n src = os.path.join(utils.get_theme_path(theme_name), 'assets')\n dst = os.path.join(kw['output_folder'], 'assets')\n for task in utils.copy_tree(src, dst):\n if task['name'] in tasks:\n continue\n tasks[task['name']] = task\n task['uptodate'] = [utils.config_changed(kw, 'nikola.plugins.task.copy_assets')]\n task['basename'] = self.name\n if code_css_input:\n if 'file_dep' not in task:\n task['file_dep'] = []\n task['file_dep'].append(code_css_input)\n yield utils.apply_filters(task, kw['filters'])\n\n # Check whether or not there is a code.css file around.\n if not code_css_input:\n def create_code_css():\n from pygments.formatters import get_formatter_by_name\n formatter = get_formatter_by_name('html', style=kw[\"code_color_scheme\"])\n utils.makedirs(os.path.dirname(code_css_path))\n with io.open(code_css_path, 'w+', encoding='utf8') as outf:\n outf.write(kw[\"code.css_head\"])\n outf.write(formatter.get_style_defs(kw[\"code.css_selectors\"]))\n outf.write(kw[\"code.css_close\"])\n\n if os.path.exists(code_css_path):\n with io.open(code_css_path, 'r', encoding='utf-8') as fh:\n testcontents = fh.read(len(kw[\"code.css_head\"])) == kw[\"code.css_head\"]\n else:\n testcontents = False\n\n task = {\n 'basename': self.name,\n 'name': code_css_path,\n 'targets': [code_css_path],\n 'uptodate': [utils.config_changed(kw, 'nikola.plugins.task.copy_assets'), testcontents],\n 'actions': [(create_code_css, [])],\n 'clean': True,\n }\n yield utils.apply_filters(task, kw['filters'])\n", "path": "nikola/plugins/task/copy_assets.py"}]}
| 1,917 | 162 |
gh_patches_debug_12867
|
rasdani/github-patches
|
git_diff
|
great-expectations__great_expectations-6650
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Use cleaner solution for non-truncating division in python 2
Prefer `from __future__ import division` to `1.*x/y`
</issue>
<code>
[start of contrib/experimental/great_expectations_experimental/expectations/expect_queried_slowly_changing_table_to_have_no_gaps.py]
1 from datetime import datetime, timedelta
2 from typing import Optional, Union
3
4 from great_expectations.core.expectation_configuration import ExpectationConfiguration
5 from great_expectations.core.util import convert_to_json_serializable
6 from great_expectations.exceptions.exceptions import (
7 InvalidExpectationConfigurationError,
8 )
9 from great_expectations.execution_engine import ExecutionEngine
10 from great_expectations.expectations.expectation import (
11 ExpectationValidationResult,
12 QueryExpectation,
13 )
14
15
16 class ExpectQueriedSlowlyChangingTableToHaveNoGaps(QueryExpectation):
17 """Expect Slowly changing table type II to have no gaps between the 'end date' of each row, and the next 'start date' in the next row.
18
19 Args:
20 template_dict: dict with the following keys: \
21 primary_key (primary key column name or multiple columns, comma separated), \
22 open_date_column (name of the column representing open date), \
23 close_date_column (name of the column representing clode date)
24 threshold: an optional parameter - default is zero. \
25 if the ratio of "gaps" to total table rows is higher than threshold - error will be raised.
26 """
27
28 metric_dependencies = ("query.template_values",)
29
30 query = """
31 SELECT SUM(CASE WHEN {close_date_column} != COALESCE(next_start_date, {close_date_column}) THEN 1 ELSE 0 END),
32 COUNT(1)
33 FROM(SELECT {primary_key}, {close_date_column}, LEAD({open_date_column}) OVER(PARTITION BY {primary_key} ORDER BY
34 {open_date_column}) AS next_start_date
35 FROM {active_batch})
36 """
37
38 success_keys = (
39 "template_dict",
40 "threshold",
41 "query",
42 )
43
44 domain_keys = (
45 "template_dict",
46 "query",
47 "batch_id",
48 "row_condition",
49 "condition_parser",
50 )
51
52 default_kwarg_values = {
53 "result_format": "BASIC",
54 "include_config": True,
55 "catch_exceptions": False,
56 "meta": None,
57 "threshold": 0,
58 "query": query,
59 }
60
61 library_metadata = {"tags": ["query-based"], "contributors": ["@itaise"]}
62
63 def _validate(
64 self,
65 configuration: ExpectationConfiguration,
66 metrics: dict,
67 runtime_configuration: dict = None,
68 execution_engine: ExecutionEngine = None,
69 ) -> Union[ExpectationValidationResult, dict]:
70 threshold = configuration["kwargs"].get("threshold")
71 if not threshold:
72 threshold = self.default_kwarg_values["threshold"]
73
74 metrics = convert_to_json_serializable(data=metrics)
75 holes_count, total_count = list(
76 metrics.get("query.template_values")[0].values()
77 )[0]
78 error_rate = holes_count / total_count
79
80 return {
81 "success": error_rate <= threshold,
82 "result": {
83 "threshold": threshold,
84 "holes_count": holes_count,
85 "total_count": total_count,
86 },
87 }
88
89 today = datetime(year=2022, month=8, day=10)
90 one_day_ago = today - timedelta(days=1)
91 two_day_ago = today - timedelta(days=2)
92 three_day_ago = today - timedelta(days=3)
93 four_day_ago = today - timedelta(days=4)
94 five_day_ago = today - timedelta(days=5)
95 six_day_ago = today - timedelta(days=6)
96 seven_day_ago = today - timedelta(days=7)
97 eight_day_ago = today - timedelta(days=8)
98 nine_day_ago = today - timedelta(days=9)
99 ten_day_ago = today - timedelta(days=10)
100
101 examples = [
102 {
103 "data": [
104 {
105 "dataset_name": "test",
106 "data": {
107 "msid": [
108 "aaa",
109 "aaa",
110 "aaa",
111 "aaa",
112 "aaa",
113 "aaa",
114 "aaa",
115 "aaa",
116 "aaa",
117 ],
118 "uuid": [
119 "aaa",
120 "aaa",
121 "aaa",
122 "aaa",
123 "aaa",
124 "aaa",
125 "aaa",
126 "aaa",
127 "aaa",
128 ],
129 "col1": [1, 2, 2, 3, 4, 5, 6, 7, 8],
130 "col2": ["a", "a", "b", "b", "a", "a", "a", "a", "a"],
131 "start_date": [
132 ten_day_ago,
133 nine_day_ago,
134 eight_day_ago,
135 seven_day_ago,
136 six_day_ago,
137 five_day_ago,
138 four_day_ago,
139 three_day_ago,
140 two_day_ago,
141 ],
142 "end_date": [
143 nine_day_ago,
144 eight_day_ago,
145 seven_day_ago,
146 six_day_ago,
147 five_day_ago,
148 four_day_ago,
149 three_day_ago,
150 two_day_ago,
151 one_day_ago,
152 ],
153 "start_date_2": [
154 ten_day_ago,
155 seven_day_ago,
156 six_day_ago,
157 five_day_ago,
158 four_day_ago,
159 three_day_ago,
160 two_day_ago,
161 two_day_ago,
162 two_day_ago,
163 ],
164 "end_date_2": [
165 nine_day_ago,
166 six_day_ago,
167 six_day_ago,
168 five_day_ago,
169 four_day_ago,
170 three_day_ago,
171 two_day_ago,
172 two_day_ago,
173 two_day_ago,
174 ],
175 },
176 },
177 ],
178 "tests": [
179 {
180 "title": "basic_positive_test",
181 "exact_match_out": False,
182 "include_in_gallery": True,
183 "in": {
184 "template_dict": {
185 "primary_key": "msid,uuid",
186 "open_date_column": "start_date",
187 "close_date_column": "end_date",
188 }
189 },
190 "out": {"success": True},
191 "only_for": ["sqlite"],
192 },
193 {
194 "title": "basic_negative_test",
195 "exact_match_out": False,
196 "include_in_gallery": True,
197 "in": {
198 "template_dict": {
199 "primary_key": "msid,uuid",
200 "open_date_column": "start_date_2",
201 "close_date_column": "end_date_2",
202 },
203 "threshold": 0.1,
204 },
205 "out": {"success": False},
206 "only_for": ["sqlite"],
207 },
208 ],
209 "test_backends": [
210 {
211 "backend": "sqlalchemy",
212 "dialects": ["sqlite"],
213 }
214 ],
215 },
216 ]
217
218 def validate_configuration(
219 self, configuration: Optional[ExpectationConfiguration]
220 ) -> None:
221 super().validate_configuration(configuration)
222 threshold = configuration["kwargs"].get("threshold")
223 if not threshold:
224 threshold = self.default_kwarg_values["threshold"]
225
226 try:
227 assert isinstance(threshold, int) or isinstance(threshold, float)
228 assert threshold >= 0
229 assert threshold <= 1
230
231 except AssertionError as e:
232 raise InvalidExpectationConfigurationError(str(e))
233
234
235 if __name__ == "__main__":
236 ExpectQueriedSlowlyChangingTableToHaveNoGaps().print_diagnostic_checklist()
237
[end of contrib/experimental/great_expectations_experimental/expectations/expect_queried_slowly_changing_table_to_have_no_gaps.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/contrib/experimental/great_expectations_experimental/expectations/expect_queried_slowly_changing_table_to_have_no_gaps.py b/contrib/experimental/great_expectations_experimental/expectations/expect_queried_slowly_changing_table_to_have_no_gaps.py
--- a/contrib/experimental/great_expectations_experimental/expectations/expect_queried_slowly_changing_table_to_have_no_gaps.py
+++ b/contrib/experimental/great_expectations_experimental/expectations/expect_queried_slowly_changing_table_to_have_no_gaps.py
@@ -72,10 +72,12 @@
threshold = self.default_kwarg_values["threshold"]
metrics = convert_to_json_serializable(data=metrics)
+ holes_count: int
+ total_count: int
holes_count, total_count = list(
metrics.get("query.template_values")[0].values()
- )[0]
- error_rate = holes_count / total_count
+ )
+ error_rate = float(holes_count) / total_count
return {
"success": error_rate <= threshold,
|
{"golden_diff": "diff --git a/contrib/experimental/great_expectations_experimental/expectations/expect_queried_slowly_changing_table_to_have_no_gaps.py b/contrib/experimental/great_expectations_experimental/expectations/expect_queried_slowly_changing_table_to_have_no_gaps.py\n--- a/contrib/experimental/great_expectations_experimental/expectations/expect_queried_slowly_changing_table_to_have_no_gaps.py\n+++ b/contrib/experimental/great_expectations_experimental/expectations/expect_queried_slowly_changing_table_to_have_no_gaps.py\n@@ -72,10 +72,12 @@\n threshold = self.default_kwarg_values[\"threshold\"]\n \n metrics = convert_to_json_serializable(data=metrics)\n+ holes_count: int\n+ total_count: int\n holes_count, total_count = list(\n metrics.get(\"query.template_values\")[0].values()\n- )[0]\n- error_rate = holes_count / total_count\n+ )\n+ error_rate = float(holes_count) / total_count\n \n return {\n \"success\": error_rate <= threshold,\n", "issue": "Use cleaner solution for non-truncating division in python 2\nPrefer `from __future__ import division` to `1.*x/y`\n", "before_files": [{"content": "from datetime import datetime, timedelta\nfrom typing import Optional, Union\n\nfrom great_expectations.core.expectation_configuration import ExpectationConfiguration\nfrom great_expectations.core.util import convert_to_json_serializable\nfrom great_expectations.exceptions.exceptions import (\n InvalidExpectationConfigurationError,\n)\nfrom great_expectations.execution_engine import ExecutionEngine\nfrom great_expectations.expectations.expectation import (\n ExpectationValidationResult,\n QueryExpectation,\n)\n\n\nclass ExpectQueriedSlowlyChangingTableToHaveNoGaps(QueryExpectation):\n \"\"\"Expect Slowly changing table type II to have no gaps between the 'end date' of each row, and the next 'start date' in the next row.\n\n Args:\n template_dict: dict with the following keys: \\\n primary_key (primary key column name or multiple columns, comma separated), \\\n open_date_column (name of the column representing open date), \\\n close_date_column (name of the column representing clode date)\n threshold: an optional parameter - default is zero. \\\n if the ratio of \"gaps\" to total table rows is higher than threshold - error will be raised.\n \"\"\"\n\n metric_dependencies = (\"query.template_values\",)\n\n query = \"\"\"\n SELECT SUM(CASE WHEN {close_date_column} != COALESCE(next_start_date, {close_date_column}) THEN 1 ELSE 0 END),\n COUNT(1)\n FROM(SELECT {primary_key}, {close_date_column}, LEAD({open_date_column}) OVER(PARTITION BY {primary_key} ORDER BY\n {open_date_column}) AS next_start_date\n FROM {active_batch})\n \"\"\"\n\n success_keys = (\n \"template_dict\",\n \"threshold\",\n \"query\",\n )\n\n domain_keys = (\n \"template_dict\",\n \"query\",\n \"batch_id\",\n \"row_condition\",\n \"condition_parser\",\n )\n\n default_kwarg_values = {\n \"result_format\": \"BASIC\",\n \"include_config\": True,\n \"catch_exceptions\": False,\n \"meta\": None,\n \"threshold\": 0,\n \"query\": query,\n }\n\n library_metadata = {\"tags\": [\"query-based\"], \"contributors\": [\"@itaise\"]}\n\n def _validate(\n self,\n configuration: ExpectationConfiguration,\n metrics: dict,\n runtime_configuration: dict = None,\n execution_engine: ExecutionEngine = None,\n ) -> Union[ExpectationValidationResult, dict]:\n threshold = configuration[\"kwargs\"].get(\"threshold\")\n if not threshold:\n threshold = self.default_kwarg_values[\"threshold\"]\n\n metrics = convert_to_json_serializable(data=metrics)\n holes_count, total_count = list(\n metrics.get(\"query.template_values\")[0].values()\n )[0]\n error_rate = holes_count / total_count\n\n return {\n \"success\": error_rate <= threshold,\n \"result\": {\n \"threshold\": threshold,\n \"holes_count\": holes_count,\n \"total_count\": total_count,\n },\n }\n\n today = datetime(year=2022, month=8, day=10)\n one_day_ago = today - timedelta(days=1)\n two_day_ago = today - timedelta(days=2)\n three_day_ago = today - timedelta(days=3)\n four_day_ago = today - timedelta(days=4)\n five_day_ago = today - timedelta(days=5)\n six_day_ago = today - timedelta(days=6)\n seven_day_ago = today - timedelta(days=7)\n eight_day_ago = today - timedelta(days=8)\n nine_day_ago = today - timedelta(days=9)\n ten_day_ago = today - timedelta(days=10)\n\n examples = [\n {\n \"data\": [\n {\n \"dataset_name\": \"test\",\n \"data\": {\n \"msid\": [\n \"aaa\",\n \"aaa\",\n \"aaa\",\n \"aaa\",\n \"aaa\",\n \"aaa\",\n \"aaa\",\n \"aaa\",\n \"aaa\",\n ],\n \"uuid\": [\n \"aaa\",\n \"aaa\",\n \"aaa\",\n \"aaa\",\n \"aaa\",\n \"aaa\",\n \"aaa\",\n \"aaa\",\n \"aaa\",\n ],\n \"col1\": [1, 2, 2, 3, 4, 5, 6, 7, 8],\n \"col2\": [\"a\", \"a\", \"b\", \"b\", \"a\", \"a\", \"a\", \"a\", \"a\"],\n \"start_date\": [\n ten_day_ago,\n nine_day_ago,\n eight_day_ago,\n seven_day_ago,\n six_day_ago,\n five_day_ago,\n four_day_ago,\n three_day_ago,\n two_day_ago,\n ],\n \"end_date\": [\n nine_day_ago,\n eight_day_ago,\n seven_day_ago,\n six_day_ago,\n five_day_ago,\n four_day_ago,\n three_day_ago,\n two_day_ago,\n one_day_ago,\n ],\n \"start_date_2\": [\n ten_day_ago,\n seven_day_ago,\n six_day_ago,\n five_day_ago,\n four_day_ago,\n three_day_ago,\n two_day_ago,\n two_day_ago,\n two_day_ago,\n ],\n \"end_date_2\": [\n nine_day_ago,\n six_day_ago,\n six_day_ago,\n five_day_ago,\n four_day_ago,\n three_day_ago,\n two_day_ago,\n two_day_ago,\n two_day_ago,\n ],\n },\n },\n ],\n \"tests\": [\n {\n \"title\": \"basic_positive_test\",\n \"exact_match_out\": False,\n \"include_in_gallery\": True,\n \"in\": {\n \"template_dict\": {\n \"primary_key\": \"msid,uuid\",\n \"open_date_column\": \"start_date\",\n \"close_date_column\": \"end_date\",\n }\n },\n \"out\": {\"success\": True},\n \"only_for\": [\"sqlite\"],\n },\n {\n \"title\": \"basic_negative_test\",\n \"exact_match_out\": False,\n \"include_in_gallery\": True,\n \"in\": {\n \"template_dict\": {\n \"primary_key\": \"msid,uuid\",\n \"open_date_column\": \"start_date_2\",\n \"close_date_column\": \"end_date_2\",\n },\n \"threshold\": 0.1,\n },\n \"out\": {\"success\": False},\n \"only_for\": [\"sqlite\"],\n },\n ],\n \"test_backends\": [\n {\n \"backend\": \"sqlalchemy\",\n \"dialects\": [\"sqlite\"],\n }\n ],\n },\n ]\n\n def validate_configuration(\n self, configuration: Optional[ExpectationConfiguration]\n ) -> None:\n super().validate_configuration(configuration)\n threshold = configuration[\"kwargs\"].get(\"threshold\")\n if not threshold:\n threshold = self.default_kwarg_values[\"threshold\"]\n\n try:\n assert isinstance(threshold, int) or isinstance(threshold, float)\n assert threshold >= 0\n assert threshold <= 1\n\n except AssertionError as e:\n raise InvalidExpectationConfigurationError(str(e))\n\n\nif __name__ == \"__main__\":\n ExpectQueriedSlowlyChangingTableToHaveNoGaps().print_diagnostic_checklist()\n", "path": "contrib/experimental/great_expectations_experimental/expectations/expect_queried_slowly_changing_table_to_have_no_gaps.py"}]}
| 2,815 | 246 |
gh_patches_debug_28251
|
rasdani/github-patches
|
git_diff
|
Lightning-AI__pytorch-lightning-3684
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
DataModules should have a reference to the Trainer
## 🚀 Feature
The DataModule should have a reference to the Trainer, so attributes on the Trainer can be accessed during prepare_data, setup, and the dataloader methods.
cc @nateraw
### Motivation
Further uncouple datamodules and lightning modules. For instance, it can be useful for DataModules to know what the current epoch is during training. If I want to use the DataModules, I don't have access to this. However this is readily available if I define my `train_dataloader` in the LightningModule instead.
### Pitch
We can follow the same setup that LightningModules have and how the trainer is initialized for them
</issue>
<code>
[start of pytorch_lightning/trainer/connectors/data_connector.py]
1 # Copyright The PyTorch Lightning team.
2 #
3 # Licensed under the Apache License, Version 2.0 (the "License");
4 # you may not use this file except in compliance with the License.
5 # You may obtain a copy of the License at
6 #
7 # http://www.apache.org/licenses/LICENSE-2.0
8 #
9 # Unless required by applicable law or agreed to in writing, software
10 # distributed under the License is distributed on an "AS IS" BASIS,
11 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12 # See the License for the specific language governing permissions and
13 # limitations under the License.
14
15 from pytorch_lightning.core.datamodule import LightningDataModule
16 from pytorch_lightning.utilities.exceptions import MisconfigurationException
17 from typing import List, Union
18 from torch.utils.data import DataLoader
19 from pytorch_lightning.utilities.model_utils import is_overridden
20
21
22 class DataConnector(object):
23
24 def __init__(self, trainer):
25 self.trainer = trainer
26
27 def on_trainer_init(self, check_val_every_n_epoch, reload_dataloaders_every_epoch, prepare_data_per_node):
28 self.trainer.datamodule = None
29 self.trainer.prepare_data_per_node = prepare_data_per_node
30
31 self.trainer.check_val_every_n_epoch = check_val_every_n_epoch
32 self.trainer.reload_dataloaders_every_epoch = reload_dataloaders_every_epoch
33 self.trainer._is_data_prepared = False
34
35 def get_profiled_train_dataloader(self, train_dataloader):
36 profiled_dl = self.trainer.profiler.profile_iterable(
37 enumerate(self._with_is_last(train_dataloader)),
38 "get_train_batch"
39 )
40 return profiled_dl
41
42 def _with_is_last(self, iterable):
43 """Pass through values from the given iterable with an added boolean indicating if this is the last item.
44 See `https://stackoverflow.com/a/1630350 <https://stackoverflow.com/a/1630350>`_"""
45 it = iter(iterable)
46 last = next(it)
47 for val in it:
48 # yield last and has next
49 yield last, False
50 last = val
51 # yield last, no longer has next
52 yield last, True
53
54 def prepare_data(self, model):
55 # on multi-gpu jobs we only want to manipulate (download, etc) on node_rank=0, local_rank=0
56 # or in the case where each node needs to do its own manipulation in which case just local_rank=0
57 if self.can_prepare_data():
58 if self.trainer.datamodule is not None:
59 self.trainer.datamodule.prepare_data()
60 model.prepare_data()
61 self.trainer._is_data_prepared = True
62
63 def can_prepare_data(self):
64 should_call_dm_prepare_data = True
65 if self.trainer.datamodule is not None and is_overridden('prepare_data', self.trainer.datamodule):
66 should_call_dm_prepare_data = not self.trainer.datamodule.has_prepared_data
67
68 if self.trainer.prepare_data_per_node:
69 return self.trainer.local_rank == 0 and should_call_dm_prepare_data
70 else:
71 return self.trainer.node_rank == 0 and self.trainer.local_rank == 0 and should_call_dm_prepare_data
72
73 def attach_data(self, model, train_dataloader, val_dataloaders, datamodule):
74 # if a datamodule comes in as the second arg, then fix it for the user
75 if isinstance(train_dataloader, LightningDataModule):
76 datamodule = train_dataloader
77 train_dataloader = None
78
79 self.__enforce_datamodule_dataloader_override(train_dataloader, val_dataloaders, datamodule)
80
81 # set up the passed in dataloaders (if needed)
82 self.attach_dataloaders(model, train_dataloader, val_dataloaders)
83 self.attach_datamodule(model, datamodule, 'fit')
84
85 def __enforce_datamodule_dataloader_override(self, train_dataloader, val_dataloaders, datamodule):
86 # If you supply a datamodule you can't supply train_dataloader or val_dataloaders
87 if (train_dataloader is not None or val_dataloaders is not None) and datamodule is not None:
88 raise MisconfigurationException(
89 'You cannot pass train_dataloader or val_dataloaders to trainer.fit if you supply a datamodule'
90 )
91
92 def attach_dataloaders(self, model, train_dataloader=None, val_dataloaders=None, test_dataloaders=None):
93 # when dataloader is passed via fit, patch the train_dataloader
94 # functions to overwrite with these implementations
95 if train_dataloader is not None:
96 model.train_dataloader = _PatchDataLoader(train_dataloader)
97
98 if val_dataloaders is not None:
99 model.val_dataloader = _PatchDataLoader(val_dataloaders)
100
101 if test_dataloaders is not None:
102 model.test_dataloader = _PatchDataLoader(test_dataloaders)
103
104 def attach_datamodule(self, model, datamodule, stage):
105
106 # We use datamodule if it's been provided on .fit or .test, otherwise we check model for it
107 datamodule = datamodule or getattr(model, 'datamodule', None)
108
109 # If we have a datamodule, attach necessary hooks + dataloaders
110 if datamodule:
111
112 # Override loader hooks
113 if is_overridden('train_dataloader', datamodule):
114 model.train_dataloader = datamodule.train_dataloader
115 if is_overridden('val_dataloader', datamodule):
116 model.val_dataloader = datamodule.val_dataloader
117 if is_overridden('test_dataloader', datamodule):
118 model.test_dataloader = datamodule.test_dataloader
119
120 # Override transfer_batch_to_device if dataset-specific to_device logic has been defined in datamodule
121 if is_overridden('transfer_batch_to_device', datamodule):
122 model.transfer_batch_to_device = datamodule.transfer_batch_to_device
123
124 self.trainer.datamodule = datamodule
125
126
127 class _PatchDataLoader(object):
128 r"""
129 Callable object for patching dataloaders passed into trainer.fit().
130 Use this class to override model.*_dataloader() and be pickle-compatible.
131
132 Args:
133 dataloader: Dataloader object to return when called.
134
135 """
136
137 def __init__(self, dataloader: Union[List[DataLoader], DataLoader]):
138 self.dataloader = dataloader
139
140 # cannot pickle __code__ so cannot verify if PatchDataloader
141 # exists which shows dataloader methods have been overwritten.
142 # so, we hack it by using the string representation
143 self.patch_loader_code = str(self.__call__.__code__)
144
145 def __call__(self) -> Union[List[DataLoader], DataLoader]:
146 return self.dataloader
147
[end of pytorch_lightning/trainer/connectors/data_connector.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/pytorch_lightning/trainer/connectors/data_connector.py b/pytorch_lightning/trainer/connectors/data_connector.py
--- a/pytorch_lightning/trainer/connectors/data_connector.py
+++ b/pytorch_lightning/trainer/connectors/data_connector.py
@@ -14,7 +14,7 @@
from pytorch_lightning.core.datamodule import LightningDataModule
from pytorch_lightning.utilities.exceptions import MisconfigurationException
-from typing import List, Union
+from typing import List, Optional, Union
from torch.utils.data import DataLoader
from pytorch_lightning.utilities.model_utils import is_overridden
@@ -101,7 +101,7 @@
if test_dataloaders is not None:
model.test_dataloader = _PatchDataLoader(test_dataloaders)
- def attach_datamodule(self, model, datamodule, stage):
+ def attach_datamodule(self, model, datamodule: Optional[LightningDataModule], stage: str) -> None:
# We use datamodule if it's been provided on .fit or .test, otherwise we check model for it
datamodule = datamodule or getattr(model, 'datamodule', None)
@@ -122,6 +122,7 @@
model.transfer_batch_to_device = datamodule.transfer_batch_to_device
self.trainer.datamodule = datamodule
+ datamodule.trainer = self.trainer
class _PatchDataLoader(object):
|
{"golden_diff": "diff --git a/pytorch_lightning/trainer/connectors/data_connector.py b/pytorch_lightning/trainer/connectors/data_connector.py\n--- a/pytorch_lightning/trainer/connectors/data_connector.py\n+++ b/pytorch_lightning/trainer/connectors/data_connector.py\n@@ -14,7 +14,7 @@\n \n from pytorch_lightning.core.datamodule import LightningDataModule\n from pytorch_lightning.utilities.exceptions import MisconfigurationException\n-from typing import List, Union\n+from typing import List, Optional, Union\n from torch.utils.data import DataLoader\n from pytorch_lightning.utilities.model_utils import is_overridden\n \n@@ -101,7 +101,7 @@\n if test_dataloaders is not None:\n model.test_dataloader = _PatchDataLoader(test_dataloaders)\n \n- def attach_datamodule(self, model, datamodule, stage):\n+ def attach_datamodule(self, model, datamodule: Optional[LightningDataModule], stage: str) -> None:\n \n # We use datamodule if it's been provided on .fit or .test, otherwise we check model for it\n datamodule = datamodule or getattr(model, 'datamodule', None)\n@@ -122,6 +122,7 @@\n model.transfer_batch_to_device = datamodule.transfer_batch_to_device\n \n self.trainer.datamodule = datamodule\n+ datamodule.trainer = self.trainer\n \n \n class _PatchDataLoader(object):\n", "issue": "DataModules should have a reference to the Trainer\n## \ud83d\ude80 Feature\r\nThe DataModule should have a reference to the Trainer, so attributes on the Trainer can be accessed during prepare_data, setup, and the dataloader methods. \r\n\r\ncc @nateraw \r\n\r\n### Motivation\r\nFurther uncouple datamodules and lightning modules. For instance, it can be useful for DataModules to know what the current epoch is during training. If I want to use the DataModules, I don't have access to this. However this is readily available if I define my `train_dataloader` in the LightningModule instead. \r\n\r\n\r\n### Pitch\r\nWe can follow the same setup that LightningModules have and how the trainer is initialized for them\r\n\n", "before_files": [{"content": "# Copyright The PyTorch Lightning team.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom pytorch_lightning.core.datamodule import LightningDataModule\nfrom pytorch_lightning.utilities.exceptions import MisconfigurationException\nfrom typing import List, Union\nfrom torch.utils.data import DataLoader\nfrom pytorch_lightning.utilities.model_utils import is_overridden\n\n\nclass DataConnector(object):\n\n def __init__(self, trainer):\n self.trainer = trainer\n\n def on_trainer_init(self, check_val_every_n_epoch, reload_dataloaders_every_epoch, prepare_data_per_node):\n self.trainer.datamodule = None\n self.trainer.prepare_data_per_node = prepare_data_per_node\n\n self.trainer.check_val_every_n_epoch = check_val_every_n_epoch\n self.trainer.reload_dataloaders_every_epoch = reload_dataloaders_every_epoch\n self.trainer._is_data_prepared = False\n\n def get_profiled_train_dataloader(self, train_dataloader):\n profiled_dl = self.trainer.profiler.profile_iterable(\n enumerate(self._with_is_last(train_dataloader)),\n \"get_train_batch\"\n )\n return profiled_dl\n\n def _with_is_last(self, iterable):\n \"\"\"Pass through values from the given iterable with an added boolean indicating if this is the last item.\n See `https://stackoverflow.com/a/1630350 <https://stackoverflow.com/a/1630350>`_\"\"\"\n it = iter(iterable)\n last = next(it)\n for val in it:\n # yield last and has next\n yield last, False\n last = val\n # yield last, no longer has next\n yield last, True\n\n def prepare_data(self, model):\n # on multi-gpu jobs we only want to manipulate (download, etc) on node_rank=0, local_rank=0\n # or in the case where each node needs to do its own manipulation in which case just local_rank=0\n if self.can_prepare_data():\n if self.trainer.datamodule is not None:\n self.trainer.datamodule.prepare_data()\n model.prepare_data()\n self.trainer._is_data_prepared = True\n\n def can_prepare_data(self):\n should_call_dm_prepare_data = True\n if self.trainer.datamodule is not None and is_overridden('prepare_data', self.trainer.datamodule):\n should_call_dm_prepare_data = not self.trainer.datamodule.has_prepared_data\n\n if self.trainer.prepare_data_per_node:\n return self.trainer.local_rank == 0 and should_call_dm_prepare_data\n else:\n return self.trainer.node_rank == 0 and self.trainer.local_rank == 0 and should_call_dm_prepare_data\n\n def attach_data(self, model, train_dataloader, val_dataloaders, datamodule):\n # if a datamodule comes in as the second arg, then fix it for the user\n if isinstance(train_dataloader, LightningDataModule):\n datamodule = train_dataloader\n train_dataloader = None\n\n self.__enforce_datamodule_dataloader_override(train_dataloader, val_dataloaders, datamodule)\n\n # set up the passed in dataloaders (if needed)\n self.attach_dataloaders(model, train_dataloader, val_dataloaders)\n self.attach_datamodule(model, datamodule, 'fit')\n\n def __enforce_datamodule_dataloader_override(self, train_dataloader, val_dataloaders, datamodule):\n # If you supply a datamodule you can't supply train_dataloader or val_dataloaders\n if (train_dataloader is not None or val_dataloaders is not None) and datamodule is not None:\n raise MisconfigurationException(\n 'You cannot pass train_dataloader or val_dataloaders to trainer.fit if you supply a datamodule'\n )\n\n def attach_dataloaders(self, model, train_dataloader=None, val_dataloaders=None, test_dataloaders=None):\n # when dataloader is passed via fit, patch the train_dataloader\n # functions to overwrite with these implementations\n if train_dataloader is not None:\n model.train_dataloader = _PatchDataLoader(train_dataloader)\n\n if val_dataloaders is not None:\n model.val_dataloader = _PatchDataLoader(val_dataloaders)\n\n if test_dataloaders is not None:\n model.test_dataloader = _PatchDataLoader(test_dataloaders)\n\n def attach_datamodule(self, model, datamodule, stage):\n\n # We use datamodule if it's been provided on .fit or .test, otherwise we check model for it\n datamodule = datamodule or getattr(model, 'datamodule', None)\n\n # If we have a datamodule, attach necessary hooks + dataloaders\n if datamodule:\n\n # Override loader hooks\n if is_overridden('train_dataloader', datamodule):\n model.train_dataloader = datamodule.train_dataloader\n if is_overridden('val_dataloader', datamodule):\n model.val_dataloader = datamodule.val_dataloader\n if is_overridden('test_dataloader', datamodule):\n model.test_dataloader = datamodule.test_dataloader\n\n # Override transfer_batch_to_device if dataset-specific to_device logic has been defined in datamodule\n if is_overridden('transfer_batch_to_device', datamodule):\n model.transfer_batch_to_device = datamodule.transfer_batch_to_device\n\n self.trainer.datamodule = datamodule\n\n\nclass _PatchDataLoader(object):\n r\"\"\"\n Callable object for patching dataloaders passed into trainer.fit().\n Use this class to override model.*_dataloader() and be pickle-compatible.\n\n Args:\n dataloader: Dataloader object to return when called.\n\n \"\"\"\n\n def __init__(self, dataloader: Union[List[DataLoader], DataLoader]):\n self.dataloader = dataloader\n\n # cannot pickle __code__ so cannot verify if PatchDataloader\n # exists which shows dataloader methods have been overwritten.\n # so, we hack it by using the string representation\n self.patch_loader_code = str(self.__call__.__code__)\n\n def __call__(self) -> Union[List[DataLoader], DataLoader]:\n return self.dataloader\n", "path": "pytorch_lightning/trainer/connectors/data_connector.py"}]}
| 2,543 | 327 |
gh_patches_debug_26319
|
rasdani/github-patches
|
git_diff
|
conan-io__conan-center-index-5256
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
[package] fontconfig/2.13.93: Please make gettext dependency optional
### Package and Environment Details (include every applicable attribute)
* Package Name/Version: **fontconfig/2.13.93**
* Operating System+version: **MacOs**
The current recipe adds an unconditional dependency on libgettext/0.20.1 on MacOs.
Since libgettext is licensed under GPLv3, it places an additional license restriction to an otherwise more liberally licensed library.
</details>
</issue>
<code>
[start of recipes/fontconfig/all/conanfile.py]
1 import os
2 import glob
3
4 from conans import ConanFile, tools, AutoToolsBuildEnvironment
5 from conans.errors import ConanInvalidConfiguration
6
7
8 class FontconfigConan(ConanFile):
9 name = "fontconfig"
10 license = "MIT"
11 url = "https://github.com/conan-io/conan-center-index"
12 description = "Fontconfig is a library for configuring and customizing font access"
13 homepage = "https://gitlab.freedesktop.org/fontconfig/fontconfig"
14 topics = ("conan", "fontconfig", "fonts", "freedesktop")
15 settings = "os", "compiler", "build_type", "arch"
16 options = {"shared": [True, False], "fPIC": [True, False]}
17 default_options = {"shared": False, "fPIC": True}
18 generators = "pkg_config"
19
20 _autotools = None
21
22 @property
23 def _source_subfolder(self):
24 return "source_subfolder"
25
26 def config_options(self):
27 if self.settings.os == "Windows":
28 del self.options.fPIC
29
30 def configure(self):
31 if self.settings.compiler == "Visual Studio":
32 raise ConanInvalidConfiguration("Visual Studio builds are not supported.")
33 if self.options.shared:
34 del self.options.fPIC
35 del self.settings.compiler.libcxx
36 del self.settings.compiler.cppstd
37
38 def requirements(self):
39 self.requires("freetype/2.10.4")
40 self.requires("expat/2.2.10")
41 if self.settings.os == "Linux":
42 self.requires("libuuid/1.0.3")
43 elif self.settings.os == "Macos":
44 self.requires("libgettext/0.20.1")
45
46 def build_requirements(self):
47 self.build_requires("gperf/3.1")
48 self.build_requires("pkgconf/1.7.3")
49 if tools.os_info.is_windows and not tools.get_env("CONAN_BASH_PATH"):
50 self.build_requires("msys2/20200517")
51
52 def source(self):
53 tools.get(**self.conan_data["sources"][self.version])
54 extrated_dir = self.name + "-" + self.version
55 os.rename(extrated_dir, self._source_subfolder)
56
57 def _configure_autotools(self):
58 if not self._autotools:
59 args = ["--enable-static=%s" % ("no" if self.options.shared else "yes"),
60 "--enable-shared=%s" % ("yes" if self.options.shared else "no"),
61 "--disable-docs"]
62 args.append("--sysconfdir=%s" % tools.unix_path(os.path.join(self.package_folder, "bin", "etc")))
63 args.append("--datadir=%s" % tools.unix_path(os.path.join(self.package_folder, "bin", "share")))
64 args.append("--datarootdir=%s" % tools.unix_path(os.path.join(self.package_folder, "bin", "share")))
65 args.append("--localstatedir=%s" % tools.unix_path(os.path.join(self.package_folder, "bin", "var")))
66 self._autotools = AutoToolsBuildEnvironment(self, win_bash=tools.os_info.is_windows)
67 self._autotools.libs = []
68 self._autotools.configure(configure_dir=self._source_subfolder, args=args)
69 tools.replace_in_file("Makefile", "po-conf test", "po-conf")
70 return self._autotools
71
72 def _patch_files(self):
73 # - fontconfig requires libtool version number, change it for the corresponding freetype one
74 tools.replace_in_file(os.path.join(self._source_subfolder, 'configure'), '21.0.15', '2.8.1')
75
76 def build(self):
77 # Patch files from dependencies
78 self._patch_files()
79 with tools.run_environment(self):
80 autotools = self._configure_autotools()
81 autotools.make()
82
83 def package(self):
84 self.copy("COPYING", dst="licenses", src=self._source_subfolder)
85 with tools.run_environment(self):
86 autotools = self._configure_autotools()
87 autotools.install()
88 os.unlink(os.path.join(self.package_folder, "lib", "libfontconfig.la"))
89 tools.rmdir(os.path.join(self.package_folder, "lib", "pkgconfig"))
90 for f in glob.glob(os.path.join(self.package_folder, "bin", "etc", "fonts", "conf.d", "*.conf")):
91 if os.path.islink(f):
92 os.unlink(f)
93 for def_file in glob.glob(os.path.join(self.package_folder, "lib", "*.def")):
94 os.remove(def_file)
95
96 def package_info(self):
97 self.cpp_info.libs = ["fontconfig"]
98 if self.settings.os in ["Linux", "FreeBSD"]:
99 self.cpp_info.system_libs.extend(["m", "pthread"])
100 self.cpp_info.names["cmake_find_package"] = "Fontconfig"
101 self.cpp_info.names["cmake_find_package_multi"] = "Fontconfig"
102
103 fontconfig_file = os.path.join(self.package_folder, "bin", "etc", "fonts", "fonts.conf")
104 self.output.info("Creating FONTCONFIG_FILE environment variable: {}".format(fontconfig_file))
105 self.env_info.FONTCONFIG_FILE = fontconfig_file
106 fontconfig_path = os.path.join(self.package_folder, "bin", "etc", "fonts")
107 self.output.info("Creating FONTCONFIG_PATH environment variable: {}".format(fontconfig_path))
108 self.env_info.FONTCONFIG_PATH = fontconfig_path
109
[end of recipes/fontconfig/all/conanfile.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/recipes/fontconfig/all/conanfile.py b/recipes/fontconfig/all/conanfile.py
--- a/recipes/fontconfig/all/conanfile.py
+++ b/recipes/fontconfig/all/conanfile.py
@@ -40,8 +40,6 @@
self.requires("expat/2.2.10")
if self.settings.os == "Linux":
self.requires("libuuid/1.0.3")
- elif self.settings.os == "Macos":
- self.requires("libgettext/0.20.1")
def build_requirements(self):
self.build_requires("gperf/3.1")
@@ -58,7 +56,9 @@
if not self._autotools:
args = ["--enable-static=%s" % ("no" if self.options.shared else "yes"),
"--enable-shared=%s" % ("yes" if self.options.shared else "no"),
- "--disable-docs"]
+ "--disable-docs",
+ "--disable-nls",
+ ]
args.append("--sysconfdir=%s" % tools.unix_path(os.path.join(self.package_folder, "bin", "etc")))
args.append("--datadir=%s" % tools.unix_path(os.path.join(self.package_folder, "bin", "share")))
args.append("--datarootdir=%s" % tools.unix_path(os.path.join(self.package_folder, "bin", "share")))
|
{"golden_diff": "diff --git a/recipes/fontconfig/all/conanfile.py b/recipes/fontconfig/all/conanfile.py\n--- a/recipes/fontconfig/all/conanfile.py\n+++ b/recipes/fontconfig/all/conanfile.py\n@@ -40,8 +40,6 @@\n self.requires(\"expat/2.2.10\")\n if self.settings.os == \"Linux\":\n self.requires(\"libuuid/1.0.3\")\n- elif self.settings.os == \"Macos\":\n- self.requires(\"libgettext/0.20.1\")\n \n def build_requirements(self):\n self.build_requires(\"gperf/3.1\")\n@@ -58,7 +56,9 @@\n if not self._autotools:\n args = [\"--enable-static=%s\" % (\"no\" if self.options.shared else \"yes\"),\n \"--enable-shared=%s\" % (\"yes\" if self.options.shared else \"no\"),\n- \"--disable-docs\"]\n+ \"--disable-docs\",\n+ \"--disable-nls\",\n+ ]\n args.append(\"--sysconfdir=%s\" % tools.unix_path(os.path.join(self.package_folder, \"bin\", \"etc\")))\n args.append(\"--datadir=%s\" % tools.unix_path(os.path.join(self.package_folder, \"bin\", \"share\")))\n args.append(\"--datarootdir=%s\" % tools.unix_path(os.path.join(self.package_folder, \"bin\", \"share\")))\n", "issue": "[package] fontconfig/2.13.93: Please make gettext dependency optional\n### Package and Environment Details (include every applicable attribute)\r\n * Package Name/Version: **fontconfig/2.13.93**\r\n * Operating System+version: **MacOs**\r\n\r\nThe current recipe adds an unconditional dependency on libgettext/0.20.1 on MacOs.\r\n\r\nSince libgettext is licensed under GPLv3, it places an additional license restriction to an otherwise more liberally licensed library.\r\n\r\n</details>\r\n\n", "before_files": [{"content": "import os\nimport glob\n\nfrom conans import ConanFile, tools, AutoToolsBuildEnvironment\nfrom conans.errors import ConanInvalidConfiguration\n\n\nclass FontconfigConan(ConanFile):\n name = \"fontconfig\"\n license = \"MIT\"\n url = \"https://github.com/conan-io/conan-center-index\"\n description = \"Fontconfig is a library for configuring and customizing font access\"\n homepage = \"https://gitlab.freedesktop.org/fontconfig/fontconfig\"\n topics = (\"conan\", \"fontconfig\", \"fonts\", \"freedesktop\")\n settings = \"os\", \"compiler\", \"build_type\", \"arch\"\n options = {\"shared\": [True, False], \"fPIC\": [True, False]}\n default_options = {\"shared\": False, \"fPIC\": True}\n generators = \"pkg_config\"\n\n _autotools = None\n\n @property\n def _source_subfolder(self):\n return \"source_subfolder\"\n\n def config_options(self):\n if self.settings.os == \"Windows\":\n del self.options.fPIC\n\n def configure(self):\n if self.settings.compiler == \"Visual Studio\":\n raise ConanInvalidConfiguration(\"Visual Studio builds are not supported.\")\n if self.options.shared:\n del self.options.fPIC\n del self.settings.compiler.libcxx\n del self.settings.compiler.cppstd\n\n def requirements(self):\n self.requires(\"freetype/2.10.4\")\n self.requires(\"expat/2.2.10\")\n if self.settings.os == \"Linux\":\n self.requires(\"libuuid/1.0.3\")\n elif self.settings.os == \"Macos\":\n self.requires(\"libgettext/0.20.1\")\n\n def build_requirements(self):\n self.build_requires(\"gperf/3.1\")\n self.build_requires(\"pkgconf/1.7.3\")\n if tools.os_info.is_windows and not tools.get_env(\"CONAN_BASH_PATH\"):\n self.build_requires(\"msys2/20200517\")\n\n def source(self):\n tools.get(**self.conan_data[\"sources\"][self.version])\n extrated_dir = self.name + \"-\" + self.version\n os.rename(extrated_dir, self._source_subfolder)\n\n def _configure_autotools(self):\n if not self._autotools:\n args = [\"--enable-static=%s\" % (\"no\" if self.options.shared else \"yes\"),\n \"--enable-shared=%s\" % (\"yes\" if self.options.shared else \"no\"),\n \"--disable-docs\"]\n args.append(\"--sysconfdir=%s\" % tools.unix_path(os.path.join(self.package_folder, \"bin\", \"etc\")))\n args.append(\"--datadir=%s\" % tools.unix_path(os.path.join(self.package_folder, \"bin\", \"share\")))\n args.append(\"--datarootdir=%s\" % tools.unix_path(os.path.join(self.package_folder, \"bin\", \"share\")))\n args.append(\"--localstatedir=%s\" % tools.unix_path(os.path.join(self.package_folder, \"bin\", \"var\")))\n self._autotools = AutoToolsBuildEnvironment(self, win_bash=tools.os_info.is_windows)\n self._autotools.libs = []\n self._autotools.configure(configure_dir=self._source_subfolder, args=args)\n tools.replace_in_file(\"Makefile\", \"po-conf test\", \"po-conf\")\n return self._autotools\n\n def _patch_files(self):\n # - fontconfig requires libtool version number, change it for the corresponding freetype one\n tools.replace_in_file(os.path.join(self._source_subfolder, 'configure'), '21.0.15', '2.8.1')\n\n def build(self):\n # Patch files from dependencies\n self._patch_files()\n with tools.run_environment(self):\n autotools = self._configure_autotools()\n autotools.make()\n\n def package(self):\n self.copy(\"COPYING\", dst=\"licenses\", src=self._source_subfolder)\n with tools.run_environment(self):\n autotools = self._configure_autotools()\n autotools.install()\n os.unlink(os.path.join(self.package_folder, \"lib\", \"libfontconfig.la\"))\n tools.rmdir(os.path.join(self.package_folder, \"lib\", \"pkgconfig\"))\n for f in glob.glob(os.path.join(self.package_folder, \"bin\", \"etc\", \"fonts\", \"conf.d\", \"*.conf\")):\n if os.path.islink(f):\n os.unlink(f)\n for def_file in glob.glob(os.path.join(self.package_folder, \"lib\", \"*.def\")):\n os.remove(def_file)\n\n def package_info(self):\n self.cpp_info.libs = [\"fontconfig\"]\n if self.settings.os in [\"Linux\", \"FreeBSD\"]:\n self.cpp_info.system_libs.extend([\"m\", \"pthread\"])\n self.cpp_info.names[\"cmake_find_package\"] = \"Fontconfig\"\n self.cpp_info.names[\"cmake_find_package_multi\"] = \"Fontconfig\"\n\n fontconfig_file = os.path.join(self.package_folder, \"bin\", \"etc\", \"fonts\", \"fonts.conf\")\n self.output.info(\"Creating FONTCONFIG_FILE environment variable: {}\".format(fontconfig_file))\n self.env_info.FONTCONFIG_FILE = fontconfig_file\n fontconfig_path = os.path.join(self.package_folder, \"bin\", \"etc\", \"fonts\")\n self.output.info(\"Creating FONTCONFIG_PATH environment variable: {}\".format(fontconfig_path))\n self.env_info.FONTCONFIG_PATH = fontconfig_path\n", "path": "recipes/fontconfig/all/conanfile.py"}]}
| 2,065 | 311 |
gh_patches_debug_2840
|
rasdani/github-patches
|
git_diff
|
conan-io__conan-center-index-2625
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
[package] libx265/3.2.1: hash of source tar.gz changed
### Package and Environment Details (include every applicable attribute)
* Package Name/Version: **libx265/3.2.1**
* Operating System+version: **Linux Ubuntu 18.04 and 20.04 **
* Compiler+version: **GCC 7.5 and 8**
* Conan version: **conan 1.28.1**
* Python version: **Python 3.6.8**
### Conan profile
```
Configuration for profile default:
[settings]
os=Linux
os_build=Linux
arch=x86_64
arch_build=x86_64
compiler=gcc
compiler.version=7.5
compiler.libcxx=libstdc++11
build_type=Release
[options]
[build_requires]
[env]
```
### Steps to reproduce (Include if Applicable)
Try to build libx265/3.2.1 from source.
### Logs (Include/Attach if Applicable)
<details><summary>Click to expand log</summary>
```
ERROR: libx265/3.2.1: Error in source() method, line 50
tools.get(**self.conan_data["sources"][self.version])
ConanException: sha256 signature failed for 'x265_3.2.1.tar.gz' file.
Provided signature: 7cf8ed2927fcb2914cdca51c974594770da705cb43288beea62b69c53725b5d7
Computed signature: fb9badcf92364fd3567f8b5aa0e5e952aeea7a39a2b864387cec31e3b58cbbcc (edited)
```
</details>
I'm surprised that this wasn't caught earlier. Indeed, when downloading the file manually and computing the sha256sum I get the same as the computed signature above. What's the best way to make a PR for this? Would we bump the recipe version or just overwrite the current 3.2.1?
</issue>
<code>
[start of recipes/libx265/all/conanfile.py]
1 from conans import CMake, ConanFile, tools
2 import os
3 import shutil
4
5
6 class Libx265Conan(ConanFile):
7 name = "libx265"
8 description = "x265 is the leading H.265 / HEVC encoder software library"
9 topics = ("conan", "libx265", "codec", "video", "H.265")
10 url = "https://github.com/conan-io/conan-center-index"
11 homepage = " https://bitbucket.org/multicoreware/x265"
12 exports_sources = ["CMakeLists.txt", "patches/*"]
13 generators = "cmake"
14 license = ("GPL-2.0-only", "commercial") # https://bitbucket.org/multicoreware/x265/src/default/COPYING
15 settings = "os", "arch", "compiler", "build_type"
16 options = {
17 "shared": [True, False],
18 "fPIC": [True, False],
19 "bit_depth": [8, 10, 12],
20 "HDR10": [True, False],
21 "SVG_HEVC_encoder": [True, False],
22 }
23 default_options = {
24 "shared": False,
25 "fPIC": True,
26 "bit_depth": 8,
27 "HDR10": False,
28 "SVG_HEVC_encoder": False,
29 }
30
31 _cmake = None
32
33 @property
34 def _source_subfolder(self):
35 return "source_subfolder"
36
37 @property
38 def _build_subfolder(self):
39 return "build_subfolder"
40
41 def config_options(self):
42 if self.settings.os == "Windows":
43 del self.options.fPIC
44
45 def configure(self):
46 if self.options.shared:
47 del self.options.fPIC
48
49 def source(self):
50 tools.get(**self.conan_data["sources"][self.version])
51 os.rename("x265_{}".format(self.version), self._source_subfolder)
52
53 def _configure_cmake(self):
54 if self._cmake:
55 return self._cmake
56 self._cmake = CMake(self)
57 self._cmake.definitions["ENABLE_SHARED"] = self.options.shared
58 self._cmake.definitions["ENABLE_LIBNUMA"] = False
59 if self.settings.os == "Macos":
60 self._cmake.definitions["CMAKE_SHARED_LINKER_FLAGS"] = "-Wl,-read_only_relocs,suppress"
61 self._cmake.definitions["HIGH_BIT_DEPTH"] = self.options.bit_depth != 8
62 self._cmake.definitions["MAIN12"] = self.options.bit_depth == 12
63 self._cmake.definitions["ENABLE_HDR10_PLUS"] = self.options.HDR10
64 self._cmake.definitions["ENABLE_SVT_HEVC"] = self.options.SVG_HEVC_encoder
65 if self.settings.compiler == "Visual Studio":
66 self._cmake.definitions["STATIC_LINK_CRT"] = "T" in str(self.settings.compiler.runtime)
67 if self.settings.os == "Linux":
68 self._cmake.definitions["PLATFORM_LIBS"] = "dl"
69 self._cmake.configure(build_folder=self._build_subfolder)
70 return self._cmake
71
72 def _patch_sources(self):
73 cmakelists = os.path.join(self._source_subfolder, "source", "CMakeLists.txt")
74 tools.replace_in_file(cmakelists,
75 "if((WIN32 AND ENABLE_CLI) OR (WIN32 AND ENABLE_SHARED))",
76 "if(FALSE)")
77 if self.settings.os == "Android":
78 tools.replace_in_file(cmakelists,
79 "list(APPEND PLATFORM_LIBS pthread)", "")
80 tools.replace_in_file(cmakelists,
81 "list(APPEND PLATFORM_LIBS rt)", "")
82 for patch in self.conan_data.get("patches", {}).get(self.version, []):
83 tools.patch(**patch)
84
85 def build(self):
86 self._patch_sources()
87 cmake = self._configure_cmake()
88 cmake.build()
89
90 def package(self):
91 self.copy("COPYING", src=self._source_subfolder, dst="licenses")
92 cmake = self._configure_cmake()
93 cmake.install()
94
95 if self.options.shared:
96 if self.settings.compiler == "Visual Studio":
97 static_lib = "x265-static.lib"
98 else:
99 static_lib = "libx265.a"
100 os.unlink(os.path.join(self.package_folder, "lib", static_lib))
101
102 if self.settings.compiler == "Visual Studio":
103 name = "libx265.lib" if self.options.shared else "x265-static.lib"
104 shutil.move(os.path.join(self.package_folder, "lib", name),
105 os.path.join(self.package_folder, "lib", "x265.lib"))
106
107 if self.settings.os != "Windows" or not self.options.shared:
108 tools.rmdir(os.path.join(self.package_folder, "bin"))
109 else:
110 for file in os.listdir(os.path.join(self.package_folder, "bin")):
111 if not file.endswith(".dll"):
112 os.unlink(os.path.join(self.package_folder, "bin", file))
113 tools.rmdir(os.path.join(self.package_folder, "lib", "pkgconfig"))
114
115 def package_info(self):
116 self.cpp_info.names["pkg_config"] = "x265"
117 self.cpp_info.libs = ["x265"]
118 if self.settings.os == "Linux":
119 self.cpp_info.system_libs.extend(["dl", "pthread", "m"])
120 if self.settings.os == "Android":
121 self.cpp_info.libs.extend(["dl", "m"])
122 libcxx = self.settings.get_safe("compiler.libcxx")
123 if libcxx in ["libstdc++", "libstdc++11"]:
124 self.cpp_info.system_libs.append("stdc++")
125 elif libcxx == "libc++":
126 self.cpp_info.system_libs.append("c++")
127 elif libcxx in ["c++_static", "c++_shared"]:
128 self.cpp_info.system_libs.extend([libcxx, "c++abi"])
129
[end of recipes/libx265/all/conanfile.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/recipes/libx265/all/conanfile.py b/recipes/libx265/all/conanfile.py
--- a/recipes/libx265/all/conanfile.py
+++ b/recipes/libx265/all/conanfile.py
@@ -48,7 +48,7 @@
def source(self):
tools.get(**self.conan_data["sources"][self.version])
- os.rename("x265_{}".format(self.version), self._source_subfolder)
+ os.rename("x265-{}".format(self.version), self._source_subfolder)
def _configure_cmake(self):
if self._cmake:
|
{"golden_diff": "diff --git a/recipes/libx265/all/conanfile.py b/recipes/libx265/all/conanfile.py\n--- a/recipes/libx265/all/conanfile.py\n+++ b/recipes/libx265/all/conanfile.py\n@@ -48,7 +48,7 @@\n \n def source(self):\n tools.get(**self.conan_data[\"sources\"][self.version])\n- os.rename(\"x265_{}\".format(self.version), self._source_subfolder)\n+ os.rename(\"x265-{}\".format(self.version), self._source_subfolder)\n \n def _configure_cmake(self):\n if self._cmake:\n", "issue": "[package] libx265/3.2.1: hash of source tar.gz changed\n### Package and Environment Details (include every applicable attribute)\r\n * Package Name/Version: **libx265/3.2.1**\r\n * Operating System+version: **Linux Ubuntu 18.04 and 20.04 **\r\n * Compiler+version: **GCC 7.5 and 8**\r\n * Conan version: **conan 1.28.1**\r\n * Python version: **Python 3.6.8**\r\n\r\n\r\n### Conan profile\r\n```\r\nConfiguration for profile default:\r\n\r\n[settings]\r\nos=Linux\r\nos_build=Linux\r\narch=x86_64\r\narch_build=x86_64\r\ncompiler=gcc\r\ncompiler.version=7.5\r\ncompiler.libcxx=libstdc++11\r\nbuild_type=Release\r\n[options]\r\n[build_requires]\r\n[env]\r\n```\r\n\r\n\r\n### Steps to reproduce (Include if Applicable)\r\nTry to build libx265/3.2.1 from source. \r\n\r\n\r\n### Logs (Include/Attach if Applicable)\r\n<details><summary>Click to expand log</summary>\r\n\r\n```\r\nERROR: libx265/3.2.1: Error in source() method, line 50\r\n\ttools.get(**self.conan_data[\"sources\"][self.version])\r\n\tConanException: sha256 signature failed for 'x265_3.2.1.tar.gz' file. \r\n Provided signature: 7cf8ed2927fcb2914cdca51c974594770da705cb43288beea62b69c53725b5d7 \r\n Computed signature: fb9badcf92364fd3567f8b5aa0e5e952aeea7a39a2b864387cec31e3b58cbbcc (edited) \r\n```\r\n\r\n</details>\r\n\r\nI'm surprised that this wasn't caught earlier. Indeed, when downloading the file manually and computing the sha256sum I get the same as the computed signature above. What's the best way to make a PR for this? Would we bump the recipe version or just overwrite the current 3.2.1?\n", "before_files": [{"content": "from conans import CMake, ConanFile, tools\nimport os\nimport shutil\n\n\nclass Libx265Conan(ConanFile):\n name = \"libx265\"\n description = \"x265 is the leading H.265 / HEVC encoder software library\"\n topics = (\"conan\", \"libx265\", \"codec\", \"video\", \"H.265\")\n url = \"https://github.com/conan-io/conan-center-index\"\n homepage = \" https://bitbucket.org/multicoreware/x265\"\n exports_sources = [\"CMakeLists.txt\", \"patches/*\"]\n generators = \"cmake\"\n license = (\"GPL-2.0-only\", \"commercial\") # https://bitbucket.org/multicoreware/x265/src/default/COPYING\n settings = \"os\", \"arch\", \"compiler\", \"build_type\"\n options = {\n \"shared\": [True, False],\n \"fPIC\": [True, False],\n \"bit_depth\": [8, 10, 12],\n \"HDR10\": [True, False],\n \"SVG_HEVC_encoder\": [True, False],\n }\n default_options = {\n \"shared\": False,\n \"fPIC\": True,\n \"bit_depth\": 8,\n \"HDR10\": False,\n \"SVG_HEVC_encoder\": False,\n }\n\n _cmake = None\n\n @property\n def _source_subfolder(self):\n return \"source_subfolder\"\n\n @property\n def _build_subfolder(self):\n return \"build_subfolder\"\n\n def config_options(self):\n if self.settings.os == \"Windows\":\n del self.options.fPIC\n\n def configure(self):\n if self.options.shared:\n del self.options.fPIC\n\n def source(self):\n tools.get(**self.conan_data[\"sources\"][self.version])\n os.rename(\"x265_{}\".format(self.version), self._source_subfolder)\n\n def _configure_cmake(self):\n if self._cmake:\n return self._cmake\n self._cmake = CMake(self)\n self._cmake.definitions[\"ENABLE_SHARED\"] = self.options.shared\n self._cmake.definitions[\"ENABLE_LIBNUMA\"] = False\n if self.settings.os == \"Macos\":\n self._cmake.definitions[\"CMAKE_SHARED_LINKER_FLAGS\"] = \"-Wl,-read_only_relocs,suppress\"\n self._cmake.definitions[\"HIGH_BIT_DEPTH\"] = self.options.bit_depth != 8\n self._cmake.definitions[\"MAIN12\"] = self.options.bit_depth == 12\n self._cmake.definitions[\"ENABLE_HDR10_PLUS\"] = self.options.HDR10\n self._cmake.definitions[\"ENABLE_SVT_HEVC\"] = self.options.SVG_HEVC_encoder\n if self.settings.compiler == \"Visual Studio\":\n self._cmake.definitions[\"STATIC_LINK_CRT\"] = \"T\" in str(self.settings.compiler.runtime)\n if self.settings.os == \"Linux\":\n self._cmake.definitions[\"PLATFORM_LIBS\"] = \"dl\"\n self._cmake.configure(build_folder=self._build_subfolder)\n return self._cmake\n\n def _patch_sources(self):\n cmakelists = os.path.join(self._source_subfolder, \"source\", \"CMakeLists.txt\")\n tools.replace_in_file(cmakelists,\n \"if((WIN32 AND ENABLE_CLI) OR (WIN32 AND ENABLE_SHARED))\",\n \"if(FALSE)\")\n if self.settings.os == \"Android\":\n tools.replace_in_file(cmakelists,\n \"list(APPEND PLATFORM_LIBS pthread)\", \"\")\n tools.replace_in_file(cmakelists,\n \"list(APPEND PLATFORM_LIBS rt)\", \"\")\n for patch in self.conan_data.get(\"patches\", {}).get(self.version, []):\n tools.patch(**patch)\n\n def build(self):\n self._patch_sources()\n cmake = self._configure_cmake()\n cmake.build()\n\n def package(self):\n self.copy(\"COPYING\", src=self._source_subfolder, dst=\"licenses\")\n cmake = self._configure_cmake()\n cmake.install()\n\n if self.options.shared:\n if self.settings.compiler == \"Visual Studio\":\n static_lib = \"x265-static.lib\"\n else:\n static_lib = \"libx265.a\"\n os.unlink(os.path.join(self.package_folder, \"lib\", static_lib))\n\n if self.settings.compiler == \"Visual Studio\":\n name = \"libx265.lib\" if self.options.shared else \"x265-static.lib\"\n shutil.move(os.path.join(self.package_folder, \"lib\", name),\n os.path.join(self.package_folder, \"lib\", \"x265.lib\"))\n\n if self.settings.os != \"Windows\" or not self.options.shared:\n tools.rmdir(os.path.join(self.package_folder, \"bin\"))\n else:\n for file in os.listdir(os.path.join(self.package_folder, \"bin\")):\n if not file.endswith(\".dll\"):\n os.unlink(os.path.join(self.package_folder, \"bin\", file))\n tools.rmdir(os.path.join(self.package_folder, \"lib\", \"pkgconfig\"))\n\n def package_info(self):\n self.cpp_info.names[\"pkg_config\"] = \"x265\"\n self.cpp_info.libs = [\"x265\"]\n if self.settings.os == \"Linux\":\n self.cpp_info.system_libs.extend([\"dl\", \"pthread\", \"m\"])\n if self.settings.os == \"Android\":\n self.cpp_info.libs.extend([\"dl\", \"m\"])\n libcxx = self.settings.get_safe(\"compiler.libcxx\")\n if libcxx in [\"libstdc++\", \"libstdc++11\"]:\n self.cpp_info.system_libs.append(\"stdc++\")\n elif libcxx == \"libc++\":\n self.cpp_info.system_libs.append(\"c++\")\n elif libcxx in [\"c++_static\", \"c++_shared\"]:\n self.cpp_info.system_libs.extend([libcxx, \"c++abi\"])\n", "path": "recipes/libx265/all/conanfile.py"}]}
| 2,658 | 150 |
gh_patches_debug_12144
|
rasdani/github-patches
|
git_diff
|
apluslms__a-plus-1183
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Hide exercises in the maintenance mode from the course table of contents
Exercises in maintenance mode (LearningObject.status maintenance) should be hidden from course table of contents.
Exercises with the status hidden are already hidden in the table of contents. The maintenance mode could be handled in the code in the same way.
</issue>
<code>
[start of exercise/cache/hierarchy.py]
1 from django.http.response import Http404
2 from course.models import CourseModule, LearningObjectCategory
3 from ..models import LearningObject
4
5
6 class NoSuchContent(Exception):
7 pass
8
9
10 class HierarchyIterator:
11 # pylint: disable-next=too-many-arguments
12 def __init__(self, children, idx=None, tree=None, visited=False, enclosed=True):
13 if idx is None:
14 self._default_start(children)
15 else:
16 self.idx = idx.copy()
17 self.levels = [children]
18 if tree and len(tree) > 1:
19 for entry in tree[:-1]:
20 self.levels.append(entry['children'])
21 self.visited = visited
22 self.enclose_begun = not enclosed
23 self.enclose_ended = not enclosed
24
25 def __iter__(self): # pylint: disable=non-iterator-returned
26 return self
27
28
29 class NextIterator(HierarchyIterator):
30
31 def _default_start(self, children):
32 self.idx = [0]
33 self.levels = [children]
34
35 def __next__(self):
36 if not self.enclose_begun:
37 self.enclose_begun = True
38 return {'type':'level','down':True}
39 i = self.idx[-1]
40 level = self.levels[-1]
41 if not self.visited:
42 if i < len(level):
43 self.visited = True
44 return level[i]
45 else:
46 children = level[i].get('children')
47 if children:
48 self.levels.append(children)
49 self.idx.append(0)
50 self.visited = False
51 return {'type':'level','down':True}
52 i += 1
53 if i < len(level):
54 self.idx[-1] = i
55 return level[i]
56 if len(self.idx) > 1:
57 self.idx = self.idx[:-1]
58 self.levels = self.levels[:-1]
59 self.idx[-1] += 1
60 self.visited = False
61 return {'type':'level','up':True}
62 if not self.enclose_ended:
63 self.enclose_ended = True
64 return {'type':'level','up':True}
65 raise StopIteration()
66
67
68 class PreviousIterator(HierarchyIterator):
69
70 def _default_start(self, children):
71 self.idx = []
72 self.levels = []
73 self._goto_last(children)
74
75 def _goto_last(self, children):
76 level = children
77 while level:
78 i = len(level) - 1
79 self.idx.append(i)
80 self.levels.append(level)
81 level = level[i].get('children')
82
83 def __next__(self):
84 i = self.idx[-1]
85 level = self.levels[-1]
86 if not self.visited:
87 self.visited = True
88 return level[i]
89 if i > 0:
90 i -= 1
91 self.idx[-1] = i
92 self._goto_last(level[i].get('children'))
93 return self.levels[-1][self.idx[-1]]
94 if len(self.idx) > 1:
95 self.idx = self.idx[:-1]
96 self.levels = self.levels[:-1]
97 return self.levels[-1][self.idx[-1]]
98 raise StopIteration()
99
100
101 class ContentMixin:
102
103 def created(self):
104 return self.data['created']
105
106 def total(self):
107 return self.data['total']
108
109 def modules(self):
110 return self.data['modules']
111
112 def modules_flatted(self):
113 for module in self.data['modules']:
114 module['flatted'] = self.flat_module(module)
115 return self.data['modules']
116
117 def categories(self):
118 categories = list(self.data['categories'].values())
119 categories.sort(key=lambda entry: entry['name'])
120 return categories
121
122 def flat_module(self, module, enclosed=True):
123 modules = self.modules()
124 idx = self._model_idx(module)
125 tree = self._by_idx(modules, idx)
126 return NextIterator(tree[0]['children'], enclosed=enclosed)
127
128 def flat_full(self):
129 return NextIterator(self.modules(), enclosed=False)
130
131 def begin(self):
132 for entry in self.flat_full():
133 if entry['type'] == 'exercise':
134 return entry
135 return None
136
137 def find_path(self, module_id, path):
138 paths = self.data['paths'].get(module_id, {})
139 if path in paths:
140 return paths[path]
141 raise NoSuchContent()
142
143 def find_number(self, number):
144 hit = None
145 search = self.modules()
146 parts = number.split('.')
147 for i in range(len(parts)):
148 number = '.'.join(parts[0:i+1])
149 for s in search:
150 if s['number'] == number:
151 hit = s
152 search = hit['children']
153 break
154 else:
155 raise NoSuchContent()
156 return hit
157
158 def find_category(self, category_id):
159 categories = self.data['categories']
160 if category_id in categories:
161 return categories[category_id]
162 raise NoSuchContent()
163
164 def find(self, model):
165 modules = self.modules()
166 idx = self._model_idx(model)
167 tree = self._by_idx(modules, idx)
168 return (
169 tree[-1],
170 tree,
171 self._previous(idx, tree),
172 self._next(idx, tree),
173 )
174
175 def search_exercises(self, **kwargs):
176 _, entries = self.search_entries(**kwargs)
177 return [e for e in entries if e['type'] == 'exercise']
178
179 def search_entries(self, number=None, category_id=None, module_id=None, # noqa: MC0001
180 exercise_id=None, filter_for_assistant=False, best=False, # pylint: disable=unused-argument
181 raise_404=True):
182 entry = None
183 if number:
184 try:
185 entry = self.find_number(number)
186 if entry['type'] == 'module':
187 module_id = entry['id']
188 elif entry['type'] == 'exercise':
189 exercise_id = entry['id']
190 except NoSuchContent:
191 if raise_404:
192 raise Http404() # pylint: disable=raise-missing-from
193 raise
194 search = None
195 if exercise_id is not None:
196 search = { 'type': 'exercise', 'id': int(exercise_id) }
197 elif module_id is not None:
198 search = { 'type': 'module', 'id': int(module_id) }
199 if search:
200 try:
201 idx = self._model_idx(search)
202 except NoSuchContent:
203 if raise_404:
204 raise Http404() # pylint: disable=raise-missing-from
205 raise
206 tree = self._by_idx(self.modules(), idx)
207 if not entry:
208 entry = tree[-1]
209 else:
210 tree = [{ 'type': 'all', 'children': self.modules() }]
211 exercises = []
212
213 def recursion(entry):
214 if (
215 entry['type'] == 'module' or ( # pylint: disable=too-many-boolean-expressions
216 entry['type'] == 'exercise' and
217 (category_id is None or entry['category_id'] == category_id) and
218 (not filter_for_assistant or entry['allow_assistant_viewing'])
219 )
220 ):
221 exercises.append(entry)
222 for child in entry['children']:
223 recursion(child)
224 recursion(tree[-1])
225 return entry, exercises
226
227 def _previous(self, idx, tree):
228 for entry in PreviousIterator(self.modules(), idx, tree, visited=True):
229 if self.is_listed(entry):
230 return entry
231 return None
232
233 def _next(self, idx, tree):
234 for entry in NextIterator(self.modules(), idx, tree, visited=True, enclosed=False):
235 if self.is_listed(entry):
236 return entry
237 return None
238
239 def _model_idx(self, model):
240 def find(index, search):
241 if search in index:
242 return index[search]
243 raise NoSuchContent()
244 entry_type = None
245 if isinstance(model, dict):
246 entry_type = model.get('type', None)
247 if entry_type == 'module':
248 return find(self.data['module_index'], model['id'])
249 if entry_type == 'exercise':
250 return find(self.data['exercise_index'], model['id'])
251 elif isinstance(model, CourseModule):
252 return find(self.data['module_index'], model.id)
253 elif isinstance(model, LearningObject):
254 return find(self.data['exercise_index'], model.id)
255 else:
256 raise NoSuchContent()
257
258 @classmethod
259 def _by_idx(cls, hierarchy, idx):
260 tree = []
261 for i in idx:
262 entry = hierarchy[i]
263 hierarchy = entry['children']
264 tree.append(entry)
265 return tree
266
267 @classmethod
268 def _add_by_difficulty(cls, to, difficulty, points):
269 if difficulty in to:
270 to[difficulty] += points
271 else:
272 to[difficulty] = points
273
274 @classmethod
275 def is_visible(cls, entry):
276 t = entry['type']
277 if t == 'exercise':
278 return (
279 entry.get('category_status') != LearningObjectCategory.STATUS.HIDDEN
280 and entry.get('module_status') != CourseModule.STATUS.HIDDEN
281 and not entry['status'] in (
282 LearningObject.STATUS.HIDDEN,
283 LearningObject.STATUS.ENROLLMENT,
284 LearningObject.STATUS.ENROLLMENT_EXTERNAL,
285 )
286 )
287 if t == 'module':
288 return entry['status'] != CourseModule.STATUS.HIDDEN
289 if t == 'category':
290 return not entry['status'] in (
291 LearningObjectCategory.STATUS.HIDDEN,
292 LearningObjectCategory.STATUS.NOTOTAL,
293 )
294 return False
295
296 @classmethod
297 def is_listed(cls, entry):
298 if not cls.is_visible(entry):
299 return False
300 t = entry['type']
301 if t == 'exercise':
302 return (
303 entry.get('category_status') != LearningObjectCategory.STATUS.HIDDEN
304 and entry.get('module_status') != CourseModule.STATUS.UNLISTED
305 and entry['status'] != LearningObject.STATUS.UNLISTED
306 )
307 if t == 'module':
308 return entry['status'] != CourseModule.STATUS.UNLISTED
309 if t == 'category':
310 return entry['status'] != LearningObjectCategory.STATUS.HIDDEN
311 return True
312
313 @classmethod
314 def is_in_maintenance(cls, entry):
315 t = entry['type']
316 if t == 'exercise':
317 return (
318 entry['module_status'] == CourseModule.STATUS.MAINTENANCE
319 or entry['status'] == LearningObject.STATUS.MAINTENANCE
320 )
321 if t == 'module':
322 return entry['status'] == CourseModule.STATUS.MAINTENANCE
323 return False
324
[end of exercise/cache/hierarchy.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/exercise/cache/hierarchy.py b/exercise/cache/hierarchy.py
--- a/exercise/cache/hierarchy.py
+++ b/exercise/cache/hierarchy.py
@@ -302,7 +302,10 @@
return (
entry.get('category_status') != LearningObjectCategory.STATUS.HIDDEN
and entry.get('module_status') != CourseModule.STATUS.UNLISTED
- and entry['status'] != LearningObject.STATUS.UNLISTED
+ and not entry['status'] in (
+ LearningObject.STATUS.UNLISTED,
+ LearningObject.STATUS.MAINTENANCE,
+ )
)
if t == 'module':
return entry['status'] != CourseModule.STATUS.UNLISTED
|
{"golden_diff": "diff --git a/exercise/cache/hierarchy.py b/exercise/cache/hierarchy.py\n--- a/exercise/cache/hierarchy.py\n+++ b/exercise/cache/hierarchy.py\n@@ -302,7 +302,10 @@\n return (\n entry.get('category_status') != LearningObjectCategory.STATUS.HIDDEN\n and entry.get('module_status') != CourseModule.STATUS.UNLISTED\n- and entry['status'] != LearningObject.STATUS.UNLISTED\n+ and not entry['status'] in (\n+ LearningObject.STATUS.UNLISTED,\n+ LearningObject.STATUS.MAINTENANCE,\n+ )\n )\n if t == 'module':\n return entry['status'] != CourseModule.STATUS.UNLISTED\n", "issue": "Hide exercises in the maintenance mode from the course table of contents\nExercises in maintenance mode (LearningObject.status maintenance) should be hidden from course table of contents.\r\n\r\nExercises with the status hidden are already hidden in the table of contents. The maintenance mode could be handled in the code in the same way.\r\n\n", "before_files": [{"content": "from django.http.response import Http404\nfrom course.models import CourseModule, LearningObjectCategory\nfrom ..models import LearningObject\n\n\nclass NoSuchContent(Exception):\n pass\n\n\nclass HierarchyIterator:\n # pylint: disable-next=too-many-arguments\n def __init__(self, children, idx=None, tree=None, visited=False, enclosed=True):\n if idx is None:\n self._default_start(children)\n else:\n self.idx = idx.copy()\n self.levels = [children]\n if tree and len(tree) > 1:\n for entry in tree[:-1]:\n self.levels.append(entry['children'])\n self.visited = visited\n self.enclose_begun = not enclosed\n self.enclose_ended = not enclosed\n\n def __iter__(self): # pylint: disable=non-iterator-returned\n return self\n\n\nclass NextIterator(HierarchyIterator):\n\n def _default_start(self, children):\n self.idx = [0]\n self.levels = [children]\n\n def __next__(self):\n if not self.enclose_begun:\n self.enclose_begun = True\n return {'type':'level','down':True}\n i = self.idx[-1]\n level = self.levels[-1]\n if not self.visited:\n if i < len(level):\n self.visited = True\n return level[i]\n else:\n children = level[i].get('children')\n if children:\n self.levels.append(children)\n self.idx.append(0)\n self.visited = False\n return {'type':'level','down':True}\n i += 1\n if i < len(level):\n self.idx[-1] = i\n return level[i]\n if len(self.idx) > 1:\n self.idx = self.idx[:-1]\n self.levels = self.levels[:-1]\n self.idx[-1] += 1\n self.visited = False\n return {'type':'level','up':True}\n if not self.enclose_ended:\n self.enclose_ended = True\n return {'type':'level','up':True}\n raise StopIteration()\n\n\nclass PreviousIterator(HierarchyIterator):\n\n def _default_start(self, children):\n self.idx = []\n self.levels = []\n self._goto_last(children)\n\n def _goto_last(self, children):\n level = children\n while level:\n i = len(level) - 1\n self.idx.append(i)\n self.levels.append(level)\n level = level[i].get('children')\n\n def __next__(self):\n i = self.idx[-1]\n level = self.levels[-1]\n if not self.visited:\n self.visited = True\n return level[i]\n if i > 0:\n i -= 1\n self.idx[-1] = i\n self._goto_last(level[i].get('children'))\n return self.levels[-1][self.idx[-1]]\n if len(self.idx) > 1:\n self.idx = self.idx[:-1]\n self.levels = self.levels[:-1]\n return self.levels[-1][self.idx[-1]]\n raise StopIteration()\n\n\nclass ContentMixin:\n\n def created(self):\n return self.data['created']\n\n def total(self):\n return self.data['total']\n\n def modules(self):\n return self.data['modules']\n\n def modules_flatted(self):\n for module in self.data['modules']:\n module['flatted'] = self.flat_module(module)\n return self.data['modules']\n\n def categories(self):\n categories = list(self.data['categories'].values())\n categories.sort(key=lambda entry: entry['name'])\n return categories\n\n def flat_module(self, module, enclosed=True):\n modules = self.modules()\n idx = self._model_idx(module)\n tree = self._by_idx(modules, idx)\n return NextIterator(tree[0]['children'], enclosed=enclosed)\n\n def flat_full(self):\n return NextIterator(self.modules(), enclosed=False)\n\n def begin(self):\n for entry in self.flat_full():\n if entry['type'] == 'exercise':\n return entry\n return None\n\n def find_path(self, module_id, path):\n paths = self.data['paths'].get(module_id, {})\n if path in paths:\n return paths[path]\n raise NoSuchContent()\n\n def find_number(self, number):\n hit = None\n search = self.modules()\n parts = number.split('.')\n for i in range(len(parts)):\n number = '.'.join(parts[0:i+1])\n for s in search:\n if s['number'] == number:\n hit = s\n search = hit['children']\n break\n else:\n raise NoSuchContent()\n return hit\n\n def find_category(self, category_id):\n categories = self.data['categories']\n if category_id in categories:\n return categories[category_id]\n raise NoSuchContent()\n\n def find(self, model):\n modules = self.modules()\n idx = self._model_idx(model)\n tree = self._by_idx(modules, idx)\n return (\n tree[-1],\n tree,\n self._previous(idx, tree),\n self._next(idx, tree),\n )\n\n def search_exercises(self, **kwargs):\n _, entries = self.search_entries(**kwargs)\n return [e for e in entries if e['type'] == 'exercise']\n\n def search_entries(self, number=None, category_id=None, module_id=None, # noqa: MC0001\n exercise_id=None, filter_for_assistant=False, best=False, # pylint: disable=unused-argument\n raise_404=True):\n entry = None\n if number:\n try:\n entry = self.find_number(number)\n if entry['type'] == 'module':\n module_id = entry['id']\n elif entry['type'] == 'exercise':\n exercise_id = entry['id']\n except NoSuchContent:\n if raise_404:\n raise Http404() # pylint: disable=raise-missing-from\n raise\n search = None\n if exercise_id is not None:\n search = { 'type': 'exercise', 'id': int(exercise_id) }\n elif module_id is not None:\n search = { 'type': 'module', 'id': int(module_id) }\n if search:\n try:\n idx = self._model_idx(search)\n except NoSuchContent:\n if raise_404:\n raise Http404() # pylint: disable=raise-missing-from\n raise\n tree = self._by_idx(self.modules(), idx)\n if not entry:\n entry = tree[-1]\n else:\n tree = [{ 'type': 'all', 'children': self.modules() }]\n exercises = []\n\n def recursion(entry):\n if (\n entry['type'] == 'module' or ( # pylint: disable=too-many-boolean-expressions\n entry['type'] == 'exercise' and\n (category_id is None or entry['category_id'] == category_id) and\n (not filter_for_assistant or entry['allow_assistant_viewing'])\n )\n ):\n exercises.append(entry)\n for child in entry['children']:\n recursion(child)\n recursion(tree[-1])\n return entry, exercises\n\n def _previous(self, idx, tree):\n for entry in PreviousIterator(self.modules(), idx, tree, visited=True):\n if self.is_listed(entry):\n return entry\n return None\n\n def _next(self, idx, tree):\n for entry in NextIterator(self.modules(), idx, tree, visited=True, enclosed=False):\n if self.is_listed(entry):\n return entry\n return None\n\n def _model_idx(self, model):\n def find(index, search):\n if search in index:\n return index[search]\n raise NoSuchContent()\n entry_type = None\n if isinstance(model, dict):\n entry_type = model.get('type', None)\n if entry_type == 'module':\n return find(self.data['module_index'], model['id'])\n if entry_type == 'exercise':\n return find(self.data['exercise_index'], model['id'])\n elif isinstance(model, CourseModule):\n return find(self.data['module_index'], model.id)\n elif isinstance(model, LearningObject):\n return find(self.data['exercise_index'], model.id)\n else:\n raise NoSuchContent()\n\n @classmethod\n def _by_idx(cls, hierarchy, idx):\n tree = []\n for i in idx:\n entry = hierarchy[i]\n hierarchy = entry['children']\n tree.append(entry)\n return tree\n\n @classmethod\n def _add_by_difficulty(cls, to, difficulty, points):\n if difficulty in to:\n to[difficulty] += points\n else:\n to[difficulty] = points\n\n @classmethod\n def is_visible(cls, entry):\n t = entry['type']\n if t == 'exercise':\n return (\n entry.get('category_status') != LearningObjectCategory.STATUS.HIDDEN\n and entry.get('module_status') != CourseModule.STATUS.HIDDEN\n and not entry['status'] in (\n LearningObject.STATUS.HIDDEN,\n LearningObject.STATUS.ENROLLMENT,\n LearningObject.STATUS.ENROLLMENT_EXTERNAL,\n )\n )\n if t == 'module':\n return entry['status'] != CourseModule.STATUS.HIDDEN\n if t == 'category':\n return not entry['status'] in (\n LearningObjectCategory.STATUS.HIDDEN,\n LearningObjectCategory.STATUS.NOTOTAL,\n )\n return False\n\n @classmethod\n def is_listed(cls, entry):\n if not cls.is_visible(entry):\n return False\n t = entry['type']\n if t == 'exercise':\n return (\n entry.get('category_status') != LearningObjectCategory.STATUS.HIDDEN\n and entry.get('module_status') != CourseModule.STATUS.UNLISTED\n and entry['status'] != LearningObject.STATUS.UNLISTED\n )\n if t == 'module':\n return entry['status'] != CourseModule.STATUS.UNLISTED\n if t == 'category':\n return entry['status'] != LearningObjectCategory.STATUS.HIDDEN\n return True\n\n @classmethod\n def is_in_maintenance(cls, entry):\n t = entry['type']\n if t == 'exercise':\n return (\n entry['module_status'] == CourseModule.STATUS.MAINTENANCE\n or entry['status'] == LearningObject.STATUS.MAINTENANCE\n )\n if t == 'module':\n return entry['status'] == CourseModule.STATUS.MAINTENANCE\n return False\n", "path": "exercise/cache/hierarchy.py"}]}
| 3,757 | 156 |
gh_patches_debug_37168
|
rasdani/github-patches
|
git_diff
|
conan-io__conan-center-index-2696
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
[request] sentry-native/0.4.1
### Package Details
* Package Name/Version: **sentry-native/0.4.1**
* Changelog: **https://github.com/getsentry/sentry-native/blob/0.4.1/CHANGELOG.md**
https://github.com/getsentry/sentry-native/tree/0.4.1
The above mentioned version is newly released by the upstream project and not yet available as a recipe. Please add this version.
Also, **please add windows support.**
</issue>
<code>
[start of recipes/sentry-native/all/conanfile.py]
1 import os
2 from conans import ConanFile, CMake, tools
3 from conans.errors import ConanInvalidConfiguration
4
5
6 class SentryNativeConan(ConanFile):
7 name = "sentry-native"
8 description = "The Sentry Native SDK is an error and crash reporting client for native applications,\n" \
9 "optimized for C and C++. Sentry allows to add tags,\n" \
10 "breadcrumbs and arbitrary custom context to enrich error reports."
11 url = "https://github.com/conan-io/conan-center-index"
12 homepage = "https://github.com/getsentry/sentry-native"
13 license = "MIT"
14 topics = ("conan", "breakpad", "crashpad",
15 "error-reporting", "crash-reporting")
16 exports_sources = ["CMakeLists.txt"]
17 generators = "cmake", "cmake_find_package"
18 settings = "os", "arch", "compiler", "build_type"
19 options = {
20 "shared": [True, False],
21 "fPIC": [True, False],
22 "backend": ["none", "inproc", "crashpad", "breakpad"],
23 "transport": ["none", "curl", "winhttp"],
24 }
25 default_options = {
26 "shared": False,
27 "fPIC": True,
28 "backend": "inproc",
29 "transport": "curl"
30 }
31
32 @property
33 def _source_subfolder(self):
34 return "source_subfolder"
35
36 _cmake = None
37
38 def requirements(self):
39 if self.options.transport == "curl":
40 self.requires("libcurl/7.68.0")
41
42 if self.options.backend == "crashpad":
43 raise ConanInvalidConfiguration("crashpad not available yet in CCI")
44 if self.options.backend == "breakpad":
45 raise ConanInvalidConfiguration("breakpad not available yet in CCI")
46
47 def config_options(self):
48 if self.settings.os == "Windows":
49 del self.options.fPIC
50
51 def source(self):
52 tools.get(**self.conan_data["sources"][self.version])
53 extracted_dir = self.name + "-" + self.version
54 os.rename(extracted_dir, self._source_subfolder)
55
56 def configure(self):
57 if self.options.backend == "inproc" and self.settings.os == "Windows":
58 raise ConanInvalidConfiguration("The in-process backend is not supported on Windows")
59
60 def _configure_cmake(self):
61 if self._cmake:
62 return self._cmake
63 self._cmake = CMake(self)
64 self._cmake.definitions["SENTRY_BACKEND"] = self.options.backend
65 self._cmake.definitions["SENTRY_ENABLE_INSTALL"] = True
66 self._cmake.definitions["SENTRY_TRANSPORT"] = self.options.transport
67 self._cmake.definitions["SENTRY_PIC"] = self.options.get_safe("fPIC", False)
68 self._cmake.configure()
69 return self._cmake
70
71 def build(self):
72 cmake = self._configure_cmake()
73 cmake.build()
74
75 def package(self):
76 self.copy("LICENSE", dst="licenses", src=self._source_subfolder)
77 cmake = self._configure_cmake()
78 cmake.install()
79 tools.rmdir(os.path.join(self.package_folder, "lib", "cmake"))
80
81 def package_info(self):
82 self.cpp_info.libs = ["sentry"]
83 if self.settings.os in ("Android", "Windows"):
84 self.cpp_info.exelinkflags= ["--build-id=sha1"]
85 self.cpp_info.sharedlinkflags = ["--build-id=sha1"]
86 if self.settings.os == "Linux":
87 self.cpp_info.system_libs = ["pthread", "dl"]
88 elif self.settings.os == "Windows":
89 self.cpp_info.system_libs = ["winhttp", "dbghelp", "pathcch"]
90
91 if not self.options.shared:
92 self.cpp_info.defines = ["SENTRY_BUILD_STATIC"]
93
[end of recipes/sentry-native/all/conanfile.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/recipes/sentry-native/all/conanfile.py b/recipes/sentry-native/all/conanfile.py
--- a/recipes/sentry-native/all/conanfile.py
+++ b/recipes/sentry-native/all/conanfile.py
@@ -1,4 +1,5 @@
import os
+import glob
from conans import ConanFile, CMake, tools
from conans.errors import ConanInvalidConfiguration
@@ -37,8 +38,8 @@
def requirements(self):
if self.options.transport == "curl":
- self.requires("libcurl/7.68.0")
-
+ self.requires("libcurl/7.71.0")
+
if self.options.backend == "crashpad":
raise ConanInvalidConfiguration("crashpad not available yet in CCI")
if self.options.backend == "breakpad":
@@ -54,7 +55,7 @@
os.rename(extracted_dir, self._source_subfolder)
def configure(self):
- if self.options.backend == "inproc" and self.settings.os == "Windows":
+ if self.options.backend == "inproc" and self.settings.os == "Windows" and tools.Version(self.version) < "0.4":
raise ConanInvalidConfiguration("The in-process backend is not supported on Windows")
def _configure_cmake(self):
@@ -77,16 +78,18 @@
cmake = self._configure_cmake()
cmake.install()
tools.rmdir(os.path.join(self.package_folder, "lib", "cmake"))
+ for pdb in glob.glob(os.path.join(self.package_folder, "bin", "*.pdb")):
+ os.unlink(pdb)
def package_info(self):
self.cpp_info.libs = ["sentry"]
- if self.settings.os in ("Android", "Windows"):
- self.cpp_info.exelinkflags= ["--build-id=sha1"]
- self.cpp_info.sharedlinkflags = ["--build-id=sha1"]
+ if self.settings.os in ("Android", "Linux"):
+ self.cpp_info.exelinkflags = ["-Wl,-E,--build-id=sha1"]
+ self.cpp_info.sharedlinkflags = ["-Wl,-E,--build-id=sha1"]
if self.settings.os == "Linux":
self.cpp_info.system_libs = ["pthread", "dl"]
elif self.settings.os == "Windows":
- self.cpp_info.system_libs = ["winhttp", "dbghelp", "pathcch"]
+ self.cpp_info.system_libs = ["winhttp", "dbghelp", "pathcch", "shlwapi"]
if not self.options.shared:
self.cpp_info.defines = ["SENTRY_BUILD_STATIC"]
|
{"golden_diff": "diff --git a/recipes/sentry-native/all/conanfile.py b/recipes/sentry-native/all/conanfile.py\n--- a/recipes/sentry-native/all/conanfile.py\n+++ b/recipes/sentry-native/all/conanfile.py\n@@ -1,4 +1,5 @@\n import os\n+import glob\n from conans import ConanFile, CMake, tools\n from conans.errors import ConanInvalidConfiguration\n \n@@ -37,8 +38,8 @@\n \n def requirements(self):\n if self.options.transport == \"curl\":\n- self.requires(\"libcurl/7.68.0\")\n- \n+ self.requires(\"libcurl/7.71.0\")\n+\n if self.options.backend == \"crashpad\":\n raise ConanInvalidConfiguration(\"crashpad not available yet in CCI\")\n if self.options.backend == \"breakpad\":\n@@ -54,7 +55,7 @@\n os.rename(extracted_dir, self._source_subfolder)\n \n def configure(self):\n- if self.options.backend == \"inproc\" and self.settings.os == \"Windows\":\n+ if self.options.backend == \"inproc\" and self.settings.os == \"Windows\" and tools.Version(self.version) < \"0.4\":\n raise ConanInvalidConfiguration(\"The in-process backend is not supported on Windows\")\n \n def _configure_cmake(self):\n@@ -77,16 +78,18 @@\n cmake = self._configure_cmake()\n cmake.install()\n tools.rmdir(os.path.join(self.package_folder, \"lib\", \"cmake\"))\n+ for pdb in glob.glob(os.path.join(self.package_folder, \"bin\", \"*.pdb\")):\n+ os.unlink(pdb)\n \n def package_info(self):\n self.cpp_info.libs = [\"sentry\"]\n- if self.settings.os in (\"Android\", \"Windows\"):\n- self.cpp_info.exelinkflags= [\"--build-id=sha1\"]\n- self.cpp_info.sharedlinkflags = [\"--build-id=sha1\"]\n+ if self.settings.os in (\"Android\", \"Linux\"):\n+ self.cpp_info.exelinkflags = [\"-Wl,-E,--build-id=sha1\"]\n+ self.cpp_info.sharedlinkflags = [\"-Wl,-E,--build-id=sha1\"]\n if self.settings.os == \"Linux\":\n self.cpp_info.system_libs = [\"pthread\", \"dl\"]\n elif self.settings.os == \"Windows\":\n- self.cpp_info.system_libs = [\"winhttp\", \"dbghelp\", \"pathcch\"]\n+ self.cpp_info.system_libs = [\"winhttp\", \"dbghelp\", \"pathcch\", \"shlwapi\"]\n \n if not self.options.shared:\n self.cpp_info.defines = [\"SENTRY_BUILD_STATIC\"]\n", "issue": "[request] sentry-native/0.4.1\n### Package Details\r\n * Package Name/Version: **sentry-native/0.4.1**\r\n * Changelog: **https://github.com/getsentry/sentry-native/blob/0.4.1/CHANGELOG.md**\r\n\r\nhttps://github.com/getsentry/sentry-native/tree/0.4.1\r\n\r\nThe above mentioned version is newly released by the upstream project and not yet available as a recipe. Please add this version.\r\nAlso, **please add windows support.**\n", "before_files": [{"content": "import os\nfrom conans import ConanFile, CMake, tools\nfrom conans.errors import ConanInvalidConfiguration\n\n\nclass SentryNativeConan(ConanFile):\n name = \"sentry-native\"\n description = \"The Sentry Native SDK is an error and crash reporting client for native applications,\\n\" \\\n \"optimized for C and C++. Sentry allows to add tags,\\n\" \\\n \"breadcrumbs and arbitrary custom context to enrich error reports.\"\n url = \"https://github.com/conan-io/conan-center-index\"\n homepage = \"https://github.com/getsentry/sentry-native\"\n license = \"MIT\"\n topics = (\"conan\", \"breakpad\", \"crashpad\",\n \"error-reporting\", \"crash-reporting\")\n exports_sources = [\"CMakeLists.txt\"]\n generators = \"cmake\", \"cmake_find_package\"\n settings = \"os\", \"arch\", \"compiler\", \"build_type\"\n options = {\n \"shared\": [True, False],\n \"fPIC\": [True, False],\n \"backend\": [\"none\", \"inproc\", \"crashpad\", \"breakpad\"],\n \"transport\": [\"none\", \"curl\", \"winhttp\"],\n }\n default_options = {\n \"shared\": False,\n \"fPIC\": True,\n \"backend\": \"inproc\",\n \"transport\": \"curl\"\n }\n\n @property\n def _source_subfolder(self):\n return \"source_subfolder\"\n\n _cmake = None\n\n def requirements(self):\n if self.options.transport == \"curl\":\n self.requires(\"libcurl/7.68.0\")\n \n if self.options.backend == \"crashpad\":\n raise ConanInvalidConfiguration(\"crashpad not available yet in CCI\")\n if self.options.backend == \"breakpad\":\n raise ConanInvalidConfiguration(\"breakpad not available yet in CCI\")\n\n def config_options(self):\n if self.settings.os == \"Windows\":\n del self.options.fPIC\n\n def source(self):\n tools.get(**self.conan_data[\"sources\"][self.version])\n extracted_dir = self.name + \"-\" + self.version\n os.rename(extracted_dir, self._source_subfolder)\n\n def configure(self):\n if self.options.backend == \"inproc\" and self.settings.os == \"Windows\":\n raise ConanInvalidConfiguration(\"The in-process backend is not supported on Windows\")\n\n def _configure_cmake(self):\n if self._cmake:\n return self._cmake\n self._cmake = CMake(self)\n self._cmake.definitions[\"SENTRY_BACKEND\"] = self.options.backend\n self._cmake.definitions[\"SENTRY_ENABLE_INSTALL\"] = True\n self._cmake.definitions[\"SENTRY_TRANSPORT\"] = self.options.transport\n self._cmake.definitions[\"SENTRY_PIC\"] = self.options.get_safe(\"fPIC\", False)\n self._cmake.configure()\n return self._cmake\n\n def build(self):\n cmake = self._configure_cmake()\n cmake.build()\n\n def package(self):\n self.copy(\"LICENSE\", dst=\"licenses\", src=self._source_subfolder)\n cmake = self._configure_cmake()\n cmake.install()\n tools.rmdir(os.path.join(self.package_folder, \"lib\", \"cmake\"))\n\n def package_info(self):\n self.cpp_info.libs = [\"sentry\"]\n if self.settings.os in (\"Android\", \"Windows\"):\n self.cpp_info.exelinkflags= [\"--build-id=sha1\"]\n self.cpp_info.sharedlinkflags = [\"--build-id=sha1\"]\n if self.settings.os == \"Linux\":\n self.cpp_info.system_libs = [\"pthread\", \"dl\"]\n elif self.settings.os == \"Windows\":\n self.cpp_info.system_libs = [\"winhttp\", \"dbghelp\", \"pathcch\"]\n\n if not self.options.shared:\n self.cpp_info.defines = [\"SENTRY_BUILD_STATIC\"]\n", "path": "recipes/sentry-native/all/conanfile.py"}]}
| 1,678 | 595 |
gh_patches_debug_4349
|
rasdani/github-patches
|
git_diff
|
dbt-labs__dbt-core-4890
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
[CT-381] [HOTFIX] Homebrew Incident Resolution
### What Happened
dbt-core depends on the dbt-extractor package, and the dbt-extractor package depends on tree-sitter-jinja2. dbt-extractor specifies tree-sitter-jinja2 via a github link using the git protocol. Github security rules changed to require this link to use https which caused cargo to fail to build the dbt-extractor.
### Who Is Affected
Everyone attempting to build dbt-core from source after the github security rules took affect. This primarily affects homebrew users since homebrew builds dbt from source locally.
### Solution:
- release new dbt-extractor (0.4.1). The fix is already in main
- dbt-labs/dbt-extractor#51
- release new dbt-core patch from branch [1.0.4-hotfix](https://github.com/dbt-labs/dbt-core/tree/1.0.4-hotfix) which depends on this new version and accepts all future patch releases so we can skip this step in the future. This branch is only the 3 necessary commits ahead of 1.0.3 to fix this incident.
- main: #4890
- backport is directly on branch [1.0.4-hotfix](https://github.com/dbt-labs/dbt-core/tree/1.0.4-hotfix) because of complications with running the bump-version workflow for a hotfix branch.
Getting the release out has been delayed due to complications with github actions due to an [ongoing GitHub incident](https://www.githubstatus.com/incidents/dcnvr6zym66r).
</issue>
<code>
[start of core/setup.py]
1 #!/usr/bin/env python
2 import os
3 import sys
4
5 if sys.version_info < (3, 7, 2):
6 print("Error: dbt does not support this version of Python.")
7 print("Please upgrade to Python 3.7.2 or higher.")
8 sys.exit(1)
9
10
11 from setuptools import setup
12
13 try:
14 from setuptools import find_namespace_packages
15 except ImportError:
16 # the user has a downlevel version of setuptools.
17 print("Error: dbt requires setuptools v40.1.0 or higher.")
18 print('Please upgrade setuptools with "pip install --upgrade setuptools" ' "and try again")
19 sys.exit(1)
20
21
22 this_directory = os.path.abspath(os.path.dirname(__file__))
23 with open(os.path.join(this_directory, "README.md")) as f:
24 long_description = f.read()
25
26
27 package_name = "dbt-core"
28 package_version = "1.0.1"
29 description = """With dbt, data analysts and engineers can build analytics \
30 the way engineers build applications."""
31
32
33 setup(
34 name=package_name,
35 version=package_version,
36 description=description,
37 long_description=long_description,
38 long_description_content_type="text/markdown",
39 author="dbt Labs",
40 author_email="[email protected]",
41 url="https://github.com/dbt-labs/dbt-core",
42 packages=find_namespace_packages(include=["dbt", "dbt.*"]),
43 include_package_data=True,
44 test_suite="test",
45 entry_points={
46 "console_scripts": [
47 "dbt = dbt.main:main",
48 ],
49 },
50 scripts=[
51 "scripts/dbt",
52 ],
53 install_requires=[
54 "Jinja2==2.11.3",
55 "MarkupSafe==2.0.1",
56 "agate>=1.6,<1.6.4",
57 "click>=7.0,<9",
58 "colorama>=0.3.9,<0.4.5",
59 "hologram==0.0.14",
60 "isodate>=0.6,<0.7",
61 "logbook>=1.5,<1.6",
62 "mashumaro==2.9",
63 "minimal-snowplow-tracker==0.0.2",
64 "networkx>=2.3,<3",
65 "packaging>=20.9,<22.0",
66 "sqlparse>=0.2.3,<0.5",
67 "dbt-extractor==0.4.0",
68 "typing-extensions>=3.7.4,<4.2",
69 "werkzeug>=1,<3",
70 # the following are all to match snowflake-connector-python
71 "requests<3.0.0",
72 "idna>=2.5,<4",
73 "cffi>=1.9,<2.0.0",
74 ],
75 zip_safe=False,
76 classifiers=[
77 "Development Status :: 5 - Production/Stable",
78 "License :: OSI Approved :: Apache Software License",
79 "Operating System :: Microsoft :: Windows",
80 "Operating System :: MacOS :: MacOS X",
81 "Operating System :: POSIX :: Linux",
82 "Programming Language :: Python :: 3.7",
83 "Programming Language :: Python :: 3.8",
84 "Programming Language :: Python :: 3.9",
85 ],
86 python_requires=">=3.7.2",
87 )
88
[end of core/setup.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/core/setup.py b/core/setup.py
--- a/core/setup.py
+++ b/core/setup.py
@@ -64,7 +64,7 @@
"networkx>=2.3,<3",
"packaging>=20.9,<22.0",
"sqlparse>=0.2.3,<0.5",
- "dbt-extractor==0.4.0",
+ "dbt-extractor~=0.4.1",
"typing-extensions>=3.7.4,<4.2",
"werkzeug>=1,<3",
# the following are all to match snowflake-connector-python
|
{"golden_diff": "diff --git a/core/setup.py b/core/setup.py\n--- a/core/setup.py\n+++ b/core/setup.py\n@@ -64,7 +64,7 @@\n \"networkx>=2.3,<3\",\n \"packaging>=20.9,<22.0\",\n \"sqlparse>=0.2.3,<0.5\",\n- \"dbt-extractor==0.4.0\",\n+ \"dbt-extractor~=0.4.1\",\n \"typing-extensions>=3.7.4,<4.2\",\n \"werkzeug>=1,<3\",\n # the following are all to match snowflake-connector-python\n", "issue": "[CT-381] [HOTFIX] Homebrew Incident Resolution\n### What Happened\r\n\r\ndbt-core depends on the dbt-extractor package, and the dbt-extractor package depends on tree-sitter-jinja2. dbt-extractor specifies tree-sitter-jinja2 via a github link using the git protocol. Github security rules changed to require this link to use https which caused cargo to fail to build the dbt-extractor.\r\n\r\n### Who Is Affected\r\n\r\nEveryone attempting to build dbt-core from source after the github security rules took affect. This primarily affects homebrew users since homebrew builds dbt from source locally.\r\n\r\n### Solution:\r\n- release new dbt-extractor (0.4.1). The fix is already in main\r\n - dbt-labs/dbt-extractor#51\r\n- release new dbt-core patch from branch [1.0.4-hotfix](https://github.com/dbt-labs/dbt-core/tree/1.0.4-hotfix) which depends on this new version and accepts all future patch releases so we can skip this step in the future. This branch is only the 3 necessary commits ahead of 1.0.3 to fix this incident.\r\n - main: #4890\r\n - backport is directly on branch [1.0.4-hotfix](https://github.com/dbt-labs/dbt-core/tree/1.0.4-hotfix) because of complications with running the bump-version workflow for a hotfix branch.\r\n \r\nGetting the release out has been delayed due to complications with github actions due to an [ongoing GitHub incident](https://www.githubstatus.com/incidents/dcnvr6zym66r).\r\n \n", "before_files": [{"content": "#!/usr/bin/env python\nimport os\nimport sys\n\nif sys.version_info < (3, 7, 2):\n print(\"Error: dbt does not support this version of Python.\")\n print(\"Please upgrade to Python 3.7.2 or higher.\")\n sys.exit(1)\n\n\nfrom setuptools import setup\n\ntry:\n from setuptools import find_namespace_packages\nexcept ImportError:\n # the user has a downlevel version of setuptools.\n print(\"Error: dbt requires setuptools v40.1.0 or higher.\")\n print('Please upgrade setuptools with \"pip install --upgrade setuptools\" ' \"and try again\")\n sys.exit(1)\n\n\nthis_directory = os.path.abspath(os.path.dirname(__file__))\nwith open(os.path.join(this_directory, \"README.md\")) as f:\n long_description = f.read()\n\n\npackage_name = \"dbt-core\"\npackage_version = \"1.0.1\"\ndescription = \"\"\"With dbt, data analysts and engineers can build analytics \\\nthe way engineers build applications.\"\"\"\n\n\nsetup(\n name=package_name,\n version=package_version,\n description=description,\n long_description=long_description,\n long_description_content_type=\"text/markdown\",\n author=\"dbt Labs\",\n author_email=\"[email protected]\",\n url=\"https://github.com/dbt-labs/dbt-core\",\n packages=find_namespace_packages(include=[\"dbt\", \"dbt.*\"]),\n include_package_data=True,\n test_suite=\"test\",\n entry_points={\n \"console_scripts\": [\n \"dbt = dbt.main:main\",\n ],\n },\n scripts=[\n \"scripts/dbt\",\n ],\n install_requires=[\n \"Jinja2==2.11.3\",\n \"MarkupSafe==2.0.1\",\n \"agate>=1.6,<1.6.4\",\n \"click>=7.0,<9\",\n \"colorama>=0.3.9,<0.4.5\",\n \"hologram==0.0.14\",\n \"isodate>=0.6,<0.7\",\n \"logbook>=1.5,<1.6\",\n \"mashumaro==2.9\",\n \"minimal-snowplow-tracker==0.0.2\",\n \"networkx>=2.3,<3\",\n \"packaging>=20.9,<22.0\",\n \"sqlparse>=0.2.3,<0.5\",\n \"dbt-extractor==0.4.0\",\n \"typing-extensions>=3.7.4,<4.2\",\n \"werkzeug>=1,<3\",\n # the following are all to match snowflake-connector-python\n \"requests<3.0.0\",\n \"idna>=2.5,<4\",\n \"cffi>=1.9,<2.0.0\",\n ],\n zip_safe=False,\n classifiers=[\n \"Development Status :: 5 - Production/Stable\",\n \"License :: OSI Approved :: Apache Software License\",\n \"Operating System :: Microsoft :: Windows\",\n \"Operating System :: MacOS :: MacOS X\",\n \"Operating System :: POSIX :: Linux\",\n \"Programming Language :: Python :: 3.7\",\n \"Programming Language :: Python :: 3.8\",\n \"Programming Language :: Python :: 3.9\",\n ],\n python_requires=\">=3.7.2\",\n)\n", "path": "core/setup.py"}]}
| 1,785 | 144 |
gh_patches_debug_36247
|
rasdani/github-patches
|
git_diff
|
TabbycatDebate__tabbycat-840
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Include QR code in private URL distribution pages
It is a bit odd to hand people a physical item directing them to the internet, but typing a random string is not really user-friendly. QR codes that directs to the URL could be added to the URL sheets for ease of use. The type of barcode that is in use can't encode as long a string, but a QR code is recognized to be scanned to lead to a website and seems to be well supported by phones.
This may require another dependency to generate these codes; `jsBarcode` can't generate them.
</issue>
<code>
[start of tabbycat/printing/views.py]
1 import json
2
3 from django.db.models import Q
4 from django.utils.translation import gettext as _
5 from django.views.generic.base import TemplateView
6
7 from adjfeedback.models import AdjudicatorFeedbackQuestion
8 from adjfeedback.utils import expected_feedback_targets
9 from checkins.models import DebateIdentifier
10 from checkins.utils import create_identifiers
11 from draw.models import Debate, DebateTeam
12 from options.utils import use_team_code_names
13 from participants.models import Person
14 from results.utils import side_and_position_names
15 from tournaments.mixins import (CurrentRoundMixin, OptionalAssistantTournamentPageMixin,
16 RoundMixin, TournamentMixin)
17 from tournaments.models import Tournament
18 from utils.mixins import AdministratorMixin
19 from venues.models import VenueCategory
20
21
22 class MasterSheetsListView(AdministratorMixin, RoundMixin, TemplateView):
23 template_name = 'division_sheets_list.html'
24
25 def get_context_data(self, **kwargs):
26 kwargs['standings'] = VenueCategory.objects.all()
27 kwargs['venue_categories'] = VenueCategory.objects.all()
28 return super().get_context_data(**kwargs)
29
30
31 class MasterSheetsView(AdministratorMixin, RoundMixin, TemplateView):
32 template_name = 'master_sheets_view.html'
33
34 def get_context_data(self, **kwargs):
35 venue_category_id = self.kwargs['venue_category_id']
36 base_venue_category = VenueCategory.objects.get(id=venue_category_id)
37 active_tournaments = Tournament.objects.filter(active=True)
38 for tournament in list(active_tournaments):
39 tournament.debates = Debate.objects.select_related(
40 'division', 'division__venue_category', 'round',
41 'round__tournament').filter(
42 # All Debates, with a matching round, at the same venue category name
43 round__seq=self.round.seq,
44 round__tournament=tournament,
45 # Hack - remove when venue category are unified
46 division__venue_category__name=base_venue_category.name
47 ).order_by('round', 'division__venue_category__name', 'division')
48
49 kwargs['base_venue_category'] = base_venue_category
50 kwargs['active_tournaments'] = active_tournaments
51 return super().get_context_data(**kwargs)
52
53
54 class RoomSheetsView(AdministratorMixin, RoundMixin, TemplateView):
55 template_name = 'room_sheets_view.html'
56
57 def get_context_data(self, **kwargs):
58 venue_category_id = self.kwargs['venue_category_id']
59 base_venue_category = VenueCategory.objects.get(id=venue_category_id)
60 venues_list = []
61
62 # Get a unique list of venue names (avoid getting duplicates across tournaments)
63 for venue in set(base_venue_category.venues.order_by('name').values_list('name', flat=True)):
64 venues_list.append({'name': venue, 'debates': []})
65 # All Debates, with a matching round, at the same venue category
66 venues_list[-1]['debates'] = Debate.objects.filter(
67 round__seq=self.round.seq, venue__name=venue).order_by('round__tournament__seq').all()
68 print(venues_list[-1])
69
70 kwargs['base_venue_category'] = base_venue_category
71 kwargs['venues'] = venues_list
72 return super().get_context_data(**kwargs)
73
74
75 class BasePrintFeedbackFormsView(RoundMixin, TemplateView):
76
77 template_name = 'feedback_list.html'
78
79 def add_defaults(self):
80 default_questions = []
81
82 if self.tournament.pref('feedback_introduction'):
83 default_scale_info = AdjudicatorFeedbackQuestion(
84 text=self.tournament.pref('feedback_introduction'), seq=0,
85 answer_type='comment', # Custom type just for print display
86 required=True, from_team=True, from_adj=True
87 )
88 default_questions.append(default_scale_info.serialize())
89
90 default_scale_question = AdjudicatorFeedbackQuestion(
91 text=_("Overall Score"), seq=0,
92 answer_type=AdjudicatorFeedbackQuestion.ANSWER_TYPE_INTEGER_SCALE,
93 required=True, from_team=True, from_adj=True,
94 min_value=self.tournament.pref('adj_min_score'),
95 max_value=self.tournament.pref('adj_max_score')
96 )
97 default_questions.append(default_scale_question.serialize())
98
99 return default_questions
100
101 def questions_dict(self):
102 questions = self.add_defaults()
103 for question in self.tournament.adj_feedback_questions:
104 questions.append(question.serialize())
105
106 return questions
107
108 def construct_info(self, venue, source, source_p, target, target_p):
109 if hasattr(source, 'name'):
110 source_n = source.name
111 elif use_team_code_names(self.tournament, False):
112 source_n = source.code_name
113 else:
114 source_n = source.short_name
115
116 return {
117 'venue': venue.serialize() if venue else '',
118 'authorInstitution': source.institution.code if source.institution else _("Unaffiliated"),
119 'author': source_n, 'authorPosition': source_p,
120 'target': target.name, 'targetPosition': target_p,
121 }
122
123 def get_team_feedbacks(self, debate, team):
124 if len(debate.adjudicators) is 0:
125 return []
126
127 team_paths = self.tournament.pref('feedback_from_teams')
128 ballots = []
129
130 if team_paths == 'orallist' and debate.adjudicators.chair:
131 ballots.append(self.construct_info(debate.venue, team, _("Team"),
132 debate.adjudicators.chair, ""))
133 elif team_paths == 'all-adjs':
134 for target in debate.adjudicators.all():
135 ballots.append(self.construct_info(debate.venue, team, _("Team"), target, ""))
136
137 return ballots
138
139 def get_adj_feedbacks(self, debate):
140 adj_paths = self.tournament.pref('feedback_paths')
141 ballots = []
142
143 debateadjs = debate.debateadjudicator_set.all()
144 for debateadj in debateadjs:
145 sadj = debateadj.adjudicator
146 spos = debate.adjudicators.get_position(sadj)
147 targets = expected_feedback_targets(debateadj, feedback_paths=adj_paths, debate=debate)
148 for tadj, tpos in targets:
149 ballots.append(self.construct_info(debate.venue, sadj, spos, tadj, tpos))
150
151 return ballots
152
153 def get_context_data(self, **kwargs):
154 draw = self.round.debate_set_with_prefetches(institutions=True)
155 draw = sorted(draw, key=lambda d: d.venue.display_name if d.venue else "")
156
157 ballots = []
158 for debate in draw:
159 for team in debate.teams:
160 ballots.extend(self.get_team_feedbacks(debate, team))
161 ballots.extend(self.get_adj_feedbacks(debate))
162
163 kwargs['ballots'] = json.dumps(ballots)
164 kwargs['questions'] = json.dumps(self.questions_dict())
165
166 kwargs['team_questions_exist'] = self.tournament.adjudicatorfeedbackquestion_set.filter(from_team=True).exists()
167 kwargs['adj_questions_exist'] = self.tournament.adjudicatorfeedbackquestion_set.filter(from_adj=True).exists()
168
169 return super().get_context_data(**kwargs)
170
171
172 class AdminPrintFeedbackFormsView(AdministratorMixin, BasePrintFeedbackFormsView):
173 pass
174
175
176 class AssistantPrintFeedbackFormsView(CurrentRoundMixin, OptionalAssistantTournamentPageMixin, BasePrintFeedbackFormsView):
177 assistant_page_permissions = ['all_areas', 'results_draw']
178
179
180 class BasePrintScoresheetsView(RoundMixin, TemplateView):
181
182 template_name = 'scoresheet_list.html'
183
184 def get_ballots_dicts(self):
185 draw = self.round.debate_set_with_prefetches()
186
187 # Create the DebateIdentifiers for the ballots if needed
188 create_identifiers(DebateIdentifier, draw)
189 identifiers = DebateIdentifier.objects.values('debate_id', 'barcode')
190
191 draw = sorted(draw, key=lambda d: d.venue.display_name if d.venue else "")
192 ballots_dicts = []
193
194 # Force translation before JSON serialization
195 sides_and_positions = [(side, [str(pos) for pos in positions])
196 for side, positions in side_and_position_names(self.tournament)]
197
198 for debate in draw:
199 debate_dict = {}
200
201 if debate.venue:
202 debate_dict['venue'] = {'display_name': debate.venue.display_name}
203 else:
204 debate_dict['venue'] = None
205
206 debate_dict['barcode'] = next((i['barcode'] for i in identifiers if i['debate_id'] == debate.id), None)
207
208 debate_dict['debateTeams'] = []
209 for side, (side_name, positions) in zip(self.tournament.sides, sides_and_positions):
210 dt_dict = {'side_name': side_name, 'positions': positions}
211 try:
212 team = debate.get_team(side)
213 dt_dict['team'] = {
214 'short_name': team.short_name,
215 'code_name': team.code_name,
216 'speakers': [{'name': s.name} for s in team.speakers],
217 }
218 except DebateTeam.DoesNotExist:
219 dt_dict['team'] = None
220 debate_dict['debateTeams'].append(dt_dict)
221
222 debate_dict['debateAdjudicators'] = []
223 for adj, pos in debate.adjudicators.with_positions():
224 da_dict = {'position': pos}
225 da_dict['adjudicator'] = {
226 'name': adj.name,
227 'institution': {'code': adj.institution.code if adj.institution else _("Unaffiliated")},
228 }
229 debate_dict['debateAdjudicators'].append(da_dict)
230
231 if self.round.ballots_per_debate == 'per-adj':
232 authors = list(debate.adjudicators.voting_with_positions())
233 else:
234 authors = [(debate.adjudicators.chair, debate.adjudicators.POSITION_CHAIR)]
235
236 blank_author_dict = {
237 'author': "_______________________________________________",
238 'authorInstitution': "",
239 'authorPosition': "",
240 }
241
242 # Add a ballot for each author
243 for author, pos in authors:
244 if author:
245 ballot_dict = {
246 'author': author.name,
247 'authorInstitution': author.institution.code if author.institution else _("Unaffiliated"),
248 'authorPosition': pos,
249 }
250 else:
251 ballot_dict = blank_author_dict
252
253 ballot_dict.update(debate_dict)
254 ballots_dicts.append(ballot_dict)
255
256 if len(authors) == 0:
257 ballot_dict = blank_author_dict
258 ballot_dict.update(debate_dict)
259 ballots_dicts.append(ballot_dict)
260
261 return ballots_dicts
262
263 def get_context_data(self, **kwargs):
264 kwargs['ballots'] = json.dumps(self.get_ballots_dicts())
265 motions = self.round.motion_set.order_by('seq')
266 kwargs['motions'] = json.dumps([{'seq': m.seq, 'text': m.text} for m in motions])
267 kwargs['use_team_code_names'] = use_team_code_names(self.tournament, False)
268 return super().get_context_data(**kwargs)
269
270
271 class AdminPrintScoresheetsView(AdministratorMixin, BasePrintScoresheetsView):
272 pass
273
274
275 class AssistantPrintScoresheetsView(CurrentRoundMixin, OptionalAssistantTournamentPageMixin, BasePrintScoresheetsView):
276 assistant_page_permissions = ['all_areas']
277
278
279 class PrintableRandomisedURLs(TournamentMixin, AdministratorMixin, TemplateView):
280
281 template_name = 'randomised_url_sheets.html'
282
283 def get_context_data(self, **kwargs):
284 kwargs['tournament_slug'] = self.tournament.slug
285
286 if not self.tournament.pref('share_adjs'):
287 kwargs['parts'] = self.tournament.participants.filter(
288 url_key__isnull=False).select_related(
289 'speaker', 'speaker__team', 'adjudicator__institution',
290 'adjudicator')
291 else:
292 kwargs['parts'] = Person.objects.filter(Q(speaker__team__tournament=self.tournament) | Q(adjudicator__tournament__isnull=True) & Q(url_key__isnull=False))
293
294 kwargs['exists'] = self.tournament.participants.filter(url_key__isnull=False).exists()
295
296 return super().get_context_data(**kwargs)
297
[end of tabbycat/printing/views.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/tabbycat/printing/views.py b/tabbycat/printing/views.py
--- a/tabbycat/printing/views.py
+++ b/tabbycat/printing/views.py
@@ -1,5 +1,8 @@
import json
+import qrcode
+from qrcode.image import svg
+
from django.db.models import Q
from django.utils.translation import gettext as _
from django.views.generic.base import TemplateView
@@ -15,6 +18,7 @@
from tournaments.mixins import (CurrentRoundMixin, OptionalAssistantTournamentPageMixin,
RoundMixin, TournamentMixin)
from tournaments.models import Tournament
+from utils.misc import reverse_tournament
from utils.mixins import AdministratorMixin
from venues.models import VenueCategory
@@ -280,16 +284,28 @@
template_name = 'randomised_url_sheets.html'
+ def add_urls(self, participants):
+ for participant in participants:
+ url = reverse_tournament('privateurls-person-index', self.tournament, kwargs={'url_key': participant['url_key']})
+ abs_url = self.request.build_absolute_uri(url)
+ qr_code = qrcode.make(abs_url, image_factory=svg.SvgPathImage)
+
+ participant['url'] = abs_url
+ participant['qr'] = ' '.join(qr_code._generate_subpaths())
+
+ return participants
+
def get_context_data(self, **kwargs):
- kwargs['tournament_slug'] = self.tournament.slug
if not self.tournament.pref('share_adjs'):
- kwargs['parts'] = self.tournament.participants.filter(
- url_key__isnull=False).select_related(
- 'speaker', 'speaker__team', 'adjudicator__institution',
- 'adjudicator')
+ participants = self.tournament.participants.filter(url_key__isnull=False)
else:
- kwargs['parts'] = Person.objects.filter(Q(speaker__team__tournament=self.tournament) | Q(adjudicator__tournament__isnull=True) & Q(url_key__isnull=False))
+ participants = Person.objects.filter(
+ Q(speaker__team__tournament=self.tournament) | Q(adjudicator__tournament__isnull=True) & Q(url_key__isnull=False))
+
+ participants_array = list(participants.select_related('speaker', 'speaker__team', 'adjudicator__institution', 'adjudicator')
+ .values('name', 'speaker__team__short_name', 'adjudicator__institution__code', 'url_key'))
+ kwargs['parts'] = self.add_urls(participants_array)
kwargs['exists'] = self.tournament.participants.filter(url_key__isnull=False).exists()
|
{"golden_diff": "diff --git a/tabbycat/printing/views.py b/tabbycat/printing/views.py\n--- a/tabbycat/printing/views.py\n+++ b/tabbycat/printing/views.py\n@@ -1,5 +1,8 @@\n import json\n \n+import qrcode\n+from qrcode.image import svg\n+\n from django.db.models import Q\n from django.utils.translation import gettext as _\n from django.views.generic.base import TemplateView\n@@ -15,6 +18,7 @@\n from tournaments.mixins import (CurrentRoundMixin, OptionalAssistantTournamentPageMixin,\n RoundMixin, TournamentMixin)\n from tournaments.models import Tournament\n+from utils.misc import reverse_tournament\n from utils.mixins import AdministratorMixin\n from venues.models import VenueCategory\n \n@@ -280,16 +284,28 @@\n \n template_name = 'randomised_url_sheets.html'\n \n+ def add_urls(self, participants):\n+ for participant in participants:\n+ url = reverse_tournament('privateurls-person-index', self.tournament, kwargs={'url_key': participant['url_key']})\n+ abs_url = self.request.build_absolute_uri(url)\n+ qr_code = qrcode.make(abs_url, image_factory=svg.SvgPathImage)\n+\n+ participant['url'] = abs_url\n+ participant['qr'] = ' '.join(qr_code._generate_subpaths())\n+\n+ return participants\n+\n def get_context_data(self, **kwargs):\n- kwargs['tournament_slug'] = self.tournament.slug\n \n if not self.tournament.pref('share_adjs'):\n- kwargs['parts'] = self.tournament.participants.filter(\n- url_key__isnull=False).select_related(\n- 'speaker', 'speaker__team', 'adjudicator__institution',\n- 'adjudicator')\n+ participants = self.tournament.participants.filter(url_key__isnull=False)\n else:\n- kwargs['parts'] = Person.objects.filter(Q(speaker__team__tournament=self.tournament) | Q(adjudicator__tournament__isnull=True) & Q(url_key__isnull=False))\n+ participants = Person.objects.filter(\n+ Q(speaker__team__tournament=self.tournament) | Q(adjudicator__tournament__isnull=True) & Q(url_key__isnull=False))\n+\n+ participants_array = list(participants.select_related('speaker', 'speaker__team', 'adjudicator__institution', 'adjudicator')\n+ .values('name', 'speaker__team__short_name', 'adjudicator__institution__code', 'url_key'))\n+ kwargs['parts'] = self.add_urls(participants_array)\n \n kwargs['exists'] = self.tournament.participants.filter(url_key__isnull=False).exists()\n", "issue": "Include QR code in private URL distribution pages\nIt is a bit odd to hand people a physical item directing them to the internet, but typing a random string is not really user-friendly. QR codes that directs to the URL could be added to the URL sheets for ease of use. The type of barcode that is in use can't encode as long a string, but a QR code is recognized to be scanned to lead to a website and seems to be well supported by phones.\r\n\r\nThis may require another dependency to generate these codes; `jsBarcode` can't generate them.\n", "before_files": [{"content": "import json\n\nfrom django.db.models import Q\nfrom django.utils.translation import gettext as _\nfrom django.views.generic.base import TemplateView\n\nfrom adjfeedback.models import AdjudicatorFeedbackQuestion\nfrom adjfeedback.utils import expected_feedback_targets\nfrom checkins.models import DebateIdentifier\nfrom checkins.utils import create_identifiers\nfrom draw.models import Debate, DebateTeam\nfrom options.utils import use_team_code_names\nfrom participants.models import Person\nfrom results.utils import side_and_position_names\nfrom tournaments.mixins import (CurrentRoundMixin, OptionalAssistantTournamentPageMixin,\n RoundMixin, TournamentMixin)\nfrom tournaments.models import Tournament\nfrom utils.mixins import AdministratorMixin\nfrom venues.models import VenueCategory\n\n\nclass MasterSheetsListView(AdministratorMixin, RoundMixin, TemplateView):\n template_name = 'division_sheets_list.html'\n\n def get_context_data(self, **kwargs):\n kwargs['standings'] = VenueCategory.objects.all()\n kwargs['venue_categories'] = VenueCategory.objects.all()\n return super().get_context_data(**kwargs)\n\n\nclass MasterSheetsView(AdministratorMixin, RoundMixin, TemplateView):\n template_name = 'master_sheets_view.html'\n\n def get_context_data(self, **kwargs):\n venue_category_id = self.kwargs['venue_category_id']\n base_venue_category = VenueCategory.objects.get(id=venue_category_id)\n active_tournaments = Tournament.objects.filter(active=True)\n for tournament in list(active_tournaments):\n tournament.debates = Debate.objects.select_related(\n 'division', 'division__venue_category', 'round',\n 'round__tournament').filter(\n # All Debates, with a matching round, at the same venue category name\n round__seq=self.round.seq,\n round__tournament=tournament,\n # Hack - remove when venue category are unified\n division__venue_category__name=base_venue_category.name\n ).order_by('round', 'division__venue_category__name', 'division')\n\n kwargs['base_venue_category'] = base_venue_category\n kwargs['active_tournaments'] = active_tournaments\n return super().get_context_data(**kwargs)\n\n\nclass RoomSheetsView(AdministratorMixin, RoundMixin, TemplateView):\n template_name = 'room_sheets_view.html'\n\n def get_context_data(self, **kwargs):\n venue_category_id = self.kwargs['venue_category_id']\n base_venue_category = VenueCategory.objects.get(id=venue_category_id)\n venues_list = []\n\n # Get a unique list of venue names (avoid getting duplicates across tournaments)\n for venue in set(base_venue_category.venues.order_by('name').values_list('name', flat=True)):\n venues_list.append({'name': venue, 'debates': []})\n # All Debates, with a matching round, at the same venue category\n venues_list[-1]['debates'] = Debate.objects.filter(\n round__seq=self.round.seq, venue__name=venue).order_by('round__tournament__seq').all()\n print(venues_list[-1])\n\n kwargs['base_venue_category'] = base_venue_category\n kwargs['venues'] = venues_list\n return super().get_context_data(**kwargs)\n\n\nclass BasePrintFeedbackFormsView(RoundMixin, TemplateView):\n\n template_name = 'feedback_list.html'\n\n def add_defaults(self):\n default_questions = []\n\n if self.tournament.pref('feedback_introduction'):\n default_scale_info = AdjudicatorFeedbackQuestion(\n text=self.tournament.pref('feedback_introduction'), seq=0,\n answer_type='comment', # Custom type just for print display\n required=True, from_team=True, from_adj=True\n )\n default_questions.append(default_scale_info.serialize())\n\n default_scale_question = AdjudicatorFeedbackQuestion(\n text=_(\"Overall Score\"), seq=0,\n answer_type=AdjudicatorFeedbackQuestion.ANSWER_TYPE_INTEGER_SCALE,\n required=True, from_team=True, from_adj=True,\n min_value=self.tournament.pref('adj_min_score'),\n max_value=self.tournament.pref('adj_max_score')\n )\n default_questions.append(default_scale_question.serialize())\n\n return default_questions\n\n def questions_dict(self):\n questions = self.add_defaults()\n for question in self.tournament.adj_feedback_questions:\n questions.append(question.serialize())\n\n return questions\n\n def construct_info(self, venue, source, source_p, target, target_p):\n if hasattr(source, 'name'):\n source_n = source.name\n elif use_team_code_names(self.tournament, False):\n source_n = source.code_name\n else:\n source_n = source.short_name\n\n return {\n 'venue': venue.serialize() if venue else '',\n 'authorInstitution': source.institution.code if source.institution else _(\"Unaffiliated\"),\n 'author': source_n, 'authorPosition': source_p,\n 'target': target.name, 'targetPosition': target_p,\n }\n\n def get_team_feedbacks(self, debate, team):\n if len(debate.adjudicators) is 0:\n return []\n\n team_paths = self.tournament.pref('feedback_from_teams')\n ballots = []\n\n if team_paths == 'orallist' and debate.adjudicators.chair:\n ballots.append(self.construct_info(debate.venue, team, _(\"Team\"),\n debate.adjudicators.chair, \"\"))\n elif team_paths == 'all-adjs':\n for target in debate.adjudicators.all():\n ballots.append(self.construct_info(debate.venue, team, _(\"Team\"), target, \"\"))\n\n return ballots\n\n def get_adj_feedbacks(self, debate):\n adj_paths = self.tournament.pref('feedback_paths')\n ballots = []\n\n debateadjs = debate.debateadjudicator_set.all()\n for debateadj in debateadjs:\n sadj = debateadj.adjudicator\n spos = debate.adjudicators.get_position(sadj)\n targets = expected_feedback_targets(debateadj, feedback_paths=adj_paths, debate=debate)\n for tadj, tpos in targets:\n ballots.append(self.construct_info(debate.venue, sadj, spos, tadj, tpos))\n\n return ballots\n\n def get_context_data(self, **kwargs):\n draw = self.round.debate_set_with_prefetches(institutions=True)\n draw = sorted(draw, key=lambda d: d.venue.display_name if d.venue else \"\")\n\n ballots = []\n for debate in draw:\n for team in debate.teams:\n ballots.extend(self.get_team_feedbacks(debate, team))\n ballots.extend(self.get_adj_feedbacks(debate))\n\n kwargs['ballots'] = json.dumps(ballots)\n kwargs['questions'] = json.dumps(self.questions_dict())\n\n kwargs['team_questions_exist'] = self.tournament.adjudicatorfeedbackquestion_set.filter(from_team=True).exists()\n kwargs['adj_questions_exist'] = self.tournament.adjudicatorfeedbackquestion_set.filter(from_adj=True).exists()\n\n return super().get_context_data(**kwargs)\n\n\nclass AdminPrintFeedbackFormsView(AdministratorMixin, BasePrintFeedbackFormsView):\n pass\n\n\nclass AssistantPrintFeedbackFormsView(CurrentRoundMixin, OptionalAssistantTournamentPageMixin, BasePrintFeedbackFormsView):\n assistant_page_permissions = ['all_areas', 'results_draw']\n\n\nclass BasePrintScoresheetsView(RoundMixin, TemplateView):\n\n template_name = 'scoresheet_list.html'\n\n def get_ballots_dicts(self):\n draw = self.round.debate_set_with_prefetches()\n\n # Create the DebateIdentifiers for the ballots if needed\n create_identifiers(DebateIdentifier, draw)\n identifiers = DebateIdentifier.objects.values('debate_id', 'barcode')\n\n draw = sorted(draw, key=lambda d: d.venue.display_name if d.venue else \"\")\n ballots_dicts = []\n\n # Force translation before JSON serialization\n sides_and_positions = [(side, [str(pos) for pos in positions])\n for side, positions in side_and_position_names(self.tournament)]\n\n for debate in draw:\n debate_dict = {}\n\n if debate.venue:\n debate_dict['venue'] = {'display_name': debate.venue.display_name}\n else:\n debate_dict['venue'] = None\n\n debate_dict['barcode'] = next((i['barcode'] for i in identifiers if i['debate_id'] == debate.id), None)\n\n debate_dict['debateTeams'] = []\n for side, (side_name, positions) in zip(self.tournament.sides, sides_and_positions):\n dt_dict = {'side_name': side_name, 'positions': positions}\n try:\n team = debate.get_team(side)\n dt_dict['team'] = {\n 'short_name': team.short_name,\n 'code_name': team.code_name,\n 'speakers': [{'name': s.name} for s in team.speakers],\n }\n except DebateTeam.DoesNotExist:\n dt_dict['team'] = None\n debate_dict['debateTeams'].append(dt_dict)\n\n debate_dict['debateAdjudicators'] = []\n for adj, pos in debate.adjudicators.with_positions():\n da_dict = {'position': pos}\n da_dict['adjudicator'] = {\n 'name': adj.name,\n 'institution': {'code': adj.institution.code if adj.institution else _(\"Unaffiliated\")},\n }\n debate_dict['debateAdjudicators'].append(da_dict)\n\n if self.round.ballots_per_debate == 'per-adj':\n authors = list(debate.adjudicators.voting_with_positions())\n else:\n authors = [(debate.adjudicators.chair, debate.adjudicators.POSITION_CHAIR)]\n\n blank_author_dict = {\n 'author': \"_______________________________________________\",\n 'authorInstitution': \"\",\n 'authorPosition': \"\",\n }\n\n # Add a ballot for each author\n for author, pos in authors:\n if author:\n ballot_dict = {\n 'author': author.name,\n 'authorInstitution': author.institution.code if author.institution else _(\"Unaffiliated\"),\n 'authorPosition': pos,\n }\n else:\n ballot_dict = blank_author_dict\n\n ballot_dict.update(debate_dict)\n ballots_dicts.append(ballot_dict)\n\n if len(authors) == 0:\n ballot_dict = blank_author_dict\n ballot_dict.update(debate_dict)\n ballots_dicts.append(ballot_dict)\n\n return ballots_dicts\n\n def get_context_data(self, **kwargs):\n kwargs['ballots'] = json.dumps(self.get_ballots_dicts())\n motions = self.round.motion_set.order_by('seq')\n kwargs['motions'] = json.dumps([{'seq': m.seq, 'text': m.text} for m in motions])\n kwargs['use_team_code_names'] = use_team_code_names(self.tournament, False)\n return super().get_context_data(**kwargs)\n\n\nclass AdminPrintScoresheetsView(AdministratorMixin, BasePrintScoresheetsView):\n pass\n\n\nclass AssistantPrintScoresheetsView(CurrentRoundMixin, OptionalAssistantTournamentPageMixin, BasePrintScoresheetsView):\n assistant_page_permissions = ['all_areas']\n\n\nclass PrintableRandomisedURLs(TournamentMixin, AdministratorMixin, TemplateView):\n\n template_name = 'randomised_url_sheets.html'\n\n def get_context_data(self, **kwargs):\n kwargs['tournament_slug'] = self.tournament.slug\n\n if not self.tournament.pref('share_adjs'):\n kwargs['parts'] = self.tournament.participants.filter(\n url_key__isnull=False).select_related(\n 'speaker', 'speaker__team', 'adjudicator__institution',\n 'adjudicator')\n else:\n kwargs['parts'] = Person.objects.filter(Q(speaker__team__tournament=self.tournament) | Q(adjudicator__tournament__isnull=True) & Q(url_key__isnull=False))\n\n kwargs['exists'] = self.tournament.participants.filter(url_key__isnull=False).exists()\n\n return super().get_context_data(**kwargs)\n", "path": "tabbycat/printing/views.py"}]}
| 3,993 | 583 |
gh_patches_debug_30485
|
rasdani/github-patches
|
git_diff
|
googleapis__google-cloud-python-4819
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Runtimeconfig: Update time fails to parse
OS: Linux (Debian 4.9.65)
Python: 3.5.3 / 2.7.13
Version: google-cloud-runtimeconfig==0.28.0, google-cloud==0.32.0
Steps to reproduce:
1. Using gcloud, set a runtimeconfig variable.
2. In Python, load that variable
3. Try to access its update_time property.
```pycon
>>> from google.cloud import runtimeconfig
>>> client = runtimeconfig.Client()
>>> config = client.config('testconfig')
>>> var = config.get_variable('foo')
>>> var.update_time
Traceback (most recent call last):
File "<stdin>", line 1, in <module>
File "/redacted/lib/python3.5/site-packages/google/cloud/runtimeconfig/variable.py", line 160, in update_time
value = _rfc3339_to_datetime(value)
File "/redacted/lib/python3.5/site-packages/google/cloud/_helpers.py", line 274, in _rfc3339_to_datetime
dt_str, _RFC3339_MICROS).replace(tzinfo=UTC)
File "/usr/lib/python3.5/_strptime.py", line 510, in _strptime_datetime
tt, fraction = _strptime(data_string, format)
File "/usr/lib/python3.5/_strptime.py", line 343, in _strptime
(data_string, format))
ValueError: time data '2018-01-22T21:39:44.095040522Z' does not match format '%Y-%m-%dT%H:%M:%S.%fZ'
>>> var._properties
{'text': '43', 'updateTime': '2018-01-22T21:39:44.095040522Z'}
```
Observation: The `%f` format accepts microseconds, not fractions of arbitrary precision. That means six digits at most, but I see nine in the data.
</issue>
<code>
[start of runtimeconfig/google/cloud/runtimeconfig/variable.py]
1 # Copyright 2016 Google LLC
2 #
3 # Licensed under the Apache License, Version 2.0 (the "License");
4 # you may not use this file except in compliance with the License.
5 # You may obtain a copy of the License at
6 #
7 # http://www.apache.org/licenses/LICENSE-2.0
8 #
9 # Unless required by applicable law or agreed to in writing, software
10 # distributed under the License is distributed on an "AS IS" BASIS,
11 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12 # See the License for the specific language governing permissions and
13 # limitations under the License.
14
15 """Create / interact with Google Cloud RuntimeConfig variables.
16
17 .. data:: STATE_UNSPECIFIED
18
19 The default variable state. See
20 https://cloud.google.com/deployment-manager/runtime-configurator/reference/rest/v1beta1/projects.configs.variables#VariableState
21
22 .. data:: STATE_UPDATED
23
24 Indicates the variable was updated, while `variables.watch` was executing.
25 See
26 https://cloud.google.com/deployment-manager/runtime-configurator/reference/rest/v1beta1/projects.configs.variables#VariableState
27
28 .. data:: STATE_DELETED
29
30 Indicates the variable was deleted, while `variables.watch`_ was executing.
31 See
32 https://cloud.google.com/deployment-manager/runtime-configurator/reference/rest/v1beta1/projects.configs.variables#VariableState
33
34 .. _variables.watch:
35 https://cloud.google.com/deployment-manager/runtime-configurator/reference/rest/v1beta1/projects.configs.variables/watch
36 """
37
38 import base64
39
40 from google.cloud._helpers import _rfc3339_to_datetime
41 from google.cloud.exceptions import NotFound
42 from google.cloud.runtimeconfig._helpers import variable_name_from_full_name
43
44
45 STATE_UNSPECIFIED = 'VARIABLE_STATE_UNSPECIFIED'
46 STATE_UPDATED = 'UPDATED'
47 STATE_DELETED = 'DELETED'
48
49
50 class Variable(object):
51 """A variable in the Cloud RuntimeConfig service.
52
53 See
54 https://cloud.google.com/deployment-manager/runtime-configurator/reference/rest/v1beta1/projects.configs.variables
55
56 :type name: str
57 :param name: The name of the variable. This corresponds to the
58 unique path of the variable in the config.
59
60 :type config: :class:`google.cloud.runtimeconfig.config.Config`
61 :param config: The config to which this variable belongs.
62 """
63
64 def __init__(self, name, config):
65 self.name = name
66 self.config = config
67 self._properties = {}
68
69 @classmethod
70 def from_api_repr(cls, resource, config):
71 """Factory: construct a Variable given its API representation
72
73 :type resource: dict
74 :param resource: change set representation returned from the API.
75
76 :type config: :class:`google.cloud.runtimeconfig.config.Config`
77 :param config: The config to which this variable belongs.
78
79 :rtype: :class:`google.cloud.runtimeconfig.variable.Variable`
80 :returns: Variable parsed from ``resource``.
81 """
82 name = variable_name_from_full_name(resource.get('name'))
83 variable = cls(name=name, config=config)
84 variable._set_properties(resource=resource)
85 return variable
86
87 @property
88 def full_name(self):
89 """Fully-qualified name of this variable.
90
91 Example:
92 ``projects/my-project/configs/my-config/variables/my-var``
93
94 :rtype: str
95 :returns: The full name based on config and variable names.
96
97 :raises: :class:`ValueError` if the variable is missing a name.
98 """
99 if not self.name:
100 raise ValueError('Missing variable name.')
101 return '%s/variables/%s' % (self.config.full_name, self.name)
102
103 @property
104 def path(self):
105 """URL path for the variable's APIs.
106
107 :rtype: str
108 :returns: The URL path based on config and variable names.
109 """
110 return '/%s' % (self.full_name,)
111
112 @property
113 def client(self):
114 """The client bound to this variable."""
115 return self.config.client
116
117 @property
118 def value(self):
119 """Value of the variable, as bytes.
120
121 See
122 https://cloud.google.com/deployment-manager/runtime-configurator/reference/rest/v1beta1/projects.configs.variables
123
124 :rtype: bytes or ``NoneType``
125 :returns: The value of the variable or ``None`` if the property
126 is not set locally.
127 """
128 value = self._properties.get('value')
129 if value is not None:
130 value = base64.b64decode(value)
131 return value
132
133 @property
134 def state(self):
135 """Retrieve the state of the variable.
136
137 See
138 https://cloud.google.com/deployment-manager/runtime-configurator/reference/rest/v1beta1/projects.configs.variables#VariableState
139
140 :rtype: str
141 :returns:
142 If set, one of "UPDATED", "DELETED", or defaults to
143 "VARIABLE_STATE_UNSPECIFIED".
144 """
145 return self._properties.get('state', STATE_UNSPECIFIED)
146
147 @property
148 def update_time(self):
149 """Retrieve the timestamp at which the variable was updated.
150
151 See
152 https://cloud.google.com/deployment-manager/runtime-configurator/reference/rest/v1beta1/projects.configs.variables
153
154 :rtype: :class:`datetime.datetime` or ``NoneType``
155 :returns: Datetime object parsed from RFC3339 valid timestamp, or
156 ``None`` if the property is not set locally.
157 """
158 value = self._properties.get('updateTime')
159 if value is not None:
160 value = _rfc3339_to_datetime(value)
161 return value
162
163 def _require_client(self, client):
164 """Check client or verify over-ride.
165
166 :type client: :class:`google.cloud.runtimconfig.client.Client`
167 :param client:
168 (Optional) The client to use. If not passed, falls back to the
169 ``client`` stored on the current zone.
170
171 :rtype: :class:`google.cloud.runtimeconfig.client.Client`
172 :returns: The client passed in or the currently bound client.
173 """
174 if client is None:
175 client = self.client
176 return client
177
178 def _set_properties(self, resource):
179 """Update properties from resource in body of ``api_response``
180
181 :type resource: dict
182 :param resource: variable representation returned from the API.
183 """
184 self._properties.clear()
185 cleaned = resource.copy()
186 if 'name' in cleaned:
187 self.name = variable_name_from_full_name(cleaned.pop('name'))
188 self._properties.update(cleaned)
189
190 def exists(self, client=None):
191 """API call: test for the existence of the variable via a GET request
192
193 See
194 https://cloud.google.com/deployment-manager/runtime-configurator/reference/rest/v1beta1/projects.configs.variables/get
195
196 :type client: :class:`~google.cloud.runtimeconfig.client.Client`
197 :param client:
198 (Optional) The client to use. If not passed, falls back to the
199 ``client`` stored on the variable's config.
200
201 :rtype: bool
202 :returns: True if the variable exists in Cloud RuntimeConfig.
203 """
204 client = self._require_client(client)
205 try:
206 # We only need the status code (200 or not) so we seek to
207 # minimize the returned payload.
208 query_params = {'fields': 'name'}
209 client._connection.api_request(method='GET', path=self.path,
210 query_params=query_params)
211 return True
212 except NotFound:
213 return False
214
215 def reload(self, client=None):
216 """API call: reload the variable via a ``GET`` request.
217
218 This method will reload the newest data for the variable.
219
220 See
221 https://cloud.google.com/deployment-manager/runtime-configurator/reference/rest/v1beta1/projects.configs/get
222
223 :type client: :class:`google.cloud.runtimeconfig.client.Client`
224 :param client:
225 (Optional) The client to use. If not passed, falls back to the
226 client stored on the current config.
227 """
228 client = self._require_client(client)
229
230 # We assume the variable exists. If it doesn't it will raise a NotFound
231 # exception.
232 resp = client._connection.api_request(method='GET', path=self.path)
233 self._set_properties(resource=resp)
234
[end of runtimeconfig/google/cloud/runtimeconfig/variable.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/runtimeconfig/google/cloud/runtimeconfig/variable.py b/runtimeconfig/google/cloud/runtimeconfig/variable.py
--- a/runtimeconfig/google/cloud/runtimeconfig/variable.py
+++ b/runtimeconfig/google/cloud/runtimeconfig/variable.py
@@ -36,8 +36,11 @@
"""
import base64
+import datetime
-from google.cloud._helpers import _rfc3339_to_datetime
+import pytz
+
+from google.api_core import datetime_helpers
from google.cloud.exceptions import NotFound
from google.cloud.runtimeconfig._helpers import variable_name_from_full_name
@@ -151,13 +154,28 @@
See
https://cloud.google.com/deployment-manager/runtime-configurator/reference/rest/v1beta1/projects.configs.variables
- :rtype: :class:`datetime.datetime` or ``NoneType``
- :returns: Datetime object parsed from RFC3339 valid timestamp, or
- ``None`` if the property is not set locally.
+ Returns:
+ :class:`~api_core.datetime_helpers.DatetimeWithNanoseconds`,
+ :class:`datetime.datetime` or ``NoneType``:
+ Datetime object parsed from RFC3339 valid timestamp, or
+ ``None`` if the property is not set locally.
+
+ Raises:
+ ValueError: if value is not a valid RFC3339 timestamp
"""
value = self._properties.get('updateTime')
if value is not None:
- value = _rfc3339_to_datetime(value)
+ try:
+ value = datetime.datetime.strptime(
+ value, datetime_helpers._RFC3339_MICROS)
+ except ValueError:
+ DatetimeNS = datetime_helpers.DatetimeWithNanoseconds
+ value = DatetimeNS.from_rfc3339(value)
+ naive = (
+ value.tzinfo is None
+ or value.tzinfo.utcoffset(value) is None)
+ if naive:
+ value = pytz.utc.localize(value)
return value
def _require_client(self, client):
|
{"golden_diff": "diff --git a/runtimeconfig/google/cloud/runtimeconfig/variable.py b/runtimeconfig/google/cloud/runtimeconfig/variable.py\n--- a/runtimeconfig/google/cloud/runtimeconfig/variable.py\n+++ b/runtimeconfig/google/cloud/runtimeconfig/variable.py\n@@ -36,8 +36,11 @@\n \"\"\"\n \n import base64\n+import datetime\n \n-from google.cloud._helpers import _rfc3339_to_datetime\n+import pytz\n+\n+from google.api_core import datetime_helpers\n from google.cloud.exceptions import NotFound\n from google.cloud.runtimeconfig._helpers import variable_name_from_full_name\n \n@@ -151,13 +154,28 @@\n See\n https://cloud.google.com/deployment-manager/runtime-configurator/reference/rest/v1beta1/projects.configs.variables\n \n- :rtype: :class:`datetime.datetime` or ``NoneType``\n- :returns: Datetime object parsed from RFC3339 valid timestamp, or\n- ``None`` if the property is not set locally.\n+ Returns:\n+ :class:`~api_core.datetime_helpers.DatetimeWithNanoseconds`,\n+ :class:`datetime.datetime` or ``NoneType``:\n+ Datetime object parsed from RFC3339 valid timestamp, or\n+ ``None`` if the property is not set locally.\n+\n+ Raises:\n+ ValueError: if value is not a valid RFC3339 timestamp\n \"\"\"\n value = self._properties.get('updateTime')\n if value is not None:\n- value = _rfc3339_to_datetime(value)\n+ try:\n+ value = datetime.datetime.strptime(\n+ value, datetime_helpers._RFC3339_MICROS)\n+ except ValueError:\n+ DatetimeNS = datetime_helpers.DatetimeWithNanoseconds\n+ value = DatetimeNS.from_rfc3339(value)\n+ naive = (\n+ value.tzinfo is None\n+ or value.tzinfo.utcoffset(value) is None)\n+ if naive:\n+ value = pytz.utc.localize(value)\n return value\n \n def _require_client(self, client):\n", "issue": "Runtimeconfig: Update time fails to parse\nOS: Linux (Debian 4.9.65)\r\nPython: 3.5.3 / 2.7.13\r\nVersion: google-cloud-runtimeconfig==0.28.0, google-cloud==0.32.0\r\n\r\nSteps to reproduce:\r\n1. Using gcloud, set a runtimeconfig variable.\r\n2. In Python, load that variable\r\n3. Try to access its update_time property.\r\n\r\n```pycon\r\n>>> from google.cloud import runtimeconfig\r\n>>> client = runtimeconfig.Client()\r\n>>> config = client.config('testconfig')\r\n>>> var = config.get_variable('foo')\r\n>>> var.update_time\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"/redacted/lib/python3.5/site-packages/google/cloud/runtimeconfig/variable.py\", line 160, in update_time\r\n value = _rfc3339_to_datetime(value)\r\n File \"/redacted/lib/python3.5/site-packages/google/cloud/_helpers.py\", line 274, in _rfc3339_to_datetime\r\n dt_str, _RFC3339_MICROS).replace(tzinfo=UTC)\r\n File \"/usr/lib/python3.5/_strptime.py\", line 510, in _strptime_datetime\r\n tt, fraction = _strptime(data_string, format)\r\n File \"/usr/lib/python3.5/_strptime.py\", line 343, in _strptime\r\n (data_string, format))\r\nValueError: time data '2018-01-22T21:39:44.095040522Z' does not match format '%Y-%m-%dT%H:%M:%S.%fZ'\r\n>>> var._properties\r\n{'text': '43', 'updateTime': '2018-01-22T21:39:44.095040522Z'}\r\n```\r\n\r\nObservation: The `%f` format accepts microseconds, not fractions of arbitrary precision. That means six digits at most, but I see nine in the data.\n", "before_files": [{"content": "# Copyright 2016 Google LLC\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\n\"\"\"Create / interact with Google Cloud RuntimeConfig variables.\n\n.. data:: STATE_UNSPECIFIED\n\n The default variable state. See\n https://cloud.google.com/deployment-manager/runtime-configurator/reference/rest/v1beta1/projects.configs.variables#VariableState\n\n.. data:: STATE_UPDATED\n\n Indicates the variable was updated, while `variables.watch` was executing.\n See\n https://cloud.google.com/deployment-manager/runtime-configurator/reference/rest/v1beta1/projects.configs.variables#VariableState\n\n.. data:: STATE_DELETED\n\n Indicates the variable was deleted, while `variables.watch`_ was executing.\n See\n https://cloud.google.com/deployment-manager/runtime-configurator/reference/rest/v1beta1/projects.configs.variables#VariableState\n\n.. _variables.watch:\n https://cloud.google.com/deployment-manager/runtime-configurator/reference/rest/v1beta1/projects.configs.variables/watch\n\"\"\"\n\nimport base64\n\nfrom google.cloud._helpers import _rfc3339_to_datetime\nfrom google.cloud.exceptions import NotFound\nfrom google.cloud.runtimeconfig._helpers import variable_name_from_full_name\n\n\nSTATE_UNSPECIFIED = 'VARIABLE_STATE_UNSPECIFIED'\nSTATE_UPDATED = 'UPDATED'\nSTATE_DELETED = 'DELETED'\n\n\nclass Variable(object):\n \"\"\"A variable in the Cloud RuntimeConfig service.\n\n See\n https://cloud.google.com/deployment-manager/runtime-configurator/reference/rest/v1beta1/projects.configs.variables\n\n :type name: str\n :param name: The name of the variable. This corresponds to the\n unique path of the variable in the config.\n\n :type config: :class:`google.cloud.runtimeconfig.config.Config`\n :param config: The config to which this variable belongs.\n \"\"\"\n\n def __init__(self, name, config):\n self.name = name\n self.config = config\n self._properties = {}\n\n @classmethod\n def from_api_repr(cls, resource, config):\n \"\"\"Factory: construct a Variable given its API representation\n\n :type resource: dict\n :param resource: change set representation returned from the API.\n\n :type config: :class:`google.cloud.runtimeconfig.config.Config`\n :param config: The config to which this variable belongs.\n\n :rtype: :class:`google.cloud.runtimeconfig.variable.Variable`\n :returns: Variable parsed from ``resource``.\n \"\"\"\n name = variable_name_from_full_name(resource.get('name'))\n variable = cls(name=name, config=config)\n variable._set_properties(resource=resource)\n return variable\n\n @property\n def full_name(self):\n \"\"\"Fully-qualified name of this variable.\n\n Example:\n ``projects/my-project/configs/my-config/variables/my-var``\n\n :rtype: str\n :returns: The full name based on config and variable names.\n\n :raises: :class:`ValueError` if the variable is missing a name.\n \"\"\"\n if not self.name:\n raise ValueError('Missing variable name.')\n return '%s/variables/%s' % (self.config.full_name, self.name)\n\n @property\n def path(self):\n \"\"\"URL path for the variable's APIs.\n\n :rtype: str\n :returns: The URL path based on config and variable names.\n \"\"\"\n return '/%s' % (self.full_name,)\n\n @property\n def client(self):\n \"\"\"The client bound to this variable.\"\"\"\n return self.config.client\n\n @property\n def value(self):\n \"\"\"Value of the variable, as bytes.\n\n See\n https://cloud.google.com/deployment-manager/runtime-configurator/reference/rest/v1beta1/projects.configs.variables\n\n :rtype: bytes or ``NoneType``\n :returns: The value of the variable or ``None`` if the property\n is not set locally.\n \"\"\"\n value = self._properties.get('value')\n if value is not None:\n value = base64.b64decode(value)\n return value\n\n @property\n def state(self):\n \"\"\"Retrieve the state of the variable.\n\n See\n https://cloud.google.com/deployment-manager/runtime-configurator/reference/rest/v1beta1/projects.configs.variables#VariableState\n\n :rtype: str\n :returns:\n If set, one of \"UPDATED\", \"DELETED\", or defaults to\n \"VARIABLE_STATE_UNSPECIFIED\".\n \"\"\"\n return self._properties.get('state', STATE_UNSPECIFIED)\n\n @property\n def update_time(self):\n \"\"\"Retrieve the timestamp at which the variable was updated.\n\n See\n https://cloud.google.com/deployment-manager/runtime-configurator/reference/rest/v1beta1/projects.configs.variables\n\n :rtype: :class:`datetime.datetime` or ``NoneType``\n :returns: Datetime object parsed from RFC3339 valid timestamp, or\n ``None`` if the property is not set locally.\n \"\"\"\n value = self._properties.get('updateTime')\n if value is not None:\n value = _rfc3339_to_datetime(value)\n return value\n\n def _require_client(self, client):\n \"\"\"Check client or verify over-ride.\n\n :type client: :class:`google.cloud.runtimconfig.client.Client`\n :param client:\n (Optional) The client to use. If not passed, falls back to the\n ``client`` stored on the current zone.\n\n :rtype: :class:`google.cloud.runtimeconfig.client.Client`\n :returns: The client passed in or the currently bound client.\n \"\"\"\n if client is None:\n client = self.client\n return client\n\n def _set_properties(self, resource):\n \"\"\"Update properties from resource in body of ``api_response``\n\n :type resource: dict\n :param resource: variable representation returned from the API.\n \"\"\"\n self._properties.clear()\n cleaned = resource.copy()\n if 'name' in cleaned:\n self.name = variable_name_from_full_name(cleaned.pop('name'))\n self._properties.update(cleaned)\n\n def exists(self, client=None):\n \"\"\"API call: test for the existence of the variable via a GET request\n\n See\n https://cloud.google.com/deployment-manager/runtime-configurator/reference/rest/v1beta1/projects.configs.variables/get\n\n :type client: :class:`~google.cloud.runtimeconfig.client.Client`\n :param client:\n (Optional) The client to use. If not passed, falls back to the\n ``client`` stored on the variable's config.\n\n :rtype: bool\n :returns: True if the variable exists in Cloud RuntimeConfig.\n \"\"\"\n client = self._require_client(client)\n try:\n # We only need the status code (200 or not) so we seek to\n # minimize the returned payload.\n query_params = {'fields': 'name'}\n client._connection.api_request(method='GET', path=self.path,\n query_params=query_params)\n return True\n except NotFound:\n return False\n\n def reload(self, client=None):\n \"\"\"API call: reload the variable via a ``GET`` request.\n\n This method will reload the newest data for the variable.\n\n See\n https://cloud.google.com/deployment-manager/runtime-configurator/reference/rest/v1beta1/projects.configs/get\n\n :type client: :class:`google.cloud.runtimeconfig.client.Client`\n :param client:\n (Optional) The client to use. If not passed, falls back to the\n client stored on the current config.\n \"\"\"\n client = self._require_client(client)\n\n # We assume the variable exists. If it doesn't it will raise a NotFound\n # exception.\n resp = client._connection.api_request(method='GET', path=self.path)\n self._set_properties(resource=resp)\n", "path": "runtimeconfig/google/cloud/runtimeconfig/variable.py"}]}
| 3,422 | 457 |
gh_patches_debug_12822
|
rasdani/github-patches
|
git_diff
|
conan-io__conan-center-index-22763
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
[question] Using redis-plus-plus with TLS/SSL on
### What is your question?
Hi, I'm having difficulty using Conan to include a version of redis-plus-plus built with TLS enabled in my project. It seems like redis-plus-plus's conan recipe has a `with_tls` option that should enable the support. However, this depends on hiredis being built with SSL enabled. The conan recipe for hiredis seems to have a `with_ssl` option that's supposed to do exactly that, but I could not get it to work.
Here's my conanfile.py:
```
from conan import ConanFile, tools
from conan.tools.cmake import CMake, CMakeDeps, CMakeToolchain, cmake_layout
class HelloConan(ConanFile):
name = "myproject"
version = "0.0.1"
settings = "os", "compiler", "build_type", "arch"
def layout(self):
cmake_layout(self)
def generate(self):
deps = CMakeDeps(self)
deps.generate()
tc = CMakeToolchain(self)
tc.generate()
def configure(self):
self.options["hiredis"]["with_ssl"] = True
self.options["redis-plus-plus"]["with_tls"] = True
def build(self):
cmake = CMake(self)
cmake.configure()
cmake.build()
def package(self):
cmake = CMake(self)
cmake.install()
def requirements(self):
self.requires("redis-plus-plus/1.3.11")
self.requires("hiredis/1.2.0")
```
The error message I get is
```
CMake Error at
/home/ago/.conan2/p/b/redis08573fb8e514a/b/build/Release/generators/cmakedeps_macros.cmake:67 (message):
Library 'hiredis_ssl' not found in package. If 'hiredis_ssl' is a system
library, declare it with 'cpp_info.system_libs' property
```
I also tried not including hiredis in the requirements and/or configure functions in my conanfile.py. The result was
```
ERROR: There are invalid packages:
redis-plus-plus/1.3.11: Invalid: redis-plus-plus:with_tls=True requires hiredis:with_ssl=True
```
I'd really appreciate if anyone could guide me in the right direction to get redis-plus-plus built with the with_tls option enabled.
</issue>
<code>
[start of recipes/hiredis/all/conanfile.py]
1 from conan import ConanFile
2 from conan.tools.cmake import CMake, CMakeDeps, CMakeToolchain, cmake_layout
3 from conan.tools.files import apply_conandata_patches, copy, export_conandata_patches, get, rmdir, rm
4 from conan.tools.scm import Version
5 from conan.tools.microsoft import is_msvc
6 import os
7
8 required_conan_version = ">=1.53.0"
9
10
11 class HiredisConan(ConanFile):
12 name = "hiredis"
13 description = "Hiredis is a minimalistic C client library for the Redis database."
14 license = "BSD-3-Clause"
15 topics = ("hiredis", "redis", "client", "database")
16 homepage = "https://github.com/redis/hiredis"
17 url = "https://github.com/conan-io/conan-center-index"
18 package_type = "library"
19 settings = "os", "arch", "compiler", "build_type"
20 options = {
21 "shared": [True, False],
22 "fPIC": [True, False],
23 "with_ssl": [True, False],
24 }
25 default_options = {
26 "shared": False,
27 "fPIC": True,
28 "with_ssl": False,
29 }
30
31 def export_sources(self):
32 export_conandata_patches(self)
33
34 def config_options(self):
35 if self.settings.os == "Windows":
36 del self.options.fPIC
37
38 def configure(self):
39 if self.options.shared:
40 self.options.rm_safe("fPIC")
41 self.settings.compiler.rm_safe("libcxx")
42 self.settings.compiler.rm_safe("cppstd")
43
44 def layout(self):
45 cmake_layout(self, src_folder="src")
46
47 def requirements(self):
48 if self.options.with_ssl:
49 self.requires("openssl/[>=1.1 <4]")
50
51 def source(self):
52 get(self, **self.conan_data["sources"][self.version], strip_root=True)
53
54 def generate(self):
55 tc = CMakeToolchain(self)
56 # Since 1.2.0, BUILD_SHARED_LIBS has been defined by option()
57 if Version(self.version) >= "1.2.0":
58 tc.cache_variables["BUILD_SHARED_LIBS"] = self.options.shared
59 tc.variables["ENABLE_SSL"] = self.options.with_ssl
60 tc.variables["DISABLE_TESTS"] = True
61 tc.variables["ENABLE_EXAMPLES"] = False
62 tc.generate()
63 deps = CMakeDeps(self)
64 deps.generate()
65
66 def build(self):
67 apply_conandata_patches(self)
68 cmake = CMake(self)
69 cmake.configure()
70 cmake.build()
71
72 def package(self):
73 copy(self, "COPYING", src=self.source_folder, dst=os.path.join(self.package_folder, "licenses"))
74 cmake = CMake(self)
75 cmake.install()
76 rmdir(self, os.path.join(self.package_folder, "lib", "pkgconfig"))
77 rmdir(self, os.path.join(self.package_folder, "lib", "cmake"))
78 rmdir(self, os.path.join(self.package_folder, "share"))
79 rmdir(self, os.path.join(self.package_folder, "build"))
80 rm(self, "*.pdb", os.path.join(self.package_folder, "bin"))
81
82 def package_info(self):
83 self.cpp_info.set_property("cmake_file_name", "hiredis")
84
85 suffix = ""
86 if Version(self.version) >= "1.1.0":
87 if is_msvc(self) and not self.options.shared and Version(self.version) < "1.2.0":
88 suffix += "_static"
89 if self.settings.build_type == "Debug":
90 suffix += "d"
91
92 # hiredis
93 self.cpp_info.components["hiredislib"].set_property("cmake_target_name", "hiredis::hiredis")
94 self.cpp_info.components["hiredislib"].set_property("pkg_config_name", "hiredis")
95 self.cpp_info.components["hiredislib"].names["cmake_find_package"] = "hiredis"
96 self.cpp_info.components["hiredislib"].names["cmake_find_package_multi"] = "hiredis"
97 self.cpp_info.components["hiredislib"].libs = [f"hiredis{suffix}"]
98 if self.settings.os == "Windows":
99 self.cpp_info.components["hiredislib"].system_libs = ["ws2_32"]
100 # hiredis_ssl
101 if self.options.with_ssl:
102 self.cpp_info.components["hiredis_ssl"].set_property("cmake_target_name", "hiredis::hiredis_ssl")
103 self.cpp_info.components["hiredis_ssl"].set_property("pkg_config_name", "hiredis_ssl")
104 self.cpp_info.components["hiredis_ssl"].names["cmake_find_package"] = "hiredis_ssl"
105 self.cpp_info.components["hiredis_ssl"].names["cmake_find_package_multi"] = "hiredis_ssl"
106 self.cpp_info.components["hiredis_ssl"].libs = [f"hiredis_ssl{suffix}"]
107 self.cpp_info.components["hiredis_ssl"].requires = ["openssl::ssl"]
108 if self.settings.os == "Windows":
109 self.cpp_info.components["hiredis_ssl"].requires.append("hiredislib")
110
111 # These cmake_target_name and pkg_config_name are unofficial. It avoids conflicts
112 # in conan generators between global target/pkg-config and hiredislib component.
113 # TODO: eventually remove the cmake_target_name trick if conan can implement smarter logic
114 # in CMakeDeps when a downstream recipe requires another recipe globally
115 # (link to all components directly instead of global target)
116 self.cpp_info.set_property("cmake_target_name", "hiredis::hiredis_all_unofficial")
117 self.cpp_info.set_property("pkg_config_name", "hiredis_all_unofficial")
118
[end of recipes/hiredis/all/conanfile.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/recipes/hiredis/all/conanfile.py b/recipes/hiredis/all/conanfile.py
--- a/recipes/hiredis/all/conanfile.py
+++ b/recipes/hiredis/all/conanfile.py
@@ -56,9 +56,9 @@
# Since 1.2.0, BUILD_SHARED_LIBS has been defined by option()
if Version(self.version) >= "1.2.0":
tc.cache_variables["BUILD_SHARED_LIBS"] = self.options.shared
- tc.variables["ENABLE_SSL"] = self.options.with_ssl
- tc.variables["DISABLE_TESTS"] = True
- tc.variables["ENABLE_EXAMPLES"] = False
+ tc.cache_variables["ENABLE_SSL"] = self.options.with_ssl
+ tc.cache_variables["DISABLE_TESTS"] = True
+ tc.cache_variables["ENABLE_EXAMPLES"] = False
tc.generate()
deps = CMakeDeps(self)
deps.generate()
|
{"golden_diff": "diff --git a/recipes/hiredis/all/conanfile.py b/recipes/hiredis/all/conanfile.py\n--- a/recipes/hiredis/all/conanfile.py\n+++ b/recipes/hiredis/all/conanfile.py\n@@ -56,9 +56,9 @@\n # Since 1.2.0, BUILD_SHARED_LIBS has been defined by option()\n if Version(self.version) >= \"1.2.0\":\n tc.cache_variables[\"BUILD_SHARED_LIBS\"] = self.options.shared\n- tc.variables[\"ENABLE_SSL\"] = self.options.with_ssl\n- tc.variables[\"DISABLE_TESTS\"] = True\n- tc.variables[\"ENABLE_EXAMPLES\"] = False\n+ tc.cache_variables[\"ENABLE_SSL\"] = self.options.with_ssl\n+ tc.cache_variables[\"DISABLE_TESTS\"] = True\n+ tc.cache_variables[\"ENABLE_EXAMPLES\"] = False\n tc.generate()\n deps = CMakeDeps(self)\n deps.generate()\n", "issue": "[question] Using redis-plus-plus with TLS/SSL on\n### What is your question?\n\nHi, I'm having difficulty using Conan to include a version of redis-plus-plus built with TLS enabled in my project. It seems like redis-plus-plus's conan recipe has a `with_tls` option that should enable the support. However, this depends on hiredis being built with SSL enabled. The conan recipe for hiredis seems to have a `with_ssl` option that's supposed to do exactly that, but I could not get it to work. \r\n\r\nHere's my conanfile.py:\r\n\r\n```\r\nfrom conan import ConanFile, tools\r\nfrom conan.tools.cmake import CMake, CMakeDeps, CMakeToolchain, cmake_layout\r\n\r\nclass HelloConan(ConanFile):\r\n name = \"myproject\"\r\n version = \"0.0.1\"\r\n settings = \"os\", \"compiler\", \"build_type\", \"arch\"\r\n\r\n\r\n def layout(self):\r\n cmake_layout(self)\r\n\r\n def generate(self):\r\n deps = CMakeDeps(self)\r\n deps.generate()\r\n tc = CMakeToolchain(self)\r\n tc.generate()\r\n\r\n def configure(self):\r\n self.options[\"hiredis\"][\"with_ssl\"] = True\r\n self.options[\"redis-plus-plus\"][\"with_tls\"] = True\r\n\r\n def build(self):\r\n cmake = CMake(self)\r\n cmake.configure()\r\n cmake.build()\r\n\r\n def package(self):\r\n cmake = CMake(self)\r\n cmake.install()\r\n\r\n def requirements(self):\r\n self.requires(\"redis-plus-plus/1.3.11\")\r\n self.requires(\"hiredis/1.2.0\")\r\n```\r\n\r\nThe error message I get is \r\n```\r\nCMake Error at \r\n/home/ago/.conan2/p/b/redis08573fb8e514a/b/build/Release/generators/cmakedeps_macros.cmake:67 (message):\r\n Library 'hiredis_ssl' not found in package. If 'hiredis_ssl' is a system\r\n library, declare it with 'cpp_info.system_libs' property\r\n```\r\n\r\nI also tried not including hiredis in the requirements and/or configure functions in my conanfile.py. The result was \r\n\r\n```\r\nERROR: There are invalid packages:\r\nredis-plus-plus/1.3.11: Invalid: redis-plus-plus:with_tls=True requires hiredis:with_ssl=True\r\n```\r\n\r\nI'd really appreciate if anyone could guide me in the right direction to get redis-plus-plus built with the with_tls option enabled.\n", "before_files": [{"content": "from conan import ConanFile\nfrom conan.tools.cmake import CMake, CMakeDeps, CMakeToolchain, cmake_layout\nfrom conan.tools.files import apply_conandata_patches, copy, export_conandata_patches, get, rmdir, rm\nfrom conan.tools.scm import Version\nfrom conan.tools.microsoft import is_msvc\nimport os\n\nrequired_conan_version = \">=1.53.0\"\n\n\nclass HiredisConan(ConanFile):\n name = \"hiredis\"\n description = \"Hiredis is a minimalistic C client library for the Redis database.\"\n license = \"BSD-3-Clause\"\n topics = (\"hiredis\", \"redis\", \"client\", \"database\")\n homepage = \"https://github.com/redis/hiredis\"\n url = \"https://github.com/conan-io/conan-center-index\"\n package_type = \"library\"\n settings = \"os\", \"arch\", \"compiler\", \"build_type\"\n options = {\n \"shared\": [True, False],\n \"fPIC\": [True, False],\n \"with_ssl\": [True, False],\n }\n default_options = {\n \"shared\": False,\n \"fPIC\": True,\n \"with_ssl\": False,\n }\n\n def export_sources(self):\n export_conandata_patches(self)\n\n def config_options(self):\n if self.settings.os == \"Windows\":\n del self.options.fPIC\n\n def configure(self):\n if self.options.shared:\n self.options.rm_safe(\"fPIC\")\n self.settings.compiler.rm_safe(\"libcxx\")\n self.settings.compiler.rm_safe(\"cppstd\")\n\n def layout(self):\n cmake_layout(self, src_folder=\"src\")\n\n def requirements(self):\n if self.options.with_ssl:\n self.requires(\"openssl/[>=1.1 <4]\")\n\n def source(self):\n get(self, **self.conan_data[\"sources\"][self.version], strip_root=True)\n\n def generate(self):\n tc = CMakeToolchain(self)\n # Since 1.2.0, BUILD_SHARED_LIBS has been defined by option()\n if Version(self.version) >= \"1.2.0\":\n tc.cache_variables[\"BUILD_SHARED_LIBS\"] = self.options.shared\n tc.variables[\"ENABLE_SSL\"] = self.options.with_ssl\n tc.variables[\"DISABLE_TESTS\"] = True\n tc.variables[\"ENABLE_EXAMPLES\"] = False\n tc.generate()\n deps = CMakeDeps(self)\n deps.generate()\n\n def build(self):\n apply_conandata_patches(self)\n cmake = CMake(self)\n cmake.configure()\n cmake.build()\n\n def package(self):\n copy(self, \"COPYING\", src=self.source_folder, dst=os.path.join(self.package_folder, \"licenses\"))\n cmake = CMake(self)\n cmake.install()\n rmdir(self, os.path.join(self.package_folder, \"lib\", \"pkgconfig\"))\n rmdir(self, os.path.join(self.package_folder, \"lib\", \"cmake\"))\n rmdir(self, os.path.join(self.package_folder, \"share\"))\n rmdir(self, os.path.join(self.package_folder, \"build\"))\n rm(self, \"*.pdb\", os.path.join(self.package_folder, \"bin\"))\n\n def package_info(self):\n self.cpp_info.set_property(\"cmake_file_name\", \"hiredis\")\n\n suffix = \"\"\n if Version(self.version) >= \"1.1.0\":\n if is_msvc(self) and not self.options.shared and Version(self.version) < \"1.2.0\":\n suffix += \"_static\"\n if self.settings.build_type == \"Debug\":\n suffix += \"d\"\n\n # hiredis\n self.cpp_info.components[\"hiredislib\"].set_property(\"cmake_target_name\", \"hiredis::hiredis\")\n self.cpp_info.components[\"hiredislib\"].set_property(\"pkg_config_name\", \"hiredis\")\n self.cpp_info.components[\"hiredislib\"].names[\"cmake_find_package\"] = \"hiredis\"\n self.cpp_info.components[\"hiredislib\"].names[\"cmake_find_package_multi\"] = \"hiredis\"\n self.cpp_info.components[\"hiredislib\"].libs = [f\"hiredis{suffix}\"]\n if self.settings.os == \"Windows\":\n self.cpp_info.components[\"hiredislib\"].system_libs = [\"ws2_32\"]\n # hiredis_ssl\n if self.options.with_ssl:\n self.cpp_info.components[\"hiredis_ssl\"].set_property(\"cmake_target_name\", \"hiredis::hiredis_ssl\")\n self.cpp_info.components[\"hiredis_ssl\"].set_property(\"pkg_config_name\", \"hiredis_ssl\")\n self.cpp_info.components[\"hiredis_ssl\"].names[\"cmake_find_package\"] = \"hiredis_ssl\"\n self.cpp_info.components[\"hiredis_ssl\"].names[\"cmake_find_package_multi\"] = \"hiredis_ssl\"\n self.cpp_info.components[\"hiredis_ssl\"].libs = [f\"hiredis_ssl{suffix}\"]\n self.cpp_info.components[\"hiredis_ssl\"].requires = [\"openssl::ssl\"]\n if self.settings.os == \"Windows\":\n self.cpp_info.components[\"hiredis_ssl\"].requires.append(\"hiredislib\")\n\n # These cmake_target_name and pkg_config_name are unofficial. It avoids conflicts\n # in conan generators between global target/pkg-config and hiredislib component.\n # TODO: eventually remove the cmake_target_name trick if conan can implement smarter logic\n # in CMakeDeps when a downstream recipe requires another recipe globally\n # (link to all components directly instead of global target)\n self.cpp_info.set_property(\"cmake_target_name\", \"hiredis::hiredis_all_unofficial\")\n self.cpp_info.set_property(\"pkg_config_name\", \"hiredis_all_unofficial\")\n", "path": "recipes/hiredis/all/conanfile.py"}]}
| 2,573 | 209 |
gh_patches_debug_36232
|
rasdani/github-patches
|
git_diff
|
pymeasure__pymeasure-350
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Error in examples/Notebook Experiments/script2.ipynb
script.ipynb runs fine but in script2.ipynb I hit the following error at `experiment = Experiment('test', procedure, analyse)`:
```python
C:\ProgramData\Anaconda3\lib\site-packages\matplotlib\__init__.py in __setitem__(self, key, val)
927 raise KeyError(
928 '%s is not a valid rc parameter. See rcParams.keys() for a '
--> 929 'list of valid parameters.' % (key,))
930
931 def __getitem__(self, key):
KeyError: 'axes.color_cycle is not a valid rc parameter. See rcParams.keys() for a list of valid parameters.'
```
Error in examples/Notebook Experiments/script2.ipynb
script.ipynb runs fine but in script2.ipynb I hit the following error at `experiment = Experiment('test', procedure, analyse)`:
```python
C:\ProgramData\Anaconda3\lib\site-packages\matplotlib\__init__.py in __setitem__(self, key, val)
927 raise KeyError(
928 '%s is not a valid rc parameter. See rcParams.keys() for a '
--> 929 'list of valid parameters.' % (key,))
930
931 def __getitem__(self, key):
KeyError: 'axes.color_cycle is not a valid rc parameter. See rcParams.keys() for a list of valid parameters.'
```
</issue>
<code>
[start of pymeasure/experiment/config.py]
1 #
2 # This file is part of the PyMeasure package.
3 #
4 # Copyright (c) 2013-2020 PyMeasure Developers
5 #
6 # Permission is hereby granted, free of charge, to any person obtaining a copy
7 # of this software and associated documentation files (the "Software"), to deal
8 # in the Software without restriction, including without limitation the rights
9 # to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
10 # copies of the Software, and to permit persons to whom the Software is
11 # furnished to do so, subject to the following conditions:
12 #
13 # The above copyright notice and this permission notice shall be included in
14 # all copies or substantial portions of the Software.
15 #
16 # THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
17 # IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
18 # FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
19 # AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
20 # LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
21 # OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
22 # THE SOFTWARE.
23 #
24
25 import configparser
26 import logging
27 import os
28
29 log = logging.getLogger(__name__)
30 log.addHandler(logging.NullHandler())
31
32
33 def set_file(filename):
34 os.environ['CONFIG'] = filename
35
36
37 def get_config(filename='default_config.ini'):
38 if 'CONFIG' in os.environ.keys():
39 filename = os.environ['CONFIG']
40 config = configparser.ConfigParser()
41 config.read(filename)
42 return config
43
44
45 # noinspection PyProtectedMember
46 def set_mpl_rcparams(config):
47 if 'matplotlib.rcParams' in config._sections.keys():
48 import matplotlib
49 for key in config._sections['matplotlib.rcParams']:
50 matplotlib.rcParams[key] = eval(config._sections['matplotlib.rcParams'][key])
51
[end of pymeasure/experiment/config.py]
[start of examples/Notebook Experiments/procedures.py]
1 #
2 # This file is part of the PyMeasure package.
3 #
4 # Copyright (c) 2013-2016 PyMeasure Developers
5 #
6 # Permission is hereby granted, free of charge, to any person obtaining a copy
7 # of this software and associated documentation files (the "Software"), to deal
8 # in the Software without restriction, including without limitation the rights
9 # to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
10 # copies of the Software, and to permit persons to whom the Software is
11 # furnished to do so, subject to the following conditions:
12 #
13 # The above copyright notice and this permission notice shall be included in
14 # all copies or substantial portions of the Software.
15 #
16 # THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
17 # IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
18 # FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
19 # AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
20 # LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
21 # OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
22 # THE SOFTWARE.
23 #
24
25 import random
26 from time import sleep
27 from pymeasure.experiment import Procedure, IntegerParameter, Parameter, FloatParameter
28 import logging
29 log = logging.getLogger(__name__)
30 log.addHandler(logging.NullHandler())
31
32 class TestProcedure(Procedure):
33
34 iterations = IntegerParameter('Loop Iterations', default=100)
35 delay = FloatParameter('Delay Time', units='s', default=0.2)
36 seed = Parameter('Random Seed', default='12345')
37
38 DATA_COLUMNS = ['Iteration', 'Random Number']
39
40 def startup(self):
41 log.info("Setting up random number generator")
42 random.seed(self.seed)
43
44 def execute(self):
45 log.info("Starting to generate numbers")
46 for i in range(self.iterations):
47 data = {
48 'Iteration': i,
49 'Random Number': random.random()
50 }
51 log.debug("Produced numbers: %s" % data)
52 self.emit('results', data)
53 self.emit('progress', 100.*i/self.iterations)
54 sleep(self.delay)
55 if self.should_stop():
56 log.warning("Catch stop command in procedure")
57 break
58
59 def shutdown(self):
60 log.info("Finished")
[end of examples/Notebook Experiments/procedures.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/examples/Notebook Experiments/procedures.py b/examples/Notebook Experiments/procedures.py
deleted file mode 100644
--- a/examples/Notebook Experiments/procedures.py
+++ /dev/null
@@ -1,60 +0,0 @@
-#
-# This file is part of the PyMeasure package.
-#
-# Copyright (c) 2013-2016 PyMeasure Developers
-#
-# Permission is hereby granted, free of charge, to any person obtaining a copy
-# of this software and associated documentation files (the "Software"), to deal
-# in the Software without restriction, including without limitation the rights
-# to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
-# copies of the Software, and to permit persons to whom the Software is
-# furnished to do so, subject to the following conditions:
-#
-# The above copyright notice and this permission notice shall be included in
-# all copies or substantial portions of the Software.
-#
-# THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
-# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
-# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
-# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
-# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
-# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
-# THE SOFTWARE.
-#
-
-import random
-from time import sleep
-from pymeasure.experiment import Procedure, IntegerParameter, Parameter, FloatParameter
-import logging
-log = logging.getLogger(__name__)
-log.addHandler(logging.NullHandler())
-
-class TestProcedure(Procedure):
-
- iterations = IntegerParameter('Loop Iterations', default=100)
- delay = FloatParameter('Delay Time', units='s', default=0.2)
- seed = Parameter('Random Seed', default='12345')
-
- DATA_COLUMNS = ['Iteration', 'Random Number']
-
- def startup(self):
- log.info("Setting up random number generator")
- random.seed(self.seed)
-
- def execute(self):
- log.info("Starting to generate numbers")
- for i in range(self.iterations):
- data = {
- 'Iteration': i,
- 'Random Number': random.random()
- }
- log.debug("Produced numbers: %s" % data)
- self.emit('results', data)
- self.emit('progress', 100.*i/self.iterations)
- sleep(self.delay)
- if self.should_stop():
- log.warning("Catch stop command in procedure")
- break
-
- def shutdown(self):
- log.info("Finished")
\ No newline at end of file
diff --git a/pymeasure/experiment/config.py b/pymeasure/experiment/config.py
--- a/pymeasure/experiment/config.py
+++ b/pymeasure/experiment/config.py
@@ -46,5 +46,6 @@
def set_mpl_rcparams(config):
if 'matplotlib.rcParams' in config._sections.keys():
import matplotlib
+ from cycler import cycler
for key in config._sections['matplotlib.rcParams']:
matplotlib.rcParams[key] = eval(config._sections['matplotlib.rcParams'][key])
|
{"golden_diff": "diff --git a/examples/Notebook Experiments/procedures.py b/examples/Notebook Experiments/procedures.py\ndeleted file mode 100644\n--- a/examples/Notebook Experiments/procedures.py\t\n+++ /dev/null\n@@ -1,60 +0,0 @@\n-#\n-# This file is part of the PyMeasure package.\n-#\n-# Copyright (c) 2013-2016 PyMeasure Developers\n-#\n-# Permission is hereby granted, free of charge, to any person obtaining a copy\n-# of this software and associated documentation files (the \"Software\"), to deal\n-# in the Software without restriction, including without limitation the rights\n-# to use, copy, modify, merge, publish, distribute, sublicense, and/or sell\n-# copies of the Software, and to permit persons to whom the Software is\n-# furnished to do so, subject to the following conditions:\n-#\n-# The above copyright notice and this permission notice shall be included in\n-# all copies or substantial portions of the Software.\n-#\n-# THE SOFTWARE IS PROVIDED \"AS IS\", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR\n-# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,\n-# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE\n-# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER\n-# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,\n-# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN\n-# THE SOFTWARE.\n-#\n-\n-import random\n-from time import sleep\n-from pymeasure.experiment import Procedure, IntegerParameter, Parameter, FloatParameter\n-import logging\n-log = logging.getLogger(__name__)\n-log.addHandler(logging.NullHandler())\n-\n-class TestProcedure(Procedure):\n-\n- iterations = IntegerParameter('Loop Iterations', default=100)\n- delay = FloatParameter('Delay Time', units='s', default=0.2)\n- seed = Parameter('Random Seed', default='12345')\n- \n- DATA_COLUMNS = ['Iteration', 'Random Number']\n-\n- def startup(self):\n- log.info(\"Setting up random number generator\")\n- random.seed(self.seed)\n-\n- def execute(self):\n- log.info(\"Starting to generate numbers\")\n- for i in range(self.iterations):\n- data = {\n- 'Iteration': i,\n- 'Random Number': random.random()\n- }\n- log.debug(\"Produced numbers: %s\" % data)\n- self.emit('results', data)\n- self.emit('progress', 100.*i/self.iterations)\n- sleep(self.delay)\n- if self.should_stop():\n- log.warning(\"Catch stop command in procedure\")\n- break\n-\n- def shutdown(self):\n- log.info(\"Finished\")\n\\ No newline at end of file\ndiff --git a/pymeasure/experiment/config.py b/pymeasure/experiment/config.py\n--- a/pymeasure/experiment/config.py\n+++ b/pymeasure/experiment/config.py\n@@ -46,5 +46,6 @@\n def set_mpl_rcparams(config):\n if 'matplotlib.rcParams' in config._sections.keys():\n import matplotlib\n+ from cycler import cycler\n for key in config._sections['matplotlib.rcParams']:\n matplotlib.rcParams[key] = eval(config._sections['matplotlib.rcParams'][key])\n", "issue": "Error in examples/Notebook Experiments/script2.ipynb\nscript.ipynb runs fine but in script2.ipynb I hit the following error at `experiment = Experiment('test', procedure, analyse)`:\r\n\r\n```python\r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\site-packages\\matplotlib\\__init__.py in __setitem__(self, key, val)\r\n 927 raise KeyError(\r\n 928 '%s is not a valid rc parameter. See rcParams.keys() for a '\r\n--> 929 'list of valid parameters.' % (key,))\r\n 930 \r\n 931 def __getitem__(self, key):\r\n\r\nKeyError: 'axes.color_cycle is not a valid rc parameter. See rcParams.keys() for a list of valid parameters.'\r\n```\r\n\nError in examples/Notebook Experiments/script2.ipynb\nscript.ipynb runs fine but in script2.ipynb I hit the following error at `experiment = Experiment('test', procedure, analyse)`:\r\n\r\n```python\r\n\r\nC:\\ProgramData\\Anaconda3\\lib\\site-packages\\matplotlib\\__init__.py in __setitem__(self, key, val)\r\n 927 raise KeyError(\r\n 928 '%s is not a valid rc parameter. See rcParams.keys() for a '\r\n--> 929 'list of valid parameters.' % (key,))\r\n 930 \r\n 931 def __getitem__(self, key):\r\n\r\nKeyError: 'axes.color_cycle is not a valid rc parameter. See rcParams.keys() for a list of valid parameters.'\r\n```\r\n\n", "before_files": [{"content": "#\n# This file is part of the PyMeasure package.\n#\n# Copyright (c) 2013-2020 PyMeasure Developers\n#\n# Permission is hereby granted, free of charge, to any person obtaining a copy\n# of this software and associated documentation files (the \"Software\"), to deal\n# in the Software without restriction, including without limitation the rights\n# to use, copy, modify, merge, publish, distribute, sublicense, and/or sell\n# copies of the Software, and to permit persons to whom the Software is\n# furnished to do so, subject to the following conditions:\n#\n# The above copyright notice and this permission notice shall be included in\n# all copies or substantial portions of the Software.\n#\n# THE SOFTWARE IS PROVIDED \"AS IS\", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR\n# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,\n# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE\n# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER\n# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,\n# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN\n# THE SOFTWARE.\n#\n\nimport configparser\nimport logging\nimport os\n\nlog = logging.getLogger(__name__)\nlog.addHandler(logging.NullHandler())\n\n\ndef set_file(filename):\n os.environ['CONFIG'] = filename\n\n\ndef get_config(filename='default_config.ini'):\n if 'CONFIG' in os.environ.keys():\n filename = os.environ['CONFIG']\n config = configparser.ConfigParser()\n config.read(filename)\n return config\n\n\n# noinspection PyProtectedMember\ndef set_mpl_rcparams(config):\n if 'matplotlib.rcParams' in config._sections.keys():\n import matplotlib\n for key in config._sections['matplotlib.rcParams']:\n matplotlib.rcParams[key] = eval(config._sections['matplotlib.rcParams'][key])\n", "path": "pymeasure/experiment/config.py"}, {"content": "#\n# This file is part of the PyMeasure package.\n#\n# Copyright (c) 2013-2016 PyMeasure Developers\n#\n# Permission is hereby granted, free of charge, to any person obtaining a copy\n# of this software and associated documentation files (the \"Software\"), to deal\n# in the Software without restriction, including without limitation the rights\n# to use, copy, modify, merge, publish, distribute, sublicense, and/or sell\n# copies of the Software, and to permit persons to whom the Software is\n# furnished to do so, subject to the following conditions:\n#\n# The above copyright notice and this permission notice shall be included in\n# all copies or substantial portions of the Software.\n#\n# THE SOFTWARE IS PROVIDED \"AS IS\", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR\n# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,\n# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE\n# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER\n# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,\n# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN\n# THE SOFTWARE.\n#\n\nimport random\nfrom time import sleep\nfrom pymeasure.experiment import Procedure, IntegerParameter, Parameter, FloatParameter\nimport logging\nlog = logging.getLogger(__name__)\nlog.addHandler(logging.NullHandler())\n\nclass TestProcedure(Procedure):\n\n iterations = IntegerParameter('Loop Iterations', default=100)\n delay = FloatParameter('Delay Time', units='s', default=0.2)\n seed = Parameter('Random Seed', default='12345')\n \n DATA_COLUMNS = ['Iteration', 'Random Number']\n\n def startup(self):\n log.info(\"Setting up random number generator\")\n random.seed(self.seed)\n\n def execute(self):\n log.info(\"Starting to generate numbers\")\n for i in range(self.iterations):\n data = {\n 'Iteration': i,\n 'Random Number': random.random()\n }\n log.debug(\"Produced numbers: %s\" % data)\n self.emit('results', data)\n self.emit('progress', 100.*i/self.iterations)\n sleep(self.delay)\n if self.should_stop():\n log.warning(\"Catch stop command in procedure\")\n break\n\n def shutdown(self):\n log.info(\"Finished\")", "path": "examples/Notebook Experiments/procedures.py"}]}
| 2,026 | 736 |
gh_patches_debug_4412
|
rasdani/github-patches
|
git_diff
|
mytardis__mytardis-1134
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
/search/parameter_field_list/ failures with Unicode first/last names in Users table
This can occur:
```
Internal Server Error: /search/parameter_field_list/
Traceback (most recent call last):
File "/home/mytardis/virtualenvs/mytardis/local/lib/python2.7/site-packages/django/core/handlers/base.py", line 132, in get_response
response = wrapped_callback(request, *callback_args, **callback_kwargs)
File "/home/mytardis/mytardis/tardis/search/views.py", line 282, in retrieve_field_list
fieldList = '+'.join([str(fn) for fn in auto_list])
UnicodeEncodeError: 'ascii' codec can't encode character u'\xfc' in position 8: ordinal not in range(128)
```
if there's a user with a non-ASCII character in their first or last name. The relevant code in tardis/search/views.py is:
```
usernames = [u.first_name + ' ' + u.last_name + ':username' for u in users]
...
auto_list = usernames + searchableFields
fieldList = '+'.join([str(fn) for fn in auto_list])
```
</issue>
<code>
[start of tardis/search/views.py]
1 """
2 views relevant to search
3 """
4 import logging
5 import warnings
6
7 from django.contrib.auth.models import User
8 from django.http import HttpResponse
9 from haystack.generic_views import SearchView
10
11 from tardis.search.forms import GroupedSearchForm
12 from tardis.search.utils import SearchQueryString
13 from tardis.tardis_portal.auth import decorators as authz
14 from tardis.tardis_portal.deprecations import RemovedInMyTardis311Warning
15 from tardis.tardis_portal.forms import createSearchDatafileSelectionForm
16 from tardis.tardis_portal.hacks import oracle_dbops_hack
17 from tardis.tardis_portal.models import Experiment
18 from tardis.tardis_portal.shortcuts import render_response_search, \
19 render_response_index
20 from tardis.tardis_portal.views.utils import __forwardToSearchExperimentFormPage, \
21 __getSearchExperimentForm, __processExperimentParameters, \
22 __getSearchDatafileForm, __processDatafileParameters, \
23 __forwardToSearchDatafileFormPage
24
25 logger = logging.getLogger(__name__)
26
27
28 def getNewSearchDatafileSelectionForm(initial=None):
29 DatafileSelectionForm = createSearchDatafileSelectionForm(initial)
30 return DatafileSelectionForm()
31
32
33 @oracle_dbops_hack
34 def search_experiment(request):
35
36 """Either show the search experiment form or the result of the search
37 experiment query.
38
39 """
40
41 if not request.GET:
42 return __forwardToSearchExperimentFormPage(request)
43
44 form = __getSearchExperimentForm(request)
45 experiments = __processExperimentParameters(request, form)
46
47 # check if the submitted form is valid
48 if experiments is not None:
49 bodyclass = 'list'
50 else:
51 return __forwardToSearchExperimentFormPage(request)
52
53 # remove information from previous searches from session
54 if 'datafileResults' in request.session:
55 del request.session['datafileResults']
56
57 results = []
58 for e in experiments:
59 result = {}
60 result['sr'] = e
61 result['dataset_hit'] = False
62 result['datafile_hit'] = False
63 result['experiment_hit'] = True
64 results.append(result)
65 c = {'header': 'Search Experiment',
66 'experiments': results,
67 'bodyclass': bodyclass}
68 url = 'tardis_portal/search_experiment_results.html'
69 return HttpResponse(render_response_search(request, url, c))
70
71
72 def search_quick(request):
73 get = False
74 experiments = Experiment.objects.all().order_by('title')
75
76 if 'results' in request.GET:
77 get = True
78 if 'quicksearch' in request.GET and request.GET['quicksearch']:
79 experiments = \
80 experiments.filter(
81 title__icontains=request.GET['quicksearch']) | \
82 experiments.filter(
83 institution_name__icontains=request.GET['quicksearch']) | \
84 experiments.filter(
85 experimentauthor__author__name__icontains=request.GET[
86 'quicksearch']) | \
87 experiments.filter(
88 pdbid__pdbid__icontains=request.GET['quicksearch'])
89
90 experiments = experiments.distinct()
91
92 logger.debug(experiments)
93
94 c = {'submitted': get, 'experiments': experiments,
95 'subtitle': 'Search Experiments'}
96 return HttpResponse(render_response_index(request,
97 'tardis_portal/search_experiment.html', c))
98
99
100 def search_datafile(request): # too complex # noqa
101 """Either show the search datafile form or the result of the search
102 datafile query.
103
104 """
105 warnings.warn(
106 "The old DataFile search form (triggered by the /search/datafile/ "
107 "URL) was only useful for X-Ray Diffraction data. It needs to be "
108 "rewritten if it is to be useful for other data types.",
109 RemovedInMyTardis311Warning
110 )
111 if 'type' in request.GET:
112 searchQueryType = request.GET.get('type')
113 else:
114 # for now we'll default to MX if nothing is provided
115 # TODO: should we forward the page to experiment search page if
116 # nothing is provided in the future?
117 searchQueryType = 'mx'
118 logger.info('search_datafile: searchQueryType {0}'.format(searchQueryType))
119 # TODO: check if going to /search/datafile will flag an error in unit test
120 bodyclass = None
121
122 if 'page' not in request.GET and 'type' in request.GET and \
123 len(request.GET) > 1:
124 # display the 1st page of the results
125
126 form = __getSearchDatafileForm(request, searchQueryType)
127 datafile_results = __processDatafileParameters(
128 request, searchQueryType, form)
129 if datafile_results is not None:
130 bodyclass = 'list'
131 else:
132 return __forwardToSearchDatafileFormPage(
133 request, searchQueryType, form)
134
135 else:
136 if 'page' in request.GET:
137 # succeeding pages of pagination
138 if 'datafileResults' in request.session:
139 datafile_results = request.session['datafileResults']
140 else:
141 form = __getSearchDatafileForm(request, searchQueryType)
142 datafile_results = __processDatafileParameters(
143 request, searchQueryType, form)
144 if datafile_results is not None:
145 bodyclass = 'list'
146 else:
147 return __forwardToSearchDatafileFormPage(
148 request, searchQueryType, form)
149 else:
150 # display the form
151 if 'datafileResults' in request.session:
152 del request.session['datafileResults']
153 return __forwardToSearchDatafileFormPage(request, searchQueryType)
154
155 # process the files to be displayed by the paginator...
156 # paginator = Paginator(datafile_results,
157 # constants.DATAFILE_RESULTS_PER_PAGE)
158
159 # try:
160 # page = int(request.GET.get('page', '1'))
161 # except ValueError:
162 # page = 1
163
164 # If page request (9999) is out of :range, deliver last page of results.
165 # try:
166 # datafiles = paginator.page(page)
167 # except (EmptyPage, InvalidPage):
168 # datafiles = paginator.page(paginator.num_pages)
169
170 import re
171 cleanedUpQueryString = re.sub('&page=\d+', '',
172 request.META['QUERY_STRING'])
173
174 # get experiments associated with datafiles
175 if datafile_results:
176 experiment_pks = list(set(datafile_results.values_list(
177 'dataset__experiments', flat=True)))
178 experiments = Experiment.safe.in_bulk(experiment_pks)
179 else:
180 experiments = {}
181
182 results = []
183 for key, e in experiments.items():
184 result = {}
185 result['sr'] = e
186 result['dataset_hit'] = False
187 result['datafile_hit'] = True
188 result['experiment_hit'] = False
189 results.append(result)
190
191 c = {
192 'experiments': results,
193 'datafiles': datafile_results,
194 # 'paginator': paginator,
195 'query_string': cleanedUpQueryString,
196 'subtitle': 'Search Datafiles',
197 'nav': [{'name': 'Search Datafile', 'link': '/search/datafile/'}],
198 'bodyclass': bodyclass,
199 'search_pressed': True,
200 'searchDatafileSelectionForm': getNewSearchDatafileSelectionForm()}
201 url = 'tardis_portal/search_experiment_results.html'
202 return HttpResponse(render_response_search(request, url, c))
203
204
205 class ExperimentSearchView(SearchView):
206 def __name__(self):
207 return "ExperimentSearchView"
208
209 def extra_context(self):
210 extra = super(ExperimentSearchView, self).extra_context()
211 # Results may contain Experiments, Datasets and DataFiles.
212 # Group them into experiments, noting whether or not the search
213 # hits were in the Dataset(s) or DataFile(s)
214 results = self.results
215 facets = results.facet_counts()
216 if facets:
217 experiment_facets = facets['fields']['experiment_id_stored']
218 experiment_ids = [int(f[0])
219 for f in experiment_facets if int(f[1]) > 0]
220 else:
221 experiment_ids = []
222
223 access_list = []
224
225 if self.request.user.is_authenticated():
226 access_list.extend(
227 [e.pk for e in
228 authz.get_accessible_experiments(self.request)])
229
230 access_list.extend(
231 [e.pk for e in Experiment.objects
232 .exclude(public_access=Experiment.PUBLIC_ACCESS_NONE)
233 .exclude(public_access=Experiment.PUBLIC_ACCESS_EMBARGO)])
234
235 ids = list(set(experiment_ids) & set(access_list))
236 experiments = Experiment.objects.filter(pk__in=ids)\
237 .order_by('-update_time')
238
239 results = []
240 for e in experiments:
241 result = {'sr': e, 'dataset_hit': False, 'datafile_hit': False,
242 'experiment_hit': False}
243 results.append(result)
244
245 extra['experiments'] = results
246 return extra
247
248 # override SearchView's method in order to
249 # return a ResponseContext
250 def create_response(self):
251 (paginator, page) = self.build_page()
252
253 # Remove unnecessary whitespace
254 # TODO this should just be done in the form clean...
255 query = SearchQueryString(self.query)
256 context = {
257 'search_query': query,
258 'form': self.form,
259 'page': page,
260 'paginator': paginator,
261 }
262 context.update(self.extra_context())
263
264 return render_response_index(self.request, self.template, context)
265
266
267 def retrieve_field_list(request):
268
269 from tardis.search.search_indexes import DataFileIndex
270
271 # Get all of the fields in the indexes
272 #
273 # TODO: these should be onl read from registered indexes
274 #
275 allFields = DataFileIndex.fields.items()
276
277 users = User.objects.all()
278
279 usernames = [u.first_name + ' ' + u.last_name + ':username' for u in users]
280
281 # Collect all of the indexed (searchable) fields, except
282 # for the main search document ('text')
283 searchableFields = ([key + ':search_field' for key, f in allFields
284 if f.indexed is True and key != 'text'])
285
286 auto_list = usernames + searchableFields
287
288 fieldList = '+'.join([str(fn) for fn in auto_list])
289 return HttpResponse(fieldList)
290
291
292 class SingleSearchView(SearchView):
293 form_class = GroupedSearchForm
294 template_name = 'search/search.html'
295
296 def form_valid(self, form):
297 sqs = form.search(user=self.request.user)
298 context = self.get_context_data(**{
299 self.form_name: form,
300 'query': form.cleaned_data.get(self.search_field),
301 'object_list': sqs,
302 })
303 return self.render_to_response(context)
304
[end of tardis/search/views.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/tardis/search/views.py b/tardis/search/views.py
--- a/tardis/search/views.py
+++ b/tardis/search/views.py
@@ -276,7 +276,11 @@
users = User.objects.all()
- usernames = [u.first_name + ' ' + u.last_name + ':username' for u in users]
+ usernames = [
+ u.first_name.encode('utf-8') +
+ ' ' +
+ u.last_name.encode('utf-8') +
+ ':username' for u in users]
# Collect all of the indexed (searchable) fields, except
# for the main search document ('text')
|
{"golden_diff": "diff --git a/tardis/search/views.py b/tardis/search/views.py\n--- a/tardis/search/views.py\n+++ b/tardis/search/views.py\n@@ -276,7 +276,11 @@\n \n users = User.objects.all()\n \n- usernames = [u.first_name + ' ' + u.last_name + ':username' for u in users]\n+ usernames = [\n+ u.first_name.encode('utf-8') +\n+ ' ' +\n+ u.last_name.encode('utf-8') +\n+ ':username' for u in users]\n \n # Collect all of the indexed (searchable) fields, except\n # for the main search document ('text')\n", "issue": "/search/parameter_field_list/ failures with Unicode first/last names in Users table\nThis can occur:\r\n```\r\nInternal Server Error: /search/parameter_field_list/\r\nTraceback (most recent call last):\r\n File \"/home/mytardis/virtualenvs/mytardis/local/lib/python2.7/site-packages/django/core/handlers/base.py\", line 132, in get_response\r\n response = wrapped_callback(request, *callback_args, **callback_kwargs)\r\n File \"/home/mytardis/mytardis/tardis/search/views.py\", line 282, in retrieve_field_list\r\n fieldList = '+'.join([str(fn) for fn in auto_list])\r\nUnicodeEncodeError: 'ascii' codec can't encode character u'\\xfc' in position 8: ordinal not in range(128)\r\n```\r\n\r\nif there's a user with a non-ASCII character in their first or last name. The relevant code in tardis/search/views.py is:\r\n\r\n```\r\nusernames = [u.first_name + ' ' + u.last_name + ':username' for u in users]\r\n...\r\nauto_list = usernames + searchableFields\r\nfieldList = '+'.join([str(fn) for fn in auto_list])\r\n```\r\n\n", "before_files": [{"content": "\"\"\"\nviews relevant to search\n\"\"\"\nimport logging\nimport warnings\n\nfrom django.contrib.auth.models import User\nfrom django.http import HttpResponse\nfrom haystack.generic_views import SearchView\n\nfrom tardis.search.forms import GroupedSearchForm\nfrom tardis.search.utils import SearchQueryString\nfrom tardis.tardis_portal.auth import decorators as authz\nfrom tardis.tardis_portal.deprecations import RemovedInMyTardis311Warning\nfrom tardis.tardis_portal.forms import createSearchDatafileSelectionForm\nfrom tardis.tardis_portal.hacks import oracle_dbops_hack\nfrom tardis.tardis_portal.models import Experiment\nfrom tardis.tardis_portal.shortcuts import render_response_search, \\\n render_response_index\nfrom tardis.tardis_portal.views.utils import __forwardToSearchExperimentFormPage, \\\n __getSearchExperimentForm, __processExperimentParameters, \\\n __getSearchDatafileForm, __processDatafileParameters, \\\n __forwardToSearchDatafileFormPage\n\nlogger = logging.getLogger(__name__)\n\n\ndef getNewSearchDatafileSelectionForm(initial=None):\n DatafileSelectionForm = createSearchDatafileSelectionForm(initial)\n return DatafileSelectionForm()\n\n\n@oracle_dbops_hack\ndef search_experiment(request):\n\n \"\"\"Either show the search experiment form or the result of the search\n experiment query.\n\n \"\"\"\n\n if not request.GET:\n return __forwardToSearchExperimentFormPage(request)\n\n form = __getSearchExperimentForm(request)\n experiments = __processExperimentParameters(request, form)\n\n # check if the submitted form is valid\n if experiments is not None:\n bodyclass = 'list'\n else:\n return __forwardToSearchExperimentFormPage(request)\n\n # remove information from previous searches from session\n if 'datafileResults' in request.session:\n del request.session['datafileResults']\n\n results = []\n for e in experiments:\n result = {}\n result['sr'] = e\n result['dataset_hit'] = False\n result['datafile_hit'] = False\n result['experiment_hit'] = True\n results.append(result)\n c = {'header': 'Search Experiment',\n 'experiments': results,\n 'bodyclass': bodyclass}\n url = 'tardis_portal/search_experiment_results.html'\n return HttpResponse(render_response_search(request, url, c))\n\n\ndef search_quick(request):\n get = False\n experiments = Experiment.objects.all().order_by('title')\n\n if 'results' in request.GET:\n get = True\n if 'quicksearch' in request.GET and request.GET['quicksearch']:\n experiments = \\\n experiments.filter(\n title__icontains=request.GET['quicksearch']) | \\\n experiments.filter(\n institution_name__icontains=request.GET['quicksearch']) | \\\n experiments.filter(\n experimentauthor__author__name__icontains=request.GET[\n 'quicksearch']) | \\\n experiments.filter(\n pdbid__pdbid__icontains=request.GET['quicksearch'])\n\n experiments = experiments.distinct()\n\n logger.debug(experiments)\n\n c = {'submitted': get, 'experiments': experiments,\n 'subtitle': 'Search Experiments'}\n return HttpResponse(render_response_index(request,\n 'tardis_portal/search_experiment.html', c))\n\n\ndef search_datafile(request): # too complex # noqa\n \"\"\"Either show the search datafile form or the result of the search\n datafile query.\n\n \"\"\"\n warnings.warn(\n \"The old DataFile search form (triggered by the /search/datafile/ \"\n \"URL) was only useful for X-Ray Diffraction data. It needs to be \"\n \"rewritten if it is to be useful for other data types.\",\n RemovedInMyTardis311Warning\n )\n if 'type' in request.GET:\n searchQueryType = request.GET.get('type')\n else:\n # for now we'll default to MX if nothing is provided\n # TODO: should we forward the page to experiment search page if\n # nothing is provided in the future?\n searchQueryType = 'mx'\n logger.info('search_datafile: searchQueryType {0}'.format(searchQueryType))\n # TODO: check if going to /search/datafile will flag an error in unit test\n bodyclass = None\n\n if 'page' not in request.GET and 'type' in request.GET and \\\n len(request.GET) > 1:\n # display the 1st page of the results\n\n form = __getSearchDatafileForm(request, searchQueryType)\n datafile_results = __processDatafileParameters(\n request, searchQueryType, form)\n if datafile_results is not None:\n bodyclass = 'list'\n else:\n return __forwardToSearchDatafileFormPage(\n request, searchQueryType, form)\n\n else:\n if 'page' in request.GET:\n # succeeding pages of pagination\n if 'datafileResults' in request.session:\n datafile_results = request.session['datafileResults']\n else:\n form = __getSearchDatafileForm(request, searchQueryType)\n datafile_results = __processDatafileParameters(\n request, searchQueryType, form)\n if datafile_results is not None:\n bodyclass = 'list'\n else:\n return __forwardToSearchDatafileFormPage(\n request, searchQueryType, form)\n else:\n # display the form\n if 'datafileResults' in request.session:\n del request.session['datafileResults']\n return __forwardToSearchDatafileFormPage(request, searchQueryType)\n\n # process the files to be displayed by the paginator...\n # paginator = Paginator(datafile_results,\n # constants.DATAFILE_RESULTS_PER_PAGE)\n\n # try:\n # page = int(request.GET.get('page', '1'))\n # except ValueError:\n # page = 1\n\n # If page request (9999) is out of :range, deliver last page of results.\n # try:\n # datafiles = paginator.page(page)\n # except (EmptyPage, InvalidPage):\n # datafiles = paginator.page(paginator.num_pages)\n\n import re\n cleanedUpQueryString = re.sub('&page=\\d+', '',\n request.META['QUERY_STRING'])\n\n # get experiments associated with datafiles\n if datafile_results:\n experiment_pks = list(set(datafile_results.values_list(\n 'dataset__experiments', flat=True)))\n experiments = Experiment.safe.in_bulk(experiment_pks)\n else:\n experiments = {}\n\n results = []\n for key, e in experiments.items():\n result = {}\n result['sr'] = e\n result['dataset_hit'] = False\n result['datafile_hit'] = True\n result['experiment_hit'] = False\n results.append(result)\n\n c = {\n 'experiments': results,\n 'datafiles': datafile_results,\n # 'paginator': paginator,\n 'query_string': cleanedUpQueryString,\n 'subtitle': 'Search Datafiles',\n 'nav': [{'name': 'Search Datafile', 'link': '/search/datafile/'}],\n 'bodyclass': bodyclass,\n 'search_pressed': True,\n 'searchDatafileSelectionForm': getNewSearchDatafileSelectionForm()}\n url = 'tardis_portal/search_experiment_results.html'\n return HttpResponse(render_response_search(request, url, c))\n\n\nclass ExperimentSearchView(SearchView):\n def __name__(self):\n return \"ExperimentSearchView\"\n\n def extra_context(self):\n extra = super(ExperimentSearchView, self).extra_context()\n # Results may contain Experiments, Datasets and DataFiles.\n # Group them into experiments, noting whether or not the search\n # hits were in the Dataset(s) or DataFile(s)\n results = self.results\n facets = results.facet_counts()\n if facets:\n experiment_facets = facets['fields']['experiment_id_stored']\n experiment_ids = [int(f[0])\n for f in experiment_facets if int(f[1]) > 0]\n else:\n experiment_ids = []\n\n access_list = []\n\n if self.request.user.is_authenticated():\n access_list.extend(\n [e.pk for e in\n authz.get_accessible_experiments(self.request)])\n\n access_list.extend(\n [e.pk for e in Experiment.objects\n .exclude(public_access=Experiment.PUBLIC_ACCESS_NONE)\n .exclude(public_access=Experiment.PUBLIC_ACCESS_EMBARGO)])\n\n ids = list(set(experiment_ids) & set(access_list))\n experiments = Experiment.objects.filter(pk__in=ids)\\\n .order_by('-update_time')\n\n results = []\n for e in experiments:\n result = {'sr': e, 'dataset_hit': False, 'datafile_hit': False,\n 'experiment_hit': False}\n results.append(result)\n\n extra['experiments'] = results\n return extra\n\n # override SearchView's method in order to\n # return a ResponseContext\n def create_response(self):\n (paginator, page) = self.build_page()\n\n # Remove unnecessary whitespace\n # TODO this should just be done in the form clean...\n query = SearchQueryString(self.query)\n context = {\n 'search_query': query,\n 'form': self.form,\n 'page': page,\n 'paginator': paginator,\n }\n context.update(self.extra_context())\n\n return render_response_index(self.request, self.template, context)\n\n\ndef retrieve_field_list(request):\n\n from tardis.search.search_indexes import DataFileIndex\n\n # Get all of the fields in the indexes\n #\n # TODO: these should be onl read from registered indexes\n #\n allFields = DataFileIndex.fields.items()\n\n users = User.objects.all()\n\n usernames = [u.first_name + ' ' + u.last_name + ':username' for u in users]\n\n # Collect all of the indexed (searchable) fields, except\n # for the main search document ('text')\n searchableFields = ([key + ':search_field' for key, f in allFields\n if f.indexed is True and key != 'text'])\n\n auto_list = usernames + searchableFields\n\n fieldList = '+'.join([str(fn) for fn in auto_list])\n return HttpResponse(fieldList)\n\n\nclass SingleSearchView(SearchView):\n form_class = GroupedSearchForm\n template_name = 'search/search.html'\n\n def form_valid(self, form):\n sqs = form.search(user=self.request.user)\n context = self.get_context_data(**{\n self.form_name: form,\n 'query': form.cleaned_data.get(self.search_field),\n 'object_list': sqs,\n })\n return self.render_to_response(context)\n", "path": "tardis/search/views.py"}]}
| 3,921 | 154 |
gh_patches_debug_28969
|
rasdani/github-patches
|
git_diff
|
optuna__optuna-4077
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
The plot and report of Bayesmark scripts are inconsistent
### Expected behavior
Consistent results in the plots and `report.md`.
### Environment
GitHub Actions: https://github.com/optuna/optuna/blob/1a520bd5daa9ff0af09fb060464bb157f8af891b/.github/workflows/performance-benchmarks-bayesmark.yml
### Error messages, stack traces, or logs
```shell
https://github.com/optuna/optuna/actions/runs/3143078497
(See Additional context)
```
### Steps to reproduce
Run `Actions -> Performance benchmarks with bayesmark` with the default arguments.
### Additional context (optional)
The values in the plots and `report.md` seem to be inconsistent.
Example:


</issue>
<code>
[start of benchmarks/run_bayesmark.py]
1 import argparse
2 import json
3 import os
4 import subprocess
5 from typing import Dict
6 from typing import List
7
8 from matplotlib import cm
9 from matplotlib import colors
10 from matplotlib.axes import Axes
11 import matplotlib.pyplot as plt
12 import numpy as np
13 import pandas as pd
14 from xarray import Dataset
15
16
17 _DB = "bo_optuna_run"
18
19
20 def run_benchmark(args: argparse.Namespace) -> None:
21
22 sampler_list = args.sampler_list.split()
23 sampler_kwargs_list = args.sampler_kwargs_list.split()
24 pruner_list = args.pruner_list.split()
25 pruner_kwargs_list = args.pruner_kwargs_list.split()
26
27 config = dict()
28 for sampler, sampler_kwargs in zip(sampler_list, sampler_kwargs_list):
29 for pruner, pruner_kwargs in zip(pruner_list, pruner_kwargs_list):
30 optimizer_name = f"{sampler}-{pruner}"
31 optimizer_kwargs = {
32 "sampler": sampler,
33 "sampler_kwargs": json.loads(sampler_kwargs),
34 "pruner": pruner,
35 "pruner_kwargs": json.loads(pruner_kwargs),
36 }
37 # We need to dynamically generate config.json sice sampler pruner combos (solvers)
38 # are parametrized by user. Following sample config schema.
39 # https://github.com/uber/bayesmark/blob/master/example_opt_root/config.json
40 config[optimizer_name] = ["optuna_optimizer.py", optimizer_kwargs]
41
42 with open(os.path.join("benchmarks", "bayesmark", "config.json"), "w") as file:
43 json.dump(config, file, indent=4)
44
45 samplers = " ".join(config.keys())
46 metric = "nll" if args.dataset in ["breast", "iris", "wine", "digits"] else "mse"
47 cmd = (
48 f"bayesmark-launch -n {args.budget} -r {args.n_runs} "
49 f"-dir runs -b {_DB} "
50 f"-o {samplers} "
51 f"-c {args.model} -d {args.dataset} "
52 f"-m {metric} --opt-root benchmarks/bayesmark"
53 )
54 subprocess.run(cmd, shell=True)
55
56
57 def make_plots(args: argparse.Namespace) -> None:
58
59 filename = f"{args.dataset}-{args.model}-partial-report.json"
60 df = pd.read_json(os.path.join("partial", filename))
61 summaries = (
62 df.groupby(["opt", "iter"])
63 .generalization.agg(["mean", "std"])
64 .rename(columns={"mean": "best_mean", "std": "best_std"})
65 .reset_index()
66 )
67
68 fig, ax = plt.subplots()
69 fig.set_size_inches(12, 8)
70 warmup = json.loads(args.plot_warmup)
71 metric = df.metric[0]
72 color_lookup = build_color_dict(sorted(df["opt"].unique()))
73
74 for optimizer, summary in summaries.groupby("opt"):
75 color = color_lookup[optimizer]
76 make_plot(summary, ax, optimizer, metric, warmup, color)
77
78 handles, labels = ax.get_legend_handles_labels()
79 fig.legend(handles, labels)
80 fig.suptitle(f"Bayesmark-{args.dataset.capitalize()}-{args.model}")
81 fig.savefig(os.path.join("plots", f"optuna-{args.dataset}-{args.model}-sumamry.png"))
82
83
84 def make_plot(
85 summary: pd.DataFrame,
86 ax: Axes,
87 optimizer: str,
88 metric: str,
89 plot_warmup: bool,
90 color: np.ndarray,
91 ) -> None:
92
93 start = 0 if plot_warmup else 10
94 argpos = summary.best_mean.expanding().apply(np.argmin).astype(int)
95 best_found = summary.best_mean.values[argpos.values]
96 sdev = summary.best_std.values[argpos.values]
97
98 if len(best_found) <= start:
99 return
100
101 ax.fill_between(
102 np.arange(len(best_found))[start:],
103 (best_found - sdev)[start:],
104 (best_found + sdev)[start:],
105 color=color,
106 alpha=0.25,
107 step="mid",
108 )
109
110 ax.plot(
111 np.arange(len(best_found))[start:],
112 best_found[start:],
113 color=color,
114 label=optimizer,
115 drawstyle="steps-mid",
116 )
117
118 ax.set_xlabel("Budget", fontsize=10)
119 ax.set_ylabel(f"Validation {metric.upper()}", fontsize=10)
120 ax.grid(alpha=0.2)
121
122
123 def build_color_dict(names: List[str]) -> Dict[str, np.ndarray]:
124
125 norm = colors.Normalize(vmin=0, vmax=1)
126 m = cm.ScalarMappable(norm, cm.tab20)
127 color_dict = m.to_rgba(np.linspace(0, 1, len(names)))
128 color_dict = dict(zip(names, color_dict))
129
130 return color_dict
131
132
133 def partial_report(args: argparse.Namespace) -> None:
134
135 eval_path = os.path.join("runs", _DB, "eval")
136 time_path = os.path.join("runs", _DB, "time")
137 studies = os.listdir(eval_path)
138 summaries: List[pd.DataFrame] = []
139
140 for study in studies:
141 table_buffer: List[pd.DataFrame] = []
142 column_buffer: List[str] = []
143 for path in [eval_path, time_path]:
144 with open(os.path.join(path, study), "r") as file:
145 data = json.load(file)
146 df = Dataset.from_dict(data["data"]).to_dataframe()
147 df = df.droplevel("suggestion")
148
149 for argument, meatadata in data["meta"]["args"].items():
150 colname = argument[2:] if argument.startswith("--") else argument
151 if colname not in column_buffer:
152 df[colname] = meatadata
153 column_buffer.append(colname)
154
155 table_buffer.append(df)
156
157 summary = pd.merge(*table_buffer, left_index=True, right_index=True)
158 summaries.append(summary.reset_index())
159
160 filename = f"{args.dataset}-{args.model}-partial-report.json"
161 pd.concat(summaries).reset_index(drop=True).to_json(os.path.join("partial", filename))
162
163
164 if __name__ == "__main__":
165 parser = argparse.ArgumentParser()
166 parser.add_argument("--dataset", type=str, default="iris")
167 parser.add_argument("--model", type=str, default="kNN")
168 parser.add_argument("--budget", type=int, default=80)
169 parser.add_argument("--n-runs", type=int, default=10)
170 parser.add_argument("--sampler-list", type=str, default="TPESampler CmaEsSampler")
171 parser.add_argument(
172 "--sampler-kwargs-list",
173 type=str,
174 default='{"multivariate":true,"constant_liar":true} {}',
175 )
176 parser.add_argument("--pruner-list", type=str, default="NopPruner")
177 parser.add_argument("--pruner-kwargs-list", type=str, default="{}")
178 parser.add_argument("--plot-warmup", type=str, default="true")
179
180 args = parser.parse_args()
181 os.makedirs("runs", exist_ok=True)
182 os.makedirs("plots", exist_ok=True)
183 os.makedirs("partial", exist_ok=True)
184
185 run_benchmark(args)
186 partial_report(args)
187 make_plots(args)
188
[end of benchmarks/run_bayesmark.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/benchmarks/run_bayesmark.py b/benchmarks/run_bayesmark.py
--- a/benchmarks/run_bayesmark.py
+++ b/benchmarks/run_bayesmark.py
@@ -58,9 +58,10 @@
filename = f"{args.dataset}-{args.model}-partial-report.json"
df = pd.read_json(os.path.join("partial", filename))
+ df["best_value"] = df.groupby(["opt", "uuid"]).generalization.cummin()
summaries = (
df.groupby(["opt", "iter"])
- .generalization.agg(["mean", "std"])
+ .best_value.agg(["mean", "std"])
.rename(columns={"mean": "best_mean", "std": "best_std"})
.reset_index()
)
@@ -91,25 +92,21 @@
) -> None:
start = 0 if plot_warmup else 10
- argpos = summary.best_mean.expanding().apply(np.argmin).astype(int)
- best_found = summary.best_mean.values[argpos.values]
- sdev = summary.best_std.values[argpos.values]
-
- if len(best_found) <= start:
+ if len(summary.best_mean) <= start:
return
ax.fill_between(
- np.arange(len(best_found))[start:],
- (best_found - sdev)[start:],
- (best_found + sdev)[start:],
+ np.arange(len(summary.best_mean))[start:],
+ (summary.best_mean - summary.best_std)[start:],
+ (summary.best_mean + summary.best_std)[start:],
color=color,
alpha=0.25,
step="mid",
)
ax.plot(
- np.arange(len(best_found))[start:],
- best_found[start:],
+ np.arange(len(summary.best_mean))[start:],
+ summary.best_mean[start:],
color=color,
label=optimizer,
drawstyle="steps-mid",
|
{"golden_diff": "diff --git a/benchmarks/run_bayesmark.py b/benchmarks/run_bayesmark.py\n--- a/benchmarks/run_bayesmark.py\n+++ b/benchmarks/run_bayesmark.py\n@@ -58,9 +58,10 @@\n \n filename = f\"{args.dataset}-{args.model}-partial-report.json\"\n df = pd.read_json(os.path.join(\"partial\", filename))\n+ df[\"best_value\"] = df.groupby([\"opt\", \"uuid\"]).generalization.cummin()\n summaries = (\n df.groupby([\"opt\", \"iter\"])\n- .generalization.agg([\"mean\", \"std\"])\n+ .best_value.agg([\"mean\", \"std\"])\n .rename(columns={\"mean\": \"best_mean\", \"std\": \"best_std\"})\n .reset_index()\n )\n@@ -91,25 +92,21 @@\n ) -> None:\n \n start = 0 if plot_warmup else 10\n- argpos = summary.best_mean.expanding().apply(np.argmin).astype(int)\n- best_found = summary.best_mean.values[argpos.values]\n- sdev = summary.best_std.values[argpos.values]\n-\n- if len(best_found) <= start:\n+ if len(summary.best_mean) <= start:\n return\n \n ax.fill_between(\n- np.arange(len(best_found))[start:],\n- (best_found - sdev)[start:],\n- (best_found + sdev)[start:],\n+ np.arange(len(summary.best_mean))[start:],\n+ (summary.best_mean - summary.best_std)[start:],\n+ (summary.best_mean + summary.best_std)[start:],\n color=color,\n alpha=0.25,\n step=\"mid\",\n )\n \n ax.plot(\n- np.arange(len(best_found))[start:],\n- best_found[start:],\n+ np.arange(len(summary.best_mean))[start:],\n+ summary.best_mean[start:],\n color=color,\n label=optimizer,\n drawstyle=\"steps-mid\",\n", "issue": "The plot and report of Bayesmark scripts are inconsistent\n### Expected behavior\r\n\r\nConsistent results in the plots and `report.md`.\r\n\r\n### Environment\r\n\r\nGitHub Actions: https://github.com/optuna/optuna/blob/1a520bd5daa9ff0af09fb060464bb157f8af891b/.github/workflows/performance-benchmarks-bayesmark.yml\r\n\r\n### Error messages, stack traces, or logs\r\n\r\n```shell\r\nhttps://github.com/optuna/optuna/actions/runs/3143078497\r\n\r\n(See Additional context)\r\n```\r\n\r\n\r\n### Steps to reproduce\r\n\r\nRun `Actions -> Performance benchmarks with bayesmark` with the default arguments.\r\n\r\n### Additional context (optional)\r\n\r\nThe values in the plots and `report.md` seem to be inconsistent.\r\nExample:\r\n\r\n\n", "before_files": [{"content": "import argparse\nimport json\nimport os\nimport subprocess\nfrom typing import Dict\nfrom typing import List\n\nfrom matplotlib import cm\nfrom matplotlib import colors\nfrom matplotlib.axes import Axes\nimport matplotlib.pyplot as plt\nimport numpy as np\nimport pandas as pd\nfrom xarray import Dataset\n\n\n_DB = \"bo_optuna_run\"\n\n\ndef run_benchmark(args: argparse.Namespace) -> None:\n\n sampler_list = args.sampler_list.split()\n sampler_kwargs_list = args.sampler_kwargs_list.split()\n pruner_list = args.pruner_list.split()\n pruner_kwargs_list = args.pruner_kwargs_list.split()\n\n config = dict()\n for sampler, sampler_kwargs in zip(sampler_list, sampler_kwargs_list):\n for pruner, pruner_kwargs in zip(pruner_list, pruner_kwargs_list):\n optimizer_name = f\"{sampler}-{pruner}\"\n optimizer_kwargs = {\n \"sampler\": sampler,\n \"sampler_kwargs\": json.loads(sampler_kwargs),\n \"pruner\": pruner,\n \"pruner_kwargs\": json.loads(pruner_kwargs),\n }\n # We need to dynamically generate config.json sice sampler pruner combos (solvers)\n # are parametrized by user. Following sample config schema.\n # https://github.com/uber/bayesmark/blob/master/example_opt_root/config.json\n config[optimizer_name] = [\"optuna_optimizer.py\", optimizer_kwargs]\n\n with open(os.path.join(\"benchmarks\", \"bayesmark\", \"config.json\"), \"w\") as file:\n json.dump(config, file, indent=4)\n\n samplers = \" \".join(config.keys())\n metric = \"nll\" if args.dataset in [\"breast\", \"iris\", \"wine\", \"digits\"] else \"mse\"\n cmd = (\n f\"bayesmark-launch -n {args.budget} -r {args.n_runs} \"\n f\"-dir runs -b {_DB} \"\n f\"-o {samplers} \"\n f\"-c {args.model} -d {args.dataset} \"\n f\"-m {metric} --opt-root benchmarks/bayesmark\"\n )\n subprocess.run(cmd, shell=True)\n\n\ndef make_plots(args: argparse.Namespace) -> None:\n\n filename = f\"{args.dataset}-{args.model}-partial-report.json\"\n df = pd.read_json(os.path.join(\"partial\", filename))\n summaries = (\n df.groupby([\"opt\", \"iter\"])\n .generalization.agg([\"mean\", \"std\"])\n .rename(columns={\"mean\": \"best_mean\", \"std\": \"best_std\"})\n .reset_index()\n )\n\n fig, ax = plt.subplots()\n fig.set_size_inches(12, 8)\n warmup = json.loads(args.plot_warmup)\n metric = df.metric[0]\n color_lookup = build_color_dict(sorted(df[\"opt\"].unique()))\n\n for optimizer, summary in summaries.groupby(\"opt\"):\n color = color_lookup[optimizer]\n make_plot(summary, ax, optimizer, metric, warmup, color)\n\n handles, labels = ax.get_legend_handles_labels()\n fig.legend(handles, labels)\n fig.suptitle(f\"Bayesmark-{args.dataset.capitalize()}-{args.model}\")\n fig.savefig(os.path.join(\"plots\", f\"optuna-{args.dataset}-{args.model}-sumamry.png\"))\n\n\ndef make_plot(\n summary: pd.DataFrame,\n ax: Axes,\n optimizer: str,\n metric: str,\n plot_warmup: bool,\n color: np.ndarray,\n) -> None:\n\n start = 0 if plot_warmup else 10\n argpos = summary.best_mean.expanding().apply(np.argmin).astype(int)\n best_found = summary.best_mean.values[argpos.values]\n sdev = summary.best_std.values[argpos.values]\n\n if len(best_found) <= start:\n return\n\n ax.fill_between(\n np.arange(len(best_found))[start:],\n (best_found - sdev)[start:],\n (best_found + sdev)[start:],\n color=color,\n alpha=0.25,\n step=\"mid\",\n )\n\n ax.plot(\n np.arange(len(best_found))[start:],\n best_found[start:],\n color=color,\n label=optimizer,\n drawstyle=\"steps-mid\",\n )\n\n ax.set_xlabel(\"Budget\", fontsize=10)\n ax.set_ylabel(f\"Validation {metric.upper()}\", fontsize=10)\n ax.grid(alpha=0.2)\n\n\ndef build_color_dict(names: List[str]) -> Dict[str, np.ndarray]:\n\n norm = colors.Normalize(vmin=0, vmax=1)\n m = cm.ScalarMappable(norm, cm.tab20)\n color_dict = m.to_rgba(np.linspace(0, 1, len(names)))\n color_dict = dict(zip(names, color_dict))\n\n return color_dict\n\n\ndef partial_report(args: argparse.Namespace) -> None:\n\n eval_path = os.path.join(\"runs\", _DB, \"eval\")\n time_path = os.path.join(\"runs\", _DB, \"time\")\n studies = os.listdir(eval_path)\n summaries: List[pd.DataFrame] = []\n\n for study in studies:\n table_buffer: List[pd.DataFrame] = []\n column_buffer: List[str] = []\n for path in [eval_path, time_path]:\n with open(os.path.join(path, study), \"r\") as file:\n data = json.load(file)\n df = Dataset.from_dict(data[\"data\"]).to_dataframe()\n df = df.droplevel(\"suggestion\")\n\n for argument, meatadata in data[\"meta\"][\"args\"].items():\n colname = argument[2:] if argument.startswith(\"--\") else argument\n if colname not in column_buffer:\n df[colname] = meatadata\n column_buffer.append(colname)\n\n table_buffer.append(df)\n\n summary = pd.merge(*table_buffer, left_index=True, right_index=True)\n summaries.append(summary.reset_index())\n\n filename = f\"{args.dataset}-{args.model}-partial-report.json\"\n pd.concat(summaries).reset_index(drop=True).to_json(os.path.join(\"partial\", filename))\n\n\nif __name__ == \"__main__\":\n parser = argparse.ArgumentParser()\n parser.add_argument(\"--dataset\", type=str, default=\"iris\")\n parser.add_argument(\"--model\", type=str, default=\"kNN\")\n parser.add_argument(\"--budget\", type=int, default=80)\n parser.add_argument(\"--n-runs\", type=int, default=10)\n parser.add_argument(\"--sampler-list\", type=str, default=\"TPESampler CmaEsSampler\")\n parser.add_argument(\n \"--sampler-kwargs-list\",\n type=str,\n default='{\"multivariate\":true,\"constant_liar\":true} {}',\n )\n parser.add_argument(\"--pruner-list\", type=str, default=\"NopPruner\")\n parser.add_argument(\"--pruner-kwargs-list\", type=str, default=\"{}\")\n parser.add_argument(\"--plot-warmup\", type=str, default=\"true\")\n\n args = parser.parse_args()\n os.makedirs(\"runs\", exist_ok=True)\n os.makedirs(\"plots\", exist_ok=True)\n os.makedirs(\"partial\", exist_ok=True)\n\n run_benchmark(args)\n partial_report(args)\n make_plots(args)\n", "path": "benchmarks/run_bayesmark.py"}]}
| 2,863 | 439 |
gh_patches_debug_57667
|
rasdani/github-patches
|
git_diff
|
evennia__evennia-3042
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
[BUG] "evennia xyzgrid help" causes TypeError: NoneType takes no arguments
#### Describe the bug
Fresh migration from git master to main and then installing xyzgrid prevents evennia xyzgrid commands from working. For example, "evennia xyzgrid help" causes TypeError: NoneType takes no arguments
#### To Reproduce
1. Migrated from git master branch to main branch for 1.x release of Evennia.
2. Installed [extra](use to be in requirements_extra).
At this point, I can run the server and log in.
3. Added the xyzgrid command set and restarted.
'path', 'goto', 'map' are seen in the command list. The Limbo room does not have a map. Everything appears to work fine.
4. Modify the server/conf/settings.py.
xyzgrid is now available.
When I use xyzgrid, such as 'evennia xyzgrid help', or any other xyzgrid command:
TypeError: NoneType takes no arguments
#### Expected behavior
'evennia xyzgrid <command>' should call the xyzgrid command.
#### Environment, Evennia version, OS etc
Evennia 1.0.1 (rev 38011cc48d)
OS: nt
Python: 3.11.1
Twisted: 22.10.0
Django: 4.1.4
#### Additional context
This is based off helix4's message in #general on discord. I added my current steps that seem to reproduce the same issue down below. Here is the original message from helix4, with steps for reproducing on the older version of the code.
I am trying to test XYZGrid on a brand new install but failing. 1. cloned the single branch of evennia-develop, and initiated an evennia game. 2. installed requirements_extra, and migrated. I can run the server and log in. 3. i added the command set and reloadead, i see path, goto, map ingame. the Limbo room does not have a map. seems to work well. 4. modify the server/conf/settings.py, xyzgrid is now available.
When I use xyzgrid, such as evennia xyzgrid help, or any other xyzgrid command:
from evennia.utils.eveditor import EvEditor
File "/home/ubuntu/3ProjectMUD/evennia/evennia/utils/eveditor.py", line 201, in <module>
class SaveYesNoCmdSet(CmdSet):
TypeError: NoneType takes no arguments
Original message
https://discord.com/channels/246323978879107073/246323978879107073/937578545704730624
Griatch's response
https://discord.com/channels/246323978879107073/246323978879107073/937610453184561183
Steps:
1. Migrated from git master branch to main branch for 1.x release of Evennia.
2. Installed [extra](use to be in requirements_extra).
At this point, I can run the server and log in.
3. Added the xyzgrid command set and restarted.
'path', 'goto', 'map' are seen in the command list. The Limbo room does not have a map. Everything appears to work fine.
4. Modify the server/conf/settings.py.
xyzgrid is now available.
When I use xyzgrid, such as 'evennia xyzgrid help', or any other xyzgrid command:
Traceback (most recent call last):
File "C:\muddev\evenv\Scripts\evennia_launcher.py", line 18, in <module>
main()
File "C:\muddev\evennia\evennia\server\evennia_launcher.py", line 2422, in main
if run_custom_commands(option, *unknown_args):
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\muddev\evennia\evennia\server\evennia_launcher.py", line 2023, in run_custom_commands
mod = importlib.import_module(modpath)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
...
File "<frozen importlib._bootstrap>", line 1206, in _gcd_import
File "<frozen importlib._bootstrap>", line 1178, in _find_and_load
File "<frozen importlib._bootstrap>", line 1128, in _find_and_load_unlocked
File "<frozen importlib._bootstrap>", line 241, in _call_with_frames_removed
File "<frozen importlib._bootstrap>", line 1206, in _gcd_import
File "<frozen importlib._bootstrap>", line 1178, in _find_and_load
File "<frozen importlib._bootstrap>", line 1149, in _find_and_load_unlocked
File "<frozen importlib._bootstrap>", line 690, in _load_unlocked
File "<frozen importlib._bootstrap_external>", line 940, in exec_module
File "<frozen importlib._bootstrap>", line 241, in _call_with_frames_removed
File "C:\muddev\evennia\evennia\contrib\grid\xyzgrid\__init__.py", line 6, in <module>
from . import commands # noqa
^^^^^^^^^^^^^^^^^^^^^^
File "C:\muddev\evennia\evennia\contrib\grid\xyzgrid\commands.py", line 15, in <module>
from evennia.commands.default import building
File "C:\muddev\evennia\evennia\commands\default\building.py", line 14, in <module>
from evennia.prototypes import menus as olc_menus
File "C:\muddev\evennia\evennia\prototypes\menus.py", line 20, in <module>
from evennia.utils.evmenu import EvMenu, list_node
File "C:\muddev\evennia\evennia\utils\evmenu.py", line 350, in <module>
class CmdEvMenuNode(Command):
TypeError: NoneType takes no arguments
</issue>
<code>
[start of evennia/contrib/grid/xyzgrid/__init__.py]
1 """
2 XYZGrid - Griatch 2021
3
4 """
5
6 from . import commands # noqa
7 from . import example # noqa
8 from . import launchcmd # noqa
9 from . import prototypes # noqa
10 from . import tests # noqa
11 from . import utils # noqa
12 from . import xymap # noqa
13 from . import xymap_legend # noqa
14 from . import xyzgrid # noqa
15 from . import xyzroom # noqa
16
[end of evennia/contrib/grid/xyzgrid/__init__.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/evennia/contrib/grid/xyzgrid/__init__.py b/evennia/contrib/grid/xyzgrid/__init__.py
--- a/evennia/contrib/grid/xyzgrid/__init__.py
+++ b/evennia/contrib/grid/xyzgrid/__init__.py
@@ -2,14 +2,15 @@
XYZGrid - Griatch 2021
"""
-
-from . import commands # noqa
-from . import example # noqa
-from . import launchcmd # noqa
-from . import prototypes # noqa
-from . import tests # noqa
-from . import utils # noqa
-from . import xymap # noqa
-from . import xymap_legend # noqa
-from . import xyzgrid # noqa
-from . import xyzroom # noqa
+from . import (
+ example,
+ launchcmd,
+ prototypes,
+ tests,
+ utils,
+ xymap,
+ xymap_legend,
+ xyzgrid,
+ xyzroom,
+ commands,
+)
|
{"golden_diff": "diff --git a/evennia/contrib/grid/xyzgrid/__init__.py b/evennia/contrib/grid/xyzgrid/__init__.py\n--- a/evennia/contrib/grid/xyzgrid/__init__.py\n+++ b/evennia/contrib/grid/xyzgrid/__init__.py\n@@ -2,14 +2,15 @@\n XYZGrid - Griatch 2021\n \n \"\"\"\n-\n-from . import commands # noqa\n-from . import example # noqa\n-from . import launchcmd # noqa\n-from . import prototypes # noqa\n-from . import tests # noqa\n-from . import utils # noqa\n-from . import xymap # noqa\n-from . import xymap_legend # noqa\n-from . import xyzgrid # noqa\n-from . import xyzroom # noqa\n+from . import (\n+ example,\n+ launchcmd,\n+ prototypes,\n+ tests,\n+ utils,\n+ xymap,\n+ xymap_legend,\n+ xyzgrid,\n+ xyzroom,\n+ commands,\n+)\n", "issue": "[BUG] \"evennia xyzgrid help\" causes TypeError: NoneType takes no arguments\n#### Describe the bug\r\nFresh migration from git master to main and then installing xyzgrid prevents evennia xyzgrid commands from working. For example, \"evennia xyzgrid help\" causes TypeError: NoneType takes no arguments\r\n\r\n#### To Reproduce\r\n1. Migrated from git master branch to main branch for 1.x release of Evennia.\r\n2. Installed [extra](use to be in requirements_extra). \r\n\r\nAt this point, I can run the server and log in.\r\n\r\n3. Added the xyzgrid command set and restarted. \r\n\r\n'path', 'goto', 'map' are seen in the command list. The Limbo room does not have a map. Everything appears to work fine.\r\n\r\n4. Modify the server/conf/settings.py.\r\n\r\nxyzgrid is now available.\r\n\r\nWhen I use xyzgrid, such as 'evennia xyzgrid help', or any other xyzgrid command:\r\nTypeError: NoneType takes no arguments\r\n\r\n#### Expected behavior\r\n'evennia xyzgrid <command>' should call the xyzgrid command.\r\n\r\n#### Environment, Evennia version, OS etc\r\n\r\n Evennia 1.0.1 (rev 38011cc48d)\r\n OS: nt\r\n Python: 3.11.1\r\n Twisted: 22.10.0\r\n Django: 4.1.4\r\n\r\n#### Additional context\r\n\r\nThis is based off helix4's message in #general on discord. I added my current steps that seem to reproduce the same issue down below. Here is the original message from helix4, with steps for reproducing on the older version of the code.\r\n\r\nI am trying to test XYZGrid on a brand new install but failing. 1. cloned the single branch of evennia-develop, and initiated an evennia game. 2. installed requirements_extra, and migrated. I can run the server and log in. 3. i added the command set and reloadead, i see path, goto, map ingame. the Limbo room does not have a map. seems to work well. 4. modify the server/conf/settings.py, xyzgrid is now available.\r\n\r\nWhen I use xyzgrid, such as evennia xyzgrid help, or any other xyzgrid command:\r\n from evennia.utils.eveditor import EvEditor\r\n File \"/home/ubuntu/3ProjectMUD/evennia/evennia/utils/eveditor.py\", line 201, in <module>\r\n class SaveYesNoCmdSet(CmdSet):\r\nTypeError: NoneType takes no arguments\r\n\r\nOriginal message\r\n\r\nhttps://discord.com/channels/246323978879107073/246323978879107073/937578545704730624\r\n\r\nGriatch's response\r\n\r\nhttps://discord.com/channels/246323978879107073/246323978879107073/937610453184561183\r\n\r\nSteps:\r\n\r\n1. Migrated from git master branch to main branch for 1.x release of Evennia.\r\n2. Installed [extra](use to be in requirements_extra). \r\n\r\nAt this point, I can run the server and log in.\r\n\r\n3. Added the xyzgrid command set and restarted. \r\n\r\n'path', 'goto', 'map' are seen in the command list. The Limbo room does not have a map. Everything appears to work fine.\r\n\r\n4. Modify the server/conf/settings.py.\r\n\r\nxyzgrid is now available.\r\n\r\nWhen I use xyzgrid, such as 'evennia xyzgrid help', or any other xyzgrid command:\r\n\r\nTraceback (most recent call last):\r\n File \"C:\\muddev\\evenv\\Scripts\\evennia_launcher.py\", line 18, in <module>\r\n main()\r\n File \"C:\\muddev\\evennia\\evennia\\server\\evennia_launcher.py\", line 2422, in main\r\n if run_custom_commands(option, *unknown_args):\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File \"C:\\muddev\\evennia\\evennia\\server\\evennia_launcher.py\", line 2023, in run_custom_commands\r\n mod = importlib.import_module(modpath)\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n...\r\n\r\n File \"<frozen importlib._bootstrap>\", line 1206, in _gcd_import\r\n File \"<frozen importlib._bootstrap>\", line 1178, in _find_and_load\r\n File \"<frozen importlib._bootstrap>\", line 1128, in _find_and_load_unlocked\r\n File \"<frozen importlib._bootstrap>\", line 241, in _call_with_frames_removed\r\n File \"<frozen importlib._bootstrap>\", line 1206, in _gcd_import\r\n File \"<frozen importlib._bootstrap>\", line 1178, in _find_and_load\r\n File \"<frozen importlib._bootstrap>\", line 1149, in _find_and_load_unlocked\r\n File \"<frozen importlib._bootstrap>\", line 690, in _load_unlocked\r\n File \"<frozen importlib._bootstrap_external>\", line 940, in exec_module\r\n File \"<frozen importlib._bootstrap>\", line 241, in _call_with_frames_removed\r\n File \"C:\\muddev\\evennia\\evennia\\contrib\\grid\\xyzgrid\\__init__.py\", line 6, in <module>\r\n from . import commands # noqa\r\n ^^^^^^^^^^^^^^^^^^^^^^\r\n File \"C:\\muddev\\evennia\\evennia\\contrib\\grid\\xyzgrid\\commands.py\", line 15, in <module>\r\n from evennia.commands.default import building\r\n File \"C:\\muddev\\evennia\\evennia\\commands\\default\\building.py\", line 14, in <module>\r\n from evennia.prototypes import menus as olc_menus\r\n File \"C:\\muddev\\evennia\\evennia\\prototypes\\menus.py\", line 20, in <module>\r\n from evennia.utils.evmenu import EvMenu, list_node\r\n File \"C:\\muddev\\evennia\\evennia\\utils\\evmenu.py\", line 350, in <module>\r\n class CmdEvMenuNode(Command):\r\nTypeError: NoneType takes no arguments\r\n\r\n\n", "before_files": [{"content": "\"\"\"\nXYZGrid - Griatch 2021\n\n\"\"\"\n\nfrom . import commands # noqa\nfrom . import example # noqa\nfrom . import launchcmd # noqa\nfrom . import prototypes # noqa\nfrom . import tests # noqa\nfrom . import utils # noqa\nfrom . import xymap # noqa\nfrom . import xymap_legend # noqa\nfrom . import xyzgrid # noqa\nfrom . import xyzroom # noqa\n", "path": "evennia/contrib/grid/xyzgrid/__init__.py"}]}
| 2,100 | 232 |
gh_patches_debug_36164
|
rasdani/github-patches
|
git_diff
|
PaddlePaddle__models-1472
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
ocr_recognition 的参数说明与代码不相符
说明中的 --train_images 与 --train_list 两个参数,在代码中并不存在;运行时,带上这两个参数会报错。
</issue>
<code>
[start of fluid/PaddleCV/ocr_recognition/train.py]
1 """Trainer for OCR CTC or attention model."""
2 from __future__ import absolute_import
3 from __future__ import division
4 from __future__ import print_function
5 import paddle.fluid as fluid
6 from utility import add_arguments, print_arguments, to_lodtensor, get_ctc_feeder_data, get_attention_feeder_data
7 import paddle.fluid.profiler as profiler
8 from crnn_ctc_model import ctc_train_net
9 from attention_model import attention_train_net
10 import data_reader
11 import argparse
12 import functools
13 import sys
14 import time
15 import os
16 import numpy as np
17
18 parser = argparse.ArgumentParser(description=__doc__)
19 add_arg = functools.partial(add_arguments, argparser=parser)
20 # yapf: disable
21 add_arg('batch_size', int, 32, "Minibatch size.")
22 add_arg('total_step', int, 720000, "The number of iterations. Zero or less means whole training set. More than 0 means the training set might be looped until # of iterations is reached.")
23 add_arg('log_period', int, 1000, "Log period.")
24 add_arg('save_model_period', int, 15000, "Save model period. '-1' means never saving the model.")
25 add_arg('eval_period', int, 15000, "Evaluate period. '-1' means never evaluating the model.")
26 add_arg('save_model_dir', str, "./models", "The directory the model to be saved to.")
27 add_arg('model', str, "crnn_ctc", "Which type of network to be used. 'crnn_ctc' or 'attention'")
28 add_arg('init_model', str, None, "The init model file of directory.")
29 add_arg('use_gpu', bool, True, "Whether use GPU to train.")
30 add_arg('min_average_window',int, 10000, "Min average window.")
31 add_arg('max_average_window',int, 12500, "Max average window. It is proposed to be set as the number of minibatch in a pass.")
32 add_arg('average_window', float, 0.15, "Average window.")
33 add_arg('parallel', bool, False, "Whether use parallel training.")
34 add_arg('profile', bool, False, "Whether to use profiling.")
35 add_arg('skip_batch_num', int, 0, "The number of first minibatches to skip as warm-up for better performance test.")
36 add_arg('skip_test', bool, False, "Whether to skip test phase.")
37 # yapf: enable
38
39
40 def train(args):
41 """OCR training"""
42
43 if args.model == "crnn_ctc":
44 train_net = ctc_train_net
45 get_feeder_data = get_ctc_feeder_data
46 else:
47 train_net = attention_train_net
48 get_feeder_data = get_attention_feeder_data
49
50 num_classes = None
51 train_images = None
52 train_list = None
53 test_images = None
54 test_list = None
55 num_classes = data_reader.num_classes(
56 ) if num_classes is None else num_classes
57 data_shape = data_reader.data_shape()
58 # define network
59 sum_cost, error_evaluator, inference_program, model_average = train_net(
60 args, data_shape, num_classes)
61
62 # data reader
63 train_reader = data_reader.train(
64 args.batch_size,
65 train_images_dir=train_images,
66 train_list_file=train_list,
67 cycle=args.total_step > 0,
68 model=args.model)
69 test_reader = data_reader.test(
70 test_images_dir=test_images, test_list_file=test_list, model=args.model)
71
72 # prepare environment
73 place = fluid.CPUPlace()
74 if args.use_gpu:
75 place = fluid.CUDAPlace(0)
76 exe = fluid.Executor(place)
77
78 if 'ce_mode' in os.environ:
79 fluid.default_startup_program().random_seed = 90
80
81 exe.run(fluid.default_startup_program())
82
83 # load init model
84 if args.init_model is not None:
85 model_dir = args.init_model
86 model_file_name = None
87 if not os.path.isdir(args.init_model):
88 model_dir = os.path.dirname(args.init_model)
89 model_file_name = os.path.basename(args.init_model)
90 fluid.io.load_params(exe, dirname=model_dir, filename=model_file_name)
91 print("Init model from: %s." % args.init_model)
92
93 train_exe = exe
94 error_evaluator.reset(exe)
95 if args.parallel:
96 train_exe = fluid.ParallelExecutor(
97 use_cuda=True if args.use_gpu else False, loss_name=sum_cost.name)
98
99 fetch_vars = [sum_cost] + error_evaluator.metrics
100
101 def train_one_batch(data):
102 var_names = [var.name for var in fetch_vars]
103 if args.parallel:
104 results = train_exe.run(var_names,
105 feed=get_feeder_data(data, place))
106 results = [np.array(result).sum() for result in results]
107 else:
108 results = train_exe.run(feed=get_feeder_data(data, place),
109 fetch_list=fetch_vars)
110 results = [result[0] for result in results]
111 return results
112
113 def test(iter_num):
114 error_evaluator.reset(exe)
115 for data in test_reader():
116 exe.run(inference_program, feed=get_feeder_data(data, place))
117 _, test_seq_error = error_evaluator.eval(exe)
118 print("\nTime: %s; Iter[%d]; Test seq error: %s.\n" % (
119 time.time(), iter_num, str(test_seq_error[0])))
120
121 #Note: The following logs are special for CE monitoring.
122 #Other situations do not need to care about these logs.
123 print("kpis test_acc %f" % (1 - test_seq_error[0]))
124
125 def save_model(args, exe, iter_num):
126 filename = "model_%05d" % iter_num
127 fluid.io.save_params(
128 exe, dirname=args.save_model_dir, filename=filename)
129 print("Saved model to: %s/%s." % (args.save_model_dir, filename))
130
131 iter_num = 0
132 stop = False
133 start_time = time.time()
134 while not stop:
135 total_loss = 0.0
136 total_seq_error = 0.0
137 batch_times = []
138 # train a pass
139 for data in train_reader():
140 if args.total_step > 0 and iter_num == args.total_step + args.skip_batch_num:
141 stop = True
142 break
143 if iter_num < args.skip_batch_num:
144 print("Warm-up iteration")
145 if iter_num == args.skip_batch_num:
146 profiler.reset_profiler()
147 start = time.time()
148 results = train_one_batch(data)
149 batch_time = time.time() - start
150 fps = args.batch_size / batch_time
151 batch_times.append(batch_time)
152 total_loss += results[0]
153 total_seq_error += results[2]
154
155 iter_num += 1
156 # training log
157 if iter_num % args.log_period == 0:
158 print("\nTime: %s; Iter[%d]; Avg loss: %.3f; Avg seq err: %.3f" % (
159 time.time(), iter_num,
160 total_loss / (args.log_period * args.batch_size),
161 total_seq_error / (args.log_period * args.batch_size)))
162 print("kpis train_cost %f" % (total_loss / (args.log_period *
163 args.batch_size)))
164 print("kpis train_acc %f" % (
165 1 - total_seq_error / (args.log_period * args.batch_size)))
166 total_loss = 0.0
167 total_seq_error = 0.0
168
169 # evaluate
170 if not args.skip_test and iter_num % args.eval_period == 0:
171 if model_average:
172 with model_average.apply(exe):
173 test(iter_num)
174 else:
175 test(iter_num)
176
177 # save model
178 if iter_num % args.save_model_period == 0:
179 if model_average:
180 with model_average.apply(exe):
181 save_model(args, exe, iter_num)
182 else:
183 save_model(args, exe, iter_num)
184 end_time = time.time()
185 print("kpis train_duration %f" % (end_time - start_time))
186 # Postprocess benchmark data
187 latencies = batch_times[args.skip_batch_num:]
188 latency_avg = np.average(latencies)
189 latency_pc99 = np.percentile(latencies, 99)
190 fpses = np.divide(args.batch_size, latencies)
191 fps_avg = np.average(fpses)
192 fps_pc99 = np.percentile(fpses, 1)
193
194 # Benchmark output
195 print('\nTotal examples (incl. warm-up): %d' %
196 (iter_num * args.batch_size))
197 print('average latency: %.5f s, 99pc latency: %.5f s' % (latency_avg,
198 latency_pc99))
199 print('average fps: %.5f, fps for 99pc latency: %.5f' % (fps_avg,
200 fps_pc99))
201
202
203 def main():
204 args = parser.parse_args()
205 print_arguments(args)
206 if args.profile:
207 if args.use_gpu:
208 with profiler.cuda_profiler("cuda_profiler.txt", 'csv') as nvprof:
209 train(args)
210 else:
211 with profiler.profiler("CPU", sorted_key='total') as cpuprof:
212 train(args)
213 else:
214 train(args)
215
216
217 if __name__ == "__main__":
218 main()
219
[end of fluid/PaddleCV/ocr_recognition/train.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/fluid/PaddleCV/ocr_recognition/train.py b/fluid/PaddleCV/ocr_recognition/train.py
--- a/fluid/PaddleCV/ocr_recognition/train.py
+++ b/fluid/PaddleCV/ocr_recognition/train.py
@@ -24,6 +24,10 @@
add_arg('save_model_period', int, 15000, "Save model period. '-1' means never saving the model.")
add_arg('eval_period', int, 15000, "Evaluate period. '-1' means never evaluating the model.")
add_arg('save_model_dir', str, "./models", "The directory the model to be saved to.")
+add_arg('train_images', str, None, "The directory of images to be used for training.")
+add_arg('train_list', str, None, "The list file of images to be used for training.")
+add_arg('test_images', str, None, "The directory of images to be used for test.")
+add_arg('test_list', str, None, "The list file of images to be used for training.")
add_arg('model', str, "crnn_ctc", "Which type of network to be used. 'crnn_ctc' or 'attention'")
add_arg('init_model', str, None, "The init model file of directory.")
add_arg('use_gpu', bool, True, "Whether use GPU to train.")
@@ -48,10 +52,6 @@
get_feeder_data = get_attention_feeder_data
num_classes = None
- train_images = None
- train_list = None
- test_images = None
- test_list = None
num_classes = data_reader.num_classes(
) if num_classes is None else num_classes
data_shape = data_reader.data_shape()
@@ -62,12 +62,12 @@
# data reader
train_reader = data_reader.train(
args.batch_size,
- train_images_dir=train_images,
- train_list_file=train_list,
+ train_images_dir=args.train_images,
+ train_list_file=args.train_list,
cycle=args.total_step > 0,
model=args.model)
test_reader = data_reader.test(
- test_images_dir=test_images, test_list_file=test_list, model=args.model)
+ test_images_dir=args.test_images, test_list_file=args.test_list, model=args.model)
# prepare environment
place = fluid.CPUPlace()
|
{"golden_diff": "diff --git a/fluid/PaddleCV/ocr_recognition/train.py b/fluid/PaddleCV/ocr_recognition/train.py\n--- a/fluid/PaddleCV/ocr_recognition/train.py\n+++ b/fluid/PaddleCV/ocr_recognition/train.py\n@@ -24,6 +24,10 @@\n add_arg('save_model_period', int, 15000, \"Save model period. '-1' means never saving the model.\")\n add_arg('eval_period', int, 15000, \"Evaluate period. '-1' means never evaluating the model.\")\n add_arg('save_model_dir', str, \"./models\", \"The directory the model to be saved to.\")\n+add_arg('train_images', str, None, \"The directory of images to be used for training.\")\n+add_arg('train_list', str, None, \"The list file of images to be used for training.\")\n+add_arg('test_images', str, None, \"The directory of images to be used for test.\")\n+add_arg('test_list', str, None, \"The list file of images to be used for training.\")\n add_arg('model', str, \"crnn_ctc\", \"Which type of network to be used. 'crnn_ctc' or 'attention'\")\n add_arg('init_model', str, None, \"The init model file of directory.\")\n add_arg('use_gpu', bool, True, \"Whether use GPU to train.\")\n@@ -48,10 +52,6 @@\n get_feeder_data = get_attention_feeder_data\n \n num_classes = None\n- train_images = None\n- train_list = None\n- test_images = None\n- test_list = None\n num_classes = data_reader.num_classes(\n ) if num_classes is None else num_classes\n data_shape = data_reader.data_shape()\n@@ -62,12 +62,12 @@\n # data reader\n train_reader = data_reader.train(\n args.batch_size,\n- train_images_dir=train_images,\n- train_list_file=train_list,\n+ train_images_dir=args.train_images,\n+ train_list_file=args.train_list,\n cycle=args.total_step > 0,\n model=args.model)\n test_reader = data_reader.test(\n- test_images_dir=test_images, test_list_file=test_list, model=args.model)\n+ test_images_dir=args.test_images, test_list_file=args.test_list, model=args.model)\n \n # prepare environment\n place = fluid.CPUPlace()\n", "issue": "ocr_recognition \u7684\u53c2\u6570\u8bf4\u660e\u4e0e\u4ee3\u7801\u4e0d\u76f8\u7b26\n\u8bf4\u660e\u4e2d\u7684 --train_images \u4e0e --train_list \u4e24\u4e2a\u53c2\u6570\uff0c\u5728\u4ee3\u7801\u4e2d\u5e76\u4e0d\u5b58\u5728\uff1b\u8fd0\u884c\u65f6\uff0c\u5e26\u4e0a\u8fd9\u4e24\u4e2a\u53c2\u6570\u4f1a\u62a5\u9519\u3002\n", "before_files": [{"content": "\"\"\"Trainer for OCR CTC or attention model.\"\"\"\nfrom __future__ import absolute_import\nfrom __future__ import division\nfrom __future__ import print_function\nimport paddle.fluid as fluid\nfrom utility import add_arguments, print_arguments, to_lodtensor, get_ctc_feeder_data, get_attention_feeder_data\nimport paddle.fluid.profiler as profiler\nfrom crnn_ctc_model import ctc_train_net\nfrom attention_model import attention_train_net\nimport data_reader\nimport argparse\nimport functools\nimport sys\nimport time\nimport os\nimport numpy as np\n\nparser = argparse.ArgumentParser(description=__doc__)\nadd_arg = functools.partial(add_arguments, argparser=parser)\n# yapf: disable\nadd_arg('batch_size', int, 32, \"Minibatch size.\")\nadd_arg('total_step', int, 720000, \"The number of iterations. Zero or less means whole training set. More than 0 means the training set might be looped until # of iterations is reached.\")\nadd_arg('log_period', int, 1000, \"Log period.\")\nadd_arg('save_model_period', int, 15000, \"Save model period. '-1' means never saving the model.\")\nadd_arg('eval_period', int, 15000, \"Evaluate period. '-1' means never evaluating the model.\")\nadd_arg('save_model_dir', str, \"./models\", \"The directory the model to be saved to.\")\nadd_arg('model', str, \"crnn_ctc\", \"Which type of network to be used. 'crnn_ctc' or 'attention'\")\nadd_arg('init_model', str, None, \"The init model file of directory.\")\nadd_arg('use_gpu', bool, True, \"Whether use GPU to train.\")\nadd_arg('min_average_window',int, 10000, \"Min average window.\")\nadd_arg('max_average_window',int, 12500, \"Max average window. It is proposed to be set as the number of minibatch in a pass.\")\nadd_arg('average_window', float, 0.15, \"Average window.\")\nadd_arg('parallel', bool, False, \"Whether use parallel training.\")\nadd_arg('profile', bool, False, \"Whether to use profiling.\")\nadd_arg('skip_batch_num', int, 0, \"The number of first minibatches to skip as warm-up for better performance test.\")\nadd_arg('skip_test', bool, False, \"Whether to skip test phase.\")\n# yapf: enable\n\n\ndef train(args):\n \"\"\"OCR training\"\"\"\n\n if args.model == \"crnn_ctc\":\n train_net = ctc_train_net\n get_feeder_data = get_ctc_feeder_data\n else:\n train_net = attention_train_net\n get_feeder_data = get_attention_feeder_data\n\n num_classes = None\n train_images = None\n train_list = None\n test_images = None\n test_list = None\n num_classes = data_reader.num_classes(\n ) if num_classes is None else num_classes\n data_shape = data_reader.data_shape()\n # define network\n sum_cost, error_evaluator, inference_program, model_average = train_net(\n args, data_shape, num_classes)\n\n # data reader\n train_reader = data_reader.train(\n args.batch_size,\n train_images_dir=train_images,\n train_list_file=train_list,\n cycle=args.total_step > 0,\n model=args.model)\n test_reader = data_reader.test(\n test_images_dir=test_images, test_list_file=test_list, model=args.model)\n\n # prepare environment\n place = fluid.CPUPlace()\n if args.use_gpu:\n place = fluid.CUDAPlace(0)\n exe = fluid.Executor(place)\n\n if 'ce_mode' in os.environ:\n fluid.default_startup_program().random_seed = 90\n\n exe.run(fluid.default_startup_program())\n\n # load init model\n if args.init_model is not None:\n model_dir = args.init_model\n model_file_name = None\n if not os.path.isdir(args.init_model):\n model_dir = os.path.dirname(args.init_model)\n model_file_name = os.path.basename(args.init_model)\n fluid.io.load_params(exe, dirname=model_dir, filename=model_file_name)\n print(\"Init model from: %s.\" % args.init_model)\n\n train_exe = exe\n error_evaluator.reset(exe)\n if args.parallel:\n train_exe = fluid.ParallelExecutor(\n use_cuda=True if args.use_gpu else False, loss_name=sum_cost.name)\n\n fetch_vars = [sum_cost] + error_evaluator.metrics\n\n def train_one_batch(data):\n var_names = [var.name for var in fetch_vars]\n if args.parallel:\n results = train_exe.run(var_names,\n feed=get_feeder_data(data, place))\n results = [np.array(result).sum() for result in results]\n else:\n results = train_exe.run(feed=get_feeder_data(data, place),\n fetch_list=fetch_vars)\n results = [result[0] for result in results]\n return results\n\n def test(iter_num):\n error_evaluator.reset(exe)\n for data in test_reader():\n exe.run(inference_program, feed=get_feeder_data(data, place))\n _, test_seq_error = error_evaluator.eval(exe)\n print(\"\\nTime: %s; Iter[%d]; Test seq error: %s.\\n\" % (\n time.time(), iter_num, str(test_seq_error[0])))\n\n #Note: The following logs are special for CE monitoring.\n #Other situations do not need to care about these logs.\n print(\"kpis\ttest_acc\t%f\" % (1 - test_seq_error[0]))\n\n def save_model(args, exe, iter_num):\n filename = \"model_%05d\" % iter_num\n fluid.io.save_params(\n exe, dirname=args.save_model_dir, filename=filename)\n print(\"Saved model to: %s/%s.\" % (args.save_model_dir, filename))\n\n iter_num = 0\n stop = False\n start_time = time.time()\n while not stop:\n total_loss = 0.0\n total_seq_error = 0.0\n batch_times = []\n # train a pass\n for data in train_reader():\n if args.total_step > 0 and iter_num == args.total_step + args.skip_batch_num:\n stop = True\n break\n if iter_num < args.skip_batch_num:\n print(\"Warm-up iteration\")\n if iter_num == args.skip_batch_num:\n profiler.reset_profiler()\n start = time.time()\n results = train_one_batch(data)\n batch_time = time.time() - start\n fps = args.batch_size / batch_time\n batch_times.append(batch_time)\n total_loss += results[0]\n total_seq_error += results[2]\n\n iter_num += 1\n # training log\n if iter_num % args.log_period == 0:\n print(\"\\nTime: %s; Iter[%d]; Avg loss: %.3f; Avg seq err: %.3f\" % (\n time.time(), iter_num,\n total_loss / (args.log_period * args.batch_size),\n total_seq_error / (args.log_period * args.batch_size)))\n print(\"kpis\ttrain_cost\t%f\" % (total_loss / (args.log_period *\n args.batch_size)))\n print(\"kpis\ttrain_acc\t%f\" % (\n 1 - total_seq_error / (args.log_period * args.batch_size)))\n total_loss = 0.0\n total_seq_error = 0.0\n\n # evaluate\n if not args.skip_test and iter_num % args.eval_period == 0:\n if model_average:\n with model_average.apply(exe):\n test(iter_num)\n else:\n test(iter_num)\n\n # save model\n if iter_num % args.save_model_period == 0:\n if model_average:\n with model_average.apply(exe):\n save_model(args, exe, iter_num)\n else:\n save_model(args, exe, iter_num)\n end_time = time.time()\n print(\"kpis\ttrain_duration\t%f\" % (end_time - start_time))\n # Postprocess benchmark data\n latencies = batch_times[args.skip_batch_num:]\n latency_avg = np.average(latencies)\n latency_pc99 = np.percentile(latencies, 99)\n fpses = np.divide(args.batch_size, latencies)\n fps_avg = np.average(fpses)\n fps_pc99 = np.percentile(fpses, 1)\n\n # Benchmark output\n print('\\nTotal examples (incl. warm-up): %d' %\n (iter_num * args.batch_size))\n print('average latency: %.5f s, 99pc latency: %.5f s' % (latency_avg,\n latency_pc99))\n print('average fps: %.5f, fps for 99pc latency: %.5f' % (fps_avg,\n fps_pc99))\n\n\ndef main():\n args = parser.parse_args()\n print_arguments(args)\n if args.profile:\n if args.use_gpu:\n with profiler.cuda_profiler(\"cuda_profiler.txt\", 'csv') as nvprof:\n train(args)\n else:\n with profiler.profiler(\"CPU\", sorted_key='total') as cpuprof:\n train(args)\n else:\n train(args)\n\n\nif __name__ == \"__main__\":\n main()\n", "path": "fluid/PaddleCV/ocr_recognition/train.py"}]}
| 3,230 | 562 |
gh_patches_debug_12467
|
rasdani/github-patches
|
git_diff
|
deis__deis-571
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Make EC2 AMIs for 0.5.1 release
</issue>
<code>
[start of provider/ec2.py]
1 """
2 Deis cloud provider implementation for Amazon EC2.
3 """
4
5 from __future__ import unicode_literals
6
7 import json
8 import time
9
10 from boto import ec2
11 from boto.exception import EC2ResponseError
12
13 # from api.ssh import connect_ssh, exec_ssh
14 from deis import settings
15
16
17 # Deis-optimized EC2 amis -- with 3.8 kernel, chef 11 deps,
18 # and large docker images (e.g. buildstep) pre-installed
19 IMAGE_MAP = {
20 'ap-northeast-1': 'ami-5d432d5c',
21 'ap-southeast-1': 'ami-b4c493e6',
22 'ap-southeast-2': 'ami-d59d03ef',
23 'eu-west-1': 'ami-ce30c5b9',
24 'sa-east-1': 'ami-61b1117c',
25 'us-east-1': 'ami-8df9c9e4',
26 'us-west-1': 'ami-62477527',
27 'us-west-2': 'ami-ea6001da',
28 }
29
30
31 def seed_flavors():
32 """Seed the database with default flavors for each EC2 region.
33
34 :rtype: list of dicts containing flavor data
35 """
36 flavors = []
37 for r in ('us-east-1', 'us-west-1', 'us-west-2', 'eu-west-1',
38 'ap-northeast-1', 'ap-southeast-1', 'ap-southeast-2',
39 'sa-east-1'):
40 flavors.append({'id': 'ec2-{}'.format(r),
41 'provider': 'ec2',
42 'params': json.dumps({
43 'region': r,
44 'image': IMAGE_MAP[r],
45 'zone': 'any',
46 'size': 'm1.medium'})})
47 return flavors
48
49
50 def build_layer(layer):
51 """
52 Build a layer.
53
54 :param layer: a dict containing formation, id, params, and creds info
55 """
56 region = layer['params'].get('region', 'us-east-1')
57 conn = _create_ec2_connection(layer['creds'], region)
58 # create a new sg and authorize all ports
59 # use iptables on the host to firewall ports
60 name = "{formation}-{id}".format(**layer)
61 sg = conn.create_security_group(name, 'Created by Deis')
62 # import a new keypair using the layer key material
63 conn.import_key_pair(name, layer['ssh_public_key'])
64 # loop until the sg is *actually* there
65 for i in xrange(10):
66 try:
67 sg.authorize(ip_protocol='tcp', from_port=1, to_port=65535,
68 cidr_ip='0.0.0.0/0')
69 break
70 except EC2ResponseError:
71 if i < 10:
72 time.sleep(1.5)
73 continue
74 else:
75 raise RuntimeError('Failed to authorize security group')
76
77
78 def destroy_layer(layer):
79 """
80 Destroy a layer.
81
82 :param layer: a dict containing formation, id, params, and creds info
83 """
84 region = layer['params'].get('region', 'us-east-1')
85 name = "{formation}-{id}".format(**layer)
86 conn = _create_ec2_connection(layer['creds'], region)
87 conn.delete_key_pair(name)
88 # there's an ec2 race condition on instances terminating
89 # successfully but still holding a lock on the security group
90 for i in range(5):
91 # let's take a nap
92 time.sleep(i ** 1.25) # 1, 2.4, 3.9, 5.6, 7.4
93 try:
94 conn.delete_security_group(name)
95 return
96 except EC2ResponseError as err:
97 if err.code == 'InvalidGroup.NotFound':
98 return
99 elif err.code in ('InvalidGroup.InUse',
100 'DependencyViolation') and i < 4:
101 continue # retry
102 else:
103 raise
104
105
106 def build_node(node):
107 """
108 Build a node.
109
110 :param node: a dict containing formation, layer, params, and creds info.
111 :rtype: a tuple of (provider_id, fully_qualified_domain_name, metadata)
112 """
113 params, creds = node['params'], node['creds']
114 region = params.setdefault('region', 'us-east-1')
115 conn = _create_ec2_connection(creds, region)
116 name = "{formation}-{layer}".format(**node)
117 params['key_name'] = name
118 sg = conn.get_all_security_groups(name)[0]
119 params.setdefault('security_groups', []).append(sg.name)
120 image_id = params.get(
121 'image', getattr(settings, 'IMAGE_MAP', IMAGE_MAP)[region])
122 images = conn.get_all_images([image_id])
123 if len(images) != 1:
124 raise LookupError('Could not find AMI: %s' % image_id)
125 image = images[0]
126 kwargs = _prepare_run_kwargs(params)
127 reservation = image.run(**kwargs)
128 instances = reservation.instances
129 boto = instances[0]
130 # sleep before tagging
131 time.sleep(10)
132 boto.update()
133 boto.add_tag('Name', node['id'])
134 # loop until running
135 while(True):
136 time.sleep(2)
137 boto.update()
138 if boto.state == 'running':
139 break
140 # prepare return values
141 provider_id = boto.id
142 fqdn = boto.public_dns_name
143 metadata = _format_metadata(boto)
144 return provider_id, fqdn, metadata
145
146
147 def destroy_node(node):
148 """
149 Destroy a node.
150
151 :param node: a dict containing a node's provider_id, params, and creds
152 """
153 provider_id = node['provider_id']
154 region = node['params'].get('region', 'us-east-1')
155 conn = _create_ec2_connection(node['creds'], region)
156 if provider_id:
157 try:
158 conn.terminate_instances([provider_id])
159 i = conn.get_all_instances([provider_id])[0].instances[0]
160 while(True):
161 time.sleep(2)
162 i.update()
163 if i.state == "terminated":
164 break
165 except EC2ResponseError as e:
166 if e.code not in ('InvalidInstanceID.NotFound',):
167 raise
168
169
170 def _create_ec2_connection(creds, region):
171 """
172 Connect to an EC2 region with the given credentials.
173
174 :param creds: a dict containing an EC2 access_key and secret_key
175 :region: the name of an EC2 region, such as "us-west-2"
176 :rtype: a connected :class:`~boto.ec2.connection.EC2Connection`
177 :raises EnvironmentError: if no credentials are provided
178 """
179 if not creds:
180 raise EnvironmentError('No credentials provided')
181 return ec2.connect_to_region(region,
182 aws_access_key_id=creds['access_key'],
183 aws_secret_access_key=creds['secret_key'])
184
185
186 def _prepare_run_kwargs(params):
187 # start with sane defaults
188 kwargs = {
189 'min_count': 1, 'max_count': 1,
190 'user_data': None, 'addressing_type': None,
191 'instance_type': None, 'placement': None,
192 'kernel_id': None, 'ramdisk_id': None,
193 'monitoring_enabled': False, 'subnet_id': None,
194 'block_device_map': None,
195 }
196 # convert zone "any" to NoneType
197 requested_zone = params.get('zone')
198 if requested_zone and requested_zone.lower() == 'any':
199 requested_zone = None
200 # lookup kwargs from params
201 param_kwargs = {
202 'instance_type': params.get('size', 'm1.medium'),
203 'security_groups': params['security_groups'],
204 'placement': requested_zone,
205 'key_name': params['key_name'],
206 'kernel_id': params.get('kernel', None),
207 }
208 # add user_data if provided in params
209 user_data = params.get('user_data')
210 if user_data:
211 kwargs.update({'user_data': user_data})
212 # params override defaults
213 kwargs.update(param_kwargs)
214 return kwargs
215
216
217 def _format_metadata(boto):
218 return {
219 'architecture': boto.architecture,
220 'block_device_mapping': {
221 k: v.volume_id for k, v in boto.block_device_mapping.items()
222 },
223 'client_token': boto.client_token,
224 'dns_name': boto.dns_name,
225 'ebs_optimized': boto.ebs_optimized,
226 'eventsSet': boto.eventsSet,
227 'group_name': boto.group_name,
228 'groups': [g.id for g in boto.groups],
229 'hypervisor': boto.hypervisor,
230 'id': boto.id,
231 'image_id': boto.image_id,
232 'instance_profile': boto.instance_profile,
233 'instance_type': boto.instance_type,
234 'interfaces': list(boto.interfaces),
235 'ip_address': boto.ip_address,
236 'kernel': boto.kernel,
237 'key_name': boto.key_name,
238 'launch_time': boto.launch_time,
239 'monitored': boto.monitored,
240 'monitoring_state': boto.monitoring_state,
241 'persistent': boto.persistent,
242 'placement': boto.placement,
243 'placement_group': boto.placement_group,
244 'placement_tenancy': boto.placement_tenancy,
245 'previous_state': boto.previous_state,
246 'private_dns_name': boto.private_dns_name,
247 'private_ip_address': boto.private_ip_address,
248 'public_dns_name': boto.public_dns_name,
249 'ramdisk': boto.ramdisk,
250 'region': boto.region.name,
251 'root_device_name': boto.root_device_name,
252 'root_device_type': boto.root_device_type,
253 'spot_instance_request_id': boto.spot_instance_request_id,
254 'state': boto.state,
255 'state_code': boto.state_code,
256 'state_reason': boto.state_reason,
257 'subnet_id': boto.subnet_id,
258 'tags': dict(boto.tags),
259 'virtualization_type': boto.virtualization_type,
260 'vpc_id': boto.vpc_id,
261 }
262
[end of provider/ec2.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/provider/ec2.py b/provider/ec2.py
--- a/provider/ec2.py
+++ b/provider/ec2.py
@@ -17,14 +17,14 @@
# Deis-optimized EC2 amis -- with 3.8 kernel, chef 11 deps,
# and large docker images (e.g. buildstep) pre-installed
IMAGE_MAP = {
- 'ap-northeast-1': 'ami-5d432d5c',
- 'ap-southeast-1': 'ami-b4c493e6',
- 'ap-southeast-2': 'ami-d59d03ef',
- 'eu-west-1': 'ami-ce30c5b9',
- 'sa-east-1': 'ami-61b1117c',
- 'us-east-1': 'ami-8df9c9e4',
- 'us-west-1': 'ami-62477527',
- 'us-west-2': 'ami-ea6001da',
+ 'ap-northeast-1': 'ami-f37503f2',
+ 'ap-southeast-1': 'ami-7867362a',
+ 'ap-southeast-2': 'ami-71bb224b',
+ 'eu-west-1': 'ami-af31c3d8',
+ 'sa-east-1': 'ami-2b64c736',
+ 'us-east-1': 'ami-c1080aa8',
+ 'us-west-1': 'ami-1c8bb759',
+ 'us-west-2': 'ami-f282eec2',
}
|
{"golden_diff": "diff --git a/provider/ec2.py b/provider/ec2.py\n--- a/provider/ec2.py\n+++ b/provider/ec2.py\n@@ -17,14 +17,14 @@\n # Deis-optimized EC2 amis -- with 3.8 kernel, chef 11 deps,\n # and large docker images (e.g. buildstep) pre-installed\n IMAGE_MAP = {\n- 'ap-northeast-1': 'ami-5d432d5c',\n- 'ap-southeast-1': 'ami-b4c493e6',\n- 'ap-southeast-2': 'ami-d59d03ef',\n- 'eu-west-1': 'ami-ce30c5b9',\n- 'sa-east-1': 'ami-61b1117c',\n- 'us-east-1': 'ami-8df9c9e4',\n- 'us-west-1': 'ami-62477527',\n- 'us-west-2': 'ami-ea6001da',\n+ 'ap-northeast-1': 'ami-f37503f2',\n+ 'ap-southeast-1': 'ami-7867362a',\n+ 'ap-southeast-2': 'ami-71bb224b',\n+ 'eu-west-1': 'ami-af31c3d8',\n+ 'sa-east-1': 'ami-2b64c736',\n+ 'us-east-1': 'ami-c1080aa8',\n+ 'us-west-1': 'ami-1c8bb759',\n+ 'us-west-2': 'ami-f282eec2',\n }\n", "issue": "Make EC2 AMIs for 0.5.1 release\n\n", "before_files": [{"content": "\"\"\"\nDeis cloud provider implementation for Amazon EC2.\n\"\"\"\n\nfrom __future__ import unicode_literals\n\nimport json\nimport time\n\nfrom boto import ec2\nfrom boto.exception import EC2ResponseError\n\n# from api.ssh import connect_ssh, exec_ssh\nfrom deis import settings\n\n\n# Deis-optimized EC2 amis -- with 3.8 kernel, chef 11 deps,\n# and large docker images (e.g. buildstep) pre-installed\nIMAGE_MAP = {\n 'ap-northeast-1': 'ami-5d432d5c',\n 'ap-southeast-1': 'ami-b4c493e6',\n 'ap-southeast-2': 'ami-d59d03ef',\n 'eu-west-1': 'ami-ce30c5b9',\n 'sa-east-1': 'ami-61b1117c',\n 'us-east-1': 'ami-8df9c9e4',\n 'us-west-1': 'ami-62477527',\n 'us-west-2': 'ami-ea6001da',\n}\n\n\ndef seed_flavors():\n \"\"\"Seed the database with default flavors for each EC2 region.\n\n :rtype: list of dicts containing flavor data\n \"\"\"\n flavors = []\n for r in ('us-east-1', 'us-west-1', 'us-west-2', 'eu-west-1',\n 'ap-northeast-1', 'ap-southeast-1', 'ap-southeast-2',\n 'sa-east-1'):\n flavors.append({'id': 'ec2-{}'.format(r),\n 'provider': 'ec2',\n 'params': json.dumps({\n 'region': r,\n 'image': IMAGE_MAP[r],\n 'zone': 'any',\n 'size': 'm1.medium'})})\n return flavors\n\n\ndef build_layer(layer):\n \"\"\"\n Build a layer.\n\n :param layer: a dict containing formation, id, params, and creds info\n \"\"\"\n region = layer['params'].get('region', 'us-east-1')\n conn = _create_ec2_connection(layer['creds'], region)\n # create a new sg and authorize all ports\n # use iptables on the host to firewall ports\n name = \"{formation}-{id}\".format(**layer)\n sg = conn.create_security_group(name, 'Created by Deis')\n # import a new keypair using the layer key material\n conn.import_key_pair(name, layer['ssh_public_key'])\n # loop until the sg is *actually* there\n for i in xrange(10):\n try:\n sg.authorize(ip_protocol='tcp', from_port=1, to_port=65535,\n cidr_ip='0.0.0.0/0')\n break\n except EC2ResponseError:\n if i < 10:\n time.sleep(1.5)\n continue\n else:\n raise RuntimeError('Failed to authorize security group')\n\n\ndef destroy_layer(layer):\n \"\"\"\n Destroy a layer.\n\n :param layer: a dict containing formation, id, params, and creds info\n \"\"\"\n region = layer['params'].get('region', 'us-east-1')\n name = \"{formation}-{id}\".format(**layer)\n conn = _create_ec2_connection(layer['creds'], region)\n conn.delete_key_pair(name)\n # there's an ec2 race condition on instances terminating\n # successfully but still holding a lock on the security group\n for i in range(5):\n # let's take a nap\n time.sleep(i ** 1.25) # 1, 2.4, 3.9, 5.6, 7.4\n try:\n conn.delete_security_group(name)\n return\n except EC2ResponseError as err:\n if err.code == 'InvalidGroup.NotFound':\n return\n elif err.code in ('InvalidGroup.InUse',\n 'DependencyViolation') and i < 4:\n continue # retry\n else:\n raise\n\n\ndef build_node(node):\n \"\"\"\n Build a node.\n\n :param node: a dict containing formation, layer, params, and creds info.\n :rtype: a tuple of (provider_id, fully_qualified_domain_name, metadata)\n \"\"\"\n params, creds = node['params'], node['creds']\n region = params.setdefault('region', 'us-east-1')\n conn = _create_ec2_connection(creds, region)\n name = \"{formation}-{layer}\".format(**node)\n params['key_name'] = name\n sg = conn.get_all_security_groups(name)[0]\n params.setdefault('security_groups', []).append(sg.name)\n image_id = params.get(\n 'image', getattr(settings, 'IMAGE_MAP', IMAGE_MAP)[region])\n images = conn.get_all_images([image_id])\n if len(images) != 1:\n raise LookupError('Could not find AMI: %s' % image_id)\n image = images[0]\n kwargs = _prepare_run_kwargs(params)\n reservation = image.run(**kwargs)\n instances = reservation.instances\n boto = instances[0]\n # sleep before tagging\n time.sleep(10)\n boto.update()\n boto.add_tag('Name', node['id'])\n # loop until running\n while(True):\n time.sleep(2)\n boto.update()\n if boto.state == 'running':\n break\n # prepare return values\n provider_id = boto.id\n fqdn = boto.public_dns_name\n metadata = _format_metadata(boto)\n return provider_id, fqdn, metadata\n\n\ndef destroy_node(node):\n \"\"\"\n Destroy a node.\n\n :param node: a dict containing a node's provider_id, params, and creds\n \"\"\"\n provider_id = node['provider_id']\n region = node['params'].get('region', 'us-east-1')\n conn = _create_ec2_connection(node['creds'], region)\n if provider_id:\n try:\n conn.terminate_instances([provider_id])\n i = conn.get_all_instances([provider_id])[0].instances[0]\n while(True):\n time.sleep(2)\n i.update()\n if i.state == \"terminated\":\n break\n except EC2ResponseError as e:\n if e.code not in ('InvalidInstanceID.NotFound',):\n raise\n\n\ndef _create_ec2_connection(creds, region):\n \"\"\"\n Connect to an EC2 region with the given credentials.\n\n :param creds: a dict containing an EC2 access_key and secret_key\n :region: the name of an EC2 region, such as \"us-west-2\"\n :rtype: a connected :class:`~boto.ec2.connection.EC2Connection`\n :raises EnvironmentError: if no credentials are provided\n \"\"\"\n if not creds:\n raise EnvironmentError('No credentials provided')\n return ec2.connect_to_region(region,\n aws_access_key_id=creds['access_key'],\n aws_secret_access_key=creds['secret_key'])\n\n\ndef _prepare_run_kwargs(params):\n # start with sane defaults\n kwargs = {\n 'min_count': 1, 'max_count': 1,\n 'user_data': None, 'addressing_type': None,\n 'instance_type': None, 'placement': None,\n 'kernel_id': None, 'ramdisk_id': None,\n 'monitoring_enabled': False, 'subnet_id': None,\n 'block_device_map': None,\n }\n # convert zone \"any\" to NoneType\n requested_zone = params.get('zone')\n if requested_zone and requested_zone.lower() == 'any':\n requested_zone = None\n # lookup kwargs from params\n param_kwargs = {\n 'instance_type': params.get('size', 'm1.medium'),\n 'security_groups': params['security_groups'],\n 'placement': requested_zone,\n 'key_name': params['key_name'],\n 'kernel_id': params.get('kernel', None),\n }\n # add user_data if provided in params\n user_data = params.get('user_data')\n if user_data:\n kwargs.update({'user_data': user_data})\n # params override defaults\n kwargs.update(param_kwargs)\n return kwargs\n\n\ndef _format_metadata(boto):\n return {\n 'architecture': boto.architecture,\n 'block_device_mapping': {\n k: v.volume_id for k, v in boto.block_device_mapping.items()\n },\n 'client_token': boto.client_token,\n 'dns_name': boto.dns_name,\n 'ebs_optimized': boto.ebs_optimized,\n 'eventsSet': boto.eventsSet,\n 'group_name': boto.group_name,\n 'groups': [g.id for g in boto.groups],\n 'hypervisor': boto.hypervisor,\n 'id': boto.id,\n 'image_id': boto.image_id,\n 'instance_profile': boto.instance_profile,\n 'instance_type': boto.instance_type,\n 'interfaces': list(boto.interfaces),\n 'ip_address': boto.ip_address,\n 'kernel': boto.kernel,\n 'key_name': boto.key_name,\n 'launch_time': boto.launch_time,\n 'monitored': boto.monitored,\n 'monitoring_state': boto.monitoring_state,\n 'persistent': boto.persistent,\n 'placement': boto.placement,\n 'placement_group': boto.placement_group,\n 'placement_tenancy': boto.placement_tenancy,\n 'previous_state': boto.previous_state,\n 'private_dns_name': boto.private_dns_name,\n 'private_ip_address': boto.private_ip_address,\n 'public_dns_name': boto.public_dns_name,\n 'ramdisk': boto.ramdisk,\n 'region': boto.region.name,\n 'root_device_name': boto.root_device_name,\n 'root_device_type': boto.root_device_type,\n 'spot_instance_request_id': boto.spot_instance_request_id,\n 'state': boto.state,\n 'state_code': boto.state_code,\n 'state_reason': boto.state_reason,\n 'subnet_id': boto.subnet_id,\n 'tags': dict(boto.tags),\n 'virtualization_type': boto.virtualization_type,\n 'vpc_id': boto.vpc_id,\n }\n", "path": "provider/ec2.py"}]}
| 3,438 | 400 |
gh_patches_debug_29525
|
rasdani/github-patches
|
git_diff
|
netbox-community__netbox-13826
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Validate all headers/fields when bulk import objects
### NetBox version
v3.4.3
### Python version
3.8
### Steps to Reproduce
1. Open Prefix Bulk Import
2. Specify all required field options and add some dummy data
`prefix,status,testa,testb`
`10.10.10.0/24,active,a,b`
4. Press Import
5. Import is done without any complains
### Expected Behavior
Fields which are not stated in Field Options should throw an error, or a warning that some data was ignored.
### Observed Behavior
Import is done without any complains, Looks like all data was imported correct, although some fields were ignored
</issue>
<code>
[start of netbox/utilities/forms/forms.py]
1 import re
2
3 from django import forms
4 from django.utils.translation import gettext as _
5 from .mixins import BootstrapMixin
6
7 __all__ = (
8 'BulkEditForm',
9 'BulkRenameForm',
10 'ConfirmationForm',
11 'CSVModelForm',
12 'FilterForm',
13 'TableConfigForm',
14 )
15
16
17 class ConfirmationForm(BootstrapMixin, forms.Form):
18 """
19 A generic confirmation form. The form is not valid unless the `confirm` field is checked.
20 """
21 return_url = forms.CharField(
22 required=False,
23 widget=forms.HiddenInput()
24 )
25 confirm = forms.BooleanField(
26 required=True,
27 widget=forms.HiddenInput(),
28 initial=True
29 )
30
31
32 class BulkEditForm(BootstrapMixin, forms.Form):
33 """
34 Provides bulk edit support for objects.
35 """
36 nullable_fields = ()
37
38
39 class BulkRenameForm(BootstrapMixin, forms.Form):
40 """
41 An extendable form to be used for renaming objects in bulk.
42 """
43 find = forms.CharField(
44 strip=False
45 )
46 replace = forms.CharField(
47 strip=False,
48 required=False
49 )
50 use_regex = forms.BooleanField(
51 required=False,
52 initial=True,
53 label=_('Use regular expressions')
54 )
55
56 def clean(self):
57 super().clean()
58
59 # Validate regular expression in "find" field
60 if self.cleaned_data['use_regex']:
61 try:
62 re.compile(self.cleaned_data['find'])
63 except re.error:
64 raise forms.ValidationError({
65 'find': "Invalid regular expression"
66 })
67
68
69 class CSVModelForm(forms.ModelForm):
70 """
71 ModelForm used for the import of objects in CSV format.
72 """
73 def __init__(self, *args, headers=None, fields=None, **kwargs):
74 headers = headers or {}
75 fields = fields or []
76 super().__init__(*args, **kwargs)
77
78 # Modify the model form to accommodate any customized to_field_name properties
79 for field, to_field in headers.items():
80 if to_field is not None:
81 self.fields[field].to_field_name = to_field
82
83 # Omit any fields not specified (e.g. because the form is being used to
84 # updated rather than create objects)
85 if fields:
86 for field in list(self.fields.keys()):
87 if field not in fields:
88 del self.fields[field]
89
90
91 class FilterForm(BootstrapMixin, forms.Form):
92 """
93 Base Form class for FilterSet forms.
94 """
95 q = forms.CharField(
96 required=False,
97 label=_('Search')
98 )
99
100
101 class TableConfigForm(BootstrapMixin, forms.Form):
102 """
103 Form for configuring user's table preferences.
104 """
105 available_columns = forms.MultipleChoiceField(
106 choices=[],
107 required=False,
108 widget=forms.SelectMultiple(
109 attrs={'size': 10, 'class': 'form-select'}
110 ),
111 label=_('Available Columns')
112 )
113 columns = forms.MultipleChoiceField(
114 choices=[],
115 required=False,
116 widget=forms.SelectMultiple(
117 attrs={'size': 10, 'class': 'form-select'}
118 ),
119 label=_('Selected Columns')
120 )
121
122 def __init__(self, table, *args, **kwargs):
123 self.table = table
124
125 super().__init__(*args, **kwargs)
126
127 # Initialize columns field based on table attributes
128 self.fields['available_columns'].choices = table.available_columns
129 self.fields['columns'].choices = table.selected_columns
130
131 @property
132 def table_name(self):
133 return self.table.__class__.__name__
134
[end of netbox/utilities/forms/forms.py]
[start of netbox/utilities/forms/bulk_import.py]
1 import csv
2 import json
3 from io import StringIO
4
5 import yaml
6 from django import forms
7 from django.utils.translation import gettext as _
8
9 from core.forms.mixins import SyncedDataMixin
10 from utilities.choices import CSVDelimiterChoices, ImportFormatChoices, ImportMethodChoices
11 from utilities.constants import CSV_DELIMITERS
12 from utilities.forms.utils import parse_csv
13 from .mixins import BootstrapMixin
14
15
16 class BulkImportForm(BootstrapMixin, SyncedDataMixin, forms.Form):
17 import_method = forms.ChoiceField(
18 choices=ImportMethodChoices,
19 required=False
20 )
21 data = forms.CharField(
22 required=False,
23 widget=forms.Textarea(attrs={'class': 'font-monospace'}),
24 help_text=_("Enter object data in CSV, JSON or YAML format.")
25 )
26 upload_file = forms.FileField(
27 label=_("Data file"),
28 required=False
29 )
30 format = forms.ChoiceField(
31 choices=ImportFormatChoices,
32 initial=ImportFormatChoices.AUTO
33 )
34 csv_delimiter = forms.ChoiceField(
35 choices=CSVDelimiterChoices,
36 initial=CSVDelimiterChoices.AUTO,
37 label=_("CSV delimiter"),
38 help_text=_("The character which delimits CSV fields. Applies only to CSV format."),
39 required=False
40 )
41
42 data_field = 'data'
43
44 def clean(self):
45 super().clean()
46
47 # Determine import method
48 import_method = self.cleaned_data.get('import_method') or ImportMethodChoices.DIRECT
49
50 # Determine whether we're reading from form data or an uploaded file
51 if self.cleaned_data['data'] and import_method != ImportMethodChoices.DIRECT:
52 raise forms.ValidationError("Form data must be empty when uploading/selecting a file.")
53 if import_method == ImportMethodChoices.UPLOAD:
54 self.upload_file = 'upload_file'
55 file = self.files.get('upload_file')
56 data = file.read().decode('utf-8-sig')
57 elif import_method == ImportMethodChoices.DATA_FILE:
58 data = self.cleaned_data['data_file'].data_as_string
59 else:
60 data = self.cleaned_data['data']
61
62 # Determine the data format
63 if self.cleaned_data['format'] == ImportFormatChoices.AUTO:
64 if self.cleaned_data['csv_delimiter'] != CSVDelimiterChoices.AUTO:
65 # Specifying the CSV delimiter implies CSV format
66 format = ImportFormatChoices.CSV
67 else:
68 format = self._detect_format(data)
69 else:
70 format = self.cleaned_data['format']
71
72 # Process data according to the selected format
73 if format == ImportFormatChoices.CSV:
74 delimiter = self.cleaned_data.get('csv_delimiter', CSVDelimiterChoices.AUTO)
75 self.cleaned_data['data'] = self._clean_csv(data, delimiter=delimiter)
76 elif format == ImportFormatChoices.JSON:
77 self.cleaned_data['data'] = self._clean_json(data)
78 elif format == ImportFormatChoices.YAML:
79 self.cleaned_data['data'] = self._clean_yaml(data)
80 else:
81 raise forms.ValidationError(f"Unknown data format: {format}")
82
83 def _detect_format(self, data):
84 """
85 Attempt to automatically detect the format (CSV, JSON, or YAML) of the given data, or raise
86 a ValidationError.
87 """
88 try:
89 if data[0] in ('{', '['):
90 return ImportFormatChoices.JSON
91 if data.startswith('---') or data.startswith('- '):
92 return ImportFormatChoices.YAML
93 # Look for any of the CSV delimiters in the first line (ignoring the default 'auto' choice)
94 first_line = data.split('\n', 1)[0]
95 csv_delimiters = CSV_DELIMITERS.values()
96 if any(x in first_line for x in csv_delimiters):
97 return ImportFormatChoices.CSV
98 except IndexError:
99 pass
100 raise forms.ValidationError({
101 'format': _('Unable to detect data format. Please specify.')
102 })
103
104 def _clean_csv(self, data, delimiter=CSVDelimiterChoices.AUTO):
105 """
106 Clean CSV-formatted data. The first row will be treated as column headers.
107 """
108 # Determine the CSV dialect
109 if delimiter == CSVDelimiterChoices.AUTO:
110 # This uses a rough heuristic to detect the CSV dialect based on the presence of supported delimiting
111 # characters. If the data is malformed, we'll fall back to the default Excel dialect.
112 delimiters = ''.join(CSV_DELIMITERS.values())
113 try:
114 dialect = csv.Sniffer().sniff(data.strip(), delimiters=delimiters)
115 except csv.Error:
116 dialect = csv.excel
117 elif delimiter in (CSVDelimiterChoices.COMMA, CSVDelimiterChoices.SEMICOLON):
118 dialect = csv.excel
119 dialect.delimiter = delimiter
120 elif delimiter == CSVDelimiterChoices.TAB:
121 dialect = csv.excel_tab
122 else:
123 raise forms.ValidationError({
124 'csv_delimiter': _('Invalid CSV delimiter'),
125 })
126
127 stream = StringIO(data.strip())
128 reader = csv.reader(stream, dialect=dialect)
129 headers, records = parse_csv(reader)
130
131 # Set CSV headers for reference by the model form
132 self._csv_headers = headers
133
134 return records
135
136 def _clean_json(self, data):
137 """
138 Clean JSON-formatted data. If only a single object is defined, it will be encapsulated as a list.
139 """
140 try:
141 data = json.loads(data)
142 # Accommodate for users entering single objects
143 if type(data) is not list:
144 data = [data]
145 return data
146 except json.decoder.JSONDecodeError as err:
147 raise forms.ValidationError({
148 self.data_field: f"Invalid JSON data: {err}"
149 })
150
151 def _clean_yaml(self, data):
152 """
153 Clean YAML-formatted data. Data must be either
154 a) A single document comprising a list of dictionaries (each representing an object), or
155 b) Multiple documents, separated with the '---' token
156 """
157 records = []
158 try:
159 for data in yaml.load_all(data, Loader=yaml.SafeLoader):
160 if type(data) is list:
161 records.extend(data)
162 elif type(data) is dict:
163 records.append(data)
164 else:
165 raise forms.ValidationError({
166 self.data_field: _(
167 "Invalid YAML data. Data must be in the form of multiple documents, or a single document "
168 "comprising a list of dictionaries."
169 )
170 })
171 except yaml.error.YAMLError as err:
172 raise forms.ValidationError({
173 self.data_field: f"Invalid YAML data: {err}"
174 })
175
176 return records
177
[end of netbox/utilities/forms/bulk_import.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/netbox/utilities/forms/bulk_import.py b/netbox/utilities/forms/bulk_import.py
--- a/netbox/utilities/forms/bulk_import.py
+++ b/netbox/utilities/forms/bulk_import.py
@@ -129,6 +129,7 @@
headers, records = parse_csv(reader)
# Set CSV headers for reference by the model form
+ headers.pop('id', None)
self._csv_headers = headers
return records
diff --git a/netbox/utilities/forms/forms.py b/netbox/utilities/forms/forms.py
--- a/netbox/utilities/forms/forms.py
+++ b/netbox/utilities/forms/forms.py
@@ -70,22 +70,24 @@
"""
ModelForm used for the import of objects in CSV format.
"""
- def __init__(self, *args, headers=None, fields=None, **kwargs):
- headers = headers or {}
- fields = fields or []
+ def __init__(self, *args, headers=None, **kwargs):
+ self.headers = headers or {}
super().__init__(*args, **kwargs)
# Modify the model form to accommodate any customized to_field_name properties
- for field, to_field in headers.items():
+ for field, to_field in self.headers.items():
if to_field is not None:
self.fields[field].to_field_name = to_field
- # Omit any fields not specified (e.g. because the form is being used to
- # updated rather than create objects)
- if fields:
- for field in list(self.fields.keys()):
- if field not in fields:
- del self.fields[field]
+ def clean(self):
+ # Flag any invalid CSV headers
+ for header in self.headers:
+ if header not in self.fields:
+ raise forms.ValidationError(
+ _("Unrecognized header: {name}").format(name=header)
+ )
+
+ return super().clean()
class FilterForm(BootstrapMixin, forms.Form):
|
{"golden_diff": "diff --git a/netbox/utilities/forms/bulk_import.py b/netbox/utilities/forms/bulk_import.py\n--- a/netbox/utilities/forms/bulk_import.py\n+++ b/netbox/utilities/forms/bulk_import.py\n@@ -129,6 +129,7 @@\n headers, records = parse_csv(reader)\n \n # Set CSV headers for reference by the model form\n+ headers.pop('id', None)\n self._csv_headers = headers\n \n return records\ndiff --git a/netbox/utilities/forms/forms.py b/netbox/utilities/forms/forms.py\n--- a/netbox/utilities/forms/forms.py\n+++ b/netbox/utilities/forms/forms.py\n@@ -70,22 +70,24 @@\n \"\"\"\n ModelForm used for the import of objects in CSV format.\n \"\"\"\n- def __init__(self, *args, headers=None, fields=None, **kwargs):\n- headers = headers or {}\n- fields = fields or []\n+ def __init__(self, *args, headers=None, **kwargs):\n+ self.headers = headers or {}\n super().__init__(*args, **kwargs)\n \n # Modify the model form to accommodate any customized to_field_name properties\n- for field, to_field in headers.items():\n+ for field, to_field in self.headers.items():\n if to_field is not None:\n self.fields[field].to_field_name = to_field\n \n- # Omit any fields not specified (e.g. because the form is being used to\n- # updated rather than create objects)\n- if fields:\n- for field in list(self.fields.keys()):\n- if field not in fields:\n- del self.fields[field]\n+ def clean(self):\n+ # Flag any invalid CSV headers\n+ for header in self.headers:\n+ if header not in self.fields:\n+ raise forms.ValidationError(\n+ _(\"Unrecognized header: {name}\").format(name=header)\n+ )\n+\n+ return super().clean()\n \n \n class FilterForm(BootstrapMixin, forms.Form):\n", "issue": "Validate all headers/fields when bulk import objects\n### NetBox version\n\nv3.4.3\n\n### Python version\n\n3.8\n\n### Steps to Reproduce\n\n1. Open Prefix Bulk Import\r\n2. Specify all required field options and add some dummy data\r\n`prefix,status,testa,testb`\r\n`10.10.10.0/24,active,a,b`\r\n4. Press Import\r\n5. Import is done without any complains\r\n\n\n### Expected Behavior\n\nFields which are not stated in Field Options should throw an error, or a warning that some data was ignored.\r\n\n\n### Observed Behavior\n\nImport is done without any complains, Looks like all data was imported correct, although some fields were ignored\n", "before_files": [{"content": "import re\n\nfrom django import forms\nfrom django.utils.translation import gettext as _\nfrom .mixins import BootstrapMixin\n\n__all__ = (\n 'BulkEditForm',\n 'BulkRenameForm',\n 'ConfirmationForm',\n 'CSVModelForm',\n 'FilterForm',\n 'TableConfigForm',\n)\n\n\nclass ConfirmationForm(BootstrapMixin, forms.Form):\n \"\"\"\n A generic confirmation form. The form is not valid unless the `confirm` field is checked.\n \"\"\"\n return_url = forms.CharField(\n required=False,\n widget=forms.HiddenInput()\n )\n confirm = forms.BooleanField(\n required=True,\n widget=forms.HiddenInput(),\n initial=True\n )\n\n\nclass BulkEditForm(BootstrapMixin, forms.Form):\n \"\"\"\n Provides bulk edit support for objects.\n \"\"\"\n nullable_fields = ()\n\n\nclass BulkRenameForm(BootstrapMixin, forms.Form):\n \"\"\"\n An extendable form to be used for renaming objects in bulk.\n \"\"\"\n find = forms.CharField(\n strip=False\n )\n replace = forms.CharField(\n strip=False,\n required=False\n )\n use_regex = forms.BooleanField(\n required=False,\n initial=True,\n label=_('Use regular expressions')\n )\n\n def clean(self):\n super().clean()\n\n # Validate regular expression in \"find\" field\n if self.cleaned_data['use_regex']:\n try:\n re.compile(self.cleaned_data['find'])\n except re.error:\n raise forms.ValidationError({\n 'find': \"Invalid regular expression\"\n })\n\n\nclass CSVModelForm(forms.ModelForm):\n \"\"\"\n ModelForm used for the import of objects in CSV format.\n \"\"\"\n def __init__(self, *args, headers=None, fields=None, **kwargs):\n headers = headers or {}\n fields = fields or []\n super().__init__(*args, **kwargs)\n\n # Modify the model form to accommodate any customized to_field_name properties\n for field, to_field in headers.items():\n if to_field is not None:\n self.fields[field].to_field_name = to_field\n\n # Omit any fields not specified (e.g. because the form is being used to\n # updated rather than create objects)\n if fields:\n for field in list(self.fields.keys()):\n if field not in fields:\n del self.fields[field]\n\n\nclass FilterForm(BootstrapMixin, forms.Form):\n \"\"\"\n Base Form class for FilterSet forms.\n \"\"\"\n q = forms.CharField(\n required=False,\n label=_('Search')\n )\n\n\nclass TableConfigForm(BootstrapMixin, forms.Form):\n \"\"\"\n Form for configuring user's table preferences.\n \"\"\"\n available_columns = forms.MultipleChoiceField(\n choices=[],\n required=False,\n widget=forms.SelectMultiple(\n attrs={'size': 10, 'class': 'form-select'}\n ),\n label=_('Available Columns')\n )\n columns = forms.MultipleChoiceField(\n choices=[],\n required=False,\n widget=forms.SelectMultiple(\n attrs={'size': 10, 'class': 'form-select'}\n ),\n label=_('Selected Columns')\n )\n\n def __init__(self, table, *args, **kwargs):\n self.table = table\n\n super().__init__(*args, **kwargs)\n\n # Initialize columns field based on table attributes\n self.fields['available_columns'].choices = table.available_columns\n self.fields['columns'].choices = table.selected_columns\n\n @property\n def table_name(self):\n return self.table.__class__.__name__\n", "path": "netbox/utilities/forms/forms.py"}, {"content": "import csv\nimport json\nfrom io import StringIO\n\nimport yaml\nfrom django import forms\nfrom django.utils.translation import gettext as _\n\nfrom core.forms.mixins import SyncedDataMixin\nfrom utilities.choices import CSVDelimiterChoices, ImportFormatChoices, ImportMethodChoices\nfrom utilities.constants import CSV_DELIMITERS\nfrom utilities.forms.utils import parse_csv\nfrom .mixins import BootstrapMixin\n\n\nclass BulkImportForm(BootstrapMixin, SyncedDataMixin, forms.Form):\n import_method = forms.ChoiceField(\n choices=ImportMethodChoices,\n required=False\n )\n data = forms.CharField(\n required=False,\n widget=forms.Textarea(attrs={'class': 'font-monospace'}),\n help_text=_(\"Enter object data in CSV, JSON or YAML format.\")\n )\n upload_file = forms.FileField(\n label=_(\"Data file\"),\n required=False\n )\n format = forms.ChoiceField(\n choices=ImportFormatChoices,\n initial=ImportFormatChoices.AUTO\n )\n csv_delimiter = forms.ChoiceField(\n choices=CSVDelimiterChoices,\n initial=CSVDelimiterChoices.AUTO,\n label=_(\"CSV delimiter\"),\n help_text=_(\"The character which delimits CSV fields. Applies only to CSV format.\"),\n required=False\n )\n\n data_field = 'data'\n\n def clean(self):\n super().clean()\n\n # Determine import method\n import_method = self.cleaned_data.get('import_method') or ImportMethodChoices.DIRECT\n\n # Determine whether we're reading from form data or an uploaded file\n if self.cleaned_data['data'] and import_method != ImportMethodChoices.DIRECT:\n raise forms.ValidationError(\"Form data must be empty when uploading/selecting a file.\")\n if import_method == ImportMethodChoices.UPLOAD:\n self.upload_file = 'upload_file'\n file = self.files.get('upload_file')\n data = file.read().decode('utf-8-sig')\n elif import_method == ImportMethodChoices.DATA_FILE:\n data = self.cleaned_data['data_file'].data_as_string\n else:\n data = self.cleaned_data['data']\n\n # Determine the data format\n if self.cleaned_data['format'] == ImportFormatChoices.AUTO:\n if self.cleaned_data['csv_delimiter'] != CSVDelimiterChoices.AUTO:\n # Specifying the CSV delimiter implies CSV format\n format = ImportFormatChoices.CSV\n else:\n format = self._detect_format(data)\n else:\n format = self.cleaned_data['format']\n\n # Process data according to the selected format\n if format == ImportFormatChoices.CSV:\n delimiter = self.cleaned_data.get('csv_delimiter', CSVDelimiterChoices.AUTO)\n self.cleaned_data['data'] = self._clean_csv(data, delimiter=delimiter)\n elif format == ImportFormatChoices.JSON:\n self.cleaned_data['data'] = self._clean_json(data)\n elif format == ImportFormatChoices.YAML:\n self.cleaned_data['data'] = self._clean_yaml(data)\n else:\n raise forms.ValidationError(f\"Unknown data format: {format}\")\n\n def _detect_format(self, data):\n \"\"\"\n Attempt to automatically detect the format (CSV, JSON, or YAML) of the given data, or raise\n a ValidationError.\n \"\"\"\n try:\n if data[0] in ('{', '['):\n return ImportFormatChoices.JSON\n if data.startswith('---') or data.startswith('- '):\n return ImportFormatChoices.YAML\n # Look for any of the CSV delimiters in the first line (ignoring the default 'auto' choice)\n first_line = data.split('\\n', 1)[0]\n csv_delimiters = CSV_DELIMITERS.values()\n if any(x in first_line for x in csv_delimiters):\n return ImportFormatChoices.CSV\n except IndexError:\n pass\n raise forms.ValidationError({\n 'format': _('Unable to detect data format. Please specify.')\n })\n\n def _clean_csv(self, data, delimiter=CSVDelimiterChoices.AUTO):\n \"\"\"\n Clean CSV-formatted data. The first row will be treated as column headers.\n \"\"\"\n # Determine the CSV dialect\n if delimiter == CSVDelimiterChoices.AUTO:\n # This uses a rough heuristic to detect the CSV dialect based on the presence of supported delimiting\n # characters. If the data is malformed, we'll fall back to the default Excel dialect.\n delimiters = ''.join(CSV_DELIMITERS.values())\n try:\n dialect = csv.Sniffer().sniff(data.strip(), delimiters=delimiters)\n except csv.Error:\n dialect = csv.excel\n elif delimiter in (CSVDelimiterChoices.COMMA, CSVDelimiterChoices.SEMICOLON):\n dialect = csv.excel\n dialect.delimiter = delimiter\n elif delimiter == CSVDelimiterChoices.TAB:\n dialect = csv.excel_tab\n else:\n raise forms.ValidationError({\n 'csv_delimiter': _('Invalid CSV delimiter'),\n })\n\n stream = StringIO(data.strip())\n reader = csv.reader(stream, dialect=dialect)\n headers, records = parse_csv(reader)\n\n # Set CSV headers for reference by the model form\n self._csv_headers = headers\n\n return records\n\n def _clean_json(self, data):\n \"\"\"\n Clean JSON-formatted data. If only a single object is defined, it will be encapsulated as a list.\n \"\"\"\n try:\n data = json.loads(data)\n # Accommodate for users entering single objects\n if type(data) is not list:\n data = [data]\n return data\n except json.decoder.JSONDecodeError as err:\n raise forms.ValidationError({\n self.data_field: f\"Invalid JSON data: {err}\"\n })\n\n def _clean_yaml(self, data):\n \"\"\"\n Clean YAML-formatted data. Data must be either\n a) A single document comprising a list of dictionaries (each representing an object), or\n b) Multiple documents, separated with the '---' token\n \"\"\"\n records = []\n try:\n for data in yaml.load_all(data, Loader=yaml.SafeLoader):\n if type(data) is list:\n records.extend(data)\n elif type(data) is dict:\n records.append(data)\n else:\n raise forms.ValidationError({\n self.data_field: _(\n \"Invalid YAML data. Data must be in the form of multiple documents, or a single document \"\n \"comprising a list of dictionaries.\"\n )\n })\n except yaml.error.YAMLError as err:\n raise forms.ValidationError({\n self.data_field: f\"Invalid YAML data: {err}\"\n })\n\n return records\n", "path": "netbox/utilities/forms/bulk_import.py"}]}
| 3,590 | 440 |
gh_patches_debug_34336
|
rasdani/github-patches
|
git_diff
|
CTFd__CTFd-461
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Optimize top10 call
https://github.com/CTFd/CTFd/blob/master/CTFd/scoreboard.py#L125-L127
This code seems pretty wasteful and is likely getting hit fairly often. Optimizing this to be a single database query is likely a good idea.
</issue>
<code>
[start of CTFd/scoreboard.py]
1 from flask import render_template, jsonify, Blueprint, redirect, url_for, request
2 from sqlalchemy.sql.expression import union_all
3
4 from CTFd.models import db, Teams, Solves, Awards, Challenges
5
6 from CTFd import utils
7
8 scoreboard = Blueprint('scoreboard', __name__)
9
10
11 def get_standings(admin=False, count=None):
12 scores = db.session.query(
13 Solves.teamid.label('teamid'),
14 db.func.sum(Challenges.value).label('score'),
15 db.func.max(Solves.id).label('id'),
16 db.func.max(Solves.date).label('date')
17 ).join(Challenges).group_by(Solves.teamid)
18
19 awards = db.session.query(
20 Awards.teamid.label('teamid'),
21 db.func.sum(Awards.value).label('score'),
22 db.func.max(Awards.id).label('id'),
23 db.func.max(Awards.date).label('date')
24 ).group_by(Awards.teamid)
25
26 """
27 Filter out solves and awards that are before a specific time point.
28 """
29 freeze = utils.get_config('freeze')
30 if not admin and freeze:
31 scores = scores.filter(Solves.date < utils.unix_time_to_utc(freeze))
32 awards = awards.filter(Awards.date < utils.unix_time_to_utc(freeze))
33
34 """
35 Combine awards and solves with a union. They should have the same amount of columns
36 """
37 results = union_all(scores, awards).alias('results')
38
39 """
40 Sum each of the results by the team id to get their score.
41 """
42 sumscores = db.session.query(
43 results.columns.teamid,
44 db.func.sum(results.columns.score).label('score'),
45 db.func.max(results.columns.id).label('id'),
46 db.func.max(results.columns.date).label('date')
47 ).group_by(results.columns.teamid).subquery()
48
49 """
50 Admins can see scores for all users but the public cannot see banned users.
51
52 Filters out banned users.
53 Properly resolves value ties by ID.
54
55 Different databases treat time precision differently so resolve by the row ID instead.
56 """
57 if admin:
58 standings_query = db.session.query(
59 Teams.id.label('teamid'),
60 Teams.name.label('name'),
61 Teams.banned, sumscores.columns.score
62 )\
63 .join(sumscores, Teams.id == sumscores.columns.teamid) \
64 .order_by(sumscores.columns.score.desc(), sumscores.columns.id)
65 else:
66 standings_query = db.session.query(
67 Teams.id.label('teamid'),
68 Teams.name.label('name'),
69 sumscores.columns.score
70 )\
71 .join(sumscores, Teams.id == sumscores.columns.teamid) \
72 .filter(Teams.banned == False) \
73 .order_by(sumscores.columns.score.desc(), sumscores.columns.id)
74
75 """
76 Only select a certain amount of users if asked.
77 """
78 if count is None:
79 standings = standings_query.all()
80 else:
81 standings = standings_query.limit(count).all()
82 db.session.close()
83
84 return standings
85
86
87 @scoreboard.route('/scoreboard')
88 def scoreboard_view():
89 if utils.get_config('view_scoreboard_if_authed') and not utils.authed():
90 return redirect(url_for('auth.login', next=request.path))
91 if utils.hide_scores():
92 return render_template('scoreboard.html', errors=['Scores are currently hidden'])
93 standings = get_standings()
94 return render_template('scoreboard.html', teams=standings, score_frozen=utils.is_scoreboard_frozen())
95
96
97 @scoreboard.route('/scores')
98 def scores():
99 json = {'standings': []}
100 if utils.get_config('view_scoreboard_if_authed') and not utils.authed():
101 return redirect(url_for('auth.login', next=request.path))
102 if utils.hide_scores():
103 return jsonify(json)
104
105 standings = get_standings()
106
107 for i, x in enumerate(standings):
108 json['standings'].append({'pos': i + 1, 'id': x.teamid, 'team': x.name, 'score': int(x.score)})
109 return jsonify(json)
110
111
112 @scoreboard.route('/top/<int:count>')
113 def topteams(count):
114 json = {'places': {}}
115 if utils.get_config('view_scoreboard_if_authed') and not utils.authed():
116 return redirect(url_for('auth.login', next=request.path))
117 if utils.hide_scores():
118 return jsonify(json)
119
120 if count > 20 or count < 0:
121 count = 10
122
123 standings = get_standings(count=count)
124
125 for i, team in enumerate(standings):
126 solves = Solves.query.filter_by(teamid=team.teamid)
127 awards = Awards.query.filter_by(teamid=team.teamid)
128
129 freeze = utils.get_config('freeze')
130
131 if freeze:
132 solves = solves.filter(Solves.date < utils.unix_time_to_utc(freeze))
133 awards = awards.filter(Awards.date < utils.unix_time_to_utc(freeze))
134
135 solves = solves.all()
136 awards = awards.all()
137
138 json['places'][i + 1] = {
139 'id': team.teamid,
140 'name': team.name,
141 'solves': []
142 }
143 for x in solves:
144 json['places'][i + 1]['solves'].append({
145 'chal': x.chalid,
146 'team': x.teamid,
147 'value': x.chal.value,
148 'time': utils.unix_time(x.date)
149 })
150 for award in awards:
151 json['places'][i + 1]['solves'].append({
152 'chal': None,
153 'team': award.teamid,
154 'value': award.value,
155 'time': utils.unix_time(award.date)
156 })
157 json['places'][i + 1]['solves'] = sorted(json['places'][i + 1]['solves'], key=lambda k: k['time'])
158 return jsonify(json)
159
[end of CTFd/scoreboard.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/CTFd/scoreboard.py b/CTFd/scoreboard.py
--- a/CTFd/scoreboard.py
+++ b/CTFd/scoreboard.py
@@ -122,37 +122,42 @@
standings = get_standings(count=count)
- for i, team in enumerate(standings):
- solves = Solves.query.filter_by(teamid=team.teamid)
- awards = Awards.query.filter_by(teamid=team.teamid)
+ team_ids = [team.teamid for team in standings]
- freeze = utils.get_config('freeze')
+ solves = Solves.query.filter(Solves.teamid.in_(team_ids))
+ awards = Awards.query.filter(Awards.teamid.in_(team_ids))
- if freeze:
- solves = solves.filter(Solves.date < utils.unix_time_to_utc(freeze))
- awards = awards.filter(Awards.date < utils.unix_time_to_utc(freeze))
+ freeze = utils.get_config('freeze')
+
+ if freeze:
+ solves = solves.filter(Solves.date < utils.unix_time_to_utc(freeze))
+ awards = awards.filter(Awards.date < utils.unix_time_to_utc(freeze))
- solves = solves.all()
- awards = awards.all()
+ solves = solves.all()
+ awards = awards.all()
+ for i, team in enumerate(team_ids):
json['places'][i + 1] = {
- 'id': team.teamid,
- 'name': team.name,
+ 'id': standings[i].teamid,
+ 'name': standings[i].name,
'solves': []
}
- for x in solves:
- json['places'][i + 1]['solves'].append({
- 'chal': x.chalid,
- 'team': x.teamid,
- 'value': x.chal.value,
- 'time': utils.unix_time(x.date)
- })
+ for solve in solves:
+ if solve.teamid == team:
+ json['places'][i + 1]['solves'].append({
+ 'chal': solve.chalid,
+ 'team': solve.teamid,
+ 'value': solve.chal.value,
+ 'time': utils.unix_time(solve.date)
+ })
for award in awards:
- json['places'][i + 1]['solves'].append({
- 'chal': None,
- 'team': award.teamid,
- 'value': award.value,
- 'time': utils.unix_time(award.date)
- })
+ if award.teamid == team:
+ json['places'][i + 1]['solves'].append({
+ 'chal': None,
+ 'team': award.teamid,
+ 'value': award.value,
+ 'time': utils.unix_time(award.date)
+ })
json['places'][i + 1]['solves'] = sorted(json['places'][i + 1]['solves'], key=lambda k: k['time'])
+
return jsonify(json)
|
{"golden_diff": "diff --git a/CTFd/scoreboard.py b/CTFd/scoreboard.py\n--- a/CTFd/scoreboard.py\n+++ b/CTFd/scoreboard.py\n@@ -122,37 +122,42 @@\n \n standings = get_standings(count=count)\n \n- for i, team in enumerate(standings):\n- solves = Solves.query.filter_by(teamid=team.teamid)\n- awards = Awards.query.filter_by(teamid=team.teamid)\n+ team_ids = [team.teamid for team in standings]\n \n- freeze = utils.get_config('freeze')\n+ solves = Solves.query.filter(Solves.teamid.in_(team_ids))\n+ awards = Awards.query.filter(Awards.teamid.in_(team_ids))\n \n- if freeze:\n- solves = solves.filter(Solves.date < utils.unix_time_to_utc(freeze))\n- awards = awards.filter(Awards.date < utils.unix_time_to_utc(freeze))\n+ freeze = utils.get_config('freeze')\n+\n+ if freeze:\n+ solves = solves.filter(Solves.date < utils.unix_time_to_utc(freeze))\n+ awards = awards.filter(Awards.date < utils.unix_time_to_utc(freeze))\n \n- solves = solves.all()\n- awards = awards.all()\n+ solves = solves.all()\n+ awards = awards.all()\n \n+ for i, team in enumerate(team_ids):\n json['places'][i + 1] = {\n- 'id': team.teamid,\n- 'name': team.name,\n+ 'id': standings[i].teamid,\n+ 'name': standings[i].name,\n 'solves': []\n }\n- for x in solves:\n- json['places'][i + 1]['solves'].append({\n- 'chal': x.chalid,\n- 'team': x.teamid,\n- 'value': x.chal.value,\n- 'time': utils.unix_time(x.date)\n- })\n+ for solve in solves:\n+ if solve.teamid == team:\n+ json['places'][i + 1]['solves'].append({\n+ 'chal': solve.chalid,\n+ 'team': solve.teamid,\n+ 'value': solve.chal.value,\n+ 'time': utils.unix_time(solve.date)\n+ })\n for award in awards:\n- json['places'][i + 1]['solves'].append({\n- 'chal': None,\n- 'team': award.teamid,\n- 'value': award.value,\n- 'time': utils.unix_time(award.date)\n- })\n+ if award.teamid == team:\n+ json['places'][i + 1]['solves'].append({\n+ 'chal': None,\n+ 'team': award.teamid,\n+ 'value': award.value,\n+ 'time': utils.unix_time(award.date)\n+ })\n json['places'][i + 1]['solves'] = sorted(json['places'][i + 1]['solves'], key=lambda k: k['time'])\n+\n return jsonify(json)\n", "issue": "Optimize top10 call\nhttps://github.com/CTFd/CTFd/blob/master/CTFd/scoreboard.py#L125-L127\r\n\r\nThis code seems pretty wasteful and is likely getting hit fairly often. Optimizing this to be a single database query is likely a good idea. \n", "before_files": [{"content": "from flask import render_template, jsonify, Blueprint, redirect, url_for, request\nfrom sqlalchemy.sql.expression import union_all\n\nfrom CTFd.models import db, Teams, Solves, Awards, Challenges\n\nfrom CTFd import utils\n\nscoreboard = Blueprint('scoreboard', __name__)\n\n\ndef get_standings(admin=False, count=None):\n scores = db.session.query(\n Solves.teamid.label('teamid'),\n db.func.sum(Challenges.value).label('score'),\n db.func.max(Solves.id).label('id'),\n db.func.max(Solves.date).label('date')\n ).join(Challenges).group_by(Solves.teamid)\n\n awards = db.session.query(\n Awards.teamid.label('teamid'),\n db.func.sum(Awards.value).label('score'),\n db.func.max(Awards.id).label('id'),\n db.func.max(Awards.date).label('date')\n ).group_by(Awards.teamid)\n\n \"\"\"\n Filter out solves and awards that are before a specific time point.\n \"\"\"\n freeze = utils.get_config('freeze')\n if not admin and freeze:\n scores = scores.filter(Solves.date < utils.unix_time_to_utc(freeze))\n awards = awards.filter(Awards.date < utils.unix_time_to_utc(freeze))\n\n \"\"\"\n Combine awards and solves with a union. They should have the same amount of columns\n \"\"\"\n results = union_all(scores, awards).alias('results')\n\n \"\"\"\n Sum each of the results by the team id to get their score.\n \"\"\"\n sumscores = db.session.query(\n results.columns.teamid,\n db.func.sum(results.columns.score).label('score'),\n db.func.max(results.columns.id).label('id'),\n db.func.max(results.columns.date).label('date')\n ).group_by(results.columns.teamid).subquery()\n\n \"\"\"\n Admins can see scores for all users but the public cannot see banned users.\n\n Filters out banned users.\n Properly resolves value ties by ID.\n\n Different databases treat time precision differently so resolve by the row ID instead.\n \"\"\"\n if admin:\n standings_query = db.session.query(\n Teams.id.label('teamid'),\n Teams.name.label('name'),\n Teams.banned, sumscores.columns.score\n )\\\n .join(sumscores, Teams.id == sumscores.columns.teamid) \\\n .order_by(sumscores.columns.score.desc(), sumscores.columns.id)\n else:\n standings_query = db.session.query(\n Teams.id.label('teamid'),\n Teams.name.label('name'),\n sumscores.columns.score\n )\\\n .join(sumscores, Teams.id == sumscores.columns.teamid) \\\n .filter(Teams.banned == False) \\\n .order_by(sumscores.columns.score.desc(), sumscores.columns.id)\n\n \"\"\"\n Only select a certain amount of users if asked.\n \"\"\"\n if count is None:\n standings = standings_query.all()\n else:\n standings = standings_query.limit(count).all()\n db.session.close()\n\n return standings\n\n\[email protected]('/scoreboard')\ndef scoreboard_view():\n if utils.get_config('view_scoreboard_if_authed') and not utils.authed():\n return redirect(url_for('auth.login', next=request.path))\n if utils.hide_scores():\n return render_template('scoreboard.html', errors=['Scores are currently hidden'])\n standings = get_standings()\n return render_template('scoreboard.html', teams=standings, score_frozen=utils.is_scoreboard_frozen())\n\n\[email protected]('/scores')\ndef scores():\n json = {'standings': []}\n if utils.get_config('view_scoreboard_if_authed') and not utils.authed():\n return redirect(url_for('auth.login', next=request.path))\n if utils.hide_scores():\n return jsonify(json)\n\n standings = get_standings()\n\n for i, x in enumerate(standings):\n json['standings'].append({'pos': i + 1, 'id': x.teamid, 'team': x.name, 'score': int(x.score)})\n return jsonify(json)\n\n\[email protected]('/top/<int:count>')\ndef topteams(count):\n json = {'places': {}}\n if utils.get_config('view_scoreboard_if_authed') and not utils.authed():\n return redirect(url_for('auth.login', next=request.path))\n if utils.hide_scores():\n return jsonify(json)\n\n if count > 20 or count < 0:\n count = 10\n\n standings = get_standings(count=count)\n\n for i, team in enumerate(standings):\n solves = Solves.query.filter_by(teamid=team.teamid)\n awards = Awards.query.filter_by(teamid=team.teamid)\n\n freeze = utils.get_config('freeze')\n\n if freeze:\n solves = solves.filter(Solves.date < utils.unix_time_to_utc(freeze))\n awards = awards.filter(Awards.date < utils.unix_time_to_utc(freeze))\n\n solves = solves.all()\n awards = awards.all()\n\n json['places'][i + 1] = {\n 'id': team.teamid,\n 'name': team.name,\n 'solves': []\n }\n for x in solves:\n json['places'][i + 1]['solves'].append({\n 'chal': x.chalid,\n 'team': x.teamid,\n 'value': x.chal.value,\n 'time': utils.unix_time(x.date)\n })\n for award in awards:\n json['places'][i + 1]['solves'].append({\n 'chal': None,\n 'team': award.teamid,\n 'value': award.value,\n 'time': utils.unix_time(award.date)\n })\n json['places'][i + 1]['solves'] = sorted(json['places'][i + 1]['solves'], key=lambda k: k['time'])\n return jsonify(json)\n", "path": "CTFd/scoreboard.py"}]}
| 2,248 | 677 |
gh_patches_debug_25120
|
rasdani/github-patches
|
git_diff
|
bridgecrewio__checkov-936
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
ERROR:checkov.terraform.checks.resource.gcp.CloudStorageSelfLogging:Failed to run check: Bucket should not log to itself for configuration
**Describe the bug**
Error during check with rule CKV_GCP_63 (not clearly identify)
Here is the result of my terraform plan for my GCS resource using [terraform google storage module](https://github.com/terraform-google-modules/terraform-google-cloud-storage)
```
resource "google_storage_bucket" "buckets" {
bucket_policy_only = (known after apply)
force_destroy = false
id = (known after apply)
labels = {
"name" = "xxxxxxxx-bucket"
}
location = "EU"
name = "xxxxxxxx-bucket"
project = "xxxxxxxx-project"
self_link = (known after apply)
storage_class = "STANDARD"
uniform_bucket_level_access = true
url = (known after apply)
versioning {
enabled = true
}
}
```
```
ERROR:checkov.terraform.checks.resource.gcp.CloudStorageSelfLogging:Failed to run check: Bucket should not log to itself for configuration
[[]], 'default_event_based_hold': [None], 'encryption': [[]], 'force_destroy': [False], 'labels': [{'name': ['xxxxxxxx-bucket'], 'start_line': [158], 'end_line': [160]}], 'lifecycle_rule': [[]], 'location': ['EU'], 'logging': [[]], 'name': ['xxxxxxxx-bucket'], 'project': ['xxxxxxxx-project'], 'requester_pays': [None], 'retention_policy': [[]], 'storage_class': ['STANDARD'], 'uniform_bucket_level_access': [True], 'versioning': [{'enabled': [True], 'start_line': [171], 'end_line': [173]}], 'website': [[]], 'start_line': [153], 'end_line': [176]} at file: /checkov.tfplan.json
Traceback (most recent call last):
File "/usr/local/bin/checkov", line 5, in <module>
run()
File "/usr/local/lib/python3.8/site-packages/checkov/main.py", line 96, in run
scan_reports = runner_registry.run(external_checks_dir=external_checks_dir, files=args.file,
File "/usr/local/lib/python3.8/site-packages/checkov/common/runners/runner_registry.py", line 34, in run
scan_report = runner.run(root_folder, external_checks_dir=external_checks_dir, files=files,
File "/usr/local/lib/python3.8/site-packages/checkov/terraform/plan_runner.py", line 65, in run
self.check_tf_definition(report, runner_filter)
File "/usr/local/lib/python3.8/site-packages/checkov/terraform/plan_runner.py", line 79, in check_tf_definition
self.run_block(definition[block_type], full_file_path, report, scanned_file,
File "/usr/local/lib/python3.8/site-packages/checkov/terraform/plan_runner.py", line 95, in run_block
results = registry.scan(scanned_file, entity, [], runner_filter)
File "/usr/local/lib/python3.8/site-packages/checkov/common/checks/base_check_registry.py", line 109, in scan
result = self.run_check(check, entity_configuration, entity_name, entity_type, scanned_file, skip_info)
File "/usr/local/lib/python3.8/site-packages/checkov/common/checks/base_check_registry.py", line 115, in run_check
result = check.run(scanned_file=scanned_file, entity_configuration=entity_configuration,
File "/usr/local/lib/python3.8/site-packages/checkov/common/checks/base_check.py", line 62, in run
raise e
File "/usr/local/lib/python3.8/site-packages/checkov/common/checks/base_check.py", line 42, in run
check_result['result'] = self.scan_entity_conf(entity_configuration, entity_type)
File "/usr/local/lib/python3.8/site-packages/checkov/terraform/checks/resource/base_resource_check.py", line 17, in scan_entity_conf
return self.scan_resource_conf(conf, entity_type)
File "/usr/local/lib/python3.8/site-packages/checkov/terraform/checks/resource/base_resource_check.py", line 33, in wrapper
return wrapped(self, conf)
File "/usr/local/lib/python3.8/site-packages/checkov/terraform/checks/resource/gcp/CloudStorageSelfLogging.py", line 17, in scan_resource_conf
if conf['logging'][0]['log_bucket']:
TypeError: list indices must be integers or slices, not str
```
**To Reproduce**
Steps to reproduce the behavior:
1. Using [terraform GCS module](https://github.com/terraform-google-modules/terraform-google-cloud-storage) in your terraform
2. Run checkov scan
3. See error
**Informations**
- Checkov Version 1.0.799 (from docker image)
</issue>
<code>
[start of checkov/terraform/checks/resource/gcp/CloudStorageLogging.py]
1 from checkov.common.models.enums import CheckResult, CheckCategories
2 from checkov.terraform.checks.resource.base_resource_check import BaseResourceCheck
3
4
5 class CloudStorageLogging(BaseResourceCheck):
6 def __init__(self):
7 name = "Bucket should log access"
8 id = "CKV_GCP_62"
9 supported_resources = ['google_storage_bucket']
10 categories = [CheckCategories.LOGGING]
11 super().__init__(name=name, id=id, categories=categories, supported_resources=supported_resources)
12
13 def scan_resource_conf(self, conf):
14 #check fot logging
15 if 'logging' in conf.keys():
16 if conf['logging'][0]['log_bucket']:
17 log_bucket_name = conf['logging'][0]['log_bucket']
18 if log_bucket_name != None:
19 return CheckResult.PASSED
20 else:
21 return CheckResult.FAILED
22 else:
23 return CheckResult.FAILED
24 return CheckResult.FAILED
25 else:
26 return CheckResult.FAILED
27 return CheckResult.FAILED
28
29 check = CloudStorageLogging()
30
[end of checkov/terraform/checks/resource/gcp/CloudStorageLogging.py]
[start of checkov/terraform/checks/resource/gcp/CloudStorageSelfLogging.py]
1 from checkov.common.models.enums import CheckResult, CheckCategories
2 from checkov.terraform.checks.resource.base_resource_check import BaseResourceCheck
3
4
5 class CloudStorageSelfLogging(BaseResourceCheck):
6 def __init__(self):
7 name = "Bucket should not log to itself"
8 id = "CKV_GCP_63"
9 supported_resources = ['google_storage_bucket']
10 categories = [CheckCategories.LOGGING]
11 super().__init__(name=name, id=id, categories=categories, supported_resources=supported_resources)
12
13 def scan_resource_conf(self, conf):
14 bucket_name = conf['name']
15 #check fot logging
16 if 'logging' in conf.keys():
17 if conf['logging'][0]['log_bucket']:
18 log_bucket_name = conf['logging'][0]['log_bucket']
19 if log_bucket_name != bucket_name:
20 return CheckResult.PASSED
21 else:
22 return CheckResult.FAILED
23 else:
24 return CheckResult.FAILED
25 return CheckResult.FAILED
26 return CheckResult.UNKNOWN
27
28 check = CloudStorageSelfLogging()
29
[end of checkov/terraform/checks/resource/gcp/CloudStorageSelfLogging.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/checkov/terraform/checks/resource/gcp/CloudStorageLogging.py b/checkov/terraform/checks/resource/gcp/CloudStorageLogging.py
--- a/checkov/terraform/checks/resource/gcp/CloudStorageLogging.py
+++ b/checkov/terraform/checks/resource/gcp/CloudStorageLogging.py
@@ -11,9 +11,9 @@
super().__init__(name=name, id=id, categories=categories, supported_resources=supported_resources)
def scan_resource_conf(self, conf):
- #check fot logging
- if 'logging' in conf.keys():
- if conf['logging'][0]['log_bucket']:
+ #check for logging
+ if 'logging' in conf:
+ if conf['logging'][0]:
log_bucket_name = conf['logging'][0]['log_bucket']
if log_bucket_name != None:
return CheckResult.PASSED
diff --git a/checkov/terraform/checks/resource/gcp/CloudStorageSelfLogging.py b/checkov/terraform/checks/resource/gcp/CloudStorageSelfLogging.py
--- a/checkov/terraform/checks/resource/gcp/CloudStorageSelfLogging.py
+++ b/checkov/terraform/checks/resource/gcp/CloudStorageSelfLogging.py
@@ -12,9 +12,9 @@
def scan_resource_conf(self, conf):
bucket_name = conf['name']
- #check fot logging
- if 'logging' in conf.keys():
- if conf['logging'][0]['log_bucket']:
+ #check for logging
+ if 'logging' in conf:
+ if conf['logging'][0]:
log_bucket_name = conf['logging'][0]['log_bucket']
if log_bucket_name != bucket_name:
return CheckResult.PASSED
|
{"golden_diff": "diff --git a/checkov/terraform/checks/resource/gcp/CloudStorageLogging.py b/checkov/terraform/checks/resource/gcp/CloudStorageLogging.py\n--- a/checkov/terraform/checks/resource/gcp/CloudStorageLogging.py\n+++ b/checkov/terraform/checks/resource/gcp/CloudStorageLogging.py\n@@ -11,9 +11,9 @@\n super().__init__(name=name, id=id, categories=categories, supported_resources=supported_resources)\n \n def scan_resource_conf(self, conf):\n- #check fot logging\n- if 'logging' in conf.keys():\n- if conf['logging'][0]['log_bucket']:\n+ #check for logging\n+ if 'logging' in conf:\n+ if conf['logging'][0]:\n log_bucket_name = conf['logging'][0]['log_bucket']\n if log_bucket_name != None:\n return CheckResult.PASSED\ndiff --git a/checkov/terraform/checks/resource/gcp/CloudStorageSelfLogging.py b/checkov/terraform/checks/resource/gcp/CloudStorageSelfLogging.py\n--- a/checkov/terraform/checks/resource/gcp/CloudStorageSelfLogging.py\n+++ b/checkov/terraform/checks/resource/gcp/CloudStorageSelfLogging.py\n@@ -12,9 +12,9 @@\n \n def scan_resource_conf(self, conf):\n bucket_name = conf['name']\n- #check fot logging\n- if 'logging' in conf.keys():\n- if conf['logging'][0]['log_bucket']:\n+ #check for logging\n+ if 'logging' in conf:\n+ if conf['logging'][0]:\n log_bucket_name = conf['logging'][0]['log_bucket']\n if log_bucket_name != bucket_name:\n return CheckResult.PASSED\n", "issue": "ERROR:checkov.terraform.checks.resource.gcp.CloudStorageSelfLogging:Failed to run check: Bucket should not log to itself for configuration\n**Describe the bug**\r\nError during check with rule CKV_GCP_63 (not clearly identify)\r\n\r\nHere is the result of my terraform plan for my GCS resource using [terraform google storage module](https://github.com/terraform-google-modules/terraform-google-cloud-storage)\r\n```\r\nresource \"google_storage_bucket\" \"buckets\" {\r\n bucket_policy_only = (known after apply)\r\n force_destroy = false\r\n id = (known after apply)\r\n labels = {\r\n \"name\" = \"xxxxxxxx-bucket\"\r\n }\r\n location = \"EU\"\r\n name = \"xxxxxxxx-bucket\"\r\n project = \"xxxxxxxx-project\"\r\n self_link = (known after apply)\r\n storage_class = \"STANDARD\"\r\n uniform_bucket_level_access = true\r\n url = (known after apply)\r\n versioning {\r\n enabled = true\r\n }\r\n }\r\n```\r\n\r\n```\r\nERROR:checkov.terraform.checks.resource.gcp.CloudStorageSelfLogging:Failed to run check: Bucket should not log to itself for configuration\r\n[[]], 'default_event_based_hold': [None], 'encryption': [[]], 'force_destroy': [False], 'labels': [{'name': ['xxxxxxxx-bucket'], 'start_line': [158], 'end_line': [160]}], 'lifecycle_rule': [[]], 'location': ['EU'], 'logging': [[]], 'name': ['xxxxxxxx-bucket'], 'project': ['xxxxxxxx-project'], 'requester_pays': [None], 'retention_policy': [[]], 'storage_class': ['STANDARD'], 'uniform_bucket_level_access': [True], 'versioning': [{'enabled': [True], 'start_line': [171], 'end_line': [173]}], 'website': [[]], 'start_line': [153], 'end_line': [176]} at file: /checkov.tfplan.json\r\nTraceback (most recent call last):\r\n File \"/usr/local/bin/checkov\", line 5, in <module>\r\n run()\r\n File \"/usr/local/lib/python3.8/site-packages/checkov/main.py\", line 96, in run\r\n scan_reports = runner_registry.run(external_checks_dir=external_checks_dir, files=args.file,\r\n File \"/usr/local/lib/python3.8/site-packages/checkov/common/runners/runner_registry.py\", line 34, in run\r\n scan_report = runner.run(root_folder, external_checks_dir=external_checks_dir, files=files,\r\n File \"/usr/local/lib/python3.8/site-packages/checkov/terraform/plan_runner.py\", line 65, in run\r\n self.check_tf_definition(report, runner_filter)\r\n File \"/usr/local/lib/python3.8/site-packages/checkov/terraform/plan_runner.py\", line 79, in check_tf_definition\r\n self.run_block(definition[block_type], full_file_path, report, scanned_file,\r\n File \"/usr/local/lib/python3.8/site-packages/checkov/terraform/plan_runner.py\", line 95, in run_block\r\n results = registry.scan(scanned_file, entity, [], runner_filter)\r\n File \"/usr/local/lib/python3.8/site-packages/checkov/common/checks/base_check_registry.py\", line 109, in scan\r\n result = self.run_check(check, entity_configuration, entity_name, entity_type, scanned_file, skip_info)\r\n File \"/usr/local/lib/python3.8/site-packages/checkov/common/checks/base_check_registry.py\", line 115, in run_check\r\n result = check.run(scanned_file=scanned_file, entity_configuration=entity_configuration,\r\n File \"/usr/local/lib/python3.8/site-packages/checkov/common/checks/base_check.py\", line 62, in run\r\n raise e\r\n File \"/usr/local/lib/python3.8/site-packages/checkov/common/checks/base_check.py\", line 42, in run\r\n check_result['result'] = self.scan_entity_conf(entity_configuration, entity_type)\r\n File \"/usr/local/lib/python3.8/site-packages/checkov/terraform/checks/resource/base_resource_check.py\", line 17, in scan_entity_conf\r\n return self.scan_resource_conf(conf, entity_type)\r\n File \"/usr/local/lib/python3.8/site-packages/checkov/terraform/checks/resource/base_resource_check.py\", line 33, in wrapper\r\n return wrapped(self, conf)\r\n File \"/usr/local/lib/python3.8/site-packages/checkov/terraform/checks/resource/gcp/CloudStorageSelfLogging.py\", line 17, in scan_resource_conf\r\n if conf['logging'][0]['log_bucket']:\r\nTypeError: list indices must be integers or slices, not str\r\n```\r\n\r\n**To Reproduce**\r\nSteps to reproduce the behavior:\r\n1. Using [terraform GCS module](https://github.com/terraform-google-modules/terraform-google-cloud-storage) in your terraform\r\n2. Run checkov scan\r\n3. See error\r\n\r\n**Informations**\r\n - Checkov Version 1.0.799 (from docker image)\n", "before_files": [{"content": "from checkov.common.models.enums import CheckResult, CheckCategories\nfrom checkov.terraform.checks.resource.base_resource_check import BaseResourceCheck\n\n\nclass CloudStorageLogging(BaseResourceCheck):\n def __init__(self):\n name = \"Bucket should log access\"\n id = \"CKV_GCP_62\"\n supported_resources = ['google_storage_bucket']\n categories = [CheckCategories.LOGGING]\n super().__init__(name=name, id=id, categories=categories, supported_resources=supported_resources)\n\n def scan_resource_conf(self, conf):\n #check fot logging\n if 'logging' in conf.keys():\n if conf['logging'][0]['log_bucket']:\n log_bucket_name = conf['logging'][0]['log_bucket']\n if log_bucket_name != None:\n return CheckResult.PASSED\n else:\n return CheckResult.FAILED\n else:\n return CheckResult.FAILED\n return CheckResult.FAILED\n else:\n return CheckResult.FAILED\n return CheckResult.FAILED\n\ncheck = CloudStorageLogging()\n", "path": "checkov/terraform/checks/resource/gcp/CloudStorageLogging.py"}, {"content": "from checkov.common.models.enums import CheckResult, CheckCategories\nfrom checkov.terraform.checks.resource.base_resource_check import BaseResourceCheck\n\n\nclass CloudStorageSelfLogging(BaseResourceCheck):\n def __init__(self):\n name = \"Bucket should not log to itself\"\n id = \"CKV_GCP_63\"\n supported_resources = ['google_storage_bucket']\n categories = [CheckCategories.LOGGING]\n super().__init__(name=name, id=id, categories=categories, supported_resources=supported_resources)\n\n def scan_resource_conf(self, conf):\n bucket_name = conf['name']\n #check fot logging\n if 'logging' in conf.keys():\n if conf['logging'][0]['log_bucket']:\n log_bucket_name = conf['logging'][0]['log_bucket']\n if log_bucket_name != bucket_name:\n return CheckResult.PASSED\n else:\n return CheckResult.FAILED\n else:\n return CheckResult.FAILED\n return CheckResult.FAILED\n return CheckResult.UNKNOWN\n\ncheck = CloudStorageSelfLogging()\n", "path": "checkov/terraform/checks/resource/gcp/CloudStorageSelfLogging.py"}]}
| 2,239 | 382 |
gh_patches_debug_38632
|
rasdani/github-patches
|
git_diff
|
kserve__kserve-549
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
XGBoost Model Perfomance Issue
/kind bug
XGboost model server perform poorly after deployed to k8s.
**What steps did you take and what happened:**
I did performance test for xgbserver using provided sample model.
`load.cfg`:
```
POST http://<target>/v1/models/default:predict
@request.json
```
`req.json`:
```
{"instances": [[1,2,3,4]]}
```
Running locally (on Mac OS, 8 core machine) with vegeta
```
cat load.cfg | vegeta attack -duration 1m -rate 1000 | vegeta report -type=text
```
I am able to hit 1000 RPS with good latencies result.
```
Requests [total, rate, throughput] 60000, 1000.02, 1000.00
Duration [total, attack, wait] 1m0.000012024s, 59.999069988s, 942.036µs
Latencies [mean, 50, 90, 95, 99, max] 1.070524ms, 946.443µs, 1.414653ms, 1.520775ms, 2.401196ms, 17.650717ms
Bytes In [total, mean] 1320000, 22.00
Bytes Out [total, mean] 1560000, 26.00
Success [ratio] 100.00%
Status Codes [code:count] 200:60000
```
But after deployed to K8S (with 4 core as CPU request and limit)
The performance deteriorates at 100 RPS
```
cat load-1.cfg | vegeta attack -rate 100 -duration 60s | vegeta report
```
```
Requests [total, rate, throughput] 6000, 100.02, 89.47
Duration [total, attack, wait] 1m7.064629048s, 59.989775627s, 7.074853421s
Latencies [mean, 50, 90, 95, 99, max] 1.524605151s, 3.683733ms, 5.575147148s, 7.007148666s, 7.789309456s, 8.606190814s
Bytes In [total, mean] 132000, 22.00
Bytes Out [total, mean] 192000, 32.00
Success [ratio] 100.00%
Status Codes [code:count] 200:6000
```
**What did you expect to happen:**
The performance should not be affected too much after deployed to K8S.
**Anything else you would like to add:**
I suspected the issue come from `nthread` in xgbserver are left to default (i.e. 0).
In:
https://github.com/kubeflow/kfserving/blob/c46a0da8c0cf168dccfcee3b79f306fea135faf4/python/xgbserver/xgbserver/model.py#L36
and
https://github.com/kubeflow/kfserving/blob/c46a0da8c0cf168dccfcee3b79f306fea135faf4/python/xgbserver/xgbserver/model.py#L42
Which will leave the decision of number of thread to use to xgboost by detecting number of cpu. However, possibly due to this https://bugs.python.org/issue36054, the `nthread` is incorrect and causing the performance issues.
I did a test with setting `nthread=1` :
```
class XGBoostModel(kfserving.KFModel):
def __init__(self, name: str, model_dir: str, booster: XGBModel = None):
super().__init__(name)
self.name = name
self.model_dir = model_dir
if not booster is None:
self._booster = booster
self.ready = True
def load(self):
model_file = os.path.join(
kfserving.Storage.download(self.model_dir), BOOSTER_FILE)
self._booster = xgb.Booster(params={"nthread" = 1},model_file=model_file)
self.ready = True
def predict(self, request: Dict) -> Dict:
try:
# Use of list as input is deprecated see https://github.com/dmlc/xgboost/pull/3970
dmatrix = xgb.DMatrix(request["instances"], nthread=1)
result: xgb.DMatrix = self._booster.predict(dmatrix)
return { "predictions": result.tolist() }
except Exception as e:
raise Exception("Failed to predict %s" % e)
```
```
cat load.cfg | vegeta attack -rate 1000 -duration 60s | vegeta report
```
and it performed much better:
```
Requests [total, rate, throughput] 60000, 1000.02, 999.99
Duration [total, attack, wait] 1m0.00064503s, 59.9988685s, 1.77653ms
Latencies [mean, 50, 90, 95, 99, max] 1.927362ms, 1.73989ms, 1.941139ms, 2.490536ms, 5.660895ms, 56.707145ms
Bytes In [total, mean] 1320000, 22.00
Bytes Out [total, mean] 1920000, 32.00
Success [ratio] 100.00%
Status Codes [code:count] 200:60000
```
The question here is:
1. Is setting `nthread=1` a good idea for other kind of workload?
2. Should this be configurable?
**Environment:**
- Istio Version: 1.1.7
- Knative Version: 0.8.1
- KFServing Version: 0.2.1
- Kubeflow version: NA
- Minikube version: NA
- Kubernetes version: (use `kubectl version`):
- OS (e.g. from `/etc/os-release`):
</issue>
<code>
[start of python/xgbserver/xgbserver/model.py]
1 # Copyright 2019 kubeflow.org.
2 #
3 # Licensed under the Apache License, Version 2.0 (the "License");
4 # you may not use this file except in compliance with the License.
5 # You may obtain a copy of the License at
6 #
7 # http://www.apache.org/licenses/LICENSE-2.0
8 #
9 # Unless required by applicable law or agreed to in writing, software
10 # distributed under the License is distributed on an "AS IS" BASIS,
11 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12 # See the License for the specific language governing permissions and
13 # limitations under the License.
14
15 import kfserving
16 import xgboost as xgb
17 from xgboost import XGBModel
18 import os
19 import numpy as np
20 from typing import List, Dict
21
22 BOOSTER_FILE = "model.bst"
23
24 class XGBoostModel(kfserving.KFModel):
25 def __init__(self, name: str, model_dir: str, booster: XGBModel = None):
26 super().__init__(name)
27 self.name = name
28 self.model_dir = model_dir
29 if not booster is None:
30 self._booster = booster
31 self.ready = True
32
33 def load(self):
34 model_file = os.path.join(
35 kfserving.Storage.download(self.model_dir), BOOSTER_FILE)
36 self._booster = xgb.Booster(model_file=model_file)
37 self.ready = True
38
39 def predict(self, request: Dict) -> Dict:
40 try:
41 # Use of list as input is deprecated see https://github.com/dmlc/xgboost/pull/3970
42 dmatrix = xgb.DMatrix(request["instances"])
43 result: xgb.DMatrix = self._booster.predict(dmatrix)
44 return { "predictions": result.tolist() }
45 except Exception as e:
46 raise Exception("Failed to predict %s" % e)
47
[end of python/xgbserver/xgbserver/model.py]
[start of python/xgbserver/xgbserver/__main__.py]
1 # Copyright 2019 kubeflow.org.
2 #
3 # Licensed under the Apache License, Version 2.0 (the "License");
4 # you may not use this file except in compliance with the License.
5 # You may obtain a copy of the License at
6 #
7 # http://www.apache.org/licenses/LICENSE-2.0
8 #
9 # Unless required by applicable law or agreed to in writing, software
10 # distributed under the License is distributed on an "AS IS" BASIS,
11 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12 # See the License for the specific language governing permissions and
13 # limitations under the License.
14
15 import kfserving
16 import argparse
17
18 from xgbserver import XGBoostModel
19
20 DEFAULT_MODEL_NAME = "default"
21 DEFAULT_LOCAL_MODEL_DIR = "/tmp/model"
22
23 parser = argparse.ArgumentParser(parents=[kfserving.kfserver.parser]) #pylint:disable=c-extension-no-member
24 parser.add_argument('--model_dir', required=True,
25 help='A URI pointer to the model directory')
26 parser.add_argument('--model_name', default=DEFAULT_MODEL_NAME,
27 help='The name that the model is served under.')
28 args, _ = parser.parse_known_args()
29
30 if __name__ == "__main__":
31 model = XGBoostModel(args.model_name, args.model_dir)
32 model.load()
33 kfserving.KFServer().start([model]) #pylint:disable=c-extension-no-member
34
[end of python/xgbserver/xgbserver/__main__.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/python/xgbserver/xgbserver/__main__.py b/python/xgbserver/xgbserver/__main__.py
--- a/python/xgbserver/xgbserver/__main__.py
+++ b/python/xgbserver/xgbserver/__main__.py
@@ -19,15 +19,18 @@
DEFAULT_MODEL_NAME = "default"
DEFAULT_LOCAL_MODEL_DIR = "/tmp/model"
+DEFAULT_NTHREAD = 1
parser = argparse.ArgumentParser(parents=[kfserving.kfserver.parser]) #pylint:disable=c-extension-no-member
parser.add_argument('--model_dir', required=True,
help='A URI pointer to the model directory')
parser.add_argument('--model_name', default=DEFAULT_MODEL_NAME,
help='The name that the model is served under.')
+parser.add_argument('--nthread', default=DEFAULT_NTHREAD,
+ help='Number of threads to use by XGBoost.')
args, _ = parser.parse_known_args()
if __name__ == "__main__":
- model = XGBoostModel(args.model_name, args.model_dir)
+ model = XGBoostModel(args.model_name, args.model_dir, args.nthread)
model.load()
kfserving.KFServer().start([model]) #pylint:disable=c-extension-no-member
diff --git a/python/xgbserver/xgbserver/model.py b/python/xgbserver/xgbserver/model.py
--- a/python/xgbserver/xgbserver/model.py
+++ b/python/xgbserver/xgbserver/model.py
@@ -22,10 +22,12 @@
BOOSTER_FILE = "model.bst"
class XGBoostModel(kfserving.KFModel):
- def __init__(self, name: str, model_dir: str, booster: XGBModel = None):
+ def __init__(self, name: str, model_dir: str, nthread: int, booster: \
+ XGBModel = None):
super().__init__(name)
self.name = name
self.model_dir = model_dir
+ self.nthread = nthread
if not booster is None:
self._booster = booster
self.ready = True
@@ -33,13 +35,14 @@
def load(self):
model_file = os.path.join(
kfserving.Storage.download(self.model_dir), BOOSTER_FILE)
- self._booster = xgb.Booster(model_file=model_file)
+ self._booster = xgb.Booster(params={"nthread" : self.nthread},
+ model_file=model_file)
self.ready = True
def predict(self, request: Dict) -> Dict:
try:
# Use of list as input is deprecated see https://github.com/dmlc/xgboost/pull/3970
- dmatrix = xgb.DMatrix(request["instances"])
+ dmatrix = xgb.DMatrix(request["instances"], nthread=self.nthread)
result: xgb.DMatrix = self._booster.predict(dmatrix)
return { "predictions": result.tolist() }
except Exception as e:
|
{"golden_diff": "diff --git a/python/xgbserver/xgbserver/__main__.py b/python/xgbserver/xgbserver/__main__.py\n--- a/python/xgbserver/xgbserver/__main__.py\n+++ b/python/xgbserver/xgbserver/__main__.py\n@@ -19,15 +19,18 @@\n \n DEFAULT_MODEL_NAME = \"default\"\n DEFAULT_LOCAL_MODEL_DIR = \"/tmp/model\"\n+DEFAULT_NTHREAD = 1\n \n parser = argparse.ArgumentParser(parents=[kfserving.kfserver.parser]) #pylint:disable=c-extension-no-member\n parser.add_argument('--model_dir', required=True,\n help='A URI pointer to the model directory')\n parser.add_argument('--model_name', default=DEFAULT_MODEL_NAME,\n help='The name that the model is served under.')\n+parser.add_argument('--nthread', default=DEFAULT_NTHREAD,\n+ help='Number of threads to use by XGBoost.')\n args, _ = parser.parse_known_args()\n \n if __name__ == \"__main__\":\n- model = XGBoostModel(args.model_name, args.model_dir)\n+ model = XGBoostModel(args.model_name, args.model_dir, args.nthread)\n model.load()\n kfserving.KFServer().start([model]) #pylint:disable=c-extension-no-member\ndiff --git a/python/xgbserver/xgbserver/model.py b/python/xgbserver/xgbserver/model.py\n--- a/python/xgbserver/xgbserver/model.py\n+++ b/python/xgbserver/xgbserver/model.py\n@@ -22,10 +22,12 @@\n BOOSTER_FILE = \"model.bst\"\n \n class XGBoostModel(kfserving.KFModel):\n- def __init__(self, name: str, model_dir: str, booster: XGBModel = None):\n+ def __init__(self, name: str, model_dir: str, nthread: int, booster: \\\n+ XGBModel = None):\n super().__init__(name)\n self.name = name\n self.model_dir = model_dir\n+ self.nthread = nthread\n if not booster is None:\n self._booster = booster\n self.ready = True\n@@ -33,13 +35,14 @@\n def load(self):\n model_file = os.path.join(\n kfserving.Storage.download(self.model_dir), BOOSTER_FILE)\n- self._booster = xgb.Booster(model_file=model_file)\n+ self._booster = xgb.Booster(params={\"nthread\" : self.nthread},\n+ model_file=model_file)\n self.ready = True\n \n def predict(self, request: Dict) -> Dict:\n try:\n # Use of list as input is deprecated see https://github.com/dmlc/xgboost/pull/3970\n- dmatrix = xgb.DMatrix(request[\"instances\"])\n+ dmatrix = xgb.DMatrix(request[\"instances\"], nthread=self.nthread)\n result: xgb.DMatrix = self._booster.predict(dmatrix)\n return { \"predictions\": result.tolist() }\n except Exception as e:\n", "issue": "XGBoost Model Perfomance Issue\n/kind bug\r\nXGboost model server perform poorly after deployed to k8s.\r\n\r\n**What steps did you take and what happened:**\r\nI did performance test for xgbserver using provided sample model.\r\n\r\n`load.cfg`:\r\n```\r\nPOST http://<target>/v1/models/default:predict\r\[email protected]\r\n```\r\n\r\n`req.json`:\r\n```\r\n{\"instances\": [[1,2,3,4]]}\r\n```\r\nRunning locally (on Mac OS, 8 core machine) with vegeta\r\n```\r\ncat load.cfg | vegeta attack -duration 1m -rate 1000 | vegeta report -type=text\r\n```\r\n\r\nI am able to hit 1000 RPS with good latencies result.\r\n\r\n```\r\nRequests [total, rate, throughput] 60000, 1000.02, 1000.00\r\nDuration [total, attack, wait] 1m0.000012024s, 59.999069988s, 942.036\u00b5s\r\nLatencies [mean, 50, 90, 95, 99, max] 1.070524ms, 946.443\u00b5s, 1.414653ms, 1.520775ms, 2.401196ms, 17.650717ms\r\nBytes In [total, mean] 1320000, 22.00\r\nBytes Out [total, mean] 1560000, 26.00\r\nSuccess [ratio] 100.00%\r\nStatus Codes [code:count] 200:60000 \r\n```\r\n\r\nBut after deployed to K8S (with 4 core as CPU request and limit)\r\n\r\nThe performance deteriorates at 100 RPS\r\n```\r\ncat load-1.cfg | vegeta attack -rate 100 -duration 60s | vegeta report\r\n```\r\n\r\n```\r\nRequests [total, rate, throughput] 6000, 100.02, 89.47\r\nDuration [total, attack, wait] 1m7.064629048s, 59.989775627s, 7.074853421s\r\nLatencies [mean, 50, 90, 95, 99, max] 1.524605151s, 3.683733ms, 5.575147148s, 7.007148666s, 7.789309456s, 8.606190814s\r\nBytes In [total, mean] 132000, 22.00\r\nBytes Out [total, mean] 192000, 32.00\r\nSuccess [ratio] 100.00%\r\nStatus Codes [code:count] 200:6000\r\n```\r\n\r\n**What did you expect to happen:**\r\n\r\nThe performance should not be affected too much after deployed to K8S.\r\n\r\n**Anything else you would like to add:**\r\nI suspected the issue come from `nthread` in xgbserver are left to default (i.e. 0). \r\nIn: \r\nhttps://github.com/kubeflow/kfserving/blob/c46a0da8c0cf168dccfcee3b79f306fea135faf4/python/xgbserver/xgbserver/model.py#L36\r\nand\r\nhttps://github.com/kubeflow/kfserving/blob/c46a0da8c0cf168dccfcee3b79f306fea135faf4/python/xgbserver/xgbserver/model.py#L42\r\n\r\nWhich will leave the decision of number of thread to use to xgboost by detecting number of cpu. However, possibly due to this https://bugs.python.org/issue36054, the `nthread` is incorrect and causing the performance issues.\r\n\r\nI did a test with setting `nthread=1` :\r\n```\r\nclass XGBoostModel(kfserving.KFModel):\r\n def __init__(self, name: str, model_dir: str, booster: XGBModel = None):\r\n super().__init__(name)\r\n self.name = name\r\n self.model_dir = model_dir\r\n if not booster is None:\r\n self._booster = booster\r\n self.ready = True\r\n\r\n def load(self):\r\n model_file = os.path.join(\r\n kfserving.Storage.download(self.model_dir), BOOSTER_FILE)\r\n self._booster = xgb.Booster(params={\"nthread\" = 1},model_file=model_file)\r\n self.ready = True\r\n\r\n def predict(self, request: Dict) -> Dict:\r\n try:\r\n # Use of list as input is deprecated see https://github.com/dmlc/xgboost/pull/3970\r\n dmatrix = xgb.DMatrix(request[\"instances\"], nthread=1)\r\n result: xgb.DMatrix = self._booster.predict(dmatrix)\r\n return { \"predictions\": result.tolist() }\r\n except Exception as e:\r\n raise Exception(\"Failed to predict %s\" % e)\r\n```\r\n\r\n```\r\ncat load.cfg | vegeta attack -rate 1000 -duration 60s | vegeta report\r\n```\r\n\r\nand it performed much better:\r\n\r\n```\r\nRequests [total, rate, throughput] 60000, 1000.02, 999.99\r\nDuration [total, attack, wait] 1m0.00064503s, 59.9988685s, 1.77653ms\r\nLatencies [mean, 50, 90, 95, 99, max] 1.927362ms, 1.73989ms, 1.941139ms, 2.490536ms, 5.660895ms, 56.707145ms\r\nBytes In [total, mean] 1320000, 22.00\r\nBytes Out [total, mean] 1920000, 32.00\r\nSuccess [ratio] 100.00%\r\nStatus Codes [code:count] 200:60000\r\n```\r\n\r\nThe question here is:\r\n1. Is setting `nthread=1` a good idea for other kind of workload?\r\n2. Should this be configurable?\r\n\r\n**Environment:**\r\n\r\n- Istio Version: 1.1.7\r\n- Knative Version: 0.8.1\r\n- KFServing Version: 0.2.1\r\n- Kubeflow version: NA\r\n- Minikube version: NA\r\n- Kubernetes version: (use `kubectl version`):\r\n- OS (e.g. from `/etc/os-release`):\r\n\n", "before_files": [{"content": "# Copyright 2019 kubeflow.org.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nimport kfserving\nimport xgboost as xgb\nfrom xgboost import XGBModel\nimport os\nimport numpy as np\nfrom typing import List, Dict\n\nBOOSTER_FILE = \"model.bst\"\n\nclass XGBoostModel(kfserving.KFModel):\n def __init__(self, name: str, model_dir: str, booster: XGBModel = None):\n super().__init__(name)\n self.name = name\n self.model_dir = model_dir\n if not booster is None:\n self._booster = booster\n self.ready = True\n\n def load(self):\n model_file = os.path.join(\n kfserving.Storage.download(self.model_dir), BOOSTER_FILE)\n self._booster = xgb.Booster(model_file=model_file)\n self.ready = True\n\n def predict(self, request: Dict) -> Dict:\n try:\n # Use of list as input is deprecated see https://github.com/dmlc/xgboost/pull/3970\n dmatrix = xgb.DMatrix(request[\"instances\"])\n result: xgb.DMatrix = self._booster.predict(dmatrix)\n return { \"predictions\": result.tolist() }\n except Exception as e:\n raise Exception(\"Failed to predict %s\" % e)\n", "path": "python/xgbserver/xgbserver/model.py"}, {"content": "# Copyright 2019 kubeflow.org.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nimport kfserving\nimport argparse\n\nfrom xgbserver import XGBoostModel\n\nDEFAULT_MODEL_NAME = \"default\"\nDEFAULT_LOCAL_MODEL_DIR = \"/tmp/model\"\n\nparser = argparse.ArgumentParser(parents=[kfserving.kfserver.parser]) #pylint:disable=c-extension-no-member\nparser.add_argument('--model_dir', required=True,\n help='A URI pointer to the model directory')\nparser.add_argument('--model_name', default=DEFAULT_MODEL_NAME,\n help='The name that the model is served under.')\nargs, _ = parser.parse_known_args()\n\nif __name__ == \"__main__\":\n model = XGBoostModel(args.model_name, args.model_dir)\n model.load()\n kfserving.KFServer().start([model]) #pylint:disable=c-extension-no-member\n", "path": "python/xgbserver/xgbserver/__main__.py"}]}
| 3,093 | 665 |
gh_patches_debug_18255
|
rasdani/github-patches
|
git_diff
|
xonsh__xonsh-1823
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Prompt `curr_branch` breaks when git config has status.short
If git config contains:
``` ini
[status]
short = true
```
then [the current method of getting the git branch](https://github.com/xonsh/xonsh/blob/619ae11486f2e4febfc2218c0d53c7b3f3c9ea49/xonsh/prompt/vc_branch.py#L16-L25) fails, even if `status.branch = true` is also set.
`git status` is a heavy-handed way of doing this anyway; better would be `git rev-parse --abbrev-ref HEAD` IMO. (PR to follow.)
</issue>
<code>
[start of xonsh/prompt/vc_branch.py]
1 # -*- coding: utf-8 -*-
2 """Prompt formatter for simple version control branchs"""
3
4 import os
5 import sys
6 import time
7 import queue
8 import builtins
9 import warnings
10 import threading
11 import subprocess
12
13 import xonsh.tools as xt
14
15
16 def _get_git_branch(q):
17 try:
18 status = subprocess.check_output(['git', 'status'],
19 stderr=subprocess.DEVNULL)
20 except (subprocess.CalledProcessError, OSError):
21 q.put(None)
22 else:
23 info = xt.decode_bytes(status)
24 branch = info.splitlines()[0].split()[-1]
25 q.put(branch)
26
27
28 def get_git_branch():
29 """Attempts to find the current git branch. If this could not
30 be determined (timeout, not in a git repo, etc.) then this returns None.
31 """
32 branch = None
33 timeout = builtins.__xonsh_env__.get('VC_BRANCH_TIMEOUT')
34 q = queue.Queue()
35
36 t = threading.Thread(target=_get_git_branch, args=(q,))
37 t.start()
38 t.join(timeout=timeout)
39 try:
40 branch = q.get_nowait()
41 except queue.Empty:
42 branch = None
43 return branch
44
45
46 def _get_parent_dir_for(path, dir_name, timeout):
47 # walk up the directory tree to see if we are inside an hg repo
48 # the timeout makes sure that we don't thrash the file system
49 previous_path = ''
50 t0 = time.time()
51 while path != previous_path and ((time.time() - t0) < timeout):
52 if os.path.isdir(os.path.join(path, dir_name)):
53 return path
54 previous_path = path
55 path, _ = os.path.split(path)
56 return (path == previous_path)
57
58
59 def get_hg_branch(cwd=None, root=None):
60 env = builtins.__xonsh_env__
61 cwd = env['PWD']
62 root = _get_parent_dir_for(cwd, '.hg', env['VC_BRANCH_TIMEOUT'])
63 if not isinstance(root, str):
64 # Bail if we are not in a repo or we timed out
65 if root:
66 return None
67 else:
68 return subprocess.TimeoutExpired(['hg'], env['VC_BRANCH_TIMEOUT'])
69 if env.get('VC_HG_SHOW_BRANCH') is True:
70 # get branch name
71 branch_path = os.path.sep.join([root, '.hg', 'branch'])
72 if os.path.exists(branch_path):
73 with open(branch_path, 'r') as branch_file:
74 branch = branch_file.read()
75 else:
76 branch = 'default'
77 else:
78 branch = ''
79 # add bookmark, if we can
80 bookmark_path = os.path.sep.join([root, '.hg', 'bookmarks.current'])
81 if os.path.exists(bookmark_path):
82 with open(bookmark_path, 'r') as bookmark_file:
83 active_bookmark = bookmark_file.read()
84 if env.get('VC_HG_SHOW_BRANCH') is True:
85 branch = "{0}, {1}".format(*(b.strip(os.linesep) for b in
86 (branch, active_bookmark)))
87 else:
88 branch = active_bookmark.strip(os.linesep)
89 else:
90 branch = branch.strip(os.linesep)
91 return branch
92
93
94 _FIRST_BRANCH_TIMEOUT = True
95
96
97 def _first_branch_timeout_message():
98 global _FIRST_BRANCH_TIMEOUT
99 sbtm = builtins.__xonsh_env__['SUPPRESS_BRANCH_TIMEOUT_MESSAGE']
100 if not _FIRST_BRANCH_TIMEOUT or sbtm:
101 return
102 _FIRST_BRANCH_TIMEOUT = False
103 print('xonsh: branch timeout: computing the branch name, color, or both '
104 'timed out while formatting the prompt. You may avoid this by '
105 'increaing the value of $VC_BRANCH_TIMEOUT or by removing branch '
106 'fields, like {curr_branch}, from your $PROMPT. See the FAQ '
107 'for more details. This message will be suppressed for the remainder '
108 'of this session. To suppress this message permanently, set '
109 '$SUPPRESS_BRANCH_TIMEOUT_MESSAGE = True in your xonshrc file.',
110 file=sys.stderr)
111
112
113 def current_branch(pad=NotImplemented):
114 """Gets the branch for a current working directory. Returns an empty string
115 if the cwd is not a repository. This currently only works for git and hg
116 and should be extended in the future. If a timeout occurred, the string
117 '<branch-timeout>' is returned.
118 """
119 if pad is not NotImplemented:
120 warnings.warn("The pad argument of current_branch has no effect now "
121 "and will be removed in the future")
122 branch = None
123 cmds = builtins.__xonsh_commands_cache__
124 if cmds.lazy_locate_binary('git') or cmds.is_empty():
125 branch = get_git_branch()
126 if (cmds.lazy_locate_binary('hg') or cmds.is_empty()) and not branch:
127 branch = get_hg_branch()
128 if isinstance(branch, subprocess.TimeoutExpired):
129 branch = '<branch-timeout>'
130 _first_branch_timeout_message()
131 return branch or None
132
133
134 def _git_dirty_working_directory(q, include_untracked):
135 status = None
136 try:
137 cmd = ['git', 'status', '--porcelain']
138 if include_untracked:
139 cmd.append('--untracked-files=normal')
140 else:
141 cmd.append('--untracked-files=no')
142 status = subprocess.check_output(cmd, stderr=subprocess.DEVNULL)
143 except (subprocess.CalledProcessError, OSError):
144 q.put(None)
145 if status is not None:
146 return q.put(bool(status))
147
148
149 def git_dirty_working_directory(include_untracked=False):
150 """Returns whether or not the git directory is dirty. If this could not
151 be determined (timeout, file not found, etc.) then this returns None.
152 """
153 timeout = builtins.__xonsh_env__.get("VC_BRANCH_TIMEOUT")
154 q = queue.Queue()
155 t = threading.Thread(target=_git_dirty_working_directory,
156 args=(q, include_untracked))
157 t.start()
158 t.join(timeout=timeout)
159 try:
160 return q.get_nowait()
161 except queue.Empty:
162 return None
163
164
165 def hg_dirty_working_directory():
166 """Computes whether or not the mercurial working directory is dirty or not.
167 If this cannot be deterimined, None is returned.
168 """
169 env = builtins.__xonsh_env__
170 cwd = env['PWD']
171 denv = env.detype()
172 vcbt = env['VC_BRANCH_TIMEOUT']
173 # Override user configurations settings and aliases
174 denv['HGRCPATH'] = ''
175 try:
176 s = subprocess.check_output(['hg', 'identify', '--id'],
177 stderr=subprocess.PIPE, cwd=cwd,
178 timeout=vcbt, universal_newlines=True,
179 env=denv)
180 return s.strip(os.linesep).endswith('+')
181 except (subprocess.CalledProcessError, subprocess.TimeoutExpired,
182 FileNotFoundError):
183 return None
184
185
186 def dirty_working_directory(cwd=None):
187 """Returns a boolean as to whether there are uncommitted files in version
188 control repository we are inside. If this cannot be determined, returns
189 None. Currently supports git and hg.
190 """
191 dwd = None
192 cmds = builtins.__xonsh_commands_cache__
193 if cmds.lazy_locate_binary('git') or cmds.is_empty():
194 dwd = git_dirty_working_directory()
195 if (cmds.lazy_locate_binary('hg') or cmds.is_empty()) and (dwd is None):
196 dwd = hg_dirty_working_directory()
197 return dwd
198
199
200 def branch_color():
201 """Return red if the current branch is dirty, yellow if the dirtiness can
202 not be determined, and green if it clean. These are bold, intense colors
203 for the foreground.
204 """
205 dwd = dirty_working_directory()
206 if dwd is None:
207 color = '{BOLD_INTENSE_YELLOW}'
208 elif dwd:
209 color = '{BOLD_INTENSE_RED}'
210 else:
211 color = '{BOLD_INTENSE_GREEN}'
212 return color
213
214
215 def branch_bg_color():
216 """Return red if the current branch is dirty, yellow if the dirtiness can
217 not be determined, and green if it clean. These are bacground colors.
218 """
219 dwd = dirty_working_directory()
220 if dwd is None:
221 color = '{BACKGROUND_YELLOW}'
222 elif dwd:
223 color = '{BACKGROUND_RED}'
224 else:
225 color = '{BACKGROUND_GREEN}'
226 return color
227
[end of xonsh/prompt/vc_branch.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/xonsh/prompt/vc_branch.py b/xonsh/prompt/vc_branch.py
--- a/xonsh/prompt/vc_branch.py
+++ b/xonsh/prompt/vc_branch.py
@@ -15,15 +15,25 @@
def _get_git_branch(q):
try:
- status = subprocess.check_output(['git', 'status'],
- stderr=subprocess.DEVNULL)
+ branches = xt.decode_bytes(subprocess.check_output(
+ ['git', 'branch'],
+ stderr=subprocess.DEVNULL
+ )).splitlines()
except (subprocess.CalledProcessError, OSError):
q.put(None)
else:
- info = xt.decode_bytes(status)
- branch = info.splitlines()[0].split()[-1]
- q.put(branch)
-
+ for branch in branches:
+ if not branch.startswith('* '):
+ continue
+ elif branch.endswith(')'):
+ branch = branch.split()[-1][:-1]
+ else:
+ branch = branch.split()[-1]
+
+ q.put(branch)
+ break
+ else:
+ q.put(None)
def get_git_branch():
"""Attempts to find the current git branch. If this could not
|
{"golden_diff": "diff --git a/xonsh/prompt/vc_branch.py b/xonsh/prompt/vc_branch.py\n--- a/xonsh/prompt/vc_branch.py\n+++ b/xonsh/prompt/vc_branch.py\n@@ -15,15 +15,25 @@\n \n def _get_git_branch(q):\n try:\n- status = subprocess.check_output(['git', 'status'],\n- stderr=subprocess.DEVNULL)\n+ branches = xt.decode_bytes(subprocess.check_output(\n+ ['git', 'branch'],\n+ stderr=subprocess.DEVNULL\n+ )).splitlines()\n except (subprocess.CalledProcessError, OSError):\n q.put(None)\n else:\n- info = xt.decode_bytes(status)\n- branch = info.splitlines()[0].split()[-1]\n- q.put(branch)\n-\n+ for branch in branches:\n+ if not branch.startswith('* '):\n+ continue\n+ elif branch.endswith(')'):\n+ branch = branch.split()[-1][:-1]\n+ else:\n+ branch = branch.split()[-1]\n+\n+ q.put(branch)\n+ break\n+ else:\n+ q.put(None)\n \n def get_git_branch():\n \"\"\"Attempts to find the current git branch. If this could not\n", "issue": "Prompt `curr_branch` breaks when git config has status.short\nIf git config contains:\n\n``` ini\n[status]\n short = true\n```\n\nthen [the current method of getting the git branch](https://github.com/xonsh/xonsh/blob/619ae11486f2e4febfc2218c0d53c7b3f3c9ea49/xonsh/prompt/vc_branch.py#L16-L25) fails, even if `status.branch = true` is also set.\n\n`git status` is a heavy-handed way of doing this anyway; better would be `git rev-parse --abbrev-ref HEAD` IMO. (PR to follow.)\n\n", "before_files": [{"content": "# -*- coding: utf-8 -*-\n\"\"\"Prompt formatter for simple version control branchs\"\"\"\n\nimport os\nimport sys\nimport time\nimport queue\nimport builtins\nimport warnings\nimport threading\nimport subprocess\n\nimport xonsh.tools as xt\n\n\ndef _get_git_branch(q):\n try:\n status = subprocess.check_output(['git', 'status'],\n stderr=subprocess.DEVNULL)\n except (subprocess.CalledProcessError, OSError):\n q.put(None)\n else:\n info = xt.decode_bytes(status)\n branch = info.splitlines()[0].split()[-1]\n q.put(branch)\n\n\ndef get_git_branch():\n \"\"\"Attempts to find the current git branch. If this could not\n be determined (timeout, not in a git repo, etc.) then this returns None.\n \"\"\"\n branch = None\n timeout = builtins.__xonsh_env__.get('VC_BRANCH_TIMEOUT')\n q = queue.Queue()\n\n t = threading.Thread(target=_get_git_branch, args=(q,))\n t.start()\n t.join(timeout=timeout)\n try:\n branch = q.get_nowait()\n except queue.Empty:\n branch = None\n return branch\n\n\ndef _get_parent_dir_for(path, dir_name, timeout):\n # walk up the directory tree to see if we are inside an hg repo\n # the timeout makes sure that we don't thrash the file system\n previous_path = ''\n t0 = time.time()\n while path != previous_path and ((time.time() - t0) < timeout):\n if os.path.isdir(os.path.join(path, dir_name)):\n return path\n previous_path = path\n path, _ = os.path.split(path)\n return (path == previous_path)\n\n\ndef get_hg_branch(cwd=None, root=None):\n env = builtins.__xonsh_env__\n cwd = env['PWD']\n root = _get_parent_dir_for(cwd, '.hg', env['VC_BRANCH_TIMEOUT'])\n if not isinstance(root, str):\n # Bail if we are not in a repo or we timed out\n if root:\n return None\n else:\n return subprocess.TimeoutExpired(['hg'], env['VC_BRANCH_TIMEOUT'])\n if env.get('VC_HG_SHOW_BRANCH') is True:\n # get branch name\n branch_path = os.path.sep.join([root, '.hg', 'branch'])\n if os.path.exists(branch_path):\n with open(branch_path, 'r') as branch_file:\n branch = branch_file.read()\n else:\n branch = 'default'\n else:\n branch = ''\n # add bookmark, if we can\n bookmark_path = os.path.sep.join([root, '.hg', 'bookmarks.current'])\n if os.path.exists(bookmark_path):\n with open(bookmark_path, 'r') as bookmark_file:\n active_bookmark = bookmark_file.read()\n if env.get('VC_HG_SHOW_BRANCH') is True:\n branch = \"{0}, {1}\".format(*(b.strip(os.linesep) for b in\n (branch, active_bookmark)))\n else:\n branch = active_bookmark.strip(os.linesep)\n else:\n branch = branch.strip(os.linesep)\n return branch\n\n\n_FIRST_BRANCH_TIMEOUT = True\n\n\ndef _first_branch_timeout_message():\n global _FIRST_BRANCH_TIMEOUT\n sbtm = builtins.__xonsh_env__['SUPPRESS_BRANCH_TIMEOUT_MESSAGE']\n if not _FIRST_BRANCH_TIMEOUT or sbtm:\n return\n _FIRST_BRANCH_TIMEOUT = False\n print('xonsh: branch timeout: computing the branch name, color, or both '\n 'timed out while formatting the prompt. You may avoid this by '\n 'increaing the value of $VC_BRANCH_TIMEOUT or by removing branch '\n 'fields, like {curr_branch}, from your $PROMPT. See the FAQ '\n 'for more details. This message will be suppressed for the remainder '\n 'of this session. To suppress this message permanently, set '\n '$SUPPRESS_BRANCH_TIMEOUT_MESSAGE = True in your xonshrc file.',\n file=sys.stderr)\n\n\ndef current_branch(pad=NotImplemented):\n \"\"\"Gets the branch for a current working directory. Returns an empty string\n if the cwd is not a repository. This currently only works for git and hg\n and should be extended in the future. If a timeout occurred, the string\n '<branch-timeout>' is returned.\n \"\"\"\n if pad is not NotImplemented:\n warnings.warn(\"The pad argument of current_branch has no effect now \"\n \"and will be removed in the future\")\n branch = None\n cmds = builtins.__xonsh_commands_cache__\n if cmds.lazy_locate_binary('git') or cmds.is_empty():\n branch = get_git_branch()\n if (cmds.lazy_locate_binary('hg') or cmds.is_empty()) and not branch:\n branch = get_hg_branch()\n if isinstance(branch, subprocess.TimeoutExpired):\n branch = '<branch-timeout>'\n _first_branch_timeout_message()\n return branch or None\n\n\ndef _git_dirty_working_directory(q, include_untracked):\n status = None\n try:\n cmd = ['git', 'status', '--porcelain']\n if include_untracked:\n cmd.append('--untracked-files=normal')\n else:\n cmd.append('--untracked-files=no')\n status = subprocess.check_output(cmd, stderr=subprocess.DEVNULL)\n except (subprocess.CalledProcessError, OSError):\n q.put(None)\n if status is not None:\n return q.put(bool(status))\n\n\ndef git_dirty_working_directory(include_untracked=False):\n \"\"\"Returns whether or not the git directory is dirty. If this could not\n be determined (timeout, file not found, etc.) then this returns None.\n \"\"\"\n timeout = builtins.__xonsh_env__.get(\"VC_BRANCH_TIMEOUT\")\n q = queue.Queue()\n t = threading.Thread(target=_git_dirty_working_directory,\n args=(q, include_untracked))\n t.start()\n t.join(timeout=timeout)\n try:\n return q.get_nowait()\n except queue.Empty:\n return None\n\n\ndef hg_dirty_working_directory():\n \"\"\"Computes whether or not the mercurial working directory is dirty or not.\n If this cannot be deterimined, None is returned.\n \"\"\"\n env = builtins.__xonsh_env__\n cwd = env['PWD']\n denv = env.detype()\n vcbt = env['VC_BRANCH_TIMEOUT']\n # Override user configurations settings and aliases\n denv['HGRCPATH'] = ''\n try:\n s = subprocess.check_output(['hg', 'identify', '--id'],\n stderr=subprocess.PIPE, cwd=cwd,\n timeout=vcbt, universal_newlines=True,\n env=denv)\n return s.strip(os.linesep).endswith('+')\n except (subprocess.CalledProcessError, subprocess.TimeoutExpired,\n FileNotFoundError):\n return None\n\n\ndef dirty_working_directory(cwd=None):\n \"\"\"Returns a boolean as to whether there are uncommitted files in version\n control repository we are inside. If this cannot be determined, returns\n None. Currently supports git and hg.\n \"\"\"\n dwd = None\n cmds = builtins.__xonsh_commands_cache__\n if cmds.lazy_locate_binary('git') or cmds.is_empty():\n dwd = git_dirty_working_directory()\n if (cmds.lazy_locate_binary('hg') or cmds.is_empty()) and (dwd is None):\n dwd = hg_dirty_working_directory()\n return dwd\n\n\ndef branch_color():\n \"\"\"Return red if the current branch is dirty, yellow if the dirtiness can\n not be determined, and green if it clean. These are bold, intense colors\n for the foreground.\n \"\"\"\n dwd = dirty_working_directory()\n if dwd is None:\n color = '{BOLD_INTENSE_YELLOW}'\n elif dwd:\n color = '{BOLD_INTENSE_RED}'\n else:\n color = '{BOLD_INTENSE_GREEN}'\n return color\n\n\ndef branch_bg_color():\n \"\"\"Return red if the current branch is dirty, yellow if the dirtiness can\n not be determined, and green if it clean. These are bacground colors.\n \"\"\"\n dwd = dirty_working_directory()\n if dwd is None:\n color = '{BACKGROUND_YELLOW}'\n elif dwd:\n color = '{BACKGROUND_RED}'\n else:\n color = '{BACKGROUND_GREEN}'\n return color\n", "path": "xonsh/prompt/vc_branch.py"}]}
| 3,075 | 277 |
gh_patches_debug_31071
|
rasdani/github-patches
|
git_diff
|
nonebot__nonebot2-772
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Bug: Message.template format 问题
**描述问题:**
无法使用format
**如何复现?**
```py
from nonebot.adapters.onebot.v11 import Message
file_list = [
"file:///file1.image",
"file:///file2.image",
...
]
text = "img1: {:image} img2: {:image} ..."
msg = Message.template(text).format(*file_list)
```
**期望的结果**
`msg` 为可以直接发送的 包含文字和图片的消息
**环境信息:**
- OS: Linux CentOS8
- Python Version: 3.9
- Nonebot Version: 2.0.0-beta.1
**截图或日志**
```log
Traceback (most recent call last):
File "/data/nb-test/test/re_test.py", line 23, in <module>
msg = Message.template(text).format(*file_list)
File "/root/.cache/pypoetry/virtualenvs/nb-test-NOSw4fD6-py3.9/lib/python3.9/site-packages/nonebot/adapters/_template.py", line 80, in format
msg += self.vformat(self.template, args, kwargs)
File "/root/.cache/pypoetry/virtualenvs/nb-test-NOSw4fD6-py3.9/lib/python3.9/site-packages/nonebot/adapters/_template.py", line 94, in vformat
result, _ = self._vformat(format_string, args, kwargs, used_args, 2)
File "/root/.cache/pypoetry/virtualenvs/nb-test-NOSw4fD6-py3.9/lib/python3.9/site-packages/nonebot/adapters/_template.py", line 165, in _vformat
formatted_text = self.format_field(obj, str(format_control))
File "/root/.cache/pypoetry/virtualenvs/nb-test-NOSw4fD6-py3.9/lib/python3.9/site-packages/nonebot/adapters/_template.py", line 181, in format_field
super().format_field(value, format_spec)
File "/root/miniconda3/lib/python3.9/string.py", line 235, in format_field
return format(value, format_spec)
ValueError: Invalid format specifier
```
</issue>
<code>
[start of nonebot/internal/adapter/template.py]
1 import inspect
2 import functools
3 from string import Formatter
4 from typing import (
5 TYPE_CHECKING,
6 Any,
7 Set,
8 Dict,
9 List,
10 Type,
11 Tuple,
12 Union,
13 Generic,
14 Mapping,
15 TypeVar,
16 Callable,
17 Optional,
18 Sequence,
19 cast,
20 overload,
21 )
22
23 if TYPE_CHECKING:
24 from .message import Message, MessageSegment
25
26 TM = TypeVar("TM", bound="Message")
27 TF = TypeVar("TF", str, "Message")
28
29 FormatSpecFunc = Callable[[Any], str]
30 FormatSpecFunc_T = TypeVar("FormatSpecFunc_T", bound=FormatSpecFunc)
31
32
33 class MessageTemplate(Formatter, Generic[TF]):
34 """消息模板格式化实现类。
35
36 参数:
37 template: 模板
38 factory: 消息构造类型,默认为 `str`
39 """
40
41 @overload
42 def __init__(
43 self: "MessageTemplate[str]", template: str, factory: Type[str] = str
44 ) -> None:
45 ...
46
47 @overload
48 def __init__(
49 self: "MessageTemplate[TM]", template: Union[str, TM], factory: Type[TM]
50 ) -> None:
51 ...
52
53 def __init__(self, template, factory=str) -> None:
54 self.template: TF = template
55 self.factory: Type[TF] = factory
56 self.format_specs: Dict[str, FormatSpecFunc] = {}
57
58 def add_format_spec(
59 self, spec: FormatSpecFunc_T, name: Optional[str] = None
60 ) -> FormatSpecFunc_T:
61 name = name or spec.__name__
62 if name in self.format_specs:
63 raise ValueError(f"Format spec {name} already exists!")
64 self.format_specs[name] = spec
65 return spec
66
67 def format(self, *args: Any, **kwargs: Any) -> TF:
68 """根据模板和参数生成消息对象"""
69 msg = self.factory()
70 if isinstance(self.template, str):
71 msg += self.vformat(self.template, args, kwargs)
72 elif isinstance(self.template, self.factory):
73 template = cast("Message[MessageSegment]", self.template)
74 for seg in template:
75 msg += self.vformat(str(seg), args, kwargs) if seg.is_text() else seg
76 else:
77 raise TypeError("template must be a string or instance of Message!")
78
79 return msg # type:ignore
80
81 def vformat(
82 self, format_string: str, args: Sequence[Any], kwargs: Mapping[str, Any]
83 ) -> TF:
84 used_args = set()
85 result, _ = self._vformat(format_string, args, kwargs, used_args, 2)
86 self.check_unused_args(list(used_args), args, kwargs)
87 return result
88
89 def _vformat(
90 self,
91 format_string: str,
92 args: Sequence[Any],
93 kwargs: Mapping[str, Any],
94 used_args: Set[Union[int, str]],
95 recursion_depth: int,
96 auto_arg_index: int = 0,
97 ) -> Tuple[TF, int]:
98 if recursion_depth < 0:
99 raise ValueError("Max string recursion exceeded")
100
101 results: List[Any] = []
102
103 for (literal_text, field_name, format_spec, conversion) in self.parse(
104 format_string
105 ):
106
107 # output the literal text
108 if literal_text:
109 results.append(literal_text)
110
111 # if there's a field, output it
112 if field_name is not None:
113 # this is some markup, find the object and do
114 # the formatting
115
116 # handle arg indexing when empty field_names are given.
117 if field_name == "":
118 if auto_arg_index is False:
119 raise ValueError(
120 "cannot switch from manual field specification to "
121 "automatic field numbering"
122 )
123 field_name = str(auto_arg_index)
124 auto_arg_index += 1
125 elif field_name.isdigit():
126 if auto_arg_index:
127 raise ValueError(
128 "cannot switch from manual field specification to "
129 "automatic field numbering"
130 )
131 # disable auto arg incrementing, if it gets
132 # used later on, then an exception will be raised
133 auto_arg_index = False
134
135 # given the field_name, find the object it references
136 # and the argument it came from
137 obj, arg_used = self.get_field(field_name, args, kwargs)
138 used_args.add(arg_used)
139
140 assert format_spec is not None
141
142 # do any conversion on the resulting object
143 obj = self.convert_field(obj, conversion) if conversion else obj
144
145 # expand the format spec, if needed
146 format_control, auto_arg_index = self._vformat(
147 format_spec,
148 args,
149 kwargs,
150 used_args,
151 recursion_depth - 1,
152 auto_arg_index,
153 )
154
155 # format the object and append to the result
156 formatted_text = self.format_field(obj, str(format_control))
157 results.append(formatted_text)
158
159 return (
160 self.factory(functools.reduce(self._add, results or [""])),
161 auto_arg_index,
162 )
163
164 def format_field(self, value: Any, format_spec: str) -> Any:
165 formatter: Optional[FormatSpecFunc] = self.format_specs.get(format_spec)
166 if formatter is None and not issubclass(self.factory, str):
167 segment_class: Type["MessageSegment"] = self.factory.get_segment_class()
168 method = getattr(segment_class, format_spec, None)
169 if inspect.ismethod(method):
170 formatter = getattr(segment_class, format_spec)
171 return (
172 super().format_field(value, format_spec)
173 if formatter is None
174 else formatter(value)
175 )
176
177 def _add(self, a: Any, b: Any) -> Any:
178 try:
179 return a + b
180 except TypeError:
181 return a + str(b)
182
[end of nonebot/internal/adapter/template.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/nonebot/internal/adapter/template.py b/nonebot/internal/adapter/template.py
--- a/nonebot/internal/adapter/template.py
+++ b/nonebot/internal/adapter/template.py
@@ -1,4 +1,3 @@
-import inspect
import functools
from string import Formatter
from typing import (
@@ -35,7 +34,7 @@
参数:
template: 模板
- factory: 消息构造类型,默认为 `str`
+ factory: 消息类型工厂,默认为 `str`
"""
@overload
@@ -64,8 +63,15 @@
self.format_specs[name] = spec
return spec
- def format(self, *args: Any, **kwargs: Any) -> TF:
- """根据模板和参数生成消息对象"""
+ def format(self, *args, **kwargs):
+ """根据传入参数和模板生成消息对象"""
+ return self._format(args, kwargs)
+
+ def format_map(self, mapping: Mapping[str, Any]) -> TF:
+ """根据传入字典和模板生成消息对象, 在传入字段名不是有效标识符时有用"""
+ return self._format([], mapping)
+
+ def _format(self, args: Sequence[Any], kwargs: Mapping[str, Any]) -> TF:
msg = self.factory()
if isinstance(self.template, str):
msg += self.vformat(self.template, args, kwargs)
@@ -166,7 +172,7 @@
if formatter is None and not issubclass(self.factory, str):
segment_class: Type["MessageSegment"] = self.factory.get_segment_class()
method = getattr(segment_class, format_spec, None)
- if inspect.ismethod(method):
+ if callable(method) and not cast(str, method.__name__).startswith("_"):
formatter = getattr(segment_class, format_spec)
return (
super().format_field(value, format_spec)
|
{"golden_diff": "diff --git a/nonebot/internal/adapter/template.py b/nonebot/internal/adapter/template.py\n--- a/nonebot/internal/adapter/template.py\n+++ b/nonebot/internal/adapter/template.py\n@@ -1,4 +1,3 @@\n-import inspect\n import functools\n from string import Formatter\n from typing import (\n@@ -35,7 +34,7 @@\n \n \u53c2\u6570:\n template: \u6a21\u677f\n- factory: \u6d88\u606f\u6784\u9020\u7c7b\u578b\uff0c\u9ed8\u8ba4\u4e3a `str`\n+ factory: \u6d88\u606f\u7c7b\u578b\u5de5\u5382\uff0c\u9ed8\u8ba4\u4e3a `str`\n \"\"\"\n \n @overload\n@@ -64,8 +63,15 @@\n self.format_specs[name] = spec\n return spec\n \n- def format(self, *args: Any, **kwargs: Any) -> TF:\n- \"\"\"\u6839\u636e\u6a21\u677f\u548c\u53c2\u6570\u751f\u6210\u6d88\u606f\u5bf9\u8c61\"\"\"\n+ def format(self, *args, **kwargs):\n+ \"\"\"\u6839\u636e\u4f20\u5165\u53c2\u6570\u548c\u6a21\u677f\u751f\u6210\u6d88\u606f\u5bf9\u8c61\"\"\"\n+ return self._format(args, kwargs)\n+\n+ def format_map(self, mapping: Mapping[str, Any]) -> TF:\n+ \"\"\"\u6839\u636e\u4f20\u5165\u5b57\u5178\u548c\u6a21\u677f\u751f\u6210\u6d88\u606f\u5bf9\u8c61, \u5728\u4f20\u5165\u5b57\u6bb5\u540d\u4e0d\u662f\u6709\u6548\u6807\u8bc6\u7b26\u65f6\u6709\u7528\"\"\"\n+ return self._format([], mapping)\n+\n+ def _format(self, args: Sequence[Any], kwargs: Mapping[str, Any]) -> TF:\n msg = self.factory()\n if isinstance(self.template, str):\n msg += self.vformat(self.template, args, kwargs)\n@@ -166,7 +172,7 @@\n if formatter is None and not issubclass(self.factory, str):\n segment_class: Type[\"MessageSegment\"] = self.factory.get_segment_class()\n method = getattr(segment_class, format_spec, None)\n- if inspect.ismethod(method):\n+ if callable(method) and not cast(str, method.__name__).startswith(\"_\"):\n formatter = getattr(segment_class, format_spec)\n return (\n super().format_field(value, format_spec)\n", "issue": "Bug: Message.template format \u95ee\u9898\n**\u63cf\u8ff0\u95ee\u9898\uff1a**\r\n\r\n\u65e0\u6cd5\u4f7f\u7528format\r\n\r\n**\u5982\u4f55\u590d\u73b0\uff1f**\r\n\r\n```py\r\nfrom nonebot.adapters.onebot.v11 import Message\r\nfile_list = [\r\n \"file:///file1.image\",\r\n \"file:///file2.image\",\r\n ...\r\n]\r\ntext = \"img1: {:image} img2: {:image} ...\"\r\n\r\nmsg = Message.template(text).format(*file_list)\r\n```\r\n\r\n\r\n**\u671f\u671b\u7684\u7ed3\u679c**\r\n\r\n`msg` \u4e3a\u53ef\u4ee5\u76f4\u63a5\u53d1\u9001\u7684 \u5305\u542b\u6587\u5b57\u548c\u56fe\u7247\u7684\u6d88\u606f\r\n\r\n**\u73af\u5883\u4fe1\u606f\uff1a**\r\n\r\n - OS: Linux CentOS8\r\n - Python Version: 3.9\r\n - Nonebot Version: 2.0.0-beta.1\r\n\r\n\r\n**\u622a\u56fe\u6216\u65e5\u5fd7**\r\n```log\r\nTraceback (most recent call last):\r\n File \"/data/nb-test/test/re_test.py\", line 23, in <module>\r\n msg = Message.template(text).format(*file_list)\r\n File \"/root/.cache/pypoetry/virtualenvs/nb-test-NOSw4fD6-py3.9/lib/python3.9/site-packages/nonebot/adapters/_template.py\", line 80, in format\r\n msg += self.vformat(self.template, args, kwargs)\r\n File \"/root/.cache/pypoetry/virtualenvs/nb-test-NOSw4fD6-py3.9/lib/python3.9/site-packages/nonebot/adapters/_template.py\", line 94, in vformat\r\n result, _ = self._vformat(format_string, args, kwargs, used_args, 2)\r\n File \"/root/.cache/pypoetry/virtualenvs/nb-test-NOSw4fD6-py3.9/lib/python3.9/site-packages/nonebot/adapters/_template.py\", line 165, in _vformat\r\n formatted_text = self.format_field(obj, str(format_control))\r\n File \"/root/.cache/pypoetry/virtualenvs/nb-test-NOSw4fD6-py3.9/lib/python3.9/site-packages/nonebot/adapters/_template.py\", line 181, in format_field\r\n super().format_field(value, format_spec)\r\n File \"/root/miniconda3/lib/python3.9/string.py\", line 235, in format_field\r\n return format(value, format_spec)\r\nValueError: Invalid format specifier\r\n```\n", "before_files": [{"content": "import inspect\nimport functools\nfrom string import Formatter\nfrom typing import (\n TYPE_CHECKING,\n Any,\n Set,\n Dict,\n List,\n Type,\n Tuple,\n Union,\n Generic,\n Mapping,\n TypeVar,\n Callable,\n Optional,\n Sequence,\n cast,\n overload,\n)\n\nif TYPE_CHECKING:\n from .message import Message, MessageSegment\n\nTM = TypeVar(\"TM\", bound=\"Message\")\nTF = TypeVar(\"TF\", str, \"Message\")\n\nFormatSpecFunc = Callable[[Any], str]\nFormatSpecFunc_T = TypeVar(\"FormatSpecFunc_T\", bound=FormatSpecFunc)\n\n\nclass MessageTemplate(Formatter, Generic[TF]):\n \"\"\"\u6d88\u606f\u6a21\u677f\u683c\u5f0f\u5316\u5b9e\u73b0\u7c7b\u3002\n\n \u53c2\u6570:\n template: \u6a21\u677f\n factory: \u6d88\u606f\u6784\u9020\u7c7b\u578b\uff0c\u9ed8\u8ba4\u4e3a `str`\n \"\"\"\n\n @overload\n def __init__(\n self: \"MessageTemplate[str]\", template: str, factory: Type[str] = str\n ) -> None:\n ...\n\n @overload\n def __init__(\n self: \"MessageTemplate[TM]\", template: Union[str, TM], factory: Type[TM]\n ) -> None:\n ...\n\n def __init__(self, template, factory=str) -> None:\n self.template: TF = template\n self.factory: Type[TF] = factory\n self.format_specs: Dict[str, FormatSpecFunc] = {}\n\n def add_format_spec(\n self, spec: FormatSpecFunc_T, name: Optional[str] = None\n ) -> FormatSpecFunc_T:\n name = name or spec.__name__\n if name in self.format_specs:\n raise ValueError(f\"Format spec {name} already exists!\")\n self.format_specs[name] = spec\n return spec\n\n def format(self, *args: Any, **kwargs: Any) -> TF:\n \"\"\"\u6839\u636e\u6a21\u677f\u548c\u53c2\u6570\u751f\u6210\u6d88\u606f\u5bf9\u8c61\"\"\"\n msg = self.factory()\n if isinstance(self.template, str):\n msg += self.vformat(self.template, args, kwargs)\n elif isinstance(self.template, self.factory):\n template = cast(\"Message[MessageSegment]\", self.template)\n for seg in template:\n msg += self.vformat(str(seg), args, kwargs) if seg.is_text() else seg\n else:\n raise TypeError(\"template must be a string or instance of Message!\")\n\n return msg # type:ignore\n\n def vformat(\n self, format_string: str, args: Sequence[Any], kwargs: Mapping[str, Any]\n ) -> TF:\n used_args = set()\n result, _ = self._vformat(format_string, args, kwargs, used_args, 2)\n self.check_unused_args(list(used_args), args, kwargs)\n return result\n\n def _vformat(\n self,\n format_string: str,\n args: Sequence[Any],\n kwargs: Mapping[str, Any],\n used_args: Set[Union[int, str]],\n recursion_depth: int,\n auto_arg_index: int = 0,\n ) -> Tuple[TF, int]:\n if recursion_depth < 0:\n raise ValueError(\"Max string recursion exceeded\")\n\n results: List[Any] = []\n\n for (literal_text, field_name, format_spec, conversion) in self.parse(\n format_string\n ):\n\n # output the literal text\n if literal_text:\n results.append(literal_text)\n\n # if there's a field, output it\n if field_name is not None:\n # this is some markup, find the object and do\n # the formatting\n\n # handle arg indexing when empty field_names are given.\n if field_name == \"\":\n if auto_arg_index is False:\n raise ValueError(\n \"cannot switch from manual field specification to \"\n \"automatic field numbering\"\n )\n field_name = str(auto_arg_index)\n auto_arg_index += 1\n elif field_name.isdigit():\n if auto_arg_index:\n raise ValueError(\n \"cannot switch from manual field specification to \"\n \"automatic field numbering\"\n )\n # disable auto arg incrementing, if it gets\n # used later on, then an exception will be raised\n auto_arg_index = False\n\n # given the field_name, find the object it references\n # and the argument it came from\n obj, arg_used = self.get_field(field_name, args, kwargs)\n used_args.add(arg_used)\n\n assert format_spec is not None\n\n # do any conversion on the resulting object\n obj = self.convert_field(obj, conversion) if conversion else obj\n\n # expand the format spec, if needed\n format_control, auto_arg_index = self._vformat(\n format_spec,\n args,\n kwargs,\n used_args,\n recursion_depth - 1,\n auto_arg_index,\n )\n\n # format the object and append to the result\n formatted_text = self.format_field(obj, str(format_control))\n results.append(formatted_text)\n\n return (\n self.factory(functools.reduce(self._add, results or [\"\"])),\n auto_arg_index,\n )\n\n def format_field(self, value: Any, format_spec: str) -> Any:\n formatter: Optional[FormatSpecFunc] = self.format_specs.get(format_spec)\n if formatter is None and not issubclass(self.factory, str):\n segment_class: Type[\"MessageSegment\"] = self.factory.get_segment_class()\n method = getattr(segment_class, format_spec, None)\n if inspect.ismethod(method):\n formatter = getattr(segment_class, format_spec)\n return (\n super().format_field(value, format_spec)\n if formatter is None\n else formatter(value)\n )\n\n def _add(self, a: Any, b: Any) -> Any:\n try:\n return a + b\n except TypeError:\n return a + str(b)\n", "path": "nonebot/internal/adapter/template.py"}]}
| 2,750 | 427 |
gh_patches_debug_32582
|
rasdani/github-patches
|
git_diff
|
googleapis__google-auth-library-python-1243
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Password prompt opened by `google.auth.default()` when credentials have expired
When you have a reauthentication policy set in Google Cloud and your crendentials have expired, calling `google.auth.default()` will open a `Please enter your password:` password prompt, instead if raising a `DefaultCredentialsError` as I believe would be expected.
The password prompt interrupts the flow of the program, with no way to catch or handle it actively.
If credentials are _revoked_ (instead of expired), you don't get the password prompt, and instead get an exception as expected.
For example, when creating a `bigquery.Client`, its `__init__` calls `super(Client, self).__init__()` which in turn calls `google.auth.default()`, which in turn [calls the external program `gcloud config config-helper --format json` in a `subprocess`](https://github.com/googleapis/google-auth-library-python/blob/main/google/auth/_cloud_sdk.py#L108), which finally opens the prompt.
Full traceback after starting it with `python -m pdb ...` and hitting Ctrl-C from when the prompt is open:
```
Traceback (most recent call last):
".../code_under_test.py", line 90, in _bqclient
client = bigquery.Client(
File ".../.venv/lib/python3.10/site-packages/google/cloud/bigquery/client.py", line 244, in __init__
super(Client, self).__init__(
File ".../.venv/lib/python3.10/site-packages/google/cloud/client/__init__.py", line 321, in __init__
Client.__init__(
File ".../.venv/lib/python3.10/site-packages/google/cloud/client/__init__.py", line 178, in __init__
credentials, _ = google.auth.default(scopes=scopes)
File ".../.venv/lib/python3.10/site-packages/google/auth/_default.py", line 620, in default
credentials, project_id = checker()
File ".../.venv/lib/python3.10/site-packages/google/auth/_default.py", line 614, in <lambda>
lambda: _get_gcloud_sdk_credentials(quota_project_id=quota_project_id),
File ".../.venv/lib/python3.10/site-packages/google/auth/_default.py", line 205, in _get_gcloud_sdk_credentials
project_id = _cloud_sdk.get_project_id()
File ".../.venv/lib/python3.10/site-packages/google/auth/_cloud_sdk.py", line 108, in get_project_id
output = _run_subprocess_ignore_stderr((command,) + _CLOUD_SDK_CONFIG_COMMAND)
File ".../.venv/lib/python3.10/site-packages/google/auth/_cloud_sdk.py", line 90, in _run_subprocess_ignore_stderr
output = subprocess.check_output(command, stderr=devnull)
...
```
#### Environment details
- OS: macOS 13.2.1
- Python version: 3.10.8
- pip version: 23.0.1
- `google-auth` version: 2.16.1
#### Steps to reproduce
1. Set a [Reauthentication policy](https://admin.google.com/ac/security/reauth/admin-tools) in Google Cloud.
2. Authenticate using `gcloud auth login --update-adc`.
3. Let your access token expire.
4. Now, `python -c "import google.auth; google.auth.default()"` will prompt you with `Please enter your password:`.
If instead of letting your token expire, you actively revoke it with `gcloud auth application-default revoke` and `gcloud auth revoke`, then you get the following exception raised:
```
Traceback (most recent call last):
File "<string>", line 1, in <module>
File ".../.venv/lib/python3.10/site-packages/google/auth/_default.py", line 653, in default
raise exceptions.DefaultCredentialsError(_HELP_MESSAGE)
google.auth.exceptions.DefaultCredentialsError: Could not automatically determine credentials. Please set GOOGLE_APPLICATION_CREDENTIALS or explicitly create credentials and re-run the application. For more information, please see https://cloud.google.com/docs/authentication/getting-started
```
Also, for some reason the password prompt is **not** opened if you capture `stdin` like `subprocess.check_output(["gcloud", "config", "config-helper", "--format", "json"], stdin=subprocess.PIPE)`. This has the interesting effect that if you run code in `pytest`(that captures `stdin`), you don't get the password prompt, but if you run it with `pytest -s` you do.
</issue>
<code>
[start of google/auth/_cloud_sdk.py]
1 # Copyright 2015 Google Inc.
2 #
3 # Licensed under the Apache License, Version 2.0 (the "License");
4 # you may not use this file except in compliance with the License.
5 # You may obtain a copy of the License at
6 #
7 # http://www.apache.org/licenses/LICENSE-2.0
8 #
9 # Unless required by applicable law or agreed to in writing, software
10 # distributed under the License is distributed on an "AS IS" BASIS,
11 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12 # See the License for the specific language governing permissions and
13 # limitations under the License.
14
15 """Helpers for reading the Google Cloud SDK's configuration."""
16
17 import json
18 import os
19 import subprocess
20
21 import six
22
23 from google.auth import environment_vars
24 from google.auth import exceptions
25
26
27 # The ~/.config subdirectory containing gcloud credentials.
28 _CONFIG_DIRECTORY = "gcloud"
29 # Windows systems store config at %APPDATA%\gcloud
30 _WINDOWS_CONFIG_ROOT_ENV_VAR = "APPDATA"
31 # The name of the file in the Cloud SDK config that contains default
32 # credentials.
33 _CREDENTIALS_FILENAME = "application_default_credentials.json"
34 # The name of the Cloud SDK shell script
35 _CLOUD_SDK_POSIX_COMMAND = "gcloud"
36 _CLOUD_SDK_WINDOWS_COMMAND = "gcloud.cmd"
37 # The command to get the Cloud SDK configuration
38 _CLOUD_SDK_CONFIG_COMMAND = ("config", "config-helper", "--format", "json")
39 # The command to get google user access token
40 _CLOUD_SDK_USER_ACCESS_TOKEN_COMMAND = ("auth", "print-access-token")
41 # Cloud SDK's application-default client ID
42 CLOUD_SDK_CLIENT_ID = (
43 "764086051850-6qr4p6gpi6hn506pt8ejuq83di341hur.apps.googleusercontent.com"
44 )
45
46
47 def get_config_path():
48 """Returns the absolute path the the Cloud SDK's configuration directory.
49
50 Returns:
51 str: The Cloud SDK config path.
52 """
53 # If the path is explicitly set, return that.
54 try:
55 return os.environ[environment_vars.CLOUD_SDK_CONFIG_DIR]
56 except KeyError:
57 pass
58
59 # Non-windows systems store this at ~/.config/gcloud
60 if os.name != "nt":
61 return os.path.join(os.path.expanduser("~"), ".config", _CONFIG_DIRECTORY)
62 # Windows systems store config at %APPDATA%\gcloud
63 else:
64 try:
65 return os.path.join(
66 os.environ[_WINDOWS_CONFIG_ROOT_ENV_VAR], _CONFIG_DIRECTORY
67 )
68 except KeyError:
69 # This should never happen unless someone is really
70 # messing with things, but we'll cover the case anyway.
71 drive = os.environ.get("SystemDrive", "C:")
72 return os.path.join(drive, "\\", _CONFIG_DIRECTORY)
73
74
75 def get_application_default_credentials_path():
76 """Gets the path to the application default credentials file.
77
78 The path may or may not exist.
79
80 Returns:
81 str: The full path to application default credentials.
82 """
83 config_path = get_config_path()
84 return os.path.join(config_path, _CREDENTIALS_FILENAME)
85
86
87 def _run_subprocess_ignore_stderr(command):
88 """ Return subprocess.check_output with the given command and ignores stderr."""
89 with open(os.devnull, "w") as devnull:
90 output = subprocess.check_output(command, stderr=devnull)
91 return output
92
93
94 def get_project_id():
95 """Gets the project ID from the Cloud SDK.
96
97 Returns:
98 Optional[str]: The project ID.
99 """
100 if os.name == "nt":
101 command = _CLOUD_SDK_WINDOWS_COMMAND
102 else:
103 command = _CLOUD_SDK_POSIX_COMMAND
104
105 try:
106 # Ignore the stderr coming from gcloud, so it won't be mixed into the output.
107 # https://github.com/googleapis/google-auth-library-python/issues/673
108 output = _run_subprocess_ignore_stderr((command,) + _CLOUD_SDK_CONFIG_COMMAND)
109 except (subprocess.CalledProcessError, OSError, IOError):
110 return None
111
112 try:
113 configuration = json.loads(output.decode("utf-8"))
114 except ValueError:
115 return None
116
117 try:
118 return configuration["configuration"]["properties"]["core"]["project"]
119 except KeyError:
120 return None
121
122
123 def get_auth_access_token(account=None):
124 """Load user access token with the ``gcloud auth print-access-token`` command.
125
126 Args:
127 account (Optional[str]): Account to get the access token for. If not
128 specified, the current active account will be used.
129
130 Returns:
131 str: The user access token.
132
133 Raises:
134 google.auth.exceptions.UserAccessTokenError: if failed to get access
135 token from gcloud.
136 """
137 if os.name == "nt":
138 command = _CLOUD_SDK_WINDOWS_COMMAND
139 else:
140 command = _CLOUD_SDK_POSIX_COMMAND
141
142 try:
143 if account:
144 command = (
145 (command,)
146 + _CLOUD_SDK_USER_ACCESS_TOKEN_COMMAND
147 + ("--account=" + account,)
148 )
149 else:
150 command = (command,) + _CLOUD_SDK_USER_ACCESS_TOKEN_COMMAND
151
152 access_token = subprocess.check_output(command, stderr=subprocess.STDOUT)
153 # remove the trailing "\n"
154 return access_token.decode("utf-8").strip()
155 except (subprocess.CalledProcessError, OSError, IOError) as caught_exc:
156 new_exc = exceptions.UserAccessTokenError(
157 "Failed to obtain access token", caught_exc
158 )
159 six.raise_from(new_exc, caught_exc)
160
[end of google/auth/_cloud_sdk.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/google/auth/_cloud_sdk.py b/google/auth/_cloud_sdk.py
--- a/google/auth/_cloud_sdk.py
+++ b/google/auth/_cloud_sdk.py
@@ -14,12 +14,12 @@
"""Helpers for reading the Google Cloud SDK's configuration."""
-import json
import os
import subprocess
import six
+from google.auth import _helpers
from google.auth import environment_vars
from google.auth import exceptions
@@ -35,7 +35,7 @@
_CLOUD_SDK_POSIX_COMMAND = "gcloud"
_CLOUD_SDK_WINDOWS_COMMAND = "gcloud.cmd"
# The command to get the Cloud SDK configuration
-_CLOUD_SDK_CONFIG_COMMAND = ("config", "config-helper", "--format", "json")
+_CLOUD_SDK_CONFIG_GET_PROJECT_COMMAND = ("config", "get", "project")
# The command to get google user access token
_CLOUD_SDK_USER_ACCESS_TOKEN_COMMAND = ("auth", "print-access-token")
# Cloud SDK's application-default client ID
@@ -105,18 +105,14 @@
try:
# Ignore the stderr coming from gcloud, so it won't be mixed into the output.
# https://github.com/googleapis/google-auth-library-python/issues/673
- output = _run_subprocess_ignore_stderr((command,) + _CLOUD_SDK_CONFIG_COMMAND)
- except (subprocess.CalledProcessError, OSError, IOError):
- return None
-
- try:
- configuration = json.loads(output.decode("utf-8"))
- except ValueError:
- return None
+ project = _run_subprocess_ignore_stderr(
+ (command,) + _CLOUD_SDK_CONFIG_GET_PROJECT_COMMAND
+ )
- try:
- return configuration["configuration"]["properties"]["core"]["project"]
- except KeyError:
+ # Turn bytes into a string and remove "\n"
+ project = _helpers.from_bytes(project).strip()
+ return project if project else None
+ except (subprocess.CalledProcessError, OSError, IOError):
return None
|
{"golden_diff": "diff --git a/google/auth/_cloud_sdk.py b/google/auth/_cloud_sdk.py\n--- a/google/auth/_cloud_sdk.py\n+++ b/google/auth/_cloud_sdk.py\n@@ -14,12 +14,12 @@\n \n \"\"\"Helpers for reading the Google Cloud SDK's configuration.\"\"\"\n \n-import json\n import os\n import subprocess\n \n import six\n \n+from google.auth import _helpers\n from google.auth import environment_vars\n from google.auth import exceptions\n \n@@ -35,7 +35,7 @@\n _CLOUD_SDK_POSIX_COMMAND = \"gcloud\"\n _CLOUD_SDK_WINDOWS_COMMAND = \"gcloud.cmd\"\n # The command to get the Cloud SDK configuration\n-_CLOUD_SDK_CONFIG_COMMAND = (\"config\", \"config-helper\", \"--format\", \"json\")\n+_CLOUD_SDK_CONFIG_GET_PROJECT_COMMAND = (\"config\", \"get\", \"project\")\n # The command to get google user access token\n _CLOUD_SDK_USER_ACCESS_TOKEN_COMMAND = (\"auth\", \"print-access-token\")\n # Cloud SDK's application-default client ID\n@@ -105,18 +105,14 @@\n try:\n # Ignore the stderr coming from gcloud, so it won't be mixed into the output.\n # https://github.com/googleapis/google-auth-library-python/issues/673\n- output = _run_subprocess_ignore_stderr((command,) + _CLOUD_SDK_CONFIG_COMMAND)\n- except (subprocess.CalledProcessError, OSError, IOError):\n- return None\n-\n- try:\n- configuration = json.loads(output.decode(\"utf-8\"))\n- except ValueError:\n- return None\n+ project = _run_subprocess_ignore_stderr(\n+ (command,) + _CLOUD_SDK_CONFIG_GET_PROJECT_COMMAND\n+ )\n \n- try:\n- return configuration[\"configuration\"][\"properties\"][\"core\"][\"project\"]\n- except KeyError:\n+ # Turn bytes into a string and remove \"\\n\"\n+ project = _helpers.from_bytes(project).strip()\n+ return project if project else None\n+ except (subprocess.CalledProcessError, OSError, IOError):\n return None\n", "issue": "Password prompt opened by `google.auth.default()` when credentials have expired\nWhen you have a reauthentication policy set in Google Cloud and your crendentials have expired, calling `google.auth.default()` will open a `Please enter your password:` password prompt, instead if raising a `DefaultCredentialsError` as I believe would be expected.\r\n\r\nThe password prompt interrupts the flow of the program, with no way to catch or handle it actively.\r\n\r\nIf credentials are _revoked_ (instead of expired), you don't get the password prompt, and instead get an exception as expected.\r\n\r\nFor example, when creating a `bigquery.Client`, its `__init__` calls `super(Client, self).__init__()` which in turn calls `google.auth.default()`, which in turn [calls the external program `gcloud config config-helper --format json` in a `subprocess`](https://github.com/googleapis/google-auth-library-python/blob/main/google/auth/_cloud_sdk.py#L108), which finally opens the prompt.\r\n\r\nFull traceback after starting it with `python -m pdb ...` and hitting Ctrl-C from when the prompt is open:\r\n\r\n```\r\nTraceback (most recent call last):\r\n \".../code_under_test.py\", line 90, in _bqclient\r\n client = bigquery.Client(\r\n File \".../.venv/lib/python3.10/site-packages/google/cloud/bigquery/client.py\", line 244, in __init__\r\n super(Client, self).__init__(\r\n File \".../.venv/lib/python3.10/site-packages/google/cloud/client/__init__.py\", line 321, in __init__\r\n Client.__init__(\r\n File \".../.venv/lib/python3.10/site-packages/google/cloud/client/__init__.py\", line 178, in __init__\r\n credentials, _ = google.auth.default(scopes=scopes)\r\n File \".../.venv/lib/python3.10/site-packages/google/auth/_default.py\", line 620, in default\r\n credentials, project_id = checker()\r\n File \".../.venv/lib/python3.10/site-packages/google/auth/_default.py\", line 614, in <lambda>\r\n lambda: _get_gcloud_sdk_credentials(quota_project_id=quota_project_id),\r\n File \".../.venv/lib/python3.10/site-packages/google/auth/_default.py\", line 205, in _get_gcloud_sdk_credentials\r\n project_id = _cloud_sdk.get_project_id()\r\n File \".../.venv/lib/python3.10/site-packages/google/auth/_cloud_sdk.py\", line 108, in get_project_id\r\n output = _run_subprocess_ignore_stderr((command,) + _CLOUD_SDK_CONFIG_COMMAND)\r\n File \".../.venv/lib/python3.10/site-packages/google/auth/_cloud_sdk.py\", line 90, in _run_subprocess_ignore_stderr\r\n output = subprocess.check_output(command, stderr=devnull)\r\n ...\r\n```\r\n\r\n#### Environment details\r\n\r\n - OS: macOS 13.2.1\r\n - Python version: 3.10.8\r\n - pip version: 23.0.1\r\n - `google-auth` version: 2.16.1\r\n\r\n#### Steps to reproduce\r\n\r\n 1. Set a [Reauthentication policy](https://admin.google.com/ac/security/reauth/admin-tools) in Google Cloud.\r\n 2. Authenticate using `gcloud auth login --update-adc`.\r\n 3. Let your access token expire.\r\n 4. Now, `python -c \"import google.auth; google.auth.default()\"` will prompt you with `Please enter your password:`.\r\n\r\nIf instead of letting your token expire, you actively revoke it with `gcloud auth application-default revoke` and `gcloud auth revoke`, then you get the following exception raised:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"<string>\", line 1, in <module>\r\n File \".../.venv/lib/python3.10/site-packages/google/auth/_default.py\", line 653, in default\r\n raise exceptions.DefaultCredentialsError(_HELP_MESSAGE)\r\ngoogle.auth.exceptions.DefaultCredentialsError: Could not automatically determine credentials. Please set GOOGLE_APPLICATION_CREDENTIALS or explicitly create credentials and re-run the application. For more information, please see https://cloud.google.com/docs/authentication/getting-started\r\n```\r\n\r\nAlso, for some reason the password prompt is **not** opened if you capture `stdin` like `subprocess.check_output([\"gcloud\", \"config\", \"config-helper\", \"--format\", \"json\"], stdin=subprocess.PIPE)`. This has the interesting effect that if you run code in `pytest`(that captures `stdin`), you don't get the password prompt, but if you run it with `pytest -s` you do.\r\n\n", "before_files": [{"content": "# Copyright 2015 Google Inc.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\n\"\"\"Helpers for reading the Google Cloud SDK's configuration.\"\"\"\n\nimport json\nimport os\nimport subprocess\n\nimport six\n\nfrom google.auth import environment_vars\nfrom google.auth import exceptions\n\n\n# The ~/.config subdirectory containing gcloud credentials.\n_CONFIG_DIRECTORY = \"gcloud\"\n# Windows systems store config at %APPDATA%\\gcloud\n_WINDOWS_CONFIG_ROOT_ENV_VAR = \"APPDATA\"\n# The name of the file in the Cloud SDK config that contains default\n# credentials.\n_CREDENTIALS_FILENAME = \"application_default_credentials.json\"\n# The name of the Cloud SDK shell script\n_CLOUD_SDK_POSIX_COMMAND = \"gcloud\"\n_CLOUD_SDK_WINDOWS_COMMAND = \"gcloud.cmd\"\n# The command to get the Cloud SDK configuration\n_CLOUD_SDK_CONFIG_COMMAND = (\"config\", \"config-helper\", \"--format\", \"json\")\n# The command to get google user access token\n_CLOUD_SDK_USER_ACCESS_TOKEN_COMMAND = (\"auth\", \"print-access-token\")\n# Cloud SDK's application-default client ID\nCLOUD_SDK_CLIENT_ID = (\n \"764086051850-6qr4p6gpi6hn506pt8ejuq83di341hur.apps.googleusercontent.com\"\n)\n\n\ndef get_config_path():\n \"\"\"Returns the absolute path the the Cloud SDK's configuration directory.\n\n Returns:\n str: The Cloud SDK config path.\n \"\"\"\n # If the path is explicitly set, return that.\n try:\n return os.environ[environment_vars.CLOUD_SDK_CONFIG_DIR]\n except KeyError:\n pass\n\n # Non-windows systems store this at ~/.config/gcloud\n if os.name != \"nt\":\n return os.path.join(os.path.expanduser(\"~\"), \".config\", _CONFIG_DIRECTORY)\n # Windows systems store config at %APPDATA%\\gcloud\n else:\n try:\n return os.path.join(\n os.environ[_WINDOWS_CONFIG_ROOT_ENV_VAR], _CONFIG_DIRECTORY\n )\n except KeyError:\n # This should never happen unless someone is really\n # messing with things, but we'll cover the case anyway.\n drive = os.environ.get(\"SystemDrive\", \"C:\")\n return os.path.join(drive, \"\\\\\", _CONFIG_DIRECTORY)\n\n\ndef get_application_default_credentials_path():\n \"\"\"Gets the path to the application default credentials file.\n\n The path may or may not exist.\n\n Returns:\n str: The full path to application default credentials.\n \"\"\"\n config_path = get_config_path()\n return os.path.join(config_path, _CREDENTIALS_FILENAME)\n\n\ndef _run_subprocess_ignore_stderr(command):\n \"\"\" Return subprocess.check_output with the given command and ignores stderr.\"\"\"\n with open(os.devnull, \"w\") as devnull:\n output = subprocess.check_output(command, stderr=devnull)\n return output\n\n\ndef get_project_id():\n \"\"\"Gets the project ID from the Cloud SDK.\n\n Returns:\n Optional[str]: The project ID.\n \"\"\"\n if os.name == \"nt\":\n command = _CLOUD_SDK_WINDOWS_COMMAND\n else:\n command = _CLOUD_SDK_POSIX_COMMAND\n\n try:\n # Ignore the stderr coming from gcloud, so it won't be mixed into the output.\n # https://github.com/googleapis/google-auth-library-python/issues/673\n output = _run_subprocess_ignore_stderr((command,) + _CLOUD_SDK_CONFIG_COMMAND)\n except (subprocess.CalledProcessError, OSError, IOError):\n return None\n\n try:\n configuration = json.loads(output.decode(\"utf-8\"))\n except ValueError:\n return None\n\n try:\n return configuration[\"configuration\"][\"properties\"][\"core\"][\"project\"]\n except KeyError:\n return None\n\n\ndef get_auth_access_token(account=None):\n \"\"\"Load user access token with the ``gcloud auth print-access-token`` command.\n\n Args:\n account (Optional[str]): Account to get the access token for. If not\n specified, the current active account will be used.\n\n Returns:\n str: The user access token.\n\n Raises:\n google.auth.exceptions.UserAccessTokenError: if failed to get access\n token from gcloud.\n \"\"\"\n if os.name == \"nt\":\n command = _CLOUD_SDK_WINDOWS_COMMAND\n else:\n command = _CLOUD_SDK_POSIX_COMMAND\n\n try:\n if account:\n command = (\n (command,)\n + _CLOUD_SDK_USER_ACCESS_TOKEN_COMMAND\n + (\"--account=\" + account,)\n )\n else:\n command = (command,) + _CLOUD_SDK_USER_ACCESS_TOKEN_COMMAND\n\n access_token = subprocess.check_output(command, stderr=subprocess.STDOUT)\n # remove the trailing \"\\n\"\n return access_token.decode(\"utf-8\").strip()\n except (subprocess.CalledProcessError, OSError, IOError) as caught_exc:\n new_exc = exceptions.UserAccessTokenError(\n \"Failed to obtain access token\", caught_exc\n )\n six.raise_from(new_exc, caught_exc)\n", "path": "google/auth/_cloud_sdk.py"}]}
| 3,117 | 447 |
gh_patches_debug_38421
|
rasdani/github-patches
|
git_diff
|
Cloud-CV__EvalAI-1636
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Modify database seeding method
## Deliverables
- [ ] Modify the seed command to accept an argument `-nc` which denotes the number of challenges that a developer wants to create for testing.
- [ ] Modify the seed script in such a way that it can be run multiple times (right now, if you run it once and then you try to run it again, then it throws an error).
</issue>
<code>
[start of apps/base/management/commands/seed.py]
1 from django.core.management import BaseCommand, call_command
2
3
4 class Command(BaseCommand):
5
6 help = "Seeds the database with random but sensible values."
7
8 def handle(self, *args, **options):
9 self.stdout.write(self.style.SUCCESS('Starting the database seeder. Hang on...'))
10 call_command('runscript', 'seed', '--settings', 'settings.dev')
11
[end of apps/base/management/commands/seed.py]
[start of scripts/seed.py]
1 # Command to run : python manage.py shell < scripts/seed.py
2 import os
3
4 from datetime import timedelta
5
6 from django.conf import settings
7 from django.contrib.auth.models import User
8 from django.core.files.uploadedfile import SimpleUploadedFile
9 from django.utils import timezone
10
11 from allauth.account.models import EmailAddress
12 from faker import Factory
13
14 from challenges.models import Challenge, ChallengePhase, DatasetSplit, Leaderboard, ChallengePhaseSplit
15 from hosts.models import ChallengeHostTeam, ChallengeHost
16 from participants.models import Participant, ParticipantTeam
17
18 fake = Factory.create()
19
20 NUMBER_OF_CHALLENGES = 1
21 NUMBER_OF_PHASES = 2
22 NUMBER_OF_DATASET_SPLITS = 2
23 DATASET_SPLIT_ITERATOR = 0
24
25 try:
26 xrange # Python 2
27 except NameError:
28 xrange = range # Python 3
29
30
31 def check_database():
32 if len(EmailAddress.objects.all()) > 0:
33 print("Are you sure you want to wipe the existing development database and reseed it? (Y/N)")
34 if raw_input().lower() == "y":
35 destroy_database()
36 else:
37 return False
38 else:
39 return True
40
41
42 def destroy_database():
43 print("Destroying existing database...")
44 print("Destroying Participant objects...")
45 Participant.objects.all().delete()
46 print("Destroying ParticipantTeam objects...")
47 ParticipantTeam.objects.all().delete()
48 print("Destroying ChallengePhaseSplit objects...")
49 ChallengePhaseSplit.objects.all().delete()
50 print("Destroying DatasetSplit objects...")
51 DatasetSplit.objects.all().delete()
52 print("Destroying ChallengePhase objects...")
53 ChallengePhase.objects.all().delete()
54 print("Destroying Leaderboard objects...")
55 Leaderboard.objects.all().delete()
56 print("Destroying Challenge objects...")
57 Challenge.objects.all().delete()
58 print("Destroying ChallengeHostTeam objects...")
59 ChallengeHostTeam.objects.all().delete()
60 print("Destroying ChallengeHost objects...")
61 ChallengeHost.objects.all().delete()
62 print("Destroying User objects...")
63 User.objects.all().delete()
64 print("Destroying EmailAddress objects...")
65 EmailAddress.objects.all().delete()
66 return True
67
68
69 def create_user(is_admin, username=""):
70 """
71 Creates superuser, participant user, host user and returns it.
72 """
73 if is_admin:
74 username = "admin"
75 email = "[email protected]"
76 else:
77 email = "%[email protected]" % (username)
78 user = User.objects.create_user(
79 email=email,
80 username=username,
81 password="password",
82 is_staff=is_admin,
83 is_superuser=is_admin,
84 )
85 EmailAddress.objects.create(user=user, email=email, verified=True, primary=True)
86 print("{} was created with username: {} password: password".format("Super user" if is_admin else "User", username))
87 return user
88
89
90 def create_challenge_host_team(user):
91 """
92 Creates challenge host team and returns it.
93 """
94 team_name = "{} Host Team".format(fake.city())
95 team = ChallengeHostTeam.objects.create(
96 team_name=team_name,
97 created_by=user,
98 )
99 print("Challenge Host Team created with team_name: {} created_by: {}".format(team_name, user.username))
100 ChallengeHost.objects.create(user=user, team_name=team, status=ChallengeHost.SELF, permissions=ChallengeHost.ADMIN)
101 print("Challenge Host created with user: {} team_name: {}".format(user.username, team_name))
102 return team
103
104
105 def create_challenges(number_of_challenges=3, host_team=None):
106 """
107 Creates past challenge, on-going challenge and upcoming challenge.
108 """
109 for i in xrange(number_of_challenges):
110 if (i % 3 == 0):
111 create_challenge("{} Challenge".format(fake.first_name()),
112 timezone.now() - timedelta(days=100),
113 timezone.now() + timedelta(days=500),
114 host_team
115 )
116 elif (i % 3 == 1):
117 create_challenge("{} Challenge" % (fake.first_name()),
118 timezone.now() - timedelta(days=500),
119 timezone.now() - timedelta(days=100),
120 host_team
121 )
122 elif (i % 3 == 2):
123 create_challenge("{} Challenge" % (fake.first_name()),
124 timezone.now() + timedelta(days=100),
125 timezone.now() + timedelta(days=500),
126 host_team
127 )
128
129
130 def create_challenge(title, start_date, end_date, host_team):
131 """
132 Creates a challenge.
133 """
134 evaluation_script = open(os.path.join(settings.BASE_DIR, 'examples', 'example1', 'string_matching.zip'), 'rb')
135 Challenge.objects.create(
136 title=title,
137 short_description=fake.paragraph(),
138 description=fake.paragraph(),
139 terms_and_conditions=fake.paragraph(),
140 submission_guidelines=fake.paragraph(),
141 evaluation_details=fake.paragraph(),
142 evaluation_script=SimpleUploadedFile(evaluation_script.name, evaluation_script.read()),
143 approved_by_admin=True,
144 creator=host_team,
145 published=True,
146 enable_forum=True,
147 anonymous_leaderboard=False,
148 start_date=start_date,
149 end_date=end_date,
150 )
151 print("Challenge created with title: {} creator: {} start_date: {} end_date: {}".format(title,
152 host_team.team_name,
153 start_date, end_date))
154
155
156 def create_challenge_phases(challenge, number_of_phases=1):
157 """
158 Creates challenge phases for the created challenges and returns it.
159 """
160 challenge_phases = []
161 for i in range(number_of_phases):
162 name = "{} Phase".format(fake.first_name())
163 with open(os.path.join(settings.BASE_DIR, 'examples', 'example1', 'test_annotation.txt'), 'rb') as data_file:
164 data = data_file.read()
165 data = data or None
166 challenge_phase = ChallengePhase.objects.create(
167 name=name,
168 description=fake.paragraph(),
169 leaderboard_public=True,
170 is_public=True,
171 start_date=challenge.start_date,
172 end_date=challenge.end_date,
173 challenge=challenge,
174 test_annotation=SimpleUploadedFile(fake.file_name(extension="txt"), data, content_type="text/plain"),
175 codename="{}{}".format("phase", i + 1),
176 )
177 challenge_phases.append(challenge_phase)
178 print("Challenge Phase created with name: {} challenge: {}".format(name, challenge.title))
179 return challenge_phases
180
181
182 def create_leaderboard():
183 """
184 Creates Leaderboard schema and returns it.
185 """
186 schema = {
187 'labels': ['score', ],
188 'default_order_by': 'score',
189 }
190 leaderboard = Leaderboard.objects.create(
191 schema=schema
192 )
193 print("Leaderboard created")
194 return leaderboard
195
196
197 def create_dataset_splits(number_of_splits):
198 """
199 Creates dataset splits and returns it.
200 """
201 dataset_splits = []
202 for i in range(number_of_splits):
203 global DATASET_SPLIT_ITERATOR
204 name = "Split {}".format(DATASET_SPLIT_ITERATOR + 1)
205 codename = "{}{}".format('split', DATASET_SPLIT_ITERATOR + 1)
206 dataset_split = DatasetSplit.objects.create(
207 name=name,
208 codename=codename,
209 )
210 dataset_splits.append(dataset_split)
211 DATASET_SPLIT_ITERATOR += 1
212 print("Dataset Split created with name: {} codename: {}".format(name, codename))
213 return dataset_splits
214
215
216 def create_challenge_phase_splits(challenge_phase, leaderboard, dataset_split):
217 """
218 Creates a challenge phase split.
219 """
220 ChallengePhaseSplit.objects.create(
221 challenge_phase=challenge_phase,
222 leaderboard=leaderboard,
223 dataset_split=dataset_split,
224 visibility=ChallengePhaseSplit.PUBLIC
225 )
226 print("Challenge Phase Split created with challenge_phase: {} dataset_split: {}".format(challenge_phase.name,
227 dataset_split.name))
228
229
230 def create_participant_team(user):
231 """
232 Creates participant team and returns it.
233 """
234 team_name = "{} Participant Team".format(fake.city())
235 team = ParticipantTeam.objects.create(
236 team_name=team_name,
237 created_by=user,
238 )
239 print("Participant Team created with team_name: {} created_by: {}".format(team_name, user.username))
240 Participant.objects.create(user=user, team=team, status="Self")
241 print("Participant created with user: {} team_name: {}".format(user.username, team_name))
242 return team
243
244
245 def run():
246 status = check_database()
247 if status is False:
248 print("Seeding aborted.")
249 return 0
250 print("Seeding...")
251 # Create superuser
252 create_user(is_admin=True)
253 # Create host user
254 host_user = create_user(is_admin=False, username="host")
255 # Create challenge host team with challenge host
256 challenge_host_team = create_challenge_host_team(user=host_user)
257 # Create challenge
258 create_challenges(number_of_challenges=NUMBER_OF_CHALLENGES, host_team=challenge_host_team)
259
260 # Fetch all the created challenges
261 challenges = Challenge.objects.all()
262 for challenge in challenges:
263 # Create a leaderboard object for each challenge
264 leaderboard = create_leaderboard()
265 # Create Phases for a challenge
266 challenge_phases = create_challenge_phases(challenge, number_of_phases=NUMBER_OF_PHASES)
267 # Create Dataset Split for each Challenge
268 dataset_splits = create_dataset_splits(number_of_splits=NUMBER_OF_DATASET_SPLITS)
269 # Create Challenge Phase Split for each Phase and Dataset Split
270 for challenge_phase in challenge_phases:
271 for dataset_split in dataset_splits:
272 create_challenge_phase_splits(challenge_phase, leaderboard, dataset_split)
273 participant_user = create_user(is_admin=False, username="participant")
274 create_participant_team(user=participant_user)
275 print('Database successfully seeded.')
276
[end of scripts/seed.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/apps/base/management/commands/seed.py b/apps/base/management/commands/seed.py
--- a/apps/base/management/commands/seed.py
+++ b/apps/base/management/commands/seed.py
@@ -5,6 +5,10 @@
help = "Seeds the database with random but sensible values."
+ def add_arguments(self, parser):
+ parser.add_argument('-nc', nargs='?', default=1, type=int, help='Number of challenges.')
+
def handle(self, *args, **options):
+ self.nc = options['nc']
self.stdout.write(self.style.SUCCESS('Starting the database seeder. Hang on...'))
- call_command('runscript', 'seed', '--settings', 'settings.dev')
+ call_command('runscript', 'seed', '--script-args', self.nc)
diff --git a/scripts/seed.py b/scripts/seed.py
--- a/scripts/seed.py
+++ b/scripts/seed.py
@@ -31,7 +31,7 @@
def check_database():
if len(EmailAddress.objects.all()) > 0:
print("Are you sure you want to wipe the existing development database and reseed it? (Y/N)")
- if raw_input().lower() == "y":
+ if settings.TEST or raw_input().lower() == "y":
destroy_database()
else:
return False
@@ -102,7 +102,7 @@
return team
-def create_challenges(number_of_challenges=3, host_team=None):
+def create_challenges(number_of_challenges, host_team=None):
"""
Creates past challenge, on-going challenge and upcoming challenge.
"""
@@ -114,13 +114,13 @@
host_team
)
elif (i % 3 == 1):
- create_challenge("{} Challenge" % (fake.first_name()),
+ create_challenge("{} Challenge".format(fake.first_name()),
timezone.now() - timedelta(days=500),
timezone.now() - timedelta(days=100),
host_team
)
elif (i % 3 == 2):
- create_challenge("{} Challenge" % (fake.first_name()),
+ create_challenge("{} Challenge".format(fake.first_name()),
timezone.now() + timedelta(days=100),
timezone.now() + timedelta(days=500),
host_team
@@ -242,7 +242,8 @@
return team
-def run():
+def run(*args):
+ NUMBER_OF_CHALLENGES = int(args[0])
status = check_database()
if status is False:
print("Seeding aborted.")
|
{"golden_diff": "diff --git a/apps/base/management/commands/seed.py b/apps/base/management/commands/seed.py\n--- a/apps/base/management/commands/seed.py\n+++ b/apps/base/management/commands/seed.py\n@@ -5,6 +5,10 @@\n \n help = \"Seeds the database with random but sensible values.\"\n \n+ def add_arguments(self, parser):\n+ parser.add_argument('-nc', nargs='?', default=1, type=int, help='Number of challenges.')\n+\n def handle(self, *args, **options):\n+ self.nc = options['nc']\n self.stdout.write(self.style.SUCCESS('Starting the database seeder. Hang on...'))\n- call_command('runscript', 'seed', '--settings', 'settings.dev')\n+ call_command('runscript', 'seed', '--script-args', self.nc)\ndiff --git a/scripts/seed.py b/scripts/seed.py\n--- a/scripts/seed.py\n+++ b/scripts/seed.py\n@@ -31,7 +31,7 @@\n def check_database():\n if len(EmailAddress.objects.all()) > 0:\n print(\"Are you sure you want to wipe the existing development database and reseed it? (Y/N)\")\n- if raw_input().lower() == \"y\":\n+ if settings.TEST or raw_input().lower() == \"y\":\n destroy_database()\n else:\n return False\n@@ -102,7 +102,7 @@\n return team\n \n \n-def create_challenges(number_of_challenges=3, host_team=None):\n+def create_challenges(number_of_challenges, host_team=None):\n \"\"\"\n Creates past challenge, on-going challenge and upcoming challenge.\n \"\"\"\n@@ -114,13 +114,13 @@\n host_team\n )\n elif (i % 3 == 1):\n- create_challenge(\"{} Challenge\" % (fake.first_name()),\n+ create_challenge(\"{} Challenge\".format(fake.first_name()),\n timezone.now() - timedelta(days=500),\n timezone.now() - timedelta(days=100),\n host_team\n )\n elif (i % 3 == 2):\n- create_challenge(\"{} Challenge\" % (fake.first_name()),\n+ create_challenge(\"{} Challenge\".format(fake.first_name()),\n timezone.now() + timedelta(days=100),\n timezone.now() + timedelta(days=500),\n host_team\n@@ -242,7 +242,8 @@\n return team\n \n \n-def run():\n+def run(*args):\n+ NUMBER_OF_CHALLENGES = int(args[0])\n status = check_database()\n if status is False:\n print(\"Seeding aborted.\")\n", "issue": "Modify database seeding method\n## Deliverables\r\n\r\n- [ ] Modify the seed command to accept an argument `-nc` which denotes the number of challenges that a developer wants to create for testing.\r\n- [ ] Modify the seed script in such a way that it can be run multiple times (right now, if you run it once and then you try to run it again, then it throws an error). \n", "before_files": [{"content": "from django.core.management import BaseCommand, call_command\n\n\nclass Command(BaseCommand):\n\n help = \"Seeds the database with random but sensible values.\"\n\n def handle(self, *args, **options):\n self.stdout.write(self.style.SUCCESS('Starting the database seeder. Hang on...'))\n call_command('runscript', 'seed', '--settings', 'settings.dev')\n", "path": "apps/base/management/commands/seed.py"}, {"content": "# Command to run : python manage.py shell < scripts/seed.py\nimport os\n\nfrom datetime import timedelta\n\nfrom django.conf import settings\nfrom django.contrib.auth.models import User\nfrom django.core.files.uploadedfile import SimpleUploadedFile\nfrom django.utils import timezone\n\nfrom allauth.account.models import EmailAddress\nfrom faker import Factory\n\nfrom challenges.models import Challenge, ChallengePhase, DatasetSplit, Leaderboard, ChallengePhaseSplit\nfrom hosts.models import ChallengeHostTeam, ChallengeHost\nfrom participants.models import Participant, ParticipantTeam\n\nfake = Factory.create()\n\nNUMBER_OF_CHALLENGES = 1\nNUMBER_OF_PHASES = 2\nNUMBER_OF_DATASET_SPLITS = 2\nDATASET_SPLIT_ITERATOR = 0\n\ntry:\n xrange # Python 2\nexcept NameError:\n xrange = range # Python 3\n\n\ndef check_database():\n if len(EmailAddress.objects.all()) > 0:\n print(\"Are you sure you want to wipe the existing development database and reseed it? (Y/N)\")\n if raw_input().lower() == \"y\":\n destroy_database()\n else:\n return False\n else:\n return True\n\n\ndef destroy_database():\n print(\"Destroying existing database...\")\n print(\"Destroying Participant objects...\")\n Participant.objects.all().delete()\n print(\"Destroying ParticipantTeam objects...\")\n ParticipantTeam.objects.all().delete()\n print(\"Destroying ChallengePhaseSplit objects...\")\n ChallengePhaseSplit.objects.all().delete()\n print(\"Destroying DatasetSplit objects...\")\n DatasetSplit.objects.all().delete()\n print(\"Destroying ChallengePhase objects...\")\n ChallengePhase.objects.all().delete()\n print(\"Destroying Leaderboard objects...\")\n Leaderboard.objects.all().delete()\n print(\"Destroying Challenge objects...\")\n Challenge.objects.all().delete()\n print(\"Destroying ChallengeHostTeam objects...\")\n ChallengeHostTeam.objects.all().delete()\n print(\"Destroying ChallengeHost objects...\")\n ChallengeHost.objects.all().delete()\n print(\"Destroying User objects...\")\n User.objects.all().delete()\n print(\"Destroying EmailAddress objects...\")\n EmailAddress.objects.all().delete()\n return True\n\n\ndef create_user(is_admin, username=\"\"):\n \"\"\"\n Creates superuser, participant user, host user and returns it.\n \"\"\"\n if is_admin:\n username = \"admin\"\n email = \"[email protected]\"\n else:\n email = \"%[email protected]\" % (username)\n user = User.objects.create_user(\n email=email,\n username=username,\n password=\"password\",\n is_staff=is_admin,\n is_superuser=is_admin,\n )\n EmailAddress.objects.create(user=user, email=email, verified=True, primary=True)\n print(\"{} was created with username: {} password: password\".format(\"Super user\" if is_admin else \"User\", username))\n return user\n\n\ndef create_challenge_host_team(user):\n \"\"\"\n Creates challenge host team and returns it.\n \"\"\"\n team_name = \"{} Host Team\".format(fake.city())\n team = ChallengeHostTeam.objects.create(\n team_name=team_name,\n created_by=user,\n )\n print(\"Challenge Host Team created with team_name: {} created_by: {}\".format(team_name, user.username))\n ChallengeHost.objects.create(user=user, team_name=team, status=ChallengeHost.SELF, permissions=ChallengeHost.ADMIN)\n print(\"Challenge Host created with user: {} team_name: {}\".format(user.username, team_name))\n return team\n\n\ndef create_challenges(number_of_challenges=3, host_team=None):\n \"\"\"\n Creates past challenge, on-going challenge and upcoming challenge.\n \"\"\"\n for i in xrange(number_of_challenges):\n if (i % 3 == 0):\n create_challenge(\"{} Challenge\".format(fake.first_name()),\n timezone.now() - timedelta(days=100),\n timezone.now() + timedelta(days=500),\n host_team\n )\n elif (i % 3 == 1):\n create_challenge(\"{} Challenge\" % (fake.first_name()),\n timezone.now() - timedelta(days=500),\n timezone.now() - timedelta(days=100),\n host_team\n )\n elif (i % 3 == 2):\n create_challenge(\"{} Challenge\" % (fake.first_name()),\n timezone.now() + timedelta(days=100),\n timezone.now() + timedelta(days=500),\n host_team\n )\n\n\ndef create_challenge(title, start_date, end_date, host_team):\n \"\"\"\n Creates a challenge.\n \"\"\"\n evaluation_script = open(os.path.join(settings.BASE_DIR, 'examples', 'example1', 'string_matching.zip'), 'rb')\n Challenge.objects.create(\n title=title,\n short_description=fake.paragraph(),\n description=fake.paragraph(),\n terms_and_conditions=fake.paragraph(),\n submission_guidelines=fake.paragraph(),\n evaluation_details=fake.paragraph(),\n evaluation_script=SimpleUploadedFile(evaluation_script.name, evaluation_script.read()),\n approved_by_admin=True,\n creator=host_team,\n published=True,\n enable_forum=True,\n anonymous_leaderboard=False,\n start_date=start_date,\n end_date=end_date,\n )\n print(\"Challenge created with title: {} creator: {} start_date: {} end_date: {}\".format(title,\n host_team.team_name,\n start_date, end_date))\n\n\ndef create_challenge_phases(challenge, number_of_phases=1):\n \"\"\"\n Creates challenge phases for the created challenges and returns it.\n \"\"\"\n challenge_phases = []\n for i in range(number_of_phases):\n name = \"{} Phase\".format(fake.first_name())\n with open(os.path.join(settings.BASE_DIR, 'examples', 'example1', 'test_annotation.txt'), 'rb') as data_file:\n data = data_file.read()\n data = data or None\n challenge_phase = ChallengePhase.objects.create(\n name=name,\n description=fake.paragraph(),\n leaderboard_public=True,\n is_public=True,\n start_date=challenge.start_date,\n end_date=challenge.end_date,\n challenge=challenge,\n test_annotation=SimpleUploadedFile(fake.file_name(extension=\"txt\"), data, content_type=\"text/plain\"),\n codename=\"{}{}\".format(\"phase\", i + 1),\n )\n challenge_phases.append(challenge_phase)\n print(\"Challenge Phase created with name: {} challenge: {}\".format(name, challenge.title))\n return challenge_phases\n\n\ndef create_leaderboard():\n \"\"\"\n Creates Leaderboard schema and returns it.\n \"\"\"\n schema = {\n 'labels': ['score', ],\n 'default_order_by': 'score',\n }\n leaderboard = Leaderboard.objects.create(\n schema=schema\n )\n print(\"Leaderboard created\")\n return leaderboard\n\n\ndef create_dataset_splits(number_of_splits):\n \"\"\"\n Creates dataset splits and returns it.\n \"\"\"\n dataset_splits = []\n for i in range(number_of_splits):\n global DATASET_SPLIT_ITERATOR\n name = \"Split {}\".format(DATASET_SPLIT_ITERATOR + 1)\n codename = \"{}{}\".format('split', DATASET_SPLIT_ITERATOR + 1)\n dataset_split = DatasetSplit.objects.create(\n name=name,\n codename=codename,\n )\n dataset_splits.append(dataset_split)\n DATASET_SPLIT_ITERATOR += 1\n print(\"Dataset Split created with name: {} codename: {}\".format(name, codename))\n return dataset_splits\n\n\ndef create_challenge_phase_splits(challenge_phase, leaderboard, dataset_split):\n \"\"\"\n Creates a challenge phase split.\n \"\"\"\n ChallengePhaseSplit.objects.create(\n challenge_phase=challenge_phase,\n leaderboard=leaderboard,\n dataset_split=dataset_split,\n visibility=ChallengePhaseSplit.PUBLIC\n )\n print(\"Challenge Phase Split created with challenge_phase: {} dataset_split: {}\".format(challenge_phase.name,\n dataset_split.name))\n\n\ndef create_participant_team(user):\n \"\"\"\n Creates participant team and returns it.\n \"\"\"\n team_name = \"{} Participant Team\".format(fake.city())\n team = ParticipantTeam.objects.create(\n team_name=team_name,\n created_by=user,\n )\n print(\"Participant Team created with team_name: {} created_by: {}\".format(team_name, user.username))\n Participant.objects.create(user=user, team=team, status=\"Self\")\n print(\"Participant created with user: {} team_name: {}\".format(user.username, team_name))\n return team\n\n\ndef run():\n status = check_database()\n if status is False:\n print(\"Seeding aborted.\")\n return 0\n print(\"Seeding...\")\n # Create superuser\n create_user(is_admin=True)\n # Create host user\n host_user = create_user(is_admin=False, username=\"host\")\n # Create challenge host team with challenge host\n challenge_host_team = create_challenge_host_team(user=host_user)\n # Create challenge\n create_challenges(number_of_challenges=NUMBER_OF_CHALLENGES, host_team=challenge_host_team)\n\n # Fetch all the created challenges\n challenges = Challenge.objects.all()\n for challenge in challenges:\n # Create a leaderboard object for each challenge\n leaderboard = create_leaderboard()\n # Create Phases for a challenge\n challenge_phases = create_challenge_phases(challenge, number_of_phases=NUMBER_OF_PHASES)\n # Create Dataset Split for each Challenge\n dataset_splits = create_dataset_splits(number_of_splits=NUMBER_OF_DATASET_SPLITS)\n # Create Challenge Phase Split for each Phase and Dataset Split\n for challenge_phase in challenge_phases:\n for dataset_split in dataset_splits:\n create_challenge_phase_splits(challenge_phase, leaderboard, dataset_split)\n participant_user = create_user(is_admin=False, username=\"participant\")\n create_participant_team(user=participant_user)\n print('Database successfully seeded.')\n", "path": "scripts/seed.py"}]}
| 3,524 | 577 |
gh_patches_debug_15645
|
rasdani/github-patches
|
git_diff
|
netbox-community__netbox-7928
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Don't fetch LDAP user and groups on all API request when FIND_GROUP_PERMS is disabled
### NetBox version
V3.0.9
### Feature type
Change to existing functionality
### Proposed functionality
Currently when using the LDAP backend for authentication, the AD is queried on every API request, regardless of other settings and regardless if the user is local or has been created by the LDAP backend. Additionally the LDAP cache built into django-auth-ldap does not function when using populate_user.
As the user is not actually authenticated against the AD when using the API (the token is used), I propose that the local user and it's group assignments are used when FIND_GROUP_PERMISSIONS is disabled.
I have a change ready for pull request if the issue is accepted.
For more info, please see the discussion I created: https://github.com/netbox-community/netbox/discussions/7708
This issue would also partly fix #6926 - it will not fix the caching, but the user who reported the issue is not using FIND_GROUP_PERMISSIONS.
### Use case
The end goal is vastly improved API performance when using the LDAP backend in most cases.
The above changes will result in the following changes for users:
**Not using the LDAP backend:**
No changes
**FIND_GROUP_PERMS = True:**
No changes
**MIRROR_GROUPS = True and FIND_GROUP_PERMS = True:**
No changes
**MIRROR_GROUPS = True and FIND_GROUP_PERMS = False:**
Local user and group assignments will be used when calling the API and the user and groups are never reloaded from the LDAP server during API calls. This means that LDAP users utilizing the API will have to login to the web ui to update group memberships. The change also allows one to use locally created users to call the API with querying the LDAP server.
**MIRROR_GROUPS = False and FIND_GROUP_PERMS = False:**
The user performing the API request has to be locally assigned groups or have local user object permissions.
### Database changes
No database changes
### External dependencies
_No response_
</issue>
<code>
[start of netbox/netbox/api/authentication.py]
1 from django.conf import settings
2 from rest_framework import authentication, exceptions
3 from rest_framework.permissions import BasePermission, DjangoObjectPermissions, SAFE_METHODS
4
5 from users.models import Token
6
7
8 class TokenAuthentication(authentication.TokenAuthentication):
9 """
10 A custom authentication scheme which enforces Token expiration times.
11 """
12 model = Token
13
14 def authenticate_credentials(self, key):
15 model = self.get_model()
16 try:
17 token = model.objects.prefetch_related('user').get(key=key)
18 except model.DoesNotExist:
19 raise exceptions.AuthenticationFailed("Invalid token")
20
21 # Enforce the Token's expiration time, if one has been set.
22 if token.is_expired:
23 raise exceptions.AuthenticationFailed("Token expired")
24
25 if not token.user.is_active:
26 raise exceptions.AuthenticationFailed("User inactive")
27
28 # When LDAP authentication is active try to load user data from LDAP directory
29 if settings.REMOTE_AUTH_BACKEND == 'netbox.authentication.LDAPBackend':
30 from netbox.authentication import LDAPBackend
31 ldap_backend = LDAPBackend()
32 user = ldap_backend.populate_user(token.user.username)
33 # If the user is found in the LDAP directory use it, if not fallback to the local user
34 if user:
35 return user, token
36
37 return token.user, token
38
39
40 class TokenPermissions(DjangoObjectPermissions):
41 """
42 Custom permissions handler which extends the built-in DjangoModelPermissions to validate a Token's write ability
43 for unsafe requests (POST/PUT/PATCH/DELETE).
44 """
45 # Override the stock perm_map to enforce view permissions
46 perms_map = {
47 'GET': ['%(app_label)s.view_%(model_name)s'],
48 'OPTIONS': [],
49 'HEAD': ['%(app_label)s.view_%(model_name)s'],
50 'POST': ['%(app_label)s.add_%(model_name)s'],
51 'PUT': ['%(app_label)s.change_%(model_name)s'],
52 'PATCH': ['%(app_label)s.change_%(model_name)s'],
53 'DELETE': ['%(app_label)s.delete_%(model_name)s'],
54 }
55
56 def __init__(self):
57
58 # LOGIN_REQUIRED determines whether read-only access is provided to anonymous users.
59 self.authenticated_users_only = settings.LOGIN_REQUIRED
60
61 super().__init__()
62
63 def _verify_write_permission(self, request):
64
65 # If token authentication is in use, verify that the token allows write operations (for unsafe methods).
66 if request.method in SAFE_METHODS or request.auth.write_enabled:
67 return True
68
69 def has_permission(self, request, view):
70
71 # Enforce Token write ability
72 if isinstance(request.auth, Token) and not self._verify_write_permission(request):
73 return False
74
75 return super().has_permission(request, view)
76
77 def has_object_permission(self, request, view, obj):
78
79 # Enforce Token write ability
80 if isinstance(request.auth, Token) and not self._verify_write_permission(request):
81 return False
82
83 return super().has_object_permission(request, view, obj)
84
85
86 class IsAuthenticatedOrLoginNotRequired(BasePermission):
87 """
88 Returns True if the user is authenticated or LOGIN_REQUIRED is False.
89 """
90 def has_permission(self, request, view):
91 if not settings.LOGIN_REQUIRED:
92 return True
93 return request.user.is_authenticated
94
[end of netbox/netbox/api/authentication.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/netbox/netbox/api/authentication.py b/netbox/netbox/api/authentication.py
--- a/netbox/netbox/api/authentication.py
+++ b/netbox/netbox/api/authentication.py
@@ -29,10 +29,13 @@
if settings.REMOTE_AUTH_BACKEND == 'netbox.authentication.LDAPBackend':
from netbox.authentication import LDAPBackend
ldap_backend = LDAPBackend()
- user = ldap_backend.populate_user(token.user.username)
- # If the user is found in the LDAP directory use it, if not fallback to the local user
- if user:
- return user, token
+
+ # Load from LDAP if FIND_GROUP_PERMS is active
+ if ldap_backend.settings.FIND_GROUP_PERMS:
+ user = ldap_backend.populate_user(token.user.username)
+ # If the user is found in the LDAP directory use it, if not fallback to the local user
+ if user:
+ return user, token
return token.user, token
|
{"golden_diff": "diff --git a/netbox/netbox/api/authentication.py b/netbox/netbox/api/authentication.py\n--- a/netbox/netbox/api/authentication.py\n+++ b/netbox/netbox/api/authentication.py\n@@ -29,10 +29,13 @@\n if settings.REMOTE_AUTH_BACKEND == 'netbox.authentication.LDAPBackend':\n from netbox.authentication import LDAPBackend\n ldap_backend = LDAPBackend()\n- user = ldap_backend.populate_user(token.user.username)\n- # If the user is found in the LDAP directory use it, if not fallback to the local user\n- if user:\n- return user, token\n+\n+ # Load from LDAP if FIND_GROUP_PERMS is active\n+ if ldap_backend.settings.FIND_GROUP_PERMS:\n+ user = ldap_backend.populate_user(token.user.username)\n+ # If the user is found in the LDAP directory use it, if not fallback to the local user\n+ if user:\n+ return user, token\n \n return token.user, token\n", "issue": "Don't fetch LDAP user and groups on all API request when FIND_GROUP_PERMS is disabled\n### NetBox version\n\nV3.0.9\n\n### Feature type\n\nChange to existing functionality\n\n### Proposed functionality\n\nCurrently when using the LDAP backend for authentication, the AD is queried on every API request, regardless of other settings and regardless if the user is local or has been created by the LDAP backend. Additionally the LDAP cache built into django-auth-ldap does not function when using populate_user.\r\n\r\nAs the user is not actually authenticated against the AD when using the API (the token is used), I propose that the local user and it's group assignments are used when FIND_GROUP_PERMISSIONS is disabled.\r\n\r\nI have a change ready for pull request if the issue is accepted.\r\n\r\nFor more info, please see the discussion I created: https://github.com/netbox-community/netbox/discussions/7708\r\n\r\nThis issue would also partly fix #6926 - it will not fix the caching, but the user who reported the issue is not using FIND_GROUP_PERMISSIONS.\n\n### Use case\n\nThe end goal is vastly improved API performance when using the LDAP backend in most cases.\r\n\r\nThe above changes will result in the following changes for users:\r\n\r\n**Not using the LDAP backend:**\r\n\r\nNo changes\r\n\r\n**FIND_GROUP_PERMS = True:**\r\n\r\nNo changes\r\n\r\n**MIRROR_GROUPS = True and FIND_GROUP_PERMS = True:**\r\n\r\nNo changes\r\n\r\n**MIRROR_GROUPS = True and FIND_GROUP_PERMS = False:**\r\n\r\nLocal user and group assignments will be used when calling the API and the user and groups are never reloaded from the LDAP server during API calls. This means that LDAP users utilizing the API will have to login to the web ui to update group memberships. The change also allows one to use locally created users to call the API with querying the LDAP server.\r\n\r\n**MIRROR_GROUPS = False and FIND_GROUP_PERMS = False:**\r\n\r\nThe user performing the API request has to be locally assigned groups or have local user object permissions.\n\n### Database changes\n\nNo database changes\n\n### External dependencies\n\n_No response_\n", "before_files": [{"content": "from django.conf import settings\nfrom rest_framework import authentication, exceptions\nfrom rest_framework.permissions import BasePermission, DjangoObjectPermissions, SAFE_METHODS\n\nfrom users.models import Token\n\n\nclass TokenAuthentication(authentication.TokenAuthentication):\n \"\"\"\n A custom authentication scheme which enforces Token expiration times.\n \"\"\"\n model = Token\n\n def authenticate_credentials(self, key):\n model = self.get_model()\n try:\n token = model.objects.prefetch_related('user').get(key=key)\n except model.DoesNotExist:\n raise exceptions.AuthenticationFailed(\"Invalid token\")\n\n # Enforce the Token's expiration time, if one has been set.\n if token.is_expired:\n raise exceptions.AuthenticationFailed(\"Token expired\")\n\n if not token.user.is_active:\n raise exceptions.AuthenticationFailed(\"User inactive\")\n\n # When LDAP authentication is active try to load user data from LDAP directory\n if settings.REMOTE_AUTH_BACKEND == 'netbox.authentication.LDAPBackend':\n from netbox.authentication import LDAPBackend\n ldap_backend = LDAPBackend()\n user = ldap_backend.populate_user(token.user.username)\n # If the user is found in the LDAP directory use it, if not fallback to the local user\n if user:\n return user, token\n\n return token.user, token\n\n\nclass TokenPermissions(DjangoObjectPermissions):\n \"\"\"\n Custom permissions handler which extends the built-in DjangoModelPermissions to validate a Token's write ability\n for unsafe requests (POST/PUT/PATCH/DELETE).\n \"\"\"\n # Override the stock perm_map to enforce view permissions\n perms_map = {\n 'GET': ['%(app_label)s.view_%(model_name)s'],\n 'OPTIONS': [],\n 'HEAD': ['%(app_label)s.view_%(model_name)s'],\n 'POST': ['%(app_label)s.add_%(model_name)s'],\n 'PUT': ['%(app_label)s.change_%(model_name)s'],\n 'PATCH': ['%(app_label)s.change_%(model_name)s'],\n 'DELETE': ['%(app_label)s.delete_%(model_name)s'],\n }\n\n def __init__(self):\n\n # LOGIN_REQUIRED determines whether read-only access is provided to anonymous users.\n self.authenticated_users_only = settings.LOGIN_REQUIRED\n\n super().__init__()\n\n def _verify_write_permission(self, request):\n\n # If token authentication is in use, verify that the token allows write operations (for unsafe methods).\n if request.method in SAFE_METHODS or request.auth.write_enabled:\n return True\n\n def has_permission(self, request, view):\n\n # Enforce Token write ability\n if isinstance(request.auth, Token) and not self._verify_write_permission(request):\n return False\n\n return super().has_permission(request, view)\n\n def has_object_permission(self, request, view, obj):\n\n # Enforce Token write ability\n if isinstance(request.auth, Token) and not self._verify_write_permission(request):\n return False\n\n return super().has_object_permission(request, view, obj)\n\n\nclass IsAuthenticatedOrLoginNotRequired(BasePermission):\n \"\"\"\n Returns True if the user is authenticated or LOGIN_REQUIRED is False.\n \"\"\"\n def has_permission(self, request, view):\n if not settings.LOGIN_REQUIRED:\n return True\n return request.user.is_authenticated\n", "path": "netbox/netbox/api/authentication.py"}]}
| 1,839 | 215 |
gh_patches_debug_11790
|
rasdani/github-patches
|
git_diff
|
deepset-ai__haystack-4567
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Duplicate tests for WhisperTranscriber
`test_whisper_api_transcribe` and `test_whisper_api_transcribe_with_params` in `test/nodes/test_whisper.py` have the exact same code.
We should remove `test_whisper_api_transcribe_with_params`.
**Alternatives considered**
Change `transcribe_test_helper(w)` to `transcribe_test_helper(w, language="en")` in `test_whisper_api_transcribe_with_params`. This is probably how it was intended. However, we're only testing the API then and don't check the output taking into account that parameter.
</issue>
<code>
[start of haystack/nodes/audio/whisper_transcriber.py]
1 import json
2
3 from typing import List, Optional, Dict, Any, Union, BinaryIO, Literal
4
5 import requests
6 import torch
7 from requests import PreparedRequest
8
9 from haystack import MultiLabel, Document
10 from haystack.errors import OpenAIError, OpenAIRateLimitError
11 from haystack.nodes.base import BaseComponent
12 from haystack.utils.import_utils import is_whisper_available
13
14 WhisperModel = Literal["tiny", "small", "medium", "large", "large-v2"]
15
16
17 class WhisperTranscriber(BaseComponent):
18 """
19 Transcribes audio files using OpenAI's Whisper. This class supports two underlying implementations:
20
21 - API (default): Uses the OpenAI API and requires an API key. See blog
22 [post](https://beta.openai.com/docs/api-reference/whisper for more details.) for more details.
23 - Local (requires installation of whisper): Uses the local installation
24 of [whisper](https://github.com/openai/whisper).
25
26 If you are using local installation of whisper, install whisper following the instructions available on
27 the Whisper [github repo](https://github.com/openai/whisper) and omit the api_key parameter.
28
29 If you are using the API implementation, you need to provide an api_key. You can get one by signing up
30 for an OpenAI account [here](https://beta.openai.com/).
31
32 For the supported audio formats, languages and other parameters, see the Whisper API
33 [documentation](https://platform.openai.com/docs/guides/speech-to-text) and the official Whisper
34 [github repo](https://github.com/openai/whisper).
35 """
36
37 # If it's not a decision component, there is only one outgoing edge
38 outgoing_edges = 1
39
40 def __init__(
41 self,
42 api_key: Optional[str] = None,
43 model_name_or_path: WhisperModel = "medium",
44 device: Optional[Union[str, torch.device]] = None,
45 ) -> None:
46 """
47 Creates a WhisperTranscriber instance.
48
49 :param api_key: OpenAI API key. If None, local installation of whisper is used.
50 :param model_name_or_path: Name of the model to use. If using local installation of whisper, this
51 value has to be one of the following: "tiny", "small", "medium", "large", "large-v2". If using
52 the API, this value has to be "whisper-1" (default).
53 :param device: Device to use for inference. This parameter is only used if you are using local
54 installation of whisper. If None, the device is automatically selected.
55 """
56 super().__init__()
57 self.api_key = api_key
58
59 self.use_local_whisper = is_whisper_available() and self.api_key is None
60
61 if self.use_local_whisper:
62 import whisper
63
64 self._model = whisper.load_model(model_name_or_path, device=device)
65 else:
66 if api_key is None:
67 raise ValueError(
68 "Please provide a valid api_key for OpenAI API. Alternatively, "
69 "install OpenAI whisper (see https://github.com/openai/whisper for more details)."
70 )
71
72 def transcribe(
73 self,
74 audio_file: Union[str, BinaryIO],
75 language: Optional[str] = None,
76 return_segments: bool = False,
77 translate: bool = False,
78 **kwargs,
79 ) -> Dict[str, Any]:
80 """
81 Transcribe audio file.
82
83 :param audio_file: Path to audio file or a binary file-like object.
84 :param language: Language of the audio file. If None, the language is automatically detected.
85 :param return_segments: If True, returns the transcription for each segment of the audio file.
86 :param translate: If True, translates the transcription to English.
87
88 """
89 transcript: Dict[str, Any] = {}
90
91 new_kwargs = {k: v for k, v in kwargs.items() if v is not None}
92 if language is not None:
93 new_kwargs["language"] = language
94
95 if self.use_local_whisper:
96 new_kwargs["return_segments"] = return_segments
97 transcript = self._invoke_local(audio_file, translate, **new_kwargs)
98 elif self.api_key:
99 transcript = self._invoke_api(audio_file, translate, **new_kwargs)
100 return transcript
101
102 def _invoke_api(
103 self, audio_file: Union[str, BinaryIO], translate: Optional[bool] = False, **kwargs
104 ) -> Dict[str, Any]:
105 if isinstance(audio_file, str):
106 with open(audio_file, "rb") as f:
107 return self._invoke_api(f, translate, **kwargs)
108 else:
109 headers = {"Authorization": f"Bearer {self.api_key}"}
110 request = PreparedRequest()
111 url: str = (
112 "https://api.openai.com/v1/audio/transcriptions"
113 if not translate
114 else "https://api.openai.com/v1/audio/translations"
115 )
116
117 request.prepare(
118 method="POST",
119 url=url,
120 headers=headers,
121 data={"model": "whisper-1", **kwargs},
122 files=[("file", (audio_file.name, audio_file, "application/octet-stream"))],
123 )
124 response = requests.post(url, data=request.body, headers=request.headers, timeout=600)
125
126 if response.status_code != 200:
127 openai_error: OpenAIError
128 if response.status_code == 429:
129 openai_error = OpenAIRateLimitError(f"API rate limit exceeded: {response.text}")
130 else:
131 openai_error = OpenAIError(
132 f"OpenAI returned an error.\n"
133 f"Status code: {response.status_code}\n"
134 f"Response body: {response.text}",
135 status_code=response.status_code,
136 )
137 raise openai_error
138
139 return json.loads(response.content)
140
141 def _invoke_local(
142 self, audio_file: Union[str, BinaryIO], translate: Optional[bool] = False, **kwargs
143 ) -> Dict[str, Any]:
144 if isinstance(audio_file, str):
145 with open(audio_file, "rb") as f:
146 return self._invoke_local(f, translate, **kwargs)
147 else:
148 return_segments = kwargs.pop("return_segments", None)
149 kwargs["task"] = "translate" if translate else "transcribe"
150 transcription = self._model.transcribe(audio_file.name, **kwargs)
151 if not return_segments:
152 transcription.pop("segments", None)
153
154 return transcription
155
156 def run(
157 self,
158 query: Optional[str] = None,
159 file_paths: Optional[List[str]] = None,
160 labels: Optional[MultiLabel] = None,
161 documents: Optional[List[Document]] = None,
162 meta: Optional[dict] = None,
163 ): # type: ignore
164 """
165 Transcribe audio files.
166
167 :param query: Ignored
168 :param file_paths: List of paths to audio files.
169 :param labels: Ignored
170 :param documents: Ignored
171 :param meta: Ignored
172 """
173 transcribed_documents: List[Document] = []
174 if file_paths:
175 for file_path in file_paths:
176 transcription = self.transcribe(file_path)
177 d = Document.from_dict(transcription, field_map={"text": "content"})
178 transcribed_documents.append(d)
179
180 output = {"documents": transcribed_documents}
181 return output, "output_1"
182
183 def run_batch(
184 self,
185 queries: Optional[Union[str, List[str]]] = None,
186 file_paths: Optional[List[str]] = None,
187 labels: Optional[Union[MultiLabel, List[MultiLabel]]] = None,
188 documents: Optional[Union[List[Document], List[List[Document]]]] = None,
189 meta: Optional[Union[Dict[str, Any], List[Dict[str, Any]]]] = None,
190 params: Optional[dict] = None,
191 debug: Optional[bool] = None,
192 ): # type: ignore
193 """
194 Transcribe audio files.
195
196 :param queries: Ignored
197 :param file_paths: List of paths to audio files.
198 :param labels: Ignored
199 :param documents: Ignored
200 :param meta: Ignored
201 :param params: Ignored
202 :param debug: Ignored
203 """
204 if file_paths and isinstance(file_paths[0], list):
205 all_files = []
206 for files_list in file_paths:
207 all_files += files_list
208 return self.run(file_paths=all_files)
209 return self.run(file_paths=file_paths)
210
[end of haystack/nodes/audio/whisper_transcriber.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/haystack/nodes/audio/whisper_transcriber.py b/haystack/nodes/audio/whisper_transcriber.py
--- a/haystack/nodes/audio/whisper_transcriber.py
+++ b/haystack/nodes/audio/whisper_transcriber.py
@@ -82,7 +82,8 @@
:param audio_file: Path to audio file or a binary file-like object.
:param language: Language of the audio file. If None, the language is automatically detected.
- :param return_segments: If True, returns the transcription for each segment of the audio file.
+ :param return_segments: If True, returns the transcription for each segment of the audio file. Supported with
+ local installation of whisper only.
:param translate: If True, translates the transcription to English.
"""
|
{"golden_diff": "diff --git a/haystack/nodes/audio/whisper_transcriber.py b/haystack/nodes/audio/whisper_transcriber.py\n--- a/haystack/nodes/audio/whisper_transcriber.py\n+++ b/haystack/nodes/audio/whisper_transcriber.py\n@@ -82,7 +82,8 @@\n \n :param audio_file: Path to audio file or a binary file-like object.\n :param language: Language of the audio file. If None, the language is automatically detected.\n- :param return_segments: If True, returns the transcription for each segment of the audio file.\n+ :param return_segments: If True, returns the transcription for each segment of the audio file. Supported with\n+ local installation of whisper only.\n :param translate: If True, translates the transcription to English.\n \n \"\"\"\n", "issue": "Duplicate tests for WhisperTranscriber\n`test_whisper_api_transcribe` and `test_whisper_api_transcribe_with_params` in `test/nodes/test_whisper.py` have the exact same code.\r\nWe should remove `test_whisper_api_transcribe_with_params`.\r\n\r\n**Alternatives considered**\r\nChange `transcribe_test_helper(w)` to `transcribe_test_helper(w, language=\"en\")` in `test_whisper_api_transcribe_with_params`. This is probably how it was intended. However, we're only testing the API then and don't check the output taking into account that parameter.\r\n\n", "before_files": [{"content": "import json\n\nfrom typing import List, Optional, Dict, Any, Union, BinaryIO, Literal\n\nimport requests\nimport torch\nfrom requests import PreparedRequest\n\nfrom haystack import MultiLabel, Document\nfrom haystack.errors import OpenAIError, OpenAIRateLimitError\nfrom haystack.nodes.base import BaseComponent\nfrom haystack.utils.import_utils import is_whisper_available\n\nWhisperModel = Literal[\"tiny\", \"small\", \"medium\", \"large\", \"large-v2\"]\n\n\nclass WhisperTranscriber(BaseComponent):\n \"\"\"\n Transcribes audio files using OpenAI's Whisper. This class supports two underlying implementations:\n\n - API (default): Uses the OpenAI API and requires an API key. See blog\n [post](https://beta.openai.com/docs/api-reference/whisper for more details.) for more details.\n - Local (requires installation of whisper): Uses the local installation\n of [whisper](https://github.com/openai/whisper).\n\n If you are using local installation of whisper, install whisper following the instructions available on\n the Whisper [github repo](https://github.com/openai/whisper) and omit the api_key parameter.\n\n If you are using the API implementation, you need to provide an api_key. You can get one by signing up\n for an OpenAI account [here](https://beta.openai.com/).\n\n For the supported audio formats, languages and other parameters, see the Whisper API\n [documentation](https://platform.openai.com/docs/guides/speech-to-text) and the official Whisper\n [github repo](https://github.com/openai/whisper).\n \"\"\"\n\n # If it's not a decision component, there is only one outgoing edge\n outgoing_edges = 1\n\n def __init__(\n self,\n api_key: Optional[str] = None,\n model_name_or_path: WhisperModel = \"medium\",\n device: Optional[Union[str, torch.device]] = None,\n ) -> None:\n \"\"\"\n Creates a WhisperTranscriber instance.\n\n :param api_key: OpenAI API key. If None, local installation of whisper is used.\n :param model_name_or_path: Name of the model to use. If using local installation of whisper, this\n value has to be one of the following: \"tiny\", \"small\", \"medium\", \"large\", \"large-v2\". If using\n the API, this value has to be \"whisper-1\" (default).\n :param device: Device to use for inference. This parameter is only used if you are using local\n installation of whisper. If None, the device is automatically selected.\n \"\"\"\n super().__init__()\n self.api_key = api_key\n\n self.use_local_whisper = is_whisper_available() and self.api_key is None\n\n if self.use_local_whisper:\n import whisper\n\n self._model = whisper.load_model(model_name_or_path, device=device)\n else:\n if api_key is None:\n raise ValueError(\n \"Please provide a valid api_key for OpenAI API. Alternatively, \"\n \"install OpenAI whisper (see https://github.com/openai/whisper for more details).\"\n )\n\n def transcribe(\n self,\n audio_file: Union[str, BinaryIO],\n language: Optional[str] = None,\n return_segments: bool = False,\n translate: bool = False,\n **kwargs,\n ) -> Dict[str, Any]:\n \"\"\"\n Transcribe audio file.\n\n :param audio_file: Path to audio file or a binary file-like object.\n :param language: Language of the audio file. If None, the language is automatically detected.\n :param return_segments: If True, returns the transcription for each segment of the audio file.\n :param translate: If True, translates the transcription to English.\n\n \"\"\"\n transcript: Dict[str, Any] = {}\n\n new_kwargs = {k: v for k, v in kwargs.items() if v is not None}\n if language is not None:\n new_kwargs[\"language\"] = language\n\n if self.use_local_whisper:\n new_kwargs[\"return_segments\"] = return_segments\n transcript = self._invoke_local(audio_file, translate, **new_kwargs)\n elif self.api_key:\n transcript = self._invoke_api(audio_file, translate, **new_kwargs)\n return transcript\n\n def _invoke_api(\n self, audio_file: Union[str, BinaryIO], translate: Optional[bool] = False, **kwargs\n ) -> Dict[str, Any]:\n if isinstance(audio_file, str):\n with open(audio_file, \"rb\") as f:\n return self._invoke_api(f, translate, **kwargs)\n else:\n headers = {\"Authorization\": f\"Bearer {self.api_key}\"}\n request = PreparedRequest()\n url: str = (\n \"https://api.openai.com/v1/audio/transcriptions\"\n if not translate\n else \"https://api.openai.com/v1/audio/translations\"\n )\n\n request.prepare(\n method=\"POST\",\n url=url,\n headers=headers,\n data={\"model\": \"whisper-1\", **kwargs},\n files=[(\"file\", (audio_file.name, audio_file, \"application/octet-stream\"))],\n )\n response = requests.post(url, data=request.body, headers=request.headers, timeout=600)\n\n if response.status_code != 200:\n openai_error: OpenAIError\n if response.status_code == 429:\n openai_error = OpenAIRateLimitError(f\"API rate limit exceeded: {response.text}\")\n else:\n openai_error = OpenAIError(\n f\"OpenAI returned an error.\\n\"\n f\"Status code: {response.status_code}\\n\"\n f\"Response body: {response.text}\",\n status_code=response.status_code,\n )\n raise openai_error\n\n return json.loads(response.content)\n\n def _invoke_local(\n self, audio_file: Union[str, BinaryIO], translate: Optional[bool] = False, **kwargs\n ) -> Dict[str, Any]:\n if isinstance(audio_file, str):\n with open(audio_file, \"rb\") as f:\n return self._invoke_local(f, translate, **kwargs)\n else:\n return_segments = kwargs.pop(\"return_segments\", None)\n kwargs[\"task\"] = \"translate\" if translate else \"transcribe\"\n transcription = self._model.transcribe(audio_file.name, **kwargs)\n if not return_segments:\n transcription.pop(\"segments\", None)\n\n return transcription\n\n def run(\n self,\n query: Optional[str] = None,\n file_paths: Optional[List[str]] = None,\n labels: Optional[MultiLabel] = None,\n documents: Optional[List[Document]] = None,\n meta: Optional[dict] = None,\n ): # type: ignore\n \"\"\"\n Transcribe audio files.\n\n :param query: Ignored\n :param file_paths: List of paths to audio files.\n :param labels: Ignored\n :param documents: Ignored\n :param meta: Ignored\n \"\"\"\n transcribed_documents: List[Document] = []\n if file_paths:\n for file_path in file_paths:\n transcription = self.transcribe(file_path)\n d = Document.from_dict(transcription, field_map={\"text\": \"content\"})\n transcribed_documents.append(d)\n\n output = {\"documents\": transcribed_documents}\n return output, \"output_1\"\n\n def run_batch(\n self,\n queries: Optional[Union[str, List[str]]] = None,\n file_paths: Optional[List[str]] = None,\n labels: Optional[Union[MultiLabel, List[MultiLabel]]] = None,\n documents: Optional[Union[List[Document], List[List[Document]]]] = None,\n meta: Optional[Union[Dict[str, Any], List[Dict[str, Any]]]] = None,\n params: Optional[dict] = None,\n debug: Optional[bool] = None,\n ): # type: ignore\n \"\"\"\n Transcribe audio files.\n\n :param queries: Ignored\n :param file_paths: List of paths to audio files.\n :param labels: Ignored\n :param documents: Ignored\n :param meta: Ignored\n :param params: Ignored\n :param debug: Ignored\n \"\"\"\n if file_paths and isinstance(file_paths[0], list):\n all_files = []\n for files_list in file_paths:\n all_files += files_list\n return self.run(file_paths=all_files)\n return self.run(file_paths=file_paths)\n", "path": "haystack/nodes/audio/whisper_transcriber.py"}]}
| 3,047 | 184 |
gh_patches_debug_31857
|
rasdani/github-patches
|
git_diff
|
projectmesa__mesa-301
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Documentation is not reflecting latest changes wrt width-height argument order in Grid()
As many people start with reading mesa on readthedocs, the documentation should be inline with the code changes wrt width-height argument order in Grid functions.This is not yet reflected.
</issue>
<code>
[start of mesa/visualization/modules/CanvasGridVisualization.py]
1 # -*- coding: utf-8 -*-
2 """
3 Modular Canvas Rendering
4 ========================
5
6 Module for visualizing model objects in grid cells.
7
8 """
9 from collections import defaultdict
10 from mesa.visualization.ModularVisualization import VisualizationElement
11
12
13 class CanvasGrid(VisualizationElement):
14 """ A CanvasGrid object uses a user-provided portrayal method to generate a
15 portrayal for each object. A portrayal is a JSON-ready dictionary which
16 tells the relevant JavaScript code (GridDraw.js) where to draw what shape.
17
18 The render method returns a dictionary, keyed on layers, with values as
19 lists of portrayals to draw. Portrayals themselves are generated by the
20 user-provided portrayal_method, which accepts an object as an input and
21 produces a portrayal of it.
22
23 A portrayal as a dictionary with the following structure:
24 "x", "y": Coordinates for the cell in which the object is placed.
25 "Shape": Can be either "circle" or "rect"
26 For Circles:
27 "r": The radius, defined as a fraction of cell size. r=1 will
28 fill the entire cell.
29 For rectangles:
30 "w", "h": The width and height of the rectangle, which are in
31 fractions of cell width and height.
32 "Color": The color to draw the shape in; needs to be a valid HTML
33 color, e.g."Red" or "#AA08F8"
34 "Filled": either "true" or "false", and determines whether the shape is
35 filled or not.
36 "Layer": Layer number of 0 or above; higher-numbered layers are drawn
37 above lower-numbered layers.
38 "text": The text to be inscribed inside the Shape. Normally useful for
39 showing the unique_id of the agent.
40 "text_color": The color to draw the inscribed text. Should be given in
41 conjunction of "text" property.
42
43
44 Attributes:
45 portrayal_method: Function which generates portrayals from objects, as
46 described above.
47 grid_height, grid_width: Size of the grid to visualize, in cells.
48 canvas_height, canvas_width: Size, in pixels, of the grid visualization
49 to draw on the client.
50 template: "canvas_module.html" stores the module's HTML template.
51
52 """
53 package_includes = ["GridDraw.js", "CanvasModule.js"]
54 portrayal_method = None # Portrayal function
55 canvas_width = 500
56 canvas_height = 500
57
58 def __init__(self, portrayal_method, grid_width, grid_height,
59 canvas_width=500, canvas_height=500):
60 """ Instantiate a new CanvasGrid.
61
62 Args:
63 portrayal_method: function to convert each object on the grid to
64 a portrayal, as described above.
65 grid_width, grid_height: Size of the grid, in cells.
66 canvas_height, canvas_width: Size of the canvas to draw in the
67 client, in pixels. (default: 500x500)
68
69 """
70 self.portrayal_method = portrayal_method
71 self.grid_width = grid_width
72 self.grid_height = grid_height
73 self.canvas_width = canvas_width
74 self.canvas_height = canvas_height
75
76 new_element = ("new CanvasModule({}, {}, {}, {})"
77 .format(self.canvas_width, self.canvas_height,
78 self.grid_width, self.grid_height))
79
80 self.js_code = "elements.push(" + new_element + ");"
81
82 def render(self, model):
83 grid_state = defaultdict(list)
84 for x in range(model.grid.width):
85 for y in range(model.grid.height):
86 cell_objects = model.grid.get_cell_list_contents([(x, y)])
87 for obj in cell_objects:
88 portrayal = self.portrayal_method(obj)
89 if portrayal:
90 portrayal["x"] = x
91 portrayal["y"] = y
92 grid_state[portrayal["Layer"]].append(portrayal)
93
94 return grid_state
95
[end of mesa/visualization/modules/CanvasGridVisualization.py]
[start of examples/Basic/basic/server.py]
1 import random
2
3 from mesa.visualization.modules import CanvasGrid
4 from mesa.visualization.ModularVisualization import ModularServer
5
6 from basic.model import Walker, ShapesModel
7
8
9 def agent_draw(agent):
10 portrayal = None
11 if agent is None:
12 # Actually this if part is unnecessary, but still keeping it for
13 # aesthetics
14 pass
15 elif isinstance(agent, Walker):
16 print("Uid: {0}, Heading: {1}".format(agent.unique_id, agent.heading))
17 portrayal = {"Shape": "arrowHead",
18 "Filled": "true",
19 "Layer": 2,
20 "Color": "green",
21 "Filled": "true",
22 "heading0": agent.heading[0],
23 "heading1": agent.heading[1],
24 "text": agent.unique_id,
25 "text_color": "white",
26 "scale": 0.8,
27 }
28 return portrayal
29
30
31 def launch_basic():
32 width = 15
33 height = 10
34 num_agents = 2
35 pixel_ratio = 50
36 grid = CanvasGrid(agent_draw, width, height,
37 width * pixel_ratio, height * pixel_ratio)
38 server = ModularServer(ShapesModel, [grid], "Basic Example",
39 num_agents, width, height)
40 server.max_steps = 0
41 server.port = 8888
42 server.launch()
43
44 if __name__ == "__main__":
45 random.seed(3)
46 launch_basic()
47
[end of examples/Basic/basic/server.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/examples/Basic/basic/server.py b/examples/Basic/basic/server.py
--- a/examples/Basic/basic/server.py
+++ b/examples/Basic/basic/server.py
@@ -19,8 +19,8 @@
"Layer": 2,
"Color": "green",
"Filled": "true",
- "heading0": agent.heading[0],
- "heading1": agent.heading[1],
+ "heading_x": agent.heading[0],
+ "heading_y": agent.heading[1],
"text": agent.unique_id,
"text_color": "white",
"scale": 0.8,
diff --git a/mesa/visualization/modules/CanvasGridVisualization.py b/mesa/visualization/modules/CanvasGridVisualization.py
--- a/mesa/visualization/modules/CanvasGridVisualization.py
+++ b/mesa/visualization/modules/CanvasGridVisualization.py
@@ -22,13 +22,17 @@
A portrayal as a dictionary with the following structure:
"x", "y": Coordinates for the cell in which the object is placed.
- "Shape": Can be either "circle" or "rect"
+ "Shape": Can be either "circle", "rect" or "arrowHead"
For Circles:
"r": The radius, defined as a fraction of cell size. r=1 will
fill the entire cell.
- For rectangles:
+ For Rectangles:
"w", "h": The width and height of the rectangle, which are in
fractions of cell width and height.
+ For arrowHead:
+ "scale": Proportion scaling as a fraction of cell size.
+ "heading_x": represents x direction unit vector.
+ "heading_y": represents y direction unit vector.
"Color": The color to draw the shape in; needs to be a valid HTML
color, e.g."Red" or "#AA08F8"
"Filled": either "true" or "false", and determines whether the shape is
|
{"golden_diff": "diff --git a/examples/Basic/basic/server.py b/examples/Basic/basic/server.py\n--- a/examples/Basic/basic/server.py\n+++ b/examples/Basic/basic/server.py\n@@ -19,8 +19,8 @@\n \"Layer\": 2,\n \"Color\": \"green\",\n \"Filled\": \"true\",\n- \"heading0\": agent.heading[0],\n- \"heading1\": agent.heading[1],\n+ \"heading_x\": agent.heading[0],\n+ \"heading_y\": agent.heading[1],\n \"text\": agent.unique_id,\n \"text_color\": \"white\",\n \"scale\": 0.8,\ndiff --git a/mesa/visualization/modules/CanvasGridVisualization.py b/mesa/visualization/modules/CanvasGridVisualization.py\n--- a/mesa/visualization/modules/CanvasGridVisualization.py\n+++ b/mesa/visualization/modules/CanvasGridVisualization.py\n@@ -22,13 +22,17 @@\n \n A portrayal as a dictionary with the following structure:\n \"x\", \"y\": Coordinates for the cell in which the object is placed.\n- \"Shape\": Can be either \"circle\" or \"rect\"\n+ \"Shape\": Can be either \"circle\", \"rect\" or \"arrowHead\"\n For Circles:\n \"r\": The radius, defined as a fraction of cell size. r=1 will\n fill the entire cell.\n- For rectangles:\n+ For Rectangles:\n \"w\", \"h\": The width and height of the rectangle, which are in\n fractions of cell width and height.\n+ For arrowHead:\n+ \"scale\": Proportion scaling as a fraction of cell size.\n+ \"heading_x\": represents x direction unit vector.\n+ \"heading_y\": represents y direction unit vector.\n \"Color\": The color to draw the shape in; needs to be a valid HTML\n color, e.g.\"Red\" or \"#AA08F8\"\n \"Filled\": either \"true\" or \"false\", and determines whether the shape is\n", "issue": "Documentation is not reflecting latest changes wrt width-height argument order in Grid()\nAs many people start with reading mesa on readthedocs, the documentation should be inline with the code changes wrt width-height argument order in Grid functions.This is not yet reflected.\n\n", "before_files": [{"content": "# -*- coding: utf-8 -*-\n\"\"\"\nModular Canvas Rendering\n========================\n\nModule for visualizing model objects in grid cells.\n\n\"\"\"\nfrom collections import defaultdict\nfrom mesa.visualization.ModularVisualization import VisualizationElement\n\n\nclass CanvasGrid(VisualizationElement):\n \"\"\" A CanvasGrid object uses a user-provided portrayal method to generate a\n portrayal for each object. A portrayal is a JSON-ready dictionary which\n tells the relevant JavaScript code (GridDraw.js) where to draw what shape.\n\n The render method returns a dictionary, keyed on layers, with values as\n lists of portrayals to draw. Portrayals themselves are generated by the\n user-provided portrayal_method, which accepts an object as an input and\n produces a portrayal of it.\n\n A portrayal as a dictionary with the following structure:\n \"x\", \"y\": Coordinates for the cell in which the object is placed.\n \"Shape\": Can be either \"circle\" or \"rect\"\n For Circles:\n \"r\": The radius, defined as a fraction of cell size. r=1 will\n fill the entire cell.\n For rectangles:\n \"w\", \"h\": The width and height of the rectangle, which are in\n fractions of cell width and height.\n \"Color\": The color to draw the shape in; needs to be a valid HTML\n color, e.g.\"Red\" or \"#AA08F8\"\n \"Filled\": either \"true\" or \"false\", and determines whether the shape is\n filled or not.\n \"Layer\": Layer number of 0 or above; higher-numbered layers are drawn\n above lower-numbered layers.\n \"text\": The text to be inscribed inside the Shape. Normally useful for\n showing the unique_id of the agent.\n \"text_color\": The color to draw the inscribed text. Should be given in\n conjunction of \"text\" property.\n\n\n Attributes:\n portrayal_method: Function which generates portrayals from objects, as\n described above.\n grid_height, grid_width: Size of the grid to visualize, in cells.\n canvas_height, canvas_width: Size, in pixels, of the grid visualization\n to draw on the client.\n template: \"canvas_module.html\" stores the module's HTML template.\n\n \"\"\"\n package_includes = [\"GridDraw.js\", \"CanvasModule.js\"]\n portrayal_method = None # Portrayal function\n canvas_width = 500\n canvas_height = 500\n\n def __init__(self, portrayal_method, grid_width, grid_height,\n canvas_width=500, canvas_height=500):\n \"\"\" Instantiate a new CanvasGrid.\n\n Args:\n portrayal_method: function to convert each object on the grid to\n a portrayal, as described above.\n grid_width, grid_height: Size of the grid, in cells.\n canvas_height, canvas_width: Size of the canvas to draw in the\n client, in pixels. (default: 500x500)\n\n \"\"\"\n self.portrayal_method = portrayal_method\n self.grid_width = grid_width\n self.grid_height = grid_height\n self.canvas_width = canvas_width\n self.canvas_height = canvas_height\n\n new_element = (\"new CanvasModule({}, {}, {}, {})\"\n .format(self.canvas_width, self.canvas_height,\n self.grid_width, self.grid_height))\n\n self.js_code = \"elements.push(\" + new_element + \");\"\n\n def render(self, model):\n grid_state = defaultdict(list)\n for x in range(model.grid.width):\n for y in range(model.grid.height):\n cell_objects = model.grid.get_cell_list_contents([(x, y)])\n for obj in cell_objects:\n portrayal = self.portrayal_method(obj)\n if portrayal:\n portrayal[\"x\"] = x\n portrayal[\"y\"] = y\n grid_state[portrayal[\"Layer\"]].append(portrayal)\n\n return grid_state\n", "path": "mesa/visualization/modules/CanvasGridVisualization.py"}, {"content": "import random\n\nfrom mesa.visualization.modules import CanvasGrid\nfrom mesa.visualization.ModularVisualization import ModularServer\n\nfrom basic.model import Walker, ShapesModel\n\n\ndef agent_draw(agent):\n portrayal = None\n if agent is None:\n # Actually this if part is unnecessary, but still keeping it for\n # aesthetics\n pass\n elif isinstance(agent, Walker):\n print(\"Uid: {0}, Heading: {1}\".format(agent.unique_id, agent.heading))\n portrayal = {\"Shape\": \"arrowHead\",\n \"Filled\": \"true\",\n \"Layer\": 2,\n \"Color\": \"green\",\n \"Filled\": \"true\",\n \"heading0\": agent.heading[0],\n \"heading1\": agent.heading[1],\n \"text\": agent.unique_id,\n \"text_color\": \"white\",\n \"scale\": 0.8,\n }\n return portrayal\n\n\ndef launch_basic():\n width = 15\n height = 10\n num_agents = 2\n pixel_ratio = 50\n grid = CanvasGrid(agent_draw, width, height,\n width * pixel_ratio, height * pixel_ratio)\n server = ModularServer(ShapesModel, [grid], \"Basic Example\",\n num_agents, width, height)\n server.max_steps = 0\n server.port = 8888\n server.launch()\n\nif __name__ == \"__main__\":\n random.seed(3)\n launch_basic()\n", "path": "examples/Basic/basic/server.py"}]}
| 2,038 | 436 |
gh_patches_debug_44700
|
rasdani/github-patches
|
git_diff
|
feast-dev__feast-3702
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
performance issues of getting online features related to parsing protobuf data
**Is your feature request related to a problem? Please describe.**
The profiler of one application I am working on shows the `from_proto` method of the `FeatureView` class takes up 80% of the execution time. The result of cProfile is shown below.
```
ncalls tottime percall cumtime percall filename:lineno(function)
1 0.003 0.003 78.931 78.931 profiler.py:53(main)
100 0.018 0.000 78.849 0.788 get_features.py:122(get_xxxx_features)
1100/400 0.020 0.000 77.077 0.193 usage.py:274(wrapper)
444200/9800 1.713 0.000 72.206 0.007 __init__.py:1030(wrapper)
200 0.078 0.000 68.913 0.345 feature_store.py:1527(get_online_features)
200 0.087 0.000 68.835 0.344 feature_store.py:1590(_get_online_features)
3500 5.634 0.002 62.442 0.018 feature_view.py:369(from_proto)
200 0.005 0.000 59.362 0.297 feature_store.py:2149(_get_feature_views_to_use)
200 0.002 0.000 58.501 0.293 feature_store.py:281(_list_feature_views)
200 0.001 0.000 58.499 0.292 registry.py:523(list_feature_views)
200 0.016 0.000 58.495 0.292 proto_registry_utils.py:150(list_feature_views)
```
There are 3 feature views accessed by `get_xxxx_features`, however 100 calls of `get_xxxx_features` lead to 3500 calls of `from_proto`. There are 17 feature views in the feature store of this application and 3 of them are used by `get_xxxx_features`.
Environment: `continuumio/miniconda3:4.11.0` (linux/amd64) base image, `python==3.9.7`, `feast=0.31.1`, `protobuf==4.23.2`
**Describe the solution you'd like**
Instead of caching the protobuf blob of `FeatureView`, cache the `FeatureView` python object in memory.
**Describe alternatives you've considered**
modify the `get_online_features` [method](https://github.com/feast-dev/feast/blob/v0.31.1/sdk/python/feast/feature_store.py#L1528) to
```python
def get_online_features(
self,
features: Union[List[str], List[FeatureView], FeatureService],
entity_rows: List[Dict[str, Any]],
full_feature_names: bool = False,
):
```
so that the application developer has the option to cache the `FeatureView` objects and use them to get features directly (by passing the `get_feature_views_to_use` [step](https://github.com/feast-dev/feast/blob/v0.31.1/sdk/python/feast/feature_store.py#L1606-L1613) )
</issue>
<code>
[start of sdk/python/feast/infra/registry/proto_registry_utils.py]
1 import uuid
2 from typing import List, Optional
3
4 from feast import usage
5 from feast.data_source import DataSource
6 from feast.entity import Entity
7 from feast.errors import (
8 DataSourceObjectNotFoundException,
9 EntityNotFoundException,
10 FeatureServiceNotFoundException,
11 FeatureViewNotFoundException,
12 SavedDatasetNotFound,
13 ValidationReferenceNotFound,
14 )
15 from feast.feature_service import FeatureService
16 from feast.feature_view import FeatureView
17 from feast.on_demand_feature_view import OnDemandFeatureView
18 from feast.project_metadata import ProjectMetadata
19 from feast.protos.feast.core.Registry_pb2 import ProjectMetadata as ProjectMetadataProto
20 from feast.protos.feast.core.Registry_pb2 import Registry as RegistryProto
21 from feast.request_feature_view import RequestFeatureView
22 from feast.saved_dataset import SavedDataset, ValidationReference
23 from feast.stream_feature_view import StreamFeatureView
24
25
26 def init_project_metadata(cached_registry_proto: RegistryProto, project: str):
27 new_project_uuid = f"{uuid.uuid4()}"
28 usage.set_current_project_uuid(new_project_uuid)
29 cached_registry_proto.project_metadata.append(
30 ProjectMetadata(project_name=project, project_uuid=new_project_uuid).to_proto()
31 )
32
33
34 def get_project_metadata(
35 registry_proto: Optional[RegistryProto], project: str
36 ) -> Optional[ProjectMetadataProto]:
37 if not registry_proto:
38 return None
39 for pm in registry_proto.project_metadata:
40 if pm.project == project:
41 return pm
42 return None
43
44
45 def get_feature_service(
46 registry_proto: RegistryProto, name: str, project: str
47 ) -> FeatureService:
48 for feature_service_proto in registry_proto.feature_services:
49 if (
50 feature_service_proto.spec.project == project
51 and feature_service_proto.spec.name == name
52 ):
53 return FeatureService.from_proto(feature_service_proto)
54 raise FeatureServiceNotFoundException(name, project=project)
55
56
57 def get_feature_view(
58 registry_proto: RegistryProto, name: str, project: str
59 ) -> FeatureView:
60 for feature_view_proto in registry_proto.feature_views:
61 if (
62 feature_view_proto.spec.name == name
63 and feature_view_proto.spec.project == project
64 ):
65 return FeatureView.from_proto(feature_view_proto)
66 raise FeatureViewNotFoundException(name, project)
67
68
69 def get_stream_feature_view(
70 registry_proto: RegistryProto, name: str, project: str
71 ) -> StreamFeatureView:
72 for feature_view_proto in registry_proto.stream_feature_views:
73 if (
74 feature_view_proto.spec.name == name
75 and feature_view_proto.spec.project == project
76 ):
77 return StreamFeatureView.from_proto(feature_view_proto)
78 raise FeatureViewNotFoundException(name, project)
79
80
81 def get_request_feature_view(registry_proto: RegistryProto, name: str, project: str):
82 for feature_view_proto in registry_proto.feature_views:
83 if (
84 feature_view_proto.spec.name == name
85 and feature_view_proto.spec.project == project
86 ):
87 return RequestFeatureView.from_proto(feature_view_proto)
88 raise FeatureViewNotFoundException(name, project)
89
90
91 def get_on_demand_feature_view(
92 registry_proto: RegistryProto, name: str, project: str
93 ) -> OnDemandFeatureView:
94 for on_demand_feature_view in registry_proto.on_demand_feature_views:
95 if (
96 on_demand_feature_view.spec.project == project
97 and on_demand_feature_view.spec.name == name
98 ):
99 return OnDemandFeatureView.from_proto(on_demand_feature_view)
100 raise FeatureViewNotFoundException(name, project=project)
101
102
103 def get_data_source(
104 registry_proto: RegistryProto, name: str, project: str
105 ) -> DataSource:
106 for data_source in registry_proto.data_sources:
107 if data_source.project == project and data_source.name == name:
108 return DataSource.from_proto(data_source)
109 raise DataSourceObjectNotFoundException(name, project=project)
110
111
112 def get_entity(registry_proto: RegistryProto, name: str, project: str) -> Entity:
113 for entity_proto in registry_proto.entities:
114 if entity_proto.spec.name == name and entity_proto.spec.project == project:
115 return Entity.from_proto(entity_proto)
116 raise EntityNotFoundException(name, project=project)
117
118
119 def get_saved_dataset(
120 registry_proto: RegistryProto, name: str, project: str
121 ) -> SavedDataset:
122 for saved_dataset in registry_proto.saved_datasets:
123 if saved_dataset.spec.name == name and saved_dataset.spec.project == project:
124 return SavedDataset.from_proto(saved_dataset)
125 raise SavedDatasetNotFound(name, project=project)
126
127
128 def get_validation_reference(
129 registry_proto: RegistryProto, name: str, project: str
130 ) -> ValidationReference:
131 for validation_reference in registry_proto.validation_references:
132 if (
133 validation_reference.name == name
134 and validation_reference.project == project
135 ):
136 return ValidationReference.from_proto(validation_reference)
137 raise ValidationReferenceNotFound(name, project=project)
138
139
140 def list_feature_services(
141 registry_proto: RegistryProto, project: str, allow_cache: bool = False
142 ) -> List[FeatureService]:
143 feature_services = []
144 for feature_service_proto in registry_proto.feature_services:
145 if feature_service_proto.spec.project == project:
146 feature_services.append(FeatureService.from_proto(feature_service_proto))
147 return feature_services
148
149
150 def list_feature_views(
151 registry_proto: RegistryProto, project: str
152 ) -> List[FeatureView]:
153 feature_views: List[FeatureView] = []
154 for feature_view_proto in registry_proto.feature_views:
155 if feature_view_proto.spec.project == project:
156 feature_views.append(FeatureView.from_proto(feature_view_proto))
157 return feature_views
158
159
160 def list_request_feature_views(
161 registry_proto: RegistryProto, project: str
162 ) -> List[RequestFeatureView]:
163 feature_views: List[RequestFeatureView] = []
164 for request_feature_view_proto in registry_proto.request_feature_views:
165 if request_feature_view_proto.spec.project == project:
166 feature_views.append(
167 RequestFeatureView.from_proto(request_feature_view_proto)
168 )
169 return feature_views
170
171
172 def list_stream_feature_views(
173 registry_proto: RegistryProto, project: str
174 ) -> List[StreamFeatureView]:
175 stream_feature_views = []
176 for stream_feature_view in registry_proto.stream_feature_views:
177 if stream_feature_view.spec.project == project:
178 stream_feature_views.append(
179 StreamFeatureView.from_proto(stream_feature_view)
180 )
181 return stream_feature_views
182
183
184 def list_on_demand_feature_views(
185 registry_proto: RegistryProto, project: str
186 ) -> List[OnDemandFeatureView]:
187 on_demand_feature_views = []
188 for on_demand_feature_view in registry_proto.on_demand_feature_views:
189 if on_demand_feature_view.spec.project == project:
190 on_demand_feature_views.append(
191 OnDemandFeatureView.from_proto(on_demand_feature_view)
192 )
193 return on_demand_feature_views
194
195
196 def list_entities(registry_proto: RegistryProto, project: str) -> List[Entity]:
197 entities = []
198 for entity_proto in registry_proto.entities:
199 if entity_proto.spec.project == project:
200 entities.append(Entity.from_proto(entity_proto))
201 return entities
202
203
204 def list_data_sources(registry_proto: RegistryProto, project: str) -> List[DataSource]:
205 data_sources = []
206 for data_source_proto in registry_proto.data_sources:
207 if data_source_proto.project == project:
208 data_sources.append(DataSource.from_proto(data_source_proto))
209 return data_sources
210
211
212 def list_saved_datasets(
213 registry_proto: RegistryProto, project: str
214 ) -> List[SavedDataset]:
215 saved_datasets = []
216 for saved_dataset in registry_proto.saved_datasets:
217 if saved_dataset.project == project:
218 saved_datasets.append(SavedDataset.from_proto(saved_dataset))
219 return saved_datasets
220
221
222 def list_validation_references(
223 registry_proto: RegistryProto, project: str
224 ) -> List[ValidationReference]:
225 validation_references = []
226 for validation_reference in registry_proto.validation_references:
227 if validation_reference.project == project:
228 validation_references.append(
229 ValidationReference.from_proto(validation_reference)
230 )
231 return validation_references
232
233
234 def list_project_metadata(
235 registry_proto: RegistryProto, project: str
236 ) -> List[ProjectMetadata]:
237 return [
238 ProjectMetadata.from_proto(project_metadata)
239 for project_metadata in registry_proto.project_metadata
240 if project_metadata.project == project
241 ]
242
[end of sdk/python/feast/infra/registry/proto_registry_utils.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/sdk/python/feast/infra/registry/proto_registry_utils.py b/sdk/python/feast/infra/registry/proto_registry_utils.py
--- a/sdk/python/feast/infra/registry/proto_registry_utils.py
+++ b/sdk/python/feast/infra/registry/proto_registry_utils.py
@@ -1,4 +1,5 @@
import uuid
+from functools import wraps
from typing import List, Optional
from feast import usage
@@ -23,6 +24,26 @@
from feast.stream_feature_view import StreamFeatureView
+def registry_proto_cache(func):
+ cache_key = None
+ cache_value = None
+
+ @wraps(func)
+ def wrapper(registry_proto: RegistryProto, project: str):
+ nonlocal cache_key, cache_value
+
+ key = tuple([id(registry_proto), registry_proto.version_id, project])
+
+ if key == cache_key:
+ return cache_value
+ else:
+ cache_value = func(registry_proto, project)
+ cache_key = key
+ return cache_value
+
+ return wrapper
+
+
def init_project_metadata(cached_registry_proto: RegistryProto, project: str):
new_project_uuid = f"{uuid.uuid4()}"
usage.set_current_project_uuid(new_project_uuid)
@@ -137,8 +158,9 @@
raise ValidationReferenceNotFound(name, project=project)
+@registry_proto_cache
def list_feature_services(
- registry_proto: RegistryProto, project: str, allow_cache: bool = False
+ registry_proto: RegistryProto, project: str
) -> List[FeatureService]:
feature_services = []
for feature_service_proto in registry_proto.feature_services:
@@ -147,6 +169,7 @@
return feature_services
+@registry_proto_cache
def list_feature_views(
registry_proto: RegistryProto, project: str
) -> List[FeatureView]:
@@ -157,6 +180,7 @@
return feature_views
+@registry_proto_cache
def list_request_feature_views(
registry_proto: RegistryProto, project: str
) -> List[RequestFeatureView]:
@@ -169,6 +193,7 @@
return feature_views
+@registry_proto_cache
def list_stream_feature_views(
registry_proto: RegistryProto, project: str
) -> List[StreamFeatureView]:
@@ -181,6 +206,7 @@
return stream_feature_views
+@registry_proto_cache
def list_on_demand_feature_views(
registry_proto: RegistryProto, project: str
) -> List[OnDemandFeatureView]:
@@ -193,6 +219,7 @@
return on_demand_feature_views
+@registry_proto_cache
def list_entities(registry_proto: RegistryProto, project: str) -> List[Entity]:
entities = []
for entity_proto in registry_proto.entities:
@@ -201,6 +228,7 @@
return entities
+@registry_proto_cache
def list_data_sources(registry_proto: RegistryProto, project: str) -> List[DataSource]:
data_sources = []
for data_source_proto in registry_proto.data_sources:
@@ -209,6 +237,7 @@
return data_sources
+@registry_proto_cache
def list_saved_datasets(
registry_proto: RegistryProto, project: str
) -> List[SavedDataset]:
@@ -219,6 +248,7 @@
return saved_datasets
+@registry_proto_cache
def list_validation_references(
registry_proto: RegistryProto, project: str
) -> List[ValidationReference]:
@@ -231,6 +261,7 @@
return validation_references
+@registry_proto_cache
def list_project_metadata(
registry_proto: RegistryProto, project: str
) -> List[ProjectMetadata]:
|
{"golden_diff": "diff --git a/sdk/python/feast/infra/registry/proto_registry_utils.py b/sdk/python/feast/infra/registry/proto_registry_utils.py\n--- a/sdk/python/feast/infra/registry/proto_registry_utils.py\n+++ b/sdk/python/feast/infra/registry/proto_registry_utils.py\n@@ -1,4 +1,5 @@\n import uuid\n+from functools import wraps\n from typing import List, Optional\n \n from feast import usage\n@@ -23,6 +24,26 @@\n from feast.stream_feature_view import StreamFeatureView\n \n \n+def registry_proto_cache(func):\n+ cache_key = None\n+ cache_value = None\n+\n+ @wraps(func)\n+ def wrapper(registry_proto: RegistryProto, project: str):\n+ nonlocal cache_key, cache_value\n+\n+ key = tuple([id(registry_proto), registry_proto.version_id, project])\n+\n+ if key == cache_key:\n+ return cache_value\n+ else:\n+ cache_value = func(registry_proto, project)\n+ cache_key = key\n+ return cache_value\n+\n+ return wrapper\n+\n+\n def init_project_metadata(cached_registry_proto: RegistryProto, project: str):\n new_project_uuid = f\"{uuid.uuid4()}\"\n usage.set_current_project_uuid(new_project_uuid)\n@@ -137,8 +158,9 @@\n raise ValidationReferenceNotFound(name, project=project)\n \n \n+@registry_proto_cache\n def list_feature_services(\n- registry_proto: RegistryProto, project: str, allow_cache: bool = False\n+ registry_proto: RegistryProto, project: str\n ) -> List[FeatureService]:\n feature_services = []\n for feature_service_proto in registry_proto.feature_services:\n@@ -147,6 +169,7 @@\n return feature_services\n \n \n+@registry_proto_cache\n def list_feature_views(\n registry_proto: RegistryProto, project: str\n ) -> List[FeatureView]:\n@@ -157,6 +180,7 @@\n return feature_views\n \n \n+@registry_proto_cache\n def list_request_feature_views(\n registry_proto: RegistryProto, project: str\n ) -> List[RequestFeatureView]:\n@@ -169,6 +193,7 @@\n return feature_views\n \n \n+@registry_proto_cache\n def list_stream_feature_views(\n registry_proto: RegistryProto, project: str\n ) -> List[StreamFeatureView]:\n@@ -181,6 +206,7 @@\n return stream_feature_views\n \n \n+@registry_proto_cache\n def list_on_demand_feature_views(\n registry_proto: RegistryProto, project: str\n ) -> List[OnDemandFeatureView]:\n@@ -193,6 +219,7 @@\n return on_demand_feature_views\n \n \n+@registry_proto_cache\n def list_entities(registry_proto: RegistryProto, project: str) -> List[Entity]:\n entities = []\n for entity_proto in registry_proto.entities:\n@@ -201,6 +228,7 @@\n return entities\n \n \n+@registry_proto_cache\n def list_data_sources(registry_proto: RegistryProto, project: str) -> List[DataSource]:\n data_sources = []\n for data_source_proto in registry_proto.data_sources:\n@@ -209,6 +237,7 @@\n return data_sources\n \n \n+@registry_proto_cache\n def list_saved_datasets(\n registry_proto: RegistryProto, project: str\n ) -> List[SavedDataset]:\n@@ -219,6 +248,7 @@\n return saved_datasets\n \n \n+@registry_proto_cache\n def list_validation_references(\n registry_proto: RegistryProto, project: str\n ) -> List[ValidationReference]:\n@@ -231,6 +261,7 @@\n return validation_references\n \n \n+@registry_proto_cache\n def list_project_metadata(\n registry_proto: RegistryProto, project: str\n ) -> List[ProjectMetadata]:\n", "issue": "performance issues of getting online features related to parsing protobuf data\n**Is your feature request related to a problem? Please describe.**\r\n\r\nThe profiler of one application I am working on shows the `from_proto` method of the `FeatureView` class takes up 80% of the execution time. The result of cProfile is shown below.\r\n```\r\n ncalls tottime percall cumtime percall filename:lineno(function)\r\n 1 0.003 0.003 78.931 78.931 profiler.py:53(main)\r\n 100 0.018 0.000 78.849 0.788 get_features.py:122(get_xxxx_features)\r\n 1100/400 0.020 0.000 77.077 0.193 usage.py:274(wrapper)\r\n 444200/9800 1.713 0.000 72.206 0.007 __init__.py:1030(wrapper)\r\n 200 0.078 0.000 68.913 0.345 feature_store.py:1527(get_online_features)\r\n 200 0.087 0.000 68.835 0.344 feature_store.py:1590(_get_online_features)\r\n 3500 5.634 0.002 62.442 0.018 feature_view.py:369(from_proto)\r\n 200 0.005 0.000 59.362 0.297 feature_store.py:2149(_get_feature_views_to_use)\r\n 200 0.002 0.000 58.501 0.293 feature_store.py:281(_list_feature_views)\r\n 200 0.001 0.000 58.499 0.292 registry.py:523(list_feature_views)\r\n 200 0.016 0.000 58.495 0.292 proto_registry_utils.py:150(list_feature_views)\r\n```\r\nThere are 3 feature views accessed by `get_xxxx_features`, however 100 calls of `get_xxxx_features` lead to 3500 calls of `from_proto`. There are 17 feature views in the feature store of this application and 3 of them are used by `get_xxxx_features`.\r\n\r\nEnvironment: `continuumio/miniconda3:4.11.0` (linux/amd64) base image, `python==3.9.7`, `feast=0.31.1`, `protobuf==4.23.2`\r\n\r\n**Describe the solution you'd like**\r\n\r\nInstead of caching the protobuf blob of `FeatureView`, cache the `FeatureView` python object in memory.\r\n\r\n**Describe alternatives you've considered**\r\n\r\nmodify the `get_online_features` [method](https://github.com/feast-dev/feast/blob/v0.31.1/sdk/python/feast/feature_store.py#L1528) to\r\n```python\r\ndef get_online_features(\r\n self,\r\n features: Union[List[str], List[FeatureView], FeatureService],\r\n entity_rows: List[Dict[str, Any]],\r\n full_feature_names: bool = False,\r\n):\r\n``` \r\nso that the application developer has the option to cache the `FeatureView` objects and use them to get features directly (by passing the `get_feature_views_to_use` [step](https://github.com/feast-dev/feast/blob/v0.31.1/sdk/python/feast/feature_store.py#L1606-L1613) )\r\n \n", "before_files": [{"content": "import uuid\nfrom typing import List, Optional\n\nfrom feast import usage\nfrom feast.data_source import DataSource\nfrom feast.entity import Entity\nfrom feast.errors import (\n DataSourceObjectNotFoundException,\n EntityNotFoundException,\n FeatureServiceNotFoundException,\n FeatureViewNotFoundException,\n SavedDatasetNotFound,\n ValidationReferenceNotFound,\n)\nfrom feast.feature_service import FeatureService\nfrom feast.feature_view import FeatureView\nfrom feast.on_demand_feature_view import OnDemandFeatureView\nfrom feast.project_metadata import ProjectMetadata\nfrom feast.protos.feast.core.Registry_pb2 import ProjectMetadata as ProjectMetadataProto\nfrom feast.protos.feast.core.Registry_pb2 import Registry as RegistryProto\nfrom feast.request_feature_view import RequestFeatureView\nfrom feast.saved_dataset import SavedDataset, ValidationReference\nfrom feast.stream_feature_view import StreamFeatureView\n\n\ndef init_project_metadata(cached_registry_proto: RegistryProto, project: str):\n new_project_uuid = f\"{uuid.uuid4()}\"\n usage.set_current_project_uuid(new_project_uuid)\n cached_registry_proto.project_metadata.append(\n ProjectMetadata(project_name=project, project_uuid=new_project_uuid).to_proto()\n )\n\n\ndef get_project_metadata(\n registry_proto: Optional[RegistryProto], project: str\n) -> Optional[ProjectMetadataProto]:\n if not registry_proto:\n return None\n for pm in registry_proto.project_metadata:\n if pm.project == project:\n return pm\n return None\n\n\ndef get_feature_service(\n registry_proto: RegistryProto, name: str, project: str\n) -> FeatureService:\n for feature_service_proto in registry_proto.feature_services:\n if (\n feature_service_proto.spec.project == project\n and feature_service_proto.spec.name == name\n ):\n return FeatureService.from_proto(feature_service_proto)\n raise FeatureServiceNotFoundException(name, project=project)\n\n\ndef get_feature_view(\n registry_proto: RegistryProto, name: str, project: str\n) -> FeatureView:\n for feature_view_proto in registry_proto.feature_views:\n if (\n feature_view_proto.spec.name == name\n and feature_view_proto.spec.project == project\n ):\n return FeatureView.from_proto(feature_view_proto)\n raise FeatureViewNotFoundException(name, project)\n\n\ndef get_stream_feature_view(\n registry_proto: RegistryProto, name: str, project: str\n) -> StreamFeatureView:\n for feature_view_proto in registry_proto.stream_feature_views:\n if (\n feature_view_proto.spec.name == name\n and feature_view_proto.spec.project == project\n ):\n return StreamFeatureView.from_proto(feature_view_proto)\n raise FeatureViewNotFoundException(name, project)\n\n\ndef get_request_feature_view(registry_proto: RegistryProto, name: str, project: str):\n for feature_view_proto in registry_proto.feature_views:\n if (\n feature_view_proto.spec.name == name\n and feature_view_proto.spec.project == project\n ):\n return RequestFeatureView.from_proto(feature_view_proto)\n raise FeatureViewNotFoundException(name, project)\n\n\ndef get_on_demand_feature_view(\n registry_proto: RegistryProto, name: str, project: str\n) -> OnDemandFeatureView:\n for on_demand_feature_view in registry_proto.on_demand_feature_views:\n if (\n on_demand_feature_view.spec.project == project\n and on_demand_feature_view.spec.name == name\n ):\n return OnDemandFeatureView.from_proto(on_demand_feature_view)\n raise FeatureViewNotFoundException(name, project=project)\n\n\ndef get_data_source(\n registry_proto: RegistryProto, name: str, project: str\n) -> DataSource:\n for data_source in registry_proto.data_sources:\n if data_source.project == project and data_source.name == name:\n return DataSource.from_proto(data_source)\n raise DataSourceObjectNotFoundException(name, project=project)\n\n\ndef get_entity(registry_proto: RegistryProto, name: str, project: str) -> Entity:\n for entity_proto in registry_proto.entities:\n if entity_proto.spec.name == name and entity_proto.spec.project == project:\n return Entity.from_proto(entity_proto)\n raise EntityNotFoundException(name, project=project)\n\n\ndef get_saved_dataset(\n registry_proto: RegistryProto, name: str, project: str\n) -> SavedDataset:\n for saved_dataset in registry_proto.saved_datasets:\n if saved_dataset.spec.name == name and saved_dataset.spec.project == project:\n return SavedDataset.from_proto(saved_dataset)\n raise SavedDatasetNotFound(name, project=project)\n\n\ndef get_validation_reference(\n registry_proto: RegistryProto, name: str, project: str\n) -> ValidationReference:\n for validation_reference in registry_proto.validation_references:\n if (\n validation_reference.name == name\n and validation_reference.project == project\n ):\n return ValidationReference.from_proto(validation_reference)\n raise ValidationReferenceNotFound(name, project=project)\n\n\ndef list_feature_services(\n registry_proto: RegistryProto, project: str, allow_cache: bool = False\n) -> List[FeatureService]:\n feature_services = []\n for feature_service_proto in registry_proto.feature_services:\n if feature_service_proto.spec.project == project:\n feature_services.append(FeatureService.from_proto(feature_service_proto))\n return feature_services\n\n\ndef list_feature_views(\n registry_proto: RegistryProto, project: str\n) -> List[FeatureView]:\n feature_views: List[FeatureView] = []\n for feature_view_proto in registry_proto.feature_views:\n if feature_view_proto.spec.project == project:\n feature_views.append(FeatureView.from_proto(feature_view_proto))\n return feature_views\n\n\ndef list_request_feature_views(\n registry_proto: RegistryProto, project: str\n) -> List[RequestFeatureView]:\n feature_views: List[RequestFeatureView] = []\n for request_feature_view_proto in registry_proto.request_feature_views:\n if request_feature_view_proto.spec.project == project:\n feature_views.append(\n RequestFeatureView.from_proto(request_feature_view_proto)\n )\n return feature_views\n\n\ndef list_stream_feature_views(\n registry_proto: RegistryProto, project: str\n) -> List[StreamFeatureView]:\n stream_feature_views = []\n for stream_feature_view in registry_proto.stream_feature_views:\n if stream_feature_view.spec.project == project:\n stream_feature_views.append(\n StreamFeatureView.from_proto(stream_feature_view)\n )\n return stream_feature_views\n\n\ndef list_on_demand_feature_views(\n registry_proto: RegistryProto, project: str\n) -> List[OnDemandFeatureView]:\n on_demand_feature_views = []\n for on_demand_feature_view in registry_proto.on_demand_feature_views:\n if on_demand_feature_view.spec.project == project:\n on_demand_feature_views.append(\n OnDemandFeatureView.from_proto(on_demand_feature_view)\n )\n return on_demand_feature_views\n\n\ndef list_entities(registry_proto: RegistryProto, project: str) -> List[Entity]:\n entities = []\n for entity_proto in registry_proto.entities:\n if entity_proto.spec.project == project:\n entities.append(Entity.from_proto(entity_proto))\n return entities\n\n\ndef list_data_sources(registry_proto: RegistryProto, project: str) -> List[DataSource]:\n data_sources = []\n for data_source_proto in registry_proto.data_sources:\n if data_source_proto.project == project:\n data_sources.append(DataSource.from_proto(data_source_proto))\n return data_sources\n\n\ndef list_saved_datasets(\n registry_proto: RegistryProto, project: str\n) -> List[SavedDataset]:\n saved_datasets = []\n for saved_dataset in registry_proto.saved_datasets:\n if saved_dataset.project == project:\n saved_datasets.append(SavedDataset.from_proto(saved_dataset))\n return saved_datasets\n\n\ndef list_validation_references(\n registry_proto: RegistryProto, project: str\n) -> List[ValidationReference]:\n validation_references = []\n for validation_reference in registry_proto.validation_references:\n if validation_reference.project == project:\n validation_references.append(\n ValidationReference.from_proto(validation_reference)\n )\n return validation_references\n\n\ndef list_project_metadata(\n registry_proto: RegistryProto, project: str\n) -> List[ProjectMetadata]:\n return [\n ProjectMetadata.from_proto(project_metadata)\n for project_metadata in registry_proto.project_metadata\n if project_metadata.project == project\n ]\n", "path": "sdk/python/feast/infra/registry/proto_registry_utils.py"}]}
| 3,828 | 846 |
gh_patches_debug_19757
|
rasdani/github-patches
|
git_diff
|
biopython__biopython-4328
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Biopython is not installable as a git repository with Poetry
### Setup
I am reporting a problem with Biopython version, Python version, and operating
system as follows:
```shell
❯ python -c 'import sys; print(sys.version)'
3.10.9 (main, Dec 7 2022, 01:12:00) [GCC 9.4.0]
```
```shell
❯ python -c 'import platform; print(platform.python_implementation()); print(platform.platform())'
CPython
Linux-5.15.90.1-microsoft-standard-WSL2-x86_64-with-glibc2.31
```
(This is an install issue, so certainly no install)
```shell
❯ python -c 'import Bio; print(Bio.__version__)'
Traceback (most recent call last):
File "<string>", line 1, in <module>
ModuleNotFoundError: No module named 'Bio'
```
### Expected behaviour
With the following in `pyproject.toml` under `[tool.poetry.dependencies]`:
```toml
biopython = { git = "https://github.com/biopython/biopython", rev="ca64586b1c6f6c8439c88e576fed091c923ad8f1" }
```
running `poetry lock` should update with the relevant biopython package.
### Actual behaviour
```shell
❯ poetry lock
Updating dependencies
Resolving dependencies... (0.0s)
Invalid version 'Undefined' on package biopython
```
### Steps to reproduce
Follow the above steps.
### Suggested Remediation
- Option A: Add a `setup.cfg` with version information as [here](https://packaging.python.org/en/latest/guides/single-sourcing-package-version/)
- Option B: Move to [PEP 518 pyproject.toml](https://peps.python.org/pep-0518/)
</issue>
<code>
[start of setup.py]
1 """setuptools based setup script for Biopython.
2
3 This uses setuptools which is now the standard python mechanism for
4 installing packages. If you have downloaded and uncompressed the
5 Biopython source code, or fetched it from git, for the simplest
6 installation just type the command::
7
8 python setup.py install
9
10 However, you would normally install the latest Biopython release from
11 the PyPI archive with::
12
13 pip install biopython
14
15 For more in-depth instructions, see the installation section of the
16 Biopython manual, linked to from:
17
18 http://biopython.org/wiki/Documentation
19
20 Or, if all else fails, feel free to write to the sign up to the Biopython
21 mailing list and ask for help. See:
22
23 http://biopython.org/wiki/Mailing_lists
24 """
25
26 import sys
27 import os
28
29 try:
30 from setuptools import setup
31 from setuptools import Command
32 from setuptools import Extension
33 except ImportError:
34 sys.exit(
35 "We need the Python library setuptools to be installed. "
36 "Try running: python -m ensurepip"
37 )
38
39 if "bdist_wheel" in sys.argv:
40 try:
41 import wheel # noqa: F401
42 except ImportError:
43 sys.exit(
44 "We need both setuptools AND wheel packages installed "
45 "for bdist_wheel to work. Try running: pip install wheel"
46 )
47
48
49 # Make sure we have the right Python version.
50 MIN_PY_VER = (3, 8)
51 if sys.version_info[:2] < MIN_PY_VER:
52 sys.stderr.write(
53 ("ERROR: Biopython requires Python %i.%i or later. " % MIN_PY_VER)
54 + ("Python %d.%d detected.\n" % sys.version_info[:2])
55 )
56 sys.exit(1)
57
58
59 class test_biopython(Command):
60 """Run all of the tests for the package.
61
62 This is a automatic test run class to make distutils kind of act like
63 perl. With this you can do:
64
65 python setup.py build
66 python setup.py install
67 python setup.py test
68
69 """
70
71 description = "Automatically run the test suite for Biopython."
72 user_options = [("offline", None, "Don't run online tests")]
73
74 def initialize_options(self):
75 """No-op, initialise options."""
76 self.offline = None
77
78 def finalize_options(self):
79 """No-op, finalise options."""
80 pass
81
82 def run(self):
83 """Run the tests."""
84 this_dir = os.getcwd()
85
86 # change to the test dir and run the tests
87 os.chdir("Tests")
88 sys.path.insert(0, "")
89 import run_tests
90
91 if self.offline:
92 run_tests.main(["--offline"])
93 else:
94 run_tests.main([])
95
96 # change back to the current directory
97 os.chdir(this_dir)
98
99
100 def can_import(module_name):
101 """Check we can import the requested module."""
102 try:
103 return __import__(module_name)
104 except ImportError:
105 return None
106
107
108 # Using requirements.txt is preferred for an application
109 # (and likely will pin specific version numbers), using
110 # setup.py's install_requires is preferred for a library
111 # (and should try not to be overly narrow with versions).
112 REQUIRES = ["numpy"]
113
114 # --- set up the packages we are going to install
115 # standard biopython packages
116 PACKAGES = [
117 "Bio",
118 "Bio.Affy",
119 "Bio.Align",
120 "Bio.Align.Applications",
121 "Bio.Align.substitution_matrices",
122 "Bio.AlignIO",
123 "Bio.Alphabet",
124 "Bio.Application",
125 "Bio.Blast",
126 "Bio.CAPS",
127 "Bio.Cluster",
128 "Bio.codonalign",
129 "Bio.Compass",
130 "Bio.Data",
131 "Bio.Emboss",
132 "Bio.Entrez",
133 "Bio.ExPASy",
134 "Bio.GenBank",
135 "Bio.Geo",
136 "Bio.Graphics",
137 "Bio.Graphics.GenomeDiagram",
138 "Bio.HMM",
139 "Bio.KEGG",
140 "Bio.KEGG.Compound",
141 "Bio.KEGG.Enzyme",
142 "Bio.KEGG.Gene",
143 "Bio.KEGG.Map",
144 "Bio.PDB.mmtf",
145 "Bio.KEGG.KGML",
146 "Bio.Medline",
147 "Bio.motifs",
148 "Bio.motifs.applications",
149 "Bio.motifs.jaspar",
150 "Bio.Nexus",
151 "Bio.NMR",
152 "Bio.Pathway",
153 "Bio.Pathway.Rep",
154 "Bio.PDB",
155 "Bio.phenotype",
156 "Bio.PopGen",
157 "Bio.PopGen.GenePop",
158 "Bio.Restriction",
159 "Bio.SCOP",
160 "Bio.SearchIO",
161 "Bio.SearchIO._legacy",
162 "Bio.SearchIO._model",
163 "Bio.SearchIO.BlastIO",
164 "Bio.SearchIO.HHsuiteIO",
165 "Bio.SearchIO.HmmerIO",
166 "Bio.SearchIO.ExonerateIO",
167 "Bio.SearchIO.InterproscanIO",
168 "Bio.SeqIO",
169 "Bio.SeqUtils",
170 "Bio.Sequencing",
171 "Bio.Sequencing.Applications",
172 "Bio.SVDSuperimposer",
173 "Bio.PDB.QCPSuperimposer",
174 "Bio.SwissProt",
175 "Bio.TogoWS",
176 "Bio.Phylo",
177 "Bio.Phylo.Applications",
178 "Bio.Phylo.PAML",
179 "Bio.UniGene",
180 "Bio.UniProt",
181 "Bio.Wise",
182 # Other top level packages,
183 "BioSQL",
184 ]
185
186 EXTENSIONS = [
187 Extension("Bio.Align._aligners", ["Bio/Align/_aligners.c"]),
188 Extension("Bio.cpairwise2", ["Bio/cpairwise2module.c"]),
189 Extension("Bio.Nexus.cnexus", ["Bio/Nexus/cnexus.c"]),
190 Extension(
191 "Bio.PDB.QCPSuperimposer.qcprotmodule",
192 ["Bio/PDB/QCPSuperimposer/qcprotmodule.c"],
193 ),
194 Extension("Bio.motifs._pwm", ["Bio/motifs/_pwm.c"]),
195 Extension(
196 "Bio.Cluster._cluster", ["Bio/Cluster/cluster.c", "Bio/Cluster/clustermodule.c"]
197 ),
198 Extension("Bio.PDB.kdtrees", ["Bio/PDB/kdtrees.c"]),
199 Extension("Bio.PDB.ccealign", ["Bio/PDB/ccealignmodule.c"]),
200 Extension("Bio.SeqIO._twoBitIO", ["Bio/SeqIO/_twoBitIO.c"]),
201 ]
202
203 # We now define the Biopython version number in Bio/__init__.py
204 # Here we can't use "import Bio" then "Bio.__version__" as that would
205 # tell us the version of Biopython already installed (if any).
206 __version__ = "Undefined"
207 for line in open("Bio/__init__.py"):
208 if line.startswith("__version__"):
209 exec(line.strip())
210
211 # We now load in our reStructuredText README.rst file to pass explicitly in the
212 # metadata, since at time of writing PyPI did not do this for us.
213 #
214 # Must make encoding explicit to avoid any conflict with the local default.
215 # Currently keeping README as ASCII (might switch to UTF8 later if needed).
216 # If any invalid character does appear in README, this will fail and alert us.
217 with open("README.rst", encoding="ascii") as handle:
218 readme_rst = handle.read()
219
220 setup(
221 name="biopython",
222 version=__version__,
223 author="The Biopython Contributors",
224 author_email="[email protected]",
225 url="https://biopython.org/",
226 description="Freely available tools for computational molecular biology.",
227 long_description=readme_rst,
228 project_urls={
229 "Documentation": "https://biopython.org/wiki/Documentation",
230 "Source": "https://github.com/biopython/biopython/",
231 "Tracker": "https://github.com/biopython/biopython/issues",
232 },
233 classifiers=[
234 "Development Status :: 5 - Production/Stable",
235 "Intended Audience :: Developers",
236 "Intended Audience :: Science/Research",
237 "License :: Freely Distributable",
238 # Technically the "Biopython License Agreement" is not OSI approved,
239 # but is almost https://opensource.org/licenses/HPND so might put:
240 # 'License :: OSI Approved',
241 # To resolve this we are moving to dual-licensing with 3-clause BSD:
242 # 'License :: OSI Approved :: BSD License',
243 "Operating System :: OS Independent",
244 "Programming Language :: Python",
245 "Programming Language :: Python :: 3",
246 "Programming Language :: Python :: 3.8",
247 "Programming Language :: Python :: 3.9",
248 "Programming Language :: Python :: 3.10",
249 "Programming Language :: Python :: 3.11",
250 "Topic :: Scientific/Engineering",
251 "Topic :: Scientific/Engineering :: Bio-Informatics",
252 "Topic :: Software Development :: Libraries :: Python Modules",
253 ],
254 cmdclass={"test": test_biopython},
255 packages=PACKAGES,
256 ext_modules=EXTENSIONS,
257 include_package_data=True, # done via MANIFEST.in under setuptools
258 install_requires=REQUIRES,
259 python_requires=">=%i.%i" % MIN_PY_VER,
260 )
261
[end of setup.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/setup.py b/setup.py
--- a/setup.py
+++ b/setup.py
@@ -25,6 +25,7 @@
import sys
import os
+import ast
try:
from setuptools import setup
@@ -200,13 +201,16 @@
Extension("Bio.SeqIO._twoBitIO", ["Bio/SeqIO/_twoBitIO.c"]),
]
-# We now define the Biopython version number in Bio/__init__.py
-# Here we can't use "import Bio" then "Bio.__version__" as that would
-# tell us the version of Biopython already installed (if any).
-__version__ = "Undefined"
-for line in open("Bio/__init__.py"):
- if line.startswith("__version__"):
- exec(line.strip())
+
+def get_version():
+ """Get version number from __init__.py."""
+ for line in open("Bio/__init__.py"):
+ if line.startswith("__version__ = "):
+ return ast.literal_eval(line.split("=")[1].strip())
+ return "Undefined"
+
+
+__version__ = get_version()
# We now load in our reStructuredText README.rst file to pass explicitly in the
# metadata, since at time of writing PyPI did not do this for us.
|
{"golden_diff": "diff --git a/setup.py b/setup.py\n--- a/setup.py\n+++ b/setup.py\n@@ -25,6 +25,7 @@\n \n import sys\n import os\n+import ast\n \n try:\n from setuptools import setup\n@@ -200,13 +201,16 @@\n Extension(\"Bio.SeqIO._twoBitIO\", [\"Bio/SeqIO/_twoBitIO.c\"]),\n ]\n \n-# We now define the Biopython version number in Bio/__init__.py\n-# Here we can't use \"import Bio\" then \"Bio.__version__\" as that would\n-# tell us the version of Biopython already installed (if any).\n-__version__ = \"Undefined\"\n-for line in open(\"Bio/__init__.py\"):\n- if line.startswith(\"__version__\"):\n- exec(line.strip())\n+\n+def get_version():\n+ \"\"\"Get version number from __init__.py.\"\"\"\n+ for line in open(\"Bio/__init__.py\"):\n+ if line.startswith(\"__version__ = \"):\n+ return ast.literal_eval(line.split(\"=\")[1].strip())\n+ return \"Undefined\"\n+\n+\n+__version__ = get_version()\n \n # We now load in our reStructuredText README.rst file to pass explicitly in the\n # metadata, since at time of writing PyPI did not do this for us.\n", "issue": "Biopython is not installable as a git repository with Poetry\n### Setup\r\n\r\nI am reporting a problem with Biopython version, Python version, and operating\r\nsystem as follows:\r\n\r\n```shell\r\n\u276f python -c 'import sys; print(sys.version)'\r\n3.10.9 (main, Dec 7 2022, 01:12:00) [GCC 9.4.0]\r\n```\r\n\r\n```shell\r\n\u276f python -c 'import platform; print(platform.python_implementation()); print(platform.platform())'\r\nCPython\r\nLinux-5.15.90.1-microsoft-standard-WSL2-x86_64-with-glibc2.31\r\n```\r\n\r\n(This is an install issue, so certainly no install)\r\n```shell\r\n\u276f python -c 'import Bio; print(Bio.__version__)'\r\nTraceback (most recent call last):\r\n File \"<string>\", line 1, in <module>\r\nModuleNotFoundError: No module named 'Bio'\r\n```\r\n\r\n### Expected behaviour\r\n\r\nWith the following in `pyproject.toml` under `[tool.poetry.dependencies]`:\r\n\r\n```toml\r\nbiopython = { git = \"https://github.com/biopython/biopython\", rev=\"ca64586b1c6f6c8439c88e576fed091c923ad8f1\" }\r\n```\r\n\r\nrunning `poetry lock` should update with the relevant biopython package.\r\n\r\n### Actual behaviour\r\n\r\n```shell\r\n\u276f poetry lock\r\nUpdating dependencies\r\nResolving dependencies... (0.0s)\r\n\r\nInvalid version 'Undefined' on package biopython\r\n```\r\n\r\n### Steps to reproduce\r\n\r\nFollow the above steps.\r\n\r\n### Suggested Remediation\r\n\r\n- Option A: Add a `setup.cfg` with version information as [here](https://packaging.python.org/en/latest/guides/single-sourcing-package-version/)\r\n- Option B: Move to [PEP 518 pyproject.toml](https://peps.python.org/pep-0518/) \n", "before_files": [{"content": "\"\"\"setuptools based setup script for Biopython.\n\nThis uses setuptools which is now the standard python mechanism for\ninstalling packages. If you have downloaded and uncompressed the\nBiopython source code, or fetched it from git, for the simplest\ninstallation just type the command::\n\n python setup.py install\n\nHowever, you would normally install the latest Biopython release from\nthe PyPI archive with::\n\n pip install biopython\n\nFor more in-depth instructions, see the installation section of the\nBiopython manual, linked to from:\n\nhttp://biopython.org/wiki/Documentation\n\nOr, if all else fails, feel free to write to the sign up to the Biopython\nmailing list and ask for help. See:\n\nhttp://biopython.org/wiki/Mailing_lists\n\"\"\"\n\nimport sys\nimport os\n\ntry:\n from setuptools import setup\n from setuptools import Command\n from setuptools import Extension\nexcept ImportError:\n sys.exit(\n \"We need the Python library setuptools to be installed. \"\n \"Try running: python -m ensurepip\"\n )\n\nif \"bdist_wheel\" in sys.argv:\n try:\n import wheel # noqa: F401\n except ImportError:\n sys.exit(\n \"We need both setuptools AND wheel packages installed \"\n \"for bdist_wheel to work. Try running: pip install wheel\"\n )\n\n\n# Make sure we have the right Python version.\nMIN_PY_VER = (3, 8)\nif sys.version_info[:2] < MIN_PY_VER:\n sys.stderr.write(\n (\"ERROR: Biopython requires Python %i.%i or later. \" % MIN_PY_VER)\n + (\"Python %d.%d detected.\\n\" % sys.version_info[:2])\n )\n sys.exit(1)\n\n\nclass test_biopython(Command):\n \"\"\"Run all of the tests for the package.\n\n This is a automatic test run class to make distutils kind of act like\n perl. With this you can do:\n\n python setup.py build\n python setup.py install\n python setup.py test\n\n \"\"\"\n\n description = \"Automatically run the test suite for Biopython.\"\n user_options = [(\"offline\", None, \"Don't run online tests\")]\n\n def initialize_options(self):\n \"\"\"No-op, initialise options.\"\"\"\n self.offline = None\n\n def finalize_options(self):\n \"\"\"No-op, finalise options.\"\"\"\n pass\n\n def run(self):\n \"\"\"Run the tests.\"\"\"\n this_dir = os.getcwd()\n\n # change to the test dir and run the tests\n os.chdir(\"Tests\")\n sys.path.insert(0, \"\")\n import run_tests\n\n if self.offline:\n run_tests.main([\"--offline\"])\n else:\n run_tests.main([])\n\n # change back to the current directory\n os.chdir(this_dir)\n\n\ndef can_import(module_name):\n \"\"\"Check we can import the requested module.\"\"\"\n try:\n return __import__(module_name)\n except ImportError:\n return None\n\n\n# Using requirements.txt is preferred for an application\n# (and likely will pin specific version numbers), using\n# setup.py's install_requires is preferred for a library\n# (and should try not to be overly narrow with versions).\nREQUIRES = [\"numpy\"]\n\n# --- set up the packages we are going to install\n# standard biopython packages\nPACKAGES = [\n \"Bio\",\n \"Bio.Affy\",\n \"Bio.Align\",\n \"Bio.Align.Applications\",\n \"Bio.Align.substitution_matrices\",\n \"Bio.AlignIO\",\n \"Bio.Alphabet\",\n \"Bio.Application\",\n \"Bio.Blast\",\n \"Bio.CAPS\",\n \"Bio.Cluster\",\n \"Bio.codonalign\",\n \"Bio.Compass\",\n \"Bio.Data\",\n \"Bio.Emboss\",\n \"Bio.Entrez\",\n \"Bio.ExPASy\",\n \"Bio.GenBank\",\n \"Bio.Geo\",\n \"Bio.Graphics\",\n \"Bio.Graphics.GenomeDiagram\",\n \"Bio.HMM\",\n \"Bio.KEGG\",\n \"Bio.KEGG.Compound\",\n \"Bio.KEGG.Enzyme\",\n \"Bio.KEGG.Gene\",\n \"Bio.KEGG.Map\",\n \"Bio.PDB.mmtf\",\n \"Bio.KEGG.KGML\",\n \"Bio.Medline\",\n \"Bio.motifs\",\n \"Bio.motifs.applications\",\n \"Bio.motifs.jaspar\",\n \"Bio.Nexus\",\n \"Bio.NMR\",\n \"Bio.Pathway\",\n \"Bio.Pathway.Rep\",\n \"Bio.PDB\",\n \"Bio.phenotype\",\n \"Bio.PopGen\",\n \"Bio.PopGen.GenePop\",\n \"Bio.Restriction\",\n \"Bio.SCOP\",\n \"Bio.SearchIO\",\n \"Bio.SearchIO._legacy\",\n \"Bio.SearchIO._model\",\n \"Bio.SearchIO.BlastIO\",\n \"Bio.SearchIO.HHsuiteIO\",\n \"Bio.SearchIO.HmmerIO\",\n \"Bio.SearchIO.ExonerateIO\",\n \"Bio.SearchIO.InterproscanIO\",\n \"Bio.SeqIO\",\n \"Bio.SeqUtils\",\n \"Bio.Sequencing\",\n \"Bio.Sequencing.Applications\",\n \"Bio.SVDSuperimposer\",\n \"Bio.PDB.QCPSuperimposer\",\n \"Bio.SwissProt\",\n \"Bio.TogoWS\",\n \"Bio.Phylo\",\n \"Bio.Phylo.Applications\",\n \"Bio.Phylo.PAML\",\n \"Bio.UniGene\",\n \"Bio.UniProt\",\n \"Bio.Wise\",\n # Other top level packages,\n \"BioSQL\",\n]\n\nEXTENSIONS = [\n Extension(\"Bio.Align._aligners\", [\"Bio/Align/_aligners.c\"]),\n Extension(\"Bio.cpairwise2\", [\"Bio/cpairwise2module.c\"]),\n Extension(\"Bio.Nexus.cnexus\", [\"Bio/Nexus/cnexus.c\"]),\n Extension(\n \"Bio.PDB.QCPSuperimposer.qcprotmodule\",\n [\"Bio/PDB/QCPSuperimposer/qcprotmodule.c\"],\n ),\n Extension(\"Bio.motifs._pwm\", [\"Bio/motifs/_pwm.c\"]),\n Extension(\n \"Bio.Cluster._cluster\", [\"Bio/Cluster/cluster.c\", \"Bio/Cluster/clustermodule.c\"]\n ),\n Extension(\"Bio.PDB.kdtrees\", [\"Bio/PDB/kdtrees.c\"]),\n Extension(\"Bio.PDB.ccealign\", [\"Bio/PDB/ccealignmodule.c\"]),\n Extension(\"Bio.SeqIO._twoBitIO\", [\"Bio/SeqIO/_twoBitIO.c\"]),\n]\n\n# We now define the Biopython version number in Bio/__init__.py\n# Here we can't use \"import Bio\" then \"Bio.__version__\" as that would\n# tell us the version of Biopython already installed (if any).\n__version__ = \"Undefined\"\nfor line in open(\"Bio/__init__.py\"):\n if line.startswith(\"__version__\"):\n exec(line.strip())\n\n# We now load in our reStructuredText README.rst file to pass explicitly in the\n# metadata, since at time of writing PyPI did not do this for us.\n#\n# Must make encoding explicit to avoid any conflict with the local default.\n# Currently keeping README as ASCII (might switch to UTF8 later if needed).\n# If any invalid character does appear in README, this will fail and alert us.\nwith open(\"README.rst\", encoding=\"ascii\") as handle:\n readme_rst = handle.read()\n\nsetup(\n name=\"biopython\",\n version=__version__,\n author=\"The Biopython Contributors\",\n author_email=\"[email protected]\",\n url=\"https://biopython.org/\",\n description=\"Freely available tools for computational molecular biology.\",\n long_description=readme_rst,\n project_urls={\n \"Documentation\": \"https://biopython.org/wiki/Documentation\",\n \"Source\": \"https://github.com/biopython/biopython/\",\n \"Tracker\": \"https://github.com/biopython/biopython/issues\",\n },\n classifiers=[\n \"Development Status :: 5 - Production/Stable\",\n \"Intended Audience :: Developers\",\n \"Intended Audience :: Science/Research\",\n \"License :: Freely Distributable\",\n # Technically the \"Biopython License Agreement\" is not OSI approved,\n # but is almost https://opensource.org/licenses/HPND so might put:\n # 'License :: OSI Approved',\n # To resolve this we are moving to dual-licensing with 3-clause BSD:\n # 'License :: OSI Approved :: BSD License',\n \"Operating System :: OS Independent\",\n \"Programming Language :: Python\",\n \"Programming Language :: Python :: 3\",\n \"Programming Language :: Python :: 3.8\",\n \"Programming Language :: Python :: 3.9\",\n \"Programming Language :: Python :: 3.10\",\n \"Programming Language :: Python :: 3.11\",\n \"Topic :: Scientific/Engineering\",\n \"Topic :: Scientific/Engineering :: Bio-Informatics\",\n \"Topic :: Software Development :: Libraries :: Python Modules\",\n ],\n cmdclass={\"test\": test_biopython},\n packages=PACKAGES,\n ext_modules=EXTENSIONS,\n include_package_data=True, # done via MANIFEST.in under setuptools\n install_requires=REQUIRES,\n python_requires=\">=%i.%i\" % MIN_PY_VER,\n)\n", "path": "setup.py"}]}
| 3,686 | 291 |
gh_patches_debug_15
|
rasdani/github-patches
|
git_diff
|
OCHA-DAP__hdx-ckan-1748
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Shrink the spacing on the top line numbers
Proposed spacings shown here:

modified css:
.item-info {
border-top: 1px solid #cccccc;
border-bottom: 1px solid #cccccc;
padding: 20px 0;
margin-top: -1px;
color: #333333;
}
.item-info .item-info-title {
font-family: 'Gotham-Bold', sans-serif;
font-weight: 400;
font-size: 16px;
letter-spacing: 0.01em;
margin-bottom: 20px;
}
.item-info .item-info-number {
font-family: 'Gotham-Light', sans-serif;
font-size: 74px;
line-height: 1;
letter-spacing: 0.01em;
margin-bottom: 20px;
}
</issue>
<code>
[start of ckanext-hdx_theme/ckanext/hdx_theme/version.py]
1 hdx_version = 'v0.4.9'
2
[end of ckanext-hdx_theme/ckanext/hdx_theme/version.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/ckanext-hdx_theme/ckanext/hdx_theme/version.py b/ckanext-hdx_theme/ckanext/hdx_theme/version.py
--- a/ckanext-hdx_theme/ckanext/hdx_theme/version.py
+++ b/ckanext-hdx_theme/ckanext/hdx_theme/version.py
@@ -1 +1 @@
-hdx_version = 'v0.4.9'
+hdx_version = 'v0.4.10'
|
{"golden_diff": "diff --git a/ckanext-hdx_theme/ckanext/hdx_theme/version.py b/ckanext-hdx_theme/ckanext/hdx_theme/version.py\n--- a/ckanext-hdx_theme/ckanext/hdx_theme/version.py\n+++ b/ckanext-hdx_theme/ckanext/hdx_theme/version.py\n@@ -1 +1 @@\n-hdx_version = 'v0.4.9'\n+hdx_version = 'v0.4.10'\n", "issue": "Shrink the spacing on the top line numbers\nProposed spacings shown here:\n\n\n\nmodified css:\n\n.item-info {\nborder-top: 1px solid #cccccc;\nborder-bottom: 1px solid #cccccc;\npadding: 20px 0;\nmargin-top: -1px;\ncolor: #333333;\n}\n\n.item-info .item-info-title {\nfont-family: 'Gotham-Bold', sans-serif;\nfont-weight: 400;\nfont-size: 16px;\nletter-spacing: 0.01em;\nmargin-bottom: 20px;\n}\n\n.item-info .item-info-number {\nfont-family: 'Gotham-Light', sans-serif;\nfont-size: 74px;\nline-height: 1;\nletter-spacing: 0.01em;\nmargin-bottom: 20px;\n}\n\n", "before_files": [{"content": "hdx_version = 'v0.4.9'\n", "path": "ckanext-hdx_theme/ckanext/hdx_theme/version.py"}]}
| 812 | 107 |
gh_patches_debug_6535
|
rasdani/github-patches
|
git_diff
|
open-telemetry__opentelemetry-python-2745
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
AttributeError: 'NoneType' object has no attribute 'attributes'
Run a slightly modified version of the [docs example](https://github.com/open-telemetry/opentelemetry-python/tree/main/docs/examples/metrics):
``` python
from typing import Iterable
from time import sleep
from opentelemetry.exporter.otlp.proto.grpc.metric_exporter import (
OTLPMetricExporter,
)
from opentelemetry.metrics import (
CallbackOptions,
Observation,
get_meter_provider,
set_meter_provider,
)
from opentelemetry.sdk.metrics import MeterProvider
from opentelemetry.sdk.metrics.export import PeriodicExportingMetricReader
exporter = OTLPMetricExporter(insecure=True)
reader = PeriodicExportingMetricReader(exporter, export_interval_millis=1000)
provider = MeterProvider(metric_readers=[reader])
set_meter_provider(provider)
def observable_counter_func(options: CallbackOptions) -> Iterable[Observation]:
yield Observation(1, {})
def observable_up_down_counter_func(
options: CallbackOptions,
) -> Iterable[Observation]:
yield Observation(-10, {})
def observable_gauge_func(options: CallbackOptions) -> Iterable[Observation]:
yield Observation(9, {})
meter = get_meter_provider().get_meter("getting-started", "0.1.2")
# Counter
counter = meter.create_counter("counter")
counter.add(1)
# Async Counter
observable_counter = meter.create_observable_counter(
"observable_counter",
[observable_counter_func],
)
# UpDownCounter
updown_counter = meter.create_up_down_counter("updown_counter")
updown_counter.add(1)
updown_counter.add(-5)
# Async UpDownCounter
observable_updown_counter = meter.create_observable_up_down_counter(
"observable_updown_counter", [observable_up_down_counter_func]
)
# Histogram
histogram = meter.create_histogram("histogram")
histogram.record(99.9)
# Async Gauge
gauge = meter.create_observable_gauge("gauge", [observable_gauge_func])
sleep(6)
```
running this script (after activating the collector via Docker) produces these errors:
```
Exception while exporting metrics 'NoneType' object has no attribute 'attributes'
Traceback (most recent call last):
File "/home/ocelotl/github/ocelotl/opentelemetry-python/opentelemetry-sdk/src/opentelemetry/sdk/metrics/_internal/export/__init__.py", line 412, in _receive_metrics
self._exporter.export(metrics_data, timeout_millis=timeout_millis)
File "/home/ocelotl/github/ocelotl/opentelemetry-python/exporter/opentelemetry-exporter-otlp-proto-grpc/src/opentelemetry/exporter/otlp/proto/grpc/metric_exporter/__init__.py", line 203, in export
return self._export(metrics_data)
File "/home/ocelotl/github/ocelotl/opentelemetry-python/exporter/opentelemetry-exporter-otlp-proto-grpc/src/opentelemetry/exporter/otlp/proto/grpc/exporter.py", line 293, in _export
request=self._translate_data(data),
File "/home/ocelotl/github/ocelotl/opentelemetry-python/exporter/opentelemetry-exporter-otlp-proto-grpc/src/opentelemetry/exporter/otlp/proto/grpc/metric_exporter/__init__.py", line 139, in _translate_data
data_point.attributes
AttributeError: 'NoneType' object has no attribute 'attributes'
Exception while exporting metrics 'NoneType' object has no attribute 'attributes'
Traceback (most recent call last):
File "/home/ocelotl/github/ocelotl/opentelemetry-python/opentelemetry-sdk/src/opentelemetry/sdk/metrics/_internal/export/__init__.py", line 412, in _receive_metrics
self._exporter.export(metrics_data, timeout_millis=timeout_millis)
File "/home/ocelotl/github/ocelotl/opentelemetry-python/exporter/opentelemetry-exporter-otlp-proto-grpc/src/opentelemetry/exporter/otlp/proto/grpc/metric_exporter/__init__.py", line 203, in export
return self._export(metrics_data)
File "/home/ocelotl/github/ocelotl/opentelemetry-python/exporter/opentelemetry-exporter-otlp-proto-grpc/src/opentelemetry/exporter/otlp/proto/grpc/exporter.py", line 293, in _export
request=self._translate_data(data),
File "/home/ocelotl/github/ocelotl/opentelemetry-python/exporter/opentelemetry-exporter-otlp-proto-grpc/src/opentelemetry/exporter/otlp/proto/grpc/metric_exporter/__init__.py", line 139, in _translate_data
data_point.attributes
AttributeError: 'NoneType' object has no attribute 'attributes'
Exception while exporting metrics 'NoneType' object has no attribute 'attributes'
Traceback (most recent call last):
File "/home/ocelotl/github/ocelotl/opentelemetry-python/opentelemetry-sdk/src/opentelemetry/sdk/metrics/_internal/export/__init__.py", line 412, in _receive_metrics
self._exporter.export(metrics_data, timeout_millis=timeout_millis)
File "/home/ocelotl/github/ocelotl/opentelemetry-python/exporter/opentelemetry-exporter-otlp-proto-grpc/src/opentelemetry/exporter/otlp/proto/grpc/metric_exporter/__init__.py", line 203, in export
return self._export(metrics_data)
File "/home/ocelotl/github/ocelotl/opentelemetry-python/exporter/opentelemetry-exporter-otlp-proto-grpc/src/opentelemetry/exporter/otlp/proto/grpc/exporter.py", line 293, in _export
request=self._translate_data(data),
File "/home/ocelotl/github/ocelotl/opentelemetry-python/exporter/opentelemetry-exporter-otlp-proto-grpc/src/opentelemetry/exporter/otlp/proto/grpc/metric_exporter/__init__.py", line 139, in _translate_data
data_point.attributes
AttributeError: 'NoneType' object has no attribute 'attributes'
Exception while exporting metrics 'NoneType' object has no attribute 'attributes'
Traceback (most recent call last):
File "/home/ocelotl/github/ocelotl/opentelemetry-python/opentelemetry-sdk/src/opentelemetry/sdk/metrics/_internal/export/__init__.py", line 412, in _receive_metrics
self._exporter.export(metrics_data, timeout_millis=timeout_millis)
File "/home/ocelotl/github/ocelotl/opentelemetry-python/exporter/opentelemetry-exporter-otlp-proto-grpc/src/opentelemetry/exporter/otlp/proto/grpc/metric_exporter/__init__.py", line 203, in export
return self._export(metrics_data)
File "/home/ocelotl/github/ocelotl/opentelemetry-python/exporter/opentelemetry-exporter-otlp-proto-grpc/src/opentelemetry/exporter/otlp/proto/grpc/exporter.py", line 293, in _export
request=self._translate_data(data),
File "/home/ocelotl/github/ocelotl/opentelemetry-python/exporter/opentelemetry-exporter-otlp-proto-grpc/src/opentelemetry/exporter/otlp/proto/grpc/metric_exporter/__init__.py", line 139, in _translate_data
data_point.attributes
AttributeError: 'NoneType' object has no attribute 'attributes'
Exception while exporting metrics 'NoneType' object has no attribute 'attributes'
Traceback (most recent call last):
File "/home/ocelotl/github/ocelotl/opentelemetry-python/opentelemetry-sdk/src/opentelemetry/sdk/metrics/_internal/export/__init__.py", line 412, in _receive_metrics
self._exporter.export(metrics_data, timeout_millis=timeout_millis)
File "/home/ocelotl/github/ocelotl/opentelemetry-python/exporter/opentelemetry-exporter-otlp-proto-grpc/src/opentelemetry/exporter/otlp/proto/grpc/metric_exporter/__init__.py", line 203, in export
return self._export(metrics_data)
File "/home/ocelotl/github/ocelotl/opentelemetry-python/exporter/opentelemetry-exporter-otlp-proto-grpc/src/opentelemetry/exporter/otlp/proto/grpc/exporter.py", line 293, in _export
request=self._translate_data(data),
File "/home/ocelotl/github/ocelotl/opentelemetry-python/exporter/opentelemetry-exporter-otlp-proto-grpc/src/opentelemetry/exporter/otlp/proto/grpc/metric_exporter/__init__.py", line 139, in _translate_data
data_point.attributes
AttributeError: 'NoneType' object has no attribute 'attributes'
shutdown
```
</issue>
<code>
[start of opentelemetry-sdk/src/opentelemetry/sdk/metrics/_internal/_view_instrument_match.py]
1 # Copyright The OpenTelemetry Authors
2 #
3 # Licensed under the Apache License, Version 2.0 (the "License");
4 # you may not use this file except in compliance with the License.
5 # You may obtain a copy of the License at
6 #
7 # http://www.apache.org/licenses/LICENSE-2.0
8 #
9 # Unless required by applicable law or agreed to in writing, software
10 # distributed under the License is distributed on an "AS IS" BASIS,
11 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12 # See the License for the specific language governing permissions and
13 # limitations under the License.
14
15
16 from logging import getLogger
17 from threading import Lock
18 from typing import Dict, Iterable
19
20 from opentelemetry.metrics import Instrument
21 from opentelemetry.sdk.metrics._internal.aggregation import (
22 Aggregation,
23 DefaultAggregation,
24 _Aggregation,
25 _SumAggregation,
26 )
27 from opentelemetry.sdk.metrics._internal.export import AggregationTemporality
28 from opentelemetry.sdk.metrics._internal.measurement import Measurement
29 from opentelemetry.sdk.metrics._internal.point import DataPointT
30 from opentelemetry.sdk.metrics._internal.view import View
31 from opentelemetry.util._time import _time_ns
32
33 _logger = getLogger(__name__)
34
35
36 class _ViewInstrumentMatch:
37 def __init__(
38 self,
39 view: View,
40 instrument: Instrument,
41 instrument_class_aggregation: Dict[type, Aggregation],
42 ):
43 self._start_time_unix_nano = _time_ns()
44 self._view = view
45 self._instrument = instrument
46 self._attributes_aggregation: Dict[frozenset, _Aggregation] = {}
47 self._lock = Lock()
48 self._instrument_class_aggregation = instrument_class_aggregation
49 self._name = self._view._name or self._instrument.name
50 self._description = (
51 self._view._description or self._instrument.description
52 )
53 if not isinstance(self._view._aggregation, DefaultAggregation):
54 self._aggregation = self._view._aggregation._create_aggregation(
55 self._instrument, None, 0
56 )
57 else:
58 self._aggregation = self._instrument_class_aggregation[
59 self._instrument.__class__
60 ]._create_aggregation(self._instrument, None, 0)
61
62 def conflicts(self, other: "_ViewInstrumentMatch") -> bool:
63 # pylint: disable=protected-access
64
65 result = (
66 self._name == other._name
67 and self._instrument.unit == other._instrument.unit
68 # The aggregation class is being used here instead of data point
69 # type since they are functionally equivalent.
70 and self._aggregation.__class__ == other._aggregation.__class__
71 )
72 if isinstance(self._aggregation, _SumAggregation):
73 result = (
74 result
75 and self._aggregation._instrument_is_monotonic
76 == other._aggregation._instrument_is_monotonic
77 and self._aggregation._instrument_temporality
78 == other._aggregation._instrument_temporality
79 )
80
81 return result
82
83 # pylint: disable=protected-access
84 def consume_measurement(self, measurement: Measurement) -> None:
85
86 if self._view._attribute_keys is not None:
87
88 attributes = {}
89
90 for key, value in (measurement.attributes or {}).items():
91 if key in self._view._attribute_keys:
92 attributes[key] = value
93 elif measurement.attributes is not None:
94 attributes = measurement.attributes
95 else:
96 attributes = {}
97
98 aggr_key = frozenset(attributes.items())
99
100 if aggr_key not in self._attributes_aggregation:
101 with self._lock:
102 if aggr_key not in self._attributes_aggregation:
103 if not isinstance(
104 self._view._aggregation, DefaultAggregation
105 ):
106 aggregation = (
107 self._view._aggregation._create_aggregation(
108 self._instrument,
109 attributes,
110 self._start_time_unix_nano,
111 )
112 )
113 else:
114 aggregation = self._instrument_class_aggregation[
115 self._instrument.__class__
116 ]._create_aggregation(
117 self._instrument,
118 attributes,
119 self._start_time_unix_nano,
120 )
121 self._attributes_aggregation[aggr_key] = aggregation
122
123 self._attributes_aggregation[aggr_key].aggregate(measurement)
124
125 def collect(
126 self,
127 aggregation_temporality: AggregationTemporality,
128 collection_start_nanos: int,
129 ) -> Iterable[DataPointT]:
130
131 with self._lock:
132 for aggregation in self._attributes_aggregation.values():
133 yield aggregation.collect(
134 aggregation_temporality, collection_start_nanos
135 )
136
[end of opentelemetry-sdk/src/opentelemetry/sdk/metrics/_internal/_view_instrument_match.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/opentelemetry-sdk/src/opentelemetry/sdk/metrics/_internal/_view_instrument_match.py b/opentelemetry-sdk/src/opentelemetry/sdk/metrics/_internal/_view_instrument_match.py
--- a/opentelemetry-sdk/src/opentelemetry/sdk/metrics/_internal/_view_instrument_match.py
+++ b/opentelemetry-sdk/src/opentelemetry/sdk/metrics/_internal/_view_instrument_match.py
@@ -130,6 +130,8 @@
with self._lock:
for aggregation in self._attributes_aggregation.values():
- yield aggregation.collect(
+ data_point = aggregation.collect(
aggregation_temporality, collection_start_nanos
)
+ if data_point is not None:
+ yield data_point
|
{"golden_diff": "diff --git a/opentelemetry-sdk/src/opentelemetry/sdk/metrics/_internal/_view_instrument_match.py b/opentelemetry-sdk/src/opentelemetry/sdk/metrics/_internal/_view_instrument_match.py\n--- a/opentelemetry-sdk/src/opentelemetry/sdk/metrics/_internal/_view_instrument_match.py\n+++ b/opentelemetry-sdk/src/opentelemetry/sdk/metrics/_internal/_view_instrument_match.py\n@@ -130,6 +130,8 @@\n \n with self._lock:\n for aggregation in self._attributes_aggregation.values():\n- yield aggregation.collect(\n+ data_point = aggregation.collect(\n aggregation_temporality, collection_start_nanos\n )\n+ if data_point is not None:\n+ yield data_point\n", "issue": "AttributeError: 'NoneType' object has no attribute 'attributes'\nRun a slightly modified version of the [docs example](https://github.com/open-telemetry/opentelemetry-python/tree/main/docs/examples/metrics):\r\n\r\n``` python\r\nfrom typing import Iterable\r\nfrom time import sleep\r\n\r\nfrom opentelemetry.exporter.otlp.proto.grpc.metric_exporter import (\r\n OTLPMetricExporter,\r\n)\r\nfrom opentelemetry.metrics import (\r\n CallbackOptions,\r\n Observation,\r\n get_meter_provider,\r\n set_meter_provider,\r\n)\r\nfrom opentelemetry.sdk.metrics import MeterProvider\r\nfrom opentelemetry.sdk.metrics.export import PeriodicExportingMetricReader\r\n\r\nexporter = OTLPMetricExporter(insecure=True)\r\nreader = PeriodicExportingMetricReader(exporter, export_interval_millis=1000)\r\nprovider = MeterProvider(metric_readers=[reader])\r\nset_meter_provider(provider)\r\n\r\n\r\ndef observable_counter_func(options: CallbackOptions) -> Iterable[Observation]:\r\n yield Observation(1, {})\r\n\r\n\r\ndef observable_up_down_counter_func(\r\n options: CallbackOptions,\r\n) -> Iterable[Observation]:\r\n yield Observation(-10, {})\r\n\r\n\r\ndef observable_gauge_func(options: CallbackOptions) -> Iterable[Observation]:\r\n yield Observation(9, {})\r\n\r\n\r\nmeter = get_meter_provider().get_meter(\"getting-started\", \"0.1.2\")\r\n\r\n# Counter\r\ncounter = meter.create_counter(\"counter\")\r\ncounter.add(1)\r\n\r\n# Async Counter\r\nobservable_counter = meter.create_observable_counter(\r\n \"observable_counter\",\r\n [observable_counter_func],\r\n)\r\n\r\n# UpDownCounter\r\nupdown_counter = meter.create_up_down_counter(\"updown_counter\")\r\nupdown_counter.add(1)\r\nupdown_counter.add(-5)\r\n\r\n# Async UpDownCounter\r\nobservable_updown_counter = meter.create_observable_up_down_counter(\r\n \"observable_updown_counter\", [observable_up_down_counter_func]\r\n)\r\n\r\n# Histogram\r\nhistogram = meter.create_histogram(\"histogram\")\r\nhistogram.record(99.9)\r\n\r\n# Async Gauge\r\ngauge = meter.create_observable_gauge(\"gauge\", [observable_gauge_func])\r\n\r\nsleep(6)\r\n```\r\nrunning this script (after activating the collector via Docker) produces these errors:\r\n\r\n```\r\nException while exporting metrics 'NoneType' object has no attribute 'attributes'\r\nTraceback (most recent call last):\r\n File \"/home/ocelotl/github/ocelotl/opentelemetry-python/opentelemetry-sdk/src/opentelemetry/sdk/metrics/_internal/export/__init__.py\", line 412, in _receive_metrics\r\n self._exporter.export(metrics_data, timeout_millis=timeout_millis)\r\n File \"/home/ocelotl/github/ocelotl/opentelemetry-python/exporter/opentelemetry-exporter-otlp-proto-grpc/src/opentelemetry/exporter/otlp/proto/grpc/metric_exporter/__init__.py\", line 203, in export\r\n return self._export(metrics_data)\r\n File \"/home/ocelotl/github/ocelotl/opentelemetry-python/exporter/opentelemetry-exporter-otlp-proto-grpc/src/opentelemetry/exporter/otlp/proto/grpc/exporter.py\", line 293, in _export\r\n request=self._translate_data(data),\r\n File \"/home/ocelotl/github/ocelotl/opentelemetry-python/exporter/opentelemetry-exporter-otlp-proto-grpc/src/opentelemetry/exporter/otlp/proto/grpc/metric_exporter/__init__.py\", line 139, in _translate_data\r\n data_point.attributes\r\nAttributeError: 'NoneType' object has no attribute 'attributes'\r\nException while exporting metrics 'NoneType' object has no attribute 'attributes'\r\nTraceback (most recent call last):\r\n File \"/home/ocelotl/github/ocelotl/opentelemetry-python/opentelemetry-sdk/src/opentelemetry/sdk/metrics/_internal/export/__init__.py\", line 412, in _receive_metrics\r\n self._exporter.export(metrics_data, timeout_millis=timeout_millis)\r\n File \"/home/ocelotl/github/ocelotl/opentelemetry-python/exporter/opentelemetry-exporter-otlp-proto-grpc/src/opentelemetry/exporter/otlp/proto/grpc/metric_exporter/__init__.py\", line 203, in export\r\n return self._export(metrics_data)\r\n File \"/home/ocelotl/github/ocelotl/opentelemetry-python/exporter/opentelemetry-exporter-otlp-proto-grpc/src/opentelemetry/exporter/otlp/proto/grpc/exporter.py\", line 293, in _export\r\n request=self._translate_data(data),\r\n File \"/home/ocelotl/github/ocelotl/opentelemetry-python/exporter/opentelemetry-exporter-otlp-proto-grpc/src/opentelemetry/exporter/otlp/proto/grpc/metric_exporter/__init__.py\", line 139, in _translate_data\r\n data_point.attributes\r\nAttributeError: 'NoneType' object has no attribute 'attributes'\r\nException while exporting metrics 'NoneType' object has no attribute 'attributes'\r\nTraceback (most recent call last):\r\n File \"/home/ocelotl/github/ocelotl/opentelemetry-python/opentelemetry-sdk/src/opentelemetry/sdk/metrics/_internal/export/__init__.py\", line 412, in _receive_metrics\r\n self._exporter.export(metrics_data, timeout_millis=timeout_millis)\r\n File \"/home/ocelotl/github/ocelotl/opentelemetry-python/exporter/opentelemetry-exporter-otlp-proto-grpc/src/opentelemetry/exporter/otlp/proto/grpc/metric_exporter/__init__.py\", line 203, in export\r\n return self._export(metrics_data)\r\n File \"/home/ocelotl/github/ocelotl/opentelemetry-python/exporter/opentelemetry-exporter-otlp-proto-grpc/src/opentelemetry/exporter/otlp/proto/grpc/exporter.py\", line 293, in _export\r\n request=self._translate_data(data),\r\n File \"/home/ocelotl/github/ocelotl/opentelemetry-python/exporter/opentelemetry-exporter-otlp-proto-grpc/src/opentelemetry/exporter/otlp/proto/grpc/metric_exporter/__init__.py\", line 139, in _translate_data\r\n data_point.attributes\r\nAttributeError: 'NoneType' object has no attribute 'attributes'\r\nException while exporting metrics 'NoneType' object has no attribute 'attributes'\r\nTraceback (most recent call last):\r\n File \"/home/ocelotl/github/ocelotl/opentelemetry-python/opentelemetry-sdk/src/opentelemetry/sdk/metrics/_internal/export/__init__.py\", line 412, in _receive_metrics\r\n self._exporter.export(metrics_data, timeout_millis=timeout_millis)\r\n File \"/home/ocelotl/github/ocelotl/opentelemetry-python/exporter/opentelemetry-exporter-otlp-proto-grpc/src/opentelemetry/exporter/otlp/proto/grpc/metric_exporter/__init__.py\", line 203, in export\r\n return self._export(metrics_data)\r\n File \"/home/ocelotl/github/ocelotl/opentelemetry-python/exporter/opentelemetry-exporter-otlp-proto-grpc/src/opentelemetry/exporter/otlp/proto/grpc/exporter.py\", line 293, in _export\r\n request=self._translate_data(data),\r\n File \"/home/ocelotl/github/ocelotl/opentelemetry-python/exporter/opentelemetry-exporter-otlp-proto-grpc/src/opentelemetry/exporter/otlp/proto/grpc/metric_exporter/__init__.py\", line 139, in _translate_data\r\n data_point.attributes\r\nAttributeError: 'NoneType' object has no attribute 'attributes'\r\nException while exporting metrics 'NoneType' object has no attribute 'attributes'\r\nTraceback (most recent call last):\r\n File \"/home/ocelotl/github/ocelotl/opentelemetry-python/opentelemetry-sdk/src/opentelemetry/sdk/metrics/_internal/export/__init__.py\", line 412, in _receive_metrics\r\n self._exporter.export(metrics_data, timeout_millis=timeout_millis)\r\n File \"/home/ocelotl/github/ocelotl/opentelemetry-python/exporter/opentelemetry-exporter-otlp-proto-grpc/src/opentelemetry/exporter/otlp/proto/grpc/metric_exporter/__init__.py\", line 203, in export\r\n return self._export(metrics_data)\r\n File \"/home/ocelotl/github/ocelotl/opentelemetry-python/exporter/opentelemetry-exporter-otlp-proto-grpc/src/opentelemetry/exporter/otlp/proto/grpc/exporter.py\", line 293, in _export\r\n request=self._translate_data(data),\r\n File \"/home/ocelotl/github/ocelotl/opentelemetry-python/exporter/opentelemetry-exporter-otlp-proto-grpc/src/opentelemetry/exporter/otlp/proto/grpc/metric_exporter/__init__.py\", line 139, in _translate_data\r\n data_point.attributes\r\nAttributeError: 'NoneType' object has no attribute 'attributes'\r\nshutdown\r\n```\r\n\n", "before_files": [{"content": "# Copyright The OpenTelemetry Authors\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\n\nfrom logging import getLogger\nfrom threading import Lock\nfrom typing import Dict, Iterable\n\nfrom opentelemetry.metrics import Instrument\nfrom opentelemetry.sdk.metrics._internal.aggregation import (\n Aggregation,\n DefaultAggregation,\n _Aggregation,\n _SumAggregation,\n)\nfrom opentelemetry.sdk.metrics._internal.export import AggregationTemporality\nfrom opentelemetry.sdk.metrics._internal.measurement import Measurement\nfrom opentelemetry.sdk.metrics._internal.point import DataPointT\nfrom opentelemetry.sdk.metrics._internal.view import View\nfrom opentelemetry.util._time import _time_ns\n\n_logger = getLogger(__name__)\n\n\nclass _ViewInstrumentMatch:\n def __init__(\n self,\n view: View,\n instrument: Instrument,\n instrument_class_aggregation: Dict[type, Aggregation],\n ):\n self._start_time_unix_nano = _time_ns()\n self._view = view\n self._instrument = instrument\n self._attributes_aggregation: Dict[frozenset, _Aggregation] = {}\n self._lock = Lock()\n self._instrument_class_aggregation = instrument_class_aggregation\n self._name = self._view._name or self._instrument.name\n self._description = (\n self._view._description or self._instrument.description\n )\n if not isinstance(self._view._aggregation, DefaultAggregation):\n self._aggregation = self._view._aggregation._create_aggregation(\n self._instrument, None, 0\n )\n else:\n self._aggregation = self._instrument_class_aggregation[\n self._instrument.__class__\n ]._create_aggregation(self._instrument, None, 0)\n\n def conflicts(self, other: \"_ViewInstrumentMatch\") -> bool:\n # pylint: disable=protected-access\n\n result = (\n self._name == other._name\n and self._instrument.unit == other._instrument.unit\n # The aggregation class is being used here instead of data point\n # type since they are functionally equivalent.\n and self._aggregation.__class__ == other._aggregation.__class__\n )\n if isinstance(self._aggregation, _SumAggregation):\n result = (\n result\n and self._aggregation._instrument_is_monotonic\n == other._aggregation._instrument_is_monotonic\n and self._aggregation._instrument_temporality\n == other._aggregation._instrument_temporality\n )\n\n return result\n\n # pylint: disable=protected-access\n def consume_measurement(self, measurement: Measurement) -> None:\n\n if self._view._attribute_keys is not None:\n\n attributes = {}\n\n for key, value in (measurement.attributes or {}).items():\n if key in self._view._attribute_keys:\n attributes[key] = value\n elif measurement.attributes is not None:\n attributes = measurement.attributes\n else:\n attributes = {}\n\n aggr_key = frozenset(attributes.items())\n\n if aggr_key not in self._attributes_aggregation:\n with self._lock:\n if aggr_key not in self._attributes_aggregation:\n if not isinstance(\n self._view._aggregation, DefaultAggregation\n ):\n aggregation = (\n self._view._aggregation._create_aggregation(\n self._instrument,\n attributes,\n self._start_time_unix_nano,\n )\n )\n else:\n aggregation = self._instrument_class_aggregation[\n self._instrument.__class__\n ]._create_aggregation(\n self._instrument,\n attributes,\n self._start_time_unix_nano,\n )\n self._attributes_aggregation[aggr_key] = aggregation\n\n self._attributes_aggregation[aggr_key].aggregate(measurement)\n\n def collect(\n self,\n aggregation_temporality: AggregationTemporality,\n collection_start_nanos: int,\n ) -> Iterable[DataPointT]:\n\n with self._lock:\n for aggregation in self._attributes_aggregation.values():\n yield aggregation.collect(\n aggregation_temporality, collection_start_nanos\n )\n", "path": "opentelemetry-sdk/src/opentelemetry/sdk/metrics/_internal/_view_instrument_match.py"}]}
| 3,861 | 162 |
gh_patches_debug_24875
|
rasdani/github-patches
|
git_diff
|
coreproject-moe__CoreProject-Monorepo-3167
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
[`Frontend`] : Move code to specific `web-component`
https://github.com/baseplate-admin/CoreProject/blob/cd436b876f4936b61397a0cc838aa88125527a78/backend/django_core/templates/anime/index.html#L123-L205
</issue>
<code>
[start of backend/django_core/apps/pages/views/anime.py]
1 from typing import TYPE_CHECKING
2
3 from django.http import HttpResponse
4 from django.shortcuts import render
5
6 from ..data.anime import (
7 anime,
8 anime_episode,
9 icons,
10 latest_animes,
11 latest_episodes,
12 my_list,
13 )
14
15 if TYPE_CHECKING:
16 from ..request import HtmxHttpRequest
17
18
19 async def anime_home_view_partial_slider_view(
20 request: "HtmxHttpRequest",
21 pk: int,
22 ) -> HttpResponse:
23 anime = latest_animes[pk]
24 next_index = (pk + 1) % len(latest_animes)
25 previous_index = (pk - 1) % len(latest_animes)
26
27 return render(
28 request,
29 "anime/_slider.html",
30 context={
31 "anime": anime,
32 "next_index": next_index,
33 "previous_index": previous_index,
34 "current_index": pk,
35 },
36 )
37
38
39 async def anime_home_view(request: "HtmxHttpRequest") -> HttpResponse:
40 if request.htmx:
41 return render(
42 request,
43 "anime/index.html",
44 context={
45 "latest_animes": latest_animes,
46 "my_list": my_list,
47 "latest_episodes": latest_episodes,
48 },
49 )
50
51 return render(
52 request,
53 "anime/_layout.html",
54 context={
55 "icons": icons,
56 "latest_animes": latest_animes,
57 "my_list": my_list,
58 "latest_episodes": latest_episodes,
59 },
60 )
61
62
63 async def anime_explore_view(request: "HtmxHttpRequest") -> HttpResponse:
64 if request.htmx:
65 return render(request, "anime/explore/index.html")
66
67 return render(request, "anime/_layout.html", context={"icons": icons})
68
69
70 async def anime_info_view(
71 request: "HtmxHttpRequest",
72 platform: str,
73 pk: int,
74 ) -> HttpResponse:
75 if request.htmx:
76 return render(
77 request,
78 "anime/info/index.html",
79 context={"anime": anime, "episode": anime_episode},
80 )
81
82 return render(request, "anime/_layout.html", context={"icons": icons})
83
84
85 async def anime_episode_view(
86 request: "HtmxHttpRequest", platform: str, mal_id: int, pk: int
87 ) -> HttpResponse:
88 if request.htmx:
89 return render(
90 request,
91 "anime/episode/index.html",
92 context={},
93 )
94
95 return render(request, "anime/_layout.html", context={"icons": icons})
96
[end of backend/django_core/apps/pages/views/anime.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/backend/django_core/apps/pages/views/anime.py b/backend/django_core/apps/pages/views/anime.py
--- a/backend/django_core/apps/pages/views/anime.py
+++ b/backend/django_core/apps/pages/views/anime.py
@@ -1,3 +1,4 @@
+import json
from typing import TYPE_CHECKING
from django.http import HttpResponse
@@ -37,6 +38,9 @@
async def anime_home_view(request: "HtmxHttpRequest") -> HttpResponse:
+ # cant parse single quoted string
+ latest_episodes_json = json.dumps(latest_episodes)
+
if request.htmx:
return render(
request,
@@ -44,7 +48,7 @@
context={
"latest_animes": latest_animes,
"my_list": my_list,
- "latest_episodes": latest_episodes,
+ "latest_episodes": latest_episodes_json,
},
)
@@ -55,7 +59,7 @@
"icons": icons,
"latest_animes": latest_animes,
"my_list": my_list,
- "latest_episodes": latest_episodes,
+ "latest_episodes": latest_episodes_json,
},
)
|
{"golden_diff": "diff --git a/backend/django_core/apps/pages/views/anime.py b/backend/django_core/apps/pages/views/anime.py\n--- a/backend/django_core/apps/pages/views/anime.py\n+++ b/backend/django_core/apps/pages/views/anime.py\n@@ -1,3 +1,4 @@\n+import json\n from typing import TYPE_CHECKING\n \n from django.http import HttpResponse\n@@ -37,6 +38,9 @@\n \n \n async def anime_home_view(request: \"HtmxHttpRequest\") -> HttpResponse:\n+ # cant parse single quoted string\n+ latest_episodes_json = json.dumps(latest_episodes)\n+\n if request.htmx:\n return render(\n request,\n@@ -44,7 +48,7 @@\n context={\n \"latest_animes\": latest_animes,\n \"my_list\": my_list,\n- \"latest_episodes\": latest_episodes,\n+ \"latest_episodes\": latest_episodes_json,\n },\n )\n \n@@ -55,7 +59,7 @@\n \"icons\": icons,\n \"latest_animes\": latest_animes,\n \"my_list\": my_list,\n- \"latest_episodes\": latest_episodes,\n+ \"latest_episodes\": latest_episodes_json,\n },\n )\n", "issue": "[`Frontend`] : Move code to specific `web-component`\nhttps://github.com/baseplate-admin/CoreProject/blob/cd436b876f4936b61397a0cc838aa88125527a78/backend/django_core/templates/anime/index.html#L123-L205\n", "before_files": [{"content": "from typing import TYPE_CHECKING\n\nfrom django.http import HttpResponse\nfrom django.shortcuts import render\n\nfrom ..data.anime import (\n anime,\n anime_episode,\n icons,\n latest_animes,\n latest_episodes,\n my_list,\n)\n\nif TYPE_CHECKING:\n from ..request import HtmxHttpRequest\n\n\nasync def anime_home_view_partial_slider_view(\n request: \"HtmxHttpRequest\",\n pk: int,\n) -> HttpResponse:\n anime = latest_animes[pk]\n next_index = (pk + 1) % len(latest_animes)\n previous_index = (pk - 1) % len(latest_animes)\n\n return render(\n request,\n \"anime/_slider.html\",\n context={\n \"anime\": anime,\n \"next_index\": next_index,\n \"previous_index\": previous_index,\n \"current_index\": pk,\n },\n )\n\n\nasync def anime_home_view(request: \"HtmxHttpRequest\") -> HttpResponse:\n if request.htmx:\n return render(\n request,\n \"anime/index.html\",\n context={\n \"latest_animes\": latest_animes,\n \"my_list\": my_list,\n \"latest_episodes\": latest_episodes,\n },\n )\n\n return render(\n request,\n \"anime/_layout.html\",\n context={\n \"icons\": icons,\n \"latest_animes\": latest_animes,\n \"my_list\": my_list,\n \"latest_episodes\": latest_episodes,\n },\n )\n\n\nasync def anime_explore_view(request: \"HtmxHttpRequest\") -> HttpResponse:\n if request.htmx:\n return render(request, \"anime/explore/index.html\")\n\n return render(request, \"anime/_layout.html\", context={\"icons\": icons})\n\n\nasync def anime_info_view(\n request: \"HtmxHttpRequest\",\n platform: str,\n pk: int,\n) -> HttpResponse:\n if request.htmx:\n return render(\n request,\n \"anime/info/index.html\",\n context={\"anime\": anime, \"episode\": anime_episode},\n )\n\n return render(request, \"anime/_layout.html\", context={\"icons\": icons})\n\n\nasync def anime_episode_view(\n request: \"HtmxHttpRequest\", platform: str, mal_id: int, pk: int\n) -> HttpResponse:\n if request.htmx:\n return render(\n request,\n \"anime/episode/index.html\",\n context={},\n )\n\n return render(request, \"anime/_layout.html\", context={\"icons\": icons})\n", "path": "backend/django_core/apps/pages/views/anime.py"}]}
| 1,342 | 259 |
gh_patches_debug_23236
|
rasdani/github-patches
|
git_diff
|
getpelican__pelican-964
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
{category} in URLs isn't slugified
When trying to save articles using:
```
CATEGORY_SAVE_AS = "{slug}/index.html"
ARTICLE_SAVE_AS = "{category}/{slug}/index.html"
```
the `{category}` in the second instance isn't slugified, so the URLs don't line up neatly and may end up with spaces or other URL-unfriendly characters in them. Ditto `{author}`.
</issue>
<code>
[start of pelican/contents.py]
1 # -*- coding: utf-8 -*-
2 from __future__ import unicode_literals, print_function
3 import six
4
5 import copy
6 import locale
7 import logging
8 import functools
9 import os
10 import re
11 import sys
12
13 from datetime import datetime
14
15
16 from pelican import signals
17 from pelican.settings import DEFAULT_CONFIG
18 from pelican.utils import (slugify, truncate_html_words, memoized, strftime,
19 python_2_unicode_compatible, deprecated_attribute,
20 path_to_url)
21
22 # Import these so that they're avalaible when you import from pelican.contents.
23 from pelican.urlwrappers import (URLWrapper, Author, Category, Tag) # NOQA
24
25 logger = logging.getLogger(__name__)
26
27
28 class Content(object):
29 """Represents a content.
30
31 :param content: the string to parse, containing the original content.
32 :param metadata: the metadata associated to this page (optional).
33 :param settings: the settings dictionary (optional).
34 :param source_path: The location of the source of this content (if any).
35 :param context: The shared context between generators.
36
37 """
38 @deprecated_attribute(old='filename', new='source_path', since=(3, 2, 0))
39 def filename():
40 return None
41
42 def __init__(self, content, metadata=None, settings=None,
43 source_path=None, context=None):
44 if metadata is None:
45 metadata = {}
46 if settings is None:
47 settings = copy.deepcopy(DEFAULT_CONFIG)
48
49 self.settings = settings
50 self._content = content
51 if context is None:
52 context = {}
53 self._context = context
54 self.translations = []
55
56 local_metadata = dict(settings['DEFAULT_METADATA'])
57 local_metadata.update(metadata)
58
59 # set metadata as attributes
60 for key, value in local_metadata.items():
61 if key in ('save_as', 'url'):
62 key = 'override_' + key
63 setattr(self, key.lower(), value)
64
65 # also keep track of the metadata attributes available
66 self.metadata = local_metadata
67
68 #default template if it's not defined in page
69 self.template = self._get_template()
70
71 # default author to the one in settings if not defined
72 if not hasattr(self, 'author'):
73 if 'AUTHOR' in settings:
74 self.author = Author(settings['AUTHOR'], settings)
75
76 # XXX Split all the following code into pieces, there is too much here.
77
78 # manage languages
79 self.in_default_lang = True
80 if 'DEFAULT_LANG' in settings:
81 default_lang = settings['DEFAULT_LANG'].lower()
82 if not hasattr(self, 'lang'):
83 self.lang = default_lang
84
85 self.in_default_lang = (self.lang == default_lang)
86
87 # create the slug if not existing, from the title
88 if not hasattr(self, 'slug') and hasattr(self, 'title'):
89 self.slug = slugify(self.title,
90 settings.get('SLUG_SUBSTITUTIONS', ()))
91
92 self.source_path = source_path
93
94 # manage the date format
95 if not hasattr(self, 'date_format'):
96 if hasattr(self, 'lang') and self.lang in settings['DATE_FORMATS']:
97 self.date_format = settings['DATE_FORMATS'][self.lang]
98 else:
99 self.date_format = settings['DEFAULT_DATE_FORMAT']
100
101 if isinstance(self.date_format, tuple):
102 locale_string = self.date_format[0]
103 if sys.version_info < (3, ) and isinstance(locale_string,
104 six.text_type):
105 locale_string = locale_string.encode('ascii')
106 locale.setlocale(locale.LC_ALL, locale_string)
107 self.date_format = self.date_format[1]
108
109 if hasattr(self, 'date'):
110 self.locale_date = strftime(self.date, self.date_format)
111
112 # manage status
113 if not hasattr(self, 'status'):
114 self.status = settings['DEFAULT_STATUS']
115 if not settings['WITH_FUTURE_DATES']:
116 if hasattr(self, 'date') and self.date > datetime.now():
117 self.status = 'draft'
118
119 # store the summary metadata if it is set
120 if 'summary' in metadata:
121 self._summary = metadata['summary']
122
123 signals.content_object_init.send(self)
124
125 def __str__(self):
126 if self.source_path is None:
127 return repr(self)
128 elif six.PY3:
129 return self.source_path or repr(self)
130 else:
131 return str(self.source_path.encode('utf-8', 'replace'))
132
133 def check_properties(self):
134 """Test mandatory properties are set."""
135 for prop in self.mandatory_properties:
136 if not hasattr(self, prop):
137 raise NameError(prop)
138
139 @property
140 def url_format(self):
141 """Returns the URL, formatted with the proper values"""
142 metadata = copy.copy(self.metadata)
143 path = self.metadata.get('path', self.get_relative_source_path())
144 metadata.update({
145 'path': path_to_url(path),
146 'slug': getattr(self, 'slug', ''),
147 'lang': getattr(self, 'lang', 'en'),
148 'date': getattr(self, 'date', datetime.now()),
149 'author': getattr(self, 'author', ''),
150 'category': getattr(self, 'category',
151 self.settings['DEFAULT_CATEGORY']),
152 })
153 return metadata
154
155 def _expand_settings(self, key):
156 fq_key = ('%s_%s' % (self.__class__.__name__, key)).upper()
157 return self.settings[fq_key].format(**self.url_format)
158
159 def get_url_setting(self, key):
160 if hasattr(self, 'override_' + key):
161 return getattr(self, 'override_' + key)
162 key = key if self.in_default_lang else 'lang_%s' % key
163 return self._expand_settings(key)
164
165 def _update_content(self, content, siteurl):
166 """Update the content attribute.
167
168 Change all the relative paths of the content to relative paths
169 suitable for the ouput content.
170
171 :param content: content resource that will be passed to the templates.
172 :param siteurl: siteurl which is locally generated by the writer in
173 case of RELATIVE_URLS.
174 """
175 if not content:
176 return content
177
178 hrefs = re.compile(r"""
179 (?P<markup><\s*[^\>]* # match tag with src and href attr
180 (?:href|src)\s*=)
181
182 (?P<quote>["\']) # require value to be quoted
183 (?P<path>\|(?P<what>.*?)\|(?P<value>.*?)) # the url value
184 \2""", re.X)
185
186 def replacer(m):
187 what = m.group('what')
188 value = m.group('value')
189 origin = m.group('path')
190
191 # we support only filename for now. the plan is to support
192 # categories, tags, etc. in the future, but let's keep things
193 # simple for now.
194
195 # XXX Put this in a different location.
196 if what == 'filename':
197 if value.startswith('/'):
198 value = value[1:]
199 else:
200 # relative to the source path of this content
201 value = self.get_relative_source_path(
202 os.path.join(self.relative_dir, value)
203 )
204
205 if value in self._context['filenames']:
206 origin = '/'.join((siteurl,
207 self._context['filenames'][value].url))
208 origin = origin.replace('\\', '/') # Fow windows paths.
209 else:
210 logger.warning("Unable to find {fn}, skipping url"
211 " replacement".format(fn=value))
212
213 return ''.join((m.group('markup'), m.group('quote'), origin,
214 m.group('quote')))
215
216 return hrefs.sub(replacer, content)
217
218 @memoized
219 def get_content(self, siteurl):
220
221 if hasattr(self, '_get_content'):
222 content = self._get_content()
223 else:
224 content = self._content
225 return self._update_content(content, siteurl)
226
227 @property
228 def content(self):
229 return self.get_content(self._context.get('localsiteurl', ''))
230
231 def _get_summary(self):
232 """Returns the summary of an article.
233
234 This is based on the summary metadata if set, otherwise truncate the
235 content.
236 """
237 if hasattr(self, '_summary'):
238 return self._summary
239
240 if self.settings['SUMMARY_MAX_LENGTH'] is None:
241 return self.content
242
243 return truncate_html_words(self.content,
244 self.settings['SUMMARY_MAX_LENGTH'])
245
246 def _set_summary(self, summary):
247 """Dummy function"""
248 pass
249
250 summary = property(_get_summary, _set_summary, "Summary of the article."
251 "Based on the content. Can't be set")
252 url = property(functools.partial(get_url_setting, key='url'))
253 save_as = property(functools.partial(get_url_setting, key='save_as'))
254
255 def _get_template(self):
256 if hasattr(self, 'template') and self.template is not None:
257 return self.template
258 else:
259 return self.default_template
260
261 def get_relative_source_path(self, source_path=None):
262 """Return the relative path (from the content path) to the given
263 source_path.
264
265 If no source path is specified, use the source path of this
266 content object.
267 """
268 if not source_path:
269 source_path = self.source_path
270 if source_path is None:
271 return None
272
273 return os.path.relpath(
274 os.path.abspath(os.path.join(self.settings['PATH'], source_path)),
275 os.path.abspath(self.settings['PATH'])
276 )
277
278 @property
279 def relative_dir(self):
280 return os.path.dirname(os.path.relpath(
281 os.path.abspath(self.source_path),
282 os.path.abspath(self.settings['PATH']))
283 )
284
285
286 class Page(Content):
287 mandatory_properties = ('title',)
288 default_template = 'page'
289
290
291 class Article(Page):
292 mandatory_properties = ('title', 'date', 'category')
293 default_template = 'article'
294
295
296 class Quote(Page):
297 base_properties = ('author', 'date')
298
299
300 @python_2_unicode_compatible
301 class Static(Page):
302 @deprecated_attribute(old='filepath', new='source_path', since=(3, 2, 0))
303 def filepath():
304 return None
305
306 @deprecated_attribute(old='src', new='source_path', since=(3, 2, 0))
307 def src():
308 return None
309
310 @deprecated_attribute(old='dst', new='save_as', since=(3, 2, 0))
311 def dst():
312 return None
313
314
315 def is_valid_content(content, f):
316 try:
317 content.check_properties()
318 return True
319 except NameError as e:
320 logger.error("Skipping %s: could not find information about "
321 "'%s'" % (f, e))
322 return False
323
[end of pelican/contents.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/pelican/contents.py b/pelican/contents.py
--- a/pelican/contents.py
+++ b/pelican/contents.py
@@ -141,14 +141,21 @@
"""Returns the URL, formatted with the proper values"""
metadata = copy.copy(self.metadata)
path = self.metadata.get('path', self.get_relative_source_path())
+ default_category = self.settings['DEFAULT_CATEGORY']
+ slug_substitutions = self.settings.get('SLUG_SUBSTITUTIONS', ())
metadata.update({
'path': path_to_url(path),
'slug': getattr(self, 'slug', ''),
'lang': getattr(self, 'lang', 'en'),
'date': getattr(self, 'date', datetime.now()),
- 'author': getattr(self, 'author', ''),
- 'category': getattr(self, 'category',
- self.settings['DEFAULT_CATEGORY']),
+ 'author': slugify(
+ getattr(self, 'author', ''),
+ slug_substitutions
+ ),
+ 'category': slugify(
+ getattr(self, 'category', default_category),
+ slug_substitutions
+ )
})
return metadata
|
{"golden_diff": "diff --git a/pelican/contents.py b/pelican/contents.py\n--- a/pelican/contents.py\n+++ b/pelican/contents.py\n@@ -141,14 +141,21 @@\n \"\"\"Returns the URL, formatted with the proper values\"\"\"\n metadata = copy.copy(self.metadata)\n path = self.metadata.get('path', self.get_relative_source_path())\n+ default_category = self.settings['DEFAULT_CATEGORY']\n+ slug_substitutions = self.settings.get('SLUG_SUBSTITUTIONS', ())\n metadata.update({\n 'path': path_to_url(path),\n 'slug': getattr(self, 'slug', ''),\n 'lang': getattr(self, 'lang', 'en'),\n 'date': getattr(self, 'date', datetime.now()),\n- 'author': getattr(self, 'author', ''),\n- 'category': getattr(self, 'category',\n- self.settings['DEFAULT_CATEGORY']),\n+ 'author': slugify(\n+ getattr(self, 'author', ''),\n+ slug_substitutions\n+ ),\n+ 'category': slugify(\n+ getattr(self, 'category', default_category),\n+ slug_substitutions\n+ )\n })\n return metadata\n", "issue": "{category} in URLs isn't slugified\nWhen trying to save articles using:\n\n```\nCATEGORY_SAVE_AS = \"{slug}/index.html\" \nARTICLE_SAVE_AS = \"{category}/{slug}/index.html\"\n```\n\nthe `{category}` in the second instance isn't slugified, so the URLs don't line up neatly and may end up with spaces or other URL-unfriendly characters in them. Ditto `{author}`.\n\n", "before_files": [{"content": "# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals, print_function\nimport six\n\nimport copy\nimport locale\nimport logging\nimport functools\nimport os\nimport re\nimport sys\n\nfrom datetime import datetime\n\n\nfrom pelican import signals\nfrom pelican.settings import DEFAULT_CONFIG\nfrom pelican.utils import (slugify, truncate_html_words, memoized, strftime,\n python_2_unicode_compatible, deprecated_attribute,\n path_to_url)\n\n# Import these so that they're avalaible when you import from pelican.contents.\nfrom pelican.urlwrappers import (URLWrapper, Author, Category, Tag) # NOQA\n\nlogger = logging.getLogger(__name__)\n\n\nclass Content(object):\n \"\"\"Represents a content.\n\n :param content: the string to parse, containing the original content.\n :param metadata: the metadata associated to this page (optional).\n :param settings: the settings dictionary (optional).\n :param source_path: The location of the source of this content (if any).\n :param context: The shared context between generators.\n\n \"\"\"\n @deprecated_attribute(old='filename', new='source_path', since=(3, 2, 0))\n def filename():\n return None\n\n def __init__(self, content, metadata=None, settings=None,\n source_path=None, context=None):\n if metadata is None:\n metadata = {}\n if settings is None:\n settings = copy.deepcopy(DEFAULT_CONFIG)\n\n self.settings = settings\n self._content = content\n if context is None:\n context = {}\n self._context = context\n self.translations = []\n\n local_metadata = dict(settings['DEFAULT_METADATA'])\n local_metadata.update(metadata)\n\n # set metadata as attributes\n for key, value in local_metadata.items():\n if key in ('save_as', 'url'):\n key = 'override_' + key\n setattr(self, key.lower(), value)\n\n # also keep track of the metadata attributes available\n self.metadata = local_metadata\n\n #default template if it's not defined in page\n self.template = self._get_template()\n\n # default author to the one in settings if not defined\n if not hasattr(self, 'author'):\n if 'AUTHOR' in settings:\n self.author = Author(settings['AUTHOR'], settings)\n\n # XXX Split all the following code into pieces, there is too much here.\n\n # manage languages\n self.in_default_lang = True\n if 'DEFAULT_LANG' in settings:\n default_lang = settings['DEFAULT_LANG'].lower()\n if not hasattr(self, 'lang'):\n self.lang = default_lang\n\n self.in_default_lang = (self.lang == default_lang)\n\n # create the slug if not existing, from the title\n if not hasattr(self, 'slug') and hasattr(self, 'title'):\n self.slug = slugify(self.title,\n settings.get('SLUG_SUBSTITUTIONS', ()))\n\n self.source_path = source_path\n\n # manage the date format\n if not hasattr(self, 'date_format'):\n if hasattr(self, 'lang') and self.lang in settings['DATE_FORMATS']:\n self.date_format = settings['DATE_FORMATS'][self.lang]\n else:\n self.date_format = settings['DEFAULT_DATE_FORMAT']\n\n if isinstance(self.date_format, tuple):\n locale_string = self.date_format[0]\n if sys.version_info < (3, ) and isinstance(locale_string,\n six.text_type):\n locale_string = locale_string.encode('ascii')\n locale.setlocale(locale.LC_ALL, locale_string)\n self.date_format = self.date_format[1]\n\n if hasattr(self, 'date'):\n self.locale_date = strftime(self.date, self.date_format)\n\n # manage status\n if not hasattr(self, 'status'):\n self.status = settings['DEFAULT_STATUS']\n if not settings['WITH_FUTURE_DATES']:\n if hasattr(self, 'date') and self.date > datetime.now():\n self.status = 'draft'\n\n # store the summary metadata if it is set\n if 'summary' in metadata:\n self._summary = metadata['summary']\n\n signals.content_object_init.send(self)\n\n def __str__(self):\n if self.source_path is None:\n return repr(self)\n elif six.PY3:\n return self.source_path or repr(self)\n else:\n return str(self.source_path.encode('utf-8', 'replace'))\n\n def check_properties(self):\n \"\"\"Test mandatory properties are set.\"\"\"\n for prop in self.mandatory_properties:\n if not hasattr(self, prop):\n raise NameError(prop)\n\n @property\n def url_format(self):\n \"\"\"Returns the URL, formatted with the proper values\"\"\"\n metadata = copy.copy(self.metadata)\n path = self.metadata.get('path', self.get_relative_source_path())\n metadata.update({\n 'path': path_to_url(path),\n 'slug': getattr(self, 'slug', ''),\n 'lang': getattr(self, 'lang', 'en'),\n 'date': getattr(self, 'date', datetime.now()),\n 'author': getattr(self, 'author', ''),\n 'category': getattr(self, 'category',\n self.settings['DEFAULT_CATEGORY']),\n })\n return metadata\n\n def _expand_settings(self, key):\n fq_key = ('%s_%s' % (self.__class__.__name__, key)).upper()\n return self.settings[fq_key].format(**self.url_format)\n\n def get_url_setting(self, key):\n if hasattr(self, 'override_' + key):\n return getattr(self, 'override_' + key)\n key = key if self.in_default_lang else 'lang_%s' % key\n return self._expand_settings(key)\n\n def _update_content(self, content, siteurl):\n \"\"\"Update the content attribute.\n\n Change all the relative paths of the content to relative paths\n suitable for the ouput content.\n\n :param content: content resource that will be passed to the templates.\n :param siteurl: siteurl which is locally generated by the writer in\n case of RELATIVE_URLS.\n \"\"\"\n if not content:\n return content\n\n hrefs = re.compile(r\"\"\"\n (?P<markup><\\s*[^\\>]* # match tag with src and href attr\n (?:href|src)\\s*=)\n\n (?P<quote>[\"\\']) # require value to be quoted\n (?P<path>\\|(?P<what>.*?)\\|(?P<value>.*?)) # the url value\n \\2\"\"\", re.X)\n\n def replacer(m):\n what = m.group('what')\n value = m.group('value')\n origin = m.group('path')\n\n # we support only filename for now. the plan is to support\n # categories, tags, etc. in the future, but let's keep things\n # simple for now.\n\n # XXX Put this in a different location.\n if what == 'filename':\n if value.startswith('/'):\n value = value[1:]\n else:\n # relative to the source path of this content\n value = self.get_relative_source_path(\n os.path.join(self.relative_dir, value)\n )\n\n if value in self._context['filenames']:\n origin = '/'.join((siteurl,\n self._context['filenames'][value].url))\n origin = origin.replace('\\\\', '/') # Fow windows paths.\n else:\n logger.warning(\"Unable to find {fn}, skipping url\"\n \" replacement\".format(fn=value))\n\n return ''.join((m.group('markup'), m.group('quote'), origin,\n m.group('quote')))\n\n return hrefs.sub(replacer, content)\n\n @memoized\n def get_content(self, siteurl):\n\n if hasattr(self, '_get_content'):\n content = self._get_content()\n else:\n content = self._content\n return self._update_content(content, siteurl)\n\n @property\n def content(self):\n return self.get_content(self._context.get('localsiteurl', ''))\n\n def _get_summary(self):\n \"\"\"Returns the summary of an article.\n\n This is based on the summary metadata if set, otherwise truncate the\n content.\n \"\"\"\n if hasattr(self, '_summary'):\n return self._summary\n\n if self.settings['SUMMARY_MAX_LENGTH'] is None:\n return self.content\n\n return truncate_html_words(self.content,\n self.settings['SUMMARY_MAX_LENGTH'])\n\n def _set_summary(self, summary):\n \"\"\"Dummy function\"\"\"\n pass\n\n summary = property(_get_summary, _set_summary, \"Summary of the article.\"\n \"Based on the content. Can't be set\")\n url = property(functools.partial(get_url_setting, key='url'))\n save_as = property(functools.partial(get_url_setting, key='save_as'))\n\n def _get_template(self):\n if hasattr(self, 'template') and self.template is not None:\n return self.template\n else:\n return self.default_template\n\n def get_relative_source_path(self, source_path=None):\n \"\"\"Return the relative path (from the content path) to the given\n source_path.\n\n If no source path is specified, use the source path of this\n content object.\n \"\"\"\n if not source_path:\n source_path = self.source_path\n if source_path is None:\n return None\n\n return os.path.relpath(\n os.path.abspath(os.path.join(self.settings['PATH'], source_path)),\n os.path.abspath(self.settings['PATH'])\n )\n\n @property\n def relative_dir(self):\n return os.path.dirname(os.path.relpath(\n os.path.abspath(self.source_path),\n os.path.abspath(self.settings['PATH']))\n )\n\n\nclass Page(Content):\n mandatory_properties = ('title',)\n default_template = 'page'\n\n\nclass Article(Page):\n mandatory_properties = ('title', 'date', 'category')\n default_template = 'article'\n\n\nclass Quote(Page):\n base_properties = ('author', 'date')\n\n\n@python_2_unicode_compatible\nclass Static(Page):\n @deprecated_attribute(old='filepath', new='source_path', since=(3, 2, 0))\n def filepath():\n return None\n\n @deprecated_attribute(old='src', new='source_path', since=(3, 2, 0))\n def src():\n return None\n\n @deprecated_attribute(old='dst', new='save_as', since=(3, 2, 0))\n def dst():\n return None\n\n\ndef is_valid_content(content, f):\n try:\n content.check_properties()\n return True\n except NameError as e:\n logger.error(\"Skipping %s: could not find information about \"\n \"'%s'\" % (f, e))\n return False\n", "path": "pelican/contents.py"}]}
| 3,812 | 257 |
gh_patches_debug_17192
|
rasdani/github-patches
|
git_diff
|
DataDog__dd-trace-py-3151
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
aioredis 1.3.x tracing does not correctly parent spans
### Which version of dd-trace-py are you using?
0.57.1
### Which version of pip are you using?
21.3.1
### Which version of the libraries are you using?
```
aioredis==1.3.1
async-timeout==4.0.2
attrs==21.4.0
coverage==6.2
-e git+ssh://[email protected]/jalaziz/dd-trace-py.git@7e36d871f2efec84c5e49f4a0f6320a9870d88c8#egg=ddtrace
hiredis==2.0.0
hypothesis==6.36.0
iniconfig==1.1.1
mock==4.0.3
opentracing==2.4.0
packaging==21.3
pluggy==1.0.0
protobuf==3.19.3
py==1.11.0
pyparsing==3.0.6
pytest==6.2.5
pytest-asyncio==0.17.2
pytest-cov==3.0.0
pytest-mock==3.6.1
six==1.16.0
sortedcontainers==2.4.0
tenacity==8.0.1
toml==0.10.2
tomli==2.0.0
```
### How can we reproduce your problem?
This test reproduces the problem:
```
@pytest.mark.asyncio
@pytest.mark.snapshot(variants={"": aioredis_version >= (2, 0), "13": aioredis_version < (2, 0)})
async def test_parenting(redis_client):
with tracer.trace("web-request", service="test"):
await redis_client.set("blah", "boo")
await redis_client.set("foo", "bar")
await redis_client.get("blah")
await redis_client.get("foo")
```
### What is the result that you get?
The redis command spans are not correctly parented to `web-request` for `aioredis` 1.3, but they are for 2.0.
### What is the result that you expected?
The redis command spans should be correctly reported as children of `web-request`.
</issue>
<code>
[start of ddtrace/contrib/aioredis/patch.py]
1 import asyncio
2 import sys
3
4 import aioredis
5
6 from ddtrace import config
7 from ddtrace.internal.utils.wrappers import unwrap as _u
8 from ddtrace.pin import Pin
9 from ddtrace.vendor.wrapt import wrap_function_wrapper as _w
10
11 from .. import trace_utils
12 from ...constants import ANALYTICS_SAMPLE_RATE_KEY
13 from ...constants import SPAN_MEASURED_KEY
14 from ...ext import SpanTypes
15 from ...ext import net
16 from ...ext import redis as redisx
17 from ..redis.util import _trace_redis_cmd
18 from ..redis.util import _trace_redis_execute_pipeline
19 from ..redis.util import format_command_args
20
21
22 try:
23 from aioredis.commands.transaction import _RedisBuffer
24 except ImportError:
25 _RedisBuffer = None
26
27 config._add("aioredis", dict(_default_service="redis"))
28
29 aioredis_version_str = getattr(aioredis, "__version__", "0.0.0")
30 aioredis_version = tuple([int(i) for i in aioredis_version_str.split(".")])
31
32
33 def patch():
34 if getattr(aioredis, "_datadog_patch", False):
35 return
36 setattr(aioredis, "_datadog_patch", True)
37 pin = Pin()
38 if aioredis_version >= (2, 0):
39 _w("aioredis.client", "Redis.execute_command", traced_execute_command)
40 _w("aioredis.client", "Redis.pipeline", traced_pipeline)
41 _w("aioredis.client", "Pipeline.execute", traced_execute_pipeline)
42 pin.onto(aioredis.client.Redis)
43 else:
44 _w("aioredis", "Redis.execute", traced_13_execute_command)
45 _w("aioredis", "Redis.pipeline", traced_13_pipeline)
46 _w("aioredis.commands.transaction", "Pipeline.execute", traced_13_execute_pipeline)
47 pin.onto(aioredis.Redis)
48
49
50 def unpatch():
51 if not getattr(aioredis, "_datadog_patch", False):
52 return
53
54 setattr(aioredis, "_datadog_patch", False)
55 if aioredis_version >= (2, 0):
56 _u(aioredis.client.Redis, "execute_command")
57 _u(aioredis.client.Redis, "pipeline")
58 _u(aioredis.client.Pipeline, "execute")
59 else:
60 _u(aioredis.Redis, "execute")
61 _u(aioredis.Redis, "pipeline")
62 _u(aioredis.commands.transaction.Pipeline, "execute")
63
64
65 async def traced_execute_command(func, instance, args, kwargs):
66 pin = Pin.get_from(instance)
67 if not pin or not pin.enabled():
68 return await func(*args, **kwargs)
69
70 with _trace_redis_cmd(pin, config.aioredis, instance, args):
71 return await func(*args, **kwargs)
72
73
74 def traced_pipeline(func, instance, args, kwargs):
75 pipeline = func(*args, **kwargs)
76 pin = Pin.get_from(instance)
77 if pin:
78 pin.onto(pipeline)
79 return pipeline
80
81
82 async def traced_execute_pipeline(func, instance, args, kwargs):
83 pin = Pin.get_from(instance)
84 if not pin or not pin.enabled():
85 return await func(*args, **kwargs)
86
87 cmds = [format_command_args(c) for c, _ in instance.command_stack]
88 resource = "\n".join(cmds)
89 with _trace_redis_execute_pipeline(pin, config.aioredis, resource, instance):
90 return await func(*args, **kwargs)
91
92
93 def traced_13_pipeline(func, instance, args, kwargs):
94 pipeline = func(*args, **kwargs)
95 pin = Pin.get_from(instance)
96 if pin:
97 pin.onto(pipeline)
98 return pipeline
99
100
101 def traced_13_execute_command(func, instance, args, kwargs):
102 # If we have a _RedisBuffer then we are in a pipeline
103 if isinstance(instance.connection, _RedisBuffer):
104 return func(*args, **kwargs)
105
106 pin = Pin.get_from(instance)
107 if not pin or not pin.enabled():
108 return func(*args, **kwargs)
109
110 # Don't activate the span since this operation is performed as a future which concludes sometime later on in
111 # execution so subsequent operations in the stack are not necessarily semantically related
112 # (we don't want this span to be the parent of all other spans created before the future is resolved)
113 span = pin.tracer.start_span(
114 redisx.CMD, service=trace_utils.ext_service(pin, config.aioredis), span_type=SpanTypes.REDIS, activate=False
115 )
116
117 span.set_tag(SPAN_MEASURED_KEY)
118 query = format_command_args(args)
119 span.resource = query
120 span.set_tag(redisx.RAWCMD, query)
121 if pin.tags:
122 span.set_tags(pin.tags)
123
124 span.set_tags(
125 {
126 net.TARGET_HOST: instance.address[0],
127 net.TARGET_PORT: instance.address[1],
128 redisx.DB: instance.db or 0,
129 }
130 )
131 span.set_metric(redisx.ARGS_LEN, len(args))
132 # set analytics sample rate if enabled
133 span.set_tag(ANALYTICS_SAMPLE_RATE_KEY, config.aioredis.get_analytics_sample_rate())
134
135 def _finish_span(future):
136 try:
137 # Accessing the result will raise an exception if:
138 # - The future was cancelled
139 # - There was an error executing the future (`future.exception()`)
140 # - The future is in an invalid state
141 future.result()
142 except Exception:
143 span.set_exc_info(*sys.exc_info())
144 finally:
145 span.finish()
146
147 task = func(*args, **kwargs)
148 # Execute command returns a coroutine when no free connections are available
149 # https://github.com/aio-libs/aioredis-py/blob/v1.3.1/aioredis/pool.py#L191
150 task = asyncio.ensure_future(task)
151 task.add_done_callback(_finish_span)
152 return task
153
154
155 async def traced_13_execute_pipeline(func, instance, args, kwargs):
156 pin = Pin.get_from(instance)
157 if not pin or not pin.enabled():
158 return await func(*args, **kwargs)
159
160 cmds = []
161 for _, cmd, cmd_args, _ in instance._pipeline:
162 parts = [cmd]
163 parts.extend(cmd_args)
164 cmds.append(format_command_args(parts))
165 resource = "\n".join(cmds)
166 with pin.tracer.trace(
167 redisx.CMD,
168 resource=resource,
169 service=trace_utils.ext_service(pin, config.aioredis),
170 span_type=SpanTypes.REDIS,
171 ) as span:
172
173 span.set_tags(
174 {
175 net.TARGET_HOST: instance._pool_or_conn.address[0],
176 net.TARGET_PORT: instance._pool_or_conn.address[1],
177 redisx.DB: instance._pool_or_conn.db or 0,
178 }
179 )
180
181 span.set_tag(SPAN_MEASURED_KEY)
182 span.set_tag(redisx.RAWCMD, resource)
183 span.set_metric(redisx.PIPELINE_LEN, len(instance._pipeline))
184 # set analytics sample rate if enabled
185 span.set_tag(ANALYTICS_SAMPLE_RATE_KEY, config.aioredis.get_analytics_sample_rate())
186
187 return await func(*args, **kwargs)
188
[end of ddtrace/contrib/aioredis/patch.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/ddtrace/contrib/aioredis/patch.py b/ddtrace/contrib/aioredis/patch.py
--- a/ddtrace/contrib/aioredis/patch.py
+++ b/ddtrace/contrib/aioredis/patch.py
@@ -110,8 +110,13 @@
# Don't activate the span since this operation is performed as a future which concludes sometime later on in
# execution so subsequent operations in the stack are not necessarily semantically related
# (we don't want this span to be the parent of all other spans created before the future is resolved)
+ parent = pin.tracer.current_span()
span = pin.tracer.start_span(
- redisx.CMD, service=trace_utils.ext_service(pin, config.aioredis), span_type=SpanTypes.REDIS, activate=False
+ redisx.CMD,
+ service=trace_utils.ext_service(pin, config.aioredis),
+ span_type=SpanTypes.REDIS,
+ activate=False,
+ child_of=parent,
)
span.set_tag(SPAN_MEASURED_KEY)
|
{"golden_diff": "diff --git a/ddtrace/contrib/aioredis/patch.py b/ddtrace/contrib/aioredis/patch.py\n--- a/ddtrace/contrib/aioredis/patch.py\n+++ b/ddtrace/contrib/aioredis/patch.py\n@@ -110,8 +110,13 @@\n # Don't activate the span since this operation is performed as a future which concludes sometime later on in\n # execution so subsequent operations in the stack are not necessarily semantically related\n # (we don't want this span to be the parent of all other spans created before the future is resolved)\n+ parent = pin.tracer.current_span()\n span = pin.tracer.start_span(\n- redisx.CMD, service=trace_utils.ext_service(pin, config.aioredis), span_type=SpanTypes.REDIS, activate=False\n+ redisx.CMD,\n+ service=trace_utils.ext_service(pin, config.aioredis),\n+ span_type=SpanTypes.REDIS,\n+ activate=False,\n+ child_of=parent,\n )\n \n span.set_tag(SPAN_MEASURED_KEY)\n", "issue": "aioredis 1.3.x tracing does not correctly parent spans\n### Which version of dd-trace-py are you using?\r\n\r\n0.57.1\r\n\r\n### Which version of pip are you using?\r\n\r\n21.3.1\r\n\r\n### Which version of the libraries are you using?\r\n\r\n```\r\naioredis==1.3.1\r\nasync-timeout==4.0.2\r\nattrs==21.4.0\r\ncoverage==6.2\r\n-e git+ssh://[email protected]/jalaziz/dd-trace-py.git@7e36d871f2efec84c5e49f4a0f6320a9870d88c8#egg=ddtrace\r\nhiredis==2.0.0\r\nhypothesis==6.36.0\r\niniconfig==1.1.1\r\nmock==4.0.3\r\nopentracing==2.4.0\r\npackaging==21.3\r\npluggy==1.0.0\r\nprotobuf==3.19.3\r\npy==1.11.0\r\npyparsing==3.0.6\r\npytest==6.2.5\r\npytest-asyncio==0.17.2\r\npytest-cov==3.0.0\r\npytest-mock==3.6.1\r\nsix==1.16.0\r\nsortedcontainers==2.4.0\r\ntenacity==8.0.1\r\ntoml==0.10.2\r\ntomli==2.0.0\r\n```\r\n\r\n### How can we reproduce your problem?\r\n\r\nThis test reproduces the problem:\r\n```\r\[email protected]\r\[email protected](variants={\"\": aioredis_version >= (2, 0), \"13\": aioredis_version < (2, 0)})\r\nasync def test_parenting(redis_client):\r\n with tracer.trace(\"web-request\", service=\"test\"):\r\n await redis_client.set(\"blah\", \"boo\")\r\n await redis_client.set(\"foo\", \"bar\")\r\n await redis_client.get(\"blah\")\r\n await redis_client.get(\"foo\")\r\n```\r\n\r\n### What is the result that you get?\r\n\r\nThe redis command spans are not correctly parented to `web-request` for `aioredis` 1.3, but they are for 2.0.\r\n\r\n### What is the result that you expected?\r\n\r\nThe redis command spans should be correctly reported as children of `web-request`.\n", "before_files": [{"content": "import asyncio\nimport sys\n\nimport aioredis\n\nfrom ddtrace import config\nfrom ddtrace.internal.utils.wrappers import unwrap as _u\nfrom ddtrace.pin import Pin\nfrom ddtrace.vendor.wrapt import wrap_function_wrapper as _w\n\nfrom .. import trace_utils\nfrom ...constants import ANALYTICS_SAMPLE_RATE_KEY\nfrom ...constants import SPAN_MEASURED_KEY\nfrom ...ext import SpanTypes\nfrom ...ext import net\nfrom ...ext import redis as redisx\nfrom ..redis.util import _trace_redis_cmd\nfrom ..redis.util import _trace_redis_execute_pipeline\nfrom ..redis.util import format_command_args\n\n\ntry:\n from aioredis.commands.transaction import _RedisBuffer\nexcept ImportError:\n _RedisBuffer = None\n\nconfig._add(\"aioredis\", dict(_default_service=\"redis\"))\n\naioredis_version_str = getattr(aioredis, \"__version__\", \"0.0.0\")\naioredis_version = tuple([int(i) for i in aioredis_version_str.split(\".\")])\n\n\ndef patch():\n if getattr(aioredis, \"_datadog_patch\", False):\n return\n setattr(aioredis, \"_datadog_patch\", True)\n pin = Pin()\n if aioredis_version >= (2, 0):\n _w(\"aioredis.client\", \"Redis.execute_command\", traced_execute_command)\n _w(\"aioredis.client\", \"Redis.pipeline\", traced_pipeline)\n _w(\"aioredis.client\", \"Pipeline.execute\", traced_execute_pipeline)\n pin.onto(aioredis.client.Redis)\n else:\n _w(\"aioredis\", \"Redis.execute\", traced_13_execute_command)\n _w(\"aioredis\", \"Redis.pipeline\", traced_13_pipeline)\n _w(\"aioredis.commands.transaction\", \"Pipeline.execute\", traced_13_execute_pipeline)\n pin.onto(aioredis.Redis)\n\n\ndef unpatch():\n if not getattr(aioredis, \"_datadog_patch\", False):\n return\n\n setattr(aioredis, \"_datadog_patch\", False)\n if aioredis_version >= (2, 0):\n _u(aioredis.client.Redis, \"execute_command\")\n _u(aioredis.client.Redis, \"pipeline\")\n _u(aioredis.client.Pipeline, \"execute\")\n else:\n _u(aioredis.Redis, \"execute\")\n _u(aioredis.Redis, \"pipeline\")\n _u(aioredis.commands.transaction.Pipeline, \"execute\")\n\n\nasync def traced_execute_command(func, instance, args, kwargs):\n pin = Pin.get_from(instance)\n if not pin or not pin.enabled():\n return await func(*args, **kwargs)\n\n with _trace_redis_cmd(pin, config.aioredis, instance, args):\n return await func(*args, **kwargs)\n\n\ndef traced_pipeline(func, instance, args, kwargs):\n pipeline = func(*args, **kwargs)\n pin = Pin.get_from(instance)\n if pin:\n pin.onto(pipeline)\n return pipeline\n\n\nasync def traced_execute_pipeline(func, instance, args, kwargs):\n pin = Pin.get_from(instance)\n if not pin or not pin.enabled():\n return await func(*args, **kwargs)\n\n cmds = [format_command_args(c) for c, _ in instance.command_stack]\n resource = \"\\n\".join(cmds)\n with _trace_redis_execute_pipeline(pin, config.aioredis, resource, instance):\n return await func(*args, **kwargs)\n\n\ndef traced_13_pipeline(func, instance, args, kwargs):\n pipeline = func(*args, **kwargs)\n pin = Pin.get_from(instance)\n if pin:\n pin.onto(pipeline)\n return pipeline\n\n\ndef traced_13_execute_command(func, instance, args, kwargs):\n # If we have a _RedisBuffer then we are in a pipeline\n if isinstance(instance.connection, _RedisBuffer):\n return func(*args, **kwargs)\n\n pin = Pin.get_from(instance)\n if not pin or not pin.enabled():\n return func(*args, **kwargs)\n\n # Don't activate the span since this operation is performed as a future which concludes sometime later on in\n # execution so subsequent operations in the stack are not necessarily semantically related\n # (we don't want this span to be the parent of all other spans created before the future is resolved)\n span = pin.tracer.start_span(\n redisx.CMD, service=trace_utils.ext_service(pin, config.aioredis), span_type=SpanTypes.REDIS, activate=False\n )\n\n span.set_tag(SPAN_MEASURED_KEY)\n query = format_command_args(args)\n span.resource = query\n span.set_tag(redisx.RAWCMD, query)\n if pin.tags:\n span.set_tags(pin.tags)\n\n span.set_tags(\n {\n net.TARGET_HOST: instance.address[0],\n net.TARGET_PORT: instance.address[1],\n redisx.DB: instance.db or 0,\n }\n )\n span.set_metric(redisx.ARGS_LEN, len(args))\n # set analytics sample rate if enabled\n span.set_tag(ANALYTICS_SAMPLE_RATE_KEY, config.aioredis.get_analytics_sample_rate())\n\n def _finish_span(future):\n try:\n # Accessing the result will raise an exception if:\n # - The future was cancelled\n # - There was an error executing the future (`future.exception()`)\n # - The future is in an invalid state\n future.result()\n except Exception:\n span.set_exc_info(*sys.exc_info())\n finally:\n span.finish()\n\n task = func(*args, **kwargs)\n # Execute command returns a coroutine when no free connections are available\n # https://github.com/aio-libs/aioredis-py/blob/v1.3.1/aioredis/pool.py#L191\n task = asyncio.ensure_future(task)\n task.add_done_callback(_finish_span)\n return task\n\n\nasync def traced_13_execute_pipeline(func, instance, args, kwargs):\n pin = Pin.get_from(instance)\n if not pin or not pin.enabled():\n return await func(*args, **kwargs)\n\n cmds = []\n for _, cmd, cmd_args, _ in instance._pipeline:\n parts = [cmd]\n parts.extend(cmd_args)\n cmds.append(format_command_args(parts))\n resource = \"\\n\".join(cmds)\n with pin.tracer.trace(\n redisx.CMD,\n resource=resource,\n service=trace_utils.ext_service(pin, config.aioredis),\n span_type=SpanTypes.REDIS,\n ) as span:\n\n span.set_tags(\n {\n net.TARGET_HOST: instance._pool_or_conn.address[0],\n net.TARGET_PORT: instance._pool_or_conn.address[1],\n redisx.DB: instance._pool_or_conn.db or 0,\n }\n )\n\n span.set_tag(SPAN_MEASURED_KEY)\n span.set_tag(redisx.RAWCMD, resource)\n span.set_metric(redisx.PIPELINE_LEN, len(instance._pipeline))\n # set analytics sample rate if enabled\n span.set_tag(ANALYTICS_SAMPLE_RATE_KEY, config.aioredis.get_analytics_sample_rate())\n\n return await func(*args, **kwargs)\n", "path": "ddtrace/contrib/aioredis/patch.py"}]}
| 3,106 | 241 |
gh_patches_debug_13729
|
rasdani/github-patches
|
git_diff
|
aws-cloudformation__cfn-lint-2765
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Invalid Property Resources/SESConfigurationSetDestination/Properties/EventDestination/SnsDestination
### CloudFormation Lint Version
0.77.8
### What operating system are you using?
Mac/Ubuntu
### Describe the bug
cfn-lint outputs:
Invalid Property Resources/SESConfigurationSetDestination/Properties/EventDestination/SnsDestination
But this is a valid property:
https://docs.aws.amazon.com/AWSCloudFormation/latest/UserGuide/aws-properties-ses-configurationseteventdestination-eventdestination.html#cfn-ses-configurationseteventdestination-eventdestination-snsdestination
### Expected behavior
cfn-lint doesn't output an error
### Reproduction template
```yaml
Resources:
SESConfigurationSetDestination:
Type: AWS::SES::ConfigurationSetEventDestination
Properties:
ConfigurationSetName: Name
EventDestination:
Name: DestinationName
Enabled: true
MatchingEventTypes:
- send
- reject
- bounce
- complaint
- delivery
SnsDestination:
TopicARN: Arn
```
</issue>
<code>
[start of scripts/update_specs_services_from_ssm.py]
1 #!/usr/bin/env python
2 """
3 Copyright Amazon.com, Inc. or its affiliates. All Rights Reserved.
4 SPDX-License-Identifier: MIT-0
5 """
6 import json
7 import logging
8
9 import boto3
10
11 from cfnlint.helpers import REGIONS, get_url_content
12 from cfnlint.maintenance import SPEC_REGIONS
13
14 """
15 Updates our dynamic patches from SSM data
16 This script requires Boto3 and Credentials to call the SSM API
17 """
18
19 LOGGER = logging.getLogger("cfnlint")
20
21 service_map = {
22 "acm": ["AWS::CertificateManager::"],
23 "apigateway": ["AWS::ApiGateway::", "AWS::ApiGatewayV2::"],
24 # 'application-autoscaling': ['AWS::ApplicationAutoScaling::'], ## remove because SSM endpoints aren't correct
25 "appstream": ["AWS::AppStream::"],
26 "appsync": ["AWS::AppSync::"],
27 "athena": ["AWS::Athena::"],
28 "autoscaling": ["AWS::AutoScaling::"],
29 "batch": ["AWS::Batch::"],
30 "budgets": ["AWS::Budgets::"],
31 "cloud9": ["AWS::Cloud9::"],
32 "cloudfront": ["AWS::CloudFront::"],
33 "cloudtrail": ["AWS::CloudTrail::"],
34 "cloudwatch": ["AWS::CloudWatch::"],
35 "codebuild": ["AWS::CodeBuild::"],
36 "codecommit": ["AWS::CodeCommit::"],
37 "codedeploy": ["AWS::CodeDeploy::"],
38 "codepipeline": ["AWS::CodePipeline::"],
39 "cognito-identity": ["AWS::Cognito::"],
40 "config": ["AWS::Config::"],
41 "datapipeline": ["AWS::DataPipeline::"],
42 "dax": ["AWS::DAX::"],
43 "dms": ["AWS::DMS::"],
44 "docdb": ["AWS::DocDB::"],
45 "ds": ["AWS::DirectoryService::"],
46 "dynamodb": ["AWS::DynamoDB::"],
47 "ec2": ["AWS::EC2::"],
48 "ecr": ["AWS::ECR::"],
49 "ecs": ["AWS::ECS::"],
50 "efs": ["AWS::EFS::"],
51 "eks": ["AWS::EKS::"],
52 "elasticache": ["AWS::ElastiCache::"],
53 "elasticbeanstalk": ["AWS::ElasticBeanstalk::"],
54 "elb": ["AWS::ElasticLoadBalancing::", "AWS::ElasticLoadBalancingV2::"],
55 "emr": ["AWS::EMR::"],
56 "es": ["AWS::Elasticsearch::"],
57 "events": ["AWS::Events::"],
58 "firehose": ["AWS::KinesisFirehose::"],
59 "fsx": ["AWS::FSx::"],
60 "gamelift": ["AWS::GameLift::"],
61 "glue": ["AWS::Glue::"],
62 "greengrass": ["AWS::Greengrass::"],
63 "guardduty": ["AWS::GuardDuty::"],
64 "inspector": ["AWS::Inspector::"],
65 "iot": ["AWS::IoT::"],
66 "iot1click-projects": ["AWS::IoT1Click::"],
67 "iotanalytics": ["AWS::IoTAnalytics::"],
68 "kinesis": ["AWS::Kinesis::"],
69 "kinesisanalytics": ["AWS::KinesisAnalytics::", "AWS::KinesisAnalyticsV2::"],
70 "kms": ["AWS::KMS::"],
71 "lambda": ["AWS::Lambda::"],
72 "logs": ["AWS::Logs::"],
73 "mq": ["AWS::AmazonMQ::"],
74 "neptune": ["AWS::Neptune::"],
75 "opsworks": ["AWS::OpsWorks::"],
76 "opsworkscm": ["AWS::OpsWorksCM::"],
77 "ram": ["AWS::RAM::"],
78 "rds": ["AWS::RDS::"],
79 "redshift": ["AWS::Redshift::"],
80 "robomaker": ["AWS::RoboMaker::"],
81 "route53": ["AWS::Route53::"],
82 "route53resolver": [
83 "AWS::Route53Resolver::ResolverRule",
84 "AWS::Route53Resolver::ResolverEndpoint",
85 ],
86 "s3": ["AWS::S3::"],
87 "sagemaker": ["AWS::SageMaker::"],
88 "sdb": ["AWS::SDB::"],
89 "secretsmanager": ["AWS::SecretsManager::"],
90 "servicecatalog": ["AWS::ServiceCatalog::"],
91 "servicediscovery": ["AWS::ServiceDiscovery::"],
92 "ses": ["AWS::SES::"],
93 "sns": ["AWS::SNS::"],
94 "sqs": ["AWS::SQS::"],
95 "ssm": ["AWS::SSM::"],
96 "stepfunctions": ["AWS::StepFunctions::"],
97 "waf-regional": ["AWS::WAFRegional::"],
98 "workspaces": ["AWS::WorkSpaces::"],
99 }
100
101 session = boto3.session.Session()
102 client = session.client("ssm", region_name="us-east-1")
103
104
105 def configure_logging():
106 """Setup Logging"""
107 ch = logging.StreamHandler()
108 ch.setLevel(logging.INFO)
109
110 LOGGER.setLevel(logging.INFO)
111 log_formatter = logging.Formatter(
112 "%(asctime)s - %(name)s - %(levelname)s - %(message)s"
113 )
114 ch.setFormatter(log_formatter)
115
116 # make sure all other log handlers are removed before adding it back
117 for handler in LOGGER.handlers:
118 LOGGER.removeHandler(handler)
119 LOGGER.addHandler(ch)
120
121
122 def update_outputs(region, resource_type, name, outputs):
123 """update outputs with appropriate results"""
124 element = {"op": "remove", "path": "/%s/%s" % (resource_type, name)}
125 outputs[region].append(element)
126
127 return outputs
128
129
130 def get_regions_for_service(service):
131 """get regions for a service"""
132 LOGGER.info("Get the regions for service %s", service)
133 results = []
134 paginator = client.get_paginator("get_parameters_by_path")
135 page_iterator = paginator.paginate(
136 Path="/aws/service/global-infrastructure/services/{}/regions".format(service),
137 )
138
139 for page in page_iterator:
140 for region in page.get("Parameters"):
141 results.append(region.get("Value"))
142
143 return results
144
145
146 def add_spec_patch(region, services):
147 """Go through spec and determine patching"""
148 LOGGER.info("Create 06_ssm_service_removal patch for region %s", region)
149 spec = json.loads(get_url_content(SPEC_REGIONS.get(region)))
150
151 patches = []
152
153 for spec_type in ["ResourceTypes", "PropertyTypes"]:
154 for resource in sorted(spec.get(spec_type).keys()):
155 for service in services:
156 for spec_name in service_map.get(service):
157 if resource.startswith(spec_name):
158 element = {
159 "op": "remove",
160 "path": "/%s/%s" % (spec_type, resource),
161 }
162 patches.append(element)
163
164 filename = "src/cfnlint/data/ExtendedSpecs/%s/06_ssm_service_removal.json" % region
165 with open(filename, "w+", encoding="utf-8") as f:
166 json.dump(patches, f, indent=1, sort_keys=True, separators=(",", ": "))
167
168
169 def add_spec_missing_services_patch(region, services):
170 """Go through spec and determine patching"""
171 LOGGER.info("Create 07_ssm_service_addition patch for region %s", region)
172 spec_string = get_url_content(SPEC_REGIONS.get(region))
173 spec_string_standard = get_url_content(SPEC_REGIONS.get("us-east-1"))
174
175 spec = json.loads(spec_string)
176 spec_standard = json.loads(spec_string_standard)
177
178 patches = []
179
180 for spec_type in ["ResourceTypes"]:
181 for service in services:
182 found = False
183 for resource in sorted(spec.get(spec_type).keys()):
184 for spec_name in service_map.get(service):
185 if resource.startswith(spec_name):
186 found = True
187 if found is False:
188 for standard_spec_type in ["ResourceTypes", "PropertyTypes"]:
189 for resource in sorted(
190 spec_standard.get(standard_spec_type).keys()
191 ):
192 for spec_name in service_map.get(service):
193 if resource.startswith(spec_name):
194 if spec_standard.get(standard_spec_type).get(resource):
195 element = {
196 "op": "add",
197 "path": "/%s/%s"
198 % (standard_spec_type, resource),
199 "value": spec_standard.get(
200 standard_spec_type
201 ).get(resource),
202 }
203 patches.append(element)
204 elif standard_spec_type == "ResourceTypes":
205 print("patch for %s not found" % service)
206
207 if patches:
208 filename = (
209 "src/cfnlint/data/ExtendedSpecs/%s/07_ssm_service_addition.json" % region
210 )
211 with open(filename, "w+", encoding="utf-8") as f:
212 json.dump(patches, f, indent=1, sort_keys=True, separators=(",", ": "))
213
214
215 def main():
216 """main function"""
217 configure_logging()
218
219 all_regions = list(set(REGIONS))
220 region_service_removal_map = {}
221 region_service_add_map = {}
222 for region in all_regions:
223 region_service_removal_map[region] = []
224 region_service_add_map[region] = []
225 for service in service_map:
226 regions = get_regions_for_service(service)
227 if regions:
228 for region in list(set(regions)):
229 region_service_add_map[region].append(service)
230 for region in list(set(all_regions) - set(regions)):
231 region_service_removal_map[region].append(service)
232
233 for region, services in region_service_removal_map.items():
234 if services:
235 add_spec_patch(region, services)
236 for region, services in region_service_add_map.items():
237 if services:
238 add_spec_missing_services_patch(region, services)
239
240
241 if __name__ == "__main__":
242 try:
243 main()
244 except (ValueError, TypeError):
245 LOGGER.error(ValueError)
246
[end of scripts/update_specs_services_from_ssm.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/scripts/update_specs_services_from_ssm.py b/scripts/update_specs_services_from_ssm.py
--- a/scripts/update_specs_services_from_ssm.py
+++ b/scripts/update_specs_services_from_ssm.py
@@ -204,12 +204,12 @@
elif standard_spec_type == "ResourceTypes":
print("patch for %s not found" % service)
- if patches:
- filename = (
- "src/cfnlint/data/ExtendedSpecs/%s/07_ssm_service_addition.json" % region
- )
- with open(filename, "w+", encoding="utf-8") as f:
- json.dump(patches, f, indent=1, sort_keys=True, separators=(",", ": "))
+ filename = (
+ "src/cfnlint/data/ExtendedSpecs/%s/07_ssm_service_addition.json" % region
+ )
+ with open(filename, "w+", encoding="utf-8") as f:
+ json.dump(patches, f, indent=1, sort_keys=True, separators=(",", ": "))
+
def main():
|
{"golden_diff": "diff --git a/scripts/update_specs_services_from_ssm.py b/scripts/update_specs_services_from_ssm.py\n--- a/scripts/update_specs_services_from_ssm.py\n+++ b/scripts/update_specs_services_from_ssm.py\n@@ -204,12 +204,12 @@\n elif standard_spec_type == \"ResourceTypes\":\n print(\"patch for %s not found\" % service)\n \n- if patches:\n- filename = (\n- \"src/cfnlint/data/ExtendedSpecs/%s/07_ssm_service_addition.json\" % region\n- )\n- with open(filename, \"w+\", encoding=\"utf-8\") as f:\n- json.dump(patches, f, indent=1, sort_keys=True, separators=(\",\", \": \"))\n+ filename = (\n+ \"src/cfnlint/data/ExtendedSpecs/%s/07_ssm_service_addition.json\" % region\n+ )\n+ with open(filename, \"w+\", encoding=\"utf-8\") as f:\n+ json.dump(patches, f, indent=1, sort_keys=True, separators=(\",\", \": \"))\n+ \n \n \n def main():\n", "issue": "Invalid Property Resources/SESConfigurationSetDestination/Properties/EventDestination/SnsDestination\n### CloudFormation Lint Version\r\n\r\n0.77.8\r\n\r\n### What operating system are you using?\r\n\r\nMac/Ubuntu\r\n\r\n### Describe the bug\r\n\r\ncfn-lint outputs:\r\n\r\nInvalid Property Resources/SESConfigurationSetDestination/Properties/EventDestination/SnsDestination\r\n\r\nBut this is a valid property:\r\n\r\nhttps://docs.aws.amazon.com/AWSCloudFormation/latest/UserGuide/aws-properties-ses-configurationseteventdestination-eventdestination.html#cfn-ses-configurationseteventdestination-eventdestination-snsdestination\r\n\r\n### Expected behavior\r\n\r\ncfn-lint doesn't output an error\r\n\r\n### Reproduction template\r\n\r\n```yaml\r\nResources:\r\n SESConfigurationSetDestination:\r\n Type: AWS::SES::ConfigurationSetEventDestination\r\n Properties:\r\n ConfigurationSetName: Name\r\n EventDestination:\r\n Name: DestinationName\r\n Enabled: true\r\n MatchingEventTypes:\r\n - send\r\n - reject\r\n - bounce\r\n - complaint\r\n - delivery\r\n SnsDestination:\r\n TopicARN: Arn\r\n```\n", "before_files": [{"content": "#!/usr/bin/env python\n\"\"\"\nCopyright Amazon.com, Inc. or its affiliates. All Rights Reserved.\nSPDX-License-Identifier: MIT-0\n\"\"\"\nimport json\nimport logging\n\nimport boto3\n\nfrom cfnlint.helpers import REGIONS, get_url_content\nfrom cfnlint.maintenance import SPEC_REGIONS\n\n\"\"\"\n Updates our dynamic patches from SSM data\n This script requires Boto3 and Credentials to call the SSM API\n\"\"\"\n\nLOGGER = logging.getLogger(\"cfnlint\")\n\nservice_map = {\n \"acm\": [\"AWS::CertificateManager::\"],\n \"apigateway\": [\"AWS::ApiGateway::\", \"AWS::ApiGatewayV2::\"],\n # 'application-autoscaling': ['AWS::ApplicationAutoScaling::'], ## remove because SSM endpoints aren't correct\n \"appstream\": [\"AWS::AppStream::\"],\n \"appsync\": [\"AWS::AppSync::\"],\n \"athena\": [\"AWS::Athena::\"],\n \"autoscaling\": [\"AWS::AutoScaling::\"],\n \"batch\": [\"AWS::Batch::\"],\n \"budgets\": [\"AWS::Budgets::\"],\n \"cloud9\": [\"AWS::Cloud9::\"],\n \"cloudfront\": [\"AWS::CloudFront::\"],\n \"cloudtrail\": [\"AWS::CloudTrail::\"],\n \"cloudwatch\": [\"AWS::CloudWatch::\"],\n \"codebuild\": [\"AWS::CodeBuild::\"],\n \"codecommit\": [\"AWS::CodeCommit::\"],\n \"codedeploy\": [\"AWS::CodeDeploy::\"],\n \"codepipeline\": [\"AWS::CodePipeline::\"],\n \"cognito-identity\": [\"AWS::Cognito::\"],\n \"config\": [\"AWS::Config::\"],\n \"datapipeline\": [\"AWS::DataPipeline::\"],\n \"dax\": [\"AWS::DAX::\"],\n \"dms\": [\"AWS::DMS::\"],\n \"docdb\": [\"AWS::DocDB::\"],\n \"ds\": [\"AWS::DirectoryService::\"],\n \"dynamodb\": [\"AWS::DynamoDB::\"],\n \"ec2\": [\"AWS::EC2::\"],\n \"ecr\": [\"AWS::ECR::\"],\n \"ecs\": [\"AWS::ECS::\"],\n \"efs\": [\"AWS::EFS::\"],\n \"eks\": [\"AWS::EKS::\"],\n \"elasticache\": [\"AWS::ElastiCache::\"],\n \"elasticbeanstalk\": [\"AWS::ElasticBeanstalk::\"],\n \"elb\": [\"AWS::ElasticLoadBalancing::\", \"AWS::ElasticLoadBalancingV2::\"],\n \"emr\": [\"AWS::EMR::\"],\n \"es\": [\"AWS::Elasticsearch::\"],\n \"events\": [\"AWS::Events::\"],\n \"firehose\": [\"AWS::KinesisFirehose::\"],\n \"fsx\": [\"AWS::FSx::\"],\n \"gamelift\": [\"AWS::GameLift::\"],\n \"glue\": [\"AWS::Glue::\"],\n \"greengrass\": [\"AWS::Greengrass::\"],\n \"guardduty\": [\"AWS::GuardDuty::\"],\n \"inspector\": [\"AWS::Inspector::\"],\n \"iot\": [\"AWS::IoT::\"],\n \"iot1click-projects\": [\"AWS::IoT1Click::\"],\n \"iotanalytics\": [\"AWS::IoTAnalytics::\"],\n \"kinesis\": [\"AWS::Kinesis::\"],\n \"kinesisanalytics\": [\"AWS::KinesisAnalytics::\", \"AWS::KinesisAnalyticsV2::\"],\n \"kms\": [\"AWS::KMS::\"],\n \"lambda\": [\"AWS::Lambda::\"],\n \"logs\": [\"AWS::Logs::\"],\n \"mq\": [\"AWS::AmazonMQ::\"],\n \"neptune\": [\"AWS::Neptune::\"],\n \"opsworks\": [\"AWS::OpsWorks::\"],\n \"opsworkscm\": [\"AWS::OpsWorksCM::\"],\n \"ram\": [\"AWS::RAM::\"],\n \"rds\": [\"AWS::RDS::\"],\n \"redshift\": [\"AWS::Redshift::\"],\n \"robomaker\": [\"AWS::RoboMaker::\"],\n \"route53\": [\"AWS::Route53::\"],\n \"route53resolver\": [\n \"AWS::Route53Resolver::ResolverRule\",\n \"AWS::Route53Resolver::ResolverEndpoint\",\n ],\n \"s3\": [\"AWS::S3::\"],\n \"sagemaker\": [\"AWS::SageMaker::\"],\n \"sdb\": [\"AWS::SDB::\"],\n \"secretsmanager\": [\"AWS::SecretsManager::\"],\n \"servicecatalog\": [\"AWS::ServiceCatalog::\"],\n \"servicediscovery\": [\"AWS::ServiceDiscovery::\"],\n \"ses\": [\"AWS::SES::\"],\n \"sns\": [\"AWS::SNS::\"],\n \"sqs\": [\"AWS::SQS::\"],\n \"ssm\": [\"AWS::SSM::\"],\n \"stepfunctions\": [\"AWS::StepFunctions::\"],\n \"waf-regional\": [\"AWS::WAFRegional::\"],\n \"workspaces\": [\"AWS::WorkSpaces::\"],\n}\n\nsession = boto3.session.Session()\nclient = session.client(\"ssm\", region_name=\"us-east-1\")\n\n\ndef configure_logging():\n \"\"\"Setup Logging\"\"\"\n ch = logging.StreamHandler()\n ch.setLevel(logging.INFO)\n\n LOGGER.setLevel(logging.INFO)\n log_formatter = logging.Formatter(\n \"%(asctime)s - %(name)s - %(levelname)s - %(message)s\"\n )\n ch.setFormatter(log_formatter)\n\n # make sure all other log handlers are removed before adding it back\n for handler in LOGGER.handlers:\n LOGGER.removeHandler(handler)\n LOGGER.addHandler(ch)\n\n\ndef update_outputs(region, resource_type, name, outputs):\n \"\"\"update outputs with appropriate results\"\"\"\n element = {\"op\": \"remove\", \"path\": \"/%s/%s\" % (resource_type, name)}\n outputs[region].append(element)\n\n return outputs\n\n\ndef get_regions_for_service(service):\n \"\"\"get regions for a service\"\"\"\n LOGGER.info(\"Get the regions for service %s\", service)\n results = []\n paginator = client.get_paginator(\"get_parameters_by_path\")\n page_iterator = paginator.paginate(\n Path=\"/aws/service/global-infrastructure/services/{}/regions\".format(service),\n )\n\n for page in page_iterator:\n for region in page.get(\"Parameters\"):\n results.append(region.get(\"Value\"))\n\n return results\n\n\ndef add_spec_patch(region, services):\n \"\"\"Go through spec and determine patching\"\"\"\n LOGGER.info(\"Create 06_ssm_service_removal patch for region %s\", region)\n spec = json.loads(get_url_content(SPEC_REGIONS.get(region)))\n\n patches = []\n\n for spec_type in [\"ResourceTypes\", \"PropertyTypes\"]:\n for resource in sorted(spec.get(spec_type).keys()):\n for service in services:\n for spec_name in service_map.get(service):\n if resource.startswith(spec_name):\n element = {\n \"op\": \"remove\",\n \"path\": \"/%s/%s\" % (spec_type, resource),\n }\n patches.append(element)\n\n filename = \"src/cfnlint/data/ExtendedSpecs/%s/06_ssm_service_removal.json\" % region\n with open(filename, \"w+\", encoding=\"utf-8\") as f:\n json.dump(patches, f, indent=1, sort_keys=True, separators=(\",\", \": \"))\n\n\ndef add_spec_missing_services_patch(region, services):\n \"\"\"Go through spec and determine patching\"\"\"\n LOGGER.info(\"Create 07_ssm_service_addition patch for region %s\", region)\n spec_string = get_url_content(SPEC_REGIONS.get(region))\n spec_string_standard = get_url_content(SPEC_REGIONS.get(\"us-east-1\"))\n\n spec = json.loads(spec_string)\n spec_standard = json.loads(spec_string_standard)\n\n patches = []\n\n for spec_type in [\"ResourceTypes\"]:\n for service in services:\n found = False\n for resource in sorted(spec.get(spec_type).keys()):\n for spec_name in service_map.get(service):\n if resource.startswith(spec_name):\n found = True\n if found is False:\n for standard_spec_type in [\"ResourceTypes\", \"PropertyTypes\"]:\n for resource in sorted(\n spec_standard.get(standard_spec_type).keys()\n ):\n for spec_name in service_map.get(service):\n if resource.startswith(spec_name):\n if spec_standard.get(standard_spec_type).get(resource):\n element = {\n \"op\": \"add\",\n \"path\": \"/%s/%s\"\n % (standard_spec_type, resource),\n \"value\": spec_standard.get(\n standard_spec_type\n ).get(resource),\n }\n patches.append(element)\n elif standard_spec_type == \"ResourceTypes\":\n print(\"patch for %s not found\" % service)\n\n if patches:\n filename = (\n \"src/cfnlint/data/ExtendedSpecs/%s/07_ssm_service_addition.json\" % region\n )\n with open(filename, \"w+\", encoding=\"utf-8\") as f:\n json.dump(patches, f, indent=1, sort_keys=True, separators=(\",\", \": \"))\n\n\ndef main():\n \"\"\"main function\"\"\"\n configure_logging()\n\n all_regions = list(set(REGIONS))\n region_service_removal_map = {}\n region_service_add_map = {}\n for region in all_regions:\n region_service_removal_map[region] = []\n region_service_add_map[region] = []\n for service in service_map:\n regions = get_regions_for_service(service)\n if regions:\n for region in list(set(regions)):\n region_service_add_map[region].append(service)\n for region in list(set(all_regions) - set(regions)):\n region_service_removal_map[region].append(service)\n\n for region, services in region_service_removal_map.items():\n if services:\n add_spec_patch(region, services)\n for region, services in region_service_add_map.items():\n if services:\n add_spec_missing_services_patch(region, services)\n\n\nif __name__ == \"__main__\":\n try:\n main()\n except (ValueError, TypeError):\n LOGGER.error(ValueError)\n", "path": "scripts/update_specs_services_from_ssm.py"}]}
| 3,585 | 245 |
gh_patches_debug_7027
|
rasdani/github-patches
|
git_diff
|
conda__conda-10261
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
`conda env remove --dry-run` actually deletes the env
## Current Behavior
Using the `--dry-run` flag doesn't do a dry-run at all.
### Steps to Reproduce
```
$ conda env list
# conda environments:
#
base * /usr
$ conda create -n foo
Collecting package metadata (current_repodata.json): done
Solving environment: done
## Package Plan ##
environment location: /home/eric/.conda/envs/foo
Proceed ([y]/n)? y
Preparing transaction: done
Verifying transaction: done
Executing transaction: done
#
# To activate this environment, use
#
# $ conda activate foo
#
# To deactivate an active environment, use
#
# $ conda deactivate
$ conda env list
# conda environments:
#
foo /home/eric/.conda/envs/foo
base * /usr
~ $ conda env remove --dry-run -n foo
Remove all packages in environment /home/eric/.conda/envs/foo:
$ conda env list
# conda environments:
#
base * /usr
$
```
## Expected Behavior
`--dry-run` is documented as `Only display what would have been done.`, but clearly that's not true :upside_down_face:
The expected behaviour here is obviously that the environment is **not** removed :slightly_smiling_face:
## Environment Information
<details open><summary><code>`conda info`</code></summary><p>
```
active environment : None
shell level : 0
user config file : /home/eric/.condarc
populated config files : /usr/share/conda/condarc.d/defaults.yaml
/home/eric/.condarc
conda version : 4.8.2
conda-build version : not installed
python version : 3.8.5.final.0
virtual packages : __glibc=2.31
base environment : /usr (read only)
channel URLs : https://conda.anaconda.org/conda-forge/linux-64
https://conda.anaconda.org/conda-forge/noarch
https://repo.anaconda.com/pkgs/main/linux-64
https://repo.anaconda.com/pkgs/main/noarch
https://repo.anaconda.com/pkgs/r/linux-64
https://repo.anaconda.com/pkgs/r/noarch
package cache : /var/cache/conda/pkgs
/home/eric/.conda/pkgs
envs directories : /home/eric/.conda/envs
/usr/envs
platform : linux-64
user-agent : conda/4.8.2 requests/2.22.0 CPython/3.8.5 Linux/5.8.6-201.fc32.x86_64 fedora/32 glibc/2.31
UID:GID : 1000:1000
netrc file : None
offline mode : False
```
</p></details>
<details open><summary><code>`conda config --show-sources`</code></summary><p>
```
==> /usr/share/conda/condarc.d/defaults.yaml <==
pkgs_dirs:
- /var/cache/conda/pkgs
- ~/.conda/pkgs
==> /home/eric/.condarc <==
channels:
- conda-forge
- defaults
```
</p></details>
<details><summary><code>`conda list --show-channel-urls`</code></summary><p>
<!-- between the ticks below, paste the output of 'conda list --show-channel-urls' -->
```
EnvironmentLocationNotFound: Not a conda environment: /usr
```
(I expect that might also be a bug, conda shouldn't crash like that here.)
</p></details>
</issue>
<code>
[start of conda/cli/main_remove.py]
1 # -*- coding: utf-8 -*-
2 # Copyright (C) 2012 Anaconda, Inc
3 # SPDX-License-Identifier: BSD-3-Clause
4 from __future__ import absolute_import, division, print_function, unicode_literals
5
6 import logging
7 from os.path import isfile, join
8 import sys
9
10 from .common import check_non_admin, specs_from_args
11 from .install import handle_txn
12 from ..base.context import context
13 from ..core.envs_manager import unregister_env
14 from ..core.link import PrefixSetup, UnlinkLinkTransaction
15 from ..core.prefix_data import PrefixData
16 from ..core.solve import Solver
17 from ..exceptions import CondaEnvironmentError, CondaValueError, DirectoryNotACondaEnvironmentError
18 from ..gateways.disk.delete import rm_rf, path_is_clean
19 from ..models.match_spec import MatchSpec
20 from ..exceptions import PackagesNotFoundError
21
22 log = logging.getLogger(__name__)
23
24
25 def execute(args, parser):
26
27 if not (args.all or args.package_names):
28 raise CondaValueError('no package names supplied,\n'
29 ' try "conda remove -h" for more details')
30
31 prefix = context.target_prefix
32 check_non_admin()
33
34 if args.all and prefix == context.default_prefix:
35 msg = "cannot remove current environment. deactivate and run conda remove again"
36 raise CondaEnvironmentError(msg)
37
38 if args.all and path_is_clean(prefix):
39 # full environment removal was requested, but environment doesn't exist anyway
40
41 # .. but you know what? If you call `conda remove --all` you'd expect the dir
42 # not to exist afterwards, would you not? If not (fine, I can see the argument
43 # about deleting people's work in envs being a very bad thing indeed), but if
44 # being careful is the goal it would still be nice if after `conda remove --all`
45 # to be able to do `conda create` on the same environment name.
46 #
47 # try:
48 # rm_rf(prefix, clean_empty_parents=True)
49 # except:
50 # log.warning("Failed rm_rf() of partially existent env {}".format(prefix))
51
52 return 0
53
54 if args.all:
55 if prefix == context.root_prefix:
56 raise CondaEnvironmentError('cannot remove root environment,\n'
57 ' add -n NAME or -p PREFIX option')
58 if not isfile(join(prefix, 'conda-meta', 'history')):
59 raise DirectoryNotACondaEnvironmentError(prefix)
60 print("\nRemove all packages in environment %s:\n" % prefix, file=sys.stderr)
61
62 if 'package_names' in args:
63 stp = PrefixSetup(
64 target_prefix=prefix,
65 unlink_precs=tuple(PrefixData(prefix).iter_records()),
66 link_precs=(),
67 remove_specs=(),
68 update_specs=(),
69 neutered_specs={},
70 )
71 txn = UnlinkLinkTransaction(stp)
72 try:
73 handle_txn(txn, prefix, args, False, True)
74 except PackagesNotFoundError:
75 print("No packages found in %s. Continuing environment removal" % prefix)
76 rm_rf(prefix, clean_empty_parents=True)
77 unregister_env(prefix)
78
79 return
80
81 else:
82 if args.features:
83 specs = tuple(MatchSpec(track_features=f) for f in set(args.package_names))
84 else:
85 specs = specs_from_args(args.package_names)
86 channel_urls = ()
87 subdirs = ()
88 solver = Solver(prefix, channel_urls, subdirs, specs_to_remove=specs)
89 txn = solver.solve_for_transaction()
90 handle_txn(txn, prefix, args, False, True)
91
92 # Keep this code for dev reference until private envs can be re-enabled in
93 # Solver.solve_for_transaction
94
95 # specs = None
96 # if args.features:
97 # specs = [MatchSpec(track_features=f) for f in set(args.package_names)]
98 # actions = remove_actions(prefix, specs, index, pinned=not context.ignore_pinned)
99 # actions['ACTION'] = 'REMOVE_FEATURE'
100 # action_groups = (actions, index),
101 # elif args.all:
102 # if prefix == context.root_prefix:
103 # raise CondaEnvironmentError('cannot remove root environment,\n'
104 # ' add -n NAME or -p PREFIX option')
105 # actions = defaultdict(list)
106 # actions[PREFIX] = prefix
107 # for dist in sorted(iterkeys(index)):
108 # add_unlink(actions, dist)
109 # actions['ACTION'] = 'REMOVE_ALL'
110 # action_groups = (actions, index),
111 # elif prefix == context.root_prefix and not context.prefix_specified:
112 # from ..core.envs_manager import EnvsDirectory
113 # ed = EnvsDirectory(join(context.root_prefix, 'envs'))
114 # get_env = lambda s: ed.get_registered_preferred_env(MatchSpec(s).name)
115 # specs = specs_from_args(args.package_names)
116 # env_spec_map = groupby(get_env, specs)
117 # action_groups = []
118 # for env_name, spcs in iteritems(env_spec_map):
119 # pfx = ed.to_prefix(env_name)
120 # r = get_resolve_object(index.copy(), pfx)
121 # specs_to_remove = tuple(MatchSpec(s) for s in spcs)
122 # prune = pfx != context.root_prefix
123 # dists_for_unlinking, dists_for_linking = solve_for_actions(
124 # pfx, r,
125 # specs_to_remove=specs_to_remove, prune=prune,
126 # )
127 # actions = get_blank_actions(pfx)
128 # actions['UNLINK'].extend(dists_for_unlinking)
129 # actions['LINK'].extend(dists_for_linking)
130 # actions['SPECS'].extend(text_type(s) for s in specs_to_remove)
131 # actions['ACTION'] = 'REMOVE'
132 # action_groups.append((actions, r.index))
133 # action_groups = tuple(action_groups)
134 # else:
135 # specs = specs_from_args(args.package_names)
136 # if sys.prefix == abspath(prefix) and names_in_specs(ROOT_NO_RM, specs) and not args.force: # NOQA
137 # raise CondaEnvironmentError('cannot remove %s from root environment' %
138 # ', '.join(ROOT_NO_RM))
139 # action_groups = (remove_actions(prefix, list(specs), index=index,
140 # force=args.force,
141 # pinned=not context.ignore_pinned,
142 # ), index),
143 #
144 #
145 # delete_trash()
146 # if any(nothing_to_do(x[0]) for x in action_groups):
147 # if args.all:
148 # print("\nRemove all packages in environment %s:\n" % prefix, file=sys.stderr)
149 # if not context.json:
150 # confirm_yn(args)
151 # rm_rf(prefix)
152 #
153 # if context.json:
154 # stdout_json({
155 # 'success': True,
156 # 'actions': tuple(x[0] for x in action_groups)
157 # })
158 # return
159 #
160 # pkg = str(args.package_names).replace("['", "")
161 # pkg = pkg.replace("']", "")
162 #
163 # error_message = "No packages named '%s' found to remove from environment." % pkg
164 # raise PackageNotFoundError(error_message)
165 # if not context.json:
166 # for actions, ndx in action_groups:
167 # print()
168 # print("Package plan for package removal in environment %s:" % actions["PREFIX"])
169 # display_actions(actions, ndx)
170 # elif context.json and args.dry_run:
171 # stdout_json({
172 # 'success': True,
173 # 'dry_run': True,
174 # 'actions': tuple(x[0] for x in action_groups),
175 # })
176 # return
177 #
178 # if not context.json:
179 # confirm_yn(args)
180 #
181 # for actions, ndx in action_groups:
182 # if context.json and not context.quiet:
183 # with json_progress_bars():
184 # execute_actions(actions, ndx, verbose=not context.quiet)
185 # else:
186 # execute_actions(actions, ndx, verbose=not context.quiet)
187 #
188 # target_prefix = actions["PREFIX"]
189 # if is_private_env_path(target_prefix) and linked_data(target_prefix) == {}:
190 # rm_rf(target_prefix)
191 #
192 # if args.all:
193 # rm_rf(prefix)
194 #
195 # if context.json:
196 # stdout_json({
197 # 'success': True,
198 # 'actions': tuple(x[0] for x in action_groups),
199 # })
200
[end of conda/cli/main_remove.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/conda/cli/main_remove.py b/conda/cli/main_remove.py
--- a/conda/cli/main_remove.py
+++ b/conda/cli/main_remove.py
@@ -73,8 +73,9 @@
handle_txn(txn, prefix, args, False, True)
except PackagesNotFoundError:
print("No packages found in %s. Continuing environment removal" % prefix)
- rm_rf(prefix, clean_empty_parents=True)
- unregister_env(prefix)
+ if not context.dry_run:
+ rm_rf(prefix, clean_empty_parents=True)
+ unregister_env(prefix)
return
|
{"golden_diff": "diff --git a/conda/cli/main_remove.py b/conda/cli/main_remove.py\n--- a/conda/cli/main_remove.py\n+++ b/conda/cli/main_remove.py\n@@ -73,8 +73,9 @@\n handle_txn(txn, prefix, args, False, True)\n except PackagesNotFoundError:\n print(\"No packages found in %s. Continuing environment removal\" % prefix)\n- rm_rf(prefix, clean_empty_parents=True)\n- unregister_env(prefix)\n+ if not context.dry_run:\n+ rm_rf(prefix, clean_empty_parents=True)\n+ unregister_env(prefix)\n \n return\n", "issue": "`conda env remove --dry-run` actually deletes the env\n## Current Behavior\r\nUsing the `--dry-run` flag doesn't do a dry-run at all.\r\n\r\n\r\n### Steps to Reproduce\r\n```\r\n$ conda env list\r\n# conda environments:\r\n#\r\nbase * /usr\r\n\r\n$ conda create -n foo\r\nCollecting package metadata (current_repodata.json): done\r\nSolving environment: done\r\n\r\n## Package Plan ##\r\n\r\n environment location: /home/eric/.conda/envs/foo\r\n\r\n\r\n\r\nProceed ([y]/n)? y\r\n\r\nPreparing transaction: done\r\nVerifying transaction: done\r\nExecuting transaction: done\r\n#\r\n# To activate this environment, use\r\n#\r\n# $ conda activate foo\r\n#\r\n# To deactivate an active environment, use\r\n#\r\n# $ conda deactivate\r\n\r\n$ conda env list\r\n# conda environments:\r\n#\r\nfoo /home/eric/.conda/envs/foo\r\nbase * /usr\r\n\r\n~ $ conda env remove --dry-run -n foo\r\n\r\nRemove all packages in environment /home/eric/.conda/envs/foo:\r\n\r\n$ conda env list\r\n# conda environments:\r\n#\r\nbase * /usr\r\n\r\n$ \r\n```\r\n\r\n\r\n## Expected Behavior\r\n`--dry-run` is documented as `Only display what would have been done.`, but clearly that's not true :upside_down_face: \r\n\r\nThe expected behaviour here is obviously that the environment is **not** removed :slightly_smiling_face: \r\n\r\n\r\n## Environment Information\r\n<details open><summary><code>`conda info`</code></summary><p>\r\n\r\n```\r\n active environment : None\r\n shell level : 0\r\n user config file : /home/eric/.condarc\r\n populated config files : /usr/share/conda/condarc.d/defaults.yaml\r\n /home/eric/.condarc\r\n conda version : 4.8.2\r\n conda-build version : not installed\r\n python version : 3.8.5.final.0\r\n virtual packages : __glibc=2.31\r\n base environment : /usr (read only)\r\n channel URLs : https://conda.anaconda.org/conda-forge/linux-64\r\n https://conda.anaconda.org/conda-forge/noarch\r\n https://repo.anaconda.com/pkgs/main/linux-64\r\n https://repo.anaconda.com/pkgs/main/noarch\r\n https://repo.anaconda.com/pkgs/r/linux-64\r\n https://repo.anaconda.com/pkgs/r/noarch\r\n package cache : /var/cache/conda/pkgs\r\n /home/eric/.conda/pkgs\r\n envs directories : /home/eric/.conda/envs\r\n /usr/envs\r\n platform : linux-64\r\n user-agent : conda/4.8.2 requests/2.22.0 CPython/3.8.5 Linux/5.8.6-201.fc32.x86_64 fedora/32 glibc/2.31\r\n UID:GID : 1000:1000\r\n netrc file : None\r\n offline mode : False\r\n```\r\n</p></details>\r\n\r\n\r\n<details open><summary><code>`conda config --show-sources`</code></summary><p>\r\n\r\n```\r\n==> /usr/share/conda/condarc.d/defaults.yaml <==\r\npkgs_dirs:\r\n - /var/cache/conda/pkgs\r\n - ~/.conda/pkgs\r\n\r\n==> /home/eric/.condarc <==\r\nchannels:\r\n - conda-forge\r\n - defaults\r\n```\r\n</p></details>\r\n\r\n\r\n<details><summary><code>`conda list --show-channel-urls`</code></summary><p>\r\n<!-- between the ticks below, paste the output of 'conda list --show-channel-urls' -->\r\n\r\n```\r\nEnvironmentLocationNotFound: Not a conda environment: /usr\r\n```\r\n(I expect that might also be a bug, conda shouldn't crash like that here.)\r\n</p></details>\r\n\n", "before_files": [{"content": "# -*- coding: utf-8 -*-\n# Copyright (C) 2012 Anaconda, Inc\n# SPDX-License-Identifier: BSD-3-Clause\nfrom __future__ import absolute_import, division, print_function, unicode_literals\n\nimport logging\nfrom os.path import isfile, join\nimport sys\n\nfrom .common import check_non_admin, specs_from_args\nfrom .install import handle_txn\nfrom ..base.context import context\nfrom ..core.envs_manager import unregister_env\nfrom ..core.link import PrefixSetup, UnlinkLinkTransaction\nfrom ..core.prefix_data import PrefixData\nfrom ..core.solve import Solver\nfrom ..exceptions import CondaEnvironmentError, CondaValueError, DirectoryNotACondaEnvironmentError\nfrom ..gateways.disk.delete import rm_rf, path_is_clean\nfrom ..models.match_spec import MatchSpec\nfrom ..exceptions import PackagesNotFoundError\n\nlog = logging.getLogger(__name__)\n\n\ndef execute(args, parser):\n\n if not (args.all or args.package_names):\n raise CondaValueError('no package names supplied,\\n'\n ' try \"conda remove -h\" for more details')\n\n prefix = context.target_prefix\n check_non_admin()\n\n if args.all and prefix == context.default_prefix:\n msg = \"cannot remove current environment. deactivate and run conda remove again\"\n raise CondaEnvironmentError(msg)\n\n if args.all and path_is_clean(prefix):\n # full environment removal was requested, but environment doesn't exist anyway\n\n # .. but you know what? If you call `conda remove --all` you'd expect the dir\n # not to exist afterwards, would you not? If not (fine, I can see the argument\n # about deleting people's work in envs being a very bad thing indeed), but if\n # being careful is the goal it would still be nice if after `conda remove --all`\n # to be able to do `conda create` on the same environment name.\n #\n # try:\n # rm_rf(prefix, clean_empty_parents=True)\n # except:\n # log.warning(\"Failed rm_rf() of partially existent env {}\".format(prefix))\n\n return 0\n\n if args.all:\n if prefix == context.root_prefix:\n raise CondaEnvironmentError('cannot remove root environment,\\n'\n ' add -n NAME or -p PREFIX option')\n if not isfile(join(prefix, 'conda-meta', 'history')):\n raise DirectoryNotACondaEnvironmentError(prefix)\n print(\"\\nRemove all packages in environment %s:\\n\" % prefix, file=sys.stderr)\n\n if 'package_names' in args:\n stp = PrefixSetup(\n target_prefix=prefix,\n unlink_precs=tuple(PrefixData(prefix).iter_records()),\n link_precs=(),\n remove_specs=(),\n update_specs=(),\n neutered_specs={},\n )\n txn = UnlinkLinkTransaction(stp)\n try:\n handle_txn(txn, prefix, args, False, True)\n except PackagesNotFoundError:\n print(\"No packages found in %s. Continuing environment removal\" % prefix)\n rm_rf(prefix, clean_empty_parents=True)\n unregister_env(prefix)\n\n return\n\n else:\n if args.features:\n specs = tuple(MatchSpec(track_features=f) for f in set(args.package_names))\n else:\n specs = specs_from_args(args.package_names)\n channel_urls = ()\n subdirs = ()\n solver = Solver(prefix, channel_urls, subdirs, specs_to_remove=specs)\n txn = solver.solve_for_transaction()\n handle_txn(txn, prefix, args, False, True)\n\n # Keep this code for dev reference until private envs can be re-enabled in\n # Solver.solve_for_transaction\n\n # specs = None\n # if args.features:\n # specs = [MatchSpec(track_features=f) for f in set(args.package_names)]\n # actions = remove_actions(prefix, specs, index, pinned=not context.ignore_pinned)\n # actions['ACTION'] = 'REMOVE_FEATURE'\n # action_groups = (actions, index),\n # elif args.all:\n # if prefix == context.root_prefix:\n # raise CondaEnvironmentError('cannot remove root environment,\\n'\n # ' add -n NAME or -p PREFIX option')\n # actions = defaultdict(list)\n # actions[PREFIX] = prefix\n # for dist in sorted(iterkeys(index)):\n # add_unlink(actions, dist)\n # actions['ACTION'] = 'REMOVE_ALL'\n # action_groups = (actions, index),\n # elif prefix == context.root_prefix and not context.prefix_specified:\n # from ..core.envs_manager import EnvsDirectory\n # ed = EnvsDirectory(join(context.root_prefix, 'envs'))\n # get_env = lambda s: ed.get_registered_preferred_env(MatchSpec(s).name)\n # specs = specs_from_args(args.package_names)\n # env_spec_map = groupby(get_env, specs)\n # action_groups = []\n # for env_name, spcs in iteritems(env_spec_map):\n # pfx = ed.to_prefix(env_name)\n # r = get_resolve_object(index.copy(), pfx)\n # specs_to_remove = tuple(MatchSpec(s) for s in spcs)\n # prune = pfx != context.root_prefix\n # dists_for_unlinking, dists_for_linking = solve_for_actions(\n # pfx, r,\n # specs_to_remove=specs_to_remove, prune=prune,\n # )\n # actions = get_blank_actions(pfx)\n # actions['UNLINK'].extend(dists_for_unlinking)\n # actions['LINK'].extend(dists_for_linking)\n # actions['SPECS'].extend(text_type(s) for s in specs_to_remove)\n # actions['ACTION'] = 'REMOVE'\n # action_groups.append((actions, r.index))\n # action_groups = tuple(action_groups)\n # else:\n # specs = specs_from_args(args.package_names)\n # if sys.prefix == abspath(prefix) and names_in_specs(ROOT_NO_RM, specs) and not args.force: # NOQA\n # raise CondaEnvironmentError('cannot remove %s from root environment' %\n # ', '.join(ROOT_NO_RM))\n # action_groups = (remove_actions(prefix, list(specs), index=index,\n # force=args.force,\n # pinned=not context.ignore_pinned,\n # ), index),\n #\n #\n # delete_trash()\n # if any(nothing_to_do(x[0]) for x in action_groups):\n # if args.all:\n # print(\"\\nRemove all packages in environment %s:\\n\" % prefix, file=sys.stderr)\n # if not context.json:\n # confirm_yn(args)\n # rm_rf(prefix)\n #\n # if context.json:\n # stdout_json({\n # 'success': True,\n # 'actions': tuple(x[0] for x in action_groups)\n # })\n # return\n #\n # pkg = str(args.package_names).replace(\"['\", \"\")\n # pkg = pkg.replace(\"']\", \"\")\n #\n # error_message = \"No packages named '%s' found to remove from environment.\" % pkg\n # raise PackageNotFoundError(error_message)\n # if not context.json:\n # for actions, ndx in action_groups:\n # print()\n # print(\"Package plan for package removal in environment %s:\" % actions[\"PREFIX\"])\n # display_actions(actions, ndx)\n # elif context.json and args.dry_run:\n # stdout_json({\n # 'success': True,\n # 'dry_run': True,\n # 'actions': tuple(x[0] for x in action_groups),\n # })\n # return\n #\n # if not context.json:\n # confirm_yn(args)\n #\n # for actions, ndx in action_groups:\n # if context.json and not context.quiet:\n # with json_progress_bars():\n # execute_actions(actions, ndx, verbose=not context.quiet)\n # else:\n # execute_actions(actions, ndx, verbose=not context.quiet)\n #\n # target_prefix = actions[\"PREFIX\"]\n # if is_private_env_path(target_prefix) and linked_data(target_prefix) == {}:\n # rm_rf(target_prefix)\n #\n # if args.all:\n # rm_rf(prefix)\n #\n # if context.json:\n # stdout_json({\n # 'success': True,\n # 'actions': tuple(x[0] for x in action_groups),\n # })\n", "path": "conda/cli/main_remove.py"}]}
| 3,810 | 131 |
gh_patches_debug_21401
|
rasdani/github-patches
|
git_diff
|
ultrabug__py3status-551
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Runtime error (BrokenPipeError) helpers.py line 11
When restarting i3 using `i3 restart`, error bar pops up with message `py3status: Runtime error (BrokenPipeError) helpers.py line 11. Please try to fix this and reload i3wm (Mod+Shift+R)`
Everything appears to be functioning and the bar still shows.
Running Ubuntu 16.04
py3status 3.1
python 3.5.2
</issue>
<code>
[start of py3status/__init__.py]
1 import locale
2 import sys
3
4 from py3status.core import Py3statusWrapper
5
6 try:
7 from setproctitle import setproctitle
8 setproctitle('py3status')
9 except ImportError:
10 pass
11
12
13 def main():
14 try:
15 locale.setlocale(locale.LC_ALL, '')
16 except locale.Error:
17 print('No locale available')
18 sys.exit(2)
19
20 py3 = None
21 try:
22 py3 = Py3statusWrapper()
23 py3.setup()
24 except KeyboardInterrupt:
25 if py3:
26 py3.notify_user('Setup interrupted (KeyboardInterrupt).')
27 sys.exit(0)
28 except Exception as e:
29 if py3:
30 py3.report_exception('Setup error')
31 else:
32 # we cannot report this Exception
33 raise e
34 sys.exit(2)
35
36 try:
37 py3.run()
38 except Exception:
39 py3.report_exception('Runtime error')
40 sys.exit(3)
41 except KeyboardInterrupt:
42 pass
43 finally:
44 py3.stop()
45 sys.exit(0)
46
47
48 if __name__ == '__main__':
49 main()
50
[end of py3status/__init__.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/py3status/__init__.py b/py3status/__init__.py
--- a/py3status/__init__.py
+++ b/py3status/__init__.py
@@ -9,6 +9,13 @@
except ImportError:
pass
+try:
+ # python3
+ IOPipeError = BrokenPipeError
+except NameError:
+ # python2
+ IOPipeError = IOError
+
def main():
try:
@@ -21,9 +28,9 @@
try:
py3 = Py3statusWrapper()
py3.setup()
- except KeyboardInterrupt:
+ except (IOPipeError, KeyboardInterrupt):
if py3:
- py3.notify_user('Setup interrupted (KeyboardInterrupt).')
+ py3.notify_user('Setup interrupted')
sys.exit(0)
except Exception as e:
if py3:
@@ -35,11 +42,11 @@
try:
py3.run()
+ except (IOPipeError, KeyboardInterrupt):
+ pass
except Exception:
py3.report_exception('Runtime error')
sys.exit(3)
- except KeyboardInterrupt:
- pass
finally:
py3.stop()
sys.exit(0)
|
{"golden_diff": "diff --git a/py3status/__init__.py b/py3status/__init__.py\n--- a/py3status/__init__.py\n+++ b/py3status/__init__.py\n@@ -9,6 +9,13 @@\n except ImportError:\n pass\n \n+try:\n+ # python3\n+ IOPipeError = BrokenPipeError\n+except NameError:\n+ # python2\n+ IOPipeError = IOError\n+\n \n def main():\n try:\n@@ -21,9 +28,9 @@\n try:\n py3 = Py3statusWrapper()\n py3.setup()\n- except KeyboardInterrupt:\n+ except (IOPipeError, KeyboardInterrupt):\n if py3:\n- py3.notify_user('Setup interrupted (KeyboardInterrupt).')\n+ py3.notify_user('Setup interrupted')\n sys.exit(0)\n except Exception as e:\n if py3:\n@@ -35,11 +42,11 @@\n \n try:\n py3.run()\n+ except (IOPipeError, KeyboardInterrupt):\n+ pass\n except Exception:\n py3.report_exception('Runtime error')\n sys.exit(3)\n- except KeyboardInterrupt:\n- pass\n finally:\n py3.stop()\n sys.exit(0)\n", "issue": "Runtime error (BrokenPipeError) helpers.py line 11\nWhen restarting i3 using `i3 restart`, error bar pops up with message `py3status: Runtime error (BrokenPipeError) helpers.py line 11. Please try to fix this and reload i3wm (Mod+Shift+R)`\n\nEverything appears to be functioning and the bar still shows.\n\nRunning Ubuntu 16.04\npy3status 3.1\npython 3.5.2\n\n", "before_files": [{"content": "import locale\nimport sys\n\nfrom py3status.core import Py3statusWrapper\n\ntry:\n from setproctitle import setproctitle\n setproctitle('py3status')\nexcept ImportError:\n pass\n\n\ndef main():\n try:\n locale.setlocale(locale.LC_ALL, '')\n except locale.Error:\n print('No locale available')\n sys.exit(2)\n\n py3 = None\n try:\n py3 = Py3statusWrapper()\n py3.setup()\n except KeyboardInterrupt:\n if py3:\n py3.notify_user('Setup interrupted (KeyboardInterrupt).')\n sys.exit(0)\n except Exception as e:\n if py3:\n py3.report_exception('Setup error')\n else:\n # we cannot report this Exception\n raise e\n sys.exit(2)\n\n try:\n py3.run()\n except Exception:\n py3.report_exception('Runtime error')\n sys.exit(3)\n except KeyboardInterrupt:\n pass\n finally:\n py3.stop()\n sys.exit(0)\n\n\nif __name__ == '__main__':\n main()\n", "path": "py3status/__init__.py"}]}
| 973 | 276 |
gh_patches_debug_24621
|
rasdani/github-patches
|
git_diff
|
rlworkgroup__garage-2082
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Shape of env_info of HalfCheetah* environments not compatible with TrajectoryBatch
A trajectory of `HalfCheetahVel` and `HalfCheetahDir` should have an `env_info` dict of numpy array but they do not. The `task` in their `env_info` is instead a dict.
`env_info` is assumed by `TrajectoryBatch` to be a dict of numpy array.
This is a bug that will break an algorithm when it uses `HalfCheetah` together with `TrajectoryBatch`.
</issue>
<code>
[start of src/garage/envs/mujoco/half_cheetah_vel_env.py]
1 """Variant of the HalfCheetahEnv with different target velocity."""
2 import numpy as np
3
4 from garage.envs.mujoco.half_cheetah_env_meta_base import (
5 HalfCheetahEnvMetaBase) # noqa: E501
6
7
8 class HalfCheetahVelEnv(HalfCheetahEnvMetaBase):
9 """Half-cheetah environment with target velocity, as described in [1].
10
11 The code is adapted from
12 https://github.com/cbfinn/maml_rl/blob/9c8e2ebd741cb0c7b8bf2d040c4caeeb8e06cc95/rllab/envs/mujoco/half_cheetah_env_rand.py
13
14 The half-cheetah follows the dynamics from MuJoCo [2], and receives at each
15 time step a reward composed of a control cost and a penalty equal to the
16 difference between its current velocity and the target velocity. The tasks
17 are generated by sampling the target velocities from the uniform
18 distribution on [0, 2].
19
20 [1] Chelsea Finn, Pieter Abbeel, Sergey Levine, "Model-Agnostic
21 Meta-Learning for Fast Adaptation of Deep Networks", 2017
22 (https://arxiv.org/abs/1703.03400)
23 [2] Emanuel Todorov, Tom Erez, Yuval Tassa, "MuJoCo: A physics engine for
24 model-based control", 2012
25 (https://homes.cs.washington.edu/~todorov/papers/TodorovIROS12.pdf)
26
27 Args:
28 task (dict or None):
29 velocity (float): Target velocity, usually between 0 and 2.
30
31 """
32
33 def __init__(self, task=None):
34 super().__init__(task or {'velocity': 0.})
35
36 def step(self, action):
37 """Take one step in the environment.
38
39 Equivalent to step in HalfCheetahEnv, but with different rewards.
40
41 Args:
42 action (np.ndarray): The action to take in the environment.
43
44 Returns:
45 tuple:
46 * observation (np.ndarray): The observation of the environment.
47 * reward (float): The reward acquired at this time step.
48 * done (boolean): Whether the environment was completed at this
49 time step. Always False for this environment.
50 * infos (dict):
51 * reward_forward (float): Reward for moving, ignoring the
52 control cost.
53 * reward_ctrl (float): The reward for acting i.e. the
54 control cost (always negative).
55 * task_vel (float): Target velocity.
56 Usually between 0 and 2.
57
58 """
59 xposbefore = self.sim.data.qpos[0]
60 self.do_simulation(action, self.frame_skip)
61 xposafter = self.sim.data.qpos[0]
62
63 forward_vel = (xposafter - xposbefore) / self.dt
64 forward_reward = -1.0 * abs(forward_vel - self._task['velocity'])
65 ctrl_cost = 0.5 * 1e-1 * np.sum(np.square(action))
66
67 observation = self._get_obs()
68 reward = forward_reward - ctrl_cost
69 done = False
70 infos = dict(reward_forward=forward_reward,
71 reward_ctrl=-ctrl_cost,
72 task_vel=self._task['velocity'])
73 return observation, reward, done, infos
74
75 def sample_tasks(self, num_tasks):
76 """Sample a list of `num_tasks` tasks.
77
78 Args:
79 num_tasks (int): Number of tasks to sample.
80
81 Returns:
82 list[dict[str, float]]: A list of "tasks," where each task is a
83 dictionary containing a single key, "velocity", mapping to a
84 value between 0 and 2.
85
86 """
87 velocities = self.np_random.uniform(0.0, 2.0, size=(num_tasks, ))
88 tasks = [{'velocity': velocity} for velocity in velocities]
89 return tasks
90
91 def set_task(self, task):
92 """Reset with a task.
93
94 Args:
95 task (dict[str, float]): A task (a dictionary containing a single
96 key, "velocity", usually between 0 and 2).
97
98 """
99 self._task = task
100
[end of src/garage/envs/mujoco/half_cheetah_vel_env.py]
[start of src/garage/envs/mujoco/half_cheetah_dir_env.py]
1 """Variant of the HalfCheetahEnv with different target directions."""
2 import numpy as np
3
4 from garage.envs.mujoco.half_cheetah_env_meta_base import (
5 HalfCheetahEnvMetaBase) # noqa: E501
6
7
8 class HalfCheetahDirEnv(HalfCheetahEnvMetaBase):
9 """Half-cheetah environment with target direction, as described in [1].
10
11 The code is adapted from
12 https://github.com/cbfinn/maml_rl/blob/9c8e2ebd741cb0c7b8bf2d040c4caeeb8e06cc95/rllab/envs/mujoco/half_cheetah_env_rand_direc.py
13
14 The half-cheetah follows the dynamics from MuJoCo [2], and receives at each
15 time step a reward composed of a control cost and a reward equal to its
16 velocity in the target direction. The tasks are generated by sampling the
17 target directions from a Bernoulli distribution on {-1, 1} with parameter
18 0.5 (-1: backward, +1: forward).
19
20 [1] Chelsea Finn, Pieter Abbeel, Sergey Levine, "Model-Agnostic
21 Meta-Learning for Fast Adaptation of Deep Networks", 2017
22 (https://arxiv.org/abs/1703.03400)
23 [2] Emanuel Todorov, Tom Erez, Yuval Tassa, "MuJoCo: A physics engine for
24 model-based control", 2012
25 (https://homes.cs.washington.edu/~todorov/papers/TodorovIROS12.pdf)
26
27 Args:
28 task (dict or None):
29 direction (float): Target direction, either -1 or 1.
30
31 """
32
33 def __init__(self, task=None):
34 super().__init__(task or {'direction': 1.})
35
36 def step(self, action):
37 """Take one step in the environment.
38
39 Equivalent to step in HalfCheetahEnv, but with different rewards.
40
41 Args:
42 action (np.ndarray): The action to take in the environment.
43
44 Raises:
45 ValueError: If the current direction is not 1.0 or -1.0.
46
47 Returns:
48 tuple:
49 * observation (np.ndarray): The observation of the environment.
50 * reward (float): The reward acquired at this time step.
51 * done (boolean): Whether the environment was completed at this
52 time step. Always False for this environment.
53 * infos (dict):
54 * reward_forward (float): Reward for moving, ignoring the
55 control cost.
56 * reward_ctrl (float): The reward for acting i.e. the
57 control cost (always negative).
58 * task_dir (float): Target direction. 1.0 for forwards,
59 -1.0 for backwards.
60
61 """
62 xposbefore = self.sim.data.qpos[0]
63 self.do_simulation(action, self.frame_skip)
64 xposafter = self.sim.data.qpos[0]
65
66 forward_vel = (xposafter - xposbefore) / self.dt
67 forward_reward = self._task['direction'] * forward_vel
68 ctrl_cost = 0.5 * 1e-1 * np.sum(np.square(action))
69
70 observation = self._get_obs()
71 reward = forward_reward - ctrl_cost
72 done = False
73 if self._task['direction'] == 1.:
74 task_name = 'fowrad'
75 elif self._task['direction'] == -1.:
76 task_name = 'backward'
77 else:
78 raise ValueError('task direction should be 1. or -1.')
79 infos = dict(reward_forward=forward_reward,
80 reward_ctrl=-ctrl_cost,
81 task_dir=self._task['direction'],
82 task_name=task_name)
83 return observation, reward, done, infos
84
85 def sample_tasks(self, num_tasks):
86 """Sample a list of `num_tasks` tasks.
87
88 Args:
89 num_tasks (int): Number of tasks to sample.
90
91 Returns:
92 list[dict[str, float]]: A list of "tasks," where each task is a
93 dictionary containing a single key, "direction", mapping to -1
94 or 1.
95
96 """
97 directions = (
98 2 * self.np_random.binomial(1, p=0.5, size=(num_tasks, )) - 1)
99 tasks = [{'direction': direction} for direction in directions]
100 return tasks
101
102 def set_task(self, task):
103 """Reset with a task.
104
105 Args:
106 task (dict[str, float]): A task (a dictionary containing a single
107 key, "direction", mapping to -1 or 1).
108
109 """
110 self._task = task
111
[end of src/garage/envs/mujoco/half_cheetah_dir_env.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/src/garage/envs/mujoco/half_cheetah_dir_env.py b/src/garage/envs/mujoco/half_cheetah_dir_env.py
--- a/src/garage/envs/mujoco/half_cheetah_dir_env.py
+++ b/src/garage/envs/mujoco/half_cheetah_dir_env.py
@@ -76,9 +76,9 @@
task_name = 'backward'
else:
raise ValueError('task direction should be 1. or -1.')
- infos = dict(reward_forward=forward_reward,
- reward_ctrl=-ctrl_cost,
- task_dir=self._task['direction'],
+ infos = dict(reward_forward=np.asarray([forward_reward]),
+ reward_ctrl=np.asarray([-ctrl_cost]),
+ task_dir=np.asarray([self._task['direction']]),
task_name=task_name)
return observation, reward, done, infos
diff --git a/src/garage/envs/mujoco/half_cheetah_vel_env.py b/src/garage/envs/mujoco/half_cheetah_vel_env.py
--- a/src/garage/envs/mujoco/half_cheetah_vel_env.py
+++ b/src/garage/envs/mujoco/half_cheetah_vel_env.py
@@ -67,9 +67,9 @@
observation = self._get_obs()
reward = forward_reward - ctrl_cost
done = False
- infos = dict(reward_forward=forward_reward,
- reward_ctrl=-ctrl_cost,
- task_vel=self._task['velocity'])
+ infos = dict(reward_forward=np.asarray([forward_reward]),
+ reward_ctrl=np.asarray([-ctrl_cost]),
+ task_vel=np.asarray([self._task['velocity']]))
return observation, reward, done, infos
def sample_tasks(self, num_tasks):
|
{"golden_diff": "diff --git a/src/garage/envs/mujoco/half_cheetah_dir_env.py b/src/garage/envs/mujoco/half_cheetah_dir_env.py\n--- a/src/garage/envs/mujoco/half_cheetah_dir_env.py\n+++ b/src/garage/envs/mujoco/half_cheetah_dir_env.py\n@@ -76,9 +76,9 @@\n task_name = 'backward'\n else:\n raise ValueError('task direction should be 1. or -1.')\n- infos = dict(reward_forward=forward_reward,\n- reward_ctrl=-ctrl_cost,\n- task_dir=self._task['direction'],\n+ infos = dict(reward_forward=np.asarray([forward_reward]),\n+ reward_ctrl=np.asarray([-ctrl_cost]),\n+ task_dir=np.asarray([self._task['direction']]),\n task_name=task_name)\n return observation, reward, done, infos\n \ndiff --git a/src/garage/envs/mujoco/half_cheetah_vel_env.py b/src/garage/envs/mujoco/half_cheetah_vel_env.py\n--- a/src/garage/envs/mujoco/half_cheetah_vel_env.py\n+++ b/src/garage/envs/mujoco/half_cheetah_vel_env.py\n@@ -67,9 +67,9 @@\n observation = self._get_obs()\n reward = forward_reward - ctrl_cost\n done = False\n- infos = dict(reward_forward=forward_reward,\n- reward_ctrl=-ctrl_cost,\n- task_vel=self._task['velocity'])\n+ infos = dict(reward_forward=np.asarray([forward_reward]),\n+ reward_ctrl=np.asarray([-ctrl_cost]),\n+ task_vel=np.asarray([self._task['velocity']]))\n return observation, reward, done, infos\n \n def sample_tasks(self, num_tasks):\n", "issue": "Shape of env_info of HalfCheetah* environments not compatible with TrajectoryBatch\nA trajectory of `HalfCheetahVel` and `HalfCheetahDir` should have an `env_info` dict of numpy array but they do not. The `task` in their `env_info` is instead a dict. \r\n\r\n`env_info` is assumed by `TrajectoryBatch` to be a dict of numpy array. \r\n\r\nThis is a bug that will break an algorithm when it uses `HalfCheetah` together with `TrajectoryBatch`.\n", "before_files": [{"content": "\"\"\"Variant of the HalfCheetahEnv with different target velocity.\"\"\"\nimport numpy as np\n\nfrom garage.envs.mujoco.half_cheetah_env_meta_base import (\n HalfCheetahEnvMetaBase) # noqa: E501\n\n\nclass HalfCheetahVelEnv(HalfCheetahEnvMetaBase):\n \"\"\"Half-cheetah environment with target velocity, as described in [1].\n\n The code is adapted from\n https://github.com/cbfinn/maml_rl/blob/9c8e2ebd741cb0c7b8bf2d040c4caeeb8e06cc95/rllab/envs/mujoco/half_cheetah_env_rand.py\n\n The half-cheetah follows the dynamics from MuJoCo [2], and receives at each\n time step a reward composed of a control cost and a penalty equal to the\n difference between its current velocity and the target velocity. The tasks\n are generated by sampling the target velocities from the uniform\n distribution on [0, 2].\n\n [1] Chelsea Finn, Pieter Abbeel, Sergey Levine, \"Model-Agnostic\n Meta-Learning for Fast Adaptation of Deep Networks\", 2017\n (https://arxiv.org/abs/1703.03400)\n [2] Emanuel Todorov, Tom Erez, Yuval Tassa, \"MuJoCo: A physics engine for\n model-based control\", 2012\n (https://homes.cs.washington.edu/~todorov/papers/TodorovIROS12.pdf)\n\n Args:\n task (dict or None):\n velocity (float): Target velocity, usually between 0 and 2.\n\n \"\"\"\n\n def __init__(self, task=None):\n super().__init__(task or {'velocity': 0.})\n\n def step(self, action):\n \"\"\"Take one step in the environment.\n\n Equivalent to step in HalfCheetahEnv, but with different rewards.\n\n Args:\n action (np.ndarray): The action to take in the environment.\n\n Returns:\n tuple:\n * observation (np.ndarray): The observation of the environment.\n * reward (float): The reward acquired at this time step.\n * done (boolean): Whether the environment was completed at this\n time step. Always False for this environment.\n * infos (dict):\n * reward_forward (float): Reward for moving, ignoring the\n control cost.\n * reward_ctrl (float): The reward for acting i.e. the\n control cost (always negative).\n * task_vel (float): Target velocity.\n Usually between 0 and 2.\n\n \"\"\"\n xposbefore = self.sim.data.qpos[0]\n self.do_simulation(action, self.frame_skip)\n xposafter = self.sim.data.qpos[0]\n\n forward_vel = (xposafter - xposbefore) / self.dt\n forward_reward = -1.0 * abs(forward_vel - self._task['velocity'])\n ctrl_cost = 0.5 * 1e-1 * np.sum(np.square(action))\n\n observation = self._get_obs()\n reward = forward_reward - ctrl_cost\n done = False\n infos = dict(reward_forward=forward_reward,\n reward_ctrl=-ctrl_cost,\n task_vel=self._task['velocity'])\n return observation, reward, done, infos\n\n def sample_tasks(self, num_tasks):\n \"\"\"Sample a list of `num_tasks` tasks.\n\n Args:\n num_tasks (int): Number of tasks to sample.\n\n Returns:\n list[dict[str, float]]: A list of \"tasks,\" where each task is a\n dictionary containing a single key, \"velocity\", mapping to a\n value between 0 and 2.\n\n \"\"\"\n velocities = self.np_random.uniform(0.0, 2.0, size=(num_tasks, ))\n tasks = [{'velocity': velocity} for velocity in velocities]\n return tasks\n\n def set_task(self, task):\n \"\"\"Reset with a task.\n\n Args:\n task (dict[str, float]): A task (a dictionary containing a single\n key, \"velocity\", usually between 0 and 2).\n\n \"\"\"\n self._task = task\n", "path": "src/garage/envs/mujoco/half_cheetah_vel_env.py"}, {"content": "\"\"\"Variant of the HalfCheetahEnv with different target directions.\"\"\"\nimport numpy as np\n\nfrom garage.envs.mujoco.half_cheetah_env_meta_base import (\n HalfCheetahEnvMetaBase) # noqa: E501\n\n\nclass HalfCheetahDirEnv(HalfCheetahEnvMetaBase):\n \"\"\"Half-cheetah environment with target direction, as described in [1].\n\n The code is adapted from\n https://github.com/cbfinn/maml_rl/blob/9c8e2ebd741cb0c7b8bf2d040c4caeeb8e06cc95/rllab/envs/mujoco/half_cheetah_env_rand_direc.py\n\n The half-cheetah follows the dynamics from MuJoCo [2], and receives at each\n time step a reward composed of a control cost and a reward equal to its\n velocity in the target direction. The tasks are generated by sampling the\n target directions from a Bernoulli distribution on {-1, 1} with parameter\n 0.5 (-1: backward, +1: forward).\n\n [1] Chelsea Finn, Pieter Abbeel, Sergey Levine, \"Model-Agnostic\n Meta-Learning for Fast Adaptation of Deep Networks\", 2017\n (https://arxiv.org/abs/1703.03400)\n [2] Emanuel Todorov, Tom Erez, Yuval Tassa, \"MuJoCo: A physics engine for\n model-based control\", 2012\n (https://homes.cs.washington.edu/~todorov/papers/TodorovIROS12.pdf)\n\n Args:\n task (dict or None):\n direction (float): Target direction, either -1 or 1.\n\n \"\"\"\n\n def __init__(self, task=None):\n super().__init__(task or {'direction': 1.})\n\n def step(self, action):\n \"\"\"Take one step in the environment.\n\n Equivalent to step in HalfCheetahEnv, but with different rewards.\n\n Args:\n action (np.ndarray): The action to take in the environment.\n\n Raises:\n ValueError: If the current direction is not 1.0 or -1.0.\n\n Returns:\n tuple:\n * observation (np.ndarray): The observation of the environment.\n * reward (float): The reward acquired at this time step.\n * done (boolean): Whether the environment was completed at this\n time step. Always False for this environment.\n * infos (dict):\n * reward_forward (float): Reward for moving, ignoring the\n control cost.\n * reward_ctrl (float): The reward for acting i.e. the\n control cost (always negative).\n * task_dir (float): Target direction. 1.0 for forwards,\n -1.0 for backwards.\n\n \"\"\"\n xposbefore = self.sim.data.qpos[0]\n self.do_simulation(action, self.frame_skip)\n xposafter = self.sim.data.qpos[0]\n\n forward_vel = (xposafter - xposbefore) / self.dt\n forward_reward = self._task['direction'] * forward_vel\n ctrl_cost = 0.5 * 1e-1 * np.sum(np.square(action))\n\n observation = self._get_obs()\n reward = forward_reward - ctrl_cost\n done = False\n if self._task['direction'] == 1.:\n task_name = 'fowrad'\n elif self._task['direction'] == -1.:\n task_name = 'backward'\n else:\n raise ValueError('task direction should be 1. or -1.')\n infos = dict(reward_forward=forward_reward,\n reward_ctrl=-ctrl_cost,\n task_dir=self._task['direction'],\n task_name=task_name)\n return observation, reward, done, infos\n\n def sample_tasks(self, num_tasks):\n \"\"\"Sample a list of `num_tasks` tasks.\n\n Args:\n num_tasks (int): Number of tasks to sample.\n\n Returns:\n list[dict[str, float]]: A list of \"tasks,\" where each task is a\n dictionary containing a single key, \"direction\", mapping to -1\n or 1.\n\n \"\"\"\n directions = (\n 2 * self.np_random.binomial(1, p=0.5, size=(num_tasks, )) - 1)\n tasks = [{'direction': direction} for direction in directions]\n return tasks\n\n def set_task(self, task):\n \"\"\"Reset with a task.\n\n Args:\n task (dict[str, float]): A task (a dictionary containing a single\n key, \"direction\", mapping to -1 or 1).\n\n \"\"\"\n self._task = task\n", "path": "src/garage/envs/mujoco/half_cheetah_dir_env.py"}]}
| 3,123 | 403 |
gh_patches_debug_35140
|
rasdani/github-patches
|
git_diff
|
litestar-org__litestar-1669
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Bug: Annotated Types not being rendered in OpenAPI schema
### Description
I've got a `msgspec.Struct` request object that uses an `typing.Annotated` type as one of it's field types.
When viewing routes request object, I'm only seeing a str or null allowed but no mention of the max length piece. This also happens when I exclude the `| None` from the type declaration as well.
### URL to code causing the issue
_No response_
### MCVE
```python
from typing import Annotated
import msgspec
IDType = Annotated[str, msgspec.Meta(max_length=16)]
class Lookup(msgspec.Struct):
id: IDType | None
```
### Steps to reproduce
```bash
1. Create route with the above request object.
2. View the route in the /schema endpoint.
```
### Screenshots
_No response_
### Logs
_No response_
### Litestar Version
2.0.0a5
### Platform
- [X] Linux
- [ ] Mac
- [ ] Windows
- [ ] Other (Please specify in the description above)
</issue>
<code>
[start of litestar/_signature/field.py]
1 from __future__ import annotations
2
3 from dataclasses import dataclass
4 from typing import Any, Literal
5
6 from typing_extensions import get_args, get_origin
7
8 from litestar.constants import UNDEFINED_SENTINELS
9 from litestar.params import BodyKwarg, DependencyKwarg, ParameterKwarg
10 from litestar.types import Empty
11 from litestar.utils.predicates import (
12 is_any,
13 is_generic,
14 is_mapping,
15 is_non_string_iterable,
16 is_non_string_sequence,
17 is_optional_union,
18 is_union,
19 )
20 from litestar.utils.typing import make_non_optional_union
21
22 __all__ = ("SignatureField",)
23
24
25 @dataclass(unsafe_hash=True, frozen=True)
26 class SignatureField:
27 """Abstraction representing a model field. This class is meant to replace equivalent datastructures from other
28 libraries, for example, pydantic or msgspec.
29 """
30
31 __slots__ = (
32 "children",
33 "default_value",
34 "extra",
35 "field_type",
36 "kwarg_model",
37 "name",
38 )
39
40 children: tuple[SignatureField, ...] | None
41 """A mapping of subtypes, if any."""
42 default_value: Any
43 """Field name."""
44 extra: dict[str, Any]
45 """A mapping of extra values."""
46 field_type: Any
47 """The type of the kwarg."""
48 kwarg_model: ParameterKwarg | BodyKwarg | DependencyKwarg | None
49 """Kwarg Parameter."""
50 name: str
51 """Field name."""
52
53 @property
54 def is_empty(self) -> bool:
55 """Check if the default value is an empty type.
56
57 Returns:
58 True if the default_value is Empty or Ellipsis otherwise False.
59 """
60 return self.default_value is Empty or self.default_value is Ellipsis
61
62 @property
63 def is_optional(self) -> bool:
64 """Check if the field type is an Optional union.
65
66 Returns:
67 True if the field_type is an Optional union otherwise False.
68 """
69 return is_optional_union(self.field_type)
70
71 @property
72 def is_mapping(self) -> bool:
73 """Check if the field type is a Mapping."""
74 return is_mapping(self.field_type)
75
76 @property
77 def is_non_string_iterable(self) -> bool:
78 """Check if the field type is an Iterable.
79
80 If ``self.field_type`` is an optional union, only the non-optional members of the union are evaluated.
81
82 See: https://github.com/litestar-org/litestar/issues/1106
83 """
84 field_type = self.field_type
85 if self.is_optional:
86 field_type = make_non_optional_union(field_type)
87 return is_non_string_iterable(field_type)
88
89 @property
90 def is_non_string_sequence(self) -> bool:
91 """Check if the field type is a non-string Sequence.
92
93 If ``self.field_type`` is an optional union, only the non-optional members of the union are evaluated.
94
95 See: https://github.com/litestar-org/litestar/issues/1106
96 """
97 field_type = self.field_type
98 if self.is_optional:
99 field_type = make_non_optional_union(field_type)
100 return is_non_string_sequence(field_type)
101
102 @property
103 def is_any(self) -> bool:
104 """Check if the field type is Any."""
105 return is_any(self.field_type)
106
107 @property
108 def is_union(self) -> bool:
109 """Check if the field type is a Union."""
110 return is_union(self.field_type)
111
112 @property
113 def is_generic(self) -> bool:
114 """Check if the field type is a custom class extending Generic."""
115 return is_generic(self.field_type)
116
117 @property
118 def is_simple_type(self) -> bool:
119 """Check if the field type is a singleton value (e.g. int, str etc.)."""
120 return not (
121 self.is_generic or self.is_optional or self.is_union or self.is_mapping or self.is_non_string_iterable
122 )
123
124 @property
125 def is_parameter_field(self) -> bool:
126 """Check if the field type is a parameter kwarg value."""
127 return self.kwarg_model is not None and isinstance(self.kwarg_model, ParameterKwarg)
128
129 @property
130 def is_const(self) -> bool:
131 """Check if the field is defined as constant value."""
132 return bool(self.kwarg_model and getattr(self.kwarg_model, "const", False))
133
134 @property
135 def is_required(self) -> bool:
136 """Check if the field should be marked as a required parameter."""
137 if isinstance(self.kwarg_model, ParameterKwarg) and self.kwarg_model.required is not None:
138 return self.kwarg_model.required
139
140 return not (self.is_optional or self.is_any) and (self.is_empty or self.default_value is None)
141
142 @property
143 def is_literal(self) -> bool:
144 """Check if the field type is Literal."""
145 return get_origin(self.field_type) is Literal
146
147 @classmethod
148 def create(
149 cls,
150 field_type: Any,
151 name: str = "",
152 default_value: Any = Empty,
153 children: tuple[SignatureField, ...] | None = None,
154 kwarg_model: ParameterKwarg | BodyKwarg | DependencyKwarg | None = None,
155 extra: dict[str, Any] | None = None,
156 ) -> SignatureField:
157 """Create a new SignatureModel instance.
158
159 Args:
160 field_type: The type of the kwarg.
161 name: Field name.
162 default_value: A default value.
163 children: A mapping of subtypes, if any.
164 kwarg_model: Kwarg Parameter.
165 extra: A mapping of extra values.
166
167 Returns:
168 SignatureField instance.
169 """
170 if kwarg_model and default_value is Empty:
171 default_value = kwarg_model.default
172
173 if not children and get_origin(field_type) and (type_args := get_args(field_type)):
174 children = tuple(SignatureField.create(arg) for arg in type_args)
175
176 return SignatureField(
177 name=name,
178 field_type=field_type if field_type is not Empty else Any,
179 default_value=default_value if default_value not in UNDEFINED_SENTINELS else Empty,
180 children=children,
181 kwarg_model=kwarg_model,
182 extra=extra or {},
183 )
184
[end of litestar/_signature/field.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/litestar/_signature/field.py b/litestar/_signature/field.py
--- a/litestar/_signature/field.py
+++ b/litestar/_signature/field.py
@@ -3,7 +3,8 @@
from dataclasses import dataclass
from typing import Any, Literal
-from typing_extensions import get_args, get_origin
+from msgspec import Meta
+from typing_extensions import Annotated, get_args, get_origin
from litestar.constants import UNDEFINED_SENTINELS
from litestar.params import BodyKwarg, DependencyKwarg, ParameterKwarg
@@ -22,6 +23,26 @@
__all__ = ("SignatureField",)
+def _create_metadata_from_type(
+ value: Any, model: type[ParameterKwarg] | type[BodyKwarg], field_type: Any
+) -> ParameterKwarg | BodyKwarg | None:
+ if isinstance(value, Meta):
+ is_sequence_container = is_non_string_sequence(field_type)
+ return model(
+ gt=value.gt,
+ ge=value.ge,
+ lt=value.lt,
+ le=value.le,
+ multiple_of=value.multiple_of,
+ regex=value.pattern,
+ min_length=value.min_length if not is_sequence_container else None,
+ max_length=value.max_length if not is_sequence_container else None,
+ min_items=value.min_length if is_sequence_container else None,
+ max_items=value.max_length if is_sequence_container else None,
+ )
+ return None
+
+
@dataclass(unsafe_hash=True, frozen=True)
class SignatureField:
"""Abstraction representing a model field. This class is meant to replace equivalent datastructures from other
@@ -170,8 +191,17 @@
if kwarg_model and default_value is Empty:
default_value = kwarg_model.default
- if not children and get_origin(field_type) and (type_args := get_args(field_type)):
- children = tuple(SignatureField.create(arg) for arg in type_args)
+ origin = get_origin(field_type)
+
+ if not children and origin and (type_args := get_args(field_type)):
+ if origin is Annotated:
+ field_type = type_args[0]
+ if not kwarg_model:
+ kwarg_model = _create_metadata_from_type(
+ type_args[1], BodyKwarg if name == "data" else ParameterKwarg, field_type=field_type
+ )
+ else:
+ children = tuple(SignatureField.create(arg) for arg in type_args)
return SignatureField(
name=name,
|
{"golden_diff": "diff --git a/litestar/_signature/field.py b/litestar/_signature/field.py\n--- a/litestar/_signature/field.py\n+++ b/litestar/_signature/field.py\n@@ -3,7 +3,8 @@\n from dataclasses import dataclass\n from typing import Any, Literal\n \n-from typing_extensions import get_args, get_origin\n+from msgspec import Meta\n+from typing_extensions import Annotated, get_args, get_origin\n \n from litestar.constants import UNDEFINED_SENTINELS\n from litestar.params import BodyKwarg, DependencyKwarg, ParameterKwarg\n@@ -22,6 +23,26 @@\n __all__ = (\"SignatureField\",)\n \n \n+def _create_metadata_from_type(\n+ value: Any, model: type[ParameterKwarg] | type[BodyKwarg], field_type: Any\n+) -> ParameterKwarg | BodyKwarg | None:\n+ if isinstance(value, Meta):\n+ is_sequence_container = is_non_string_sequence(field_type)\n+ return model(\n+ gt=value.gt,\n+ ge=value.ge,\n+ lt=value.lt,\n+ le=value.le,\n+ multiple_of=value.multiple_of,\n+ regex=value.pattern,\n+ min_length=value.min_length if not is_sequence_container else None,\n+ max_length=value.max_length if not is_sequence_container else None,\n+ min_items=value.min_length if is_sequence_container else None,\n+ max_items=value.max_length if is_sequence_container else None,\n+ )\n+ return None\n+\n+\n @dataclass(unsafe_hash=True, frozen=True)\n class SignatureField:\n \"\"\"Abstraction representing a model field. This class is meant to replace equivalent datastructures from other\n@@ -170,8 +191,17 @@\n if kwarg_model and default_value is Empty:\n default_value = kwarg_model.default\n \n- if not children and get_origin(field_type) and (type_args := get_args(field_type)):\n- children = tuple(SignatureField.create(arg) for arg in type_args)\n+ origin = get_origin(field_type)\n+\n+ if not children and origin and (type_args := get_args(field_type)):\n+ if origin is Annotated:\n+ field_type = type_args[0]\n+ if not kwarg_model:\n+ kwarg_model = _create_metadata_from_type(\n+ type_args[1], BodyKwarg if name == \"data\" else ParameterKwarg, field_type=field_type\n+ )\n+ else:\n+ children = tuple(SignatureField.create(arg) for arg in type_args)\n \n return SignatureField(\n name=name,\n", "issue": "Bug: Annotated Types not being rendered in OpenAPI schema\n### Description\n\nI've got a `msgspec.Struct` request object that uses an `typing.Annotated` type as one of it's field types.\r\n\r\nWhen viewing routes request object, I'm only seeing a str or null allowed but no mention of the max length piece. This also happens when I exclude the `| None` from the type declaration as well.\n\n### URL to code causing the issue\n\n_No response_\n\n### MCVE\n\n```python\nfrom typing import Annotated\r\nimport msgspec\r\n\r\nIDType = Annotated[str, msgspec.Meta(max_length=16)]\r\n\r\nclass Lookup(msgspec.Struct):\r\n id: IDType | None\n```\n\n\n### Steps to reproduce\n\n```bash\n1. Create route with the above request object.\r\n2. View the route in the /schema endpoint.\n```\n\n\n### Screenshots\n\n_No response_\n\n### Logs\n\n_No response_\n\n### Litestar Version\n\n2.0.0a5\n\n### Platform\n\n- [X] Linux\n- [ ] Mac\n- [ ] Windows\n- [ ] Other (Please specify in the description above)\n", "before_files": [{"content": "from __future__ import annotations\n\nfrom dataclasses import dataclass\nfrom typing import Any, Literal\n\nfrom typing_extensions import get_args, get_origin\n\nfrom litestar.constants import UNDEFINED_SENTINELS\nfrom litestar.params import BodyKwarg, DependencyKwarg, ParameterKwarg\nfrom litestar.types import Empty\nfrom litestar.utils.predicates import (\n is_any,\n is_generic,\n is_mapping,\n is_non_string_iterable,\n is_non_string_sequence,\n is_optional_union,\n is_union,\n)\nfrom litestar.utils.typing import make_non_optional_union\n\n__all__ = (\"SignatureField\",)\n\n\n@dataclass(unsafe_hash=True, frozen=True)\nclass SignatureField:\n \"\"\"Abstraction representing a model field. This class is meant to replace equivalent datastructures from other\n libraries, for example, pydantic or msgspec.\n \"\"\"\n\n __slots__ = (\n \"children\",\n \"default_value\",\n \"extra\",\n \"field_type\",\n \"kwarg_model\",\n \"name\",\n )\n\n children: tuple[SignatureField, ...] | None\n \"\"\"A mapping of subtypes, if any.\"\"\"\n default_value: Any\n \"\"\"Field name.\"\"\"\n extra: dict[str, Any]\n \"\"\"A mapping of extra values.\"\"\"\n field_type: Any\n \"\"\"The type of the kwarg.\"\"\"\n kwarg_model: ParameterKwarg | BodyKwarg | DependencyKwarg | None\n \"\"\"Kwarg Parameter.\"\"\"\n name: str\n \"\"\"Field name.\"\"\"\n\n @property\n def is_empty(self) -> bool:\n \"\"\"Check if the default value is an empty type.\n\n Returns:\n True if the default_value is Empty or Ellipsis otherwise False.\n \"\"\"\n return self.default_value is Empty or self.default_value is Ellipsis\n\n @property\n def is_optional(self) -> bool:\n \"\"\"Check if the field type is an Optional union.\n\n Returns:\n True if the field_type is an Optional union otherwise False.\n \"\"\"\n return is_optional_union(self.field_type)\n\n @property\n def is_mapping(self) -> bool:\n \"\"\"Check if the field type is a Mapping.\"\"\"\n return is_mapping(self.field_type)\n\n @property\n def is_non_string_iterable(self) -> bool:\n \"\"\"Check if the field type is an Iterable.\n\n If ``self.field_type`` is an optional union, only the non-optional members of the union are evaluated.\n\n See: https://github.com/litestar-org/litestar/issues/1106\n \"\"\"\n field_type = self.field_type\n if self.is_optional:\n field_type = make_non_optional_union(field_type)\n return is_non_string_iterable(field_type)\n\n @property\n def is_non_string_sequence(self) -> bool:\n \"\"\"Check if the field type is a non-string Sequence.\n\n If ``self.field_type`` is an optional union, only the non-optional members of the union are evaluated.\n\n See: https://github.com/litestar-org/litestar/issues/1106\n \"\"\"\n field_type = self.field_type\n if self.is_optional:\n field_type = make_non_optional_union(field_type)\n return is_non_string_sequence(field_type)\n\n @property\n def is_any(self) -> bool:\n \"\"\"Check if the field type is Any.\"\"\"\n return is_any(self.field_type)\n\n @property\n def is_union(self) -> bool:\n \"\"\"Check if the field type is a Union.\"\"\"\n return is_union(self.field_type)\n\n @property\n def is_generic(self) -> bool:\n \"\"\"Check if the field type is a custom class extending Generic.\"\"\"\n return is_generic(self.field_type)\n\n @property\n def is_simple_type(self) -> bool:\n \"\"\"Check if the field type is a singleton value (e.g. int, str etc.).\"\"\"\n return not (\n self.is_generic or self.is_optional or self.is_union or self.is_mapping or self.is_non_string_iterable\n )\n\n @property\n def is_parameter_field(self) -> bool:\n \"\"\"Check if the field type is a parameter kwarg value.\"\"\"\n return self.kwarg_model is not None and isinstance(self.kwarg_model, ParameterKwarg)\n\n @property\n def is_const(self) -> bool:\n \"\"\"Check if the field is defined as constant value.\"\"\"\n return bool(self.kwarg_model and getattr(self.kwarg_model, \"const\", False))\n\n @property\n def is_required(self) -> bool:\n \"\"\"Check if the field should be marked as a required parameter.\"\"\"\n if isinstance(self.kwarg_model, ParameterKwarg) and self.kwarg_model.required is not None:\n return self.kwarg_model.required\n\n return not (self.is_optional or self.is_any) and (self.is_empty or self.default_value is None)\n\n @property\n def is_literal(self) -> bool:\n \"\"\"Check if the field type is Literal.\"\"\"\n return get_origin(self.field_type) is Literal\n\n @classmethod\n def create(\n cls,\n field_type: Any,\n name: str = \"\",\n default_value: Any = Empty,\n children: tuple[SignatureField, ...] | None = None,\n kwarg_model: ParameterKwarg | BodyKwarg | DependencyKwarg | None = None,\n extra: dict[str, Any] | None = None,\n ) -> SignatureField:\n \"\"\"Create a new SignatureModel instance.\n\n Args:\n field_type: The type of the kwarg.\n name: Field name.\n default_value: A default value.\n children: A mapping of subtypes, if any.\n kwarg_model: Kwarg Parameter.\n extra: A mapping of extra values.\n\n Returns:\n SignatureField instance.\n \"\"\"\n if kwarg_model and default_value is Empty:\n default_value = kwarg_model.default\n\n if not children and get_origin(field_type) and (type_args := get_args(field_type)):\n children = tuple(SignatureField.create(arg) for arg in type_args)\n\n return SignatureField(\n name=name,\n field_type=field_type if field_type is not Empty else Any,\n default_value=default_value if default_value not in UNDEFINED_SENTINELS else Empty,\n children=children,\n kwarg_model=kwarg_model,\n extra=extra or {},\n )\n", "path": "litestar/_signature/field.py"}]}
| 2,605 | 575 |
gh_patches_debug_34970
|
rasdani/github-patches
|
git_diff
|
vyperlang__vyper-2235
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Cannot use empty for an array inside an event log
### What's your issue about?
The following code:
```python
event Foo:
a: bytes32[2]
@external
def test():
log Foo(empty(bytes32[2]))
```
Produces the following error:
```python
File "/home/computer/code/python/vyper/vyper/parser/events.py", line 208, in pack_args_by_32
for arg2 in arg.elements:
AttributeError: 'LLLnode' object has no attribute 'elements'
```
The issue happens with any array type.
### How can it be fixed?
I think that this code should compile, so we need to look at what's happening in parser that prevents it.
</issue>
<code>
[start of vyper/parser/events.py]
1 from vyper import ast as vy_ast
2 from vyper.exceptions import TypeMismatch
3 from vyper.parser.expr import Expr
4 from vyper.parser.lll_node import LLLnode
5 from vyper.parser.parser_utils import (
6 base_type_conversion,
7 getpos,
8 make_byte_array_copier,
9 make_setter,
10 unwrap_location,
11 zero_pad,
12 )
13 from vyper.types.types import (
14 BaseType,
15 ByteArrayLike,
16 ListType,
17 get_size_of_type,
18 )
19 from vyper.utils import bytes_to_int, ceil32, keccak256
20
21
22 def pack_logging_topics(event_id, args, expected_topics, context, pos):
23 topics = [event_id]
24 code_pos = pos
25 for pos, expected_topic in enumerate(expected_topics):
26 expected_type = expected_topic.typ
27 arg = args[pos]
28 value = Expr(arg, context).lll_node
29 arg_type = value.typ
30
31 if isinstance(arg_type, ByteArrayLike) and isinstance(expected_type, ByteArrayLike):
32 if arg_type.maxlen > expected_type.maxlen:
33 raise TypeMismatch(
34 f"Topic input bytes are too big: {arg_type} {expected_type}", code_pos
35 )
36
37 if isinstance(arg, (vy_ast.Str, vy_ast.Bytes)):
38 # for literals, generate the topic at compile time
39 value = arg.value
40 if isinstance(value, str):
41 value = value.encode()
42 topics.append(bytes_to_int(keccak256(value)))
43
44 elif value.location == "memory":
45 topics.append(["sha3", ["add", value, 32], ["mload", value]])
46
47 else:
48 # storage or calldata
49 placeholder = context.new_internal_variable(value.typ)
50 placeholder_node = LLLnode.from_list(placeholder, typ=value.typ, location="memory")
51 copier = make_byte_array_copier(
52 placeholder_node,
53 LLLnode.from_list("_sub", typ=value.typ, location=value.location),
54 )
55 lll_node = [
56 "with",
57 "_sub",
58 value,
59 ["seq", copier, ["sha3", ["add", placeholder, 32], ["mload", placeholder]]],
60 ]
61 topics.append(lll_node)
62
63 elif isinstance(arg_type, ListType) and isinstance(expected_type, ListType):
64 size = get_size_of_type(value.typ) * 32
65 if value.location == "memory":
66 topics.append(["sha3", value, size])
67
68 else:
69 # storage or calldata
70 placeholder = context.new_internal_variable(value.typ)
71 placeholder_node = LLLnode.from_list(placeholder, typ=value.typ, location="memory")
72 setter = make_setter(placeholder_node, value, "memory", value.pos)
73 lll_node = ["seq", setter, ["sha3", placeholder, size]]
74 topics.append(lll_node)
75
76 else:
77 if arg_type != expected_type:
78 raise TypeMismatch(
79 f"Invalid type for logging topic, got {arg_type} expected {expected_type}",
80 value.pos,
81 )
82 value = unwrap_location(value)
83 value = base_type_conversion(value, arg_type, expected_type, pos=code_pos)
84 topics.append(value)
85
86 return topics
87
88
89 def pack_args_by_32(
90 holder,
91 maxlen,
92 arg,
93 typ,
94 context,
95 placeholder,
96 dynamic_offset_counter=None,
97 datamem_start=None,
98 pos=None,
99 ):
100 """
101 Copy necessary variables to pre-allocated memory section.
102
103 :param holder: Complete holder for all args
104 :param maxlen: Total length in bytes of the full arg section (static + dynamic).
105 :param arg: Current arg to pack
106 :param context: Context of arg
107 :param placeholder: Static placeholder for static argument part.
108 :param dynamic_offset_counter: position counter stored in static args.
109 :param dynamic_placeholder: pointer to current position in memory to write dynamic values to.
110 :param datamem_start: position where the whole datemem section starts.
111 """
112
113 if isinstance(typ, BaseType):
114 if isinstance(arg, LLLnode):
115 value = unwrap_location(arg)
116 else:
117 value = Expr(arg, context).lll_node
118 value = base_type_conversion(value, value.typ, typ, pos)
119 holder.append(LLLnode.from_list(["mstore", placeholder, value], typ=typ, location="memory"))
120 elif isinstance(typ, ByteArrayLike):
121
122 if isinstance(arg, LLLnode): # Is prealloacted variable.
123 source_lll = arg
124 else:
125 source_lll = Expr(arg, context).lll_node
126
127 # Set static offset, in arg slot.
128 holder.append(LLLnode.from_list(["mstore", placeholder, ["mload", dynamic_offset_counter]]))
129 # Get the biginning to write the ByteArray to.
130 dest_placeholder = LLLnode.from_list(
131 ["add", datamem_start, ["mload", dynamic_offset_counter]],
132 typ=typ,
133 location="memory",
134 annotation="pack_args_by_32:dest_placeholder",
135 )
136 copier = make_byte_array_copier(dest_placeholder, source_lll, pos=pos)
137 holder.append(copier)
138 # Add zero padding.
139 holder.append(zero_pad(dest_placeholder))
140
141 # Increment offset counter.
142 increment_counter = LLLnode.from_list(
143 [
144 "mstore",
145 dynamic_offset_counter,
146 [
147 "add",
148 [
149 "add",
150 ["mload", dynamic_offset_counter],
151 ["ceil32", ["mload", dest_placeholder]],
152 ],
153 32,
154 ],
155 ],
156 annotation="Increment dynamic offset counter",
157 )
158 holder.append(increment_counter)
159 elif isinstance(typ, ListType):
160 maxlen += (typ.count - 1) * 32
161 typ = typ.subtype
162
163 def check_list_type_match(provided): # Check list types match.
164 if provided != typ:
165 raise TypeMismatch(
166 f"Log list type '{provided}' does not match provided, expected '{typ}'"
167 )
168
169 # NOTE: Below code could be refactored into iterators/getter functions for each type of
170 # repetitive loop. But seeing how each one is a unique for loop, and in which way
171 # the sub value makes the difference in each type of list clearer.
172
173 # List from storage
174 if isinstance(arg, vy_ast.Attribute) and arg.value.id == "self":
175 stor_list = context.globals[arg.attr]
176 check_list_type_match(stor_list.typ.subtype)
177 size = stor_list.typ.count
178 mem_offset = 0
179 for i in range(0, size):
180 storage_offset = i
181 arg2 = LLLnode.from_list(
182 ["sload", ["add", ["sha3_32", Expr(arg, context).lll_node], storage_offset]],
183 typ=typ,
184 )
185 holder, maxlen = pack_args_by_32(
186 holder, maxlen, arg2, typ, context, placeholder + mem_offset, pos=pos,
187 )
188 mem_offset += get_size_of_type(typ) * 32
189
190 # List from variable.
191 elif isinstance(arg, vy_ast.Name):
192 size = context.vars[arg.id].size
193 pos = context.vars[arg.id].pos
194 check_list_type_match(context.vars[arg.id].typ.subtype)
195 mem_offset = 0
196 for _ in range(0, size):
197 arg2 = LLLnode.from_list(
198 pos + mem_offset, typ=typ, location=context.vars[arg.id].location
199 )
200 holder, maxlen = pack_args_by_32(
201 holder, maxlen, arg2, typ, context, placeholder + mem_offset, pos=pos,
202 )
203 mem_offset += get_size_of_type(typ) * 32
204
205 # List from list literal.
206 else:
207 mem_offset = 0
208 for arg2 in arg.elements:
209 holder, maxlen = pack_args_by_32(
210 holder, maxlen, arg2, typ, context, placeholder + mem_offset, pos=pos,
211 )
212 mem_offset += get_size_of_type(typ) * 32
213 return holder, maxlen
214
215
216 # Pack logging data arguments
217 def pack_logging_data(expected_data, args, context, pos):
218 # Checks to see if there's any data
219 if not args:
220 return ["seq"], 0, None, 0
221 holder = ["seq"]
222 maxlen = len(args) * 32 # total size of all packed args (upper limit)
223
224 # Unroll any function calls, to temp variables.
225 prealloacted = {}
226 for idx, (arg, _expected_arg) in enumerate(zip(args, expected_data)):
227
228 if isinstance(arg, (vy_ast.Str, vy_ast.Call)):
229 expr = Expr(arg, context)
230 source_lll = expr.lll_node
231 typ = source_lll.typ
232
233 if isinstance(arg, vy_ast.Str):
234 if len(arg.s) > typ.maxlen:
235 raise TypeMismatch(f"Data input bytes are to big: {len(arg.s)} {typ}", pos)
236
237 tmp_variable = context.new_internal_variable(source_lll.typ)
238 tmp_variable_node = LLLnode.from_list(
239 tmp_variable,
240 typ=source_lll.typ,
241 pos=getpos(arg),
242 location="memory",
243 annotation=f"log_prealloacted {source_lll.typ}",
244 )
245 # Store len.
246 # holder.append(['mstore', len_placeholder, ['mload', unwrap_location(source_lll)]])
247 # Copy bytes.
248
249 holder.append(
250 make_setter(tmp_variable_node, source_lll, pos=getpos(arg), location="memory")
251 )
252 prealloacted[idx] = tmp_variable_node
253
254 # Create internal variables for for dynamic and static args.
255 static_types = []
256 for data in expected_data:
257 static_types.append(data.typ if not isinstance(data.typ, ByteArrayLike) else BaseType(32))
258
259 requires_dynamic_offset = any(isinstance(data.typ, ByteArrayLike) for data in expected_data)
260
261 dynamic_offset_counter = None
262 if requires_dynamic_offset:
263 dynamic_offset_counter = context.new_internal_variable(BaseType(32))
264 dynamic_placeholder = context.new_internal_variable(BaseType(32))
265
266 static_vars = [context.new_internal_variable(i) for i in static_types]
267
268 # Populate static placeholders.
269 for i, (arg, data) in enumerate(zip(args, expected_data)):
270 typ = data.typ
271 placeholder = static_vars[i]
272 if not isinstance(typ, ByteArrayLike):
273 holder, maxlen = pack_args_by_32(
274 holder, maxlen, prealloacted.get(i, arg), typ, context, placeholder, pos=pos,
275 )
276
277 # Dynamic position starts right after the static args.
278 if requires_dynamic_offset:
279 holder.append(LLLnode.from_list(["mstore", dynamic_offset_counter, maxlen]))
280
281 # Calculate maximum dynamic offset placeholders, used for gas estimation.
282 for _arg, data in zip(args, expected_data):
283 typ = data.typ
284 if isinstance(typ, ByteArrayLike):
285 maxlen += 32 + ceil32(typ.maxlen)
286
287 if requires_dynamic_offset:
288 datamem_start = dynamic_placeholder + 32
289 else:
290 datamem_start = static_vars[0]
291
292 # Copy necessary data into allocated dynamic section.
293 for i, (arg, data) in enumerate(zip(args, expected_data)):
294 typ = data.typ
295 if isinstance(typ, ByteArrayLike):
296 pack_args_by_32(
297 holder=holder,
298 maxlen=maxlen,
299 arg=prealloacted.get(i, arg),
300 typ=typ,
301 context=context,
302 placeholder=static_vars[i],
303 datamem_start=datamem_start,
304 dynamic_offset_counter=dynamic_offset_counter,
305 pos=pos,
306 )
307
308 return holder, maxlen, dynamic_offset_counter, datamem_start
309
[end of vyper/parser/events.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/vyper/parser/events.py b/vyper/parser/events.py
--- a/vyper/parser/events.py
+++ b/vyper/parser/events.py
@@ -1,5 +1,5 @@
from vyper import ast as vy_ast
-from vyper.exceptions import TypeMismatch
+from vyper.exceptions import StructureException, TypeMismatch
from vyper.parser.expr import Expr
from vyper.parser.lll_node import LLLnode
from vyper.parser.parser_utils import (
@@ -7,6 +7,7 @@
getpos,
make_byte_array_copier,
make_setter,
+ mzero,
unwrap_location,
zero_pad,
)
@@ -157,6 +158,13 @@
)
holder.append(increment_counter)
elif isinstance(typ, ListType):
+
+ if isinstance(arg, vy_ast.Call) and arg.func.get("id") == "empty":
+ # special case for `empty()` with a static-sized array
+ holder.append(mzero(placeholder, get_size_of_type(typ) * 32))
+ maxlen += (get_size_of_type(typ) - 1) * 32
+ return holder, maxlen
+
maxlen += (typ.count - 1) * 32
typ = typ.subtype
@@ -225,7 +233,7 @@
prealloacted = {}
for idx, (arg, _expected_arg) in enumerate(zip(args, expected_data)):
- if isinstance(arg, (vy_ast.Str, vy_ast.Call)):
+ if isinstance(arg, (vy_ast.Str, vy_ast.Call)) and arg.get("func.id") != "empty":
expr = Expr(arg, context)
source_lll = expr.lll_node
typ = source_lll.typ
@@ -293,6 +301,11 @@
for i, (arg, data) in enumerate(zip(args, expected_data)):
typ = data.typ
if isinstance(typ, ByteArrayLike):
+ if isinstance(arg, vy_ast.Call) and arg.func.get("id") == "empty":
+ # TODO add support for this
+ raise StructureException(
+ "Cannot use `empty` on Bytes or String types within an event log", arg
+ )
pack_args_by_32(
holder=holder,
maxlen=maxlen,
|
{"golden_diff": "diff --git a/vyper/parser/events.py b/vyper/parser/events.py\n--- a/vyper/parser/events.py\n+++ b/vyper/parser/events.py\n@@ -1,5 +1,5 @@\n from vyper import ast as vy_ast\n-from vyper.exceptions import TypeMismatch\n+from vyper.exceptions import StructureException, TypeMismatch\n from vyper.parser.expr import Expr\n from vyper.parser.lll_node import LLLnode\n from vyper.parser.parser_utils import (\n@@ -7,6 +7,7 @@\n getpos,\n make_byte_array_copier,\n make_setter,\n+ mzero,\n unwrap_location,\n zero_pad,\n )\n@@ -157,6 +158,13 @@\n )\n holder.append(increment_counter)\n elif isinstance(typ, ListType):\n+\n+ if isinstance(arg, vy_ast.Call) and arg.func.get(\"id\") == \"empty\":\n+ # special case for `empty()` with a static-sized array\n+ holder.append(mzero(placeholder, get_size_of_type(typ) * 32))\n+ maxlen += (get_size_of_type(typ) - 1) * 32\n+ return holder, maxlen\n+\n maxlen += (typ.count - 1) * 32\n typ = typ.subtype\n \n@@ -225,7 +233,7 @@\n prealloacted = {}\n for idx, (arg, _expected_arg) in enumerate(zip(args, expected_data)):\n \n- if isinstance(arg, (vy_ast.Str, vy_ast.Call)):\n+ if isinstance(arg, (vy_ast.Str, vy_ast.Call)) and arg.get(\"func.id\") != \"empty\":\n expr = Expr(arg, context)\n source_lll = expr.lll_node\n typ = source_lll.typ\n@@ -293,6 +301,11 @@\n for i, (arg, data) in enumerate(zip(args, expected_data)):\n typ = data.typ\n if isinstance(typ, ByteArrayLike):\n+ if isinstance(arg, vy_ast.Call) and arg.func.get(\"id\") == \"empty\":\n+ # TODO add support for this\n+ raise StructureException(\n+ \"Cannot use `empty` on Bytes or String types within an event log\", arg\n+ )\n pack_args_by_32(\n holder=holder,\n maxlen=maxlen,\n", "issue": "Cannot use empty for an array inside an event log\n### What's your issue about?\r\n\r\nThe following code:\r\n\r\n```python\r\nevent Foo:\r\n a: bytes32[2]\r\n\r\n@external\r\ndef test():\r\n log Foo(empty(bytes32[2]))\r\n```\r\n\r\nProduces the following error:\r\n\r\n```python\r\nFile \"/home/computer/code/python/vyper/vyper/parser/events.py\", line 208, in pack_args_by_32\r\n for arg2 in arg.elements:\r\nAttributeError: 'LLLnode' object has no attribute 'elements'\r\n```\r\n\r\nThe issue happens with any array type.\r\n\r\n### How can it be fixed?\r\nI think that this code should compile, so we need to look at what's happening in parser that prevents it.\r\n\n", "before_files": [{"content": "from vyper import ast as vy_ast\nfrom vyper.exceptions import TypeMismatch\nfrom vyper.parser.expr import Expr\nfrom vyper.parser.lll_node import LLLnode\nfrom vyper.parser.parser_utils import (\n base_type_conversion,\n getpos,\n make_byte_array_copier,\n make_setter,\n unwrap_location,\n zero_pad,\n)\nfrom vyper.types.types import (\n BaseType,\n ByteArrayLike,\n ListType,\n get_size_of_type,\n)\nfrom vyper.utils import bytes_to_int, ceil32, keccak256\n\n\ndef pack_logging_topics(event_id, args, expected_topics, context, pos):\n topics = [event_id]\n code_pos = pos\n for pos, expected_topic in enumerate(expected_topics):\n expected_type = expected_topic.typ\n arg = args[pos]\n value = Expr(arg, context).lll_node\n arg_type = value.typ\n\n if isinstance(arg_type, ByteArrayLike) and isinstance(expected_type, ByteArrayLike):\n if arg_type.maxlen > expected_type.maxlen:\n raise TypeMismatch(\n f\"Topic input bytes are too big: {arg_type} {expected_type}\", code_pos\n )\n\n if isinstance(arg, (vy_ast.Str, vy_ast.Bytes)):\n # for literals, generate the topic at compile time\n value = arg.value\n if isinstance(value, str):\n value = value.encode()\n topics.append(bytes_to_int(keccak256(value)))\n\n elif value.location == \"memory\":\n topics.append([\"sha3\", [\"add\", value, 32], [\"mload\", value]])\n\n else:\n # storage or calldata\n placeholder = context.new_internal_variable(value.typ)\n placeholder_node = LLLnode.from_list(placeholder, typ=value.typ, location=\"memory\")\n copier = make_byte_array_copier(\n placeholder_node,\n LLLnode.from_list(\"_sub\", typ=value.typ, location=value.location),\n )\n lll_node = [\n \"with\",\n \"_sub\",\n value,\n [\"seq\", copier, [\"sha3\", [\"add\", placeholder, 32], [\"mload\", placeholder]]],\n ]\n topics.append(lll_node)\n\n elif isinstance(arg_type, ListType) and isinstance(expected_type, ListType):\n size = get_size_of_type(value.typ) * 32\n if value.location == \"memory\":\n topics.append([\"sha3\", value, size])\n\n else:\n # storage or calldata\n placeholder = context.new_internal_variable(value.typ)\n placeholder_node = LLLnode.from_list(placeholder, typ=value.typ, location=\"memory\")\n setter = make_setter(placeholder_node, value, \"memory\", value.pos)\n lll_node = [\"seq\", setter, [\"sha3\", placeholder, size]]\n topics.append(lll_node)\n\n else:\n if arg_type != expected_type:\n raise TypeMismatch(\n f\"Invalid type for logging topic, got {arg_type} expected {expected_type}\",\n value.pos,\n )\n value = unwrap_location(value)\n value = base_type_conversion(value, arg_type, expected_type, pos=code_pos)\n topics.append(value)\n\n return topics\n\n\ndef pack_args_by_32(\n holder,\n maxlen,\n arg,\n typ,\n context,\n placeholder,\n dynamic_offset_counter=None,\n datamem_start=None,\n pos=None,\n):\n \"\"\"\n Copy necessary variables to pre-allocated memory section.\n\n :param holder: Complete holder for all args\n :param maxlen: Total length in bytes of the full arg section (static + dynamic).\n :param arg: Current arg to pack\n :param context: Context of arg\n :param placeholder: Static placeholder for static argument part.\n :param dynamic_offset_counter: position counter stored in static args.\n :param dynamic_placeholder: pointer to current position in memory to write dynamic values to.\n :param datamem_start: position where the whole datemem section starts.\n \"\"\"\n\n if isinstance(typ, BaseType):\n if isinstance(arg, LLLnode):\n value = unwrap_location(arg)\n else:\n value = Expr(arg, context).lll_node\n value = base_type_conversion(value, value.typ, typ, pos)\n holder.append(LLLnode.from_list([\"mstore\", placeholder, value], typ=typ, location=\"memory\"))\n elif isinstance(typ, ByteArrayLike):\n\n if isinstance(arg, LLLnode): # Is prealloacted variable.\n source_lll = arg\n else:\n source_lll = Expr(arg, context).lll_node\n\n # Set static offset, in arg slot.\n holder.append(LLLnode.from_list([\"mstore\", placeholder, [\"mload\", dynamic_offset_counter]]))\n # Get the biginning to write the ByteArray to.\n dest_placeholder = LLLnode.from_list(\n [\"add\", datamem_start, [\"mload\", dynamic_offset_counter]],\n typ=typ,\n location=\"memory\",\n annotation=\"pack_args_by_32:dest_placeholder\",\n )\n copier = make_byte_array_copier(dest_placeholder, source_lll, pos=pos)\n holder.append(copier)\n # Add zero padding.\n holder.append(zero_pad(dest_placeholder))\n\n # Increment offset counter.\n increment_counter = LLLnode.from_list(\n [\n \"mstore\",\n dynamic_offset_counter,\n [\n \"add\",\n [\n \"add\",\n [\"mload\", dynamic_offset_counter],\n [\"ceil32\", [\"mload\", dest_placeholder]],\n ],\n 32,\n ],\n ],\n annotation=\"Increment dynamic offset counter\",\n )\n holder.append(increment_counter)\n elif isinstance(typ, ListType):\n maxlen += (typ.count - 1) * 32\n typ = typ.subtype\n\n def check_list_type_match(provided): # Check list types match.\n if provided != typ:\n raise TypeMismatch(\n f\"Log list type '{provided}' does not match provided, expected '{typ}'\"\n )\n\n # NOTE: Below code could be refactored into iterators/getter functions for each type of\n # repetitive loop. But seeing how each one is a unique for loop, and in which way\n # the sub value makes the difference in each type of list clearer.\n\n # List from storage\n if isinstance(arg, vy_ast.Attribute) and arg.value.id == \"self\":\n stor_list = context.globals[arg.attr]\n check_list_type_match(stor_list.typ.subtype)\n size = stor_list.typ.count\n mem_offset = 0\n for i in range(0, size):\n storage_offset = i\n arg2 = LLLnode.from_list(\n [\"sload\", [\"add\", [\"sha3_32\", Expr(arg, context).lll_node], storage_offset]],\n typ=typ,\n )\n holder, maxlen = pack_args_by_32(\n holder, maxlen, arg2, typ, context, placeholder + mem_offset, pos=pos,\n )\n mem_offset += get_size_of_type(typ) * 32\n\n # List from variable.\n elif isinstance(arg, vy_ast.Name):\n size = context.vars[arg.id].size\n pos = context.vars[arg.id].pos\n check_list_type_match(context.vars[arg.id].typ.subtype)\n mem_offset = 0\n for _ in range(0, size):\n arg2 = LLLnode.from_list(\n pos + mem_offset, typ=typ, location=context.vars[arg.id].location\n )\n holder, maxlen = pack_args_by_32(\n holder, maxlen, arg2, typ, context, placeholder + mem_offset, pos=pos,\n )\n mem_offset += get_size_of_type(typ) * 32\n\n # List from list literal.\n else:\n mem_offset = 0\n for arg2 in arg.elements:\n holder, maxlen = pack_args_by_32(\n holder, maxlen, arg2, typ, context, placeholder + mem_offset, pos=pos,\n )\n mem_offset += get_size_of_type(typ) * 32\n return holder, maxlen\n\n\n# Pack logging data arguments\ndef pack_logging_data(expected_data, args, context, pos):\n # Checks to see if there's any data\n if not args:\n return [\"seq\"], 0, None, 0\n holder = [\"seq\"]\n maxlen = len(args) * 32 # total size of all packed args (upper limit)\n\n # Unroll any function calls, to temp variables.\n prealloacted = {}\n for idx, (arg, _expected_arg) in enumerate(zip(args, expected_data)):\n\n if isinstance(arg, (vy_ast.Str, vy_ast.Call)):\n expr = Expr(arg, context)\n source_lll = expr.lll_node\n typ = source_lll.typ\n\n if isinstance(arg, vy_ast.Str):\n if len(arg.s) > typ.maxlen:\n raise TypeMismatch(f\"Data input bytes are to big: {len(arg.s)} {typ}\", pos)\n\n tmp_variable = context.new_internal_variable(source_lll.typ)\n tmp_variable_node = LLLnode.from_list(\n tmp_variable,\n typ=source_lll.typ,\n pos=getpos(arg),\n location=\"memory\",\n annotation=f\"log_prealloacted {source_lll.typ}\",\n )\n # Store len.\n # holder.append(['mstore', len_placeholder, ['mload', unwrap_location(source_lll)]])\n # Copy bytes.\n\n holder.append(\n make_setter(tmp_variable_node, source_lll, pos=getpos(arg), location=\"memory\")\n )\n prealloacted[idx] = tmp_variable_node\n\n # Create internal variables for for dynamic and static args.\n static_types = []\n for data in expected_data:\n static_types.append(data.typ if not isinstance(data.typ, ByteArrayLike) else BaseType(32))\n\n requires_dynamic_offset = any(isinstance(data.typ, ByteArrayLike) for data in expected_data)\n\n dynamic_offset_counter = None\n if requires_dynamic_offset:\n dynamic_offset_counter = context.new_internal_variable(BaseType(32))\n dynamic_placeholder = context.new_internal_variable(BaseType(32))\n\n static_vars = [context.new_internal_variable(i) for i in static_types]\n\n # Populate static placeholders.\n for i, (arg, data) in enumerate(zip(args, expected_data)):\n typ = data.typ\n placeholder = static_vars[i]\n if not isinstance(typ, ByteArrayLike):\n holder, maxlen = pack_args_by_32(\n holder, maxlen, prealloacted.get(i, arg), typ, context, placeholder, pos=pos,\n )\n\n # Dynamic position starts right after the static args.\n if requires_dynamic_offset:\n holder.append(LLLnode.from_list([\"mstore\", dynamic_offset_counter, maxlen]))\n\n # Calculate maximum dynamic offset placeholders, used for gas estimation.\n for _arg, data in zip(args, expected_data):\n typ = data.typ\n if isinstance(typ, ByteArrayLike):\n maxlen += 32 + ceil32(typ.maxlen)\n\n if requires_dynamic_offset:\n datamem_start = dynamic_placeholder + 32\n else:\n datamem_start = static_vars[0]\n\n # Copy necessary data into allocated dynamic section.\n for i, (arg, data) in enumerate(zip(args, expected_data)):\n typ = data.typ\n if isinstance(typ, ByteArrayLike):\n pack_args_by_32(\n holder=holder,\n maxlen=maxlen,\n arg=prealloacted.get(i, arg),\n typ=typ,\n context=context,\n placeholder=static_vars[i],\n datamem_start=datamem_start,\n dynamic_offset_counter=dynamic_offset_counter,\n pos=pos,\n )\n\n return holder, maxlen, dynamic_offset_counter, datamem_start\n", "path": "vyper/parser/events.py"}]}
| 4,064 | 509 |
gh_patches_debug_42561
|
rasdani/github-patches
|
git_diff
|
rasterio__rasterio-270
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Add pixel size and center point to rio-info
Make these standard info properties as in gdalinfo.
</issue>
<code>
[start of rasterio/warp.py]
1 """Raster warping and reprojection"""
2
3 from rasterio._warp import _reproject, _transform, _transform_geom, RESAMPLING
4 from rasterio.transform import guard_transform
5
6
7 def transform(src_crs, dst_crs, xs, ys, zs=None):
8 """
9 Transform vectors of x, y and optionally z from source
10 coordinate reference system into target.
11
12 Parameters
13 ------------
14 src_crs: dict
15 Source coordinate reference system, in rasterio dict format.
16 Example: {'init': 'EPSG:4326'}
17 dst_crs: dict
18 Target coordinate reference system.
19 xs: array_like
20 Contains x values. Will be cast to double floating point values.
21 ys: array_like
22 Contains y values.
23 zs: array_like, optional
24 Contains z values. Assumed to be all 0 if absent.
25
26 Returns
27 ---------
28 out: tuple of array_like, (xs, ys, [zs])
29 Tuple of x, y, and optionally z vectors, transformed into the target
30 coordinate reference system.
31 """
32 return _transform(src_crs, dst_crs, xs, ys, zs)
33
34
35 def transform_geom(
36 src_crs,
37 dst_crs,
38 geom,
39 antimeridian_cutting=False,
40 antimeridian_offset=10.0,
41 precision=-1):
42 """
43 Transform geometry from source coordinate reference system into target.
44
45 Parameters
46 ------------
47 src_crs: dict
48 Source coordinate reference system, in rasterio dict format.
49 Example: {'init': 'EPSG:4326'}
50 dst_crs: dict
51 Target coordinate reference system.
52 geom: GeoJSON like dict object
53 antimeridian_cutting: bool, optional
54 If True, cut geometries at the antimeridian, otherwise geometries will
55 not be cut (default).
56 antimeridian_offset: float
57 Offset from the antimeridian in degrees (default: 10) within which
58 any geometries will be split.
59 precision: float
60 If >= 0, geometry coordinates will be rounded to this number of decimal
61 places after the transform operation, otherwise original coordinate
62 values will be preserved (default).
63
64 Returns
65 ---------
66 out: GeoJSON like dict object
67 Transformed geometry in GeoJSON dict format
68 """
69
70 return _transform_geom(
71 src_crs,
72 dst_crs,
73 geom,
74 antimeridian_cutting,
75 antimeridian_offset,
76 precision)
77
78
79 def reproject(
80 source,
81 destination,
82 src_transform=None,
83 src_crs=None,
84 dst_transform=None,
85 dst_crs=None,
86 resampling=RESAMPLING.nearest,
87 **kwargs):
88 """
89 Reproject a source raster to a destination raster.
90
91 Parameters
92 ------------
93 source: ndarray or rasterio Band
94 Source raster.
95 destination: ndarray or rasterio Band
96 Target raster.
97 src_transform: affine transform object, optional
98 Source affine transformation. Required if source and destination
99 are ndarrays. Will be derived from source if it is a rasterio Band.
100 src_crs: dict, optional
101 Source coordinate reference system, in rasterio dict format.
102 Required if source and destination are ndarrays.
103 Will be derived from source if it is a rasterio Band.
104 Example: {'init': 'EPSG:4326'}
105 dst_transform: affine transform object, optional
106 Target affine transformation. Required if source and destination
107 are ndarrays. Will be derived from target if it is a rasterio Band.
108 dst_crs: dict, optional
109 Target coordinate reference system. Required if source and destination
110 are ndarrays. Will be derived from target if it is a rasterio Band.
111 resampling: int
112 Resampling method to use. One of the following:
113 RESAMPLING.nearest,
114 RESAMPLING.bilinear,
115 RESAMPLING.cubic,
116 RESAMPLING.cubic_spline,
117 RESAMPLING.lanczos,
118 RESAMPLING.average,
119 RESAMPLING.mode
120 kwargs: dict, optional
121 Additional arguments passed to transformation function.
122
123 Returns
124 ---------
125 out: None
126 Output is written to destination.
127 """
128
129 if src_transform:
130 src_transform = guard_transform(src_transform).to_gdal()
131 if dst_transform:
132 dst_transform = guard_transform(dst_transform).to_gdal()
133
134 _reproject(
135 source,
136 destination,
137 src_transform,
138 src_crs,
139 dst_transform,
140 dst_crs,
141 resampling,
142 **kwargs)
143
[end of rasterio/warp.py]
[start of rasterio/rio/info.py]
1 # Info command.
2
3 import json
4 import logging
5 import os.path
6 import pprint
7 import sys
8
9 import click
10
11 import rasterio
12 import rasterio.crs
13 from rasterio.rio.cli import cli
14
15
16 @cli.command(short_help="Print information about the rio environment.")
17 @click.option('--formats', 'key', flag_value='formats', default=True,
18 help="Enumerate the available formats.")
19 @click.pass_context
20 def env(ctx, key):
21 """Print information about the Rasterio environment: available
22 formats, etc.
23 """
24 verbosity = (ctx.obj and ctx.obj.get('verbosity')) or 1
25 logger = logging.getLogger('rio')
26 stdout = click.get_text_stream('stdout')
27 with rasterio.drivers(CPL_DEBUG=(verbosity > 2)) as env:
28 if key == 'formats':
29 for k, v in sorted(env.drivers().items()):
30 stdout.write("%s: %s\n" % (k, v))
31 stdout.write('\n')
32
33
34 @cli.command(short_help="Print information about a data file.")
35 @click.argument('input', type=click.Path(exists=True))
36 @click.option('--meta', 'aspect', flag_value='meta', default=True,
37 help="Show data file structure (default).")
38 @click.option('--tags', 'aspect', flag_value='tags',
39 help="Show data file tags.")
40 @click.option('--namespace', help="Select a tag namespace.")
41 @click.option('--indent', default=None, type=int,
42 help="Indentation level for pretty printed output")
43 # Options to pick out a single metadata item and print it as
44 # a string.
45 @click.option('--count', 'meta_member', flag_value='count',
46 help="Print the count of bands.")
47 @click.option('--dtype', 'meta_member', flag_value='dtype',
48 help="Print the dtype name.")
49 @click.option('--nodata', 'meta_member', flag_value='nodata',
50 help="Print the nodata value.")
51 @click.option('-f', '--format', '--driver', 'meta_member', flag_value='driver',
52 help="Print the format driver.")
53 @click.option('--shape', 'meta_member', flag_value='shape',
54 help="Print the (height, width) shape.")
55 @click.option('--height', 'meta_member', flag_value='height',
56 help="Print the height (number of rows).")
57 @click.option('--width', 'meta_member', flag_value='width',
58 help="Print the width (number of columns).")
59 @click.option('--crs', 'meta_member', flag_value='crs',
60 help="Print the CRS as a PROJ.4 string.")
61 @click.option('--bounds', 'meta_member', flag_value='bounds',
62 help="Print the boundary coordinates "
63 "(left, bottom, right, top).")
64 @click.pass_context
65 def info(ctx, input, aspect, indent, namespace, meta_member):
66 """Print metadata about the dataset as JSON.
67
68 Optionally print a single metadata item as a string.
69 """
70 verbosity = (ctx.obj and ctx.obj.get('verbosity')) or 1
71 logger = logging.getLogger('rio')
72 stdout = click.get_text_stream('stdout')
73 try:
74 with rasterio.drivers(CPL_DEBUG=(verbosity > 2)):
75 with rasterio.open(input, 'r-') as src:
76 info = src.meta
77 del info['affine']
78 del info['transform']
79 info['shape'] = info['height'], info['width']
80 info['bounds'] = src.bounds
81 proj4 = rasterio.crs.to_string(src.crs)
82 if proj4.startswith('+init=epsg'):
83 proj4 = proj4.split('=')[1].upper()
84 info['crs'] = proj4
85 if aspect == 'meta':
86 if meta_member:
87 if isinstance(info[meta_member], (list, tuple)):
88 print(" ".join(map(str, info[meta_member])))
89 else:
90 print(info[meta_member])
91 else:
92 stdout.write(json.dumps(info, indent=indent))
93 stdout.write("\n")
94 elif aspect == 'tags':
95 stdout.write(json.dumps(src.tags(ns=namespace),
96 indent=indent))
97 stdout.write("\n")
98 sys.exit(0)
99 except Exception:
100 logger.exception("Failed. Exception caught")
101 sys.exit(1)
102
[end of rasterio/rio/info.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/rasterio/rio/info.py b/rasterio/rio/info.py
--- a/rasterio/rio/info.py
+++ b/rasterio/rio/info.py
@@ -61,18 +61,29 @@
@click.option('--bounds', 'meta_member', flag_value='bounds',
help="Print the boundary coordinates "
"(left, bottom, right, top).")
[email protected]('--res', 'meta_member', flag_value='res',
+ help="Print pixel width and height.")
[email protected]('--lnglat', 'meta_member', flag_value='lnglat',
+ help="Print longitude and latitude at center.")
[email protected]('--stats', 'meta_member', flag_value='stats',
+ help="Print statistics (min, max, mean) of a single band "
+ "(use --bidx).")
[email protected]('-v', '--tell-me-more', '--verbose', is_flag=True,
+ help="Output extra information.")
[email protected]('--bidx', type=int, default=1,
+ help="Input file band index (default: 1).")
@click.pass_context
-def info(ctx, input, aspect, indent, namespace, meta_member):
+def info(ctx, input, aspect, indent, namespace, meta_member, verbose, bidx):
"""Print metadata about the dataset as JSON.
Optionally print a single metadata item as a string.
"""
verbosity = (ctx.obj and ctx.obj.get('verbosity')) or 1
logger = logging.getLogger('rio')
- stdout = click.get_text_stream('stdout')
+ mode = 'r' if (verbose or meta_member == 'stats') else 'r-'
try:
with rasterio.drivers(CPL_DEBUG=(verbosity > 2)):
- with rasterio.open(input, 'r-') as src:
+ with rasterio.open(input, mode) as src:
info = src.meta
del info['affine']
del info['transform']
@@ -82,19 +93,28 @@
if proj4.startswith('+init=epsg'):
proj4 = proj4.split('=')[1].upper()
info['crs'] = proj4
+ if verbose:
+ stats = [{'min': float(b.min()),
+ 'max': float(b.max()),
+ 'mean': float(b.mean())} for b in src.read()]
+ info['stats'] = stats
if aspect == 'meta':
- if meta_member:
+ if meta_member == 'stats':
+ band = src.read(bidx)
+ click.echo('%f %f %f' % (
+ float(band.min()),
+ float(band.max()),
+ float(band.mean())))
+ elif meta_member:
if isinstance(info[meta_member], (list, tuple)):
- print(" ".join(map(str, info[meta_member])))
+ click.echo(" ".join(map(str, info[meta_member])))
else:
- print(info[meta_member])
+ click.echo(info[meta_member])
else:
- stdout.write(json.dumps(info, indent=indent))
- stdout.write("\n")
+ click.echo(json.dumps(info, indent=indent))
elif aspect == 'tags':
- stdout.write(json.dumps(src.tags(ns=namespace),
+ click.echo(json.dumps(src.tags(ns=namespace),
indent=indent))
- stdout.write("\n")
sys.exit(0)
except Exception:
logger.exception("Failed. Exception caught")
diff --git a/rasterio/warp.py b/rasterio/warp.py
--- a/rasterio/warp.py
+++ b/rasterio/warp.py
@@ -1,6 +1,7 @@
"""Raster warping and reprojection"""
-from rasterio._warp import _reproject, _transform, _transform_geom, RESAMPLING
+from rasterio._base import _transform
+from rasterio._warp import _transform_geom, _reproject, RESAMPLING
from rasterio.transform import guard_transform
|
{"golden_diff": "diff --git a/rasterio/rio/info.py b/rasterio/rio/info.py\n--- a/rasterio/rio/info.py\n+++ b/rasterio/rio/info.py\n@@ -61,18 +61,29 @@\n @click.option('--bounds', 'meta_member', flag_value='bounds',\n help=\"Print the boundary coordinates \"\n \"(left, bottom, right, top).\")\[email protected]('--res', 'meta_member', flag_value='res',\n+ help=\"Print pixel width and height.\")\[email protected]('--lnglat', 'meta_member', flag_value='lnglat',\n+ help=\"Print longitude and latitude at center.\")\[email protected]('--stats', 'meta_member', flag_value='stats',\n+ help=\"Print statistics (min, max, mean) of a single band \"\n+ \"(use --bidx).\")\[email protected]('-v', '--tell-me-more', '--verbose', is_flag=True,\n+ help=\"Output extra information.\")\[email protected]('--bidx', type=int, default=1,\n+ help=\"Input file band index (default: 1).\")\n @click.pass_context\n-def info(ctx, input, aspect, indent, namespace, meta_member):\n+def info(ctx, input, aspect, indent, namespace, meta_member, verbose, bidx):\n \"\"\"Print metadata about the dataset as JSON.\n \n Optionally print a single metadata item as a string.\n \"\"\"\n verbosity = (ctx.obj and ctx.obj.get('verbosity')) or 1\n logger = logging.getLogger('rio')\n- stdout = click.get_text_stream('stdout')\n+ mode = 'r' if (verbose or meta_member == 'stats') else 'r-'\n try:\n with rasterio.drivers(CPL_DEBUG=(verbosity > 2)):\n- with rasterio.open(input, 'r-') as src:\n+ with rasterio.open(input, mode) as src:\n info = src.meta\n del info['affine']\n del info['transform']\n@@ -82,19 +93,28 @@\n if proj4.startswith('+init=epsg'):\n proj4 = proj4.split('=')[1].upper()\n info['crs'] = proj4\n+ if verbose:\n+ stats = [{'min': float(b.min()),\n+ 'max': float(b.max()),\n+ 'mean': float(b.mean())} for b in src.read()]\n+ info['stats'] = stats\n if aspect == 'meta':\n- if meta_member:\n+ if meta_member == 'stats':\n+ band = src.read(bidx)\n+ click.echo('%f %f %f' % (\n+ float(band.min()),\n+ float(band.max()),\n+ float(band.mean())))\n+ elif meta_member:\n if isinstance(info[meta_member], (list, tuple)):\n- print(\" \".join(map(str, info[meta_member])))\n+ click.echo(\" \".join(map(str, info[meta_member])))\n else:\n- print(info[meta_member])\n+ click.echo(info[meta_member])\n else:\n- stdout.write(json.dumps(info, indent=indent))\n- stdout.write(\"\\n\")\n+ click.echo(json.dumps(info, indent=indent))\n elif aspect == 'tags':\n- stdout.write(json.dumps(src.tags(ns=namespace), \n+ click.echo(json.dumps(src.tags(ns=namespace), \n indent=indent))\n- stdout.write(\"\\n\")\n sys.exit(0)\n except Exception:\n logger.exception(\"Failed. Exception caught\")\ndiff --git a/rasterio/warp.py b/rasterio/warp.py\n--- a/rasterio/warp.py\n+++ b/rasterio/warp.py\n@@ -1,6 +1,7 @@\n \"\"\"Raster warping and reprojection\"\"\"\n \n-from rasterio._warp import _reproject, _transform, _transform_geom, RESAMPLING\n+from rasterio._base import _transform\n+from rasterio._warp import _transform_geom, _reproject, RESAMPLING\n from rasterio.transform import guard_transform\n", "issue": "Add pixel size and center point to rio-info\nMake these standard info properties as in gdalinfo.\n\n", "before_files": [{"content": "\"\"\"Raster warping and reprojection\"\"\"\n\nfrom rasterio._warp import _reproject, _transform, _transform_geom, RESAMPLING\nfrom rasterio.transform import guard_transform\n\n\ndef transform(src_crs, dst_crs, xs, ys, zs=None):\n \"\"\"\n Transform vectors of x, y and optionally z from source\n coordinate reference system into target.\n\n Parameters\n ------------\n src_crs: dict\n Source coordinate reference system, in rasterio dict format.\n Example: {'init': 'EPSG:4326'}\n dst_crs: dict\n Target coordinate reference system.\n xs: array_like\n Contains x values. Will be cast to double floating point values.\n ys: array_like\n Contains y values.\n zs: array_like, optional\n Contains z values. Assumed to be all 0 if absent.\n\n Returns\n ---------\n out: tuple of array_like, (xs, ys, [zs])\n Tuple of x, y, and optionally z vectors, transformed into the target\n coordinate reference system.\n \"\"\"\n return _transform(src_crs, dst_crs, xs, ys, zs)\n\n\ndef transform_geom(\n src_crs,\n dst_crs,\n geom,\n antimeridian_cutting=False,\n antimeridian_offset=10.0,\n precision=-1):\n \"\"\"\n Transform geometry from source coordinate reference system into target.\n\n Parameters\n ------------\n src_crs: dict\n Source coordinate reference system, in rasterio dict format.\n Example: {'init': 'EPSG:4326'}\n dst_crs: dict\n Target coordinate reference system.\n geom: GeoJSON like dict object\n antimeridian_cutting: bool, optional\n If True, cut geometries at the antimeridian, otherwise geometries will\n not be cut (default).\n antimeridian_offset: float\n Offset from the antimeridian in degrees (default: 10) within which\n any geometries will be split.\n precision: float\n If >= 0, geometry coordinates will be rounded to this number of decimal\n places after the transform operation, otherwise original coordinate\n values will be preserved (default).\n\n Returns\n ---------\n out: GeoJSON like dict object\n Transformed geometry in GeoJSON dict format\n \"\"\"\n\n return _transform_geom(\n src_crs,\n dst_crs,\n geom,\n antimeridian_cutting,\n antimeridian_offset,\n precision)\n\n\ndef reproject(\n source,\n destination,\n src_transform=None,\n src_crs=None,\n dst_transform=None,\n dst_crs=None,\n resampling=RESAMPLING.nearest,\n **kwargs):\n \"\"\"\n Reproject a source raster to a destination raster.\n\n Parameters\n ------------\n source: ndarray or rasterio Band\n Source raster.\n destination: ndarray or rasterio Band\n Target raster.\n src_transform: affine transform object, optional\n Source affine transformation. Required if source and destination\n are ndarrays. Will be derived from source if it is a rasterio Band.\n src_crs: dict, optional\n Source coordinate reference system, in rasterio dict format.\n Required if source and destination are ndarrays.\n Will be derived from source if it is a rasterio Band.\n Example: {'init': 'EPSG:4326'}\n dst_transform: affine transform object, optional\n Target affine transformation. Required if source and destination\n are ndarrays. Will be derived from target if it is a rasterio Band.\n dst_crs: dict, optional\n Target coordinate reference system. Required if source and destination\n are ndarrays. Will be derived from target if it is a rasterio Band.\n resampling: int\n Resampling method to use. One of the following:\n RESAMPLING.nearest,\n RESAMPLING.bilinear,\n RESAMPLING.cubic,\n RESAMPLING.cubic_spline,\n RESAMPLING.lanczos,\n RESAMPLING.average,\n RESAMPLING.mode\n kwargs: dict, optional\n Additional arguments passed to transformation function.\n\n Returns\n ---------\n out: None\n Output is written to destination.\n \"\"\"\n\n if src_transform:\n src_transform = guard_transform(src_transform).to_gdal()\n if dst_transform:\n dst_transform = guard_transform(dst_transform).to_gdal()\n\n _reproject(\n source,\n destination,\n src_transform,\n src_crs,\n dst_transform,\n dst_crs,\n resampling,\n **kwargs)\n", "path": "rasterio/warp.py"}, {"content": "# Info command.\n\nimport json\nimport logging\nimport os.path\nimport pprint\nimport sys\n\nimport click\n\nimport rasterio\nimport rasterio.crs\nfrom rasterio.rio.cli import cli\n\n\[email protected](short_help=\"Print information about the rio environment.\")\[email protected]('--formats', 'key', flag_value='formats', default=True,\n help=\"Enumerate the available formats.\")\[email protected]_context\ndef env(ctx, key):\n \"\"\"Print information about the Rasterio environment: available\n formats, etc.\n \"\"\"\n verbosity = (ctx.obj and ctx.obj.get('verbosity')) or 1\n logger = logging.getLogger('rio')\n stdout = click.get_text_stream('stdout')\n with rasterio.drivers(CPL_DEBUG=(verbosity > 2)) as env:\n if key == 'formats':\n for k, v in sorted(env.drivers().items()):\n stdout.write(\"%s: %s\\n\" % (k, v))\n stdout.write('\\n')\n\n\[email protected](short_help=\"Print information about a data file.\")\[email protected]('input', type=click.Path(exists=True))\[email protected]('--meta', 'aspect', flag_value='meta', default=True,\n help=\"Show data file structure (default).\")\[email protected]('--tags', 'aspect', flag_value='tags',\n help=\"Show data file tags.\")\[email protected]('--namespace', help=\"Select a tag namespace.\")\[email protected]('--indent', default=None, type=int,\n help=\"Indentation level for pretty printed output\")\n# Options to pick out a single metadata item and print it as\n# a string.\[email protected]('--count', 'meta_member', flag_value='count',\n help=\"Print the count of bands.\")\[email protected]('--dtype', 'meta_member', flag_value='dtype',\n help=\"Print the dtype name.\")\[email protected]('--nodata', 'meta_member', flag_value='nodata',\n help=\"Print the nodata value.\")\[email protected]('-f', '--format', '--driver', 'meta_member', flag_value='driver',\n help=\"Print the format driver.\")\[email protected]('--shape', 'meta_member', flag_value='shape',\n help=\"Print the (height, width) shape.\")\[email protected]('--height', 'meta_member', flag_value='height',\n help=\"Print the height (number of rows).\")\[email protected]('--width', 'meta_member', flag_value='width',\n help=\"Print the width (number of columns).\")\[email protected]('--crs', 'meta_member', flag_value='crs',\n help=\"Print the CRS as a PROJ.4 string.\")\[email protected]('--bounds', 'meta_member', flag_value='bounds',\n help=\"Print the boundary coordinates \"\n \"(left, bottom, right, top).\")\[email protected]_context\ndef info(ctx, input, aspect, indent, namespace, meta_member):\n \"\"\"Print metadata about the dataset as JSON.\n\n Optionally print a single metadata item as a string.\n \"\"\"\n verbosity = (ctx.obj and ctx.obj.get('verbosity')) or 1\n logger = logging.getLogger('rio')\n stdout = click.get_text_stream('stdout')\n try:\n with rasterio.drivers(CPL_DEBUG=(verbosity > 2)):\n with rasterio.open(input, 'r-') as src:\n info = src.meta\n del info['affine']\n del info['transform']\n info['shape'] = info['height'], info['width']\n info['bounds'] = src.bounds\n proj4 = rasterio.crs.to_string(src.crs)\n if proj4.startswith('+init=epsg'):\n proj4 = proj4.split('=')[1].upper()\n info['crs'] = proj4\n if aspect == 'meta':\n if meta_member:\n if isinstance(info[meta_member], (list, tuple)):\n print(\" \".join(map(str, info[meta_member])))\n else:\n print(info[meta_member])\n else:\n stdout.write(json.dumps(info, indent=indent))\n stdout.write(\"\\n\")\n elif aspect == 'tags':\n stdout.write(json.dumps(src.tags(ns=namespace), \n indent=indent))\n stdout.write(\"\\n\")\n sys.exit(0)\n except Exception:\n logger.exception(\"Failed. Exception caught\")\n sys.exit(1)\n", "path": "rasterio/rio/info.py"}]}
| 3,021 | 875 |
gh_patches_debug_4747
|
rasdani/github-patches
|
git_diff
|
scrapy__scrapy-5006
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Remove the use of parsel.Selector._default_type
Used at https://github.com/scrapy/scrapy/blob/58ca8bbf6d1589bd0c8cc1ebda52299346f55e8a/scrapy/selector/unified.py#L72
We should stop relying on this private class variable unless there’s a good reason for it.
[Noticed](https://github.com/scrapy/parsel/pull/181/files#r562118000) while trying out [JMESPath support for Parsel](https://github.com/scrapy/parsel/pull/181) in a real life project.
</issue>
<code>
[start of scrapy/selector/unified.py]
1 """
2 XPath selectors based on lxml
3 """
4
5 from parsel import Selector as _ParselSelector
6 from scrapy.utils.trackref import object_ref
7 from scrapy.utils.python import to_bytes
8 from scrapy.http import HtmlResponse, XmlResponse
9
10
11 __all__ = ['Selector', 'SelectorList']
12
13
14 def _st(response, st):
15 if st is None:
16 return 'xml' if isinstance(response, XmlResponse) else 'html'
17 return st
18
19
20 def _response_from_text(text, st):
21 rt = XmlResponse if st == 'xml' else HtmlResponse
22 return rt(url='about:blank', encoding='utf-8',
23 body=to_bytes(text, 'utf-8'))
24
25
26 class SelectorList(_ParselSelector.selectorlist_cls, object_ref):
27 """
28 The :class:`SelectorList` class is a subclass of the builtin ``list``
29 class, which provides a few additional methods.
30 """
31
32
33 class Selector(_ParselSelector, object_ref):
34 """
35 An instance of :class:`Selector` is a wrapper over response to select
36 certain parts of its content.
37
38 ``response`` is an :class:`~scrapy.http.HtmlResponse` or an
39 :class:`~scrapy.http.XmlResponse` object that will be used for selecting
40 and extracting data.
41
42 ``text`` is a unicode string or utf-8 encoded text for cases when a
43 ``response`` isn't available. Using ``text`` and ``response`` together is
44 undefined behavior.
45
46 ``type`` defines the selector type, it can be ``"html"``, ``"xml"``
47 or ``None`` (default).
48
49 If ``type`` is ``None``, the selector automatically chooses the best type
50 based on ``response`` type (see below), or defaults to ``"html"`` in case it
51 is used together with ``text``.
52
53 If ``type`` is ``None`` and a ``response`` is passed, the selector type is
54 inferred from the response type as follows:
55
56 * ``"html"`` for :class:`~scrapy.http.HtmlResponse` type
57 * ``"xml"`` for :class:`~scrapy.http.XmlResponse` type
58 * ``"html"`` for anything else
59
60 Otherwise, if ``type`` is set, the selector type will be forced and no
61 detection will occur.
62 """
63
64 __slots__ = ['response']
65 selectorlist_cls = SelectorList
66
67 def __init__(self, response=None, text=None, type=None, root=None, **kwargs):
68 if response is not None and text is not None:
69 raise ValueError(f'{self.__class__.__name__}.__init__() received '
70 'both response and text')
71
72 st = _st(response, type or self._default_type)
73
74 if text is not None:
75 response = _response_from_text(text, st)
76
77 if response is not None:
78 text = response.text
79 kwargs.setdefault('base_url', response.url)
80
81 self.response = response
82 super().__init__(text=text, type=st, root=root, **kwargs)
83
[end of scrapy/selector/unified.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/scrapy/selector/unified.py b/scrapy/selector/unified.py
--- a/scrapy/selector/unified.py
+++ b/scrapy/selector/unified.py
@@ -69,7 +69,7 @@
raise ValueError(f'{self.__class__.__name__}.__init__() received '
'both response and text')
- st = _st(response, type or self._default_type)
+ st = _st(response, type)
if text is not None:
response = _response_from_text(text, st)
|
{"golden_diff": "diff --git a/scrapy/selector/unified.py b/scrapy/selector/unified.py\n--- a/scrapy/selector/unified.py\n+++ b/scrapy/selector/unified.py\n@@ -69,7 +69,7 @@\n raise ValueError(f'{self.__class__.__name__}.__init__() received '\n 'both response and text')\n \n- st = _st(response, type or self._default_type)\n+ st = _st(response, type)\n \n if text is not None:\n response = _response_from_text(text, st)\n", "issue": "Remove the use of parsel.Selector._default_type\nUsed at https://github.com/scrapy/scrapy/blob/58ca8bbf6d1589bd0c8cc1ebda52299346f55e8a/scrapy/selector/unified.py#L72\r\n\r\nWe should stop relying on this private class variable unless there\u2019s a good reason for it.\r\n\r\n[Noticed](https://github.com/scrapy/parsel/pull/181/files#r562118000) while trying out [JMESPath support for Parsel](https://github.com/scrapy/parsel/pull/181) in a real life project.\n", "before_files": [{"content": "\"\"\"\nXPath selectors based on lxml\n\"\"\"\n\nfrom parsel import Selector as _ParselSelector\nfrom scrapy.utils.trackref import object_ref\nfrom scrapy.utils.python import to_bytes\nfrom scrapy.http import HtmlResponse, XmlResponse\n\n\n__all__ = ['Selector', 'SelectorList']\n\n\ndef _st(response, st):\n if st is None:\n return 'xml' if isinstance(response, XmlResponse) else 'html'\n return st\n\n\ndef _response_from_text(text, st):\n rt = XmlResponse if st == 'xml' else HtmlResponse\n return rt(url='about:blank', encoding='utf-8',\n body=to_bytes(text, 'utf-8'))\n\n\nclass SelectorList(_ParselSelector.selectorlist_cls, object_ref):\n \"\"\"\n The :class:`SelectorList` class is a subclass of the builtin ``list``\n class, which provides a few additional methods.\n \"\"\"\n\n\nclass Selector(_ParselSelector, object_ref):\n \"\"\"\n An instance of :class:`Selector` is a wrapper over response to select\n certain parts of its content.\n\n ``response`` is an :class:`~scrapy.http.HtmlResponse` or an\n :class:`~scrapy.http.XmlResponse` object that will be used for selecting\n and extracting data.\n\n ``text`` is a unicode string or utf-8 encoded text for cases when a\n ``response`` isn't available. Using ``text`` and ``response`` together is\n undefined behavior.\n\n ``type`` defines the selector type, it can be ``\"html\"``, ``\"xml\"``\n or ``None`` (default).\n\n If ``type`` is ``None``, the selector automatically chooses the best type\n based on ``response`` type (see below), or defaults to ``\"html\"`` in case it\n is used together with ``text``.\n\n If ``type`` is ``None`` and a ``response`` is passed, the selector type is\n inferred from the response type as follows:\n\n * ``\"html\"`` for :class:`~scrapy.http.HtmlResponse` type\n * ``\"xml\"`` for :class:`~scrapy.http.XmlResponse` type\n * ``\"html\"`` for anything else\n\n Otherwise, if ``type`` is set, the selector type will be forced and no\n detection will occur.\n \"\"\"\n\n __slots__ = ['response']\n selectorlist_cls = SelectorList\n\n def __init__(self, response=None, text=None, type=None, root=None, **kwargs):\n if response is not None and text is not None:\n raise ValueError(f'{self.__class__.__name__}.__init__() received '\n 'both response and text')\n\n st = _st(response, type or self._default_type)\n\n if text is not None:\n response = _response_from_text(text, st)\n\n if response is not None:\n text = response.text\n kwargs.setdefault('base_url', response.url)\n\n self.response = response\n super().__init__(text=text, type=st, root=root, **kwargs)\n", "path": "scrapy/selector/unified.py"}]}
| 1,519 | 121 |
gh_patches_debug_2707
|
rasdani/github-patches
|
git_diff
|
DataDog__dd-trace-py-1582
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
ddtrace.Pin() for multiple grpc channels doesn't work
Thanks for taking the time for reporting an issue!
Before reporting an issue on dd-trace-py, please be sure to provide all
necessary information.
If you're hitting a bug, make sure that you're using the latest version of this
library.
### Which version of dd-trace-py are you using?
0.38.2
I didn't find anything related to this issue in the release notes of the releases after this version.
### Which version of the libraries are you using?
datadog==0.36.0
### How can we reproduce your problem?
Approach 1:
servers is a list of grpc server addresses
```
for server in servers:
channel = grpc.insecure_channel(server)
Pin.override(channel, service=server)
# Do something with the channel
```
Since `Pin.override(grpc.Channel, service=server)` worked with one server, I also tried the following to see how it looks
Approach 2:
servers is a list of grpc server addresses
```
for server in servers:
Pin.override(grpc.Channel, service=server)
channel = grpc.insecure_channel(server)
# Do something with the channel
```
### What is the result that you get?
In Approach 1, Pin.override did not set the service name correctly. Everywhere in Datadog, I could see it as `grpc-client` which is the default value.
In Approach 2, since I I don't pass the channels corresponding to each server, all servers are overriden by Pin to the final server (probably because it's the last one in the loop)
### What is the result that you expected?
ddtrace.Pin() onto multiple grpc channels should work and I should be able to see the correct `service` in Datadog APM traces and Service Map
</issue>
<code>
[start of ddtrace/contrib/grpc/patch.py]
1 import os
2
3 import grpc
4
5 from ddtrace.vendor.wrapt import wrap_function_wrapper as _w
6 from ddtrace import config, Pin
7
8 from ...utils.wrappers import unwrap as _u
9
10 from . import constants
11 from .client_interceptor import create_client_interceptor, intercept_channel
12 from .server_interceptor import create_server_interceptor
13
14
15 config._add('grpc_server', dict(
16 service_name=config._get_service(default=constants.GRPC_SERVICE_SERVER),
17 distributed_tracing_enabled=True,
18 ))
19
20
21 # Precedence for the service name:
22 # 1) DD_GRPC_SERVICE if defined; or
23 # 2) For compatibility, the globally set service + "-grpc-client"; or
24 # 3) The fall-back "grpc-client"
25 if "DD_GRPC_SERVICE" in os.environ:
26 service = os.getenv("DD_GRPC_SERVICE")
27 elif config._get_service():
28 service = "{}-{}".format(config._get_service(), constants.GRPC_SERVICE_CLIENT)
29 else:
30 service = constants.GRPC_SERVICE_CLIENT
31
32
33 # TODO[tbutt]: keeping name for client config unchanged to maintain backwards
34 # compatibility but should change in future
35 config._add('grpc', dict(
36 service_name=service,
37 distributed_tracing_enabled=True,
38 ))
39
40
41 def patch():
42 _patch_client()
43 _patch_server()
44
45
46 def unpatch():
47 _unpatch_client()
48 _unpatch_server()
49
50
51 def _patch_client():
52 if getattr(constants.GRPC_PIN_MODULE_CLIENT, '__datadog_patch', False):
53 return
54 setattr(constants.GRPC_PIN_MODULE_CLIENT, '__datadog_patch', True)
55
56 Pin().onto(constants.GRPC_PIN_MODULE_CLIENT)
57
58 _w('grpc', 'insecure_channel', _client_channel_interceptor)
59 _w('grpc', 'secure_channel', _client_channel_interceptor)
60 _w('grpc', 'intercept_channel', intercept_channel)
61
62
63 def _unpatch_client():
64 if not getattr(constants.GRPC_PIN_MODULE_CLIENT, '__datadog_patch', False):
65 return
66 setattr(constants.GRPC_PIN_MODULE_CLIENT, '__datadog_patch', False)
67
68 pin = Pin.get_from(constants.GRPC_PIN_MODULE_CLIENT)
69 if pin:
70 pin.remove_from(constants.GRPC_PIN_MODULE_CLIENT)
71
72 _u(grpc, 'secure_channel')
73 _u(grpc, 'insecure_channel')
74
75
76 def _patch_server():
77 if getattr(constants.GRPC_PIN_MODULE_SERVER, '__datadog_patch', False):
78 return
79 setattr(constants.GRPC_PIN_MODULE_SERVER, '__datadog_patch', True)
80
81 Pin().onto(constants.GRPC_PIN_MODULE_SERVER)
82
83 _w('grpc', 'server', _server_constructor_interceptor)
84
85
86 def _unpatch_server():
87 if not getattr(constants.GRPC_PIN_MODULE_SERVER, '__datadog_patch', False):
88 return
89 setattr(constants.GRPC_PIN_MODULE_SERVER, '__datadog_patch', False)
90
91 pin = Pin.get_from(constants.GRPC_PIN_MODULE_SERVER)
92 if pin:
93 pin.remove_from(constants.GRPC_PIN_MODULE_SERVER)
94
95 _u(grpc, 'server')
96
97
98 def _client_channel_interceptor(wrapped, instance, args, kwargs):
99 channel = wrapped(*args, **kwargs)
100
101 pin = Pin.get_from(constants.GRPC_PIN_MODULE_CLIENT)
102 if not pin or not pin.enabled():
103 return channel
104
105 (host, port) = _parse_target_from_arguments(args, kwargs)
106
107 interceptor_function = create_client_interceptor(pin, host, port)
108 return grpc.intercept_channel(channel, interceptor_function)
109
110
111 def _server_constructor_interceptor(wrapped, instance, args, kwargs):
112 # DEV: we clone the pin on the grpc module and configure it for the server
113 # interceptor
114
115 pin = Pin.get_from(constants.GRPC_PIN_MODULE_SERVER)
116 if not pin or not pin.enabled():
117 return wrapped(*args, **kwargs)
118
119 interceptor = create_server_interceptor(pin)
120
121 # DEV: Inject our tracing interceptor first in the list of interceptors
122 if 'interceptors' in kwargs:
123 kwargs['interceptors'] = (interceptor,) + tuple(kwargs['interceptors'])
124 else:
125 kwargs['interceptors'] = (interceptor,)
126
127 return wrapped(*args, **kwargs)
128
129
130 def _parse_target_from_arguments(args, kwargs):
131 if 'target' in kwargs:
132 target = kwargs['target']
133 else:
134 target = args[0]
135
136 split = target.rsplit(':', 2)
137
138 return (split[0], split[1] if len(split) > 1 else None)
139
[end of ddtrace/contrib/grpc/patch.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/ddtrace/contrib/grpc/patch.py b/ddtrace/contrib/grpc/patch.py
--- a/ddtrace/contrib/grpc/patch.py
+++ b/ddtrace/contrib/grpc/patch.py
@@ -98,7 +98,7 @@
def _client_channel_interceptor(wrapped, instance, args, kwargs):
channel = wrapped(*args, **kwargs)
- pin = Pin.get_from(constants.GRPC_PIN_MODULE_CLIENT)
+ pin = Pin.get_from(channel)
if not pin or not pin.enabled():
return channel
|
{"golden_diff": "diff --git a/ddtrace/contrib/grpc/patch.py b/ddtrace/contrib/grpc/patch.py\n--- a/ddtrace/contrib/grpc/patch.py\n+++ b/ddtrace/contrib/grpc/patch.py\n@@ -98,7 +98,7 @@\n def _client_channel_interceptor(wrapped, instance, args, kwargs):\n channel = wrapped(*args, **kwargs)\n \n- pin = Pin.get_from(constants.GRPC_PIN_MODULE_CLIENT)\n+ pin = Pin.get_from(channel)\n if not pin or not pin.enabled():\n return channel\n", "issue": "ddtrace.Pin() for multiple grpc channels doesn't work\nThanks for taking the time for reporting an issue!\r\n\r\nBefore reporting an issue on dd-trace-py, please be sure to provide all\r\nnecessary information.\r\n\r\nIf you're hitting a bug, make sure that you're using the latest version of this\r\nlibrary.\r\n\r\n### Which version of dd-trace-py are you using?\r\n0.38.2\r\nI didn't find anything related to this issue in the release notes of the releases after this version.\r\n\r\n### Which version of the libraries are you using?\r\ndatadog==0.36.0\r\n\r\n### How can we reproduce your problem?\r\nApproach 1:\r\nservers is a list of grpc server addresses\r\n```\r\nfor server in servers:\r\n channel = grpc.insecure_channel(server)\r\n Pin.override(channel, service=server)\r\n # Do something with the channel\r\n```\r\n\r\nSince `Pin.override(grpc.Channel, service=server)` worked with one server, I also tried the following to see how it looks\r\nApproach 2:\r\nservers is a list of grpc server addresses\r\n```\r\nfor server in servers:\r\n Pin.override(grpc.Channel, service=server)\r\n channel = grpc.insecure_channel(server)\r\n # Do something with the channel\r\n```\r\n\r\n### What is the result that you get?\r\nIn Approach 1, Pin.override did not set the service name correctly. Everywhere in Datadog, I could see it as `grpc-client` which is the default value.\r\nIn Approach 2, since I I don't pass the channels corresponding to each server, all servers are overriden by Pin to the final server (probably because it's the last one in the loop)\r\n\r\n### What is the result that you expected?\r\nddtrace.Pin() onto multiple grpc channels should work and I should be able to see the correct `service` in Datadog APM traces and Service Map\n", "before_files": [{"content": "import os\n\nimport grpc\n\nfrom ddtrace.vendor.wrapt import wrap_function_wrapper as _w\nfrom ddtrace import config, Pin\n\nfrom ...utils.wrappers import unwrap as _u\n\nfrom . import constants\nfrom .client_interceptor import create_client_interceptor, intercept_channel\nfrom .server_interceptor import create_server_interceptor\n\n\nconfig._add('grpc_server', dict(\n service_name=config._get_service(default=constants.GRPC_SERVICE_SERVER),\n distributed_tracing_enabled=True,\n))\n\n\n# Precedence for the service name:\n# 1) DD_GRPC_SERVICE if defined; or\n# 2) For compatibility, the globally set service + \"-grpc-client\"; or\n# 3) The fall-back \"grpc-client\"\nif \"DD_GRPC_SERVICE\" in os.environ:\n service = os.getenv(\"DD_GRPC_SERVICE\")\nelif config._get_service():\n service = \"{}-{}\".format(config._get_service(), constants.GRPC_SERVICE_CLIENT)\nelse:\n service = constants.GRPC_SERVICE_CLIENT\n\n\n# TODO[tbutt]: keeping name for client config unchanged to maintain backwards\n# compatibility but should change in future\nconfig._add('grpc', dict(\n service_name=service,\n distributed_tracing_enabled=True,\n))\n\n\ndef patch():\n _patch_client()\n _patch_server()\n\n\ndef unpatch():\n _unpatch_client()\n _unpatch_server()\n\n\ndef _patch_client():\n if getattr(constants.GRPC_PIN_MODULE_CLIENT, '__datadog_patch', False):\n return\n setattr(constants.GRPC_PIN_MODULE_CLIENT, '__datadog_patch', True)\n\n Pin().onto(constants.GRPC_PIN_MODULE_CLIENT)\n\n _w('grpc', 'insecure_channel', _client_channel_interceptor)\n _w('grpc', 'secure_channel', _client_channel_interceptor)\n _w('grpc', 'intercept_channel', intercept_channel)\n\n\ndef _unpatch_client():\n if not getattr(constants.GRPC_PIN_MODULE_CLIENT, '__datadog_patch', False):\n return\n setattr(constants.GRPC_PIN_MODULE_CLIENT, '__datadog_patch', False)\n\n pin = Pin.get_from(constants.GRPC_PIN_MODULE_CLIENT)\n if pin:\n pin.remove_from(constants.GRPC_PIN_MODULE_CLIENT)\n\n _u(grpc, 'secure_channel')\n _u(grpc, 'insecure_channel')\n\n\ndef _patch_server():\n if getattr(constants.GRPC_PIN_MODULE_SERVER, '__datadog_patch', False):\n return\n setattr(constants.GRPC_PIN_MODULE_SERVER, '__datadog_patch', True)\n\n Pin().onto(constants.GRPC_PIN_MODULE_SERVER)\n\n _w('grpc', 'server', _server_constructor_interceptor)\n\n\ndef _unpatch_server():\n if not getattr(constants.GRPC_PIN_MODULE_SERVER, '__datadog_patch', False):\n return\n setattr(constants.GRPC_PIN_MODULE_SERVER, '__datadog_patch', False)\n\n pin = Pin.get_from(constants.GRPC_PIN_MODULE_SERVER)\n if pin:\n pin.remove_from(constants.GRPC_PIN_MODULE_SERVER)\n\n _u(grpc, 'server')\n\n\ndef _client_channel_interceptor(wrapped, instance, args, kwargs):\n channel = wrapped(*args, **kwargs)\n\n pin = Pin.get_from(constants.GRPC_PIN_MODULE_CLIENT)\n if not pin or not pin.enabled():\n return channel\n\n (host, port) = _parse_target_from_arguments(args, kwargs)\n\n interceptor_function = create_client_interceptor(pin, host, port)\n return grpc.intercept_channel(channel, interceptor_function)\n\n\ndef _server_constructor_interceptor(wrapped, instance, args, kwargs):\n # DEV: we clone the pin on the grpc module and configure it for the server\n # interceptor\n\n pin = Pin.get_from(constants.GRPC_PIN_MODULE_SERVER)\n if not pin or not pin.enabled():\n return wrapped(*args, **kwargs)\n\n interceptor = create_server_interceptor(pin)\n\n # DEV: Inject our tracing interceptor first in the list of interceptors\n if 'interceptors' in kwargs:\n kwargs['interceptors'] = (interceptor,) + tuple(kwargs['interceptors'])\n else:\n kwargs['interceptors'] = (interceptor,)\n\n return wrapped(*args, **kwargs)\n\n\ndef _parse_target_from_arguments(args, kwargs):\n if 'target' in kwargs:\n target = kwargs['target']\n else:\n target = args[0]\n\n split = target.rsplit(':', 2)\n\n return (split[0], split[1] if len(split) > 1 else None)\n", "path": "ddtrace/contrib/grpc/patch.py"}]}
| 2,217 | 118 |
gh_patches_debug_19885
|
rasdani/github-patches
|
git_diff
|
mlcommons__GaNDLF-498
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Per-label accuracy does not work for multiple batches
**Describe the bug**
When `batch_size > 1`, `per_label_accuracy` computation fails.
**To Reproduce**
Steps to reproduce the behavior:
1. Set `batch_size = 4` in any classification unit test
2. See error
**Expected behavior**
The function should compute multiple batches of accuracies.
**Screenshots**
N.A.
**GaNDLF Version**
<!-- Put the output of the following command:
python -c 'import GANDLF as g;print(g.__version__)'
-->
0.0.15-dev
**Desktop (please complete the following information):**
N.A.
**Additional context**
Reported by @brandon-edwards
</issue>
<code>
[start of GANDLF/metrics/regression.py]
1 """
2 All the metrics are to be called from here
3 """
4 import torch
5 from sklearn.metrics import balanced_accuracy_score
6 import numpy as np
7
8
9 def classification_accuracy(output, label, params):
10 """
11 This function computes the classification accuracy.
12
13 Args:
14 output (torch.Tensor): The output of the model.
15 label (torch.Tensor): The ground truth labels.
16 params (dict): The parameter dictionary containing training and data information.
17
18 Returns:
19 torch.Tensor: The classification accuracy.
20 """
21 if params["problem_type"] == "classification":
22 predicted_classes = torch.argmax(output, 1)
23 else:
24 predicted_classes = output
25
26 acc = torch.sum(predicted_classes == label.squeeze()) / len(label)
27 return acc
28
29
30 def balanced_acc_score(output, label, params):
31 """
32 This function computes the balanced accuracy.
33
34 Args:
35 output (torch.Tensor): The output of the model.
36 label (torch.Tensor): The ground truth labels.
37 params (dict): The parameter dictionary containing training and data information.
38
39 Returns:
40 torch.Tensor: The balanced accuracy.
41 """
42 if params["problem_type"] == "classification":
43 predicted_classes = torch.argmax(output, 1)
44 else:
45 predicted_classes = output
46
47 return torch.from_numpy(
48 np.array(balanced_accuracy_score(predicted_classes.cpu(), label.cpu()))
49 )
50
51
52 def per_label_accuracy(output, label, params):
53 """
54 This function computes the per class accuracy.
55
56 Args:
57 output (torch.Tensor): The output of the model.
58 label (torch.Tensor): The ground truth labels.
59 params (dict): The parameter dictionary containing training and data information.
60
61 Returns:
62 torch.Tensor: The per class accuracy.
63 """
64 if params["problem_type"] == "classification":
65 predicted_classes = np.array([0] * len(params["model"]["class_list"]))
66 label_cpu = np.array([0] * len(params["model"]["class_list"]))
67 predicted_classes[torch.argmax(output, 1).cpu().item()] = 1
68 label_cpu[label.cpu().item()] = 1
69 return torch.from_numpy((predicted_classes == label_cpu).astype(float))
70 else:
71 return balanced_acc_score(output, label, params)
72
[end of GANDLF/metrics/regression.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/GANDLF/metrics/regression.py b/GANDLF/metrics/regression.py
--- a/GANDLF/metrics/regression.py
+++ b/GANDLF/metrics/regression.py
@@ -62,10 +62,14 @@
torch.Tensor: The per class accuracy.
"""
if params["problem_type"] == "classification":
- predicted_classes = np.array([0] * len(params["model"]["class_list"]))
- label_cpu = np.array([0] * len(params["model"]["class_list"]))
- predicted_classes[torch.argmax(output, 1).cpu().item()] = 1
- label_cpu[label.cpu().item()] = 1
- return torch.from_numpy((predicted_classes == label_cpu).astype(float))
+ # ensure this works for multiple batches
+ output_accuracy = torch.zeros(len(params["model"]["class_list"]))
+ for output_batch, label_batch in zip(output, label):
+ predicted_classes = torch.Tensor([0] * len(params["model"]["class_list"]))
+ label_cpu = torch.Tensor([0] * len(params["model"]["class_list"]))
+ predicted_classes[torch.argmax(output_batch, 0).cpu().item()] = 1
+ label_cpu[label_batch.cpu().item()] = 1
+ output_accuracy += (predicted_classes == label_cpu).type(torch.float)
+ return output_accuracy / len(output)
else:
return balanced_acc_score(output, label, params)
|
{"golden_diff": "diff --git a/GANDLF/metrics/regression.py b/GANDLF/metrics/regression.py\n--- a/GANDLF/metrics/regression.py\n+++ b/GANDLF/metrics/regression.py\n@@ -62,10 +62,14 @@\n torch.Tensor: The per class accuracy.\n \"\"\"\n if params[\"problem_type\"] == \"classification\":\n- predicted_classes = np.array([0] * len(params[\"model\"][\"class_list\"]))\n- label_cpu = np.array([0] * len(params[\"model\"][\"class_list\"]))\n- predicted_classes[torch.argmax(output, 1).cpu().item()] = 1\n- label_cpu[label.cpu().item()] = 1\n- return torch.from_numpy((predicted_classes == label_cpu).astype(float))\n+ # ensure this works for multiple batches\n+ output_accuracy = torch.zeros(len(params[\"model\"][\"class_list\"]))\n+ for output_batch, label_batch in zip(output, label):\n+ predicted_classes = torch.Tensor([0] * len(params[\"model\"][\"class_list\"]))\n+ label_cpu = torch.Tensor([0] * len(params[\"model\"][\"class_list\"]))\n+ predicted_classes[torch.argmax(output_batch, 0).cpu().item()] = 1\n+ label_cpu[label_batch.cpu().item()] = 1\n+ output_accuracy += (predicted_classes == label_cpu).type(torch.float)\n+ return output_accuracy / len(output)\n else:\n return balanced_acc_score(output, label, params)\n", "issue": "Per-label accuracy does not work for multiple batches\n**Describe the bug**\r\nWhen `batch_size > 1`, `per_label_accuracy` computation fails.\r\n\r\n**To Reproduce**\r\nSteps to reproduce the behavior:\r\n1. Set `batch_size = 4` in any classification unit test\r\n2. See error\r\n\r\n**Expected behavior**\r\nThe function should compute multiple batches of accuracies.\r\n\r\n**Screenshots**\r\nN.A.\r\n\r\n**GaNDLF Version**\r\n<!-- Put the output of the following command:\r\npython -c 'import GANDLF as g;print(g.__version__)'\r\n-->\r\n0.0.15-dev\r\n\r\n**Desktop (please complete the following information):**\r\nN.A.\r\n\r\n**Additional context**\r\nReported by @brandon-edwards\n", "before_files": [{"content": "\"\"\"\nAll the metrics are to be called from here\n\"\"\"\nimport torch\nfrom sklearn.metrics import balanced_accuracy_score\nimport numpy as np\n\n\ndef classification_accuracy(output, label, params):\n \"\"\"\n This function computes the classification accuracy.\n\n Args:\n output (torch.Tensor): The output of the model.\n label (torch.Tensor): The ground truth labels.\n params (dict): The parameter dictionary containing training and data information.\n\n Returns:\n torch.Tensor: The classification accuracy.\n \"\"\"\n if params[\"problem_type\"] == \"classification\":\n predicted_classes = torch.argmax(output, 1)\n else:\n predicted_classes = output\n\n acc = torch.sum(predicted_classes == label.squeeze()) / len(label)\n return acc\n\n\ndef balanced_acc_score(output, label, params):\n \"\"\"\n This function computes the balanced accuracy.\n\n Args:\n output (torch.Tensor): The output of the model.\n label (torch.Tensor): The ground truth labels.\n params (dict): The parameter dictionary containing training and data information.\n\n Returns:\n torch.Tensor: The balanced accuracy.\n \"\"\"\n if params[\"problem_type\"] == \"classification\":\n predicted_classes = torch.argmax(output, 1)\n else:\n predicted_classes = output\n\n return torch.from_numpy(\n np.array(balanced_accuracy_score(predicted_classes.cpu(), label.cpu()))\n )\n\n\ndef per_label_accuracy(output, label, params):\n \"\"\"\n This function computes the per class accuracy.\n\n Args:\n output (torch.Tensor): The output of the model.\n label (torch.Tensor): The ground truth labels.\n params (dict): The parameter dictionary containing training and data information.\n\n Returns:\n torch.Tensor: The per class accuracy.\n \"\"\"\n if params[\"problem_type\"] == \"classification\":\n predicted_classes = np.array([0] * len(params[\"model\"][\"class_list\"]))\n label_cpu = np.array([0] * len(params[\"model\"][\"class_list\"]))\n predicted_classes[torch.argmax(output, 1).cpu().item()] = 1\n label_cpu[label.cpu().item()] = 1\n return torch.from_numpy((predicted_classes == label_cpu).astype(float))\n else:\n return balanced_acc_score(output, label, params)\n", "path": "GANDLF/metrics/regression.py"}]}
| 1,302 | 318 |
gh_patches_debug_12330
|
rasdani/github-patches
|
git_diff
|
falconry__falcon-1883
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
deprecated() utility raises AttributeError under Meinheld
The inner function of our [`deprecated()`](https://falcon.readthedocs.io/en/latest/api/util.html#falcon.deprecated) utility generator grabs the current stack frame object via [`inspect.getcurrentframe()`](https://docs.python.org/3/library/inspect.html#inspect.currentframe), and then uses its attributes to provide a more informative deprecation warning.
However, as warned in the latter's docs, this function is not guaranteed to return a valid stack frame object on all Python implementations; it may also return `None`. It seems that running Gunicorn+Meinheld workers can trigger this situation even under CPython.
Discovered using the following command line under CPython 3.7 and 3.8:
```
gunicorn --workers=8 --worker-class="egg:meinheld#gunicorn_worker" test:app
```
For instance, assigning a value to the deprecated [`Response.body`](https://falcon.readthedocs.io/en/latest/api/request_and_response_wsgi.html#falcon.Response.body) yields
```
2021-03-11 23:31:42 [FALCON] [ERROR] GET /things => Traceback (most recent call last):
File "falcon/app.py", line 361, in falcon.app.App.__call__
File "/tmp/benchmark/test3.py", line 13, in on_get
resp.body = ('\nTwo things awe me most, the starry sky '
File "falcon/util/deprecation.py", line 67, in falcon.util.deprecation.deprecated.decorator.wrapper
AttributeError: 'NoneType' object has no attribute 'f_code'
```
</issue>
<code>
[start of falcon/util/deprecation.py]
1 # Copyright 2013 by Rackspace Hosting, Inc.
2 #
3 # Licensed under the Apache License, Version 2.0 (the "License");
4 # you may not use this file except in compliance with the License.
5 # You may obtain a copy of the License at
6 #
7 # http://www.apache.org/licenses/LICENSE-2.0
8 #
9 # Unless required by applicable law or agreed to in writing, software
10 # distributed under the License is distributed on an "AS IS" BASIS,
11 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12 # See the License for the specific language governing permissions and
13 # limitations under the License.
14
15 """Miscellaneous deprecation utilities.
16
17 This module provides decorators to mark functions and classes as deprecated.
18 """
19
20 import functools
21 import inspect
22 import warnings
23
24
25 __all__ = (
26 'DeprecatedWarning',
27 'deprecated',
28 'deprecated_args',
29 )
30
31
32 # NOTE(kgriffs): We don't want our deprecations to be ignored by default,
33 # so create our own type.
34 #
35 # TODO(kgriffs): Revisit this decision if users complain.
36 class DeprecatedWarning(UserWarning):
37 pass
38
39
40 def deprecated(instructions, is_property=False):
41 """Flag a method as deprecated.
42
43 This function returns a decorator which can be used to mark deprecated
44 functions. Applying this decorator will result in a warning being
45 emitted when the function is used.
46
47 Args:
48 instructions (str): Specific guidance for the developer, e.g.:
49 'Please migrate to add_proxy(...)'
50 is_property (bool): If the deprecated object is a property. It
51 will omit the ``(...)`` from the generated documentation
52 """
53
54 def decorator(func):
55
56 object_name = 'property' if is_property else 'function'
57 post_name = '' if is_property else '(...)'
58 message = 'Call to deprecated {} {}{}. {}'.format(
59 object_name, func.__name__, post_name, instructions)
60
61 @functools.wraps(func)
62 def wrapper(*args, **kwargs):
63 frame = inspect.currentframe().f_back
64
65 warnings.warn_explicit(message,
66 category=DeprecatedWarning,
67 filename=inspect.getfile(frame.f_code),
68 lineno=frame.f_lineno)
69
70 return func(*args, **kwargs)
71
72 return wrapper
73
74 return decorator
75
76
77 def deprecated_args(*, allowed_positional, is_method=True):
78 """Flag a method call with positional args as deprecated.
79
80 Keyword Args:
81 allowed_positional (int): Number of allowed positional arguments
82 is_method (bool, optional): The decorated function is a method. Will
83 add one to the number of allowed positional args to account for
84 ``self``. Defaults to True.
85 """
86
87 template = (
88 'Calls with{} positional args are deprecated.'
89 ' Please specify them as keyword arguments instead.'
90 )
91 text = ' more than {}'.format(allowed_positional) if allowed_positional else ''
92 warn_text = template.format(text)
93 if is_method:
94 allowed_positional += 1
95
96 def deprecated_args(fn):
97 @functools.wraps(fn)
98 def wraps(*args, **kwargs):
99 if len(args) > allowed_positional:
100 warnings.warn(warn_text, DeprecatedWarning, stacklevel=2)
101 return fn(*args, **kwargs)
102
103 return wraps
104
105 return deprecated_args
106
[end of falcon/util/deprecation.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/falcon/util/deprecation.py b/falcon/util/deprecation.py
--- a/falcon/util/deprecation.py
+++ b/falcon/util/deprecation.py
@@ -18,7 +18,6 @@
"""
import functools
-import inspect
import warnings
@@ -60,12 +59,7 @@
@functools.wraps(func)
def wrapper(*args, **kwargs):
- frame = inspect.currentframe().f_back
-
- warnings.warn_explicit(message,
- category=DeprecatedWarning,
- filename=inspect.getfile(frame.f_code),
- lineno=frame.f_lineno)
+ warnings.warn(message, category=DeprecatedWarning, stacklevel=2)
return func(*args, **kwargs)
|
{"golden_diff": "diff --git a/falcon/util/deprecation.py b/falcon/util/deprecation.py\n--- a/falcon/util/deprecation.py\n+++ b/falcon/util/deprecation.py\n@@ -18,7 +18,6 @@\n \"\"\"\n \n import functools\n-import inspect\n import warnings\n \n \n@@ -60,12 +59,7 @@\n \n @functools.wraps(func)\n def wrapper(*args, **kwargs):\n- frame = inspect.currentframe().f_back\n-\n- warnings.warn_explicit(message,\n- category=DeprecatedWarning,\n- filename=inspect.getfile(frame.f_code),\n- lineno=frame.f_lineno)\n+ warnings.warn(message, category=DeprecatedWarning, stacklevel=2)\n \n return func(*args, **kwargs)\n", "issue": "deprecated() utility raises AttributeError under Meinheld\nThe inner function of our [`deprecated()`](https://falcon.readthedocs.io/en/latest/api/util.html#falcon.deprecated) utility generator grabs the current stack frame object via [`inspect.getcurrentframe()`](https://docs.python.org/3/library/inspect.html#inspect.currentframe), and then uses its attributes to provide a more informative deprecation warning.\r\n\r\nHowever, as warned in the latter's docs, this function is not guaranteed to return a valid stack frame object on all Python implementations; it may also return `None`. It seems that running Gunicorn+Meinheld workers can trigger this situation even under CPython.\r\n\r\nDiscovered using the following command line under CPython 3.7 and 3.8:\r\n```\r\ngunicorn --workers=8 --worker-class=\"egg:meinheld#gunicorn_worker\" test:app\r\n```\r\n\r\nFor instance, assigning a value to the deprecated [`Response.body`](https://falcon.readthedocs.io/en/latest/api/request_and_response_wsgi.html#falcon.Response.body) yields\r\n```\r\n2021-03-11 23:31:42 [FALCON] [ERROR] GET /things => Traceback (most recent call last):\r\n File \"falcon/app.py\", line 361, in falcon.app.App.__call__\r\n File \"/tmp/benchmark/test3.py\", line 13, in on_get\r\n resp.body = ('\\nTwo things awe me most, the starry sky '\r\n File \"falcon/util/deprecation.py\", line 67, in falcon.util.deprecation.deprecated.decorator.wrapper\r\nAttributeError: 'NoneType' object has no attribute 'f_code'\r\n```\n", "before_files": [{"content": "# Copyright 2013 by Rackspace Hosting, Inc.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\n\"\"\"Miscellaneous deprecation utilities.\n\nThis module provides decorators to mark functions and classes as deprecated.\n\"\"\"\n\nimport functools\nimport inspect\nimport warnings\n\n\n__all__ = (\n 'DeprecatedWarning',\n 'deprecated',\n 'deprecated_args',\n)\n\n\n# NOTE(kgriffs): We don't want our deprecations to be ignored by default,\n# so create our own type.\n#\n# TODO(kgriffs): Revisit this decision if users complain.\nclass DeprecatedWarning(UserWarning):\n pass\n\n\ndef deprecated(instructions, is_property=False):\n \"\"\"Flag a method as deprecated.\n\n This function returns a decorator which can be used to mark deprecated\n functions. Applying this decorator will result in a warning being\n emitted when the function is used.\n\n Args:\n instructions (str): Specific guidance for the developer, e.g.:\n 'Please migrate to add_proxy(...)'\n is_property (bool): If the deprecated object is a property. It\n will omit the ``(...)`` from the generated documentation\n \"\"\"\n\n def decorator(func):\n\n object_name = 'property' if is_property else 'function'\n post_name = '' if is_property else '(...)'\n message = 'Call to deprecated {} {}{}. {}'.format(\n object_name, func.__name__, post_name, instructions)\n\n @functools.wraps(func)\n def wrapper(*args, **kwargs):\n frame = inspect.currentframe().f_back\n\n warnings.warn_explicit(message,\n category=DeprecatedWarning,\n filename=inspect.getfile(frame.f_code),\n lineno=frame.f_lineno)\n\n return func(*args, **kwargs)\n\n return wrapper\n\n return decorator\n\n\ndef deprecated_args(*, allowed_positional, is_method=True):\n \"\"\"Flag a method call with positional args as deprecated.\n\n Keyword Args:\n allowed_positional (int): Number of allowed positional arguments\n is_method (bool, optional): The decorated function is a method. Will\n add one to the number of allowed positional args to account for\n ``self``. Defaults to True.\n \"\"\"\n\n template = (\n 'Calls with{} positional args are deprecated.'\n ' Please specify them as keyword arguments instead.'\n )\n text = ' more than {}'.format(allowed_positional) if allowed_positional else ''\n warn_text = template.format(text)\n if is_method:\n allowed_positional += 1\n\n def deprecated_args(fn):\n @functools.wraps(fn)\n def wraps(*args, **kwargs):\n if len(args) > allowed_positional:\n warnings.warn(warn_text, DeprecatedWarning, stacklevel=2)\n return fn(*args, **kwargs)\n\n return wraps\n\n return deprecated_args\n", "path": "falcon/util/deprecation.py"}]}
| 1,831 | 164 |
gh_patches_debug_44294
|
rasdani/github-patches
|
git_diff
|
readthedocs__readthedocs.org-7070
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Projects built with `htmldir` have invalid link for homepage in search results
Projects that are built with `htmldir` have invalid link in search results to homepage.
For example: https://2.python-requests.org/en/master/search/?q=%22The+Contributor%22
This is related to #5821.
We are using `processed_json.path` for the search results, but the correct path is present in `full_path`.
https://github.com/readthedocs/readthedocs.org/blob/3c04a96b3a9ea8b68ecf270670080285c73c689b/readthedocs/search/documents.py#L73-L74
</issue>
<code>
[start of readthedocs/search/api.py]
1 import itertools
2 import logging
3
4 from django.shortcuts import get_object_or_404
5 from django.utils import timezone
6 from rest_framework import generics, serializers
7 from rest_framework.exceptions import ValidationError
8 from rest_framework.pagination import PageNumberPagination
9
10 from readthedocs.api.v2.permissions import IsAuthorizedToViewVersion
11 from readthedocs.builds.models import Version
12 from readthedocs.projects.models import HTMLFile, Project
13 from readthedocs.search import tasks, utils
14 from readthedocs.search.faceted_search import PageSearch
15
16 log = logging.getLogger(__name__)
17
18
19 class SearchPagination(PageNumberPagination):
20 page_size = 50
21 page_size_query_param = 'page_size'
22 max_page_size = 100
23
24
25 class PageSearchSerializer(serializers.Serializer):
26 project = serializers.CharField()
27 version = serializers.CharField()
28 title = serializers.CharField()
29 path = serializers.CharField()
30 link = serializers.SerializerMethodField()
31 highlight = serializers.SerializerMethodField()
32 inner_hits = serializers.SerializerMethodField()
33
34 def get_link(self, obj):
35 projects_url = self.context.get('projects_url')
36 if projects_url:
37 docs_url = projects_url[obj.project]
38 return docs_url + obj.path
39
40 def get_highlight(self, obj):
41 highlight = getattr(obj.meta, 'highlight', None)
42 if highlight:
43 ret = highlight.to_dict()
44 log.debug('API Search highlight [Page title]: %s', ret)
45 return ret
46
47 def get_inner_hits(self, obj):
48 inner_hits = getattr(obj.meta, 'inner_hits', None)
49 if inner_hits:
50 sections = inner_hits.sections or []
51 domains = inner_hits.domains or []
52 all_results = itertools.chain(sections, domains)
53
54 sorted_results = utils._get_sorted_results(
55 results=all_results,
56 source_key='_source',
57 )
58
59 log.debug('[API] Sorted Results: %s', sorted_results)
60 return sorted_results
61
62
63 class PageSearchAPIView(generics.ListAPIView):
64
65 """
66 Main entry point to perform a search using Elasticsearch.
67
68 Required query params:
69 - q (search term)
70 - project
71 - version
72
73 .. note::
74
75 The methods `_get_project` and `_get_version`
76 are called many times, so a basic cache is implemented.
77 """
78
79 permission_classes = [IsAuthorizedToViewVersion]
80 pagination_class = SearchPagination
81 serializer_class = PageSearchSerializer
82
83 def _get_project(self):
84 cache_key = '_cached_project'
85 project = getattr(self, cache_key, None)
86
87 if not project:
88 project_slug = self.request.GET.get('project', None)
89 project = get_object_or_404(Project, slug=project_slug)
90 setattr(self, cache_key, project)
91
92 return project
93
94 def _get_version(self):
95 cache_key = '_cached_version'
96 version = getattr(self, cache_key, None)
97
98 if not version:
99 version_slug = self.request.GET.get('version', None)
100 project = self._get_project()
101 version = get_object_or_404(
102 project.versions.all(),
103 slug=version_slug,
104 )
105 setattr(self, cache_key, version)
106
107 return version
108
109 def get_queryset(self):
110 """
111 Return Elasticsearch DSL Search object instead of Django Queryset.
112
113 Django Queryset and elasticsearch-dsl ``Search`` object is similar pattern.
114 So for searching, its possible to return ``Search`` object instead of queryset.
115 The ``filter_backends`` and ``pagination_class`` is compatible with ``Search``
116 """
117 # Validate all the required params are there
118 self.validate_query_params()
119 query = self.request.query_params.get('q', '')
120 kwargs = {'filter_by_user': False, 'filters': {}}
121 kwargs['filters']['project'] = [p.slug for p in self.get_all_projects()]
122 kwargs['filters']['version'] = self._get_version().slug
123
124 # Check to avoid searching all projects in case these filters are empty.
125 if not kwargs['filters']['project']:
126 log.info("Unable to find a project to search")
127 return HTMLFile.objects.none()
128 if not kwargs['filters']['version']:
129 log.info("Unable to find a version to search")
130 return HTMLFile.objects.none()
131
132 user = self.request.user
133 queryset = PageSearch(
134 query=query, user=user, **kwargs
135 )
136 return queryset
137
138 def validate_query_params(self):
139 """
140 Validate all required query params are passed on the request.
141
142 Query params required are: ``q``, ``project`` and ``version``.
143
144 :rtype: None
145
146 :raises: ValidationError if one of them is missing.
147 """
148 required_query_params = {'q', 'project', 'version'} # python `set` literal is `{}`
149 request_params = set(self.request.query_params.keys())
150 missing_params = required_query_params - request_params
151 if missing_params:
152 errors = {}
153 for param in missing_params:
154 errors[param] = ["This query param is required"]
155
156 raise ValidationError(errors)
157
158 def get_serializer_context(self):
159 context = super().get_serializer_context()
160 context['projects_url'] = self.get_all_projects_url()
161 return context
162
163 def get_all_projects(self):
164 """
165 Return a list of the project itself and all its subprojects the user has permissions over.
166
167 :rtype: list
168 """
169 main_version = self._get_version()
170 main_project = self._get_project()
171
172 all_projects = [main_project]
173
174 subprojects = Project.objects.filter(
175 superprojects__parent_id=main_project.id,
176 )
177 for project in subprojects:
178 version = (
179 Version.internal
180 .public(user=self.request.user, project=project, include_hidden=False)
181 .filter(slug=main_version.slug)
182 .first()
183 )
184 if version:
185 all_projects.append(version.project)
186 return all_projects
187
188 def get_all_projects_url(self):
189 """
190 Return a dict containing the project slug and its version URL.
191
192 The dictionary contains the project and its subprojects . Each project's
193 slug is used as a key and the documentation URL for that project and
194 version as the value.
195
196 Example:
197
198 {
199 "requests": "https://requests.readthedocs.io/en/latest/",
200 "requests-oauth": "https://requests-oauth.readthedocs.io/en/latest/",
201 }
202
203 :rtype: dict
204 """
205 all_projects = self.get_all_projects()
206 version_slug = self._get_version().slug
207 projects_url = {}
208 for project in all_projects:
209 projects_url[project.slug] = project.get_docs_url(version_slug=version_slug)
210 return projects_url
211
212 def list(self, request, *args, **kwargs):
213 """Overriding ``list`` method to record query in database."""
214
215 response = super().list(request, *args, **kwargs)
216
217 project_slug = self._get_project().slug
218 version_slug = self._get_version().slug
219 total_results = response.data.get('count', 0)
220 time = timezone.now()
221
222 query = self.request.query_params.get('q', '')
223 query = query.lower().strip()
224
225 # record the search query with a celery task
226 tasks.record_search_query.delay(
227 project_slug,
228 version_slug,
229 query,
230 total_results,
231 time.isoformat(),
232 )
233
234 return response
235
[end of readthedocs/search/api.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/readthedocs/search/api.py b/readthedocs/search/api.py
--- a/readthedocs/search/api.py
+++ b/readthedocs/search/api.py
@@ -1,5 +1,6 @@
import itertools
import logging
+import re
from django.shortcuts import get_object_or_404
from django.utils import timezone
@@ -9,6 +10,7 @@
from readthedocs.api.v2.permissions import IsAuthorizedToViewVersion
from readthedocs.builds.models import Version
+from readthedocs.projects.constants import MKDOCS, SPHINX_HTMLDIR
from readthedocs.projects.models import HTMLFile, Project
from readthedocs.search import tasks, utils
from readthedocs.search.faceted_search import PageSearch
@@ -27,15 +29,28 @@
version = serializers.CharField()
title = serializers.CharField()
path = serializers.CharField()
+ full_path = serializers.CharField()
link = serializers.SerializerMethodField()
highlight = serializers.SerializerMethodField()
inner_hits = serializers.SerializerMethodField()
def get_link(self, obj):
- projects_url = self.context.get('projects_url')
- if projects_url:
- docs_url = projects_url[obj.project]
- return docs_url + obj.path
+ project_data = self.context['projects_data'].get(obj.project)
+ if not project_data:
+ return None
+
+ docs_url, doctype = project_data
+ path = obj.full_path
+
+ # Generate an appropriate link for the doctypes that use htmldir,
+ # and always end it with / so it goes directly to proxito.
+ if doctype in {SPHINX_HTMLDIR, MKDOCS}:
+ new_path = re.sub('(^|/)index.html$', '/', path)
+ # docs_url already ends with /,
+ # so path doesn't need to start with /.
+ path = new_path.lstrip('/')
+
+ return docs_url + path
def get_highlight(self, obj):
highlight = getattr(obj.meta, 'highlight', None)
@@ -157,7 +172,7 @@
def get_serializer_context(self):
context = super().get_serializer_context()
- context['projects_url'] = self.get_all_projects_url()
+ context['projects_data'] = self.get_all_projects_data()
return context
def get_all_projects(self):
@@ -185,29 +200,44 @@
all_projects.append(version.project)
return all_projects
- def get_all_projects_url(self):
+ def get_all_projects_data(self):
"""
- Return a dict containing the project slug and its version URL.
-
- The dictionary contains the project and its subprojects . Each project's
- slug is used as a key and the documentation URL for that project and
- version as the value.
+ Return a dict containing the project slug and its version URL and version's doctype.
- Example:
+ The dictionary contains the project and its subprojects. Each project's
+ slug is used as a key and a tuple with the documentation URL and doctype
+ from the version. Example:
{
- "requests": "https://requests.readthedocs.io/en/latest/",
- "requests-oauth": "https://requests-oauth.readthedocs.io/en/latest/",
+ "requests": (
+ "https://requests.readthedocs.io/en/latest/",
+ "sphinx",
+ ),
+ "requests-oauth": (
+ "https://requests-oauth.readthedocs.io/en/latest/",
+ "sphinx_htmldir",
+ ),
}
:rtype: dict
"""
all_projects = self.get_all_projects()
version_slug = self._get_version().slug
- projects_url = {}
+ project_urls = {}
for project in all_projects:
- projects_url[project.slug] = project.get_docs_url(version_slug=version_slug)
- return projects_url
+ project_urls[project.slug] = project.get_docs_url(version_slug=version_slug)
+
+ versions_doctype = (
+ Version.objects
+ .filter(project__slug__in=project_urls.keys(), slug=version_slug)
+ .values_list('project__slug', 'documentation_type')
+ )
+
+ projects_data = {
+ project_slug: (project_urls[project_slug], doctype)
+ for project_slug, doctype in versions_doctype
+ }
+ return projects_data
def list(self, request, *args, **kwargs):
"""Overriding ``list`` method to record query in database."""
|
{"golden_diff": "diff --git a/readthedocs/search/api.py b/readthedocs/search/api.py\n--- a/readthedocs/search/api.py\n+++ b/readthedocs/search/api.py\n@@ -1,5 +1,6 @@\n import itertools\n import logging\n+import re\n \n from django.shortcuts import get_object_or_404\n from django.utils import timezone\n@@ -9,6 +10,7 @@\n \n from readthedocs.api.v2.permissions import IsAuthorizedToViewVersion\n from readthedocs.builds.models import Version\n+from readthedocs.projects.constants import MKDOCS, SPHINX_HTMLDIR\n from readthedocs.projects.models import HTMLFile, Project\n from readthedocs.search import tasks, utils\n from readthedocs.search.faceted_search import PageSearch\n@@ -27,15 +29,28 @@\n version = serializers.CharField()\n title = serializers.CharField()\n path = serializers.CharField()\n+ full_path = serializers.CharField()\n link = serializers.SerializerMethodField()\n highlight = serializers.SerializerMethodField()\n inner_hits = serializers.SerializerMethodField()\n \n def get_link(self, obj):\n- projects_url = self.context.get('projects_url')\n- if projects_url:\n- docs_url = projects_url[obj.project]\n- return docs_url + obj.path\n+ project_data = self.context['projects_data'].get(obj.project)\n+ if not project_data:\n+ return None\n+\n+ docs_url, doctype = project_data\n+ path = obj.full_path\n+\n+ # Generate an appropriate link for the doctypes that use htmldir,\n+ # and always end it with / so it goes directly to proxito.\n+ if doctype in {SPHINX_HTMLDIR, MKDOCS}:\n+ new_path = re.sub('(^|/)index.html$', '/', path)\n+ # docs_url already ends with /,\n+ # so path doesn't need to start with /.\n+ path = new_path.lstrip('/')\n+\n+ return docs_url + path\n \n def get_highlight(self, obj):\n highlight = getattr(obj.meta, 'highlight', None)\n@@ -157,7 +172,7 @@\n \n def get_serializer_context(self):\n context = super().get_serializer_context()\n- context['projects_url'] = self.get_all_projects_url()\n+ context['projects_data'] = self.get_all_projects_data()\n return context\n \n def get_all_projects(self):\n@@ -185,29 +200,44 @@\n all_projects.append(version.project)\n return all_projects\n \n- def get_all_projects_url(self):\n+ def get_all_projects_data(self):\n \"\"\"\n- Return a dict containing the project slug and its version URL.\n-\n- The dictionary contains the project and its subprojects . Each project's\n- slug is used as a key and the documentation URL for that project and\n- version as the value.\n+ Return a dict containing the project slug and its version URL and version's doctype.\n \n- Example:\n+ The dictionary contains the project and its subprojects. Each project's\n+ slug is used as a key and a tuple with the documentation URL and doctype\n+ from the version. Example:\n \n {\n- \"requests\": \"https://requests.readthedocs.io/en/latest/\",\n- \"requests-oauth\": \"https://requests-oauth.readthedocs.io/en/latest/\",\n+ \"requests\": (\n+ \"https://requests.readthedocs.io/en/latest/\",\n+ \"sphinx\",\n+ ),\n+ \"requests-oauth\": (\n+ \"https://requests-oauth.readthedocs.io/en/latest/\",\n+ \"sphinx_htmldir\",\n+ ),\n }\n \n :rtype: dict\n \"\"\"\n all_projects = self.get_all_projects()\n version_slug = self._get_version().slug\n- projects_url = {}\n+ project_urls = {}\n for project in all_projects:\n- projects_url[project.slug] = project.get_docs_url(version_slug=version_slug)\n- return projects_url\n+ project_urls[project.slug] = project.get_docs_url(version_slug=version_slug)\n+\n+ versions_doctype = (\n+ Version.objects\n+ .filter(project__slug__in=project_urls.keys(), slug=version_slug)\n+ .values_list('project__slug', 'documentation_type')\n+ )\n+\n+ projects_data = {\n+ project_slug: (project_urls[project_slug], doctype)\n+ for project_slug, doctype in versions_doctype\n+ }\n+ return projects_data\n \n def list(self, request, *args, **kwargs):\n \"\"\"Overriding ``list`` method to record query in database.\"\"\"\n", "issue": "Projects built with `htmldir` have invalid link for homepage in search results\nProjects that are built with `htmldir` have invalid link in search results to homepage.\r\nFor example: https://2.python-requests.org/en/master/search/?q=%22The+Contributor%22\r\n\r\nThis is related to #5821.\r\n\r\nWe are using `processed_json.path` for the search results, but the correct path is present in `full_path`.\r\n\r\nhttps://github.com/readthedocs/readthedocs.org/blob/3c04a96b3a9ea8b68ecf270670080285c73c689b/readthedocs/search/documents.py#L73-L74\n", "before_files": [{"content": "import itertools\nimport logging\n\nfrom django.shortcuts import get_object_or_404\nfrom django.utils import timezone\nfrom rest_framework import generics, serializers\nfrom rest_framework.exceptions import ValidationError\nfrom rest_framework.pagination import PageNumberPagination\n\nfrom readthedocs.api.v2.permissions import IsAuthorizedToViewVersion\nfrom readthedocs.builds.models import Version\nfrom readthedocs.projects.models import HTMLFile, Project\nfrom readthedocs.search import tasks, utils\nfrom readthedocs.search.faceted_search import PageSearch\n\nlog = logging.getLogger(__name__)\n\n\nclass SearchPagination(PageNumberPagination):\n page_size = 50\n page_size_query_param = 'page_size'\n max_page_size = 100\n\n\nclass PageSearchSerializer(serializers.Serializer):\n project = serializers.CharField()\n version = serializers.CharField()\n title = serializers.CharField()\n path = serializers.CharField()\n link = serializers.SerializerMethodField()\n highlight = serializers.SerializerMethodField()\n inner_hits = serializers.SerializerMethodField()\n\n def get_link(self, obj):\n projects_url = self.context.get('projects_url')\n if projects_url:\n docs_url = projects_url[obj.project]\n return docs_url + obj.path\n\n def get_highlight(self, obj):\n highlight = getattr(obj.meta, 'highlight', None)\n if highlight:\n ret = highlight.to_dict()\n log.debug('API Search highlight [Page title]: %s', ret)\n return ret\n\n def get_inner_hits(self, obj):\n inner_hits = getattr(obj.meta, 'inner_hits', None)\n if inner_hits:\n sections = inner_hits.sections or []\n domains = inner_hits.domains or []\n all_results = itertools.chain(sections, domains)\n\n sorted_results = utils._get_sorted_results(\n results=all_results,\n source_key='_source',\n )\n\n log.debug('[API] Sorted Results: %s', sorted_results)\n return sorted_results\n\n\nclass PageSearchAPIView(generics.ListAPIView):\n\n \"\"\"\n Main entry point to perform a search using Elasticsearch.\n\n Required query params:\n - q (search term)\n - project\n - version\n\n .. note::\n\n The methods `_get_project` and `_get_version`\n are called many times, so a basic cache is implemented.\n \"\"\"\n\n permission_classes = [IsAuthorizedToViewVersion]\n pagination_class = SearchPagination\n serializer_class = PageSearchSerializer\n\n def _get_project(self):\n cache_key = '_cached_project'\n project = getattr(self, cache_key, None)\n\n if not project:\n project_slug = self.request.GET.get('project', None)\n project = get_object_or_404(Project, slug=project_slug)\n setattr(self, cache_key, project)\n\n return project\n\n def _get_version(self):\n cache_key = '_cached_version'\n version = getattr(self, cache_key, None)\n\n if not version:\n version_slug = self.request.GET.get('version', None)\n project = self._get_project()\n version = get_object_or_404(\n project.versions.all(),\n slug=version_slug,\n )\n setattr(self, cache_key, version)\n\n return version\n\n def get_queryset(self):\n \"\"\"\n Return Elasticsearch DSL Search object instead of Django Queryset.\n\n Django Queryset and elasticsearch-dsl ``Search`` object is similar pattern.\n So for searching, its possible to return ``Search`` object instead of queryset.\n The ``filter_backends`` and ``pagination_class`` is compatible with ``Search``\n \"\"\"\n # Validate all the required params are there\n self.validate_query_params()\n query = self.request.query_params.get('q', '')\n kwargs = {'filter_by_user': False, 'filters': {}}\n kwargs['filters']['project'] = [p.slug for p in self.get_all_projects()]\n kwargs['filters']['version'] = self._get_version().slug\n\n # Check to avoid searching all projects in case these filters are empty.\n if not kwargs['filters']['project']:\n log.info(\"Unable to find a project to search\")\n return HTMLFile.objects.none()\n if not kwargs['filters']['version']:\n log.info(\"Unable to find a version to search\")\n return HTMLFile.objects.none()\n\n user = self.request.user\n queryset = PageSearch(\n query=query, user=user, **kwargs\n )\n return queryset\n\n def validate_query_params(self):\n \"\"\"\n Validate all required query params are passed on the request.\n\n Query params required are: ``q``, ``project`` and ``version``.\n\n :rtype: None\n\n :raises: ValidationError if one of them is missing.\n \"\"\"\n required_query_params = {'q', 'project', 'version'} # python `set` literal is `{}`\n request_params = set(self.request.query_params.keys())\n missing_params = required_query_params - request_params\n if missing_params:\n errors = {}\n for param in missing_params:\n errors[param] = [\"This query param is required\"]\n\n raise ValidationError(errors)\n\n def get_serializer_context(self):\n context = super().get_serializer_context()\n context['projects_url'] = self.get_all_projects_url()\n return context\n\n def get_all_projects(self):\n \"\"\"\n Return a list of the project itself and all its subprojects the user has permissions over.\n\n :rtype: list\n \"\"\"\n main_version = self._get_version()\n main_project = self._get_project()\n\n all_projects = [main_project]\n\n subprojects = Project.objects.filter(\n superprojects__parent_id=main_project.id,\n )\n for project in subprojects:\n version = (\n Version.internal\n .public(user=self.request.user, project=project, include_hidden=False)\n .filter(slug=main_version.slug)\n .first()\n )\n if version:\n all_projects.append(version.project)\n return all_projects\n\n def get_all_projects_url(self):\n \"\"\"\n Return a dict containing the project slug and its version URL.\n\n The dictionary contains the project and its subprojects . Each project's\n slug is used as a key and the documentation URL for that project and\n version as the value.\n\n Example:\n\n {\n \"requests\": \"https://requests.readthedocs.io/en/latest/\",\n \"requests-oauth\": \"https://requests-oauth.readthedocs.io/en/latest/\",\n }\n\n :rtype: dict\n \"\"\"\n all_projects = self.get_all_projects()\n version_slug = self._get_version().slug\n projects_url = {}\n for project in all_projects:\n projects_url[project.slug] = project.get_docs_url(version_slug=version_slug)\n return projects_url\n\n def list(self, request, *args, **kwargs):\n \"\"\"Overriding ``list`` method to record query in database.\"\"\"\n\n response = super().list(request, *args, **kwargs)\n\n project_slug = self._get_project().slug\n version_slug = self._get_version().slug\n total_results = response.data.get('count', 0)\n time = timezone.now()\n\n query = self.request.query_params.get('q', '')\n query = query.lower().strip()\n\n # record the search query with a celery task\n tasks.record_search_query.delay(\n project_slug,\n version_slug,\n query,\n total_results,\n time.isoformat(),\n )\n\n return response\n", "path": "readthedocs/search/api.py"}]}
| 2,883 | 1,013 |
gh_patches_debug_1321
|
rasdani/github-patches
|
git_diff
|
pyodide__pyodide-717
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Calling yaml.load() without Loader=... is deprecated
For each built packages there is now the following deprecation warning ,
```
pyodide_build/common.py:27: YAMLLoadWarning: calling yaml.load() without Loader=... is deprecated, as the default Loader is unsafe. Please read https://msg.pyyaml.org/load for full details.
return yaml.load(fd)
```
it would be nice to fix this.
</issue>
<code>
[start of pyodide_build/common.py]
1 from pathlib import Path
2 from typing import Optional, Set
3
4
5 ROOTDIR = Path(__file__).parents[1].resolve() / "tools"
6 HOSTPYTHON = ROOTDIR / ".." / "cpython" / "build" / "3.8.2" / "host"
7 TARGETPYTHON = ROOTDIR / ".." / "cpython" / "installs" / "python-3.8.2"
8 DEFAULTCFLAGS = ""
9 DEFAULTLDFLAGS = " ".join(
10 [
11 "-O3",
12 "-s",
13 "BINARYEN_METHOD='native-wasm'",
14 "-Werror",
15 "-s",
16 "EMULATED_FUNCTION_POINTERS=1",
17 "-s",
18 "EMULATE_FUNCTION_POINTER_CASTS=1",
19 "-s",
20 "SIDE_MODULE=1",
21 "-s",
22 "WASM=1",
23 "--memory-init-file",
24 "0",
25 ]
26 )
27
28
29 def parse_package(package):
30 # Import yaml here because pywasmcross needs to run in the built native
31 # Python, which won't have PyYAML
32 import yaml
33
34 # TODO: Validate against a schema
35 with open(package) as fd:
36 return yaml.load(fd)
37
38
39 def _parse_package_subset(query: Optional[str]) -> Optional[Set[str]]:
40 """Parse the list of packages specified with PYODIDE_PACKAGES env var.
41
42 Also add the list of mandatory packages: ['micropip', 'distlib']
43
44 Returns:
45 a set of package names to build or None.
46 """
47 if query is None:
48 return None
49 packages = query.split(",")
50 packages = [el.strip() for el in packages]
51 packages = ["micropip", "distlib"] + packages
52 return set(packages)
53
[end of pyodide_build/common.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/pyodide_build/common.py b/pyodide_build/common.py
--- a/pyodide_build/common.py
+++ b/pyodide_build/common.py
@@ -33,7 +33,7 @@
# TODO: Validate against a schema
with open(package) as fd:
- return yaml.load(fd)
+ return yaml.safe_load(fd)
def _parse_package_subset(query: Optional[str]) -> Optional[Set[str]]:
|
{"golden_diff": "diff --git a/pyodide_build/common.py b/pyodide_build/common.py\n--- a/pyodide_build/common.py\n+++ b/pyodide_build/common.py\n@@ -33,7 +33,7 @@\n \n # TODO: Validate against a schema\n with open(package) as fd:\n- return yaml.load(fd)\n+ return yaml.safe_load(fd)\n \n \n def _parse_package_subset(query: Optional[str]) -> Optional[Set[str]]:\n", "issue": "Calling yaml.load() without Loader=... is deprecated\nFor each built packages there is now the following deprecation warning ,\r\n```\r\npyodide_build/common.py:27: YAMLLoadWarning: calling yaml.load() without Loader=... is deprecated, as the default Loader is unsafe. Please read https://msg.pyyaml.org/load for full details.\r\n return yaml.load(fd)\r\n```\r\nit would be nice to fix this.\n", "before_files": [{"content": "from pathlib import Path\nfrom typing import Optional, Set\n\n\nROOTDIR = Path(__file__).parents[1].resolve() / \"tools\"\nHOSTPYTHON = ROOTDIR / \"..\" / \"cpython\" / \"build\" / \"3.8.2\" / \"host\"\nTARGETPYTHON = ROOTDIR / \"..\" / \"cpython\" / \"installs\" / \"python-3.8.2\"\nDEFAULTCFLAGS = \"\"\nDEFAULTLDFLAGS = \" \".join(\n [\n \"-O3\",\n \"-s\",\n \"BINARYEN_METHOD='native-wasm'\",\n \"-Werror\",\n \"-s\",\n \"EMULATED_FUNCTION_POINTERS=1\",\n \"-s\",\n \"EMULATE_FUNCTION_POINTER_CASTS=1\",\n \"-s\",\n \"SIDE_MODULE=1\",\n \"-s\",\n \"WASM=1\",\n \"--memory-init-file\",\n \"0\",\n ]\n)\n\n\ndef parse_package(package):\n # Import yaml here because pywasmcross needs to run in the built native\n # Python, which won't have PyYAML\n import yaml\n\n # TODO: Validate against a schema\n with open(package) as fd:\n return yaml.load(fd)\n\n\ndef _parse_package_subset(query: Optional[str]) -> Optional[Set[str]]:\n \"\"\"Parse the list of packages specified with PYODIDE_PACKAGES env var.\n\n Also add the list of mandatory packages: ['micropip', 'distlib']\n\n Returns:\n a set of package names to build or None.\n \"\"\"\n if query is None:\n return None\n packages = query.split(\",\")\n packages = [el.strip() for el in packages]\n packages = [\"micropip\", \"distlib\"] + packages\n return set(packages)\n", "path": "pyodide_build/common.py"}]}
| 1,104 | 98 |
gh_patches_debug_20517
|
rasdani/github-patches
|
git_diff
|
quantumlib__Cirq-1863
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Split cirq packages into with/without contrib
Otherwise there's no way to easily pip install the contrib-requirements
</issue>
<code>
[start of setup.py]
1 # Copyright 2018 The Cirq Developers
2 #
3 # Licensed under the Apache License, Version 2.0 (the "License");
4 # you may not use this file except in compliance with the License.
5 # You may obtain a copy of the License at
6 #
7 # https://www.apache.org/licenses/LICENSE-2.0
8 #
9 # Unless required by applicable law or agreed to in writing, software
10 # distributed under the License is distributed on an "AS IS" BASIS,
11 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12 # See the License for the specific language governing permissions and
13 # limitations under the License.
14
15 import io
16 from setuptools import find_packages, setup
17
18 # This reads the __version__ variable from cirq/_version.py
19 __version__ = ''
20 exec(open('cirq/_version.py').read())
21
22 description = ('A framework for creating, editing, and invoking '
23 'Noisy Intermediate Scale Quantum (NISQ) circuits.')
24
25 # README file as long_description.
26 long_description = io.open('README.rst', encoding='utf-8').read()
27
28 # Read in requirements
29 requirements = open('requirements.txt').readlines()
30 requirements = [r.strip() for r in requirements]
31
32 cirq_packages = ['cirq'] + [
33 'cirq.' + package for package in find_packages(where='cirq')
34 ]
35
36 setup(name='cirq',
37 version=__version__,
38 url='http://github.com/quantumlib/cirq',
39 author='The Cirq Developers',
40 author_email='[email protected]',
41 python_requires=('>=3.6.0'),
42 install_requires=requirements,
43 license='Apache 2',
44 description=description,
45 long_description=long_description,
46 packages=cirq_packages,
47 package_data={
48 'cirq.api.google.v1': ['*.proto'],
49 'cirq.api.google.v2': ['*.proto'],
50 })
51
[end of setup.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/setup.py b/setup.py
--- a/setup.py
+++ b/setup.py
@@ -28,6 +28,10 @@
# Read in requirements
requirements = open('requirements.txt').readlines()
requirements = [r.strip() for r in requirements]
+contrib_requirements = open('cirq/contrib/contrib-requirements.txt').readlines()
+contrib_requirements = [r.strip() for r in contrib_requirements]
+dev_requirements = open('dev_tools/conf/pip-list-dev-tools.txt').readlines()
+dev_requirements = [r.strip() for r in dev_requirements]
cirq_packages = ['cirq'] + [
'cirq.' + package for package in find_packages(where='cirq')
@@ -40,6 +44,10 @@
author_email='[email protected]',
python_requires=('>=3.6.0'),
install_requires=requirements,
+ extras_require={
+ 'contrib': contrib_requirements,
+ 'dev': dev_requirements,
+ },
license='Apache 2',
description=description,
long_description=long_description,
|
{"golden_diff": "diff --git a/setup.py b/setup.py\n--- a/setup.py\n+++ b/setup.py\n@@ -28,6 +28,10 @@\n # Read in requirements\n requirements = open('requirements.txt').readlines()\n requirements = [r.strip() for r in requirements]\n+contrib_requirements = open('cirq/contrib/contrib-requirements.txt').readlines()\n+contrib_requirements = [r.strip() for r in contrib_requirements]\n+dev_requirements = open('dev_tools/conf/pip-list-dev-tools.txt').readlines()\n+dev_requirements = [r.strip() for r in dev_requirements]\n \n cirq_packages = ['cirq'] + [\n 'cirq.' + package for package in find_packages(where='cirq')\n@@ -40,6 +44,10 @@\n author_email='[email protected]',\n python_requires=('>=3.6.0'),\n install_requires=requirements,\n+ extras_require={\n+ 'contrib': contrib_requirements,\n+ 'dev': dev_requirements,\n+ },\n license='Apache 2',\n description=description,\n long_description=long_description,\n", "issue": "Split cirq packages into with/without contrib\nOtherwise there's no way to easily pip install the contrib-requirements\n", "before_files": [{"content": "# Copyright 2018 The Cirq Developers\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# https://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nimport io\nfrom setuptools import find_packages, setup\n\n# This reads the __version__ variable from cirq/_version.py\n__version__ = ''\nexec(open('cirq/_version.py').read())\n\ndescription = ('A framework for creating, editing, and invoking '\n 'Noisy Intermediate Scale Quantum (NISQ) circuits.')\n\n# README file as long_description.\nlong_description = io.open('README.rst', encoding='utf-8').read()\n\n# Read in requirements\nrequirements = open('requirements.txt').readlines()\nrequirements = [r.strip() for r in requirements]\n\ncirq_packages = ['cirq'] + [\n 'cirq.' + package for package in find_packages(where='cirq')\n]\n\nsetup(name='cirq',\n version=__version__,\n url='http://github.com/quantumlib/cirq',\n author='The Cirq Developers',\n author_email='[email protected]',\n python_requires=('>=3.6.0'),\n install_requires=requirements,\n license='Apache 2',\n description=description,\n long_description=long_description,\n packages=cirq_packages,\n package_data={\n 'cirq.api.google.v1': ['*.proto'],\n 'cirq.api.google.v2': ['*.proto'],\n })\n", "path": "setup.py"}]}
| 1,052 | 238 |
gh_patches_debug_2369
|
rasdani/github-patches
|
git_diff
|
Pyomo__pyomo-2265
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Consistent semantic versioning
## Summary
The most recent version of Pyomo released was 6.2, as opposed to 6.2.0. It seems inconsistent with the way many other packages are versioned (e.g. NumFocus packages), although I am unaware if there is a standard specified anywhere. Is there a benefit to the former as opposed to the latter?
## Context
Managing our dependencies, we automate pulling in new versions of packages, running them through our CI prior to upgrading. We run this in two ways - one allowing all upgrades and one allowing only compatible upgrades (PEP 440). This always requires manual review because not all packages use semantic versioning (or the same semantic versioning). One manual override we had to apply this time was pinning `Pyomo ~= 6.2.0` instead of what our script automatically pulled in `Pyomo ~= 6.2`.
Consistent semantic versioning
## Summary
The most recent version of Pyomo released was 6.2, as opposed to 6.2.0. It seems inconsistent with the way many other packages are versioned (e.g. NumFocus packages), although I am unaware if there is a standard specified anywhere. Is there a benefit to the former as opposed to the latter?
## Context
Managing our dependencies, we automate pulling in new versions of packages, running them through our CI prior to upgrading. We run this in two ways - one allowing all upgrades and one allowing only compatible upgrades (PEP 440). This always requires manual review because not all packages use semantic versioning (or the same semantic versioning). One manual override we had to apply this time was pinning `Pyomo ~= 6.2.0` instead of what our script automatically pulled in `Pyomo ~= 6.2`.
</issue>
<code>
[start of pyomo/version/info.py]
1 # ___________________________________________________________________________
2 #
3 # Pyomo: Python Optimization Modeling Objects
4 # Copyright 2017 National Technology and Engineering Solutions of Sandia, LLC
5 # Under the terms of Contract DE-NA0003525 with National Technology and
6 # Engineering Solutions of Sandia, LLC, the U.S. Government retains certain
7 # rights in this software.
8 # This software is distributed under the 3-clause BSD License.
9 # ___________________________________________________________________________
10
11 _init_url="$URL$"
12
13 # NOTE: releaselevel should be left at 'invalid' for trunk development
14 # and set to 'final' for releases. During development, the
15 # major.minor.micro should point ot the NEXT release (generally, the
16 # next micro release after the current release).
17 #
18 # Note: When cutting a release, also update the major/minor/micro in
19 #
20 # pyomo/RELEASE.txt
21 #
22 # The VOTD zipbuilder will automatically change releaselevel to "VOTD
23 # {hash}" and set the serial number to YYMMDDhhmm. The serial number
24 # should generally be left at 0, unless a downstream package is tracking
25 # main and needs a hard reference to "suitably new" development.
26 major=6
27 minor=2
28 micro=1
29 releaselevel='invalid'
30 #releaselevel='final'
31 serial=0
32
33 if releaselevel == 'final':
34 pass
35 elif '/tags/' in _init_url: #pragma:nocover
36 releaselevel = 'final'
37 elif releaselevel == 'invalid':
38 from os.path import abspath, dirname, exists, join
39 if __file__.endswith('setup.py'):
40 # This file is being sources (exec'ed) from setup.py.
41 # dirname(__file__) setup.py's scope is the root sourec directory
42 _rootdir = os.path.dirname(__file__)
43 else:
44 # Eventually this should import PYOMO_ROOT_DIR from
45 # pyomo.common instead of reimplementing that logic here.
46 #
47 # __file__ fails if script is called in different ways on Windows
48 # __file__ fails if someone does os.chdir() before
49 # sys.argv[0] also fails because it doesn't not always contains the path
50 from inspect import getfile, currentframe
51 _rootdir = join(dirname(abspath(getfile(currentframe()))), '..', '..')
52
53 if exists(join(_rootdir, '.git')):
54 try:
55 with open(join(_rootdir, '.git', 'HEAD')) as _FILE:
56 _ref = _FILE.readline().strip() #pragma:nocover
57 releaselevel = 'devel {%s}' % (
58 _ref.split('/')[-1].split('\\')[-1], ) #pragma:nocover
59 except:
60 releaselevel = 'devel' #pragma:nocover
61 elif exists(join(_rootdir, '.svn')):
62 releaselevel = 'devel {svn}' #pragma:nocover
63 else:
64 releaselevel = 'VOTD' #pragma:nocover
65
66
67 version_info = (major, minor, micro, releaselevel, serial)
68
69 version = '.'.join(str(x) for x in version_info[:(3 if micro else 2)])
70 __version__ = version
71 if releaselevel != 'final':
72 version += ' ('+releaselevel+')'
73 if releaselevel.startswith('devel'):
74 __version__ += ".dev%d" % (serial,)
75 elif releaselevel.startswith('VOTD'):
76 __version__ += "a%d" % (serial,)
77
[end of pyomo/version/info.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/pyomo/version/info.py b/pyomo/version/info.py
--- a/pyomo/version/info.py
+++ b/pyomo/version/info.py
@@ -66,7 +66,7 @@
version_info = (major, minor, micro, releaselevel, serial)
-version = '.'.join(str(x) for x in version_info[:(3 if micro else 2)])
+version = '.'.join(str(x) for x in version_info[:3])
__version__ = version
if releaselevel != 'final':
version += ' ('+releaselevel+')'
|
{"golden_diff": "diff --git a/pyomo/version/info.py b/pyomo/version/info.py\n--- a/pyomo/version/info.py\n+++ b/pyomo/version/info.py\n@@ -66,7 +66,7 @@\n \n version_info = (major, minor, micro, releaselevel, serial)\n \n-version = '.'.join(str(x) for x in version_info[:(3 if micro else 2)])\n+version = '.'.join(str(x) for x in version_info[:3])\n __version__ = version\n if releaselevel != 'final':\n version += ' ('+releaselevel+')'\n", "issue": "Consistent semantic versioning\n## Summary\r\n\r\nThe most recent version of Pyomo released was 6.2, as opposed to 6.2.0. It seems inconsistent with the way many other packages are versioned (e.g. NumFocus packages), although I am unaware if there is a standard specified anywhere. Is there a benefit to the former as opposed to the latter? \r\n\r\n## Context\r\n\r\nManaging our dependencies, we automate pulling in new versions of packages, running them through our CI prior to upgrading. We run this in two ways - one allowing all upgrades and one allowing only compatible upgrades (PEP 440). This always requires manual review because not all packages use semantic versioning (or the same semantic versioning). One manual override we had to apply this time was pinning `Pyomo ~= 6.2.0` instead of what our script automatically pulled in `Pyomo ~= 6.2`.\nConsistent semantic versioning\n## Summary\r\n\r\nThe most recent version of Pyomo released was 6.2, as opposed to 6.2.0. It seems inconsistent with the way many other packages are versioned (e.g. NumFocus packages), although I am unaware if there is a standard specified anywhere. Is there a benefit to the former as opposed to the latter? \r\n\r\n## Context\r\n\r\nManaging our dependencies, we automate pulling in new versions of packages, running them through our CI prior to upgrading. We run this in two ways - one allowing all upgrades and one allowing only compatible upgrades (PEP 440). This always requires manual review because not all packages use semantic versioning (or the same semantic versioning). One manual override we had to apply this time was pinning `Pyomo ~= 6.2.0` instead of what our script automatically pulled in `Pyomo ~= 6.2`.\n", "before_files": [{"content": "# ___________________________________________________________________________\n#\n# Pyomo: Python Optimization Modeling Objects\n# Copyright 2017 National Technology and Engineering Solutions of Sandia, LLC\n# Under the terms of Contract DE-NA0003525 with National Technology and \n# Engineering Solutions of Sandia, LLC, the U.S. Government retains certain \n# rights in this software.\n# This software is distributed under the 3-clause BSD License.\n# ___________________________________________________________________________\n\n_init_url=\"$URL$\"\n\n# NOTE: releaselevel should be left at 'invalid' for trunk development\n# and set to 'final' for releases. During development, the\n# major.minor.micro should point ot the NEXT release (generally, the\n# next micro release after the current release).\n#\n# Note: When cutting a release, also update the major/minor/micro in\n#\n# pyomo/RELEASE.txt\n#\n# The VOTD zipbuilder will automatically change releaselevel to \"VOTD\n# {hash}\" and set the serial number to YYMMDDhhmm. The serial number\n# should generally be left at 0, unless a downstream package is tracking\n# main and needs a hard reference to \"suitably new\" development.\nmajor=6\nminor=2\nmicro=1\nreleaselevel='invalid'\n#releaselevel='final'\nserial=0\n\nif releaselevel == 'final':\n pass\nelif '/tags/' in _init_url: #pragma:nocover\n releaselevel = 'final'\nelif releaselevel == 'invalid':\n from os.path import abspath, dirname, exists, join\n if __file__.endswith('setup.py'):\n # This file is being sources (exec'ed) from setup.py.\n # dirname(__file__) setup.py's scope is the root sourec directory\n _rootdir = os.path.dirname(__file__)\n else:\n # Eventually this should import PYOMO_ROOT_DIR from\n # pyomo.common instead of reimplementing that logic here.\n #\n # __file__ fails if script is called in different ways on Windows\n # __file__ fails if someone does os.chdir() before\n # sys.argv[0] also fails because it doesn't not always contains the path\n from inspect import getfile, currentframe\n _rootdir = join(dirname(abspath(getfile(currentframe()))), '..', '..')\n\n if exists(join(_rootdir, '.git')):\n try:\n with open(join(_rootdir, '.git', 'HEAD')) as _FILE:\n _ref = _FILE.readline().strip() #pragma:nocover\n releaselevel = 'devel {%s}' % (\n _ref.split('/')[-1].split('\\\\')[-1], ) #pragma:nocover\n except:\n releaselevel = 'devel' #pragma:nocover\n elif exists(join(_rootdir, '.svn')):\n releaselevel = 'devel {svn}' #pragma:nocover\n else:\n releaselevel = 'VOTD' #pragma:nocover\n\n\nversion_info = (major, minor, micro, releaselevel, serial)\n\nversion = '.'.join(str(x) for x in version_info[:(3 if micro else 2)])\n__version__ = version\nif releaselevel != 'final':\n version += ' ('+releaselevel+')'\nif releaselevel.startswith('devel'):\n __version__ += \".dev%d\" % (serial,)\nelif releaselevel.startswith('VOTD'):\n __version__ += \"a%d\" % (serial,)\n", "path": "pyomo/version/info.py"}]}
| 1,833 | 123 |
gh_patches_debug_36212
|
rasdani/github-patches
|
git_diff
|
strawberry-graphql__strawberry-403
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Error adding more than one implementation of an interface
**Observed Behaviour**: When i try to add two implementations of an interface, i get a duplicated type name exception
**Expected Behaviour**: Instead of trying to recreate the interface type again, reuse it.
**Steps to reproduce**:
1. Create an interface
2. Create two types which implement the interface
3. Launch `strawberry server app`
4. See it fails with ` Schema must contain uniquely named types but contains multiple types named '<InterfaceName>'`
**Snippet to reproduce the issue**
````python
from typing import List, Optional, Union
import strawberry
from strawberry import field
@strawberry.interface
class Person:
name: str
email: str
@strawberry.type
class Speaker(Person):
job: str
@strawberry.type
class Attendee(Person):
interests: List[str]
def get_people_by_name(name: str):
return []
@strawberry.type
class Query:
searchPeopleByName: List[Union[Speaker, Attendee]] = field(resolver=get_people_by_name)
schema = strawberry.Schema(query=Query)
````
**Full traceback:**
```
File "/mnt/c/Users/<User>/code/nerdearla/test_app.py", line 30, in <module>
schema = strawberry.Schema(query=Query)
File "/home/crow/.virtualenvs/venv/lib/python3.8/site-packages/strawberry/schema/schema.py", line 42, in __init__
self._schema = GraphQLSchema(
File "/home/crow/.virtualenvs/venv/lib/python3.8/site-packages/graphql/type/schema.py", line 240, in __init__
raise TypeError(
TypeError: Schema must contain uniquely named types but contains multiple types named 'Person'.
```
</issue>
<code>
[start of strawberry/schema/types/object_type.py]
1 from typing import Type, cast
2
3 from graphql import GraphQLInputObjectType, GraphQLObjectType
4 from graphql.type.definition import GraphQLInterfaceType
5 from strawberry.type import TypeDefinition
6
7 from .fields import get_field
8 from .types import ConcreteType, GraphQLType, TypeMap
9
10
11 def _get_object_type_for_type_definition(
12 type_definition: TypeDefinition, type_map: TypeMap
13 ) -> GraphQLType:
14
15 TypeClass: Type = GraphQLObjectType
16
17 kwargs = {}
18
19 if type_definition.is_input:
20 TypeClass = GraphQLInputObjectType
21 elif type_definition.is_interface:
22 TypeClass = GraphQLInterfaceType
23
24 if type_definition.interfaces:
25 kwargs["interfaces"] = [
26 _get_object_type_for_type_definition(interface, type_map)
27 for interface in type_definition.interfaces
28 ]
29
30 assert not type_definition.is_generic
31
32 return TypeClass(
33 name=type_definition.name,
34 fields=lambda: {
35 field.name: get_field(field, type_definition.is_input, type_map)
36 for field in type_definition.fields
37 },
38 description=type_definition.description,
39 **kwargs,
40 )
41
42
43 def get_object_type(origin: Type, type_map: TypeMap) -> GraphQLObjectType:
44 """Returns a root type (Query, Mutation, Subscription) from a decorated type"""
45
46 if not hasattr(origin, "_type_definition"):
47 raise ValueError(f"Wrong type passed to get object type {origin}")
48
49 type_definition: TypeDefinition = origin._type_definition
50
51 name = type_definition.name
52
53 if name not in type_map:
54 object_type = _get_object_type_for_type_definition(type_definition, type_map)
55
56 type_map[name] = ConcreteType(
57 definition=type_definition, implementation=object_type
58 )
59
60 return cast(GraphQLObjectType, type_map[name].implementation)
61
[end of strawberry/schema/types/object_type.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/strawberry/schema/types/object_type.py b/strawberry/schema/types/object_type.py
--- a/strawberry/schema/types/object_type.py
+++ b/strawberry/schema/types/object_type.py
@@ -12,32 +12,43 @@
type_definition: TypeDefinition, type_map: TypeMap
) -> GraphQLType:
- TypeClass: Type = GraphQLObjectType
-
- kwargs = {}
-
- if type_definition.is_input:
- TypeClass = GraphQLInputObjectType
- elif type_definition.is_interface:
- TypeClass = GraphQLInterfaceType
-
- if type_definition.interfaces:
- kwargs["interfaces"] = [
- _get_object_type_for_type_definition(interface, type_map)
- for interface in type_definition.interfaces
- ]
-
- assert not type_definition.is_generic
-
- return TypeClass(
- name=type_definition.name,
- fields=lambda: {
- field.name: get_field(field, type_definition.is_input, type_map)
- for field in type_definition.fields
- },
- description=type_definition.description,
- **kwargs,
- )
+ if type_definition.name not in type_map:
+ TypeClass: Type = GraphQLObjectType
+
+ kwargs = {}
+
+ if type_definition.is_input:
+ TypeClass = GraphQLInputObjectType
+ elif type_definition.is_interface:
+ TypeClass = GraphQLInterfaceType
+
+ if type_definition.interfaces:
+ kwargs["interfaces"] = [
+ _get_object_type_for_type_definition(interface, type_map)
+ for interface in type_definition.interfaces
+ ]
+ # this tells GraphQL core what the returned object's actual type is
+ kwargs["is_type_of"] = lambda obj, _: isinstance( # type: ignore
+ obj, type_definition.origin
+ )
+
+ assert not type_definition.is_generic
+
+ object_type = TypeClass(
+ name=type_definition.name,
+ fields=lambda: {
+ field.name: get_field(field, type_definition.is_input, type_map)
+ for field in type_definition.fields
+ },
+ description=type_definition.description,
+ **kwargs,
+ )
+
+ type_map[type_definition.name] = ConcreteType(
+ definition=type_definition, implementation=object_type
+ )
+
+ return type_map[type_definition.name].implementation
def get_object_type(origin: Type, type_map: TypeMap) -> GraphQLObjectType:
@@ -48,13 +59,7 @@
type_definition: TypeDefinition = origin._type_definition
- name = type_definition.name
-
- if name not in type_map:
- object_type = _get_object_type_for_type_definition(type_definition, type_map)
-
- type_map[name] = ConcreteType(
- definition=type_definition, implementation=object_type
- )
-
- return cast(GraphQLObjectType, type_map[name].implementation)
+ return cast(
+ GraphQLObjectType,
+ _get_object_type_for_type_definition(type_definition, type_map),
+ )
|
{"golden_diff": "diff --git a/strawberry/schema/types/object_type.py b/strawberry/schema/types/object_type.py\n--- a/strawberry/schema/types/object_type.py\n+++ b/strawberry/schema/types/object_type.py\n@@ -12,32 +12,43 @@\n type_definition: TypeDefinition, type_map: TypeMap\n ) -> GraphQLType:\n \n- TypeClass: Type = GraphQLObjectType\n-\n- kwargs = {}\n-\n- if type_definition.is_input:\n- TypeClass = GraphQLInputObjectType\n- elif type_definition.is_interface:\n- TypeClass = GraphQLInterfaceType\n-\n- if type_definition.interfaces:\n- kwargs[\"interfaces\"] = [\n- _get_object_type_for_type_definition(interface, type_map)\n- for interface in type_definition.interfaces\n- ]\n-\n- assert not type_definition.is_generic\n-\n- return TypeClass(\n- name=type_definition.name,\n- fields=lambda: {\n- field.name: get_field(field, type_definition.is_input, type_map)\n- for field in type_definition.fields\n- },\n- description=type_definition.description,\n- **kwargs,\n- )\n+ if type_definition.name not in type_map:\n+ TypeClass: Type = GraphQLObjectType\n+\n+ kwargs = {}\n+\n+ if type_definition.is_input:\n+ TypeClass = GraphQLInputObjectType\n+ elif type_definition.is_interface:\n+ TypeClass = GraphQLInterfaceType\n+\n+ if type_definition.interfaces:\n+ kwargs[\"interfaces\"] = [\n+ _get_object_type_for_type_definition(interface, type_map)\n+ for interface in type_definition.interfaces\n+ ]\n+ # this tells GraphQL core what the returned object's actual type is\n+ kwargs[\"is_type_of\"] = lambda obj, _: isinstance( # type: ignore\n+ obj, type_definition.origin\n+ )\n+\n+ assert not type_definition.is_generic\n+\n+ object_type = TypeClass(\n+ name=type_definition.name,\n+ fields=lambda: {\n+ field.name: get_field(field, type_definition.is_input, type_map)\n+ for field in type_definition.fields\n+ },\n+ description=type_definition.description,\n+ **kwargs,\n+ )\n+\n+ type_map[type_definition.name] = ConcreteType(\n+ definition=type_definition, implementation=object_type\n+ )\n+\n+ return type_map[type_definition.name].implementation\n \n \n def get_object_type(origin: Type, type_map: TypeMap) -> GraphQLObjectType:\n@@ -48,13 +59,7 @@\n \n type_definition: TypeDefinition = origin._type_definition\n \n- name = type_definition.name\n-\n- if name not in type_map:\n- object_type = _get_object_type_for_type_definition(type_definition, type_map)\n-\n- type_map[name] = ConcreteType(\n- definition=type_definition, implementation=object_type\n- )\n-\n- return cast(GraphQLObjectType, type_map[name].implementation)\n+ return cast(\n+ GraphQLObjectType,\n+ _get_object_type_for_type_definition(type_definition, type_map),\n+ )\n", "issue": "Error adding more than one implementation of an interface\n**Observed Behaviour**: When i try to add two implementations of an interface, i get a duplicated type name exception\r\n\r\n**Expected Behaviour**: Instead of trying to recreate the interface type again, reuse it.\r\n\r\n**Steps to reproduce**:\r\n1. Create an interface\r\n2. Create two types which implement the interface\r\n3. Launch `strawberry server app`\r\n4. See it fails with ` Schema must contain uniquely named types but contains multiple types named '<InterfaceName>'`\r\n\r\n**Snippet to reproduce the issue**\r\n````python\r\nfrom typing import List, Optional, Union\r\nimport strawberry\r\nfrom strawberry import field\r\n\r\n\r\[email protected]\r\nclass Person:\r\n name: str\r\n email: str\r\n\r\n\r\[email protected]\r\nclass Speaker(Person):\r\n job: str \r\n\r\n\r\[email protected]\r\nclass Attendee(Person):\r\n interests: List[str]\r\n\r\n\r\ndef get_people_by_name(name: str): \r\n return []\r\n\r\n\r\[email protected]\r\nclass Query:\r\n searchPeopleByName: List[Union[Speaker, Attendee]] = field(resolver=get_people_by_name)\r\n\r\nschema = strawberry.Schema(query=Query)\r\n````\r\n**Full traceback:**\r\n```\r\n File \"/mnt/c/Users/<User>/code/nerdearla/test_app.py\", line 30, in <module>\r\n schema = strawberry.Schema(query=Query)\r\n File \"/home/crow/.virtualenvs/venv/lib/python3.8/site-packages/strawberry/schema/schema.py\", line 42, in __init__\r\n self._schema = GraphQLSchema(\r\n File \"/home/crow/.virtualenvs/venv/lib/python3.8/site-packages/graphql/type/schema.py\", line 240, in __init__\r\n raise TypeError(\r\nTypeError: Schema must contain uniquely named types but contains multiple types named 'Person'.\r\n```\n", "before_files": [{"content": "from typing import Type, cast\n\nfrom graphql import GraphQLInputObjectType, GraphQLObjectType\nfrom graphql.type.definition import GraphQLInterfaceType\nfrom strawberry.type import TypeDefinition\n\nfrom .fields import get_field\nfrom .types import ConcreteType, GraphQLType, TypeMap\n\n\ndef _get_object_type_for_type_definition(\n type_definition: TypeDefinition, type_map: TypeMap\n) -> GraphQLType:\n\n TypeClass: Type = GraphQLObjectType\n\n kwargs = {}\n\n if type_definition.is_input:\n TypeClass = GraphQLInputObjectType\n elif type_definition.is_interface:\n TypeClass = GraphQLInterfaceType\n\n if type_definition.interfaces:\n kwargs[\"interfaces\"] = [\n _get_object_type_for_type_definition(interface, type_map)\n for interface in type_definition.interfaces\n ]\n\n assert not type_definition.is_generic\n\n return TypeClass(\n name=type_definition.name,\n fields=lambda: {\n field.name: get_field(field, type_definition.is_input, type_map)\n for field in type_definition.fields\n },\n description=type_definition.description,\n **kwargs,\n )\n\n\ndef get_object_type(origin: Type, type_map: TypeMap) -> GraphQLObjectType:\n \"\"\"Returns a root type (Query, Mutation, Subscription) from a decorated type\"\"\"\n\n if not hasattr(origin, \"_type_definition\"):\n raise ValueError(f\"Wrong type passed to get object type {origin}\")\n\n type_definition: TypeDefinition = origin._type_definition\n\n name = type_definition.name\n\n if name not in type_map:\n object_type = _get_object_type_for_type_definition(type_definition, type_map)\n\n type_map[name] = ConcreteType(\n definition=type_definition, implementation=object_type\n )\n\n return cast(GraphQLObjectType, type_map[name].implementation)\n", "path": "strawberry/schema/types/object_type.py"}]}
| 1,423 | 670 |
gh_patches_debug_26521
|
rasdani/github-patches
|
git_diff
|
internetarchive__openlibrary-7718
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Publisher search endpoint solr performance
<!-- What problem are we solving? What does the experience look like today? What are the symptoms? -->
Our /search/publishers endpoint is doing a strange roll-up and submitting many solr select queries causing performance issues. Solution presumably is to not make more than 1 solr query on /search/publishers.
### Proposal
Change the backend call for /search/publishers to make a single query to solr `publisher:(...)` query.
### Evidence / Screenshot (if possible)

<img width="775" alt="Screenshot 2023-03-23 at 12 18 55 PM" src="https://user-images.githubusercontent.com/978325/227324919-d19b91c5-d19b-4746-9908-43e0f7cf1cbd.png">
### Relevant url?
<!-- `https://openlibrary.org/...` -->
http://testing.openlibrary.org/search/publishers?q=Black%20Dolls%20And%20White%20Dolls%20From%201940%20Through%201970%3A%20Their%20Impact%20Then%20On%20Black%20And%20White%20Children%27s%20Development%20
### Related files
<!-- Files related to this issue; this is super useful for new contributors who might want to help! If you're not sure, leave this blank; a maintainer will add them. -->
https://github.com/internetarchive/openlibrary/blob/b897c8c51a79308e38f9825fac82864a5cc7d3ae/openlibrary/plugins/worksearch/publishers.py#L82
### Stakeholders
<!-- @ tag stakeholders of this bug -->
</issue>
<code>
[start of openlibrary/plugins/worksearch/publishers.py]
1 """Publisher pages
2 """
3 from infogami.utils import delegate, stats
4 from infogami.utils.view import render_template, safeint
5 import web
6 import logging
7
8 from . import subjects
9 from . import search
10
11 logger = logging.getLogger("openlibrary.worksearch")
12
13
14 class publishers(subjects.subjects):
15 path = '(/publishers/[^/]+)'
16
17 def GET(self, key):
18 key = key.replace("_", " ")
19 page = subjects.get_subject(key, details=True)
20
21 if not page or page.work_count == 0:
22 web.ctx.status = "404 Not Found"
23 return render_template('publishers/notfound.tmpl', key)
24
25 return render_template("publishers/view", page)
26
27 def is_enabled(self):
28 return "publishers" in web.ctx.features
29
30
31 class publishers_json(subjects.subjects_json):
32 path = '(/publishers/[^/]+)'
33 encoding = "json"
34
35 def is_enabled(self):
36 return "publishers" in web.ctx.features
37
38 def normalize_key(self, key):
39 return key
40
41 def process_key(self, key):
42 return key.replace("_", " ")
43
44
45 class index(delegate.page):
46 path = "/publishers"
47
48 def GET(self):
49 return render_template("publishers/index")
50
51 def is_enabled(self):
52 return "publishers" in web.ctx.features
53
54
55 class publisher_search(delegate.page):
56 path = '/search/publishers'
57
58 def GET(self):
59 i = web.input(q="")
60 solr = search.get_solr()
61 q = {"publisher": i.q}
62
63 result = solr.select(
64 q,
65 facets=["publisher_facet"],
66 fields=["publisher", "publisher_facet"],
67 rows=0,
68 )
69 result = self.process_result(result)
70 return render_template('search/publishers', i.q, result)
71
72 def process_result(self, result):
73 solr = search.get_solr()
74
75 def process(p):
76 return web.storage(
77 name=p.value,
78 key="/publishers/" + p.value.replace(" ", "_"),
79 count=solr.select({"publisher_facet": p.value}, rows=0)['num_found'],
80 )
81
82 publisher_facets = result['facets']['publisher_facet'][:25]
83 return [process(p) for p in publisher_facets]
84
85
86 class PublisherEngine(subjects.SubjectEngine):
87 def normalize_key(self, key):
88 return key
89
90 def get_ebook_count(self, name, value, publish_year):
91 # Query solr for this publish_year and publish_year combination and read the has_fulltext=true facet
92 solr = search.get_solr()
93 q = {"publisher_facet": value}
94
95 if isinstance(publish_year, list):
96 q['publish_year'] = tuple(publish_year) # range
97 elif publish_year:
98 q['publish_year'] = publish_year
99
100 result = solr.select(q, facets=["has_fulltext"], rows=0)
101 counts = {v.value: v.count for v in result["facets"]["has_fulltext"]}
102 return counts.get('true')
103
104
105 def setup():
106 subjects.SUBJECTS.append(
107 subjects.SubjectMeta(
108 name="publisher",
109 key="publishers",
110 prefix="/publishers/",
111 facet="publisher_facet",
112 facet_key="publisher_facet",
113 Engine=PublisherEngine,
114 )
115 )
116
[end of openlibrary/plugins/worksearch/publishers.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/openlibrary/plugins/worksearch/publishers.py b/openlibrary/plugins/worksearch/publishers.py
--- a/openlibrary/plugins/worksearch/publishers.py
+++ b/openlibrary/plugins/worksearch/publishers.py
@@ -57,30 +57,28 @@
def GET(self):
i = web.input(q="")
- solr = search.get_solr()
- q = {"publisher": i.q}
-
- result = solr.select(
- q,
+ result = search.get_solr().select(
+ {"publisher": i.q, "type": "work"},
facets=["publisher_facet"],
- fields=["publisher", "publisher_facet"],
+ facet_mincount=1,
+ facet_limit=25,
+ facet_contains=i.q,
+ facet_contains_ignoreCase='true',
rows=0,
)
result = self.process_result(result)
return render_template('search/publishers', i.q, result)
def process_result(self, result):
- solr = search.get_solr()
-
- def process(p):
- return web.storage(
+ publisher_facets = result['facets']['publisher_facet']
+ return [
+ web.storage(
name=p.value,
key="/publishers/" + p.value.replace(" ", "_"),
- count=solr.select({"publisher_facet": p.value}, rows=0)['num_found'],
+ count=p.count,
)
-
- publisher_facets = result['facets']['publisher_facet'][:25]
- return [process(p) for p in publisher_facets]
+ for p in publisher_facets
+ ]
class PublisherEngine(subjects.SubjectEngine):
|
{"golden_diff": "diff --git a/openlibrary/plugins/worksearch/publishers.py b/openlibrary/plugins/worksearch/publishers.py\n--- a/openlibrary/plugins/worksearch/publishers.py\n+++ b/openlibrary/plugins/worksearch/publishers.py\n@@ -57,30 +57,28 @@\n \n def GET(self):\n i = web.input(q=\"\")\n- solr = search.get_solr()\n- q = {\"publisher\": i.q}\n-\n- result = solr.select(\n- q,\n+ result = search.get_solr().select(\n+ {\"publisher\": i.q, \"type\": \"work\"},\n facets=[\"publisher_facet\"],\n- fields=[\"publisher\", \"publisher_facet\"],\n+ facet_mincount=1,\n+ facet_limit=25,\n+ facet_contains=i.q,\n+ facet_contains_ignoreCase='true',\n rows=0,\n )\n result = self.process_result(result)\n return render_template('search/publishers', i.q, result)\n \n def process_result(self, result):\n- solr = search.get_solr()\n-\n- def process(p):\n- return web.storage(\n+ publisher_facets = result['facets']['publisher_facet']\n+ return [\n+ web.storage(\n name=p.value,\n key=\"/publishers/\" + p.value.replace(\" \", \"_\"),\n- count=solr.select({\"publisher_facet\": p.value}, rows=0)['num_found'],\n+ count=p.count,\n )\n-\n- publisher_facets = result['facets']['publisher_facet'][:25]\n- return [process(p) for p in publisher_facets]\n+ for p in publisher_facets\n+ ]\n \n \n class PublisherEngine(subjects.SubjectEngine):\n", "issue": "Publisher search endpoint solr performance\n<!-- What problem are we solving? What does the experience look like today? What are the symptoms? -->\r\n\r\nOur /search/publishers endpoint is doing a strange roll-up and submitting many solr select queries causing performance issues. Solution presumably is to not make more than 1 solr query on /search/publishers.\r\n\r\n### Proposal\r\n\r\nChange the backend call for /search/publishers to make a single query to solr `publisher:(...)` query.\r\n\r\n### Evidence / Screenshot (if possible)\r\n\r\n<img width=\"775\" alt=\"Screenshot 2023-03-23 at 12 18 55 PM\" src=\"https://user-images.githubusercontent.com/978325/227324919-d19b91c5-d19b-4746-9908-43e0f7cf1cbd.png\">\r\n\r\n### Relevant url?\r\n<!-- `https://openlibrary.org/...` -->\r\n\r\nhttp://testing.openlibrary.org/search/publishers?q=Black%20Dolls%20And%20White%20Dolls%20From%201940%20Through%201970%3A%20Their%20Impact%20Then%20On%20Black%20And%20White%20Children%27s%20Development%20\r\n\r\n### Related files\r\n<!-- Files related to this issue; this is super useful for new contributors who might want to help! If you're not sure, leave this blank; a maintainer will add them. -->\r\n\r\nhttps://github.com/internetarchive/openlibrary/blob/b897c8c51a79308e38f9825fac82864a5cc7d3ae/openlibrary/plugins/worksearch/publishers.py#L82\r\n\r\n### Stakeholders\r\n<!-- @ tag stakeholders of this bug -->\r\n\n", "before_files": [{"content": "\"\"\"Publisher pages\n\"\"\"\nfrom infogami.utils import delegate, stats\nfrom infogami.utils.view import render_template, safeint\nimport web\nimport logging\n\nfrom . import subjects\nfrom . import search\n\nlogger = logging.getLogger(\"openlibrary.worksearch\")\n\n\nclass publishers(subjects.subjects):\n path = '(/publishers/[^/]+)'\n\n def GET(self, key):\n key = key.replace(\"_\", \" \")\n page = subjects.get_subject(key, details=True)\n\n if not page or page.work_count == 0:\n web.ctx.status = \"404 Not Found\"\n return render_template('publishers/notfound.tmpl', key)\n\n return render_template(\"publishers/view\", page)\n\n def is_enabled(self):\n return \"publishers\" in web.ctx.features\n\n\nclass publishers_json(subjects.subjects_json):\n path = '(/publishers/[^/]+)'\n encoding = \"json\"\n\n def is_enabled(self):\n return \"publishers\" in web.ctx.features\n\n def normalize_key(self, key):\n return key\n\n def process_key(self, key):\n return key.replace(\"_\", \" \")\n\n\nclass index(delegate.page):\n path = \"/publishers\"\n\n def GET(self):\n return render_template(\"publishers/index\")\n\n def is_enabled(self):\n return \"publishers\" in web.ctx.features\n\n\nclass publisher_search(delegate.page):\n path = '/search/publishers'\n\n def GET(self):\n i = web.input(q=\"\")\n solr = search.get_solr()\n q = {\"publisher\": i.q}\n\n result = solr.select(\n q,\n facets=[\"publisher_facet\"],\n fields=[\"publisher\", \"publisher_facet\"],\n rows=0,\n )\n result = self.process_result(result)\n return render_template('search/publishers', i.q, result)\n\n def process_result(self, result):\n solr = search.get_solr()\n\n def process(p):\n return web.storage(\n name=p.value,\n key=\"/publishers/\" + p.value.replace(\" \", \"_\"),\n count=solr.select({\"publisher_facet\": p.value}, rows=0)['num_found'],\n )\n\n publisher_facets = result['facets']['publisher_facet'][:25]\n return [process(p) for p in publisher_facets]\n\n\nclass PublisherEngine(subjects.SubjectEngine):\n def normalize_key(self, key):\n return key\n\n def get_ebook_count(self, name, value, publish_year):\n # Query solr for this publish_year and publish_year combination and read the has_fulltext=true facet\n solr = search.get_solr()\n q = {\"publisher_facet\": value}\n\n if isinstance(publish_year, list):\n q['publish_year'] = tuple(publish_year) # range\n elif publish_year:\n q['publish_year'] = publish_year\n\n result = solr.select(q, facets=[\"has_fulltext\"], rows=0)\n counts = {v.value: v.count for v in result[\"facets\"][\"has_fulltext\"]}\n return counts.get('true')\n\n\ndef setup():\n subjects.SUBJECTS.append(\n subjects.SubjectMeta(\n name=\"publisher\",\n key=\"publishers\",\n prefix=\"/publishers/\",\n facet=\"publisher_facet\",\n facet_key=\"publisher_facet\",\n Engine=PublisherEngine,\n )\n )\n", "path": "openlibrary/plugins/worksearch/publishers.py"}]}
| 2,009 | 373 |
gh_patches_debug_18174
|
rasdani/github-patches
|
git_diff
|
ephios-dev__ephios-82
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Redirect anonymous users to login view instead of raising 403
this also raises 403 if users are not logged in. this is not what we want
_Originally posted by @jeriox in https://github.com/jeriox/jep/pull/48#discussion_r479789720_
</issue>
<code>
[start of jep/permissions.py]
1 import guardian.mixins
2 from django.contrib.auth.models import Permission, Group
3 from guardian.ctypes import get_content_type
4 from guardian.utils import get_group_obj_perms_model
5
6
7 def get_groups_with_perms(obj, only_with_perms_in):
8
9 ctype = get_content_type(obj)
10 group_model = get_group_obj_perms_model(obj)
11
12 group_rel_name = group_model.group.field.related_query_name()
13
14 if group_model.objects.is_generic():
15 group_filters = {
16 "%s__content_type" % group_rel_name: ctype,
17 "%s__object_pk" % group_rel_name: obj.pk,
18 }
19 else:
20 group_filters = {"%s__content_object" % group_rel_name: obj}
21
22 permission_ids = Permission.objects.filter(
23 content_type=ctype, codename__in=only_with_perms_in
24 ).values_list("id", flat=True)
25 group_filters.update(
26 {"%s__permission_id__in" % group_rel_name: permission_ids,}
27 )
28 return Group.objects.filter(**group_filters).distinct()
29
30
31 class CustomPermissionRequiredMixin(guardian.mixins.PermissionRequiredMixin):
32 raise_exception = True
33 accept_global_perms = True
34
35 # FIXME redirect non logged in users and raise Permission for others
36
[end of jep/permissions.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/jep/permissions.py b/jep/permissions.py
--- a/jep/permissions.py
+++ b/jep/permissions.py
@@ -1,8 +1,12 @@
import guardian.mixins
+from django.contrib.auth import REDIRECT_FIELD_NAME
from django.contrib.auth.models import Permission, Group
+from django.contrib.auth.views import redirect_to_login
from guardian.ctypes import get_content_type
from guardian.utils import get_group_obj_perms_model
+from jep import settings
+
def get_groups_with_perms(obj, only_with_perms_in):
@@ -32,4 +36,10 @@
raise_exception = True
accept_global_perms = True
- # FIXME redirect non logged in users and raise Permission for others
+ def on_permission_check_fail(self, request, response, obj=None):
+ if request.user.is_authenticated:
+ return response
+ else:
+ return redirect_to_login(
+ self.request.get_full_path(), settings.LOGIN_URL, REDIRECT_FIELD_NAME
+ )
|
{"golden_diff": "diff --git a/jep/permissions.py b/jep/permissions.py\n--- a/jep/permissions.py\n+++ b/jep/permissions.py\n@@ -1,8 +1,12 @@\n import guardian.mixins\n+from django.contrib.auth import REDIRECT_FIELD_NAME\n from django.contrib.auth.models import Permission, Group\n+from django.contrib.auth.views import redirect_to_login\n from guardian.ctypes import get_content_type\n from guardian.utils import get_group_obj_perms_model\n \n+from jep import settings\n+\n \n def get_groups_with_perms(obj, only_with_perms_in):\n \n@@ -32,4 +36,10 @@\n raise_exception = True\n accept_global_perms = True\n \n- # FIXME redirect non logged in users and raise Permission for others\n+ def on_permission_check_fail(self, request, response, obj=None):\n+ if request.user.is_authenticated:\n+ return response\n+ else:\n+ return redirect_to_login(\n+ self.request.get_full_path(), settings.LOGIN_URL, REDIRECT_FIELD_NAME\n+ )\n", "issue": "Redirect anonymous users to login view instead of raising 403\nthis also raises 403 if users are not logged in. this is not what we want\r\n\r\n_Originally posted by @jeriox in https://github.com/jeriox/jep/pull/48#discussion_r479789720_\n", "before_files": [{"content": "import guardian.mixins\nfrom django.contrib.auth.models import Permission, Group\nfrom guardian.ctypes import get_content_type\nfrom guardian.utils import get_group_obj_perms_model\n\n\ndef get_groups_with_perms(obj, only_with_perms_in):\n\n ctype = get_content_type(obj)\n group_model = get_group_obj_perms_model(obj)\n\n group_rel_name = group_model.group.field.related_query_name()\n\n if group_model.objects.is_generic():\n group_filters = {\n \"%s__content_type\" % group_rel_name: ctype,\n \"%s__object_pk\" % group_rel_name: obj.pk,\n }\n else:\n group_filters = {\"%s__content_object\" % group_rel_name: obj}\n\n permission_ids = Permission.objects.filter(\n content_type=ctype, codename__in=only_with_perms_in\n ).values_list(\"id\", flat=True)\n group_filters.update(\n {\"%s__permission_id__in\" % group_rel_name: permission_ids,}\n )\n return Group.objects.filter(**group_filters).distinct()\n\n\nclass CustomPermissionRequiredMixin(guardian.mixins.PermissionRequiredMixin):\n raise_exception = True\n accept_global_perms = True\n\n # FIXME redirect non logged in users and raise Permission for others\n", "path": "jep/permissions.py"}]}
| 930 | 222 |
gh_patches_debug_65915
|
rasdani/github-patches
|
git_diff
|
ipython__ipython-2812
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Cython -la doesn't work
It's parsed as -l instead.
``` python
@magic_arguments.argument(
'-la', '--link-args', action='append', default=[],
help="Extra flags to pass to linker via the `extra_link_args` "
"Extension flag (can be specified multiple times)."
)
@magic_arguments.argument(
'-l', '--lib', action='append', default=[],
help="Add a library to link the extension against (can be specified "
"multiple times)."
)
```
</issue>
<code>
[start of IPython/extensions/cythonmagic.py]
1 # -*- coding: utf-8 -*-
2 """
3 =====================
4 Cython related magics
5 =====================
6
7 Usage
8 =====
9
10 ``%%cython``
11
12 {CYTHON_DOC}
13
14 ``%%cython_inline``
15
16 {CYTHON_INLINE_DOC}
17
18 ``%%cython_pyximport``
19
20 {CYTHON_PYXIMPORT_DOC}
21
22 Author:
23 * Brian Granger
24
25 Parts of this code were taken from Cython.inline.
26 """
27 #-----------------------------------------------------------------------------
28 # Copyright (C) 2010-2011, IPython Development Team.
29 #
30 # Distributed under the terms of the Modified BSD License.
31 #
32 # The full license is in the file COPYING.txt, distributed with this software.
33 #-----------------------------------------------------------------------------
34
35 from __future__ import print_function
36
37 import imp
38 import io
39 import os
40 import re
41 import sys
42 import time
43
44 try:
45 import hashlib
46 except ImportError:
47 import md5 as hashlib
48
49 from distutils.core import Distribution, Extension
50 from distutils.command.build_ext import build_ext
51
52 from IPython.core import display
53 from IPython.core import magic_arguments
54 from IPython.core.magic import Magics, magics_class, cell_magic
55 from IPython.testing.skipdoctest import skip_doctest
56 from IPython.utils import py3compat
57
58 import Cython
59 from Cython.Compiler.Errors import CompileError
60 from Cython.Build.Dependencies import cythonize
61
62
63 @magics_class
64 class CythonMagics(Magics):
65
66 def __init__(self, shell):
67 super(CythonMagics,self).__init__(shell)
68 self._reloads = {}
69 self._code_cache = {}
70
71 def _import_all(self, module):
72 for k,v in module.__dict__.items():
73 if not k.startswith('__'):
74 self.shell.push({k:v})
75
76 @cell_magic
77 def cython_inline(self, line, cell):
78 """Compile and run a Cython code cell using Cython.inline.
79
80 This magic simply passes the body of the cell to Cython.inline
81 and returns the result. If the variables `a` and `b` are defined
82 in the user's namespace, here is a simple example that returns
83 their sum::
84
85 %%cython_inline
86 return a+b
87
88 For most purposes, we recommend the usage of the `%%cython` magic.
89 """
90 locs = self.shell.user_global_ns
91 globs = self.shell.user_ns
92 return Cython.inline(cell, locals=locs, globals=globs)
93
94 @cell_magic
95 def cython_pyximport(self, line, cell):
96 """Compile and import a Cython code cell using pyximport.
97
98 The contents of the cell are written to a `.pyx` file in the current
99 working directory, which is then imported using `pyximport`. This
100 magic requires a module name to be passed::
101
102 %%cython_pyximport modulename
103 def f(x):
104 return 2.0*x
105
106 The compiled module is then imported and all of its symbols are
107 injected into the user's namespace. For most purposes, we recommend
108 the usage of the `%%cython` magic.
109 """
110 module_name = line.strip()
111 if not module_name:
112 raise ValueError('module name must be given')
113 fname = module_name + '.pyx'
114 with io.open(fname, 'w', encoding='utf-8') as f:
115 f.write(cell)
116 if 'pyximport' not in sys.modules:
117 import pyximport
118 pyximport.install(reload_support=True)
119 if module_name in self._reloads:
120 module = self._reloads[module_name]
121 reload(module)
122 else:
123 __import__(module_name)
124 module = sys.modules[module_name]
125 self._reloads[module_name] = module
126 self._import_all(module)
127
128 @magic_arguments.magic_arguments()
129 @magic_arguments.argument(
130 '-c', '--compile-args', action='append', default=[],
131 help="Extra flags to pass to compiler via the `extra_compile_args` "
132 "Extension flag (can be specified multiple times)."
133 )
134 @magic_arguments.argument(
135 '-la', '--link-args', action='append', default=[],
136 help="Extra flags to pass to linker via the `extra_link_args` "
137 "Extension flag (can be specified multiple times)."
138 )
139 @magic_arguments.argument(
140 '-l', '--lib', action='append', default=[],
141 help="Add a library to link the extension against (can be specified "
142 "multiple times)."
143 )
144 @magic_arguments.argument(
145 '-L', dest='library_dirs', metavar='dir', action='append', default=[],
146 help="Add a path to the list of libary directories (can be specified "
147 "multiple times)."
148 )
149 @magic_arguments.argument(
150 '-I', '--include', action='append', default=[],
151 help="Add a path to the list of include directories (can be specified "
152 "multiple times)."
153 )
154 @magic_arguments.argument(
155 '-+', '--cplus', action='store_true', default=False,
156 help="Output a C++ rather than C file."
157 )
158 @magic_arguments.argument(
159 '-f', '--force', action='store_true', default=False,
160 help="Force the compilation of a new module, even if the source has been "
161 "previously compiled."
162 )
163 @magic_arguments.argument(
164 '-a', '--annotate', action='store_true', default=False,
165 help="Produce a colorized HTML version of the source."
166 )
167 @cell_magic
168 def cython(self, line, cell):
169 """Compile and import everything from a Cython code cell.
170
171 The contents of the cell are written to a `.pyx` file in the
172 directory `IPYTHONDIR/cython` using a filename with the hash of the
173 code. This file is then cythonized and compiled. The resulting module
174 is imported and all of its symbols are injected into the user's
175 namespace. The usage is similar to that of `%%cython_pyximport` but
176 you don't have to pass a module name::
177
178 %%cython
179 def f(x):
180 return 2.0*x
181
182 To compile OpenMP codes, pass the required `--compile-args`
183 and `--link-args`. For example with gcc::
184
185 %%cython --compile-args=-fopenmp --link-args=-fopenmp
186 ...
187 """
188 args = magic_arguments.parse_argstring(self.cython, line)
189 code = cell if cell.endswith('\n') else cell+'\n'
190 lib_dir = os.path.join(self.shell.ipython_dir, 'cython')
191 quiet = True
192 key = code, sys.version_info, sys.executable, Cython.__version__
193
194 if not os.path.exists(lib_dir):
195 os.makedirs(lib_dir)
196
197 if args.force:
198 # Force a new module name by adding the current time to the
199 # key which is hashed to determine the module name.
200 key += time.time(),
201
202 module_name = "_cython_magic_" + hashlib.md5(str(key).encode('utf-8')).hexdigest()
203 module_path = os.path.join(lib_dir, module_name + self.so_ext)
204
205 have_module = os.path.isfile(module_path)
206 need_cythonize = not have_module
207
208 if args.annotate:
209 html_file = os.path.join(lib_dir, module_name + '.html')
210 if not os.path.isfile(html_file):
211 need_cythonize = True
212
213 if need_cythonize:
214 c_include_dirs = args.include
215 if 'numpy' in code:
216 import numpy
217 c_include_dirs.append(numpy.get_include())
218 pyx_file = os.path.join(lib_dir, module_name + '.pyx')
219 pyx_file = py3compat.cast_bytes_py2(pyx_file, encoding=sys.getfilesystemencoding())
220 with io.open(pyx_file, 'w', encoding='utf-8') as f:
221 f.write(code)
222 extension = Extension(
223 name = module_name,
224 sources = [pyx_file],
225 include_dirs = c_include_dirs,
226 library_dirs = args.library_dirs,
227 extra_compile_args = args.compile_args,
228 extra_link_args = args.link_args,
229 libraries = args.lib,
230 language = 'c++' if args.cplus else 'c',
231 )
232 build_extension = self._get_build_extension()
233 try:
234 opts = dict(
235 quiet=quiet,
236 annotate = args.annotate,
237 force = True,
238 )
239 build_extension.extensions = cythonize([extension], **opts)
240 except CompileError:
241 return
242
243 if not have_module:
244 build_extension.build_temp = os.path.dirname(pyx_file)
245 build_extension.build_lib = lib_dir
246 build_extension.run()
247 self._code_cache[key] = module_name
248
249 module = imp.load_dynamic(module_name, module_path)
250 self._import_all(module)
251
252 if args.annotate:
253 try:
254 with io.open(html_file, encoding='utf-8') as f:
255 annotated_html = f.read()
256 except IOError as e:
257 # File could not be opened. Most likely the user has a version
258 # of Cython before 0.15.1 (when `cythonize` learned the
259 # `force` keyword argument) and has already compiled this
260 # exact source without annotation.
261 print('Cython completed successfully but the annotated '
262 'source could not be read.', file=sys.stderr)
263 print(e, file=sys.stderr)
264 else:
265 return display.HTML(self.clean_annotated_html(annotated_html))
266
267 @property
268 def so_ext(self):
269 """The extension suffix for compiled modules."""
270 try:
271 return self._so_ext
272 except AttributeError:
273 self._so_ext = self._get_build_extension().get_ext_filename('')
274 return self._so_ext
275
276 def _get_build_extension(self):
277 dist = Distribution()
278 config_files = dist.find_config_files()
279 try:
280 config_files.remove('setup.cfg')
281 except ValueError:
282 pass
283 dist.parse_config_files(config_files)
284 build_extension = build_ext(dist)
285 build_extension.finalize_options()
286 return build_extension
287
288 @staticmethod
289 def clean_annotated_html(html):
290 """Clean up the annotated HTML source.
291
292 Strips the link to the generated C or C++ file, which we do not
293 present to the user.
294 """
295 r = re.compile('<p>Raw output: <a href="(.*)">(.*)</a>')
296 html = '\n'.join(l for l in html.splitlines() if not r.match(l))
297 return html
298
299 __doc__ = __doc__.format(
300 CYTHON_DOC = ' '*8 + CythonMagics.cython.__doc__,
301 CYTHON_INLINE_DOC = ' '*8 + CythonMagics.cython_inline.__doc__,
302 CYTHON_PYXIMPORT_DOC = ' '*8 + CythonMagics.cython_pyximport.__doc__,
303 )
304
305 def load_ipython_extension(ip):
306 """Load the extension in IPython."""
307 ip.register_magics(CythonMagics)
308
[end of IPython/extensions/cythonmagic.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/IPython/extensions/cythonmagic.py b/IPython/extensions/cythonmagic.py
--- a/IPython/extensions/cythonmagic.py
+++ b/IPython/extensions/cythonmagic.py
@@ -132,7 +132,7 @@
"Extension flag (can be specified multiple times)."
)
@magic_arguments.argument(
- '-la', '--link-args', action='append', default=[],
+ '--link-args', action='append', default=[],
help="Extra flags to pass to linker via the `extra_link_args` "
"Extension flag (can be specified multiple times)."
)
|
{"golden_diff": "diff --git a/IPython/extensions/cythonmagic.py b/IPython/extensions/cythonmagic.py\n--- a/IPython/extensions/cythonmagic.py\n+++ b/IPython/extensions/cythonmagic.py\n@@ -132,7 +132,7 @@\n \"Extension flag (can be specified multiple times).\"\n )\n @magic_arguments.argument(\n- '-la', '--link-args', action='append', default=[],\n+ '--link-args', action='append', default=[],\n help=\"Extra flags to pass to linker via the `extra_link_args` \"\n \"Extension flag (can be specified multiple times).\"\n )\n", "issue": "Cython -la doesn't work\nIt's parsed as -l instead.\n\n``` python\n @magic_arguments.argument(\n '-la', '--link-args', action='append', default=[],\n help=\"Extra flags to pass to linker via the `extra_link_args` \"\n \"Extension flag (can be specified multiple times).\"\n )\n @magic_arguments.argument(\n '-l', '--lib', action='append', default=[],\n help=\"Add a library to link the extension against (can be specified \"\n \"multiple times).\"\n )\n```\n\n", "before_files": [{"content": "# -*- coding: utf-8 -*-\n\"\"\"\n=====================\nCython related magics\n=====================\n\nUsage\n=====\n\n``%%cython``\n\n{CYTHON_DOC}\n\n``%%cython_inline``\n\n{CYTHON_INLINE_DOC}\n\n``%%cython_pyximport``\n\n{CYTHON_PYXIMPORT_DOC}\n\nAuthor:\n* Brian Granger\n\nParts of this code were taken from Cython.inline.\n\"\"\"\n#-----------------------------------------------------------------------------\n# Copyright (C) 2010-2011, IPython Development Team.\n#\n# Distributed under the terms of the Modified BSD License.\n#\n# The full license is in the file COPYING.txt, distributed with this software.\n#-----------------------------------------------------------------------------\n\nfrom __future__ import print_function\n\nimport imp\nimport io\nimport os\nimport re\nimport sys\nimport time\n\ntry:\n import hashlib\nexcept ImportError:\n import md5 as hashlib\n\nfrom distutils.core import Distribution, Extension\nfrom distutils.command.build_ext import build_ext\n\nfrom IPython.core import display\nfrom IPython.core import magic_arguments\nfrom IPython.core.magic import Magics, magics_class, cell_magic\nfrom IPython.testing.skipdoctest import skip_doctest\nfrom IPython.utils import py3compat\n\nimport Cython\nfrom Cython.Compiler.Errors import CompileError\nfrom Cython.Build.Dependencies import cythonize\n\n\n@magics_class\nclass CythonMagics(Magics):\n\n def __init__(self, shell):\n super(CythonMagics,self).__init__(shell)\n self._reloads = {}\n self._code_cache = {}\n\n def _import_all(self, module):\n for k,v in module.__dict__.items():\n if not k.startswith('__'):\n self.shell.push({k:v})\n\n @cell_magic\n def cython_inline(self, line, cell):\n \"\"\"Compile and run a Cython code cell using Cython.inline.\n\n This magic simply passes the body of the cell to Cython.inline\n and returns the result. If the variables `a` and `b` are defined\n in the user's namespace, here is a simple example that returns\n their sum::\n\n %%cython_inline\n return a+b\n\n For most purposes, we recommend the usage of the `%%cython` magic.\n \"\"\"\n locs = self.shell.user_global_ns\n globs = self.shell.user_ns\n return Cython.inline(cell, locals=locs, globals=globs)\n\n @cell_magic\n def cython_pyximport(self, line, cell):\n \"\"\"Compile and import a Cython code cell using pyximport.\n\n The contents of the cell are written to a `.pyx` file in the current\n working directory, which is then imported using `pyximport`. This\n magic requires a module name to be passed::\n\n %%cython_pyximport modulename\n def f(x):\n return 2.0*x\n\n The compiled module is then imported and all of its symbols are\n injected into the user's namespace. For most purposes, we recommend\n the usage of the `%%cython` magic.\n \"\"\"\n module_name = line.strip()\n if not module_name:\n raise ValueError('module name must be given')\n fname = module_name + '.pyx'\n with io.open(fname, 'w', encoding='utf-8') as f:\n f.write(cell)\n if 'pyximport' not in sys.modules:\n import pyximport\n pyximport.install(reload_support=True)\n if module_name in self._reloads:\n module = self._reloads[module_name]\n reload(module)\n else:\n __import__(module_name)\n module = sys.modules[module_name]\n self._reloads[module_name] = module\n self._import_all(module)\n\n @magic_arguments.magic_arguments()\n @magic_arguments.argument(\n '-c', '--compile-args', action='append', default=[],\n help=\"Extra flags to pass to compiler via the `extra_compile_args` \"\n \"Extension flag (can be specified multiple times).\"\n )\n @magic_arguments.argument(\n '-la', '--link-args', action='append', default=[],\n help=\"Extra flags to pass to linker via the `extra_link_args` \"\n \"Extension flag (can be specified multiple times).\"\n )\n @magic_arguments.argument(\n '-l', '--lib', action='append', default=[],\n help=\"Add a library to link the extension against (can be specified \"\n \"multiple times).\"\n )\n @magic_arguments.argument(\n '-L', dest='library_dirs', metavar='dir', action='append', default=[],\n help=\"Add a path to the list of libary directories (can be specified \"\n \"multiple times).\"\n )\n @magic_arguments.argument(\n '-I', '--include', action='append', default=[],\n help=\"Add a path to the list of include directories (can be specified \"\n \"multiple times).\"\n )\n @magic_arguments.argument(\n '-+', '--cplus', action='store_true', default=False,\n help=\"Output a C++ rather than C file.\"\n )\n @magic_arguments.argument(\n '-f', '--force', action='store_true', default=False,\n help=\"Force the compilation of a new module, even if the source has been \"\n \"previously compiled.\"\n )\n @magic_arguments.argument(\n '-a', '--annotate', action='store_true', default=False,\n help=\"Produce a colorized HTML version of the source.\"\n )\n @cell_magic\n def cython(self, line, cell):\n \"\"\"Compile and import everything from a Cython code cell.\n\n The contents of the cell are written to a `.pyx` file in the\n directory `IPYTHONDIR/cython` using a filename with the hash of the\n code. This file is then cythonized and compiled. The resulting module\n is imported and all of its symbols are injected into the user's\n namespace. The usage is similar to that of `%%cython_pyximport` but\n you don't have to pass a module name::\n\n %%cython\n def f(x):\n return 2.0*x\n\n To compile OpenMP codes, pass the required `--compile-args`\n and `--link-args`. For example with gcc::\n\n %%cython --compile-args=-fopenmp --link-args=-fopenmp\n ...\n \"\"\"\n args = magic_arguments.parse_argstring(self.cython, line)\n code = cell if cell.endswith('\\n') else cell+'\\n'\n lib_dir = os.path.join(self.shell.ipython_dir, 'cython')\n quiet = True\n key = code, sys.version_info, sys.executable, Cython.__version__\n\n if not os.path.exists(lib_dir):\n os.makedirs(lib_dir)\n\n if args.force:\n # Force a new module name by adding the current time to the\n # key which is hashed to determine the module name.\n key += time.time(),\n\n module_name = \"_cython_magic_\" + hashlib.md5(str(key).encode('utf-8')).hexdigest()\n module_path = os.path.join(lib_dir, module_name + self.so_ext)\n\n have_module = os.path.isfile(module_path)\n need_cythonize = not have_module\n\n if args.annotate:\n html_file = os.path.join(lib_dir, module_name + '.html')\n if not os.path.isfile(html_file):\n need_cythonize = True\n\n if need_cythonize:\n c_include_dirs = args.include\n if 'numpy' in code:\n import numpy\n c_include_dirs.append(numpy.get_include())\n pyx_file = os.path.join(lib_dir, module_name + '.pyx')\n pyx_file = py3compat.cast_bytes_py2(pyx_file, encoding=sys.getfilesystemencoding())\n with io.open(pyx_file, 'w', encoding='utf-8') as f:\n f.write(code)\n extension = Extension(\n name = module_name,\n sources = [pyx_file],\n include_dirs = c_include_dirs,\n library_dirs = args.library_dirs,\n extra_compile_args = args.compile_args,\n extra_link_args = args.link_args,\n libraries = args.lib,\n language = 'c++' if args.cplus else 'c',\n )\n build_extension = self._get_build_extension()\n try:\n opts = dict(\n quiet=quiet,\n annotate = args.annotate,\n force = True,\n )\n build_extension.extensions = cythonize([extension], **opts)\n except CompileError:\n return\n\n if not have_module:\n build_extension.build_temp = os.path.dirname(pyx_file)\n build_extension.build_lib = lib_dir\n build_extension.run()\n self._code_cache[key] = module_name\n\n module = imp.load_dynamic(module_name, module_path)\n self._import_all(module)\n\n if args.annotate:\n try:\n with io.open(html_file, encoding='utf-8') as f:\n annotated_html = f.read()\n except IOError as e:\n # File could not be opened. Most likely the user has a version\n # of Cython before 0.15.1 (when `cythonize` learned the\n # `force` keyword argument) and has already compiled this\n # exact source without annotation.\n print('Cython completed successfully but the annotated '\n 'source could not be read.', file=sys.stderr)\n print(e, file=sys.stderr)\n else:\n return display.HTML(self.clean_annotated_html(annotated_html))\n\n @property\n def so_ext(self):\n \"\"\"The extension suffix for compiled modules.\"\"\"\n try:\n return self._so_ext\n except AttributeError:\n self._so_ext = self._get_build_extension().get_ext_filename('')\n return self._so_ext\n\n def _get_build_extension(self):\n dist = Distribution()\n config_files = dist.find_config_files()\n try:\n config_files.remove('setup.cfg')\n except ValueError:\n pass\n dist.parse_config_files(config_files)\n build_extension = build_ext(dist)\n build_extension.finalize_options()\n return build_extension\n\n @staticmethod\n def clean_annotated_html(html):\n \"\"\"Clean up the annotated HTML source.\n\n Strips the link to the generated C or C++ file, which we do not\n present to the user.\n \"\"\"\n r = re.compile('<p>Raw output: <a href=\"(.*)\">(.*)</a>')\n html = '\\n'.join(l for l in html.splitlines() if not r.match(l))\n return html\n\n__doc__ = __doc__.format(\n CYTHON_DOC = ' '*8 + CythonMagics.cython.__doc__,\n CYTHON_INLINE_DOC = ' '*8 + CythonMagics.cython_inline.__doc__,\n CYTHON_PYXIMPORT_DOC = ' '*8 + CythonMagics.cython_pyximport.__doc__,\n)\n\ndef load_ipython_extension(ip):\n \"\"\"Load the extension in IPython.\"\"\"\n ip.register_magics(CythonMagics)\n", "path": "IPython/extensions/cythonmagic.py"}]}
| 3,876 | 138 |
gh_patches_debug_15802
|
rasdani/github-patches
|
git_diff
|
lutris__lutris-1179
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Have logger scroll automatically only when at the bottom
Currently the logger scrolls whenever it outputs which makes scrolling up useless unless the game is stopped. This behavior is annoying.
</issue>
<code>
[start of lutris/gui/logwindow.py]
1 from gi.repository import Gtk
2 from lutris.gui.widgets.dialogs import Dialog
3
4
5 class LogTextView(Gtk.TextView):
6 def __init__(self, buffer):
7 super(LogTextView, self).__init__()
8
9 self.set_buffer(buffer)
10 self.set_editable(False)
11 self.set_monospace(True)
12 self.set_left_margin(10)
13 self.set_wrap_mode(Gtk.WrapMode.CHAR)
14 self.get_style_context().add_class('lutris-logview')
15 self.connect("size-allocate", self.autoscroll)
16
17 def autoscroll(self, *args):
18 adj = self.get_vadjustment()
19 adj.set_value(adj.get_upper() - adj.get_page_size())
20
21
22 class LogWindow(Dialog):
23 def __init__(self, title, buffer, parent):
24 super(LogWindow, self).__init__(title, parent, 0,
25 ('_OK', Gtk.ResponseType.OK))
26 self.set_size_request(640, 480)
27 self.grid = Gtk.Grid()
28 self.buffer = buffer
29 self.logtextview = LogTextView(self.buffer)
30
31 scrolledwindow = Gtk.ScrolledWindow(hexpand=True, vexpand=True,
32 child=self.logtextview)
33 self.vbox.add(scrolledwindow)
34 self.show_all()
35
[end of lutris/gui/logwindow.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/lutris/gui/logwindow.py b/lutris/gui/logwindow.py
--- a/lutris/gui/logwindow.py
+++ b/lutris/gui/logwindow.py
@@ -10,13 +10,16 @@
self.set_editable(False)
self.set_monospace(True)
self.set_left_margin(10)
+ self.scroll_max = 0
self.set_wrap_mode(Gtk.WrapMode.CHAR)
self.get_style_context().add_class('lutris-logview')
self.connect("size-allocate", self.autoscroll)
def autoscroll(self, *args):
adj = self.get_vadjustment()
- adj.set_value(adj.get_upper() - adj.get_page_size())
+ if adj.get_value() == self.scroll_max or self.scroll_max == 0:
+ adj.set_value(adj.get_upper() - adj.get_page_size())
+ self.scroll_max = adj.get_upper() - adj.get_page_size()
class LogWindow(Dialog):
|
{"golden_diff": "diff --git a/lutris/gui/logwindow.py b/lutris/gui/logwindow.py\n--- a/lutris/gui/logwindow.py\n+++ b/lutris/gui/logwindow.py\n@@ -10,13 +10,16 @@\n self.set_editable(False)\n self.set_monospace(True)\n self.set_left_margin(10)\n+ self.scroll_max = 0\n self.set_wrap_mode(Gtk.WrapMode.CHAR)\n self.get_style_context().add_class('lutris-logview')\n self.connect(\"size-allocate\", self.autoscroll)\n \n def autoscroll(self, *args):\n adj = self.get_vadjustment()\n- adj.set_value(adj.get_upper() - adj.get_page_size())\n+ if adj.get_value() == self.scroll_max or self.scroll_max == 0:\n+ adj.set_value(adj.get_upper() - adj.get_page_size())\n+ self.scroll_max = adj.get_upper() - adj.get_page_size()\n \n \n class LogWindow(Dialog):\n", "issue": "Have logger scroll automatically only when at the bottom\nCurrently the logger scrolls whenever it outputs which makes scrolling up useless unless the game is stopped. This behavior is annoying.\n", "before_files": [{"content": "from gi.repository import Gtk\nfrom lutris.gui.widgets.dialogs import Dialog\n\n\nclass LogTextView(Gtk.TextView):\n def __init__(self, buffer):\n super(LogTextView, self).__init__()\n\n self.set_buffer(buffer)\n self.set_editable(False)\n self.set_monospace(True)\n self.set_left_margin(10)\n self.set_wrap_mode(Gtk.WrapMode.CHAR)\n self.get_style_context().add_class('lutris-logview')\n self.connect(\"size-allocate\", self.autoscroll)\n\n def autoscroll(self, *args):\n adj = self.get_vadjustment()\n adj.set_value(adj.get_upper() - adj.get_page_size())\n\n\nclass LogWindow(Dialog):\n def __init__(self, title, buffer, parent):\n super(LogWindow, self).__init__(title, parent, 0,\n ('_OK', Gtk.ResponseType.OK))\n self.set_size_request(640, 480)\n self.grid = Gtk.Grid()\n self.buffer = buffer\n self.logtextview = LogTextView(self.buffer)\n\n scrolledwindow = Gtk.ScrolledWindow(hexpand=True, vexpand=True,\n child=self.logtextview)\n self.vbox.add(scrolledwindow)\n self.show_all()\n", "path": "lutris/gui/logwindow.py"}]}
| 897 | 214 |
gh_patches_debug_16370
|
rasdani/github-patches
|
git_diff
|
open-mmlab__mmaction2-676
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Localizer train cfg & test cfg ?
</issue>
<code>
[start of configs/_base_/models/bsn_tem.py]
1 # model settings
2 model = dict(
3 type='TEM',
4 temporal_dim=100,
5 boundary_ratio=0.1,
6 tem_feat_dim=400,
7 tem_hidden_dim=512,
8 tem_match_threshold=0.5)
9 # model training and testing settings
10 train_cfg = None
11 test_cfg = dict(average_clips='score')
12
[end of configs/_base_/models/bsn_tem.py]
[start of configs/_base_/models/bsn_pem.py]
1 # model settings
2 model = dict(
3 type='PEM',
4 pem_feat_dim=32,
5 pem_hidden_dim=256,
6 pem_u_ratio_m=1,
7 pem_u_ratio_l=2,
8 pem_high_temporal_iou_threshold=0.6,
9 pem_low_temporal_iou_threshold=0.2,
10 soft_nms_alpha=0.75,
11 soft_nms_low_threshold=0.65,
12 soft_nms_high_threshold=0.9,
13 post_process_top_k=100)
14 # model training and testing settings
15 train_cfg = None
16 test_cfg = dict(average_clips='score')
17
[end of configs/_base_/models/bsn_pem.py]
[start of configs/_base_/models/bmn_400x100.py]
1 # model settings
2 model = dict(
3 type='BMN',
4 temporal_dim=100,
5 boundary_ratio=0.5,
6 num_samples=32,
7 num_samples_per_bin=3,
8 feat_dim=400,
9 soft_nms_alpha=0.4,
10 soft_nms_low_threshold=0.5,
11 soft_nms_high_threshold=0.9,
12 post_process_top_k=100)
13 # model training and testing settings
14 train_cfg = None
15 test_cfg = dict(average_clips='score')
16
[end of configs/_base_/models/bmn_400x100.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/configs/_base_/models/bmn_400x100.py b/configs/_base_/models/bmn_400x100.py
--- a/configs/_base_/models/bmn_400x100.py
+++ b/configs/_base_/models/bmn_400x100.py
@@ -10,6 +10,3 @@
soft_nms_low_threshold=0.5,
soft_nms_high_threshold=0.9,
post_process_top_k=100)
-# model training and testing settings
-train_cfg = None
-test_cfg = dict(average_clips='score')
diff --git a/configs/_base_/models/bsn_pem.py b/configs/_base_/models/bsn_pem.py
--- a/configs/_base_/models/bsn_pem.py
+++ b/configs/_base_/models/bsn_pem.py
@@ -11,6 +11,3 @@
soft_nms_low_threshold=0.65,
soft_nms_high_threshold=0.9,
post_process_top_k=100)
-# model training and testing settings
-train_cfg = None
-test_cfg = dict(average_clips='score')
diff --git a/configs/_base_/models/bsn_tem.py b/configs/_base_/models/bsn_tem.py
--- a/configs/_base_/models/bsn_tem.py
+++ b/configs/_base_/models/bsn_tem.py
@@ -6,6 +6,3 @@
tem_feat_dim=400,
tem_hidden_dim=512,
tem_match_threshold=0.5)
-# model training and testing settings
-train_cfg = None
-test_cfg = dict(average_clips='score')
|
{"golden_diff": "diff --git a/configs/_base_/models/bmn_400x100.py b/configs/_base_/models/bmn_400x100.py\n--- a/configs/_base_/models/bmn_400x100.py\n+++ b/configs/_base_/models/bmn_400x100.py\n@@ -10,6 +10,3 @@\n soft_nms_low_threshold=0.5,\n soft_nms_high_threshold=0.9,\n post_process_top_k=100)\n-# model training and testing settings\n-train_cfg = None\n-test_cfg = dict(average_clips='score')\ndiff --git a/configs/_base_/models/bsn_pem.py b/configs/_base_/models/bsn_pem.py\n--- a/configs/_base_/models/bsn_pem.py\n+++ b/configs/_base_/models/bsn_pem.py\n@@ -11,6 +11,3 @@\n soft_nms_low_threshold=0.65,\n soft_nms_high_threshold=0.9,\n post_process_top_k=100)\n-# model training and testing settings\n-train_cfg = None\n-test_cfg = dict(average_clips='score')\ndiff --git a/configs/_base_/models/bsn_tem.py b/configs/_base_/models/bsn_tem.py\n--- a/configs/_base_/models/bsn_tem.py\n+++ b/configs/_base_/models/bsn_tem.py\n@@ -6,6 +6,3 @@\n tem_feat_dim=400,\n tem_hidden_dim=512,\n tem_match_threshold=0.5)\n-# model training and testing settings\n-train_cfg = None\n-test_cfg = dict(average_clips='score')\n", "issue": "Localizer train cfg & test cfg ?\n\n", "before_files": [{"content": "# model settings\nmodel = dict(\n type='TEM',\n temporal_dim=100,\n boundary_ratio=0.1,\n tem_feat_dim=400,\n tem_hidden_dim=512,\n tem_match_threshold=0.5)\n# model training and testing settings\ntrain_cfg = None\ntest_cfg = dict(average_clips='score')\n", "path": "configs/_base_/models/bsn_tem.py"}, {"content": "# model settings\nmodel = dict(\n type='PEM',\n pem_feat_dim=32,\n pem_hidden_dim=256,\n pem_u_ratio_m=1,\n pem_u_ratio_l=2,\n pem_high_temporal_iou_threshold=0.6,\n pem_low_temporal_iou_threshold=0.2,\n soft_nms_alpha=0.75,\n soft_nms_low_threshold=0.65,\n soft_nms_high_threshold=0.9,\n post_process_top_k=100)\n# model training and testing settings\ntrain_cfg = None\ntest_cfg = dict(average_clips='score')\n", "path": "configs/_base_/models/bsn_pem.py"}, {"content": "# model settings\nmodel = dict(\n type='BMN',\n temporal_dim=100,\n boundary_ratio=0.5,\n num_samples=32,\n num_samples_per_bin=3,\n feat_dim=400,\n soft_nms_alpha=0.4,\n soft_nms_low_threshold=0.5,\n soft_nms_high_threshold=0.9,\n post_process_top_k=100)\n# model training and testing settings\ntrain_cfg = None\ntest_cfg = dict(average_clips='score')\n", "path": "configs/_base_/models/bmn_400x100.py"}]}
| 1,022 | 397 |
gh_patches_debug_16280
|
rasdani/github-patches
|
git_diff
|
mirumee__ariadne-35
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
If value from resolver is callable, call it with **kwargs.
[Apollo doc](https://www.apollographql.com/docs/graphql-tools/resolvers) for default resolver says that if `field_name` resolves to function, it will be called with query arguments:
> Calls a function on obj with the relevant field name and passes the query arguments into that function
This can be useful for situations when parent resolver returned an object with getter functions.
</issue>
<code>
[start of ariadne/resolvers.py]
1 from graphql import GraphQLObjectType, GraphQLScalarType, GraphQLSchema
2 from graphql.execution.base import ResolveInfo
3
4
5 def resolve_parent_field(parent, name: str):
6 if isinstance(parent, dict):
7 return parent.get(name)
8 return getattr(parent, name, None)
9
10
11 def default_resolver(parent, info: ResolveInfo):
12 return resolve_parent_field(parent, info.field_name)
13
14
15 def resolve_to(name: str):
16 def resolver(parent, *_):
17 return resolve_parent_field(parent, name)
18
19 return resolver
20
21
22 def add_resolve_functions_to_schema(schema: GraphQLSchema, resolvers: dict):
23 for type_name, type_object in schema.get_type_map().items():
24 if isinstance(type_object, GraphQLObjectType):
25 add_resolve_functions_to_object(type_name, type_object, resolvers)
26 if isinstance(type_object, GraphQLScalarType):
27 add_resolve_functions_to_scalar(type_name, type_object, resolvers)
28
29
30 def add_resolve_functions_to_object(name: str, obj: GraphQLObjectType, resolvers: dict):
31 type_resolvers = resolvers.get(name, {})
32 for field_name, field_object in obj.fields.items():
33 field_resolver = type_resolvers.get(field_name)
34 if field_resolver:
35 field_object.resolver = field_resolver
36 elif field_object.resolver is None:
37 field_object.resolver = default_resolver
38
39
40 def add_resolve_functions_to_scalar(name: str, obj: GraphQLObjectType, resolvers: dict):
41 scalar_resolvers = resolvers.get(name, {})
42
43 serialize = scalar_resolvers.get("serialize", obj.serialize)
44 obj.serialize = serialize
45
46 parse_literal = scalar_resolvers.get("parse_literal", obj.parse_literal)
47 obj.parse_literal = parse_literal
48
49 parse_value = scalar_resolvers.get("parse_value", obj.parse_value)
50 obj.parse_value = parse_value
51
[end of ariadne/resolvers.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/ariadne/resolvers.py b/ariadne/resolvers.py
--- a/ariadne/resolvers.py
+++ b/ariadne/resolvers.py
@@ -2,19 +2,23 @@
from graphql.execution.base import ResolveInfo
-def resolve_parent_field(parent, name: str):
+def resolve_parent_field(parent, name: str, **kwargs: dict):
if isinstance(parent, dict):
- return parent.get(name)
- return getattr(parent, name, None)
+ value = parent.get(name)
+ else:
+ value = getattr(parent, name, None)
+ if callable(value):
+ return value(**kwargs)
+ return value
-def default_resolver(parent, info: ResolveInfo):
- return resolve_parent_field(parent, info.field_name)
+def default_resolver(parent, info: ResolveInfo, **kwargs):
+ return resolve_parent_field(parent, info.field_name, **kwargs)
def resolve_to(name: str):
- def resolver(parent, *_):
- return resolve_parent_field(parent, name)
+ def resolver(parent, *_, **kwargs):
+ return resolve_parent_field(parent, name, **kwargs)
return resolver
|
{"golden_diff": "diff --git a/ariadne/resolvers.py b/ariadne/resolvers.py\n--- a/ariadne/resolvers.py\n+++ b/ariadne/resolvers.py\n@@ -2,19 +2,23 @@\n from graphql.execution.base import ResolveInfo\n \n \n-def resolve_parent_field(parent, name: str):\n+def resolve_parent_field(parent, name: str, **kwargs: dict):\n if isinstance(parent, dict):\n- return parent.get(name)\n- return getattr(parent, name, None)\n+ value = parent.get(name)\n+ else:\n+ value = getattr(parent, name, None)\n+ if callable(value):\n+ return value(**kwargs)\n+ return value\n \n \n-def default_resolver(parent, info: ResolveInfo):\n- return resolve_parent_field(parent, info.field_name)\n+def default_resolver(parent, info: ResolveInfo, **kwargs):\n+ return resolve_parent_field(parent, info.field_name, **kwargs)\n \n \n def resolve_to(name: str):\n- def resolver(parent, *_):\n- return resolve_parent_field(parent, name)\n+ def resolver(parent, *_, **kwargs):\n+ return resolve_parent_field(parent, name, **kwargs)\n \n return resolver\n", "issue": "If value from resolver is callable, call it with **kwargs.\n[Apollo doc](https://www.apollographql.com/docs/graphql-tools/resolvers) for default resolver says that if `field_name` resolves to function, it will be called with query arguments:\r\n\r\n> Calls a function on obj with the relevant field name and passes the query arguments into that function\r\n\r\nThis can be useful for situations when parent resolver returned an object with getter functions.\n", "before_files": [{"content": "from graphql import GraphQLObjectType, GraphQLScalarType, GraphQLSchema\nfrom graphql.execution.base import ResolveInfo\n\n\ndef resolve_parent_field(parent, name: str):\n if isinstance(parent, dict):\n return parent.get(name)\n return getattr(parent, name, None)\n\n\ndef default_resolver(parent, info: ResolveInfo):\n return resolve_parent_field(parent, info.field_name)\n\n\ndef resolve_to(name: str):\n def resolver(parent, *_):\n return resolve_parent_field(parent, name)\n\n return resolver\n\n\ndef add_resolve_functions_to_schema(schema: GraphQLSchema, resolvers: dict):\n for type_name, type_object in schema.get_type_map().items():\n if isinstance(type_object, GraphQLObjectType):\n add_resolve_functions_to_object(type_name, type_object, resolvers)\n if isinstance(type_object, GraphQLScalarType):\n add_resolve_functions_to_scalar(type_name, type_object, resolvers)\n\n\ndef add_resolve_functions_to_object(name: str, obj: GraphQLObjectType, resolvers: dict):\n type_resolvers = resolvers.get(name, {})\n for field_name, field_object in obj.fields.items():\n field_resolver = type_resolvers.get(field_name)\n if field_resolver:\n field_object.resolver = field_resolver\n elif field_object.resolver is None:\n field_object.resolver = default_resolver\n\n\ndef add_resolve_functions_to_scalar(name: str, obj: GraphQLObjectType, resolvers: dict):\n scalar_resolvers = resolvers.get(name, {})\n\n serialize = scalar_resolvers.get(\"serialize\", obj.serialize)\n obj.serialize = serialize\n\n parse_literal = scalar_resolvers.get(\"parse_literal\", obj.parse_literal)\n obj.parse_literal = parse_literal\n\n parse_value = scalar_resolvers.get(\"parse_value\", obj.parse_value)\n obj.parse_value = parse_value\n", "path": "ariadne/resolvers.py"}]}
| 1,101 | 259 |
gh_patches_debug_13586
|
rasdani/github-patches
|
git_diff
|
pwndbg__pwndbg-146
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
"show print elements 0" causes exceptions on stop
```
pwndbg> show print elements
Limit on string chars or array elements to print is unlimited.
Traceback (most recent call last):
File "/home/david/.pwndbg/pwndbg/events.py", line 111, in caller
func()
File "/home/david/.pwndbg/pwndbg/strings.py", line 34, in update_length
length = int(message)
File "/home/david/.pwndbg/pwndbg/inthook.py", line 44, in __new__
return _int(_int(value, *a, **kw))
ValueError: invalid literal for int() with base 10: 'unlimited'
Python Exception <class 'ValueError'> invalid literal for int() with base 10: 'unlimited':
```
</issue>
<code>
[start of pwndbg/strings.py]
1 #!/usr/bin/env python
2 # -*- coding: utf-8 -*-
3 """
4 Functionality for resolving ASCII printable strings within
5 the debuggee's address space.
6 """
7 from __future__ import absolute_import
8 from __future__ import division
9 from __future__ import print_function
10 from __future__ import unicode_literals
11
12 import string
13
14 import gdb
15
16 import pwndbg.events
17 import pwndbg.memory
18 import pwndbg.typeinfo
19
20 length = 15
21
22 @pwndbg.events.stop
23 def update_length():
24 r"""
25 Unfortunately there's not a better way to get at this info.
26
27 >>> gdb.execute('show print elements', from_tty=False, to_string=True)
28 'Limit on string chars or array elements to print is 21.\n'
29 """
30 global length
31 message = gdb.execute('show print elements', from_tty=False, to_string=True)
32 message = message.split()[-1]
33 message = message.strip('.')
34 length = int(message)
35
36 def get(address, maxlen = None):
37 if maxlen is None:
38 maxlen = length
39
40 try:
41 sz = pwndbg.memory.string(address)
42 sz = sz.decode('latin-1', 'replace')
43
44 if not sz or not all(s in string.printable for s in sz):
45 return None
46 except Exception as e:
47 return None
48
49 if len(sz) < maxlen:
50 return sz
51
52 return sz[:maxlen] + '...'
53
[end of pwndbg/strings.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/pwndbg/strings.py b/pwndbg/strings.py
--- a/pwndbg/strings.py
+++ b/pwndbg/strings.py
@@ -31,7 +31,10 @@
message = gdb.execute('show print elements', from_tty=False, to_string=True)
message = message.split()[-1]
message = message.strip('.')
- length = int(message)
+ if message == 'unlimited':
+ length = 0
+ else:
+ length = int(message)
def get(address, maxlen = None):
if maxlen is None:
@@ -46,7 +49,7 @@
except Exception as e:
return None
- if len(sz) < maxlen:
+ if len(sz) < maxlen or not maxlen:
return sz
return sz[:maxlen] + '...'
|
{"golden_diff": "diff --git a/pwndbg/strings.py b/pwndbg/strings.py\n--- a/pwndbg/strings.py\n+++ b/pwndbg/strings.py\n@@ -31,7 +31,10 @@\n message = gdb.execute('show print elements', from_tty=False, to_string=True)\n message = message.split()[-1]\n message = message.strip('.')\n- length = int(message)\n+ if message == 'unlimited':\n+ length = 0\n+ else:\n+ length = int(message)\n \n def get(address, maxlen = None):\n if maxlen is None:\n@@ -46,7 +49,7 @@\n except Exception as e:\n return None\n \n- if len(sz) < maxlen:\n+ if len(sz) < maxlen or not maxlen:\n return sz\n \n return sz[:maxlen] + '...'\n", "issue": "\"show print elements 0\" causes exceptions on stop\n```\r\npwndbg> show print elements\r\nLimit on string chars or array elements to print is unlimited.\r\nTraceback (most recent call last):\r\n File \"/home/david/.pwndbg/pwndbg/events.py\", line 111, in caller\r\n func()\r\n File \"/home/david/.pwndbg/pwndbg/strings.py\", line 34, in update_length\r\n length = int(message)\r\n File \"/home/david/.pwndbg/pwndbg/inthook.py\", line 44, in __new__\r\n return _int(_int(value, *a, **kw))\r\nValueError: invalid literal for int() with base 10: 'unlimited'\r\nPython Exception <class 'ValueError'> invalid literal for int() with base 10: 'unlimited': \r\n```\n", "before_files": [{"content": "#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\"\"\"\nFunctionality for resolving ASCII printable strings within\nthe debuggee's address space.\n\"\"\"\nfrom __future__ import absolute_import\nfrom __future__ import division\nfrom __future__ import print_function\nfrom __future__ import unicode_literals\n\nimport string\n\nimport gdb\n\nimport pwndbg.events\nimport pwndbg.memory\nimport pwndbg.typeinfo\n\nlength = 15\n\[email protected]\ndef update_length():\n r\"\"\"\n Unfortunately there's not a better way to get at this info.\n\n >>> gdb.execute('show print elements', from_tty=False, to_string=True)\n 'Limit on string chars or array elements to print is 21.\\n'\n \"\"\"\n global length\n message = gdb.execute('show print elements', from_tty=False, to_string=True)\n message = message.split()[-1]\n message = message.strip('.')\n length = int(message)\n\ndef get(address, maxlen = None):\n if maxlen is None:\n maxlen = length\n\n try:\n sz = pwndbg.memory.string(address)\n sz = sz.decode('latin-1', 'replace')\n\n if not sz or not all(s in string.printable for s in sz):\n return None\n except Exception as e:\n return None\n\n if len(sz) < maxlen:\n return sz\n\n return sz[:maxlen] + '...'\n", "path": "pwndbg/strings.py"}]}
| 1,137 | 190 |
gh_patches_debug_1981
|
rasdani/github-patches
|
git_diff
|
vyperlang__vyper-2905
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Missing @view decorator for interface ERC20Detailed.py
### Version Information
* vyper Version (output of `vyper --version`): 0.3.3
* OS: linux
* Python Version (output of `python --version`): Python 3.9.5
### What's your issue about?
**Issue**
Error using `ERC20Detailed.py` as an interface to a vyper class. Trying to compile the following snippet produces the following error.
```
# @version 0.3.3
from vyper.interfaces import ERC20Detailed
@view
@external
def getSymbol() -> String[32]:
return ERC20Detailed(0x5f3b5DfEb7B28CDbD7FAba78963EE202a494e2A2).symbol()
```
**Error**
```
vyper.exceptions.StateAccessViolation: May not call state modifying function 'symbol' within a constant
function.vyper.exceptions.StateAccessViolation: May not call state modifying function 'symbol' within a constant function.
```
**Reason**
This issue occurs because `ERC20Detailed.py` does not contain `@view` decorator for its interfaces
### How can it be fixed?
Adding `@view` decorator to interface under `vyper.builtin_interfaces.ERC20Detailed.py`
```
@external
@view
def name() -> String[1]:
pass
@external
@view
def symbol() -> String[1]:
pass
@external
@view
def decimals() -> uint8:
pass
```
**Why?**
Running `vyper -f interface examples/tokens/ERC20.vy` generates the following
```
...
@view
@external
def name() -> String[32]:
pass
@view
@external
def symbol() -> String[32]:
pass
@view
@external
def decimals() -> uint8:
pass
...
```
Adding `@view` decorator to `vyper.builtin_interfaces.ERC20Detailed.py` would make interface consistent.
</issue>
<code>
[start of vyper/builtin_interfaces/ERC20Detailed.py]
1 """
2 NOTE: interface uses `String[1]` where 1 is the lower bound of the string returned by the function.
3 For end-users this means they can't use `implements: ERC20Detailed` unless their implementation
4 uses a value n >= 1. Regardless this is fine as one can't do String[0] where n == 0.
5 """
6
7 interface_code = """
8 @external
9 def name() -> String[1]:
10 pass
11
12 @external
13 def symbol() -> String[1]:
14 pass
15
16 @external
17 def decimals() -> uint8:
18 pass
19 """
20
[end of vyper/builtin_interfaces/ERC20Detailed.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/vyper/builtin_interfaces/ERC20Detailed.py b/vyper/builtin_interfaces/ERC20Detailed.py
--- a/vyper/builtin_interfaces/ERC20Detailed.py
+++ b/vyper/builtin_interfaces/ERC20Detailed.py
@@ -5,14 +5,17 @@
"""
interface_code = """
+@view
@external
def name() -> String[1]:
pass
+@view
@external
def symbol() -> String[1]:
pass
+@view
@external
def decimals() -> uint8:
pass
|
{"golden_diff": "diff --git a/vyper/builtin_interfaces/ERC20Detailed.py b/vyper/builtin_interfaces/ERC20Detailed.py\n--- a/vyper/builtin_interfaces/ERC20Detailed.py\n+++ b/vyper/builtin_interfaces/ERC20Detailed.py\n@@ -5,14 +5,17 @@\n \"\"\"\n \n interface_code = \"\"\"\n+@view\n @external\n def name() -> String[1]:\n pass\n \n+@view\n @external\n def symbol() -> String[1]:\n pass\n \n+@view\n @external\n def decimals() -> uint8:\n pass\n", "issue": "Missing @view decorator for interface ERC20Detailed.py\n### Version Information\r\n* vyper Version (output of `vyper --version`): 0.3.3\r\n* OS: linux\r\n* Python Version (output of `python --version`): Python 3.9.5\r\n### What's your issue about?\r\n**Issue**\r\nError using `ERC20Detailed.py` as an interface to a vyper class. Trying to compile the following snippet produces the following error.\r\n```\r\n# @version 0.3.3\r\n\r\nfrom vyper.interfaces import ERC20Detailed\r\n\r\n@view\r\n@external\r\ndef getSymbol() -> String[32]:\r\n return ERC20Detailed(0x5f3b5DfEb7B28CDbD7FAba78963EE202a494e2A2).symbol()\r\n```\r\n**Error**\r\n```\r\nvyper.exceptions.StateAccessViolation: May not call state modifying function 'symbol' within a constant\r\nfunction.vyper.exceptions.StateAccessViolation: May not call state modifying function 'symbol' within a constant function.\r\n```\r\n**Reason**\r\nThis issue occurs because `ERC20Detailed.py` does not contain `@view` decorator for its interfaces\r\n### How can it be fixed?\r\nAdding `@view` decorator to interface under `vyper.builtin_interfaces.ERC20Detailed.py`\r\n```\r\n@external\r\n@view\r\ndef name() -> String[1]:\r\n pass\r\n \r\n@external\r\n@view\r\ndef symbol() -> String[1]:\r\n pass\r\n \r\n@external\r\n@view\r\ndef decimals() -> uint8:\r\n pass\r\n```\r\n**Why?**\r\nRunning `vyper -f interface examples/tokens/ERC20.vy` generates the following\r\n```\r\n...\r\n@view\r\n@external\r\ndef name() -> String[32]:\r\n pass\r\n \r\n@view\r\n@external\r\ndef symbol() -> String[32]:\r\n pass\r\n \r\n@view\r\n@external\r\ndef decimals() -> uint8:\r\n pass\r\n...\r\n```\r\n\r\nAdding `@view` decorator to `vyper.builtin_interfaces.ERC20Detailed.py` would make interface consistent.\n", "before_files": [{"content": "\"\"\"\nNOTE: interface uses `String[1]` where 1 is the lower bound of the string returned by the function.\n For end-users this means they can't use `implements: ERC20Detailed` unless their implementation\n uses a value n >= 1. Regardless this is fine as one can't do String[0] where n == 0.\n\"\"\"\n\ninterface_code = \"\"\"\n@external\ndef name() -> String[1]:\n pass\n\n@external\ndef symbol() -> String[1]:\n pass\n\n@external\ndef decimals() -> uint8:\n pass\n\"\"\"\n", "path": "vyper/builtin_interfaces/ERC20Detailed.py"}]}
| 1,158 | 127 |
gh_patches_debug_22138
|
rasdani/github-patches
|
git_diff
|
mkdocs__mkdocs-2604
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
CVE-2021-40978 - Path Traversal.
Hey!
We have verified a security flaw in the current version of MKdocs, a path traversal failure affecting the built-in dev-server.
That flaw turns the server susceptible to providing data outside the scope of the application allowing anyone to request sensitive files.
If you need further information, don't hesitate to get in touch with me.
https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-40978
https://github.com/nisdn/CVE-2021-40978
</issue>
<code>
[start of mkdocs/livereload/__init__.py]
1 import functools
2 import io
3 import logging
4 import mimetypes
5 import os
6 import os.path
7 import re
8 import socketserver
9 import threading
10 import time
11 import warnings
12 import wsgiref.simple_server
13
14 import watchdog.events
15 import watchdog.observers.polling
16
17
18 class _LoggerAdapter(logging.LoggerAdapter):
19 def process(self, msg, kwargs):
20 return time.strftime("[%H:%M:%S] ") + msg, kwargs
21
22
23 log = _LoggerAdapter(logging.getLogger(__name__), {})
24
25
26 class LiveReloadServer(socketserver.ThreadingMixIn, wsgiref.simple_server.WSGIServer):
27 daemon_threads = True
28 poll_response_timeout = 60
29
30 def __init__(
31 self,
32 builder,
33 host,
34 port,
35 root,
36 mount_path="/",
37 polling_interval=0.5,
38 shutdown_delay=0.25,
39 **kwargs,
40 ):
41 self.builder = builder
42 self.server_name = host
43 self.server_port = port
44 self.root = os.path.abspath(root)
45 self.mount_path = ("/" + mount_path.lstrip("/")).rstrip("/") + "/"
46 self.url = f"http://{self.server_name}:{self.server_port}{self.mount_path}"
47 self.build_delay = 0.1
48 self.shutdown_delay = shutdown_delay
49 # To allow custom error pages.
50 self.error_handler = lambda code: None
51
52 super().__init__((host, port), _Handler, **kwargs)
53 self.set_app(self.serve_request)
54
55 self._wanted_epoch = _timestamp() # The version of the site that started building.
56 self._visible_epoch = self._wanted_epoch # Latest fully built version of the site.
57 self._epoch_cond = threading.Condition() # Must be held when accessing _visible_epoch.
58
59 self._to_rebuild = {} # Used as an ordered set of functions to call.
60 self._rebuild_cond = threading.Condition() # Must be held when accessing _to_rebuild.
61
62 self._shutdown = False
63 self.serve_thread = threading.Thread(target=lambda: self.serve_forever(shutdown_delay))
64 self.observer = watchdog.observers.polling.PollingObserver(timeout=polling_interval)
65
66 def watch(self, path, func=None, recursive=True):
67 """Add the 'path' to watched paths, call the function and reload when any file changes under it."""
68 path = os.path.abspath(path)
69 if func in (None, self.builder):
70 func = self.builder
71 else:
72 warnings.warn(
73 "Plugins should not pass the 'func' parameter of watch(). "
74 "The ability to execute custom callbacks will be removed soon.",
75 DeprecationWarning,
76 stacklevel=2,
77 )
78
79 def callback(event):
80 if event.is_directory:
81 return
82 log.debug(str(event))
83 with self._rebuild_cond:
84 self._to_rebuild[func] = True
85 self._rebuild_cond.notify_all()
86
87 handler = watchdog.events.FileSystemEventHandler()
88 handler.on_any_event = callback
89 log.debug(f"Watching '{path}'")
90 self.observer.schedule(handler, path, recursive=recursive)
91
92 def serve(self):
93 self.observer.start()
94
95 log.info(f"Serving on {self.url}")
96 self.serve_thread.start()
97
98 self._build_loop()
99
100 def _build_loop(self):
101 while True:
102 with self._rebuild_cond:
103 while not self._rebuild_cond.wait_for(
104 lambda: self._to_rebuild or self._shutdown, timeout=self.shutdown_delay
105 ):
106 # We could have used just one wait instead of a loop + timeout, but we need
107 # occasional breaks, otherwise on Windows we can't receive KeyboardInterrupt.
108 pass
109 if self._shutdown:
110 break
111 log.info("Detected file changes")
112 while self._rebuild_cond.wait(timeout=self.build_delay):
113 log.debug("Waiting for file changes to stop happening")
114
115 self._wanted_epoch = _timestamp()
116 funcs = list(self._to_rebuild)
117 self._to_rebuild.clear()
118
119 for func in funcs:
120 func()
121
122 with self._epoch_cond:
123 log.info("Reloading browsers")
124 self._visible_epoch = self._wanted_epoch
125 self._epoch_cond.notify_all()
126
127 def shutdown(self):
128 self.observer.stop()
129 with self._rebuild_cond:
130 self._shutdown = True
131 self._rebuild_cond.notify_all()
132
133 if self.serve_thread.is_alive():
134 super().shutdown()
135 self.serve_thread.join()
136 self.observer.join()
137
138 def serve_request(self, environ, start_response):
139 try:
140 result = self._serve_request(environ, start_response)
141 except Exception:
142 code = 500
143 msg = "500 Internal Server Error"
144 log.exception(msg)
145 else:
146 if result is not None:
147 return result
148 code = 404
149 msg = "404 Not Found"
150
151 error_content = None
152 try:
153 error_content = self.error_handler(code)
154 except Exception:
155 log.exception("Failed to render an error message!")
156 if error_content is None:
157 error_content = msg.encode()
158
159 start_response(msg, [("Content-Type", "text/html")])
160 return [error_content]
161
162 def _serve_request(self, environ, start_response):
163 # https://bugs.python.org/issue16679
164 # https://github.com/bottlepy/bottle/blob/f9b1849db4/bottle.py#L984
165 path = environ["PATH_INFO"].encode("latin-1").decode("utf-8", "ignore")
166
167 m = re.fullmatch(r"/livereload/([0-9]+)/[0-9]+", path)
168 if m:
169 epoch = int(m[1])
170 start_response("200 OK", [("Content-Type", "text/plain")])
171
172 def condition():
173 return self._visible_epoch > epoch
174
175 with self._epoch_cond:
176 if not condition():
177 # Stall the browser, respond as soon as there's something new.
178 # If there's not, respond anyway after a minute.
179 self._log_poll_request(environ.get("HTTP_REFERER"), request_id=path)
180 self._epoch_cond.wait_for(condition, timeout=self.poll_response_timeout)
181 return [b"%d" % self._visible_epoch]
182
183 if path == "/js/livereload.js":
184 file_path = os.path.join(os.path.dirname(os.path.abspath(__file__)), "livereload.js")
185 elif path.startswith(self.mount_path):
186 rel_file_path = path[len(self.mount_path):].lstrip("/")
187 if path.endswith("/"):
188 rel_file_path += "index.html"
189 file_path = os.path.join(self.root, rel_file_path)
190 elif path == "/":
191 start_response("302 Found", [("Location", self.mount_path)])
192 return []
193 else:
194 return None # Not found
195
196 # Wait until the ongoing rebuild (if any) finishes, so we're not serving a half-built site.
197 with self._epoch_cond:
198 self._epoch_cond.wait_for(lambda: self._visible_epoch == self._wanted_epoch)
199 epoch = self._visible_epoch
200
201 try:
202 file = open(file_path, "rb")
203 except OSError:
204 if not path.endswith("/") and os.path.isfile(os.path.join(file_path, "index.html")):
205 start_response("302 Found", [("Location", path + "/")])
206 return []
207 return None # Not found
208
209 if file_path.endswith(".html"):
210 with file:
211 content = file.read()
212 content = self._inject_js_into_html(content, epoch)
213 file = io.BytesIO(content)
214 content_length = len(content)
215 else:
216 content_length = os.path.getsize(file_path)
217
218 content_type = self._guess_type(file_path)
219 start_response(
220 "200 OK", [("Content-Type", content_type), ("Content-Length", str(content_length))]
221 )
222 return wsgiref.util.FileWrapper(file)
223
224 @classmethod
225 def _inject_js_into_html(cls, content, epoch):
226 try:
227 body_end = content.rindex(b"</body>")
228 except ValueError:
229 body_end = len(content)
230 # The page will reload if the livereload poller returns a newer epoch than what it knows.
231 # The other timestamp becomes just a unique identifier for the initiating page.
232 return (
233 b'%b<script src="/js/livereload.js"></script><script>livereload(%d, %d);</script>%b'
234 % (content[:body_end], epoch, _timestamp(), content[body_end:])
235 )
236
237 @classmethod
238 @functools.lru_cache() # "Cache" to not repeat the same message for the same browser tab.
239 def _log_poll_request(cls, url, request_id):
240 log.info(f"Browser connected: {url}")
241
242 def _guess_type(cls, path):
243 # MkDocs only ensures a few common types (as seen in livereload_tests.py::test_mime_types).
244 # Other uncommon types will not be accepted.
245 if path.endswith((".js", ".JS")):
246 return "application/javascript"
247 if path.endswith(".gz"):
248 return "application/gzip"
249
250 guess, _ = mimetypes.guess_type(path)
251 if guess:
252 return guess
253 return "application/octet-stream"
254
255
256 class _Handler(wsgiref.simple_server.WSGIRequestHandler):
257 def log_request(self, code="-", size="-"):
258 level = logging.DEBUG if str(code) == "200" else logging.WARNING
259 log.log(level, f'"{self.requestline}" code {code}')
260
261 def log_message(self, format, *args):
262 log.debug(format, *args)
263
264
265 def _timestamp():
266 return round(time.monotonic() * 1000)
267
[end of mkdocs/livereload/__init__.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/mkdocs/livereload/__init__.py b/mkdocs/livereload/__init__.py
--- a/mkdocs/livereload/__init__.py
+++ b/mkdocs/livereload/__init__.py
@@ -4,6 +4,7 @@
import mimetypes
import os
import os.path
+import posixpath
import re
import socketserver
import threading
@@ -183,9 +184,11 @@
if path == "/js/livereload.js":
file_path = os.path.join(os.path.dirname(os.path.abspath(__file__)), "livereload.js")
elif path.startswith(self.mount_path):
- rel_file_path = path[len(self.mount_path):].lstrip("/")
+ rel_file_path = path[len(self.mount_path):]
if path.endswith("/"):
rel_file_path += "index.html"
+ # Prevent directory traversal - normalize the path.
+ rel_file_path = posixpath.normpath("/" + rel_file_path).lstrip("/")
file_path = os.path.join(self.root, rel_file_path)
elif path == "/":
start_response("302 Found", [("Location", self.mount_path)])
|
{"golden_diff": "diff --git a/mkdocs/livereload/__init__.py b/mkdocs/livereload/__init__.py\n--- a/mkdocs/livereload/__init__.py\n+++ b/mkdocs/livereload/__init__.py\n@@ -4,6 +4,7 @@\n import mimetypes\n import os\n import os.path\n+import posixpath\n import re\n import socketserver\n import threading\n@@ -183,9 +184,11 @@\n if path == \"/js/livereload.js\":\n file_path = os.path.join(os.path.dirname(os.path.abspath(__file__)), \"livereload.js\")\n elif path.startswith(self.mount_path):\n- rel_file_path = path[len(self.mount_path):].lstrip(\"/\")\n+ rel_file_path = path[len(self.mount_path):]\n if path.endswith(\"/\"):\n rel_file_path += \"index.html\"\n+ # Prevent directory traversal - normalize the path.\n+ rel_file_path = posixpath.normpath(\"/\" + rel_file_path).lstrip(\"/\")\n file_path = os.path.join(self.root, rel_file_path)\n elif path == \"/\":\n start_response(\"302 Found\", [(\"Location\", self.mount_path)])\n", "issue": "CVE-2021-40978 - Path Traversal.\nHey!\r\n\r\nWe have verified a security flaw in the current version of MKdocs, a path traversal failure affecting the built-in dev-server.\r\n\r\nThat flaw turns the server susceptible to providing data outside the scope of the application allowing anyone to request sensitive files.\r\n\r\nIf you need further information, don't hesitate to get in touch with me.\r\n\r\nhttps://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-40978\r\nhttps://github.com/nisdn/CVE-2021-40978\r\n\n", "before_files": [{"content": "import functools\nimport io\nimport logging\nimport mimetypes\nimport os\nimport os.path\nimport re\nimport socketserver\nimport threading\nimport time\nimport warnings\nimport wsgiref.simple_server\n\nimport watchdog.events\nimport watchdog.observers.polling\n\n\nclass _LoggerAdapter(logging.LoggerAdapter):\n def process(self, msg, kwargs):\n return time.strftime(\"[%H:%M:%S] \") + msg, kwargs\n\n\nlog = _LoggerAdapter(logging.getLogger(__name__), {})\n\n\nclass LiveReloadServer(socketserver.ThreadingMixIn, wsgiref.simple_server.WSGIServer):\n daemon_threads = True\n poll_response_timeout = 60\n\n def __init__(\n self,\n builder,\n host,\n port,\n root,\n mount_path=\"/\",\n polling_interval=0.5,\n shutdown_delay=0.25,\n **kwargs,\n ):\n self.builder = builder\n self.server_name = host\n self.server_port = port\n self.root = os.path.abspath(root)\n self.mount_path = (\"/\" + mount_path.lstrip(\"/\")).rstrip(\"/\") + \"/\"\n self.url = f\"http://{self.server_name}:{self.server_port}{self.mount_path}\"\n self.build_delay = 0.1\n self.shutdown_delay = shutdown_delay\n # To allow custom error pages.\n self.error_handler = lambda code: None\n\n super().__init__((host, port), _Handler, **kwargs)\n self.set_app(self.serve_request)\n\n self._wanted_epoch = _timestamp() # The version of the site that started building.\n self._visible_epoch = self._wanted_epoch # Latest fully built version of the site.\n self._epoch_cond = threading.Condition() # Must be held when accessing _visible_epoch.\n\n self._to_rebuild = {} # Used as an ordered set of functions to call.\n self._rebuild_cond = threading.Condition() # Must be held when accessing _to_rebuild.\n\n self._shutdown = False\n self.serve_thread = threading.Thread(target=lambda: self.serve_forever(shutdown_delay))\n self.observer = watchdog.observers.polling.PollingObserver(timeout=polling_interval)\n\n def watch(self, path, func=None, recursive=True):\n \"\"\"Add the 'path' to watched paths, call the function and reload when any file changes under it.\"\"\"\n path = os.path.abspath(path)\n if func in (None, self.builder):\n func = self.builder\n else:\n warnings.warn(\n \"Plugins should not pass the 'func' parameter of watch(). \"\n \"The ability to execute custom callbacks will be removed soon.\",\n DeprecationWarning,\n stacklevel=2,\n )\n\n def callback(event):\n if event.is_directory:\n return\n log.debug(str(event))\n with self._rebuild_cond:\n self._to_rebuild[func] = True\n self._rebuild_cond.notify_all()\n\n handler = watchdog.events.FileSystemEventHandler()\n handler.on_any_event = callback\n log.debug(f\"Watching '{path}'\")\n self.observer.schedule(handler, path, recursive=recursive)\n\n def serve(self):\n self.observer.start()\n\n log.info(f\"Serving on {self.url}\")\n self.serve_thread.start()\n\n self._build_loop()\n\n def _build_loop(self):\n while True:\n with self._rebuild_cond:\n while not self._rebuild_cond.wait_for(\n lambda: self._to_rebuild or self._shutdown, timeout=self.shutdown_delay\n ):\n # We could have used just one wait instead of a loop + timeout, but we need\n # occasional breaks, otherwise on Windows we can't receive KeyboardInterrupt.\n pass\n if self._shutdown:\n break\n log.info(\"Detected file changes\")\n while self._rebuild_cond.wait(timeout=self.build_delay):\n log.debug(\"Waiting for file changes to stop happening\")\n\n self._wanted_epoch = _timestamp()\n funcs = list(self._to_rebuild)\n self._to_rebuild.clear()\n\n for func in funcs:\n func()\n\n with self._epoch_cond:\n log.info(\"Reloading browsers\")\n self._visible_epoch = self._wanted_epoch\n self._epoch_cond.notify_all()\n\n def shutdown(self):\n self.observer.stop()\n with self._rebuild_cond:\n self._shutdown = True\n self._rebuild_cond.notify_all()\n\n if self.serve_thread.is_alive():\n super().shutdown()\n self.serve_thread.join()\n self.observer.join()\n\n def serve_request(self, environ, start_response):\n try:\n result = self._serve_request(environ, start_response)\n except Exception:\n code = 500\n msg = \"500 Internal Server Error\"\n log.exception(msg)\n else:\n if result is not None:\n return result\n code = 404\n msg = \"404 Not Found\"\n\n error_content = None\n try:\n error_content = self.error_handler(code)\n except Exception:\n log.exception(\"Failed to render an error message!\")\n if error_content is None:\n error_content = msg.encode()\n\n start_response(msg, [(\"Content-Type\", \"text/html\")])\n return [error_content]\n\n def _serve_request(self, environ, start_response):\n # https://bugs.python.org/issue16679\n # https://github.com/bottlepy/bottle/blob/f9b1849db4/bottle.py#L984\n path = environ[\"PATH_INFO\"].encode(\"latin-1\").decode(\"utf-8\", \"ignore\")\n\n m = re.fullmatch(r\"/livereload/([0-9]+)/[0-9]+\", path)\n if m:\n epoch = int(m[1])\n start_response(\"200 OK\", [(\"Content-Type\", \"text/plain\")])\n\n def condition():\n return self._visible_epoch > epoch\n\n with self._epoch_cond:\n if not condition():\n # Stall the browser, respond as soon as there's something new.\n # If there's not, respond anyway after a minute.\n self._log_poll_request(environ.get(\"HTTP_REFERER\"), request_id=path)\n self._epoch_cond.wait_for(condition, timeout=self.poll_response_timeout)\n return [b\"%d\" % self._visible_epoch]\n\n if path == \"/js/livereload.js\":\n file_path = os.path.join(os.path.dirname(os.path.abspath(__file__)), \"livereload.js\")\n elif path.startswith(self.mount_path):\n rel_file_path = path[len(self.mount_path):].lstrip(\"/\")\n if path.endswith(\"/\"):\n rel_file_path += \"index.html\"\n file_path = os.path.join(self.root, rel_file_path)\n elif path == \"/\":\n start_response(\"302 Found\", [(\"Location\", self.mount_path)])\n return []\n else:\n return None # Not found\n\n # Wait until the ongoing rebuild (if any) finishes, so we're not serving a half-built site.\n with self._epoch_cond:\n self._epoch_cond.wait_for(lambda: self._visible_epoch == self._wanted_epoch)\n epoch = self._visible_epoch\n\n try:\n file = open(file_path, \"rb\")\n except OSError:\n if not path.endswith(\"/\") and os.path.isfile(os.path.join(file_path, \"index.html\")):\n start_response(\"302 Found\", [(\"Location\", path + \"/\")])\n return []\n return None # Not found\n\n if file_path.endswith(\".html\"):\n with file:\n content = file.read()\n content = self._inject_js_into_html(content, epoch)\n file = io.BytesIO(content)\n content_length = len(content)\n else:\n content_length = os.path.getsize(file_path)\n\n content_type = self._guess_type(file_path)\n start_response(\n \"200 OK\", [(\"Content-Type\", content_type), (\"Content-Length\", str(content_length))]\n )\n return wsgiref.util.FileWrapper(file)\n\n @classmethod\n def _inject_js_into_html(cls, content, epoch):\n try:\n body_end = content.rindex(b\"</body>\")\n except ValueError:\n body_end = len(content)\n # The page will reload if the livereload poller returns a newer epoch than what it knows.\n # The other timestamp becomes just a unique identifier for the initiating page.\n return (\n b'%b<script src=\"/js/livereload.js\"></script><script>livereload(%d, %d);</script>%b'\n % (content[:body_end], epoch, _timestamp(), content[body_end:])\n )\n\n @classmethod\n @functools.lru_cache() # \"Cache\" to not repeat the same message for the same browser tab.\n def _log_poll_request(cls, url, request_id):\n log.info(f\"Browser connected: {url}\")\n\n def _guess_type(cls, path):\n # MkDocs only ensures a few common types (as seen in livereload_tests.py::test_mime_types).\n # Other uncommon types will not be accepted.\n if path.endswith((\".js\", \".JS\")):\n return \"application/javascript\"\n if path.endswith(\".gz\"):\n return \"application/gzip\"\n\n guess, _ = mimetypes.guess_type(path)\n if guess:\n return guess\n return \"application/octet-stream\"\n\n\nclass _Handler(wsgiref.simple_server.WSGIRequestHandler):\n def log_request(self, code=\"-\", size=\"-\"):\n level = logging.DEBUG if str(code) == \"200\" else logging.WARNING\n log.log(level, f'\"{self.requestline}\" code {code}')\n\n def log_message(self, format, *args):\n log.debug(format, *args)\n\n\ndef _timestamp():\n return round(time.monotonic() * 1000)\n", "path": "mkdocs/livereload/__init__.py"}]}
| 3,498 | 256 |
gh_patches_debug_21290
|
rasdani/github-patches
|
git_diff
|
rucio__rucio-953
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
setup_clients.py classifiers needs to be a list, not tuples
Motivation
----------
Classifiers were changed to tuple, which does not work, needs to be a list.
</issue>
<code>
[start of setup_rucio_client.py]
1 # Copyright 2014-2018 CERN for the benefit of the ATLAS collaboration.
2 #
3 # Licensed under the Apache License, Version 2.0 (the "License");
4 # you may not use this file except in compliance with the License.
5 # You may obtain a copy of the License at
6 #
7 # http://www.apache.org/licenses/LICENSE-2.0
8 #
9 # Unless required by applicable law or agreed to in writing, software
10 # distributed under the License is distributed on an "AS IS" BASIS,
11 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12 # See the License for the specific language governing permissions and
13 # limitations under the License.
14 #
15 # Authors:
16 # - Vincent Garonne <[email protected]>, 2014-2018
17 # - Martin Barisits <[email protected]>, 2017
18
19 import os
20 import re
21 import shutil
22 import subprocess
23 import sys
24
25 from distutils.command.sdist import sdist as _sdist # pylint:disable=no-name-in-module,import-error
26 from setuptools import setup
27
28 sys.path.insert(0, os.path.abspath('lib/'))
29
30 from rucio import version # noqa
31
32 if sys.version_info < (2, 5):
33 print('ERROR: Rucio requires at least Python 2.6 to run.')
34 sys.exit(1)
35 sys.path.insert(0, os.path.abspath('lib/'))
36
37
38 # Arguments to the setup script to build Basic/Lite distributions
39 COPY_ARGS = sys.argv[1:]
40 NAME = 'rucio-clients'
41 IS_RELEASE = False
42 PACKAGES = ['rucio', 'rucio.client', 'rucio.common',
43 'rucio.rse.protocols', 'rucio.rse', 'rucio.tests']
44 REQUIREMENTS_FILES = ['tools/pip-requires-client']
45 DESCRIPTION = "Rucio Client Lite Package"
46 DATA_FILES = [('etc/', ['etc/rse-accounts.cfg.template', 'etc/rucio.cfg.template', 'etc/rucio.cfg.atlas.client.template']),
47 ('tools/', ['tools/pip-requires-client', ]), ]
48
49 SCRIPTS = ['bin/rucio', 'bin/rucio-admin']
50 if os.path.exists('build/'):
51 shutil.rmtree('build/')
52 if os.path.exists('lib/rucio_clients.egg-info/'):
53 shutil.rmtree('lib/rucio_clients.egg-info/')
54 if os.path.exists('lib/rucio.egg-info/'):
55 shutil.rmtree('lib/rucio.egg-info/')
56
57 SSH_EXTRAS = ['paramiko==1.18.4']
58 KERBEROS_EXTRAS = ['kerberos>=1.2.5', 'pykerberos>=1.1.14', 'requests-kerberos>=0.11.0']
59 SWIFT_EXTRAS = ['python-swiftclient>=3.5.0', ]
60 EXTRAS_REQUIRES = dict(ssh=SSH_EXTRAS,
61 kerberos=KERBEROS_EXTRAS,
62 swift=SWIFT_EXTRAS)
63
64 if '--release' in COPY_ARGS:
65 IS_RELEASE = True
66 COPY_ARGS.remove('--release')
67
68
69 # If Sphinx is installed on the box running setup.py,
70 # enable setup.py to build the documentation, otherwise,
71 # just ignore it
72 cmdclass = {}
73
74 try:
75 from sphinx.setup_command import BuildDoc
76
77 class local_BuildDoc(BuildDoc):
78 '''
79 local_BuildDoc
80 '''
81 def run(self):
82 '''
83 run
84 '''
85 for builder in ['html']: # 'man','latex'
86 self.builder = builder
87 self.finalize_options()
88 BuildDoc.run(self)
89 cmdclass['build_sphinx'] = local_BuildDoc
90 except Exception:
91 pass
92
93
94 def get_reqs_from_file(requirements_file):
95 '''
96 get_reqs_from_file
97 '''
98 if os.path.exists(requirements_file):
99 return open(requirements_file, 'r').read().split('\n')
100 return []
101
102
103 def parse_requirements(requirements_files):
104 '''
105 parse_requirements
106 '''
107 requirements = []
108 for requirements_file in requirements_files:
109 for line in get_reqs_from_file(requirements_file):
110 if re.match(r'\s*-e\s+', line):
111 requirements.append(re.sub(r'\s*-e\s+.*#egg=(.*)$', r'\1', line))
112 elif re.match(r'\s*-f\s+', line):
113 pass
114 else:
115 requirements.append(line)
116 return requirements
117
118
119 def parse_dependency_links(requirements_files):
120 '''
121 parse_dependency_links
122 '''
123 dependency_links = []
124 for requirements_file in requirements_files:
125 for line in get_reqs_from_file(requirements_file):
126 if re.match(r'(\s*#)|(\s*$)', line):
127 continue
128 if re.match(r'\s*-[ef]\s+', line):
129 dependency_links.append(re.sub(r'\s*-[ef]\s+', '', line))
130 return dependency_links
131
132
133 def write_requirements():
134 '''
135 write_requirements
136 '''
137 venv = os.environ.get('VIRTUAL_ENV', None)
138 if venv is not None:
139 req_file = open("requirements.txt", "w")
140 output = subprocess.Popen(["pip", "freeze", "-l"], stdout=subprocess.PIPE)
141 requirements = output.communicate()[0].strip()
142 req_file.write(requirements)
143 req_file.close()
144
145
146 REQUIRES = parse_requirements(requirements_files=REQUIREMENTS_FILES)
147 DEPEND_LINKS = parse_dependency_links(requirements_files=REQUIREMENTS_FILES)
148
149
150 class CustomSdist(_sdist):
151 '''
152 CustomSdist
153 '''
154 user_options = [
155 ('packaging=', None, "Some option to indicate what should be packaged")
156 ] + _sdist.user_options
157
158 def __init__(self, *args, **kwargs):
159 '''
160 __init__
161 '''
162 _sdist.__init__(self, *args, **kwargs)
163 self.packaging = "default value for this option"
164
165 def get_file_list(self):
166 '''
167 get_file_list
168 '''
169 print("Chosen packaging option: " + NAME)
170 self.distribution.data_files = DATA_FILES
171 _sdist.get_file_list(self)
172
173
174 cmdclass['sdist'] = CustomSdist
175
176 setup(
177 name=NAME,
178 version=version.version_string(),
179 packages=PACKAGES,
180 package_dir={'': 'lib'},
181 data_files=DATA_FILES,
182 script_args=COPY_ARGS,
183 cmdclass=cmdclass,
184 include_package_data=True,
185 scripts=SCRIPTS,
186 # doc=cmdclass,
187 author="Rucio",
188 author_email="[email protected]",
189 description=DESCRIPTION,
190 license="Apache License, Version 2.0",
191 url="http://rucio.cern.ch/",
192 python_requires=">=2.6, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*",
193 classifiers=(
194 'Development Status :: 5 - Production/Stable',
195 'License :: OSI Approved :: Apache Software License',
196 'Intended Audience :: Information Technology',
197 'Intended Audience :: System Administrators',
198 'Operating System :: POSIX :: Linux',
199 'Natural Language :: English',
200 'Programming Language :: Python',
201 'Programming Language :: Python :: 2.6',
202 'Programming Language :: Python :: 2.7',
203 'Programming Language :: Python :: 3',
204 'Programming Language :: Python :: 3.4',
205 'Programming Language :: Python :: 3.5',
206 'Programming Language :: Python :: 3.6',
207 'Programming Language :: Python :: Implementation :: CPython',
208 'Programming Language :: Python :: Implementation :: PyPy',
209 'Environment :: No Input/Output (Daemon)'
210 ),
211 install_requires=REQUIRES,
212 extras_require=EXTRAS_REQUIRES,
213 dependency_links=DEPEND_LINKS,
214 )
215
[end of setup_rucio_client.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/setup_rucio_client.py b/setup_rucio_client.py
--- a/setup_rucio_client.py
+++ b/setup_rucio_client.py
@@ -190,7 +190,7 @@
license="Apache License, Version 2.0",
url="http://rucio.cern.ch/",
python_requires=">=2.6, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*",
- classifiers=(
+ classifiers=[
'Development Status :: 5 - Production/Stable',
'License :: OSI Approved :: Apache Software License',
'Intended Audience :: Information Technology',
@@ -207,7 +207,7 @@
'Programming Language :: Python :: Implementation :: CPython',
'Programming Language :: Python :: Implementation :: PyPy',
'Environment :: No Input/Output (Daemon)'
- ),
+ ],
install_requires=REQUIRES,
extras_require=EXTRAS_REQUIRES,
dependency_links=DEPEND_LINKS,
|
{"golden_diff": "diff --git a/setup_rucio_client.py b/setup_rucio_client.py\n--- a/setup_rucio_client.py\n+++ b/setup_rucio_client.py\n@@ -190,7 +190,7 @@\n license=\"Apache License, Version 2.0\",\n url=\"http://rucio.cern.ch/\",\n python_requires=\">=2.6, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*\",\n- classifiers=(\n+ classifiers=[\n 'Development Status :: 5 - Production/Stable',\n 'License :: OSI Approved :: Apache Software License',\n 'Intended Audience :: Information Technology',\n@@ -207,7 +207,7 @@\n 'Programming Language :: Python :: Implementation :: CPython',\n 'Programming Language :: Python :: Implementation :: PyPy',\n 'Environment :: No Input/Output (Daemon)'\n- ),\n+ ],\n install_requires=REQUIRES,\n extras_require=EXTRAS_REQUIRES,\n dependency_links=DEPEND_LINKS,\n", "issue": "setup_clients.py classifiers needs to be a list, not tuples\nMotivation\r\n----------\r\nClassifiers were changed to tuple, which does not work, needs to be a list.\n", "before_files": [{"content": "# Copyright 2014-2018 CERN for the benefit of the ATLAS collaboration.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n#\n# Authors:\n# - Vincent Garonne <[email protected]>, 2014-2018\n# - Martin Barisits <[email protected]>, 2017\n\nimport os\nimport re\nimport shutil\nimport subprocess\nimport sys\n\nfrom distutils.command.sdist import sdist as _sdist # pylint:disable=no-name-in-module,import-error\nfrom setuptools import setup\n\nsys.path.insert(0, os.path.abspath('lib/'))\n\nfrom rucio import version # noqa\n\nif sys.version_info < (2, 5):\n print('ERROR: Rucio requires at least Python 2.6 to run.')\n sys.exit(1)\nsys.path.insert(0, os.path.abspath('lib/'))\n\n\n# Arguments to the setup script to build Basic/Lite distributions\nCOPY_ARGS = sys.argv[1:]\nNAME = 'rucio-clients'\nIS_RELEASE = False\nPACKAGES = ['rucio', 'rucio.client', 'rucio.common',\n 'rucio.rse.protocols', 'rucio.rse', 'rucio.tests']\nREQUIREMENTS_FILES = ['tools/pip-requires-client']\nDESCRIPTION = \"Rucio Client Lite Package\"\nDATA_FILES = [('etc/', ['etc/rse-accounts.cfg.template', 'etc/rucio.cfg.template', 'etc/rucio.cfg.atlas.client.template']),\n ('tools/', ['tools/pip-requires-client', ]), ]\n\nSCRIPTS = ['bin/rucio', 'bin/rucio-admin']\nif os.path.exists('build/'):\n shutil.rmtree('build/')\nif os.path.exists('lib/rucio_clients.egg-info/'):\n shutil.rmtree('lib/rucio_clients.egg-info/')\nif os.path.exists('lib/rucio.egg-info/'):\n shutil.rmtree('lib/rucio.egg-info/')\n\nSSH_EXTRAS = ['paramiko==1.18.4']\nKERBEROS_EXTRAS = ['kerberos>=1.2.5', 'pykerberos>=1.1.14', 'requests-kerberos>=0.11.0']\nSWIFT_EXTRAS = ['python-swiftclient>=3.5.0', ]\nEXTRAS_REQUIRES = dict(ssh=SSH_EXTRAS,\n kerberos=KERBEROS_EXTRAS,\n swift=SWIFT_EXTRAS)\n\nif '--release' in COPY_ARGS:\n IS_RELEASE = True\n COPY_ARGS.remove('--release')\n\n\n# If Sphinx is installed on the box running setup.py,\n# enable setup.py to build the documentation, otherwise,\n# just ignore it\ncmdclass = {}\n\ntry:\n from sphinx.setup_command import BuildDoc\n\n class local_BuildDoc(BuildDoc):\n '''\n local_BuildDoc\n '''\n def run(self):\n '''\n run\n '''\n for builder in ['html']: # 'man','latex'\n self.builder = builder\n self.finalize_options()\n BuildDoc.run(self)\n cmdclass['build_sphinx'] = local_BuildDoc\nexcept Exception:\n pass\n\n\ndef get_reqs_from_file(requirements_file):\n '''\n get_reqs_from_file\n '''\n if os.path.exists(requirements_file):\n return open(requirements_file, 'r').read().split('\\n')\n return []\n\n\ndef parse_requirements(requirements_files):\n '''\n parse_requirements\n '''\n requirements = []\n for requirements_file in requirements_files:\n for line in get_reqs_from_file(requirements_file):\n if re.match(r'\\s*-e\\s+', line):\n requirements.append(re.sub(r'\\s*-e\\s+.*#egg=(.*)$', r'\\1', line))\n elif re.match(r'\\s*-f\\s+', line):\n pass\n else:\n requirements.append(line)\n return requirements\n\n\ndef parse_dependency_links(requirements_files):\n '''\n parse_dependency_links\n '''\n dependency_links = []\n for requirements_file in requirements_files:\n for line in get_reqs_from_file(requirements_file):\n if re.match(r'(\\s*#)|(\\s*$)', line):\n continue\n if re.match(r'\\s*-[ef]\\s+', line):\n dependency_links.append(re.sub(r'\\s*-[ef]\\s+', '', line))\n return dependency_links\n\n\ndef write_requirements():\n '''\n write_requirements\n '''\n venv = os.environ.get('VIRTUAL_ENV', None)\n if venv is not None:\n req_file = open(\"requirements.txt\", \"w\")\n output = subprocess.Popen([\"pip\", \"freeze\", \"-l\"], stdout=subprocess.PIPE)\n requirements = output.communicate()[0].strip()\n req_file.write(requirements)\n req_file.close()\n\n\nREQUIRES = parse_requirements(requirements_files=REQUIREMENTS_FILES)\nDEPEND_LINKS = parse_dependency_links(requirements_files=REQUIREMENTS_FILES)\n\n\nclass CustomSdist(_sdist):\n '''\n CustomSdist\n '''\n user_options = [\n ('packaging=', None, \"Some option to indicate what should be packaged\")\n ] + _sdist.user_options\n\n def __init__(self, *args, **kwargs):\n '''\n __init__\n '''\n _sdist.__init__(self, *args, **kwargs)\n self.packaging = \"default value for this option\"\n\n def get_file_list(self):\n '''\n get_file_list\n '''\n print(\"Chosen packaging option: \" + NAME)\n self.distribution.data_files = DATA_FILES\n _sdist.get_file_list(self)\n\n\ncmdclass['sdist'] = CustomSdist\n\nsetup(\n name=NAME,\n version=version.version_string(),\n packages=PACKAGES,\n package_dir={'': 'lib'},\n data_files=DATA_FILES,\n script_args=COPY_ARGS,\n cmdclass=cmdclass,\n include_package_data=True,\n scripts=SCRIPTS,\n # doc=cmdclass,\n author=\"Rucio\",\n author_email=\"[email protected]\",\n description=DESCRIPTION,\n license=\"Apache License, Version 2.0\",\n url=\"http://rucio.cern.ch/\",\n python_requires=\">=2.6, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*\",\n classifiers=(\n 'Development Status :: 5 - Production/Stable',\n 'License :: OSI Approved :: Apache Software License',\n 'Intended Audience :: Information Technology',\n 'Intended Audience :: System Administrators',\n 'Operating System :: POSIX :: Linux',\n 'Natural Language :: English',\n 'Programming Language :: Python',\n 'Programming Language :: Python :: 2.6',\n 'Programming Language :: Python :: 2.7',\n 'Programming Language :: Python :: 3',\n 'Programming Language :: Python :: 3.4',\n 'Programming Language :: Python :: 3.5',\n 'Programming Language :: Python :: 3.6',\n 'Programming Language :: Python :: Implementation :: CPython',\n 'Programming Language :: Python :: Implementation :: PyPy',\n 'Environment :: No Input/Output (Daemon)'\n ),\n install_requires=REQUIRES,\n extras_require=EXTRAS_REQUIRES,\n dependency_links=DEPEND_LINKS,\n)\n", "path": "setup_rucio_client.py"}]}
| 2,804 | 223 |
gh_patches_debug_41092
|
rasdani/github-patches
|
git_diff
|
jupyterhub__jupyterhub-4249
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
migrate remaining rst docs to myst
We still have several doc pages in .rst format, while most are .md.
@consideRatio proposed migrating to myst, which I agree we should do, and it may be good to get done before the docs-reorg internship gets underway in a few weeks.
The challenge is the following open PRs (mostly from the outreachy application period last month) touch these rst files:
- [x] #4209:
- docs/source/contributing/setup.rst
- [x] #4186:
- docs/source/getting-started/security-basics.rst
- [x] #4127:
- docs/source/contributing/docs.rst
- [x] #4119:
- docs/source/contributing/tests.rst
- [x] #4094:
- docs/source/admin/upgrading.rst
- docs/source/getting-started/security-basics.rst
- [x] #4084:
- docs/source/index.rst
- [ ] #4066:
- docs/source/contributing/index.rst
- docs/source/contributing/layout.rst
- [x] #4008:
- docs/source/index-admin.rst
- [x] #3053:
- docs/source/reference/index.rst
- [ ] #2726:
- docs/source/getting-started/index.rst
- docs/source/index.rst
<details>
<summary>PR list generated with</summary>
```python
from github import Github as GitHub
gh = GitHub()
repo = gh.get_repo("jupyterhub/jupyterhub")
for pull in repo.get_pulls(state="open"):
pull_files = [f.filename for f in list(pull.get_files())]
if any(f.endswith(".rst") for f in pull_files):
print(f"- [ ] #{pull.number}:")
for f in pull_files:
if f.endswith(".rst"):
print(f" - {f}")
continue
```
</details>
We should merge/close as many of these as we can to avoid conflicts (some that just touch an index are no big deal to resolve conflicts).
</issue>
<code>
[start of docs/generate-metrics.py]
1 import os
2
3 from pytablewriter import RstSimpleTableWriter
4 from pytablewriter.style import Style
5
6 import jupyterhub.metrics
7
8 HERE = os.path.abspath(os.path.dirname(__file__))
9
10
11 class Generator:
12 @classmethod
13 def create_writer(cls, table_name, headers, values):
14 writer = RstSimpleTableWriter()
15 writer.table_name = table_name
16 writer.headers = headers
17 writer.value_matrix = values
18 writer.margin = 1
19 [writer.set_style(header, Style(align="center")) for header in headers]
20 return writer
21
22 def _parse_metrics(self):
23 table_rows = []
24 for name in dir(jupyterhub.metrics):
25 obj = getattr(jupyterhub.metrics, name)
26 if obj.__class__.__module__.startswith('prometheus_client.'):
27 for metric in obj.describe():
28 table_rows.append([metric.type, metric.name, metric.documentation])
29 return table_rows
30
31 def prometheus_metrics(self):
32 generated_directory = f"{HERE}/source/reference"
33 if not os.path.exists(generated_directory):
34 os.makedirs(generated_directory)
35
36 filename = f"{generated_directory}/metrics.rst"
37 table_name = ""
38 headers = ["Type", "Name", "Description"]
39 values = self._parse_metrics()
40 writer = self.create_writer(table_name, headers, values)
41
42 title = "List of Prometheus Metrics"
43 underline = "============================"
44 content = f"{title}\n{underline}\n{writer.dumps()}"
45 with open(filename, 'w') as f:
46 f.write(content)
47 print(f"Generated {filename}.")
48
49
50 def main():
51 doc_generator = Generator()
52 doc_generator.prometheus_metrics()
53
54
55 if __name__ == "__main__":
56 main()
57
[end of docs/generate-metrics.py]
[start of docs/source/conf.py]
1 # Configuration file for Sphinx to build our documentation to HTML.
2 #
3 # Configuration reference: https://www.sphinx-doc.org/en/master/usage/configuration.html
4 #
5 import contextlib
6 import datetime
7 import io
8 import os
9 import subprocess
10
11 from docutils import nodes
12 from sphinx.directives.other import SphinxDirective
13
14 import jupyterhub
15 from jupyterhub.app import JupyterHub
16
17 # -- Project information -----------------------------------------------------
18 # ref: https://www.sphinx-doc.org/en/master/usage/configuration.html#project-information
19 #
20 project = "JupyterHub"
21 author = "Project Jupyter Contributors"
22 copyright = f"{datetime.date.today().year}, {author}"
23 version = "%i.%i" % jupyterhub.version_info[:2]
24 release = jupyterhub.__version__
25
26
27 # -- General Sphinx configuration --------------------------------------------
28 # ref: https://www.sphinx-doc.org/en/master/usage/configuration.html#general-configuration
29 #
30 extensions = [
31 "sphinx.ext.autodoc",
32 "sphinx.ext.intersphinx",
33 "sphinx.ext.napoleon",
34 "autodoc_traits",
35 "sphinx_copybutton",
36 "sphinx-jsonschema",
37 "sphinxext.opengraph",
38 "sphinxext.rediraffe",
39 "myst_parser",
40 ]
41 root_doc = "index"
42 source_suffix = [".md", ".rst"]
43 # default_role let's use use `foo` instead of ``foo`` in rST
44 default_role = "literal"
45
46
47 # -- MyST configuration ------------------------------------------------------
48 # ref: https://myst-parser.readthedocs.io/en/latest/configuration.html
49 #
50 myst_heading_anchors = 2
51 myst_enable_extensions = [
52 "colon_fence",
53 "deflist",
54 ]
55
56
57 # -- Custom directives to generate documentation -----------------------------
58 # ref: https://myst-parser.readthedocs.io/en/latest/syntax/roles-and-directives.html
59 #
60 # We define custom directives to help us generate documentation using Python on
61 # demand when referenced from our documentation files.
62 #
63
64 # Create a temp instance of JupyterHub for use by two separate directive classes
65 # to get the output from using the "--generate-config" and "--help-all" CLI
66 # flags respectively.
67 #
68 jupyterhub_app = JupyterHub()
69
70
71 class ConfigDirective(SphinxDirective):
72 """Generate the configuration file output for use in the documentation."""
73
74 has_content = False
75 required_arguments = 0
76 optional_arguments = 0
77 final_argument_whitespace = False
78 option_spec = {}
79
80 def run(self):
81 # The generated configuration file for this version
82 generated_config = jupyterhub_app.generate_config_file()
83 # post-process output
84 home_dir = os.environ["HOME"]
85 generated_config = generated_config.replace(home_dir, "$HOME", 1)
86 par = nodes.literal_block(text=generated_config)
87 return [par]
88
89
90 class HelpAllDirective(SphinxDirective):
91 """Print the output of jupyterhub help --all for use in the documentation."""
92
93 has_content = False
94 required_arguments = 0
95 optional_arguments = 0
96 final_argument_whitespace = False
97 option_spec = {}
98
99 def run(self):
100 # The output of the help command for this version
101 buffer = io.StringIO()
102 with contextlib.redirect_stdout(buffer):
103 jupyterhub_app.print_help("--help-all")
104 all_help = buffer.getvalue()
105 # post-process output
106 home_dir = os.environ["HOME"]
107 all_help = all_help.replace(home_dir, "$HOME", 1)
108 par = nodes.literal_block(text=all_help)
109 return [par]
110
111
112 def setup(app):
113 app.add_css_file("custom.css")
114 app.add_directive("jupyterhub-generate-config", ConfigDirective)
115 app.add_directive("jupyterhub-help-all", HelpAllDirective)
116
117
118 # -- Read The Docs -----------------------------------------------------------
119 #
120 # Since RTD runs sphinx-build directly without running "make html", we run the
121 # pre-requisite steps for "make html" from here if needed.
122 #
123 if os.environ.get("READTHEDOCS"):
124 docs = os.path.dirname(os.path.dirname(__file__))
125 subprocess.check_call(["make", "metrics", "scopes"], cwd=docs)
126
127
128 # -- Spell checking ----------------------------------------------------------
129 # ref: https://sphinxcontrib-spelling.readthedocs.io/en/latest/customize.html#configuration-options
130 #
131 # The "sphinxcontrib.spelling" extension is optionally enabled if its available.
132 #
133 try:
134 import sphinxcontrib.spelling # noqa
135 except ImportError:
136 pass
137 else:
138 extensions.append("sphinxcontrib.spelling")
139 spelling_word_list_filename = "spelling_wordlist.txt"
140
141
142 # -- Options for HTML output -------------------------------------------------
143 # ref: https://www.sphinx-doc.org/en/master/usage/configuration.html#options-for-html-output
144 #
145 html_logo = "_static/images/logo/logo.png"
146 html_favicon = "_static/images/logo/favicon.ico"
147 html_static_path = ["_static"]
148
149 html_theme = "pydata_sphinx_theme"
150 html_theme_options = {
151 "icon_links": [
152 {
153 "name": "GitHub",
154 "url": "https://github.com/jupyterhub/jupyterhub",
155 "icon": "fab fa-github-square",
156 },
157 {
158 "name": "Discourse",
159 "url": "https://discourse.jupyter.org/c/jupyterhub/10",
160 "icon": "fab fa-discourse",
161 },
162 ],
163 "use_edit_page_button": True,
164 "navbar_align": "left",
165 }
166 html_context = {
167 "github_user": "jupyterhub",
168 "github_repo": "jupyterhub",
169 "github_version": "main",
170 "doc_path": "docs/source",
171 }
172
173
174 # -- Options for linkcheck builder -------------------------------------------
175 # ref: https://www.sphinx-doc.org/en/master/usage/configuration.html#options-for-the-linkcheck-builder
176 #
177 linkcheck_ignore = [
178 r"(.*)github\.com(.*)#", # javascript based anchors
179 r"(.*)/#%21(.*)/(.*)", # /#!forum/jupyter - encoded anchor edge case
180 r"https://github.com/[^/]*$", # too many github usernames / searches in changelog
181 "https://github.com/jupyterhub/jupyterhub/pull/", # too many PRs in changelog
182 "https://github.com/jupyterhub/jupyterhub/compare/", # too many comparisons in changelog
183 ]
184 linkcheck_anchors_ignore = [
185 "/#!",
186 "/#%21",
187 ]
188
189
190 # -- Intersphinx -------------------------------------------------------------
191 # ref: https://www.sphinx-doc.org/en/master/usage/extensions/intersphinx.html#configuration
192 #
193 intersphinx_mapping = {
194 "python": ("https://docs.python.org/3/", None),
195 "tornado": ("https://www.tornadoweb.org/en/stable/", None),
196 }
197 # -- Options for the opengraph extension -------------------------------------
198 # ref: https://github.com/wpilibsuite/sphinxext-opengraph#options
199 #
200 # ogp_site_url is set automatically by RTD
201 ogp_image = "_static/logo.png"
202 ogp_use_first_image = True
203
204
205 # -- Options for the rediraffe extension -------------------------------------
206 # ref: https://github.com/wpilibsuite/sphinxext-rediraffe#readme
207 #
208 # This extensions help us relocated content without breaking links. If a
209 # document is moved internally, a redirect like should be configured below to
210 # help us not break links.
211 #
212 rediraffe_branch = "main"
213 rediraffe_redirects = {
214 # "old-file": "new-folder/new-file-name",
215 }
216
[end of docs/source/conf.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/docs/generate-metrics.py b/docs/generate-metrics.py
--- a/docs/generate-metrics.py
+++ b/docs/generate-metrics.py
@@ -1,7 +1,6 @@
import os
-from pytablewriter import RstSimpleTableWriter
-from pytablewriter.style import Style
+from pytablewriter import MarkdownTableWriter
import jupyterhub.metrics
@@ -11,12 +10,11 @@
class Generator:
@classmethod
def create_writer(cls, table_name, headers, values):
- writer = RstSimpleTableWriter()
+ writer = MarkdownTableWriter()
writer.table_name = table_name
writer.headers = headers
writer.value_matrix = values
writer.margin = 1
- [writer.set_style(header, Style(align="center")) for header in headers]
return writer
def _parse_metrics(self):
@@ -33,18 +31,17 @@
if not os.path.exists(generated_directory):
os.makedirs(generated_directory)
- filename = f"{generated_directory}/metrics.rst"
+ filename = f"{generated_directory}/metrics.md"
table_name = ""
headers = ["Type", "Name", "Description"]
values = self._parse_metrics()
writer = self.create_writer(table_name, headers, values)
- title = "List of Prometheus Metrics"
- underline = "============================"
- content = f"{title}\n{underline}\n{writer.dumps()}"
with open(filename, 'w') as f:
- f.write(content)
- print(f"Generated {filename}.")
+ f.write("# List of Prometheus Metrics\n\n")
+ f.write(writer.dumps())
+ f.write("\n")
+ print(f"Generated {filename}")
def main():
diff --git a/docs/source/conf.py b/docs/source/conf.py
--- a/docs/source/conf.py
+++ b/docs/source/conf.py
@@ -20,8 +20,6 @@
project = "JupyterHub"
author = "Project Jupyter Contributors"
copyright = f"{datetime.date.today().year}, {author}"
-version = "%i.%i" % jupyterhub.version_info[:2]
-release = jupyterhub.__version__
# -- General Sphinx configuration --------------------------------------------
@@ -39,7 +37,7 @@
"myst_parser",
]
root_doc = "index"
-source_suffix = [".md", ".rst"]
+source_suffix = [".md"]
# default_role let's use use `foo` instead of ``foo`` in rST
default_role = "literal"
@@ -48,11 +46,21 @@
# ref: https://myst-parser.readthedocs.io/en/latest/configuration.html
#
myst_heading_anchors = 2
+
myst_enable_extensions = [
+ # available extensions: https://myst-parser.readthedocs.io/en/latest/syntax/optional.html
"colon_fence",
"deflist",
+ "fieldlist",
+ "substitution",
]
+myst_substitutions = {
+ # date example: Dev 07, 2022
+ "date": datetime.date.today().strftime("%b %d, %Y").title(),
+ "version": jupyterhub.__version__,
+}
+
# -- Custom directives to generate documentation -----------------------------
# ref: https://myst-parser.readthedocs.io/en/latest/syntax/roles-and-directives.html
|
{"golden_diff": "diff --git a/docs/generate-metrics.py b/docs/generate-metrics.py\n--- a/docs/generate-metrics.py\n+++ b/docs/generate-metrics.py\n@@ -1,7 +1,6 @@\n import os\n \n-from pytablewriter import RstSimpleTableWriter\n-from pytablewriter.style import Style\n+from pytablewriter import MarkdownTableWriter\n \n import jupyterhub.metrics\n \n@@ -11,12 +10,11 @@\n class Generator:\n @classmethod\n def create_writer(cls, table_name, headers, values):\n- writer = RstSimpleTableWriter()\n+ writer = MarkdownTableWriter()\n writer.table_name = table_name\n writer.headers = headers\n writer.value_matrix = values\n writer.margin = 1\n- [writer.set_style(header, Style(align=\"center\")) for header in headers]\n return writer\n \n def _parse_metrics(self):\n@@ -33,18 +31,17 @@\n if not os.path.exists(generated_directory):\n os.makedirs(generated_directory)\n \n- filename = f\"{generated_directory}/metrics.rst\"\n+ filename = f\"{generated_directory}/metrics.md\"\n table_name = \"\"\n headers = [\"Type\", \"Name\", \"Description\"]\n values = self._parse_metrics()\n writer = self.create_writer(table_name, headers, values)\n \n- title = \"List of Prometheus Metrics\"\n- underline = \"============================\"\n- content = f\"{title}\\n{underline}\\n{writer.dumps()}\"\n with open(filename, 'w') as f:\n- f.write(content)\n- print(f\"Generated {filename}.\")\n+ f.write(\"# List of Prometheus Metrics\\n\\n\")\n+ f.write(writer.dumps())\n+ f.write(\"\\n\")\n+ print(f\"Generated {filename}\")\n \n \n def main():\ndiff --git a/docs/source/conf.py b/docs/source/conf.py\n--- a/docs/source/conf.py\n+++ b/docs/source/conf.py\n@@ -20,8 +20,6 @@\n project = \"JupyterHub\"\n author = \"Project Jupyter Contributors\"\n copyright = f\"{datetime.date.today().year}, {author}\"\n-version = \"%i.%i\" % jupyterhub.version_info[:2]\n-release = jupyterhub.__version__\n \n \n # -- General Sphinx configuration --------------------------------------------\n@@ -39,7 +37,7 @@\n \"myst_parser\",\n ]\n root_doc = \"index\"\n-source_suffix = [\".md\", \".rst\"]\n+source_suffix = [\".md\"]\n # default_role let's use use `foo` instead of ``foo`` in rST\n default_role = \"literal\"\n \n@@ -48,11 +46,21 @@\n # ref: https://myst-parser.readthedocs.io/en/latest/configuration.html\n #\n myst_heading_anchors = 2\n+\n myst_enable_extensions = [\n+ # available extensions: https://myst-parser.readthedocs.io/en/latest/syntax/optional.html\n \"colon_fence\",\n \"deflist\",\n+ \"fieldlist\",\n+ \"substitution\",\n ]\n \n+myst_substitutions = {\n+ # date example: Dev 07, 2022\n+ \"date\": datetime.date.today().strftime(\"%b %d, %Y\").title(),\n+ \"version\": jupyterhub.__version__,\n+}\n+\n \n # -- Custom directives to generate documentation -----------------------------\n # ref: https://myst-parser.readthedocs.io/en/latest/syntax/roles-and-directives.html\n", "issue": "migrate remaining rst docs to myst\nWe still have several doc pages in .rst format, while most are .md.\r\n\r\n@consideRatio proposed migrating to myst, which I agree we should do, and it may be good to get done before the docs-reorg internship gets underway in a few weeks.\r\n\r\nThe challenge is the following open PRs (mostly from the outreachy application period last month) touch these rst files:\r\n\r\n- [x] #4209:\r\n - docs/source/contributing/setup.rst\r\n- [x] #4186:\r\n - docs/source/getting-started/security-basics.rst\r\n- [x] #4127:\r\n - docs/source/contributing/docs.rst\r\n- [x] #4119:\r\n - docs/source/contributing/tests.rst\r\n- [x] #4094:\r\n - docs/source/admin/upgrading.rst\r\n - docs/source/getting-started/security-basics.rst\r\n- [x] #4084:\r\n - docs/source/index.rst\r\n- [ ] #4066:\r\n - docs/source/contributing/index.rst\r\n - docs/source/contributing/layout.rst\r\n- [x] #4008:\r\n - docs/source/index-admin.rst\r\n- [x] #3053:\r\n - docs/source/reference/index.rst\r\n- [ ] #2726:\r\n - docs/source/getting-started/index.rst\r\n - docs/source/index.rst\r\n\r\n<details>\r\n<summary>PR list generated with</summary>\r\n\r\n```python\r\nfrom github import Github as GitHub\r\n\r\ngh = GitHub()\r\n\r\nrepo = gh.get_repo(\"jupyterhub/jupyterhub\")\r\nfor pull in repo.get_pulls(state=\"open\"):\r\n pull_files = [f.filename for f in list(pull.get_files())]\r\n if any(f.endswith(\".rst\") for f in pull_files):\r\n print(f\"- [ ] #{pull.number}:\")\r\n for f in pull_files:\r\n if f.endswith(\".rst\"):\r\n print(f\" - {f}\")\r\n continue\r\n```\r\n\r\n</details>\r\n\r\nWe should merge/close as many of these as we can to avoid conflicts (some that just touch an index are no big deal to resolve conflicts).\r\n\r\n\r\n\n", "before_files": [{"content": "import os\n\nfrom pytablewriter import RstSimpleTableWriter\nfrom pytablewriter.style import Style\n\nimport jupyterhub.metrics\n\nHERE = os.path.abspath(os.path.dirname(__file__))\n\n\nclass Generator:\n @classmethod\n def create_writer(cls, table_name, headers, values):\n writer = RstSimpleTableWriter()\n writer.table_name = table_name\n writer.headers = headers\n writer.value_matrix = values\n writer.margin = 1\n [writer.set_style(header, Style(align=\"center\")) for header in headers]\n return writer\n\n def _parse_metrics(self):\n table_rows = []\n for name in dir(jupyterhub.metrics):\n obj = getattr(jupyterhub.metrics, name)\n if obj.__class__.__module__.startswith('prometheus_client.'):\n for metric in obj.describe():\n table_rows.append([metric.type, metric.name, metric.documentation])\n return table_rows\n\n def prometheus_metrics(self):\n generated_directory = f\"{HERE}/source/reference\"\n if not os.path.exists(generated_directory):\n os.makedirs(generated_directory)\n\n filename = f\"{generated_directory}/metrics.rst\"\n table_name = \"\"\n headers = [\"Type\", \"Name\", \"Description\"]\n values = self._parse_metrics()\n writer = self.create_writer(table_name, headers, values)\n\n title = \"List of Prometheus Metrics\"\n underline = \"============================\"\n content = f\"{title}\\n{underline}\\n{writer.dumps()}\"\n with open(filename, 'w') as f:\n f.write(content)\n print(f\"Generated {filename}.\")\n\n\ndef main():\n doc_generator = Generator()\n doc_generator.prometheus_metrics()\n\n\nif __name__ == \"__main__\":\n main()\n", "path": "docs/generate-metrics.py"}, {"content": "# Configuration file for Sphinx to build our documentation to HTML.\n#\n# Configuration reference: https://www.sphinx-doc.org/en/master/usage/configuration.html\n#\nimport contextlib\nimport datetime\nimport io\nimport os\nimport subprocess\n\nfrom docutils import nodes\nfrom sphinx.directives.other import SphinxDirective\n\nimport jupyterhub\nfrom jupyterhub.app import JupyterHub\n\n# -- Project information -----------------------------------------------------\n# ref: https://www.sphinx-doc.org/en/master/usage/configuration.html#project-information\n#\nproject = \"JupyterHub\"\nauthor = \"Project Jupyter Contributors\"\ncopyright = f\"{datetime.date.today().year}, {author}\"\nversion = \"%i.%i\" % jupyterhub.version_info[:2]\nrelease = jupyterhub.__version__\n\n\n# -- General Sphinx configuration --------------------------------------------\n# ref: https://www.sphinx-doc.org/en/master/usage/configuration.html#general-configuration\n#\nextensions = [\n \"sphinx.ext.autodoc\",\n \"sphinx.ext.intersphinx\",\n \"sphinx.ext.napoleon\",\n \"autodoc_traits\",\n \"sphinx_copybutton\",\n \"sphinx-jsonschema\",\n \"sphinxext.opengraph\",\n \"sphinxext.rediraffe\",\n \"myst_parser\",\n]\nroot_doc = \"index\"\nsource_suffix = [\".md\", \".rst\"]\n# default_role let's use use `foo` instead of ``foo`` in rST\ndefault_role = \"literal\"\n\n\n# -- MyST configuration ------------------------------------------------------\n# ref: https://myst-parser.readthedocs.io/en/latest/configuration.html\n#\nmyst_heading_anchors = 2\nmyst_enable_extensions = [\n \"colon_fence\",\n \"deflist\",\n]\n\n\n# -- Custom directives to generate documentation -----------------------------\n# ref: https://myst-parser.readthedocs.io/en/latest/syntax/roles-and-directives.html\n#\n# We define custom directives to help us generate documentation using Python on\n# demand when referenced from our documentation files.\n#\n\n# Create a temp instance of JupyterHub for use by two separate directive classes\n# to get the output from using the \"--generate-config\" and \"--help-all\" CLI\n# flags respectively.\n#\njupyterhub_app = JupyterHub()\n\n\nclass ConfigDirective(SphinxDirective):\n \"\"\"Generate the configuration file output for use in the documentation.\"\"\"\n\n has_content = False\n required_arguments = 0\n optional_arguments = 0\n final_argument_whitespace = False\n option_spec = {}\n\n def run(self):\n # The generated configuration file for this version\n generated_config = jupyterhub_app.generate_config_file()\n # post-process output\n home_dir = os.environ[\"HOME\"]\n generated_config = generated_config.replace(home_dir, \"$HOME\", 1)\n par = nodes.literal_block(text=generated_config)\n return [par]\n\n\nclass HelpAllDirective(SphinxDirective):\n \"\"\"Print the output of jupyterhub help --all for use in the documentation.\"\"\"\n\n has_content = False\n required_arguments = 0\n optional_arguments = 0\n final_argument_whitespace = False\n option_spec = {}\n\n def run(self):\n # The output of the help command for this version\n buffer = io.StringIO()\n with contextlib.redirect_stdout(buffer):\n jupyterhub_app.print_help(\"--help-all\")\n all_help = buffer.getvalue()\n # post-process output\n home_dir = os.environ[\"HOME\"]\n all_help = all_help.replace(home_dir, \"$HOME\", 1)\n par = nodes.literal_block(text=all_help)\n return [par]\n\n\ndef setup(app):\n app.add_css_file(\"custom.css\")\n app.add_directive(\"jupyterhub-generate-config\", ConfigDirective)\n app.add_directive(\"jupyterhub-help-all\", HelpAllDirective)\n\n\n# -- Read The Docs -----------------------------------------------------------\n#\n# Since RTD runs sphinx-build directly without running \"make html\", we run the\n# pre-requisite steps for \"make html\" from here if needed.\n#\nif os.environ.get(\"READTHEDOCS\"):\n docs = os.path.dirname(os.path.dirname(__file__))\n subprocess.check_call([\"make\", \"metrics\", \"scopes\"], cwd=docs)\n\n\n# -- Spell checking ----------------------------------------------------------\n# ref: https://sphinxcontrib-spelling.readthedocs.io/en/latest/customize.html#configuration-options\n#\n# The \"sphinxcontrib.spelling\" extension is optionally enabled if its available.\n#\ntry:\n import sphinxcontrib.spelling # noqa\nexcept ImportError:\n pass\nelse:\n extensions.append(\"sphinxcontrib.spelling\")\nspelling_word_list_filename = \"spelling_wordlist.txt\"\n\n\n# -- Options for HTML output -------------------------------------------------\n# ref: https://www.sphinx-doc.org/en/master/usage/configuration.html#options-for-html-output\n#\nhtml_logo = \"_static/images/logo/logo.png\"\nhtml_favicon = \"_static/images/logo/favicon.ico\"\nhtml_static_path = [\"_static\"]\n\nhtml_theme = \"pydata_sphinx_theme\"\nhtml_theme_options = {\n \"icon_links\": [\n {\n \"name\": \"GitHub\",\n \"url\": \"https://github.com/jupyterhub/jupyterhub\",\n \"icon\": \"fab fa-github-square\",\n },\n {\n \"name\": \"Discourse\",\n \"url\": \"https://discourse.jupyter.org/c/jupyterhub/10\",\n \"icon\": \"fab fa-discourse\",\n },\n ],\n \"use_edit_page_button\": True,\n \"navbar_align\": \"left\",\n}\nhtml_context = {\n \"github_user\": \"jupyterhub\",\n \"github_repo\": \"jupyterhub\",\n \"github_version\": \"main\",\n \"doc_path\": \"docs/source\",\n}\n\n\n# -- Options for linkcheck builder -------------------------------------------\n# ref: https://www.sphinx-doc.org/en/master/usage/configuration.html#options-for-the-linkcheck-builder\n#\nlinkcheck_ignore = [\n r\"(.*)github\\.com(.*)#\", # javascript based anchors\n r\"(.*)/#%21(.*)/(.*)\", # /#!forum/jupyter - encoded anchor edge case\n r\"https://github.com/[^/]*$\", # too many github usernames / searches in changelog\n \"https://github.com/jupyterhub/jupyterhub/pull/\", # too many PRs in changelog\n \"https://github.com/jupyterhub/jupyterhub/compare/\", # too many comparisons in changelog\n]\nlinkcheck_anchors_ignore = [\n \"/#!\",\n \"/#%21\",\n]\n\n\n# -- Intersphinx -------------------------------------------------------------\n# ref: https://www.sphinx-doc.org/en/master/usage/extensions/intersphinx.html#configuration\n#\nintersphinx_mapping = {\n \"python\": (\"https://docs.python.org/3/\", None),\n \"tornado\": (\"https://www.tornadoweb.org/en/stable/\", None),\n}\n# -- Options for the opengraph extension -------------------------------------\n# ref: https://github.com/wpilibsuite/sphinxext-opengraph#options\n#\n# ogp_site_url is set automatically by RTD\nogp_image = \"_static/logo.png\"\nogp_use_first_image = True\n\n\n# -- Options for the rediraffe extension -------------------------------------\n# ref: https://github.com/wpilibsuite/sphinxext-rediraffe#readme\n#\n# This extensions help us relocated content without breaking links. If a\n# document is moved internally, a redirect like should be configured below to\n# help us not break links.\n#\nrediraffe_branch = \"main\"\nrediraffe_redirects = {\n # \"old-file\": \"new-folder/new-file-name\",\n}\n", "path": "docs/source/conf.py"}]}
| 3,676 | 744 |
gh_patches_debug_7521
|
rasdani/github-patches
|
git_diff
|
sunpy__sunpy-780
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Database.query buggy with Time
Database.query works different using the Time attribute than VSO does.
</issue>
<code>
[start of sunpy/database/attrs.py]
1 # Author: Simon Liedtke <[email protected]>
2 #
3 # This module was developed with funding provided by
4 # the Google Summer of Code (2013).
5
6 from __future__ import absolute_import
7
8 from sqlalchemy import or_, and_, not_
9 from astropy.units import Unit, nm, equivalencies
10
11 from sunpy.time import parse_time
12 from sunpy.net.vso import attrs as vso_attrs
13 from sunpy.net.attr import AttrWalker, Attr, ValueAttr, AttrAnd, AttrOr
14 from sunpy.database.tables import DatabaseEntry, Tag as TableTag,\
15 FitsHeaderEntry as TableFitsHeaderEntry
16
17 __all__ = [
18 'Starred', 'Tag', 'Path', 'DownloadTime', 'FitsHeaderEntry', 'walker']
19
20
21 class _BooleanAttr(object):
22 def __init__(self, value, make):
23 self.value = bool(value)
24 self.make = make
25
26 def __and__(self, other):
27 if not isinstance(other, self.make):
28 return AttrAnd([self, other])
29 attr = self.make()
30 attr.value = self.value and other.value
31 return attr
32
33 def __or__(self, other):
34 if not isinstance(other, self.make):
35 return AttrOr([self, other])
36 attr = self.make()
37 attr.value = self.value or other.value
38 return attr
39
40 def __nonzero__(self):
41 return self.value
42
43 def __invert__(self):
44 attr = self.make()
45 attr.value = not self.value
46 return attr
47
48 def __eq__(self, other):
49 return isinstance(other, self.make) and self.value == other.value
50
51 def collides(self, other): # pragma: no cover
52 return False
53
54 def __repr__(self):
55 return '<{0}{1}()>'.format(
56 '~' if not self.value else '', self.__class__.__name__)
57
58
59 class Starred(_BooleanAttr, Attr):
60 def __init__(self):
61 _BooleanAttr.__init__(self, True, self.__class__)
62
63
64 class Tag(Attr):
65 def __init__(self, tagname):
66 self.tagname = tagname
67 self.inverted = False
68
69 def __invert__(self):
70 tag = self.__class__(self.tagname)
71 tag.inverted = True
72 return tag
73
74 def collides(self, other): # pragma: no cover
75 return False
76
77 def __repr__(self):
78 return '<{0}Tag({1!r})>'.format(
79 '~' if self.inverted else '', self.tagname)
80
81
82 class Path(Attr):
83 def __init__(self, value):
84 self.value = value
85 self.inverted = False
86
87 def __invert__(self):
88 path = self.__class__(self.value)
89 path.inverted = True
90 return path
91
92 def collides(self, other): # pragma: no cover
93 return isinstance(other, self.__class__)
94
95 def __repr__(self):
96 return '<{0}Path({1!r})>'.format(
97 '~' if self.inverted else '', self.value)
98
99
100 # TODO: support excluding ranges as soon as
101 # vso_attrs._Range.__xor__ is fixed / renamed
102 class DownloadTime(Attr, vso_attrs._Range):
103 def __init__(self, start, end):
104 self.start = parse_time(start)
105 self.end = parse_time(end)
106 self.inverted = False
107 vso_attrs._Range.__init__(self, start, end, self.__class__)
108
109 def __invert__(self):
110 download_time = self.__class__(self.start, self.end)
111 download_time.inverted = True
112 return download_time
113
114 def collides(self, other): # pragma: no cover
115 return isinstance(other, self.__class__)
116
117 def __repr__(self):
118 return '<{0}DownloadTime({1!r}, {2!r})>'.format(
119 '~' if self.inverted else '', self.start, self.end)
120
121
122 class FitsHeaderEntry(Attr):
123 def __init__(self, key, value):
124 self.key = key
125 self.value = value
126 self.inverted = False
127
128 def __invert__(self):
129 entry = self.__class__(self.key, self.value)
130 entry.inverted = True
131 return entry
132
133 def collides(self, other): # pragma: no cover
134 return False
135
136 def __repr__(self):
137 return '<{0}FitsHeaderEntry({1!r}, {2!r})>'.format(
138 '~' if self.inverted else '', self.key, self.value)
139
140
141 walker = AttrWalker()
142
143
144 @walker.add_creator(AttrOr)
145 def _create(wlk, root, session):
146 entries = set()
147 for attr in root.attrs:
148 # make sure to add only new entries to the set to avoid duplicates
149 entries.update(set(wlk.create(attr, session)) - entries)
150 return list(entries)
151
152
153 @walker.add_creator(AttrAnd)
154 def _create(wlk, root, session):
155 entries = [set(wlk.create(attr, session)) for attr in root.attrs]
156 return list(set.intersection(*entries))
157
158
159 @walker.add_creator(ValueAttr)
160 def _create(wlk, root, session):
161 query = session.query(DatabaseEntry)
162 for key, value in root.attrs.iteritems():
163 typ = key[0]
164 if typ == 'tag':
165 criterion = TableTag.name.in_([value])
166 # `key[1]` is here the `inverted` attribute of the tag. That means
167 # that if it is True, the given tag must not be included in the
168 # resulting entries.
169 if key[1]:
170 query = query.filter(~DatabaseEntry.tags.any(criterion))
171 else:
172 query = query.filter(DatabaseEntry.tags.any(criterion))
173 elif typ == 'fitsheaderentry':
174 key, val, inverted = value
175 key_criterion = TableFitsHeaderEntry.key == key
176 value_criterion = TableFitsHeaderEntry.value == val
177 if inverted:
178 query = query.filter(not_(and_(
179 DatabaseEntry.fits_header_entries.any(key_criterion),
180 DatabaseEntry.fits_header_entries.any(value_criterion))))
181 else:
182 query = query.filter(and_(
183 DatabaseEntry.fits_header_entries.any(key_criterion),
184 DatabaseEntry.fits_header_entries.any(value_criterion)))
185 elif typ == 'download time':
186 start, end, inverted = value
187 if inverted:
188 query = query.filter(
189 ~DatabaseEntry.download_time.between(start, end))
190 else:
191 query = query.filter(
192 DatabaseEntry.download_time.between(start, end))
193 elif typ == 'path':
194 path, inverted = value
195 if inverted:
196 query = query.filter(or_(
197 DatabaseEntry.path != path, DatabaseEntry.path == None))
198 else:
199 query = query.filter(DatabaseEntry.path == path)
200 elif typ == 'wave':
201 min_, max_, unit = value
202 waveunit = Unit(unit)
203 # convert min_ and max_ to nm from the unit `waveunit`
204 wavemin = waveunit.to(nm, min_, equivalencies.spectral())
205 wavemax = waveunit.to(nm, max_, equivalencies.spectral())
206 query = query.filter(and_(
207 DatabaseEntry.wavemin >= wavemin,
208 DatabaseEntry.wavemax <= wavemax))
209 elif typ == 'time':
210 start, end, near = value
211 query = query.filter(and_(
212 DatabaseEntry.observation_time_start >= start,
213 DatabaseEntry.observation_time_end <= end))
214 else:
215 query = query.filter_by(**{typ: value})
216 return query.all()
217
218
219 @walker.add_converter(Tag)
220 def _convert(attr):
221 return ValueAttr({('tag', attr.inverted): attr.tagname})
222
223
224 @walker.add_converter(Starred)
225 def _convert(attr):
226 return ValueAttr({('starred', ): attr.value})
227
228
229 @walker.add_converter(Path)
230 def _convert(attr):
231 return ValueAttr({('path', ): (attr.value, attr.inverted)})
232
233
234 @walker.add_converter(DownloadTime)
235 def _convert(attr):
236 return ValueAttr({
237 ('download time', ): (attr.start, attr.end, attr.inverted)})
238
239
240 @walker.add_converter(FitsHeaderEntry)
241 def _convert(attr):
242 return ValueAttr(
243 {('fitsheaderentry', ): (attr.key, attr.value, attr.inverted)})
244
245
246 @walker.add_converter(vso_attrs._VSOSimpleAttr)
247 def _convert(attr):
248 return ValueAttr({(attr.__class__.__name__.lower(), ): attr.value})
249
250
251 @walker.add_converter(vso_attrs.Wave)
252 def _convert(attr):
253 return ValueAttr({('wave', ): (attr.min, attr.max, attr.unit)})
254
255
256 @walker.add_converter(vso_attrs.Time)
257 def _convert(attr):
258 return ValueAttr({('time', ): (attr.start, attr.end, attr.near)})
259
[end of sunpy/database/attrs.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/sunpy/database/attrs.py b/sunpy/database/attrs.py
--- a/sunpy/database/attrs.py
+++ b/sunpy/database/attrs.py
@@ -209,8 +209,8 @@
elif typ == 'time':
start, end, near = value
query = query.filter(and_(
- DatabaseEntry.observation_time_start >= start,
- DatabaseEntry.observation_time_end <= end))
+ DatabaseEntry.observation_time_start < end,
+ DatabaseEntry.observation_time_end > start))
else:
query = query.filter_by(**{typ: value})
return query.all()
|
{"golden_diff": "diff --git a/sunpy/database/attrs.py b/sunpy/database/attrs.py\n--- a/sunpy/database/attrs.py\n+++ b/sunpy/database/attrs.py\n@@ -209,8 +209,8 @@\n elif typ == 'time':\n start, end, near = value\n query = query.filter(and_(\n- DatabaseEntry.observation_time_start >= start,\n- DatabaseEntry.observation_time_end <= end))\n+ DatabaseEntry.observation_time_start < end,\n+ DatabaseEntry.observation_time_end > start))\n else:\n query = query.filter_by(**{typ: value})\n return query.all()\n", "issue": "Database.query buggy with Time\nDatabase.query works different using the Time attribute than VSO does.\n\n", "before_files": [{"content": "# Author: Simon Liedtke <[email protected]>\n#\n# This module was developed with funding provided by\n# the Google Summer of Code (2013).\n\nfrom __future__ import absolute_import\n\nfrom sqlalchemy import or_, and_, not_\nfrom astropy.units import Unit, nm, equivalencies\n\nfrom sunpy.time import parse_time\nfrom sunpy.net.vso import attrs as vso_attrs\nfrom sunpy.net.attr import AttrWalker, Attr, ValueAttr, AttrAnd, AttrOr\nfrom sunpy.database.tables import DatabaseEntry, Tag as TableTag,\\\n FitsHeaderEntry as TableFitsHeaderEntry\n\n__all__ = [\n 'Starred', 'Tag', 'Path', 'DownloadTime', 'FitsHeaderEntry', 'walker']\n\n\nclass _BooleanAttr(object):\n def __init__(self, value, make):\n self.value = bool(value)\n self.make = make\n\n def __and__(self, other):\n if not isinstance(other, self.make):\n return AttrAnd([self, other])\n attr = self.make()\n attr.value = self.value and other.value\n return attr\n\n def __or__(self, other):\n if not isinstance(other, self.make):\n return AttrOr([self, other])\n attr = self.make()\n attr.value = self.value or other.value\n return attr\n\n def __nonzero__(self):\n return self.value\n\n def __invert__(self):\n attr = self.make()\n attr.value = not self.value\n return attr\n\n def __eq__(self, other):\n return isinstance(other, self.make) and self.value == other.value\n\n def collides(self, other): # pragma: no cover\n return False\n\n def __repr__(self):\n return '<{0}{1}()>'.format(\n '~' if not self.value else '', self.__class__.__name__)\n\n\nclass Starred(_BooleanAttr, Attr):\n def __init__(self):\n _BooleanAttr.__init__(self, True, self.__class__)\n\n\nclass Tag(Attr):\n def __init__(self, tagname):\n self.tagname = tagname\n self.inverted = False\n\n def __invert__(self):\n tag = self.__class__(self.tagname)\n tag.inverted = True\n return tag\n\n def collides(self, other): # pragma: no cover\n return False\n\n def __repr__(self):\n return '<{0}Tag({1!r})>'.format(\n '~' if self.inverted else '', self.tagname)\n\n\nclass Path(Attr):\n def __init__(self, value):\n self.value = value\n self.inverted = False\n\n def __invert__(self):\n path = self.__class__(self.value)\n path.inverted = True\n return path\n\n def collides(self, other): # pragma: no cover\n return isinstance(other, self.__class__)\n\n def __repr__(self):\n return '<{0}Path({1!r})>'.format(\n '~' if self.inverted else '', self.value)\n\n\n# TODO: support excluding ranges as soon as\n# vso_attrs._Range.__xor__ is fixed / renamed\nclass DownloadTime(Attr, vso_attrs._Range):\n def __init__(self, start, end):\n self.start = parse_time(start)\n self.end = parse_time(end)\n self.inverted = False\n vso_attrs._Range.__init__(self, start, end, self.__class__)\n\n def __invert__(self):\n download_time = self.__class__(self.start, self.end)\n download_time.inverted = True\n return download_time\n\n def collides(self, other): # pragma: no cover\n return isinstance(other, self.__class__)\n\n def __repr__(self):\n return '<{0}DownloadTime({1!r}, {2!r})>'.format(\n '~' if self.inverted else '', self.start, self.end)\n\n\nclass FitsHeaderEntry(Attr):\n def __init__(self, key, value):\n self.key = key\n self.value = value\n self.inverted = False\n\n def __invert__(self):\n entry = self.__class__(self.key, self.value)\n entry.inverted = True\n return entry\n\n def collides(self, other): # pragma: no cover\n return False\n\n def __repr__(self):\n return '<{0}FitsHeaderEntry({1!r}, {2!r})>'.format(\n '~' if self.inverted else '', self.key, self.value)\n\n\nwalker = AttrWalker()\n\n\[email protected]_creator(AttrOr)\ndef _create(wlk, root, session):\n entries = set()\n for attr in root.attrs:\n # make sure to add only new entries to the set to avoid duplicates\n entries.update(set(wlk.create(attr, session)) - entries)\n return list(entries)\n\n\[email protected]_creator(AttrAnd)\ndef _create(wlk, root, session):\n entries = [set(wlk.create(attr, session)) for attr in root.attrs]\n return list(set.intersection(*entries))\n\n\[email protected]_creator(ValueAttr)\ndef _create(wlk, root, session):\n query = session.query(DatabaseEntry)\n for key, value in root.attrs.iteritems():\n typ = key[0]\n if typ == 'tag':\n criterion = TableTag.name.in_([value])\n # `key[1]` is here the `inverted` attribute of the tag. That means\n # that if it is True, the given tag must not be included in the\n # resulting entries.\n if key[1]:\n query = query.filter(~DatabaseEntry.tags.any(criterion))\n else:\n query = query.filter(DatabaseEntry.tags.any(criterion))\n elif typ == 'fitsheaderentry':\n key, val, inverted = value\n key_criterion = TableFitsHeaderEntry.key == key\n value_criterion = TableFitsHeaderEntry.value == val\n if inverted:\n query = query.filter(not_(and_(\n DatabaseEntry.fits_header_entries.any(key_criterion),\n DatabaseEntry.fits_header_entries.any(value_criterion))))\n else:\n query = query.filter(and_(\n DatabaseEntry.fits_header_entries.any(key_criterion),\n DatabaseEntry.fits_header_entries.any(value_criterion)))\n elif typ == 'download time':\n start, end, inverted = value\n if inverted:\n query = query.filter(\n ~DatabaseEntry.download_time.between(start, end))\n else:\n query = query.filter(\n DatabaseEntry.download_time.between(start, end))\n elif typ == 'path':\n path, inverted = value\n if inverted:\n query = query.filter(or_(\n DatabaseEntry.path != path, DatabaseEntry.path == None))\n else:\n query = query.filter(DatabaseEntry.path == path)\n elif typ == 'wave':\n min_, max_, unit = value\n waveunit = Unit(unit)\n # convert min_ and max_ to nm from the unit `waveunit`\n wavemin = waveunit.to(nm, min_, equivalencies.spectral())\n wavemax = waveunit.to(nm, max_, equivalencies.spectral())\n query = query.filter(and_(\n DatabaseEntry.wavemin >= wavemin,\n DatabaseEntry.wavemax <= wavemax))\n elif typ == 'time':\n start, end, near = value\n query = query.filter(and_(\n DatabaseEntry.observation_time_start >= start,\n DatabaseEntry.observation_time_end <= end))\n else:\n query = query.filter_by(**{typ: value})\n return query.all()\n\n\[email protected]_converter(Tag)\ndef _convert(attr):\n return ValueAttr({('tag', attr.inverted): attr.tagname})\n\n\[email protected]_converter(Starred)\ndef _convert(attr):\n return ValueAttr({('starred', ): attr.value})\n\n\[email protected]_converter(Path)\ndef _convert(attr):\n return ValueAttr({('path', ): (attr.value, attr.inverted)})\n\n\[email protected]_converter(DownloadTime)\ndef _convert(attr):\n return ValueAttr({\n ('download time', ): (attr.start, attr.end, attr.inverted)})\n\n\[email protected]_converter(FitsHeaderEntry)\ndef _convert(attr):\n return ValueAttr(\n {('fitsheaderentry', ): (attr.key, attr.value, attr.inverted)})\n\n\[email protected]_converter(vso_attrs._VSOSimpleAttr)\ndef _convert(attr):\n return ValueAttr({(attr.__class__.__name__.lower(), ): attr.value})\n\n\[email protected]_converter(vso_attrs.Wave)\ndef _convert(attr):\n return ValueAttr({('wave', ): (attr.min, attr.max, attr.unit)})\n\n\[email protected]_converter(vso_attrs.Time)\ndef _convert(attr):\n return ValueAttr({('time', ): (attr.start, attr.end, attr.near)})\n", "path": "sunpy/database/attrs.py"}]}
| 3,170 | 142 |
gh_patches_debug_42442
|
rasdani/github-patches
|
git_diff
|
saulpw__visidata-2434
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
[fec] opening rows in FEC files does not seem to work
**Small description**
FEC files can be opened, but rows cannot be browsed.
**Data to reproduce**
Using [a sample file 1326070.fec](https://docquery.fec.gov/dcdev/posted/1326070.fec).
**Steps to reproduce**
**Expected result**
When the user presses `Enter`, rows should open as in [jsvine's demo video at 13 seconds].(https://asciinema.org/a/Xyh2BFsUaOF0AlHTmMUbqQZPC)
**Actual result**
When the user presses `Enter`, `open-rows` runs, for example, opening the itemization row, as:
```
column | value
------------------------
component | itemization
size | 56958
```
**Configuration**
- Does this issue reproduce without any plugins or configuration (using the `-N` CLI flag)?
Yes, it reproduces with `-N`.
- Does this issue reproduce with either the [latest release](https://www.visidata.org/releases/), or with the [develop branch](https://www.visidata.org/install/#update-visidata-from-an-existing-installation)?
Yes, the latest develop.
**Additional context**
`fecfile` 0.8.0 installed via pip
Python 3.10.12
Ubuntu 22.04.4
</issue>
<code>
[start of visidata/loaders/fec.py]
1 """
2 Filename: vdfec.py
3 Last updated: 2019-04-21
4 Home: https://github.com/jsvine/visidata-plugins
5 Author: Jeremy Singer-Vine
6
7 # Installation
8
9 - Install fecfile: `pip install fecfile`
10 - Add vdfec.py to your ~/.visidata directory
11 - Add "import vdfec" to your ~/.visidatarc file
12
13 # Usage
14
15 vdfec.py enables VisiData to load .fec files from the Federal Election Commission.
16
17 Once saved to your ~/.visidata directory, and imported via your ~/.visidatarc file,
18 you should be able to interactively explore .fec files as you would any other
19 filetype in VisiData.
20
21 From the command line:
22
23 vd path/to/my/file.fec
24
25 # Thanks
26
27 vdfec.py depends heavily on Evan Sonderegger's `fecfile` Python library: https://esonderegger.github.io/fecfile/
28
29 ... which in turn is based on Derek Willis's `Fech` Ruby library: https://github.com/dwillis/Fech
30
31 ... and Chris Zubak-Skees' transformation of `Fech`'s form-and-field mappings: https://github.com/PublicI/fec-parse/blob/master/lib/renderedmaps.js
32
33 Thanks to all who have contributed to those projects.
34
35 """
36
37 from copy import copy
38 from visidata import (
39 vd,
40 Path,
41 Sheet,
42 TextSheet,
43 Column,
44 ColumnAttr,
45 ColumnItem,
46 ENTER,
47 asyncthread,
48 Progress,
49 addGlobals,
50 )
51
52 class DiveSheet(Sheet):
53 "A deeply-diveable, quick-diving sheet."
54
55 def reload(self):
56 mapping = self.source
57
58 self.columns = []
59 self.rows = []
60
61 self.key_type = str
62 self.size = len(mapping)
63
64 if self.size == 0:
65 return
66
67 if isinstance(mapping, list):
68 first = mapping[0]
69 if isinstance(first, dict):
70 colgetter = lambda x: x.keys()
71 elif isinstance(first, list):
72 colgetter = lambda x: list(range(len(x)))
73 else:
74 mapping = dict(enumerate(mapping))
75 self.key_type = int
76 self.size = len(mapping)
77
78 if isinstance(mapping, dict):
79 self.is_keyvalue = True
80 if self.size:
81 max_key_len = max(map(len, map(str, mapping.keys())))
82 key_width = min(50, max(max_key_len + 2, 6))
83 else:
84 key_width = None
85
86 self.addColumn(ColumnItem(
87 "key",
88 width = key_width,
89 type = self.key_type
90 ))
91 self.addColumn(ColumnItem("value"))
92 self.setKeys(self.columns[:1])
93
94 for k, v in mapping.items():
95 self.addRow({ "key": k, "value": v })
96
97 elif isinstance(mapping, list):
98 self.is_keyvalue = False
99 indices = []
100 for item in mapping:
101 try:
102 cols = colgetter(item)
103 for col in cols:
104 if col not in indices:
105 self.addColumn(ColumnItem(col))
106 indices.append(col)
107
108 self.addRow(item)
109
110 except Exception as e:
111 vd.warning("Can't dive on lists with heterogenous item types.")
112 return False
113
114 def dive(self):
115 if self.is_keyvalue:
116 cell = self.cursorRow["value"]
117 name = vd.joinSheetnames(self.name, self.cursorRow["key"])
118
119 if isinstance(cell, (list, dict)):
120 vs = self.__class__(name, source = cell)
121 else:
122 vd.warning("Nothing to dive into.")
123 return
124 else:
125 name = vd.joinSheetnames(self.name, "row")
126 vs = self.__class__(name, source = self.cursorRow)
127
128 success = vs.reload()
129 if success == False:
130 return
131
132 vd.push(vs)
133
134 DiveSheet.addCommand(
135 ENTER,
136 'dive-row',
137 'vd.sheet.dive()'
138 )
139
140 class FECItemizationSheet(Sheet):
141 "A sheet to display a list of FEC itemizations from a given form/schedule."
142
143 rowtype = "itemizations"
144
145 @asyncthread
146 def reload(self):
147 self.rows = []
148 self.columns = []
149
150 if len(self.source) == 0:
151 return
152
153 for i, row in enumerate(Progress(self.source, total = len(self.source))):
154 if i == 0:
155 self.set_columns_from_row(row)
156 self.addRow(row)
157
158 def set_columns_from_row(self, row):
159 self.columns.clear()
160 for i, name in enumerate(row.keys()):
161 self.addColumn(ColumnItem(name))
162 def dive(self):
163 vs = DiveSheet(
164 vd.joinSheetnames(self.name, "detail"),
165 source = self.cursorRow
166 )
167 vs.reload()
168 vd.push(vs)
169
170 FECItemizationSheet.addCommand(
171 ENTER,
172 'dive-row',
173 'vd.sheet.dive()'
174 )
175
176 class FECScheduleSheet(Sheet):
177 "A sheet to display the list of itemized schedules in a filing."
178
179 rowtype = "schedules"
180
181 columns = [
182 ColumnAttr("schedule", "schedule_name", width = 14),
183 ColumnAttr("name", width = 0),
184 ColumnAttr("size", type = int),
185 ]
186
187 nKeys = 1
188
189 @asyncthread
190 def reload(self):
191 self.rows = []
192
193 for schedule_name in self.source.keys():
194 vs = FECItemizationSheet(
195 vd.joinSheetnames(self.name, schedule_name),
196 schedule_name = schedule_name,
197 source = self.source[schedule_name],
198 size = len(self.source[schedule_name]),
199 )
200 self.addRow(vs)
201
202 FECScheduleSheet.addCommand(
203 ENTER,
204 'dive-row',
205 'vd.push(cursorRow)'
206 )
207
208 COMPONENT_SHEET_CLASSES = {
209 "header": DiveSheet,
210 "summary": DiveSheet,
211 "itemization": FECScheduleSheet,
212 "text": FECItemizationSheet,
213 "F99_text": TextSheet,
214 }
215
216 class FECFiling(Sheet):
217 "A sheet representing an entire .fec file."
218
219 rowtype = "components"
220 filing = None
221
222 columns = [
223 ColumnAttr("component", "component_name", width = 14),
224 ColumnAttr("name", width = 0),
225 ColumnAttr("size", type = int),
226 ]
227
228 nKeys = 1
229
230 @asyncthread
231 def reload(self):
232 from fecfile import fecparser
233 self.rows = []
234
235 row_dict = { }
236 itemization_subsheets = {}
237
238 def addSheetRow(component_name):
239 "On first encountering a component, add a row to the filing sheet"
240
241 cls = COMPONENT_SHEET_CLASSES[component_name]
242
243 source_cls = list if cls in [
244 FECItemizationSheet,
245 TextSheet
246 ] else dict
247
248 vs = cls(
249 vd.joinSheetnames(self.name, component_name),
250 component_name = component_name,
251 source = source_cls(),
252 size = 0,
253 )
254
255 vs.reload()
256 row_dict[component_name] = vs
257 self.addRow(vs)
258
259 src = Path(self.source.resolve())
260
261 item_iter = fecparser.iter_lines(src, { "as_strings": True })
262
263 for item in item_iter:
264 dtype = item.data_type
265 if dtype not in row_dict.keys():
266 addSheetRow(dtype)
267
268 sheet_row = row_dict[dtype]
269
270 if dtype in [ "header", "summary" ]:
271 sheet_row.source = item.data
272 sheet_row.reload()
273
274 elif dtype == "text":
275 if len(sheet_row.source) == 0:
276 sheet_row.set_columns_from_row(item.data)
277 sheet_row.source.append(item.data)
278 sheet_row.addRow(item.data)
279 sheet_row.size += 1
280
281 elif dtype == "F99_text":
282 sheet_row.source = item.data.split("\n")
283 sheet_row.size = len(sheet_row.source)
284
285 elif dtype == "itemization":
286 form_type = item.data["form_type"]
287
288 if form_type[0] == "S":
289 form_type = "Schedule " + item.data["form_type"][1]
290
291 if form_type not in sheet_row.source:
292 sheet_row.source[form_type] = [ ]
293 subsheet = FECItemizationSheet(
294 vd.joinSheetnames(sheet_row.name, form_type),
295 schedule_name = form_type,
296 source = [ ],
297 size = 0,
298 )
299 subsheet.reload()
300 subsheet.set_columns_from_row(item.data)
301 sheet_row.addRow(subsheet)
302 itemization_subsheets[form_type] = subsheet
303 else:
304 subsheet = itemization_subsheets[form_type]
305
306 subsheet.addRow(item.data)
307 subsheet.source.append(item.data)
308 subsheet.size += 1
309
310 sheet_row.source[form_type].append(item.data)
311 sheet_row.size += 1
312
313 FECFiling.addCommand(
314 ENTER,
315 'dive-row',
316 'vd.push(cursorRow)'
317 )
318
319 def open_fec(p):
320 return FECFiling(p.base_stem, source=p)
321
322 addGlobals({
323 "open_fec": open_fec,
324 "DiveSheet": DiveSheet
325 })
326
[end of visidata/loaders/fec.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/visidata/loaders/fec.py b/visidata/loaders/fec.py
--- a/visidata/loaders/fec.py
+++ b/visidata/loaders/fec.py
@@ -37,13 +37,12 @@
from copy import copy
from visidata import (
vd,
+ VisiData,
Path,
Sheet,
TextSheet,
- Column,
ColumnAttr,
ColumnItem,
- ENTER,
asyncthread,
Progress,
addGlobals,
@@ -111,10 +110,10 @@
vd.warning("Can't dive on lists with heterogenous item types.")
return False
- def dive(self):
+ def openRow(self, row):
if self.is_keyvalue:
- cell = self.cursorRow["value"]
- name = vd.joinSheetnames(self.name, self.cursorRow["key"])
+ cell = row["value"]
+ name = vd.joinSheetnames(self.name, row["key"])
if isinstance(cell, (list, dict)):
vs = self.__class__(name, source = cell)
@@ -123,19 +122,13 @@
return
else:
name = vd.joinSheetnames(self.name, "row")
- vs = self.__class__(name, source = self.cursorRow)
+ vs = self.__class__(name, source = self.row)
success = vs.reload()
if success == False:
- return
-
- vd.push(vs)
+ vd.fail('could not reload new sheet')
+ return vs
-DiveSheet.addCommand(
- ENTER,
- 'dive-row',
- 'vd.sheet.dive()'
-)
class FECItemizationSheet(Sheet):
"A sheet to display a list of FEC itemizations from a given form/schedule."
@@ -159,19 +152,9 @@
self.columns.clear()
for i, name in enumerate(row.keys()):
self.addColumn(ColumnItem(name))
- def dive(self):
- vs = DiveSheet(
- vd.joinSheetnames(self.name, "detail"),
- source = self.cursorRow
- )
- vs.reload()
- vd.push(vs)
-
-FECItemizationSheet.addCommand(
- ENTER,
- 'dive-row',
- 'vd.sheet.dive()'
-)
+
+ def openRow(self, row):
+ return row
class FECScheduleSheet(Sheet):
"A sheet to display the list of itemized schedules in a filing."
@@ -199,11 +182,8 @@
)
self.addRow(vs)
-FECScheduleSheet.addCommand(
- ENTER,
- 'dive-row',
- 'vd.push(cursorRow)'
-)
+ def openRow(self, row):
+ return row
COMPONENT_SHEET_CLASSES = {
"header": DiveSheet,
@@ -230,7 +210,7 @@
@asyncthread
def reload(self):
from fecfile import fecparser
- self.rows = []
+ self.rows = [] # rowdef: Sheet, of a type from COMPONENT_SHEET_CLASSES.values()
row_dict = { }
itemization_subsheets = {}
@@ -310,16 +290,13 @@
sheet_row.source[form_type].append(item.data)
sheet_row.size += 1
-FECFiling.addCommand(
- ENTER,
- 'dive-row',
- 'vd.push(cursorRow)'
-)
+ def openRow(self, row):
+ return row
-def open_fec(p):
[email protected]
+def open_fec(vd, p):
return FECFiling(p.base_stem, source=p)
addGlobals({
- "open_fec": open_fec,
"DiveSheet": DiveSheet
})
|
{"golden_diff": "diff --git a/visidata/loaders/fec.py b/visidata/loaders/fec.py\n--- a/visidata/loaders/fec.py\n+++ b/visidata/loaders/fec.py\n@@ -37,13 +37,12 @@\n from copy import copy\n from visidata import (\n vd,\n+ VisiData,\n Path,\n Sheet,\n TextSheet,\n- Column,\n ColumnAttr,\n ColumnItem,\n- ENTER,\n asyncthread,\n Progress,\n addGlobals,\n@@ -111,10 +110,10 @@\n vd.warning(\"Can't dive on lists with heterogenous item types.\")\n return False\n \n- def dive(self):\n+ def openRow(self, row):\n if self.is_keyvalue:\n- cell = self.cursorRow[\"value\"]\n- name = vd.joinSheetnames(self.name, self.cursorRow[\"key\"])\n+ cell = row[\"value\"]\n+ name = vd.joinSheetnames(self.name, row[\"key\"])\n \n if isinstance(cell, (list, dict)):\n vs = self.__class__(name, source = cell)\n@@ -123,19 +122,13 @@\n return\n else:\n name = vd.joinSheetnames(self.name, \"row\")\n- vs = self.__class__(name, source = self.cursorRow)\n+ vs = self.__class__(name, source = self.row)\n \n success = vs.reload()\n if success == False:\n- return\n-\n- vd.push(vs)\n+ vd.fail('could not reload new sheet')\n+ return vs\n \n-DiveSheet.addCommand(\n- ENTER,\n- 'dive-row',\n- 'vd.sheet.dive()'\n-)\n \n class FECItemizationSheet(Sheet):\n \"A sheet to display a list of FEC itemizations from a given form/schedule.\"\n@@ -159,19 +152,9 @@\n self.columns.clear()\n for i, name in enumerate(row.keys()):\n self.addColumn(ColumnItem(name))\n- def dive(self):\n- vs = DiveSheet(\n- vd.joinSheetnames(self.name, \"detail\"),\n- source = self.cursorRow\n- )\n- vs.reload()\n- vd.push(vs)\n-\n-FECItemizationSheet.addCommand(\n- ENTER,\n- 'dive-row',\n- 'vd.sheet.dive()'\n-)\n+\n+ def openRow(self, row):\n+ return row\n \n class FECScheduleSheet(Sheet):\n \"A sheet to display the list of itemized schedules in a filing.\"\n@@ -199,11 +182,8 @@\n )\n self.addRow(vs)\n \n-FECScheduleSheet.addCommand(\n- ENTER,\n- 'dive-row',\n- 'vd.push(cursorRow)'\n-)\n+ def openRow(self, row):\n+ return row\n \n COMPONENT_SHEET_CLASSES = {\n \"header\": DiveSheet,\n@@ -230,7 +210,7 @@\n @asyncthread\n def reload(self):\n from fecfile import fecparser\n- self.rows = []\n+ self.rows = [] # rowdef: Sheet, of a type from COMPONENT_SHEET_CLASSES.values()\n \n row_dict = { }\n itemization_subsheets = {}\n@@ -310,16 +290,13 @@\n sheet_row.source[form_type].append(item.data)\n sheet_row.size += 1\n \n-FECFiling.addCommand(\n- ENTER,\n- 'dive-row',\n- 'vd.push(cursorRow)'\n-)\n+ def openRow(self, row):\n+ return row\n \n-def open_fec(p):\[email protected]\n+def open_fec(vd, p):\n return FECFiling(p.base_stem, source=p)\n \n addGlobals({\n- \"open_fec\": open_fec,\n \"DiveSheet\": DiveSheet\n })\n", "issue": "[fec] opening rows in FEC files does not seem to work\n**Small description**\r\nFEC files can be opened, but rows cannot be browsed.\r\n\r\n**Data to reproduce**\r\nUsing [a sample file 1326070.fec](https://docquery.fec.gov/dcdev/posted/1326070.fec).\r\n\r\n**Steps to reproduce**\r\n\r\n**Expected result**\r\nWhen the user presses `Enter`, rows should open as in [jsvine's demo video at 13 seconds].(https://asciinema.org/a/Xyh2BFsUaOF0AlHTmMUbqQZPC)\r\n\r\n**Actual result**\r\nWhen the user presses `Enter`, `open-rows` runs, for example, opening the itemization row, as:\r\n```\r\ncolumn | value\r\n------------------------\r\ncomponent | itemization\r\nsize | 56958\r\n```\r\n\r\n**Configuration**\r\n- Does this issue reproduce without any plugins or configuration (using the `-N` CLI flag)?\r\nYes, it reproduces with `-N`.\r\n\r\n- Does this issue reproduce with either the [latest release](https://www.visidata.org/releases/), or with the [develop branch](https://www.visidata.org/install/#update-visidata-from-an-existing-installation)?\r\nYes, the latest develop.\r\n\r\n**Additional context**\r\n`fecfile` 0.8.0 installed via pip\r\nPython 3.10.12\r\nUbuntu 22.04.4\n", "before_files": [{"content": "\"\"\"\nFilename: vdfec.py\nLast updated: 2019-04-21\nHome: https://github.com/jsvine/visidata-plugins\nAuthor: Jeremy Singer-Vine\n\n# Installation\n\n- Install fecfile: `pip install fecfile`\n- Add vdfec.py to your ~/.visidata directory\n- Add \"import vdfec\" to your ~/.visidatarc file\n\n# Usage\n\nvdfec.py enables VisiData to load .fec files from the Federal Election Commission.\n\nOnce saved to your ~/.visidata directory, and imported via your ~/.visidatarc file,\nyou should be able to interactively explore .fec files as you would any other \nfiletype in VisiData.\n\nFrom the command line:\n\n vd path/to/my/file.fec\n\n# Thanks\n\nvdfec.py depends heavily on Evan Sonderegger's `fecfile` Python library: https://esonderegger.github.io/fecfile/\n\n... which in turn is based on Derek Willis's `Fech` Ruby library: https://github.com/dwillis/Fech\n\n... and Chris Zubak-Skees' transformation of `Fech`'s form-and-field mappings: https://github.com/PublicI/fec-parse/blob/master/lib/renderedmaps.js\n\nThanks to all who have contributed to those projects.\n\n\"\"\"\n\nfrom copy import copy\nfrom visidata import (\n vd,\n Path,\n Sheet,\n TextSheet,\n Column,\n ColumnAttr,\n ColumnItem,\n ENTER,\n asyncthread,\n Progress,\n addGlobals,\n)\n\nclass DiveSheet(Sheet):\n \"A deeply-diveable, quick-diving sheet.\"\n\n def reload(self):\n mapping = self.source\n\n self.columns = []\n self.rows = []\n\n self.key_type = str\n self.size = len(mapping)\n\n if self.size == 0:\n return\n\n if isinstance(mapping, list):\n first = mapping[0]\n if isinstance(first, dict):\n colgetter = lambda x: x.keys()\n elif isinstance(first, list):\n colgetter = lambda x: list(range(len(x)))\n else:\n mapping = dict(enumerate(mapping))\n self.key_type = int\n self.size = len(mapping)\n\n if isinstance(mapping, dict):\n self.is_keyvalue = True\n if self.size:\n max_key_len = max(map(len, map(str, mapping.keys())))\n key_width = min(50, max(max_key_len + 2, 6))\n else:\n key_width = None\n\n self.addColumn(ColumnItem(\n \"key\",\n width = key_width,\n type = self.key_type\n ))\n self.addColumn(ColumnItem(\"value\"))\n self.setKeys(self.columns[:1])\n\n for k, v in mapping.items():\n self.addRow({ \"key\": k, \"value\": v })\n\n elif isinstance(mapping, list):\n self.is_keyvalue = False\n indices = [] \n for item in mapping:\n try:\n cols = colgetter(item)\n for col in cols:\n if col not in indices:\n self.addColumn(ColumnItem(col))\n indices.append(col)\n\n self.addRow(item)\n\n except Exception as e:\n vd.warning(\"Can't dive on lists with heterogenous item types.\")\n return False\n\n def dive(self):\n if self.is_keyvalue:\n cell = self.cursorRow[\"value\"]\n name = vd.joinSheetnames(self.name, self.cursorRow[\"key\"])\n\n if isinstance(cell, (list, dict)):\n vs = self.__class__(name, source = cell)\n else:\n vd.warning(\"Nothing to dive into.\")\n return\n else:\n name = vd.joinSheetnames(self.name, \"row\")\n vs = self.__class__(name, source = self.cursorRow)\n\n success = vs.reload()\n if success == False:\n return\n\n vd.push(vs)\n\nDiveSheet.addCommand(\n ENTER,\n 'dive-row',\n 'vd.sheet.dive()'\n)\n\nclass FECItemizationSheet(Sheet):\n \"A sheet to display a list of FEC itemizations from a given form/schedule.\"\n\n rowtype = \"itemizations\"\n\n @asyncthread\n def reload(self):\n self.rows = []\n self.columns = []\n\n if len(self.source) == 0:\n return\n\n for i, row in enumerate(Progress(self.source, total = len(self.source))):\n if i == 0:\n self.set_columns_from_row(row)\n self.addRow(row)\n \n def set_columns_from_row(self, row):\n self.columns.clear()\n for i, name in enumerate(row.keys()):\n self.addColumn(ColumnItem(name))\n def dive(self):\n vs = DiveSheet(\n vd.joinSheetnames(self.name, \"detail\"),\n source = self.cursorRow\n )\n vs.reload()\n vd.push(vs)\n\nFECItemizationSheet.addCommand(\n ENTER,\n 'dive-row',\n 'vd.sheet.dive()'\n)\n\nclass FECScheduleSheet(Sheet):\n \"A sheet to display the list of itemized schedules in a filing.\"\n\n rowtype = \"schedules\"\n\n columns = [\n ColumnAttr(\"schedule\", \"schedule_name\", width = 14),\n ColumnAttr(\"name\", width = 0),\n ColumnAttr(\"size\", type = int),\n ]\n\n nKeys = 1\n\n @asyncthread\n def reload(self):\n self.rows = []\n\n for schedule_name in self.source.keys():\n vs = FECItemizationSheet(\n vd.joinSheetnames(self.name, schedule_name),\n schedule_name = schedule_name,\n source = self.source[schedule_name],\n size = len(self.source[schedule_name]),\n )\n self.addRow(vs)\n\nFECScheduleSheet.addCommand(\n ENTER,\n 'dive-row',\n 'vd.push(cursorRow)'\n)\n\nCOMPONENT_SHEET_CLASSES = {\n \"header\": DiveSheet,\n \"summary\": DiveSheet,\n \"itemization\": FECScheduleSheet,\n \"text\": FECItemizationSheet,\n \"F99_text\": TextSheet,\n}\n\nclass FECFiling(Sheet):\n \"A sheet representing an entire .fec file.\"\n\n rowtype = \"components\"\n filing = None\n\n columns = [\n ColumnAttr(\"component\", \"component_name\", width = 14),\n ColumnAttr(\"name\", width = 0),\n ColumnAttr(\"size\", type = int),\n ]\n\n nKeys = 1\n\n @asyncthread\n def reload(self):\n from fecfile import fecparser\n self.rows = []\n\n row_dict = { }\n itemization_subsheets = {}\n\n def addSheetRow(component_name):\n \"On first encountering a component, add a row to the filing sheet\"\n\n cls = COMPONENT_SHEET_CLASSES[component_name]\n\n source_cls = list if cls in [\n FECItemizationSheet,\n TextSheet\n ] else dict\n\n vs = cls(\n vd.joinSheetnames(self.name, component_name),\n component_name = component_name,\n source = source_cls(),\n size = 0,\n )\n\n vs.reload()\n row_dict[component_name] = vs\n self.addRow(vs)\n\n src = Path(self.source.resolve())\n\n item_iter = fecparser.iter_lines(src, { \"as_strings\": True })\n\n for item in item_iter:\n dtype = item.data_type \n if dtype not in row_dict.keys():\n addSheetRow(dtype)\n\n sheet_row = row_dict[dtype]\n\n if dtype in [ \"header\", \"summary\" ]:\n sheet_row.source = item.data\n sheet_row.reload()\n\n elif dtype == \"text\":\n if len(sheet_row.source) == 0:\n sheet_row.set_columns_from_row(item.data)\n sheet_row.source.append(item.data)\n sheet_row.addRow(item.data)\n sheet_row.size += 1\n\n elif dtype == \"F99_text\":\n sheet_row.source = item.data.split(\"\\n\")\n sheet_row.size = len(sheet_row.source)\n\n elif dtype == \"itemization\":\n form_type = item.data[\"form_type\"]\n\n if form_type[0] == \"S\":\n form_type = \"Schedule \" + item.data[\"form_type\"][1]\n\n if form_type not in sheet_row.source:\n sheet_row.source[form_type] = [ ] \n subsheet = FECItemizationSheet(\n vd.joinSheetnames(sheet_row.name, form_type),\n schedule_name = form_type,\n source = [ ],\n size = 0,\n )\n subsheet.reload()\n subsheet.set_columns_from_row(item.data)\n sheet_row.addRow(subsheet)\n itemization_subsheets[form_type] = subsheet\n else:\n subsheet = itemization_subsheets[form_type]\n\n subsheet.addRow(item.data)\n subsheet.source.append(item.data)\n subsheet.size += 1\n\n sheet_row.source[form_type].append(item.data)\n sheet_row.size += 1\n\nFECFiling.addCommand(\n ENTER,\n 'dive-row',\n 'vd.push(cursorRow)'\n)\n\ndef open_fec(p):\n return FECFiling(p.base_stem, source=p)\n\naddGlobals({\n \"open_fec\": open_fec,\n \"DiveSheet\": DiveSheet\n})\n", "path": "visidata/loaders/fec.py"}]}
| 3,743 | 859 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.