problem_id
stringlengths 18
22
| source
stringclasses 1
value | task_type
stringclasses 1
value | in_source_id
stringlengths 13
58
| prompt
stringlengths 1.71k
18.9k
| golden_diff
stringlengths 145
5.13k
| verification_info
stringlengths 465
23.6k
| num_tokens_prompt
int64 556
4.1k
| num_tokens_diff
int64 47
1.02k
|
---|---|---|---|---|---|---|---|---|
gh_patches_debug_53192
|
rasdani/github-patches
|
git_diff
|
buildbot__buildbot-426
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Move to a ChangeHook in hgbuildbot.py (closes #2227)
Starting a Twisted reactor within a Mercurial hook is problematic
because a hook can get called multiple times (from the same Mercurial
process) which conflicts with the one-time only nature of the Twisted
reactor start/stop calls.
By moving to HTTP requests the hook is now re-entrant which fixes
ReactorNotRestartable exceptions being raised all the time.
The requests library is used to make the calls, since it has a
functional TLS implementation (unlike urllib/httplib in Python 2 until
recently).
Additionally, the codebase argument is added to the base ChangeHook, so
it can properly be used with a codebaseGenerator.
Finally, if a venv argument is used, it is now properly used to activate
the environment [1].
Possible improvements to this changeset are:
- send requests in parallel;
- switch back to urllib2 so the hook doesn't depend on anything;
- a few more cleanups.
[1] https://github.com/pypa/virtualenv/blob/master/virtualenv_embedded/activate_this.py
</issue>
<code>
[start of master/buildbot/steps/source/cvs.py]
1 # This file is part of Buildbot. Buildbot is free software: you can
2 # redistribute it and/or modify it under the terms of the GNU General Public
3 # License as published by the Free Software Foundation, version 2.
4 #
5 # This program is distributed in the hope that it will be useful, but WITHOUT
6 # ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS
7 # FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
8 # details.
9 #
10 # You should have received a copy of the GNU General Public License along with
11 # this program; if not, write to the Free Software Foundation, Inc., 51
12 # Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA.
13 #
14 # Copyright Buildbot Team Members
15
16 from email.Utils import formatdate
17 import time
18
19 from twisted.python import log
20 from twisted.internet import defer
21
22 from buildbot.process import buildstep
23 from buildbot.steps.shell import StringFileWriter
24 from buildbot.steps.source.base import Source
25 from buildbot.interfaces import BuildSlaveTooOldError
26
27 class CVS(Source):
28
29 name = "cvs"
30
31 renderables = [ "cvsroot" ]
32
33 def __init__(self, cvsroot=None, cvsmodule='', mode='incremental',
34 method=None, branch=None, global_options=[], extra_options=[],
35 login=None, **kwargs):
36
37 self.cvsroot = cvsroot
38 self.cvsmodule = cvsmodule
39 self.branch = branch
40 self.global_options = global_options
41 self.extra_options = extra_options
42 self.login = login
43 self.mode = mode
44 self.method = method
45 self.srcdir = 'source'
46 Source.__init__(self, **kwargs)
47
48 def startVC(self, branch, revision, patch):
49 self.revision = revision
50 self.stdio_log = self.addLog("stdio")
51 self.method = self._getMethod()
52 d = self.checkCvs()
53 def checkInstall(cvsInstalled):
54 if not cvsInstalled:
55 raise BuildSlaveTooOldError("CVS is not installed on slave")
56 return 0
57 d.addCallback(checkInstall)
58 d.addCallback(self.checkLogin)
59
60 if self.mode == 'incremental':
61 d.addCallback(lambda _: self.incremental())
62 elif self.mode == 'full':
63 d.addCallback(lambda _: self.full())
64
65 d.addCallback(self.parseGotRevision)
66 d.addCallback(self.finish)
67 d.addErrback(self.failed)
68 return d
69
70 @defer.inlineCallbacks
71 def incremental(self):
72 updatable = yield self._sourcedirIsUpdatable()
73 if updatable:
74 rv = yield self.doUpdate()
75 else:
76 rv = yield self.clobber()
77 defer.returnValue(rv)
78
79 @defer.inlineCallbacks
80 def full(self):
81 if self.method == 'clobber':
82 rv = yield self.clobber()
83 defer.returnValue(rv)
84 return
85
86 elif self.method == 'copy':
87 rv = yield self.copy()
88 defer.returnValue(rv)
89 return
90
91 updatable = yield self._sourcedirIsUpdatable()
92 if not updatable:
93 log.msg("CVS repo not present, making full checkout")
94 rv = yield self.doCheckout(self.workdir)
95 elif self.method == 'clean':
96 rv = yield self.clean()
97 elif self.method == 'fresh':
98 rv = yield self.fresh()
99 else:
100 raise ValueError("Unknown method, check your configuration")
101 defer.returnValue(rv)
102
103 def clobber(self):
104 cmd = buildstep.RemoteCommand('rmdir', {'dir': self.workdir,
105 'logEnviron': self.logEnviron})
106 cmd.useLog(self.stdio_log, False)
107 d = self.runCommand(cmd)
108 def checkRemoval(res):
109 if res != 0:
110 raise RuntimeError("Failed to delete directory")
111 return res
112 d.addCallback(lambda _: checkRemoval(cmd.rc))
113 d.addCallback(lambda _: self.doCheckout(self.workdir))
114 return d
115
116 def fresh(self, ):
117 d = self.purge(True)
118 d.addCallback(lambda _: self.doUpdate())
119 return d
120
121 def clean(self, ):
122 d = self.purge(False)
123 d.addCallback(lambda _: self.doUpdate())
124 return d
125
126 def copy(self):
127 cmd = buildstep.RemoteCommand('rmdir', {'dir': self.workdir,
128 'logEnviron': self.logEnviron})
129 cmd.useLog(self.stdio_log, False)
130 d = self.runCommand(cmd)
131 self.workdir = 'source'
132 d.addCallback(lambda _: self.incremental())
133 def copy(_):
134 cmd = buildstep.RemoteCommand('cpdir',
135 {'fromdir': 'source',
136 'todir':'build',
137 'logEnviron': self.logEnviron,})
138 cmd.useLog(self.stdio_log, False)
139 d = self.runCommand(cmd)
140 return d
141 d.addCallback(copy)
142 def resetWorkdir(_):
143 self.workdir = 'build'
144 return 0
145 d.addCallback(resetWorkdir)
146 return d
147
148 def purge(self, ignore_ignores):
149 command = ['cvsdiscard']
150 if ignore_ignores:
151 command += ['--ignore']
152 cmd = buildstep.RemoteShellCommand(self.workdir, command,
153 env=self.env,
154 logEnviron=self.logEnviron)
155 cmd.useLog(self.stdio_log, False)
156 d = self.runCommand(cmd)
157 def evaluate(rc):
158 if rc != 0:
159 raise buildstep.BuildStepFailed()
160 return rc
161 d.addCallback(lambda _: evaluate(cmd.rc))
162 return d
163
164 def doCheckout(self, dir):
165 command = ['-d', self.cvsroot, '-z3', 'checkout', '-d', dir ]
166 command = self.global_options + command + self.extra_options
167 if self.branch:
168 command += ['-r', self.branch]
169 if self.revision:
170 command += ['-D', self.revision]
171 command += [ self.cvsmodule ]
172 d = self._dovccmd(command, '')
173 return d
174
175 def doUpdate(self):
176 command = ['-z3', 'update', '-dP']
177 if self.branch:
178 command += ['-r', self.branch]
179 if self.revision:
180 command += ['-D', self.revision]
181 d = self._dovccmd(command)
182 return d
183
184 def finish(self, res):
185 d = defer.succeed(res)
186 def _gotResults(results):
187 self.setStatus(self.cmd, results)
188 return results
189 d.addCallback(_gotResults)
190 d.addCallbacks(self.finished, self.checkDisconnect)
191 return d
192
193 def checkLogin(self, _):
194 if self.login:
195 d = defer.succeed(0)
196 else:
197 d = self._dovccmd(['-d', self.cvsroot, 'login'])
198 def setLogin(res):
199 # this happens only if the login command succeeds.
200 self.login = True
201 return res
202 d.addCallback(setLogin)
203
204 return d
205
206 def _dovccmd(self, command, workdir=None):
207 if workdir is None:
208 workdir = self.workdir
209 if not command:
210 raise ValueError("No command specified")
211 cmd = buildstep.RemoteShellCommand(workdir, ['cvs'] +
212 command,
213 env=self.env,
214 logEnviron=self.logEnviron)
215 cmd.useLog(self.stdio_log, False)
216 d = self.runCommand(cmd)
217 def evaluateCommand(cmd):
218 if cmd.rc != 0:
219 log.msg("Source step failed while running command %s" % cmd)
220 raise buildstep.BuildStepFailed()
221 return cmd.rc
222 d.addCallback(lambda _: evaluateCommand(cmd))
223 return d
224
225 @defer.inlineCallbacks
226 def _sourcedirIsUpdatable(self):
227 myFileWriter = StringFileWriter()
228 args = {
229 'workdir': self.build.path_module.join(self.workdir, 'CVS'),
230 'writer': myFileWriter,
231 'maxsize': None,
232 'blocksize': 32*1024,
233 }
234
235 cmd = buildstep.RemoteCommand('uploadFile',
236 dict(slavesrc='Root', **args),
237 ignore_updates=True)
238 yield self.runCommand(cmd)
239 if cmd.rc is not None and cmd.rc != 0:
240 defer.returnValue(False)
241 return
242 if myFileWriter.buffer.strip() != self.cvsroot:
243 defer.returnValue(False)
244 return
245
246 myFileWriter.buffer = ""
247 cmd = buildstep.RemoteCommand('uploadFile',
248 dict(slavesrc='Repository', **args),
249 ignore_updates=True)
250 yield self.runCommand(cmd)
251 if cmd.rc is not None and cmd.rc != 0:
252 defer.returnValue(False)
253 return
254 if myFileWriter.buffer.strip() != self.cvsmodule:
255 defer.returnValue(False)
256 return
257
258 defer.returnValue(True)
259
260 def parseGotRevision(self, res):
261 revision = time.strftime("%Y-%m-%d %H:%M:%S +0000", time.gmtime())
262 self.setProperty('got_revision', revision, 'Source')
263 return res
264
265 def checkCvs(self):
266 d = self._dovccmd(['--version'])
267 def check(res):
268 if res == 0:
269 return True
270 return False
271 d.addCallback(check)
272 return d
273
274 def _getMethod(self):
275 if self.method is not None and self.mode != 'incremental':
276 return self.method
277 elif self.mode == 'incremental':
278 return None
279 elif self.method is None and self.mode == 'full':
280 return 'fresh'
281
282 def computeSourceRevision(self, changes):
283 if not changes:
284 return None
285 lastChange = max([c.when for c in changes])
286 lastSubmit = max([br.submittedAt for br in self.build.requests])
287 when = (lastChange + lastSubmit) / 2
288 return formatdate(when)
289
[end of master/buildbot/steps/source/cvs.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/master/buildbot/steps/source/cvs.py b/master/buildbot/steps/source/cvs.py
--- a/master/buildbot/steps/source/cvs.py
+++ b/master/buildbot/steps/source/cvs.py
@@ -46,6 +46,7 @@
Source.__init__(self, **kwargs)
def startVC(self, branch, revision, patch):
+ self.branch = branch
self.revision = revision
self.stdio_log = self.addLog("stdio")
self.method = self._getMethod()
|
{"golden_diff": "diff --git a/master/buildbot/steps/source/cvs.py b/master/buildbot/steps/source/cvs.py\n--- a/master/buildbot/steps/source/cvs.py\n+++ b/master/buildbot/steps/source/cvs.py\n@@ -46,6 +46,7 @@\n Source.__init__(self, **kwargs)\n \n def startVC(self, branch, revision, patch):\n+ self.branch = branch\n self.revision = revision\n self.stdio_log = self.addLog(\"stdio\")\n self.method = self._getMethod()\n", "issue": "Move to a ChangeHook in hgbuildbot.py (closes #2227)\nStarting a Twisted reactor within a Mercurial hook is problematic\nbecause a hook can get called multiple times (from the same Mercurial\nprocess) which conflicts with the one-time only nature of the Twisted\nreactor start/stop calls.\n\nBy moving to HTTP requests the hook is now re-entrant which fixes\nReactorNotRestartable exceptions being raised all the time.\n\nThe requests library is used to make the calls, since it has a\nfunctional TLS implementation (unlike urllib/httplib in Python 2 until\nrecently).\n\nAdditionally, the codebase argument is added to the base ChangeHook, so\nit can properly be used with a codebaseGenerator.\n\nFinally, if a venv argument is used, it is now properly used to activate\nthe environment [1].\n\nPossible improvements to this changeset are:\n- send requests in parallel;\n- switch back to urllib2 so the hook doesn't depend on anything;\n- a few more cleanups.\n\n[1] https://github.com/pypa/virtualenv/blob/master/virtualenv_embedded/activate_this.py\n\n", "before_files": [{"content": "# This file is part of Buildbot. Buildbot is free software: you can\n# redistribute it and/or modify it under the terms of the GNU General Public\n# License as published by the Free Software Foundation, version 2.\n#\n# This program is distributed in the hope that it will be useful, but WITHOUT\n# ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS\n# FOR A PARTICULAR PURPOSE. See the GNU General Public License for more\n# details.\n#\n# You should have received a copy of the GNU General Public License along with\n# this program; if not, write to the Free Software Foundation, Inc., 51\n# Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA.\n#\n# Copyright Buildbot Team Members\n\nfrom email.Utils import formatdate\nimport time\n\nfrom twisted.python import log\nfrom twisted.internet import defer\n\nfrom buildbot.process import buildstep\nfrom buildbot.steps.shell import StringFileWriter\nfrom buildbot.steps.source.base import Source\nfrom buildbot.interfaces import BuildSlaveTooOldError\n\nclass CVS(Source):\n\n name = \"cvs\"\n\n renderables = [ \"cvsroot\" ]\n\n def __init__(self, cvsroot=None, cvsmodule='', mode='incremental',\n method=None, branch=None, global_options=[], extra_options=[],\n login=None, **kwargs):\n\n self.cvsroot = cvsroot\n self.cvsmodule = cvsmodule\n self.branch = branch\n self.global_options = global_options\n self.extra_options = extra_options\n self.login = login\n self.mode = mode\n self.method = method\n self.srcdir = 'source'\n Source.__init__(self, **kwargs)\n\n def startVC(self, branch, revision, patch):\n self.revision = revision\n self.stdio_log = self.addLog(\"stdio\")\n self.method = self._getMethod()\n d = self.checkCvs()\n def checkInstall(cvsInstalled):\n if not cvsInstalled:\n raise BuildSlaveTooOldError(\"CVS is not installed on slave\")\n return 0\n d.addCallback(checkInstall)\n d.addCallback(self.checkLogin)\n\n if self.mode == 'incremental':\n d.addCallback(lambda _: self.incremental())\n elif self.mode == 'full':\n d.addCallback(lambda _: self.full())\n\n d.addCallback(self.parseGotRevision)\n d.addCallback(self.finish)\n d.addErrback(self.failed)\n return d\n\n @defer.inlineCallbacks\n def incremental(self):\n updatable = yield self._sourcedirIsUpdatable()\n if updatable:\n rv = yield self.doUpdate()\n else:\n rv = yield self.clobber()\n defer.returnValue(rv)\n\n @defer.inlineCallbacks\n def full(self):\n if self.method == 'clobber':\n rv = yield self.clobber()\n defer.returnValue(rv)\n return\n\n elif self.method == 'copy':\n rv = yield self.copy()\n defer.returnValue(rv)\n return\n\n updatable = yield self._sourcedirIsUpdatable()\n if not updatable:\n log.msg(\"CVS repo not present, making full checkout\")\n rv = yield self.doCheckout(self.workdir)\n elif self.method == 'clean':\n rv = yield self.clean()\n elif self.method == 'fresh':\n rv = yield self.fresh()\n else:\n raise ValueError(\"Unknown method, check your configuration\")\n defer.returnValue(rv)\n\n def clobber(self):\n cmd = buildstep.RemoteCommand('rmdir', {'dir': self.workdir,\n 'logEnviron': self.logEnviron})\n cmd.useLog(self.stdio_log, False)\n d = self.runCommand(cmd)\n def checkRemoval(res):\n if res != 0:\n raise RuntimeError(\"Failed to delete directory\")\n return res\n d.addCallback(lambda _: checkRemoval(cmd.rc))\n d.addCallback(lambda _: self.doCheckout(self.workdir))\n return d\n\n def fresh(self, ):\n d = self.purge(True)\n d.addCallback(lambda _: self.doUpdate())\n return d\n\n def clean(self, ):\n d = self.purge(False)\n d.addCallback(lambda _: self.doUpdate())\n return d\n\n def copy(self):\n cmd = buildstep.RemoteCommand('rmdir', {'dir': self.workdir,\n 'logEnviron': self.logEnviron})\n cmd.useLog(self.stdio_log, False)\n d = self.runCommand(cmd) \n self.workdir = 'source'\n d.addCallback(lambda _: self.incremental())\n def copy(_):\n cmd = buildstep.RemoteCommand('cpdir',\n {'fromdir': 'source',\n 'todir':'build',\n 'logEnviron': self.logEnviron,})\n cmd.useLog(self.stdio_log, False)\n d = self.runCommand(cmd)\n return d\n d.addCallback(copy)\n def resetWorkdir(_):\n self.workdir = 'build'\n return 0\n d.addCallback(resetWorkdir)\n return d\n \n def purge(self, ignore_ignores):\n command = ['cvsdiscard']\n if ignore_ignores:\n command += ['--ignore']\n cmd = buildstep.RemoteShellCommand(self.workdir, command,\n env=self.env,\n logEnviron=self.logEnviron)\n cmd.useLog(self.stdio_log, False)\n d = self.runCommand(cmd)\n def evaluate(rc):\n if rc != 0:\n raise buildstep.BuildStepFailed()\n return rc\n d.addCallback(lambda _: evaluate(cmd.rc))\n return d\n \n def doCheckout(self, dir):\n command = ['-d', self.cvsroot, '-z3', 'checkout', '-d', dir ]\n command = self.global_options + command + self.extra_options\n if self.branch:\n command += ['-r', self.branch]\n if self.revision:\n command += ['-D', self.revision]\n command += [ self.cvsmodule ]\n d = self._dovccmd(command, '')\n return d\n\n def doUpdate(self):\n command = ['-z3', 'update', '-dP']\n if self.branch:\n command += ['-r', self.branch]\n if self.revision:\n command += ['-D', self.revision]\n d = self._dovccmd(command)\n return d\n\n def finish(self, res):\n d = defer.succeed(res)\n def _gotResults(results):\n self.setStatus(self.cmd, results)\n return results\n d.addCallback(_gotResults)\n d.addCallbacks(self.finished, self.checkDisconnect)\n return d\n\n def checkLogin(self, _):\n if self.login:\n d = defer.succeed(0)\n else:\n d = self._dovccmd(['-d', self.cvsroot, 'login'])\n def setLogin(res):\n # this happens only if the login command succeeds.\n self.login = True\n return res\n d.addCallback(setLogin)\n\n return d\n\n def _dovccmd(self, command, workdir=None):\n if workdir is None:\n workdir = self.workdir\n if not command:\n raise ValueError(\"No command specified\")\n cmd = buildstep.RemoteShellCommand(workdir, ['cvs'] +\n command,\n env=self.env,\n logEnviron=self.logEnviron)\n cmd.useLog(self.stdio_log, False)\n d = self.runCommand(cmd)\n def evaluateCommand(cmd):\n if cmd.rc != 0:\n log.msg(\"Source step failed while running command %s\" % cmd)\n raise buildstep.BuildStepFailed()\n return cmd.rc\n d.addCallback(lambda _: evaluateCommand(cmd))\n return d\n\n @defer.inlineCallbacks\n def _sourcedirIsUpdatable(self):\n myFileWriter = StringFileWriter()\n args = {\n 'workdir': self.build.path_module.join(self.workdir, 'CVS'),\n 'writer': myFileWriter,\n 'maxsize': None,\n 'blocksize': 32*1024,\n }\n\n cmd = buildstep.RemoteCommand('uploadFile',\n dict(slavesrc='Root', **args),\n ignore_updates=True)\n yield self.runCommand(cmd)\n if cmd.rc is not None and cmd.rc != 0:\n defer.returnValue(False)\n return\n if myFileWriter.buffer.strip() != self.cvsroot:\n defer.returnValue(False)\n return\n\n myFileWriter.buffer = \"\"\n cmd = buildstep.RemoteCommand('uploadFile',\n dict(slavesrc='Repository', **args),\n ignore_updates=True)\n yield self.runCommand(cmd)\n if cmd.rc is not None and cmd.rc != 0:\n defer.returnValue(False)\n return\n if myFileWriter.buffer.strip() != self.cvsmodule:\n defer.returnValue(False)\n return\n\n defer.returnValue(True)\n\n def parseGotRevision(self, res):\n revision = time.strftime(\"%Y-%m-%d %H:%M:%S +0000\", time.gmtime())\n self.setProperty('got_revision', revision, 'Source')\n return res\n\n def checkCvs(self):\n d = self._dovccmd(['--version'])\n def check(res):\n if res == 0:\n return True\n return False\n d.addCallback(check)\n return d\n\n def _getMethod(self):\n if self.method is not None and self.mode != 'incremental':\n return self.method\n elif self.mode == 'incremental':\n return None\n elif self.method is None and self.mode == 'full':\n return 'fresh'\n\n def computeSourceRevision(self, changes):\n if not changes:\n return None\n lastChange = max([c.when for c in changes])\n lastSubmit = max([br.submittedAt for br in self.build.requests])\n when = (lastChange + lastSubmit) / 2\n return formatdate(when)\n", "path": "master/buildbot/steps/source/cvs.py"}]}
| 3,708 | 117 |
gh_patches_debug_19831
|
rasdani/github-patches
|
git_diff
|
akvo__akvo-rsr-1882
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Calculate updates from baseline, not from actual
## Test plan
There shouldn't be a difference from the original results framework. No specific testing needed.
## Issue description
Currently, the place of the updates on the slider is calculated by going back from the actual. However, this sometimes results in strange behaviour (e.g. the slider being before the baseline). We should calculate starting from the baseline and add up all the previous updates.
</issue>
<code>
[start of akvo/rsr/models/project_update.py]
1 # -*- coding: utf-8 -*-
2
3 # Akvo RSR is covered by the GNU Affero General Public License.
4 # See more details in the license.txt file located at the root folder of the Akvo RSR module.
5 # For additional details on the GNU license please see < http://www.gnu.org/licenses/agpl.html >.
6
7
8 from datetime import datetime, timedelta
9 from decimal import Decimal, InvalidOperation
10
11 from django.conf import settings
12 from django.core.exceptions import ValidationError
13 from django.db import models
14 from django.utils.translation import ugettext_lazy as _
15
16 from django_counter.models import ViewCounter
17
18 from sorl.thumbnail.fields import ImageField
19 from embed_video.fields import EmbedVideoField
20
21 from akvo.utils import rsr_image_path, to_gmt
22
23 from ..fields import ValidXMLCharField, ValidXMLTextField
24 from ..mixins import TimestampsMixin
25
26
27 def image_path(instance, file_name):
28 """Create a path like 'db/project/<update.project.id>/update/<update.id>/image_name.ext'"""
29 path = 'db/project/%d/update/%%(instance_pk)s/%%(file_name)s' % instance.project.pk
30 return rsr_image_path(instance, file_name, path)
31
32
33 class ProjectUpdate(TimestampsMixin, models.Model):
34 UPDATE_METHODS = (
35 ('W', _(u'web')),
36 ('E', _(u'e-mail')),
37 ('S', _(u'SMS')),
38 ('M', _(u'mobile')),
39 )
40
41 project = models.ForeignKey('Project', related_name='project_updates',
42 verbose_name=_(u'project'))
43 user = models.ForeignKey(settings.AUTH_USER_MODEL, verbose_name=_(u'user'))
44 title = ValidXMLCharField(_(u'title'), max_length=80, db_index=True,
45 help_text=_(u'80 characters'))
46 text = ValidXMLTextField(_(u'text'), blank=True)
47 language = ValidXMLCharField(max_length=2, choices=settings.LANGUAGES, default='en',
48 help_text=_(u'The language of the update'))
49 primary_location = models.ForeignKey('ProjectUpdateLocation', null=True, blank=True,
50 on_delete=models.SET_NULL)
51 photo = ImageField(_(u'photo'), blank=True, upload_to=image_path,
52 help_text=_(u'The image should have 4:3 height:width ratio for best '
53 u'displaying result'))
54 photo_caption = ValidXMLCharField(_(u'photo caption'), blank=True, max_length=75,
55 help_text=_(u'75 characters'))
56 photo_credit = ValidXMLCharField(_(u'photo credit'), blank=True, max_length=75,
57 help_text=_(u'75 characters'))
58 video = EmbedVideoField(_(u'video URL'), blank=True,
59 help_text=_(u'Supported providers: YouTube and Vimeo'))
60 video_caption = ValidXMLCharField(_(u'video caption'), blank=True, max_length=75,
61 help_text=_(u'75 characters'))
62 video_credit = ValidXMLCharField(_(u'video credit'), blank=True, max_length=75,
63 help_text=_(u'75 characters'))
64 update_method = ValidXMLCharField(_(u'update method'), blank=True, max_length=1,
65 choices=UPDATE_METHODS, db_index=True, default='W')
66 user_agent = ValidXMLCharField(_(u'user agent'), blank=True, max_length=200, default='')
67 uuid = ValidXMLCharField(_(u'uuid'), blank=True, max_length=40, default='', db_index=True,
68 help_text=_(u'Universally unique ID set by creating user agent'))
69 notes = ValidXMLTextField(verbose_name=_(u"Notes and comments"), blank=True, default='')
70
71 # Indicator updates
72 indicator_period = models.ForeignKey('IndicatorPeriod', related_name='updates',
73 verbose_name=_(u'indicator period'), blank=True, null=True)
74 period_update = models.DecimalField(_(u'period update'), blank=True, null=True, max_digits=14,
75 decimal_places=2)
76
77 class Meta:
78 app_label = 'rsr'
79 get_latest_by = "created_at"
80 verbose_name = _(u'project update')
81 verbose_name_plural = _(u'project updates')
82 ordering = ['-id', ]
83
84 def save(self, *args, **kwargs):
85 if self.indicator_period and self.period_update:
86 if not self.pk:
87 # Newly created update to indicator period, update the actual value.
88 self.indicator_period.update_actual_value(self.period_update)
89
90 else:
91 # Update to already existing indicator period, check if values have been changed.
92 orig_update = ProjectUpdate.objects.get(pk=self.pk)
93 if orig_update.indicator_period != self.indicator_period:
94 # Indicator period has changed. Substract value from old period, and add new
95 # value to new period.
96 try:
97 orig_update.update_actual_value(Decimal(orig_update.period_update) * -1)
98 except (InvalidOperation, TypeError):
99 pass
100 self.indicator_period.update_actual_value(self.period_update)
101
102 elif orig_update.period_update != self.period_update:
103 # Indicator value has changed. Add the difference to it.
104 try:
105 self.indicator_period.update_actual_value(
106 Decimal(self.period_update) - Decimal(orig_update.period_update)
107 )
108 except (InvalidOperation, TypeError):
109 self.indicator_period.update_actual_value(self.period_update)
110
111 super(ProjectUpdate, self).save(*args, **kwargs)
112
113 def delete(self, *args, **kwargs):
114 if self.indicator_period and self.period_update:
115 try:
116 self.indicator_period.update_actual_value(
117 Decimal(self.period_update) * -1
118 )
119 except (InvalidOperation, TypeError):
120 pass
121 super(ProjectUpdate, self).delete(*args, **kwargs)
122
123 def clean(self):
124 if hasattr(self, 'project') and hasattr(self, 'indicator_period') and \
125 hasattr(self, 'period_update'):
126 validation_errors = {}
127
128 # Don't allow an indicator period that belongs to a different project
129 if self.project and self.indicator_period:
130 if not self.indicator_period.indicator.result.project == self.project:
131 validation_errors['indicator_period'] = u'%s' % _(
132 u'Indicator period must be part of the same project'
133 )
134
135 # Don't allow an indicator update to a non-Impact project
136 if self.indicator_period and self.period_update and not self.project.is_impact_project:
137 validation_errors['project'] = u'%s' % _(
138 u'Project must be an Impact project to place indicator updates to it'
139 )
140
141 if validation_errors:
142 raise ValidationError(validation_errors)
143
144 def img(self, value=''):
145 try:
146 return self.photo.thumbnail_tag
147 except:
148 return value
149 img.allow_tags = True
150
151 def edit_window_has_expired(self):
152 """Determine whether or not update timeout window has expired.
153 The timeout is controlled by settings.PROJECT_UPDATE_TIMEOUT and
154 defaults to 30 minutes.
155 """
156 return (datetime.now() - self.created_at) > self.edit_timeout
157
158 @property
159 def expires_at(self):
160 return to_gmt(self.created_at + self.edit_timeout)
161
162 @property
163 def edit_timeout(self):
164 timeout_minutes = getattr(settings, 'PROJECT_UPDATE_TIMEOUT', 30)
165 return timedelta(minutes=timeout_minutes)
166
167 @property
168 def edit_time_remaining(self):
169 return self.edit_timeout - self.created_at
170
171 @property
172 def time_gmt(self):
173 return to_gmt(self.created_at)
174
175 @property
176 def time_last_updated_gmt(self):
177 return to_gmt(self.last_modified_at)
178
179 @property
180 def view_count(self):
181 counter = ViewCounter.objects.get_for_object(self)
182 return counter.count or 0
183
184 @models.permalink
185 def get_absolute_url(self):
186 return 'update-main', (), {'project_id': self.project.pk, 'update_id': self.pk}
187
188 def __unicode__(self):
189 return _(u'Project update for %(project_name)s') % {'project_name': self.project.title}
190
[end of akvo/rsr/models/project_update.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/akvo/rsr/models/project_update.py b/akvo/rsr/models/project_update.py
--- a/akvo/rsr/models/project_update.py
+++ b/akvo/rsr/models/project_update.py
@@ -112,12 +112,19 @@
def delete(self, *args, **kwargs):
if self.indicator_period and self.period_update:
- try:
- self.indicator_period.update_actual_value(
- Decimal(self.period_update) * -1
- )
- except (InvalidOperation, TypeError):
- pass
+ # Subsctract the value of the update from the actual value of the indicator period
+ if ProjectUpdate.objects.filter(indicator_period=self.indicator_period).\
+ exclude(pk=self.pk).exists():
+ try:
+ self.indicator_period.update_actual_value(
+ Decimal(self.period_update) * -1
+ )
+ except (InvalidOperation, TypeError):
+ pass
+ else:
+ # There's no other updates for this indicator period, remove actual value
+ self.indicator_period.actual_value = ''
+ self.indicator_period.save()
super(ProjectUpdate, self).delete(*args, **kwargs)
def clean(self):
|
{"golden_diff": "diff --git a/akvo/rsr/models/project_update.py b/akvo/rsr/models/project_update.py\n--- a/akvo/rsr/models/project_update.py\n+++ b/akvo/rsr/models/project_update.py\n@@ -112,12 +112,19 @@\n \n def delete(self, *args, **kwargs):\n if self.indicator_period and self.period_update:\n- try:\n- self.indicator_period.update_actual_value(\n- Decimal(self.period_update) * -1\n- )\n- except (InvalidOperation, TypeError):\n- pass\n+ # Subsctract the value of the update from the actual value of the indicator period\n+ if ProjectUpdate.objects.filter(indicator_period=self.indicator_period).\\\n+ exclude(pk=self.pk).exists():\n+ try:\n+ self.indicator_period.update_actual_value(\n+ Decimal(self.period_update) * -1\n+ )\n+ except (InvalidOperation, TypeError):\n+ pass\n+ else:\n+ # There's no other updates for this indicator period, remove actual value\n+ self.indicator_period.actual_value = ''\n+ self.indicator_period.save()\n super(ProjectUpdate, self).delete(*args, **kwargs)\n \n def clean(self):\n", "issue": "Calculate updates from baseline, not from actual\n## Test plan\n\nThere shouldn't be a difference from the original results framework. No specific testing needed.\n## Issue description\n\nCurrently, the place of the updates on the slider is calculated by going back from the actual. However, this sometimes results in strange behaviour (e.g. the slider being before the baseline). We should calculate starting from the baseline and add up all the previous updates.\n\n", "before_files": [{"content": "# -*- coding: utf-8 -*-\n\n# Akvo RSR is covered by the GNU Affero General Public License.\n# See more details in the license.txt file located at the root folder of the Akvo RSR module.\n# For additional details on the GNU license please see < http://www.gnu.org/licenses/agpl.html >.\n\n\nfrom datetime import datetime, timedelta\nfrom decimal import Decimal, InvalidOperation\n\nfrom django.conf import settings\nfrom django.core.exceptions import ValidationError\nfrom django.db import models\nfrom django.utils.translation import ugettext_lazy as _\n\nfrom django_counter.models import ViewCounter\n\nfrom sorl.thumbnail.fields import ImageField\nfrom embed_video.fields import EmbedVideoField\n\nfrom akvo.utils import rsr_image_path, to_gmt\n\nfrom ..fields import ValidXMLCharField, ValidXMLTextField\nfrom ..mixins import TimestampsMixin\n\n\ndef image_path(instance, file_name):\n \"\"\"Create a path like 'db/project/<update.project.id>/update/<update.id>/image_name.ext'\"\"\"\n path = 'db/project/%d/update/%%(instance_pk)s/%%(file_name)s' % instance.project.pk\n return rsr_image_path(instance, file_name, path)\n\n\nclass ProjectUpdate(TimestampsMixin, models.Model):\n UPDATE_METHODS = (\n ('W', _(u'web')),\n ('E', _(u'e-mail')),\n ('S', _(u'SMS')),\n ('M', _(u'mobile')),\n )\n\n project = models.ForeignKey('Project', related_name='project_updates',\n verbose_name=_(u'project'))\n user = models.ForeignKey(settings.AUTH_USER_MODEL, verbose_name=_(u'user'))\n title = ValidXMLCharField(_(u'title'), max_length=80, db_index=True,\n help_text=_(u'80 characters'))\n text = ValidXMLTextField(_(u'text'), blank=True)\n language = ValidXMLCharField(max_length=2, choices=settings.LANGUAGES, default='en',\n help_text=_(u'The language of the update'))\n primary_location = models.ForeignKey('ProjectUpdateLocation', null=True, blank=True,\n on_delete=models.SET_NULL)\n photo = ImageField(_(u'photo'), blank=True, upload_to=image_path,\n help_text=_(u'The image should have 4:3 height:width ratio for best '\n u'displaying result'))\n photo_caption = ValidXMLCharField(_(u'photo caption'), blank=True, max_length=75,\n help_text=_(u'75 characters'))\n photo_credit = ValidXMLCharField(_(u'photo credit'), blank=True, max_length=75,\n help_text=_(u'75 characters'))\n video = EmbedVideoField(_(u'video URL'), blank=True,\n help_text=_(u'Supported providers: YouTube and Vimeo'))\n video_caption = ValidXMLCharField(_(u'video caption'), blank=True, max_length=75,\n help_text=_(u'75 characters'))\n video_credit = ValidXMLCharField(_(u'video credit'), blank=True, max_length=75,\n help_text=_(u'75 characters'))\n update_method = ValidXMLCharField(_(u'update method'), blank=True, max_length=1,\n choices=UPDATE_METHODS, db_index=True, default='W')\n user_agent = ValidXMLCharField(_(u'user agent'), blank=True, max_length=200, default='')\n uuid = ValidXMLCharField(_(u'uuid'), blank=True, max_length=40, default='', db_index=True,\n help_text=_(u'Universally unique ID set by creating user agent'))\n notes = ValidXMLTextField(verbose_name=_(u\"Notes and comments\"), blank=True, default='')\n\n # Indicator updates\n indicator_period = models.ForeignKey('IndicatorPeriod', related_name='updates',\n verbose_name=_(u'indicator period'), blank=True, null=True)\n period_update = models.DecimalField(_(u'period update'), blank=True, null=True, max_digits=14,\n decimal_places=2)\n\n class Meta:\n app_label = 'rsr'\n get_latest_by = \"created_at\"\n verbose_name = _(u'project update')\n verbose_name_plural = _(u'project updates')\n ordering = ['-id', ]\n\n def save(self, *args, **kwargs):\n if self.indicator_period and self.period_update:\n if not self.pk:\n # Newly created update to indicator period, update the actual value.\n self.indicator_period.update_actual_value(self.period_update)\n\n else:\n # Update to already existing indicator period, check if values have been changed.\n orig_update = ProjectUpdate.objects.get(pk=self.pk)\n if orig_update.indicator_period != self.indicator_period:\n # Indicator period has changed. Substract value from old period, and add new\n # value to new period.\n try:\n orig_update.update_actual_value(Decimal(orig_update.period_update) * -1)\n except (InvalidOperation, TypeError):\n pass\n self.indicator_period.update_actual_value(self.period_update)\n\n elif orig_update.period_update != self.period_update:\n # Indicator value has changed. Add the difference to it.\n try:\n self.indicator_period.update_actual_value(\n Decimal(self.period_update) - Decimal(orig_update.period_update)\n )\n except (InvalidOperation, TypeError):\n self.indicator_period.update_actual_value(self.period_update)\n\n super(ProjectUpdate, self).save(*args, **kwargs)\n\n def delete(self, *args, **kwargs):\n if self.indicator_period and self.period_update:\n try:\n self.indicator_period.update_actual_value(\n Decimal(self.period_update) * -1\n )\n except (InvalidOperation, TypeError):\n pass\n super(ProjectUpdate, self).delete(*args, **kwargs)\n\n def clean(self):\n if hasattr(self, 'project') and hasattr(self, 'indicator_period') and \\\n hasattr(self, 'period_update'):\n validation_errors = {}\n\n # Don't allow an indicator period that belongs to a different project\n if self.project and self.indicator_period:\n if not self.indicator_period.indicator.result.project == self.project:\n validation_errors['indicator_period'] = u'%s' % _(\n u'Indicator period must be part of the same project'\n )\n\n # Don't allow an indicator update to a non-Impact project\n if self.indicator_period and self.period_update and not self.project.is_impact_project:\n validation_errors['project'] = u'%s' % _(\n u'Project must be an Impact project to place indicator updates to it'\n )\n\n if validation_errors:\n raise ValidationError(validation_errors)\n\n def img(self, value=''):\n try:\n return self.photo.thumbnail_tag\n except:\n return value\n img.allow_tags = True\n\n def edit_window_has_expired(self):\n \"\"\"Determine whether or not update timeout window has expired.\n The timeout is controlled by settings.PROJECT_UPDATE_TIMEOUT and\n defaults to 30 minutes.\n \"\"\"\n return (datetime.now() - self.created_at) > self.edit_timeout\n\n @property\n def expires_at(self):\n return to_gmt(self.created_at + self.edit_timeout)\n\n @property\n def edit_timeout(self):\n timeout_minutes = getattr(settings, 'PROJECT_UPDATE_TIMEOUT', 30)\n return timedelta(minutes=timeout_minutes)\n\n @property\n def edit_time_remaining(self):\n return self.edit_timeout - self.created_at\n\n @property\n def time_gmt(self):\n return to_gmt(self.created_at)\n\n @property\n def time_last_updated_gmt(self):\n return to_gmt(self.last_modified_at)\n\n @property\n def view_count(self):\n counter = ViewCounter.objects.get_for_object(self)\n return counter.count or 0\n\n @models.permalink\n def get_absolute_url(self):\n return 'update-main', (), {'project_id': self.project.pk, 'update_id': self.pk}\n\n def __unicode__(self):\n return _(u'Project update for %(project_name)s') % {'project_name': self.project.title}\n", "path": "akvo/rsr/models/project_update.py"}]}
| 2,848 | 274 |
gh_patches_debug_1714
|
rasdani/github-patches
|
git_diff
|
networkx__networkx-1544
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Added implementation of SimRank matrix of the graph.
</issue>
<code>
[start of networkx/relabel.py]
1 # Copyright (C) 2006-2013 by
2 # Aric Hagberg <[email protected]>
3 # Dan Schult <[email protected]>
4 # Pieter Swart <[email protected]>
5 # All rights reserved.
6 # BSD license.
7 import networkx as nx
8 __author__ = """\n""".join(['Aric Hagberg <[email protected]>',
9 'Pieter Swart ([email protected])',
10 'Dan Schult ([email protected])'])
11 __all__ = ['convert_node_labels_to_integers', 'relabel_nodes']
12
13 def relabel_nodes(G, mapping, copy=True):
14 """Relabel the nodes of the graph G.
15
16 Parameters
17 ----------
18 G : graph
19 A NetworkX graph
20
21 mapping : dictionary
22 A dictionary with the old labels as keys and new labels as values.
23 A partial mapping is allowed.
24
25 copy : bool (optional, default=True)
26 If True return a copy, or if False relabel the nodes in place.
27
28 Examples
29 --------
30 >>> G=nx.path_graph(3) # nodes 0-1-2
31 >>> mapping={0:'a',1:'b',2:'c'}
32 >>> H=nx.relabel_nodes(G,mapping)
33 >>> print(sorted(H.nodes()))
34 ['a', 'b', 'c']
35
36 >>> G=nx.path_graph(26) # nodes 0..25
37 >>> mapping=dict(zip(G.nodes(),"abcdefghijklmnopqrstuvwxyz"))
38 >>> H=nx.relabel_nodes(G,mapping) # nodes a..z
39 >>> mapping=dict(zip(G.nodes(),range(1,27)))
40 >>> G1=nx.relabel_nodes(G,mapping) # nodes 1..26
41
42 Partial in-place mapping:
43
44 >>> G=nx.path_graph(3) # nodes 0-1-2
45 >>> mapping={0:'a',1:'b'} # 0->'a' and 1->'b'
46 >>> G=nx.relabel_nodes(G,mapping, copy=False)
47
48 print(G.nodes())
49 [2, 'b', 'a']
50
51 Mapping as function:
52
53 >>> G=nx.path_graph(3)
54 >>> def mapping(x):
55 ... return x**2
56 >>> H=nx.relabel_nodes(G,mapping)
57 >>> print(H.nodes())
58 [0, 1, 4]
59
60 Notes
61 -----
62 Only the nodes specified in the mapping will be relabeled.
63
64 The keyword setting copy=False modifies the graph in place.
65 This is not always possible if the mapping is circular.
66 In that case use copy=True.
67
68 See Also
69 --------
70 convert_node_labels_to_integers
71 """
72 # you can pass a function f(old_label)->new_label
73 # but we'll just make a dictionary here regardless
74 if not hasattr(mapping,"__getitem__"):
75 m = dict((n, mapping(n)) for n in G)
76 else:
77 m = mapping
78 if copy:
79 return _relabel_copy(G, m)
80 else:
81 return _relabel_inplace(G, m)
82
83
84 def _relabel_inplace(G, mapping):
85 old_labels = set(mapping.keys())
86 new_labels = set(mapping.values())
87 if len(old_labels & new_labels) > 0:
88 # labels sets overlap
89 # can we topological sort and still do the relabeling?
90 D = nx.DiGraph(list(mapping.items()))
91 D.remove_edges_from(D.selfloop_edges())
92 try:
93 nodes = nx.topological_sort(D, reverse=True)
94 except nx.NetworkXUnfeasible:
95 raise nx.NetworkXUnfeasible('The node label sets are overlapping '
96 'and no ordering can resolve the '
97 'mapping. Use copy=True.')
98 else:
99 # non-overlapping label sets
100 nodes = old_labels
101
102 multigraph = G.is_multigraph()
103 directed = G.is_directed()
104
105 for old in nodes:
106 try:
107 new = mapping[old]
108 except KeyError:
109 continue
110 try:
111 G.add_node(new, attr_dict=G.node[old])
112 except KeyError:
113 raise KeyError("Node %s is not in the graph"%old)
114 if multigraph:
115 new_edges = [(new, new if old == target else target, key, data)
116 for (_,target,key,data)
117 in G.edges(old, data=True, keys=True)]
118 if directed:
119 new_edges += [(new if old == source else source, new, key, data)
120 for (source, _, key,data)
121 in G.in_edges(old, data=True, keys=True)]
122 else:
123 new_edges = [(new, new if old == target else target, data)
124 for (_,target,data) in G.edges(old, data=True)]
125 if directed:
126 new_edges += [(new if old == source else source,new,data)
127 for (source,_,data) in G.in_edges(old, data=True)]
128 G.remove_node(old)
129 G.add_edges_from(new_edges)
130 return G
131
132 def _relabel_copy(G, mapping):
133 H = G.__class__()
134 H.name = "(%s)" % G.name
135 if G.is_multigraph():
136 H.add_edges_from( (mapping.get(n1, n1),mapping.get(n2, n2),k,d.copy())
137 for (n1,n2,k,d) in G.edges_iter(keys=True, data=True))
138 else:
139 H.add_edges_from( (mapping.get(n1, n1),mapping.get(n2, n2),d.copy())
140 for (n1, n2, d) in G.edges_iter(data=True))
141
142 H.add_nodes_from(mapping.get(n, n) for n in G)
143 H.node.update(dict((mapping.get(n, n), d.copy()) for n,d in G.node.items()))
144 H.graph.update(G.graph.copy())
145
146 return H
147
148
149 def convert_node_labels_to_integers(G, first_label=0, ordering="default",
150 label_attribute=None):
151 """Return a copy of the graph G with the nodes relabeled using
152 consecutive integers.
153
154 Parameters
155 ----------
156 G : graph
157 A NetworkX graph
158
159 first_label : int, optional (default=0)
160 An integer specifying the starting offset in numbering nodes.
161 The new integer labels are numbered first_label, ..., n-1+first_label.
162
163 ordering : string
164 "default" : inherit node ordering from G.nodes()
165 "sorted" : inherit node ordering from sorted(G.nodes())
166 "increasing degree" : nodes are sorted by increasing degree
167 "decreasing degree" : nodes are sorted by decreasing degree
168
169 label_attribute : string, optional (default=None)
170 Name of node attribute to store old label. If None no attribute
171 is created.
172
173 Notes
174 -----
175 Node and edge attribute data are copied to the new (relabeled) graph.
176
177 See Also
178 --------
179 relabel_nodes
180 """
181 N = G.number_of_nodes()+first_label
182 if ordering == "default":
183 mapping = dict(zip(G.nodes(), range(first_label, N)))
184 elif ordering == "sorted":
185 nlist = G.nodes()
186 nlist.sort()
187 mapping = dict(zip(nlist, range(first_label, N)))
188 elif ordering == "increasing degree":
189 dv_pairs = [(d,n) for (n,d) in G.degree_iter()]
190 dv_pairs.sort() # in-place sort from lowest to highest degree
191 mapping = dict(zip([n for d,n in dv_pairs], range(first_label, N)))
192 elif ordering == "decreasing degree":
193 dv_pairs = [(d,n) for (n,d) in G.degree_iter()]
194 dv_pairs.sort() # in-place sort from lowest to highest degree
195 dv_pairs.reverse()
196 mapping = dict(zip([n for d,n in dv_pairs], range(first_label, N)))
197 else:
198 raise nx.NetworkXError('Unknown node ordering: %s'%ordering)
199 H = relabel_nodes(G, mapping)
200 H.name = "("+G.name+")_with_int_labels"
201 # create node attribute with the old label
202 if label_attribute is not None:
203 nx.set_node_attributes(H, label_attribute,
204 dict((v,k) for k,v in mapping.items()))
205 return H
206
[end of networkx/relabel.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/networkx/relabel.py b/networkx/relabel.py
--- a/networkx/relabel.py
+++ b/networkx/relabel.py
@@ -107,6 +107,8 @@
new = mapping[old]
except KeyError:
continue
+ if new == old:
+ continue
try:
G.add_node(new, attr_dict=G.node[old])
except KeyError:
|
{"golden_diff": "diff --git a/networkx/relabel.py b/networkx/relabel.py\n--- a/networkx/relabel.py\n+++ b/networkx/relabel.py\n@@ -107,6 +107,8 @@\n new = mapping[old]\n except KeyError:\n continue\n+ if new == old:\n+ continue\n try:\n G.add_node(new, attr_dict=G.node[old])\n except KeyError:\n", "issue": "Added implementation of SimRank matrix of the graph.\n\n", "before_files": [{"content": "# Copyright (C) 2006-2013 by\n# Aric Hagberg <[email protected]>\n# Dan Schult <[email protected]>\n# Pieter Swart <[email protected]>\n# All rights reserved.\n# BSD license.\nimport networkx as nx\n__author__ = \"\"\"\\n\"\"\".join(['Aric Hagberg <[email protected]>',\n 'Pieter Swart ([email protected])',\n 'Dan Schult ([email protected])'])\n__all__ = ['convert_node_labels_to_integers', 'relabel_nodes']\n\ndef relabel_nodes(G, mapping, copy=True):\n \"\"\"Relabel the nodes of the graph G.\n\n Parameters\n ----------\n G : graph\n A NetworkX graph\n\n mapping : dictionary\n A dictionary with the old labels as keys and new labels as values.\n A partial mapping is allowed.\n\n copy : bool (optional, default=True)\n If True return a copy, or if False relabel the nodes in place.\n\n Examples\n --------\n >>> G=nx.path_graph(3) # nodes 0-1-2\n >>> mapping={0:'a',1:'b',2:'c'}\n >>> H=nx.relabel_nodes(G,mapping)\n >>> print(sorted(H.nodes()))\n ['a', 'b', 'c']\n\n >>> G=nx.path_graph(26) # nodes 0..25\n >>> mapping=dict(zip(G.nodes(),\"abcdefghijklmnopqrstuvwxyz\"))\n >>> H=nx.relabel_nodes(G,mapping) # nodes a..z\n >>> mapping=dict(zip(G.nodes(),range(1,27)))\n >>> G1=nx.relabel_nodes(G,mapping) # nodes 1..26\n\n Partial in-place mapping:\n\n >>> G=nx.path_graph(3) # nodes 0-1-2\n >>> mapping={0:'a',1:'b'} # 0->'a' and 1->'b'\n >>> G=nx.relabel_nodes(G,mapping, copy=False)\n\n print(G.nodes())\n [2, 'b', 'a']\n\n Mapping as function:\n\n >>> G=nx.path_graph(3)\n >>> def mapping(x):\n ... return x**2\n >>> H=nx.relabel_nodes(G,mapping)\n >>> print(H.nodes())\n [0, 1, 4]\n\n Notes\n -----\n Only the nodes specified in the mapping will be relabeled.\n\n The keyword setting copy=False modifies the graph in place.\n This is not always possible if the mapping is circular.\n In that case use copy=True.\n\n See Also\n --------\n convert_node_labels_to_integers\n \"\"\"\n # you can pass a function f(old_label)->new_label\n # but we'll just make a dictionary here regardless\n if not hasattr(mapping,\"__getitem__\"):\n m = dict((n, mapping(n)) for n in G)\n else:\n m = mapping\n if copy:\n return _relabel_copy(G, m)\n else:\n return _relabel_inplace(G, m)\n\n\ndef _relabel_inplace(G, mapping):\n old_labels = set(mapping.keys())\n new_labels = set(mapping.values())\n if len(old_labels & new_labels) > 0:\n # labels sets overlap\n # can we topological sort and still do the relabeling?\n D = nx.DiGraph(list(mapping.items()))\n D.remove_edges_from(D.selfloop_edges())\n try:\n nodes = nx.topological_sort(D, reverse=True)\n except nx.NetworkXUnfeasible:\n raise nx.NetworkXUnfeasible('The node label sets are overlapping '\n 'and no ordering can resolve the '\n 'mapping. Use copy=True.')\n else:\n # non-overlapping label sets\n nodes = old_labels\n\n multigraph = G.is_multigraph()\n directed = G.is_directed()\n\n for old in nodes:\n try:\n new = mapping[old]\n except KeyError:\n continue\n try:\n G.add_node(new, attr_dict=G.node[old])\n except KeyError:\n raise KeyError(\"Node %s is not in the graph\"%old)\n if multigraph:\n new_edges = [(new, new if old == target else target, key, data)\n for (_,target,key,data)\n in G.edges(old, data=True, keys=True)]\n if directed:\n new_edges += [(new if old == source else source, new, key, data)\n for (source, _, key,data)\n in G.in_edges(old, data=True, keys=True)]\n else:\n new_edges = [(new, new if old == target else target, data)\n for (_,target,data) in G.edges(old, data=True)]\n if directed:\n new_edges += [(new if old == source else source,new,data)\n for (source,_,data) in G.in_edges(old, data=True)]\n G.remove_node(old)\n G.add_edges_from(new_edges)\n return G\n\ndef _relabel_copy(G, mapping):\n H = G.__class__()\n H.name = \"(%s)\" % G.name\n if G.is_multigraph():\n H.add_edges_from( (mapping.get(n1, n1),mapping.get(n2, n2),k,d.copy())\n for (n1,n2,k,d) in G.edges_iter(keys=True, data=True))\n else:\n H.add_edges_from( (mapping.get(n1, n1),mapping.get(n2, n2),d.copy())\n for (n1, n2, d) in G.edges_iter(data=True))\n\n H.add_nodes_from(mapping.get(n, n) for n in G)\n H.node.update(dict((mapping.get(n, n), d.copy()) for n,d in G.node.items()))\n H.graph.update(G.graph.copy())\n\n return H\n\n\ndef convert_node_labels_to_integers(G, first_label=0, ordering=\"default\",\n label_attribute=None):\n \"\"\"Return a copy of the graph G with the nodes relabeled using\n consecutive integers.\n\n Parameters\n ----------\n G : graph\n A NetworkX graph\n\n first_label : int, optional (default=0)\n An integer specifying the starting offset in numbering nodes.\n The new integer labels are numbered first_label, ..., n-1+first_label.\n\n ordering : string\n \"default\" : inherit node ordering from G.nodes()\n \"sorted\" : inherit node ordering from sorted(G.nodes())\n \"increasing degree\" : nodes are sorted by increasing degree\n \"decreasing degree\" : nodes are sorted by decreasing degree\n\n label_attribute : string, optional (default=None)\n Name of node attribute to store old label. If None no attribute\n is created.\n\n Notes\n -----\n Node and edge attribute data are copied to the new (relabeled) graph.\n\n See Also\n --------\n relabel_nodes\n \"\"\"\n N = G.number_of_nodes()+first_label\n if ordering == \"default\":\n mapping = dict(zip(G.nodes(), range(first_label, N)))\n elif ordering == \"sorted\":\n nlist = G.nodes()\n nlist.sort()\n mapping = dict(zip(nlist, range(first_label, N)))\n elif ordering == \"increasing degree\":\n dv_pairs = [(d,n) for (n,d) in G.degree_iter()]\n dv_pairs.sort() # in-place sort from lowest to highest degree\n mapping = dict(zip([n for d,n in dv_pairs], range(first_label, N)))\n elif ordering == \"decreasing degree\":\n dv_pairs = [(d,n) for (n,d) in G.degree_iter()]\n dv_pairs.sort() # in-place sort from lowest to highest degree\n dv_pairs.reverse()\n mapping = dict(zip([n for d,n in dv_pairs], range(first_label, N)))\n else:\n raise nx.NetworkXError('Unknown node ordering: %s'%ordering)\n H = relabel_nodes(G, mapping)\n H.name = \"(\"+G.name+\")_with_int_labels\"\n # create node attribute with the old label\n if label_attribute is not None:\n nx.set_node_attributes(H, label_attribute,\n dict((v,k) for k,v in mapping.items()))\n return H\n", "path": "networkx/relabel.py"}]}
| 2,860 | 91 |
gh_patches_debug_15337
|
rasdani/github-patches
|
git_diff
|
weecology__retriever-952
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
install to msaccess error
When I try to install a dataset to msaccess I get the error:
Couldn't create table (no module named 'pypyodbc')
I've tried doing "pip install pypyodbc --upgrade" which seems to work, but doesn't help with the retriever error.
I'm on a windows 10 machine
</issue>
<code>
[start of retriever/engines/msaccess.py]
1 from __future__ import print_function
2 from builtins import str
3 import os
4 from retriever.lib.models import Engine, no_cleanup
5 from retriever import DATA_DIR, current_platform
6
7
8 class engine(Engine):
9 """Engine instance for Microsoft Access."""
10 name = "Microsoft Access"
11 instructions = """Create a database in Microsoft Access, close Access, then \nselect your database file using this dialog."""
12 abbreviation = "msaccess"
13 datatypes = {
14 "auto": "AUTOINCREMENT",
15 "int": "INTEGER",
16 "bigint": "INTEGER",
17 "double": "NUMERIC",
18 "decimal": "NUMERIC",
19 "char": "VARCHAR",
20 "bool": "BIT",
21 }
22 required_opts = [("file",
23 "Enter the filename of your Access database",
24 os.path.join(DATA_DIR, "access.mdb"),
25 "Access databases (*.mdb, *.accdb)|*.mdb;*.accdb"),
26 ("table_name",
27 "Format of table name",
28 "[{db} {table}]"),
29 ]
30
31 def convert_data_type(self, datatype):
32 """MS Access can't handle complex Decimal types"""
33 converted = Engine.convert_data_type(self, datatype)
34 if "NUMERIC" in converted:
35 converted = "NUMERIC"
36 elif "VARCHAR" in converted:
37 try:
38 length = int(converted.split('(')[1].split(')')[0].split(',')[0])
39 if length > 255:
40 converted = "TEXT"
41 except:
42 pass
43 return converted
44
45 def create_db(self):
46 """MS Access doesn't create databases."""
47 return None
48
49 def drop_statement(self, objecttype, objectname):
50 """Returns a drop table or database SQL statement."""
51 dropstatement = "DROP %s %s" % (objecttype, objectname)
52 return dropstatement
53
54 def insert_data_from_file(self, filename):
55 """Perform a bulk insert."""
56 self.get_cursor()
57 ct = len([True for c in self.table.columns if c[1][0][:3] == "ct-"]) != 0
58 if ((self.table.cleanup.function == no_cleanup and not self.table.fixed_width and
59 self.table.header_rows < 2)
60 and (self.table.delimiter in ["\t", ","])
61 and not ct
62 and (not hasattr(self.table, "do_not_bulk_insert") or not self.table.do_not_bulk_insert)
63 ):
64 print ("Inserting data from " + os.path.basename(filename) + "...")
65
66 if self.table.delimiter == "\t":
67 fmt = "TabDelimited"
68 elif self.table.delimiter == ",":
69 fmt = "CSVDelimited"
70
71 if self.table.header_rows == 1:
72 hdr = "Yes"
73 else:
74 hdr = "No"
75
76 columns = self.table.get_insert_columns()
77
78 need_to_delete = False
79 add_to_record_id = 0
80
81 if self.table.pk and not self.table.contains_pk:
82 if '.' in os.path.basename(filename):
83 proper_name = filename.split('.')
84 newfilename = '.'.join((proper_name[0:-1]) if len(proper_name) > 0 else proper_name[0]
85 ) + "_new." + filename.split(".")[-1]
86 else:
87 newfilename = filename + "_new"
88
89 if not os.path.isfile(newfilename):
90 print("Adding index to " + os.path.abspath(newfilename) + "...")
91 read = open(filename, "rb")
92 write = open(newfilename, "wb")
93 to_write = ""
94
95 for line in read:
96 to_write += str(id) + self.table.delimiter + line.replace("\n", "\r\n")
97 add_to_record_id += 1
98 self.table.record_id += add_to_record_id
99
100 write.write(to_write)
101 write.close()
102 read.close()
103 need_to_delete = True
104 columns = "record_id, " + columns
105 else:
106 newfilename = filename
107
108 newfilename = os.path.abspath(newfilename)
109 filename_length = (len(os.path.basename(newfilename)) * -1) - 1
110 filepath = newfilename[:filename_length]
111 statement = """
112 INSERT INTO """ + self.table_name() + " (" + columns + """)
113 SELECT * FROM [""" + os.path.basename(newfilename) + ''']
114 IN "''' + filepath + '''" "Text;FMT=''' + fmt + ''';HDR=''' + hdr + ''';"'''
115
116 try:
117 self.execute(statement)
118 except:
119 print("Couldn't bulk insert. Trying manual insert.")
120 self.connection.rollback()
121
122 self.table.record_id -= add_to_record_id
123
124 return Engine.insert_data_from_file(self, filename)
125
126 if need_to_delete:
127 os.remove(newfilename)
128
129 else:
130 return Engine.insert_data_from_file(self, filename)
131
132 def table_exists(self, dbname, tablename):
133 """Determine if the table already exists in the database"""
134 if not hasattr(self, 'existing_table_names'):
135 self.existing_table_names = set()
136 for row in self.cursor.tables():
137 tableinfo = row[2]
138 if not tableinfo.startswith("MSys"):
139 # ignore system tables
140 database, table = tableinfo.split()
141 self.existing_table_names.add((database, table))
142 return self.table_name(name=tablename, dbname=dbname).lower() in self.existing_table_names
143
144 def get_connection(self):
145 """Gets the db connection."""
146 if current_platform != "windows":
147 raise Exception("MS Access can only be used in Windows.")
148 import pypyodbc as dbapi
149 self.get_input()
150 if not os.path.exists(self.opts['file']) and self.opts['file'].endswith('.mdb'):
151 dbapi.win_create_mdb(self.opts['file'])
152 connection_string = ("DRIVER={Microsoft Access Driver (*.mdb, *.accdb)};DBQ=" +
153 os.path.abspath(self.opts["file"]).replace("/", "//") + ";")
154 return dbapi.connect(connection_string, autocommit=False)
155
[end of retriever/engines/msaccess.py]
[start of setup.py]
1 """Use the following command to install retriever: python setup.py install"""
2 from __future__ import absolute_import
3 import platform
4 import os
5 from setuptools import setup
6 from pkg_resources import parse_version
7
8 current_platform = platform.system().lower()
9 extra_includes = []
10
11 if os.path.exists(".git/hooks"): # check if we are in git repo
12 os.system("cp hooks/pre-commit .git/hooks/pre-commit")
13 os.system("chmod +x .git/hooks/pre-commit")
14
15
16 __version__ = 'v2.1.dev'
17 with open(os.path.join("retriever", "_version.py"), "w") as version_file:
18 version_file.write("__version__ = " + "'" + __version__ + "'\n")
19 version_file.close()
20
21
22 def clean_version(v):
23 return parse_version(v).__repr__().lstrip("<Version('").rstrip("')>")
24
25 packages = [
26 'retriever.lib',
27 'retriever.engines',
28 'retriever',
29 ]
30
31 includes = [
32 'xlrd',
33 'future',
34 'argcomplete',
35 'pymysql',
36 'psycopg2',
37 'sqlite3',
38 ] + extra_includes
39
40 excludes = [
41 'pyreadline',
42 'doctest',
43 'pickle',
44 'pdb',
45 'pywin', 'pywin.debugger',
46 'pywin.debugger.dbgcon',
47 'pywin.dialogs', 'pywin.dialogs.list',
48 'Tkconstants', 'Tkinter', 'tcl', 'tk'
49 ]
50
51 setup(name='retriever',
52 version=clean_version(__version__),
53 description='Data Retriever',
54 author='Ben Morris, Akash Goel, Henry Senyondo, Ethan White',
55 author_email='[email protected]',
56 url='https://github.com/weecology/retriever',
57 classifiers=['Intended Audience :: Science/Research',
58 'License :: OSI Approved :: MIT License',
59 'Programming Language :: Python',
60 'Programming Language :: Python :: 2',
61 'Programming Language :: Python :: 3', ],
62 packages=packages,
63 package_dir={
64 'retriever': 'retriever'
65 },
66 entry_points={
67 'console_scripts': [
68 'retriever = retriever.__main__:main',
69 ],
70 },
71 install_requires=[
72 'xlrd',
73 'future',
74 'argcomplete'
75 ],
76
77 # py2app flags
78 app=['__main__.py'],
79 data_files=[('', ['CITATION'])],
80 setup_requires=[],
81 )
82
83 # windows doesn't have bash. No point in using bash-completion
84 if current_platform != "windows":
85 # if platform is OS X use "~/.bash_profile"
86 if current_platform == "darwin":
87 bash_file = "~/.bash_profile"
88 # if platform is Linux use "~/.bashrc
89 elif current_platform == "linux":
90 bash_file = "~/.bashrc"
91 # else write and discard
92 else:
93 bash_file = "/dev/null"
94
95 argcomplete_command = 'eval "$(register-python-argcomplete retriever)"'
96 with open(os.path.expanduser(bash_file), "a+") as bashrc:
97 bashrc.seek(0)
98 # register retriever for arg-completion if not already registered
99 # whenever a new shell is spawned
100 if argcomplete_command not in bashrc.read():
101 bashrc.write(argcomplete_command + "\n")
102 bashrc.close()
103 os.system("activate-global-python-argcomplete")
104 # register for the current shell
105 os.system(argcomplete_command)
106
107 try:
108 from retriever.compile import compile
109 from retriever.lib.repository import check_for_updates
110 compile()
111 check_for_updates()
112 except:
113 pass
114
[end of setup.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/retriever/engines/msaccess.py b/retriever/engines/msaccess.py
--- a/retriever/engines/msaccess.py
+++ b/retriever/engines/msaccess.py
@@ -27,6 +27,7 @@
"Format of table name",
"[{db} {table}]"),
]
+ placeholder = "?"
def convert_data_type(self, datatype):
"""MS Access can't handle complex Decimal types"""
diff --git a/setup.py b/setup.py
--- a/setup.py
+++ b/setup.py
@@ -7,6 +7,8 @@
current_platform = platform.system().lower()
extra_includes = []
+if current_platform == "windows":
+ extra_includes = ["pypyodbc"]
if os.path.exists(".git/hooks"): # check if we are in git repo
os.system("cp hooks/pre-commit .git/hooks/pre-commit")
|
{"golden_diff": "diff --git a/retriever/engines/msaccess.py b/retriever/engines/msaccess.py\n--- a/retriever/engines/msaccess.py\n+++ b/retriever/engines/msaccess.py\n@@ -27,6 +27,7 @@\n \"Format of table name\",\n \"[{db} {table}]\"),\n ]\n+ placeholder = \"?\"\n \n def convert_data_type(self, datatype):\n \"\"\"MS Access can't handle complex Decimal types\"\"\"\ndiff --git a/setup.py b/setup.py\n--- a/setup.py\n+++ b/setup.py\n@@ -7,6 +7,8 @@\n \n current_platform = platform.system().lower()\n extra_includes = []\n+if current_platform == \"windows\":\n+ extra_includes = [\"pypyodbc\"]\n \n if os.path.exists(\".git/hooks\"): # check if we are in git repo\n os.system(\"cp hooks/pre-commit .git/hooks/pre-commit\")\n", "issue": "install to msaccess error\nWhen I try to install a dataset to msaccess I get the error:\r\nCouldn't create table (no module named 'pypyodbc')\r\n\r\nI've tried doing \"pip install pypyodbc --upgrade\" which seems to work, but doesn't help with the retriever error.\r\n\r\nI'm on a windows 10 machine\n", "before_files": [{"content": "from __future__ import print_function\nfrom builtins import str\nimport os\nfrom retriever.lib.models import Engine, no_cleanup\nfrom retriever import DATA_DIR, current_platform\n\n\nclass engine(Engine):\n \"\"\"Engine instance for Microsoft Access.\"\"\"\n name = \"Microsoft Access\"\n instructions = \"\"\"Create a database in Microsoft Access, close Access, then \\nselect your database file using this dialog.\"\"\"\n abbreviation = \"msaccess\"\n datatypes = {\n \"auto\": \"AUTOINCREMENT\",\n \"int\": \"INTEGER\",\n \"bigint\": \"INTEGER\",\n \"double\": \"NUMERIC\",\n \"decimal\": \"NUMERIC\",\n \"char\": \"VARCHAR\",\n \"bool\": \"BIT\",\n }\n required_opts = [(\"file\",\n \"Enter the filename of your Access database\",\n os.path.join(DATA_DIR, \"access.mdb\"),\n \"Access databases (*.mdb, *.accdb)|*.mdb;*.accdb\"),\n (\"table_name\",\n \"Format of table name\",\n \"[{db} {table}]\"),\n ]\n\n def convert_data_type(self, datatype):\n \"\"\"MS Access can't handle complex Decimal types\"\"\"\n converted = Engine.convert_data_type(self, datatype)\n if \"NUMERIC\" in converted:\n converted = \"NUMERIC\"\n elif \"VARCHAR\" in converted:\n try:\n length = int(converted.split('(')[1].split(')')[0].split(',')[0])\n if length > 255:\n converted = \"TEXT\"\n except:\n pass\n return converted\n\n def create_db(self):\n \"\"\"MS Access doesn't create databases.\"\"\"\n return None\n\n def drop_statement(self, objecttype, objectname):\n \"\"\"Returns a drop table or database SQL statement.\"\"\"\n dropstatement = \"DROP %s %s\" % (objecttype, objectname)\n return dropstatement\n\n def insert_data_from_file(self, filename):\n \"\"\"Perform a bulk insert.\"\"\"\n self.get_cursor()\n ct = len([True for c in self.table.columns if c[1][0][:3] == \"ct-\"]) != 0\n if ((self.table.cleanup.function == no_cleanup and not self.table.fixed_width and\n self.table.header_rows < 2)\n and (self.table.delimiter in [\"\\t\", \",\"])\n and not ct\n and (not hasattr(self.table, \"do_not_bulk_insert\") or not self.table.do_not_bulk_insert)\n ):\n print (\"Inserting data from \" + os.path.basename(filename) + \"...\")\n\n if self.table.delimiter == \"\\t\":\n fmt = \"TabDelimited\"\n elif self.table.delimiter == \",\":\n fmt = \"CSVDelimited\"\n\n if self.table.header_rows == 1:\n hdr = \"Yes\"\n else:\n hdr = \"No\"\n\n columns = self.table.get_insert_columns()\n\n need_to_delete = False\n add_to_record_id = 0\n\n if self.table.pk and not self.table.contains_pk:\n if '.' in os.path.basename(filename):\n proper_name = filename.split('.')\n newfilename = '.'.join((proper_name[0:-1]) if len(proper_name) > 0 else proper_name[0]\n ) + \"_new.\" + filename.split(\".\")[-1]\n else:\n newfilename = filename + \"_new\"\n\n if not os.path.isfile(newfilename):\n print(\"Adding index to \" + os.path.abspath(newfilename) + \"...\")\n read = open(filename, \"rb\")\n write = open(newfilename, \"wb\")\n to_write = \"\"\n\n for line in read:\n to_write += str(id) + self.table.delimiter + line.replace(\"\\n\", \"\\r\\n\")\n add_to_record_id += 1\n self.table.record_id += add_to_record_id\n\n write.write(to_write)\n write.close()\n read.close()\n need_to_delete = True\n columns = \"record_id, \" + columns\n else:\n newfilename = filename\n\n newfilename = os.path.abspath(newfilename)\n filename_length = (len(os.path.basename(newfilename)) * -1) - 1\n filepath = newfilename[:filename_length]\n statement = \"\"\"\nINSERT INTO \"\"\" + self.table_name() + \" (\" + columns + \"\"\")\nSELECT * FROM [\"\"\" + os.path.basename(newfilename) + ''']\nIN \"''' + filepath + '''\" \"Text;FMT=''' + fmt + ''';HDR=''' + hdr + ''';\"'''\n\n try:\n self.execute(statement)\n except:\n print(\"Couldn't bulk insert. Trying manual insert.\")\n self.connection.rollback()\n\n self.table.record_id -= add_to_record_id\n\n return Engine.insert_data_from_file(self, filename)\n\n if need_to_delete:\n os.remove(newfilename)\n\n else:\n return Engine.insert_data_from_file(self, filename)\n\n def table_exists(self, dbname, tablename):\n \"\"\"Determine if the table already exists in the database\"\"\"\n if not hasattr(self, 'existing_table_names'):\n self.existing_table_names = set()\n for row in self.cursor.tables():\n tableinfo = row[2]\n if not tableinfo.startswith(\"MSys\"):\n # ignore system tables\n database, table = tableinfo.split()\n self.existing_table_names.add((database, table))\n return self.table_name(name=tablename, dbname=dbname).lower() in self.existing_table_names\n\n def get_connection(self):\n \"\"\"Gets the db connection.\"\"\"\n if current_platform != \"windows\":\n raise Exception(\"MS Access can only be used in Windows.\")\n import pypyodbc as dbapi\n self.get_input()\n if not os.path.exists(self.opts['file']) and self.opts['file'].endswith('.mdb'):\n dbapi.win_create_mdb(self.opts['file'])\n connection_string = (\"DRIVER={Microsoft Access Driver (*.mdb, *.accdb)};DBQ=\" +\n os.path.abspath(self.opts[\"file\"]).replace(\"/\", \"//\") + \";\")\n return dbapi.connect(connection_string, autocommit=False)\n", "path": "retriever/engines/msaccess.py"}, {"content": "\"\"\"Use the following command to install retriever: python setup.py install\"\"\"\nfrom __future__ import absolute_import\nimport platform\nimport os\nfrom setuptools import setup\nfrom pkg_resources import parse_version\n\ncurrent_platform = platform.system().lower()\nextra_includes = []\n\nif os.path.exists(\".git/hooks\"): # check if we are in git repo\n os.system(\"cp hooks/pre-commit .git/hooks/pre-commit\")\n os.system(\"chmod +x .git/hooks/pre-commit\")\n\n\n__version__ = 'v2.1.dev'\nwith open(os.path.join(\"retriever\", \"_version.py\"), \"w\") as version_file:\n version_file.write(\"__version__ = \" + \"'\" + __version__ + \"'\\n\")\n version_file.close()\n\n\ndef clean_version(v):\n return parse_version(v).__repr__().lstrip(\"<Version('\").rstrip(\"')>\")\n\npackages = [\n 'retriever.lib',\n 'retriever.engines',\n 'retriever',\n]\n\nincludes = [\n 'xlrd',\n 'future',\n 'argcomplete',\n 'pymysql',\n 'psycopg2',\n 'sqlite3',\n] + extra_includes\n\nexcludes = [\n 'pyreadline',\n 'doctest',\n 'pickle',\n 'pdb',\n 'pywin', 'pywin.debugger',\n 'pywin.debugger.dbgcon',\n 'pywin.dialogs', 'pywin.dialogs.list',\n 'Tkconstants', 'Tkinter', 'tcl', 'tk'\n]\n\nsetup(name='retriever',\n version=clean_version(__version__),\n description='Data Retriever',\n author='Ben Morris, Akash Goel, Henry Senyondo, Ethan White',\n author_email='[email protected]',\n url='https://github.com/weecology/retriever',\n classifiers=['Intended Audience :: Science/Research',\n 'License :: OSI Approved :: MIT License',\n 'Programming Language :: Python',\n 'Programming Language :: Python :: 2',\n 'Programming Language :: Python :: 3', ],\n packages=packages,\n package_dir={\n 'retriever': 'retriever'\n },\n entry_points={\n 'console_scripts': [\n 'retriever = retriever.__main__:main',\n ],\n },\n install_requires=[\n 'xlrd',\n 'future',\n 'argcomplete'\n ],\n\n # py2app flags\n app=['__main__.py'],\n data_files=[('', ['CITATION'])],\n setup_requires=[],\n )\n\n# windows doesn't have bash. No point in using bash-completion\nif current_platform != \"windows\":\n # if platform is OS X use \"~/.bash_profile\"\n if current_platform == \"darwin\":\n bash_file = \"~/.bash_profile\"\n # if platform is Linux use \"~/.bashrc\n elif current_platform == \"linux\":\n bash_file = \"~/.bashrc\"\n # else write and discard\n else:\n bash_file = \"/dev/null\"\n\n argcomplete_command = 'eval \"$(register-python-argcomplete retriever)\"'\n with open(os.path.expanduser(bash_file), \"a+\") as bashrc:\n bashrc.seek(0)\n # register retriever for arg-completion if not already registered\n # whenever a new shell is spawned\n if argcomplete_command not in bashrc.read():\n bashrc.write(argcomplete_command + \"\\n\")\n bashrc.close()\n os.system(\"activate-global-python-argcomplete\")\n # register for the current shell\n os.system(argcomplete_command)\n\ntry:\n from retriever.compile import compile\n from retriever.lib.repository import check_for_updates\n compile()\n check_for_updates()\nexcept:\n pass\n", "path": "setup.py"}]}
| 3,313 | 202 |
gh_patches_debug_2390
|
rasdani/github-patches
|
git_diff
|
Qiskit__qiskit-2448
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
No module named 'vcr': requirement is missing (vcrpy)
<!-- ⚠️ If you do not respect this template, your issue will be closed -->
<!-- ⚠️ Make sure to browse the opened and closed issues -->
### Information
- **Qiskit Terra version**: 0.10.1
- **Python version**: 3.7.3
- **Operating system**: windows 10
### What is the current behavior?
Fresh qiskit installation inside a new environment on windows 10.
In one of the terra tutorial (using_the_transpiler) `from qiskit.test.mock import FakeTokyo` is failing 'ModuleNotFoundError: No module named vcr'
### Suggested solutions
'pip install vcrpy'
'vcrpy' needs to be added in requirements.
</issue>
<code>
[start of qiskit/util.py]
1 # -*- coding: utf-8 -*-
2 # This code is part of Qiskit.
3 #
4 # (C) Copyright IBM 2017.
5 #
6 # This code is licensed under the Apache License, Version 2.0. You may
7 # obtain a copy of this license in the LICENSE.txt file in the root directory
8 # of this source tree or at http://www.apache.org/licenses/LICENSE-2.0.
9 #
10 # Any modifications or derivative works of this code must retain this
11 # copyright notice, and modified files need to carry a notice indicating
12 # that they have been altered from the originals.
13
14 """Common utilities for Qiskit."""
15
16 import platform
17 import re
18 import socket
19 import sys
20 import warnings
21
22 import psutil
23 from marshmallow.warnings import ChangedInMarshmallow3Warning
24
25
26 def _check_python_version():
27 """Check for Python version 3.5+."""
28 if sys.version_info < (3, 5):
29 raise Exception('Qiskit requires Python version 3.5 or greater.')
30
31
32 def _filter_deprecation_warnings():
33 """Apply filters to deprecation warnings.
34
35 Force the `DeprecationWarning` warnings to be displayed for the qiskit
36 module, overriding the system configuration as they are ignored by default
37 [1] for end-users. Additionally, silence the `ChangedInMarshmallow3Warning`
38 messages.
39
40 TODO: on Python 3.7, this might not be needed due to PEP-0565 [2].
41
42 [1] https://docs.python.org/3/library/warnings.html#default-warning-filters
43 [2] https://www.python.org/dev/peps/pep-0565/
44 """
45 deprecation_filter = ('always', None, DeprecationWarning,
46 re.compile(r'^qiskit\.*', re.UNICODE), 0)
47
48 # Instead of using warnings.simple_filter() directly, the internal
49 # _add_filter() function is used for being able to match against the
50 # module.
51 try:
52 warnings._add_filter(*deprecation_filter, append=False)
53 except AttributeError:
54 # ._add_filter is internal and not available in some Python versions.
55 pass
56
57 # Add a filter for ignoring ChangedInMarshmallow3Warning, as we depend on
58 # marhsmallow 2 explicitly. 2.17.0 introduced new deprecation warnings that
59 # are useful for eventually migrating, but too verbose for our purposes.
60 warnings.simplefilter('ignore', category=ChangedInMarshmallow3Warning)
61
62
63 _check_python_version()
64 _filter_deprecation_warnings()
65
66
67 def local_hardware_info():
68 """Basic hardware information about the local machine.
69
70 Gives actual number of CPU's in the machine, even when hyperthreading is
71 turned on. CPU count defaults to 1 when true count can't be determined.
72
73 Returns:
74 dict: The hardware information.
75 """
76 results = {
77 'os': platform.system(),
78 'memory': psutil.virtual_memory().total / (1024 ** 3),
79 'cpus': psutil.cpu_count(logical=False) or 1
80 }
81 return results
82
83
84 def _has_connection(hostname, port):
85 """Checks if internet connection exists to host via specified port.
86
87 If any exception is raised while trying to open a socket this will return
88 false.
89
90 Args:
91 hostname (str): Hostname to connect to.
92 port (int): Port to connect to
93
94 Returns:
95 bool: Has connection or not
96
97 """
98 try:
99 host = socket.gethostbyname(hostname)
100 socket.create_connection((host, port), 2)
101 return True
102 except Exception: # pylint: disable=broad-except
103 return False
104
[end of qiskit/util.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/qiskit/util.py b/qiskit/util.py
--- a/qiskit/util.py
+++ b/qiskit/util.py
@@ -97,7 +97,7 @@
"""
try:
host = socket.gethostbyname(hostname)
- socket.create_connection((host, port), 2)
+ socket.create_connection((host, port), 2).close()
return True
except Exception: # pylint: disable=broad-except
return False
|
{"golden_diff": "diff --git a/qiskit/util.py b/qiskit/util.py\n--- a/qiskit/util.py\n+++ b/qiskit/util.py\n@@ -97,7 +97,7 @@\n \"\"\"\n try:\n host = socket.gethostbyname(hostname)\n- socket.create_connection((host, port), 2)\n+ socket.create_connection((host, port), 2).close()\n return True\n except Exception: # pylint: disable=broad-except\n return False\n", "issue": "No module named 'vcr': requirement is missing (vcrpy) \n<!-- \u26a0\ufe0f If you do not respect this template, your issue will be closed -->\r\n<!-- \u26a0\ufe0f Make sure to browse the opened and closed issues -->\r\n\r\n### Information\r\n\r\n- **Qiskit Terra version**: 0.10.1\r\n- **Python version**: 3.7.3\r\n- **Operating system**: windows 10\r\n\r\n### What is the current behavior?\r\nFresh qiskit installation inside a new environment on windows 10. \r\nIn one of the terra tutorial (using_the_transpiler) `from qiskit.test.mock import FakeTokyo` is failing 'ModuleNotFoundError: No module named vcr'\r\n\r\n### Suggested solutions\r\n'pip install vcrpy' \r\n'vcrpy' needs to be added in requirements.\r\n\n", "before_files": [{"content": "# -*- coding: utf-8 -*-\n# This code is part of Qiskit.\n#\n# (C) Copyright IBM 2017.\n#\n# This code is licensed under the Apache License, Version 2.0. You may\n# obtain a copy of this license in the LICENSE.txt file in the root directory\n# of this source tree or at http://www.apache.org/licenses/LICENSE-2.0.\n#\n# Any modifications or derivative works of this code must retain this\n# copyright notice, and modified files need to carry a notice indicating\n# that they have been altered from the originals.\n\n\"\"\"Common utilities for Qiskit.\"\"\"\n\nimport platform\nimport re\nimport socket\nimport sys\nimport warnings\n\nimport psutil\nfrom marshmallow.warnings import ChangedInMarshmallow3Warning\n\n\ndef _check_python_version():\n \"\"\"Check for Python version 3.5+.\"\"\"\n if sys.version_info < (3, 5):\n raise Exception('Qiskit requires Python version 3.5 or greater.')\n\n\ndef _filter_deprecation_warnings():\n \"\"\"Apply filters to deprecation warnings.\n\n Force the `DeprecationWarning` warnings to be displayed for the qiskit\n module, overriding the system configuration as they are ignored by default\n [1] for end-users. Additionally, silence the `ChangedInMarshmallow3Warning`\n messages.\n\n TODO: on Python 3.7, this might not be needed due to PEP-0565 [2].\n\n [1] https://docs.python.org/3/library/warnings.html#default-warning-filters\n [2] https://www.python.org/dev/peps/pep-0565/\n \"\"\"\n deprecation_filter = ('always', None, DeprecationWarning,\n re.compile(r'^qiskit\\.*', re.UNICODE), 0)\n\n # Instead of using warnings.simple_filter() directly, the internal\n # _add_filter() function is used for being able to match against the\n # module.\n try:\n warnings._add_filter(*deprecation_filter, append=False)\n except AttributeError:\n # ._add_filter is internal and not available in some Python versions.\n pass\n\n # Add a filter for ignoring ChangedInMarshmallow3Warning, as we depend on\n # marhsmallow 2 explicitly. 2.17.0 introduced new deprecation warnings that\n # are useful for eventually migrating, but too verbose for our purposes.\n warnings.simplefilter('ignore', category=ChangedInMarshmallow3Warning)\n\n\n_check_python_version()\n_filter_deprecation_warnings()\n\n\ndef local_hardware_info():\n \"\"\"Basic hardware information about the local machine.\n\n Gives actual number of CPU's in the machine, even when hyperthreading is\n turned on. CPU count defaults to 1 when true count can't be determined.\n\n Returns:\n dict: The hardware information.\n \"\"\"\n results = {\n 'os': platform.system(),\n 'memory': psutil.virtual_memory().total / (1024 ** 3),\n 'cpus': psutil.cpu_count(logical=False) or 1\n }\n return results\n\n\ndef _has_connection(hostname, port):\n \"\"\"Checks if internet connection exists to host via specified port.\n\n If any exception is raised while trying to open a socket this will return\n false.\n\n Args:\n hostname (str): Hostname to connect to.\n port (int): Port to connect to\n\n Returns:\n bool: Has connection or not\n\n \"\"\"\n try:\n host = socket.gethostbyname(hostname)\n socket.create_connection((host, port), 2)\n return True\n except Exception: # pylint: disable=broad-except\n return False\n", "path": "qiskit/util.py"}]}
| 1,728 | 108 |
gh_patches_debug_34444
|
rasdani/github-patches
|
git_diff
|
google__osv.dev-439
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Bulk query API
A bulk query API would allow developers to more easily query the API without hitting rate limits. It would also help with scenarios like #257, where an SBOM will contain many dependencies.
</issue>
<code>
[start of docs/build.py]
1 # Copyright 2021 Google LLC
2 #
3 # Licensed under the Apache License, Version 2.0 (the "License");
4 # you may not use this file except in compliance with the License.
5 # You may obtain a copy of the License at
6 #
7 # http://www.apache.org/licenses/LICENSE-2.0
8 #
9 # Unless required by applicable law or agreed to in writing, software
10 # distributed under the License is distributed on an "AS IS" BASIS,
11 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12 # See the License for the specific language governing permissions and
13 # limitations under the License.
14 """Documentation builder."""
15
16 import json
17 import os
18 import shutil
19 import subprocess
20
21 _ROOT_DIR = os.path.dirname(os.path.dirname(os.path.abspath(__file__)))
22 _GENERATED_FILENAME = 'v1/osv_service_v1.swagger.json'
23
24
25 def property_description_workaround(definition):
26 """Work around an OpenAPI limitation with a field descriptions getting
27 replaced by the object descriptions."""
28 # Workaround described in https://github.com/Redocly/redoc/issues/835.
29 for value in definition['properties'].values():
30 if '$ref' in value:
31 value['allOf'] = [{'$ref': value['$ref']}]
32 del value['$ref']
33
34
35 def replace_property_name(definition, key, replacement):
36 """Replace property name."""
37 definition['properties'][replacement] = definition['properties'][key]
38 del definition['properties'][key]
39
40
41 def main():
42 api_dir = os.path.join(_ROOT_DIR, 'gcp', 'api')
43 v1_api_dir = os.path.join(api_dir, 'v1')
44 googleapis_dir = os.path.join(api_dir, 'googleapis')
45 service_proto_path = os.path.join(v1_api_dir, 'osv_service_v1.proto')
46
47 # Add OSV dependencies.
48 osv_path = os.path.join(api_dir, 'osv')
49 if os.path.exists(osv_path):
50 shutil.rmtree(osv_path)
51
52 shutil.copytree(os.path.join(_ROOT_DIR, 'lib', 'osv'), osv_path)
53
54 subprocess.run([
55 'protoc',
56 '-I',
57 api_dir,
58 '-I',
59 v1_api_dir,
60 '-I',
61 googleapis_dir,
62 '--openapiv2_out',
63 '.',
64 '--openapiv2_opt',
65 'logtostderr=true',
66 service_proto_path,
67 ],
68 check=True)
69
70 with open(_GENERATED_FILENAME) as f:
71 spec = json.load(f)
72
73 spec['host'] = 'api.osv.dev'
74 spec['info']['title'] = 'OSV'
75 spec['info']['version'] = '1.0'
76 spec['tags'] = [{
77 'name': 'api',
78 'x-displayName': 'API',
79 'description': 'The API has 2 methods:'
80 }, {
81 'name': 'vulnerability_schema',
82 'x-displayName': 'Vulnerability schema',
83 'description': 'Please see the [OpenSSF Open Source Vulnerability spec]'
84 '(https://ossf.github.io/osv-schema/).',
85 }]
86
87 spec['x-tagGroups'] = [{
88 'name': 'API',
89 'tags': ['api']
90 }, {
91 'name': 'Schema',
92 'tags': ['vulnerability_schema']
93 }]
94
95 spec['paths']['/v1/query']['post']['tags'] = ['api']
96 spec['paths']['/v1/vulns/{id}']['get']['tags'] = ['api']
97
98 spec['paths']['/v1/query']['post']['x-code-samples'] = [{
99 'lang':
100 'Curl example',
101 'source':
102 ('curl -X POST -d \\\n'
103 ' \'{"commit": "6879efc2c1596d11a6a6ad296f80063b558d5e0f"}\' \\\n'
104 ' "https://api.osv.dev/v1/query"\n\n'
105 'curl -X POST -d \\\n'
106 ' \'{"package": {"name": "mruby"}, "version": "2.1.2rc"}\' \\\n'
107 ' "https://api.osv.dev/v1/query"')
108 }]
109
110 spec['paths']['/v1/vulns/{id}']['get']['x-code-samples'] = [{
111 'lang': 'Curl example',
112 'source': 'curl "https://api.osv.dev/v1/vulns/OSV-2020-111"'
113 }]
114
115 property_description_workaround(spec['definitions']['v1Query'])
116 property_description_workaround(spec['definitions']['osvVulnerability'])
117
118 replace_property_name(spec['definitions']['osvVulnerability'],
119 'databaseSpecific', 'database_specific')
120
121 with open('sections.md') as f:
122 spec['info']['description'] = f.read()
123
124 with open(_GENERATED_FILENAME, 'w') as f:
125 f.write(json.dumps(spec, indent=2))
126
127 shutil.move(_GENERATED_FILENAME, os.path.basename(_GENERATED_FILENAME))
128
129
130 if __name__ == '__main__':
131 main()
132
[end of docs/build.py]
[start of gcp/api/osv_service_v1_pb2_grpc.py]
1 # Generated by the gRPC Python protocol compiler plugin. DO NOT EDIT!
2 """Client and server classes corresponding to protobuf-defined services."""
3 import grpc
4
5 from osv import vulnerability_pb2 as osv_dot_vulnerability__pb2
6 import osv_service_v1_pb2 as osv__service__v1__pb2
7
8
9 class OSVStub(object):
10 """Open source vulnerability database.
11 """
12
13 def __init__(self, channel):
14 """Constructor.
15
16 Args:
17 channel: A grpc.Channel.
18 """
19 self.GetVulnById = channel.unary_unary(
20 '/osv.v1.OSV/GetVulnById',
21 request_serializer=osv__service__v1__pb2.GetVulnByIdParameters.SerializeToString,
22 response_deserializer=osv_dot_vulnerability__pb2.Vulnerability.FromString,
23 )
24 self.QueryAffected = channel.unary_unary(
25 '/osv.v1.OSV/QueryAffected',
26 request_serializer=osv__service__v1__pb2.QueryAffectedParameters.SerializeToString,
27 response_deserializer=osv__service__v1__pb2.VulnerabilityList.FromString,
28 )
29 self.QueryAffectedBatch = channel.unary_unary(
30 '/osv.v1.OSV/QueryAffectedBatch',
31 request_serializer=osv__service__v1__pb2.QueryAffectedBatchParameters.SerializeToString,
32 response_deserializer=osv__service__v1__pb2.BatchVulnerabilityList.FromString,
33 )
34
35
36 class OSVServicer(object):
37 """Open source vulnerability database.
38 """
39
40 def GetVulnById(self, request, context):
41 """Return a `Vulnerability` object for a given OSV ID.
42 """
43 context.set_code(grpc.StatusCode.UNIMPLEMENTED)
44 context.set_details('Method not implemented!')
45 raise NotImplementedError('Method not implemented!')
46
47 def QueryAffected(self, request, context):
48 """Query vulnerabilities for a particular project at a given commit or
49 version.
50 """
51 context.set_code(grpc.StatusCode.UNIMPLEMENTED)
52 context.set_details('Method not implemented!')
53 raise NotImplementedError('Method not implemented!')
54
55 def QueryAffectedBatch(self, request, context):
56 """Query vulnerabilities for a particular project at given commits and
57 versions.
58 """
59 context.set_code(grpc.StatusCode.UNIMPLEMENTED)
60 context.set_details('Method not implemented!')
61 raise NotImplementedError('Method not implemented!')
62
63
64 def add_OSVServicer_to_server(servicer, server):
65 rpc_method_handlers = {
66 'GetVulnById': grpc.unary_unary_rpc_method_handler(
67 servicer.GetVulnById,
68 request_deserializer=osv__service__v1__pb2.GetVulnByIdParameters.FromString,
69 response_serializer=osv_dot_vulnerability__pb2.Vulnerability.SerializeToString,
70 ),
71 'QueryAffected': grpc.unary_unary_rpc_method_handler(
72 servicer.QueryAffected,
73 request_deserializer=osv__service__v1__pb2.QueryAffectedParameters.FromString,
74 response_serializer=osv__service__v1__pb2.VulnerabilityList.SerializeToString,
75 ),
76 'QueryAffectedBatch': grpc.unary_unary_rpc_method_handler(
77 servicer.QueryAffectedBatch,
78 request_deserializer=osv__service__v1__pb2.QueryAffectedBatchParameters.FromString,
79 response_serializer=osv__service__v1__pb2.BatchVulnerabilityList.SerializeToString,
80 ),
81 }
82 generic_handler = grpc.method_handlers_generic_handler(
83 'osv.v1.OSV', rpc_method_handlers)
84 server.add_generic_rpc_handlers((generic_handler,))
85
86
87 # This class is part of an EXPERIMENTAL API.
88 class OSV(object):
89 """Open source vulnerability database.
90 """
91
92 @staticmethod
93 def GetVulnById(request,
94 target,
95 options=(),
96 channel_credentials=None,
97 call_credentials=None,
98 insecure=False,
99 compression=None,
100 wait_for_ready=None,
101 timeout=None,
102 metadata=None):
103 return grpc.experimental.unary_unary(request, target, '/osv.v1.OSV/GetVulnById',
104 osv__service__v1__pb2.GetVulnByIdParameters.SerializeToString,
105 osv_dot_vulnerability__pb2.Vulnerability.FromString,
106 options, channel_credentials,
107 insecure, call_credentials, compression, wait_for_ready, timeout, metadata)
108
109 @staticmethod
110 def QueryAffected(request,
111 target,
112 options=(),
113 channel_credentials=None,
114 call_credentials=None,
115 insecure=False,
116 compression=None,
117 wait_for_ready=None,
118 timeout=None,
119 metadata=None):
120 return grpc.experimental.unary_unary(request, target, '/osv.v1.OSV/QueryAffected',
121 osv__service__v1__pb2.QueryAffectedParameters.SerializeToString,
122 osv__service__v1__pb2.VulnerabilityList.FromString,
123 options, channel_credentials,
124 insecure, call_credentials, compression, wait_for_ready, timeout, metadata)
125
126 @staticmethod
127 def QueryAffectedBatch(request,
128 target,
129 options=(),
130 channel_credentials=None,
131 call_credentials=None,
132 insecure=False,
133 compression=None,
134 wait_for_ready=None,
135 timeout=None,
136 metadata=None):
137 return grpc.experimental.unary_unary(request, target, '/osv.v1.OSV/QueryAffectedBatch',
138 osv__service__v1__pb2.QueryAffectedBatchParameters.SerializeToString,
139 osv__service__v1__pb2.BatchVulnerabilityList.FromString,
140 options, channel_credentials,
141 insecure, call_credentials, compression, wait_for_ready, timeout, metadata)
142
[end of gcp/api/osv_service_v1_pb2_grpc.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/docs/build.py b/docs/build.py
--- a/docs/build.py
+++ b/docs/build.py
@@ -76,7 +76,7 @@
spec['tags'] = [{
'name': 'api',
'x-displayName': 'API',
- 'description': 'The API has 2 methods:'
+ 'description': 'The API has 3 methods:'
}, {
'name': 'vulnerability_schema',
'x-displayName': 'Vulnerability schema',
@@ -93,6 +93,7 @@
}]
spec['paths']['/v1/query']['post']['tags'] = ['api']
+ spec['paths']['/v1/querybatch']['post']['tags'] = ['api']
spec['paths']['/v1/vulns/{id}']['get']['tags'] = ['api']
spec['paths']['/v1/query']['post']['x-code-samples'] = [{
@@ -107,6 +108,33 @@
' "https://api.osv.dev/v1/query"')
}]
+ spec['paths']['/v1/querybatch']['post']['x-code-samples'] = [{
+ 'lang':
+ 'Curl example',
+ 'source':
+ ("""cat <<EOF | curl -X POST -d @- "https://api.osv.dev/v1/querybatch"
+{
+ "queries": [
+ {
+ "package": {
+ "purl": "pkg:pypi/[email protected]"
+ }
+ },
+ {
+ "commit": "6879efc2c1596d11a6a6ad296f80063b558d5e0f"
+ },
+ {
+ "package": {
+ "ecosystem": "PyPI",
+ "name": "jinja2"
+ },
+ "version": "2.4.1"
+ }
+ ]
+}
+EOF""")
+ }]
+
spec['paths']['/v1/vulns/{id}']['get']['x-code-samples'] = [{
'lang': 'Curl example',
'source': 'curl "https://api.osv.dev/v1/vulns/OSV-2020-111"'
diff --git a/gcp/api/osv_service_v1_pb2_grpc.py b/gcp/api/osv_service_v1_pb2_grpc.py
--- a/gcp/api/osv_service_v1_pb2_grpc.py
+++ b/gcp/api/osv_service_v1_pb2_grpc.py
@@ -53,8 +53,7 @@
raise NotImplementedError('Method not implemented!')
def QueryAffectedBatch(self, request, context):
- """Query vulnerabilities for a particular project at given commits and
- versions.
+ """Query vulnerabilities (batched) for given package versions and commits.
"""
context.set_code(grpc.StatusCode.UNIMPLEMENTED)
context.set_details('Method not implemented!')
|
{"golden_diff": "diff --git a/docs/build.py b/docs/build.py\n--- a/docs/build.py\n+++ b/docs/build.py\n@@ -76,7 +76,7 @@\n spec['tags'] = [{\n 'name': 'api',\n 'x-displayName': 'API',\n- 'description': 'The API has 2 methods:'\n+ 'description': 'The API has 3 methods:'\n }, {\n 'name': 'vulnerability_schema',\n 'x-displayName': 'Vulnerability schema',\n@@ -93,6 +93,7 @@\n }]\n \n spec['paths']['/v1/query']['post']['tags'] = ['api']\n+ spec['paths']['/v1/querybatch']['post']['tags'] = ['api']\n spec['paths']['/v1/vulns/{id}']['get']['tags'] = ['api']\n \n spec['paths']['/v1/query']['post']['x-code-samples'] = [{\n@@ -107,6 +108,33 @@\n ' \"https://api.osv.dev/v1/query\"')\n }]\n \n+ spec['paths']['/v1/querybatch']['post']['x-code-samples'] = [{\n+ 'lang':\n+ 'Curl example',\n+ 'source':\n+ (\"\"\"cat <<EOF | curl -X POST -d @- \"https://api.osv.dev/v1/querybatch\"\n+{\n+ \"queries\": [\n+ {\n+ \"package\": {\n+ \"purl\": \"pkg:pypi/[email protected]\"\n+ }\n+ },\n+ {\n+ \"commit\": \"6879efc2c1596d11a6a6ad296f80063b558d5e0f\"\n+ },\n+ {\n+ \"package\": {\n+ \"ecosystem\": \"PyPI\",\n+ \"name\": \"jinja2\"\n+ },\n+ \"version\": \"2.4.1\"\n+ }\n+ ]\n+}\n+EOF\"\"\")\n+ }]\n+\n spec['paths']['/v1/vulns/{id}']['get']['x-code-samples'] = [{\n 'lang': 'Curl example',\n 'source': 'curl \"https://api.osv.dev/v1/vulns/OSV-2020-111\"'\ndiff --git a/gcp/api/osv_service_v1_pb2_grpc.py b/gcp/api/osv_service_v1_pb2_grpc.py\n--- a/gcp/api/osv_service_v1_pb2_grpc.py\n+++ b/gcp/api/osv_service_v1_pb2_grpc.py\n@@ -53,8 +53,7 @@\n raise NotImplementedError('Method not implemented!')\n \n def QueryAffectedBatch(self, request, context):\n- \"\"\"Query vulnerabilities for a particular project at given commits and\n- versions.\n+ \"\"\"Query vulnerabilities (batched) for given package versions and commits.\n \"\"\"\n context.set_code(grpc.StatusCode.UNIMPLEMENTED)\n context.set_details('Method not implemented!')\n", "issue": "Bulk query API\nA bulk query API would allow developers to more easily query the API without hitting rate limits. It would also help with scenarios like #257, where an SBOM will contain many dependencies.\n", "before_files": [{"content": "# Copyright 2021 Google LLC\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\"\"\"Documentation builder.\"\"\"\n\nimport json\nimport os\nimport shutil\nimport subprocess\n\n_ROOT_DIR = os.path.dirname(os.path.dirname(os.path.abspath(__file__)))\n_GENERATED_FILENAME = 'v1/osv_service_v1.swagger.json'\n\n\ndef property_description_workaround(definition):\n \"\"\"Work around an OpenAPI limitation with a field descriptions getting\n replaced by the object descriptions.\"\"\"\n # Workaround described in https://github.com/Redocly/redoc/issues/835.\n for value in definition['properties'].values():\n if '$ref' in value:\n value['allOf'] = [{'$ref': value['$ref']}]\n del value['$ref']\n\n\ndef replace_property_name(definition, key, replacement):\n \"\"\"Replace property name.\"\"\"\n definition['properties'][replacement] = definition['properties'][key]\n del definition['properties'][key]\n\n\ndef main():\n api_dir = os.path.join(_ROOT_DIR, 'gcp', 'api')\n v1_api_dir = os.path.join(api_dir, 'v1')\n googleapis_dir = os.path.join(api_dir, 'googleapis')\n service_proto_path = os.path.join(v1_api_dir, 'osv_service_v1.proto')\n\n # Add OSV dependencies.\n osv_path = os.path.join(api_dir, 'osv')\n if os.path.exists(osv_path):\n shutil.rmtree(osv_path)\n\n shutil.copytree(os.path.join(_ROOT_DIR, 'lib', 'osv'), osv_path)\n\n subprocess.run([\n 'protoc',\n '-I',\n api_dir,\n '-I',\n v1_api_dir,\n '-I',\n googleapis_dir,\n '--openapiv2_out',\n '.',\n '--openapiv2_opt',\n 'logtostderr=true',\n service_proto_path,\n ],\n check=True)\n\n with open(_GENERATED_FILENAME) as f:\n spec = json.load(f)\n\n spec['host'] = 'api.osv.dev'\n spec['info']['title'] = 'OSV'\n spec['info']['version'] = '1.0'\n spec['tags'] = [{\n 'name': 'api',\n 'x-displayName': 'API',\n 'description': 'The API has 2 methods:'\n }, {\n 'name': 'vulnerability_schema',\n 'x-displayName': 'Vulnerability schema',\n 'description': 'Please see the [OpenSSF Open Source Vulnerability spec]'\n '(https://ossf.github.io/osv-schema/).',\n }]\n\n spec['x-tagGroups'] = [{\n 'name': 'API',\n 'tags': ['api']\n }, {\n 'name': 'Schema',\n 'tags': ['vulnerability_schema']\n }]\n\n spec['paths']['/v1/query']['post']['tags'] = ['api']\n spec['paths']['/v1/vulns/{id}']['get']['tags'] = ['api']\n\n spec['paths']['/v1/query']['post']['x-code-samples'] = [{\n 'lang':\n 'Curl example',\n 'source':\n ('curl -X POST -d \\\\\\n'\n ' \\'{\"commit\": \"6879efc2c1596d11a6a6ad296f80063b558d5e0f\"}\\' \\\\\\n'\n ' \"https://api.osv.dev/v1/query\"\\n\\n'\n 'curl -X POST -d \\\\\\n'\n ' \\'{\"package\": {\"name\": \"mruby\"}, \"version\": \"2.1.2rc\"}\\' \\\\\\n'\n ' \"https://api.osv.dev/v1/query\"')\n }]\n\n spec['paths']['/v1/vulns/{id}']['get']['x-code-samples'] = [{\n 'lang': 'Curl example',\n 'source': 'curl \"https://api.osv.dev/v1/vulns/OSV-2020-111\"'\n }]\n\n property_description_workaround(spec['definitions']['v1Query'])\n property_description_workaround(spec['definitions']['osvVulnerability'])\n\n replace_property_name(spec['definitions']['osvVulnerability'],\n 'databaseSpecific', 'database_specific')\n\n with open('sections.md') as f:\n spec['info']['description'] = f.read()\n\n with open(_GENERATED_FILENAME, 'w') as f:\n f.write(json.dumps(spec, indent=2))\n\n shutil.move(_GENERATED_FILENAME, os.path.basename(_GENERATED_FILENAME))\n\n\nif __name__ == '__main__':\n main()\n", "path": "docs/build.py"}, {"content": "# Generated by the gRPC Python protocol compiler plugin. DO NOT EDIT!\n\"\"\"Client and server classes corresponding to protobuf-defined services.\"\"\"\nimport grpc\n\nfrom osv import vulnerability_pb2 as osv_dot_vulnerability__pb2\nimport osv_service_v1_pb2 as osv__service__v1__pb2\n\n\nclass OSVStub(object):\n \"\"\"Open source vulnerability database.\n \"\"\"\n\n def __init__(self, channel):\n \"\"\"Constructor.\n\n Args:\n channel: A grpc.Channel.\n \"\"\"\n self.GetVulnById = channel.unary_unary(\n '/osv.v1.OSV/GetVulnById',\n request_serializer=osv__service__v1__pb2.GetVulnByIdParameters.SerializeToString,\n response_deserializer=osv_dot_vulnerability__pb2.Vulnerability.FromString,\n )\n self.QueryAffected = channel.unary_unary(\n '/osv.v1.OSV/QueryAffected',\n request_serializer=osv__service__v1__pb2.QueryAffectedParameters.SerializeToString,\n response_deserializer=osv__service__v1__pb2.VulnerabilityList.FromString,\n )\n self.QueryAffectedBatch = channel.unary_unary(\n '/osv.v1.OSV/QueryAffectedBatch',\n request_serializer=osv__service__v1__pb2.QueryAffectedBatchParameters.SerializeToString,\n response_deserializer=osv__service__v1__pb2.BatchVulnerabilityList.FromString,\n )\n\n\nclass OSVServicer(object):\n \"\"\"Open source vulnerability database.\n \"\"\"\n\n def GetVulnById(self, request, context):\n \"\"\"Return a `Vulnerability` object for a given OSV ID.\n \"\"\"\n context.set_code(grpc.StatusCode.UNIMPLEMENTED)\n context.set_details('Method not implemented!')\n raise NotImplementedError('Method not implemented!')\n\n def QueryAffected(self, request, context):\n \"\"\"Query vulnerabilities for a particular project at a given commit or\n version.\n \"\"\"\n context.set_code(grpc.StatusCode.UNIMPLEMENTED)\n context.set_details('Method not implemented!')\n raise NotImplementedError('Method not implemented!')\n\n def QueryAffectedBatch(self, request, context):\n \"\"\"Query vulnerabilities for a particular project at given commits and\n versions.\n \"\"\"\n context.set_code(grpc.StatusCode.UNIMPLEMENTED)\n context.set_details('Method not implemented!')\n raise NotImplementedError('Method not implemented!')\n\n\ndef add_OSVServicer_to_server(servicer, server):\n rpc_method_handlers = {\n 'GetVulnById': grpc.unary_unary_rpc_method_handler(\n servicer.GetVulnById,\n request_deserializer=osv__service__v1__pb2.GetVulnByIdParameters.FromString,\n response_serializer=osv_dot_vulnerability__pb2.Vulnerability.SerializeToString,\n ),\n 'QueryAffected': grpc.unary_unary_rpc_method_handler(\n servicer.QueryAffected,\n request_deserializer=osv__service__v1__pb2.QueryAffectedParameters.FromString,\n response_serializer=osv__service__v1__pb2.VulnerabilityList.SerializeToString,\n ),\n 'QueryAffectedBatch': grpc.unary_unary_rpc_method_handler(\n servicer.QueryAffectedBatch,\n request_deserializer=osv__service__v1__pb2.QueryAffectedBatchParameters.FromString,\n response_serializer=osv__service__v1__pb2.BatchVulnerabilityList.SerializeToString,\n ),\n }\n generic_handler = grpc.method_handlers_generic_handler(\n 'osv.v1.OSV', rpc_method_handlers)\n server.add_generic_rpc_handlers((generic_handler,))\n\n\n # This class is part of an EXPERIMENTAL API.\nclass OSV(object):\n \"\"\"Open source vulnerability database.\n \"\"\"\n\n @staticmethod\n def GetVulnById(request,\n target,\n options=(),\n channel_credentials=None,\n call_credentials=None,\n insecure=False,\n compression=None,\n wait_for_ready=None,\n timeout=None,\n metadata=None):\n return grpc.experimental.unary_unary(request, target, '/osv.v1.OSV/GetVulnById',\n osv__service__v1__pb2.GetVulnByIdParameters.SerializeToString,\n osv_dot_vulnerability__pb2.Vulnerability.FromString,\n options, channel_credentials,\n insecure, call_credentials, compression, wait_for_ready, timeout, metadata)\n\n @staticmethod\n def QueryAffected(request,\n target,\n options=(),\n channel_credentials=None,\n call_credentials=None,\n insecure=False,\n compression=None,\n wait_for_ready=None,\n timeout=None,\n metadata=None):\n return grpc.experimental.unary_unary(request, target, '/osv.v1.OSV/QueryAffected',\n osv__service__v1__pb2.QueryAffectedParameters.SerializeToString,\n osv__service__v1__pb2.VulnerabilityList.FromString,\n options, channel_credentials,\n insecure, call_credentials, compression, wait_for_ready, timeout, metadata)\n\n @staticmethod\n def QueryAffectedBatch(request,\n target,\n options=(),\n channel_credentials=None,\n call_credentials=None,\n insecure=False,\n compression=None,\n wait_for_ready=None,\n timeout=None,\n metadata=None):\n return grpc.experimental.unary_unary(request, target, '/osv.v1.OSV/QueryAffectedBatch',\n osv__service__v1__pb2.QueryAffectedBatchParameters.SerializeToString,\n osv__service__v1__pb2.BatchVulnerabilityList.FromString,\n options, channel_credentials,\n insecure, call_credentials, compression, wait_for_ready, timeout, metadata)\n", "path": "gcp/api/osv_service_v1_pb2_grpc.py"}]}
| 3,549 | 677 |
gh_patches_debug_10980
|
rasdani/github-patches
|
git_diff
|
rucio__rucio-1478
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
readthedocs build fails due to pycurl
Motivation
----------
Currently the build process on readthedocs fails.
```
Collecting pycurl>=7.19 (from fts3-rest-API==3.7.1->-r requirements.readthedocs.txt (line 44))
Using cached https://files.pythonhosted.org/packages/e8/e4/0dbb8735407189f00b33d84122b9be52c790c7c3b25286826f4e1bdb7bde/pycurl-7.43.0.2.tar.gz
Complete output from command python setup.py egg_info:
Traceback (most recent call last):
File "<string>", line 1, in <module>
File "/tmp/pip-build-jSh8ZH/pycurl/setup.py", line 913, in <module>
ext = get_extension(sys.argv, split_extension_source=split_extension_source)
File "/tmp/pip-build-jSh8ZH/pycurl/setup.py", line 582, in get_extension
ext_config = ExtensionConfiguration(argv)
File "/tmp/pip-build-jSh8ZH/pycurl/setup.py", line 99, in __init__
self.configure()
File "/tmp/pip-build-jSh8ZH/pycurl/setup.py", line 227, in configure_unix
raise ConfigurationError(msg)
__main__.ConfigurationError: Could not run curl-config: [Errno 2] No such file or directory
----------------------------------------
Command "python setup.py egg_info" failed with error code 1 in /tmp/pip-build-jSh8ZH/pycurl/
You are using pip version 9.0.1, however version 18.0 is available.
You should consider upgrading via the 'pip install --upgrade pip' command.
```
</issue>
<code>
[start of doc/source/conf.py]
1 # -*- coding: utf-8 -*-
2 #
3 # Rucio documentation build configuration file, created by
4 # sphinx-quickstart on Fri Oct 27 14:25:40 2017.
5 #
6 # This file is execfile()d with the current directory set to its
7 # containing dir.
8 #
9 # Note that not all possible configuration values are present in this
10 # autogenerated file.
11 #
12 # All configuration values have a default; values that are commented out
13 # serve to show the default.
14
15 # If extensions (or modules to document with autodoc) are in another directory,
16 # add these directories to sys.path here. If the directory is relative to the
17 # documentation root, use os.path.abspath to make it absolute, like shown here.
18 #
19 import os
20 import sys
21 # sys.path.insert(0, os.path.abspath('.'))
22
23 from mock import Mock as MagicMock
24
25
26 class Mock(MagicMock):
27 @classmethod
28 def __getattr__(cls, name):
29 return Mock()
30
31 @classmethod
32 def __getitem__(cls, name):
33 return Mock()
34
35
36 MOCK_MODULES = ['pycurl']
37 sys.modules.update((mod_name, Mock()) for mod_name in MOCK_MODULES)
38
39 # -- General configuration ------------------------------------------------
40
41 # If your documentation needs a minimal Sphinx version, state it here.
42 #
43 # needs_sphinx = '1.0'
44
45 # Add any Sphinx extension module names here, as strings. They can be
46 # extensions coming with Sphinx (named 'sphinx.ext.*') or your custom
47 # ones.
48 extensions = ['sphinx.ext.autodoc',
49 'sphinx.ext.doctest',
50 'sphinx.ext.todo',
51 'rucio.common.doc.argparse.ext',
52 'sphinxcontrib.httpdomain',
53 'sphinxcontrib.autohttp.flask',
54 'sphinxcontrib.autohttp.flaskqref']
55
56
57 on_rtd = os.environ.get('READTHEDOCS', None) == 'True'
58
59 if not on_rtd: # only import and set the theme if we're building docs locally
60 import sphinx_rtd_theme
61 html_theme = 'sphinx_rtd_theme'
62 html_theme_path = [sphinx_rtd_theme.get_html_theme_path()]
63
64
65 # Add any paths that contain templates here, relative to this directory.
66 templates_path = ['_templates']
67
68 # The suffix(es) of source filenames.
69 # You can specify multiple suffix as a list of string:
70 #
71 # source_suffix = ['.rst', '.md']
72 source_suffix = '.rst'
73
74 # The master toctree document.
75 master_doc = 'index'
76
77 # General information about the project.
78 project = u'Rucio'
79 copyright = u'2012-2018 CERN for the benefit of the ATLAS collaboration'
80 author = u'[email protected]'
81
82 # The version info for the project you're documenting, acts as replacement for
83 # |version| and |release|, also used in various other places throughout the
84 # built documents.
85 #
86 # The short X.Y version.
87 version = u''
88 # The full version, including alpha/beta/rc tags.
89 release = u'1.2'
90
91 # The language for content autogenerated by Sphinx. Refer to documentation
92 # for a list of supported languages.
93 #
94 # This is also used if you do content translation via gettext catalogs.
95 # Usually you set "language" from the command line for these cases.
96 language = None
97
98 # List of patterns, relative to source directory, that match files and
99 # directories to ignore when looking for source files.
100 # This patterns also effect to html_static_path and html_extra_path
101 exclude_patterns = []
102
103 # The name of the Pygments (syntax highlighting) style to use.
104 pygments_style = 'sphinx'
105
106 # If true, `todo` and `todoList` produce output, else they produce nothing.
107 todo_include_todos = True
108
109
110 # -- Options for HTML output ----------------------------------------------
111
112 # The theme to use for HTML and HTML Help pages. See the documentation for
113 # a list of builtin themes.
114 #
115 # html_theme = 'alabaster'
116
117 # Theme options are theme-specific and customize the look and feel of a theme
118 # further. For a list of options available for each theme, see the
119 # documentation.
120 #
121 # html_theme_options = {}
122
123 # Add any paths that contain custom static files (such as style sheets) here,
124 # relative to this directory. They are copied after the builtin static files,
125 # so a file named "default.css" will overwrite the builtin "default.css".
126 # html_static_path = ['_static']
127
128 # Custom sidebar templates, must be a dictionary that maps document names
129 # to template names.
130 #
131 # This is required for the alabaster theme
132 # refs: http://alabaster.readthedocs.io/en/latest/installation.html#sidebars
133 # html_sidebars = {
134 # '**': [
135 # 'about.html',
136 # 'navigation.html',
137 # 'relations.html', # needs 'show_related': True theme option to display
138 # 'searchbox.html',
139 # 'donate.html',
140 # ]
141 # }
142
143
144 # -- Options for HTMLHelp output ------------------------------------------
145
146 # Output file base name for HTML help builder.
147 htmlhelp_basename = 'Ruciodoc'
148
149
150 # -- Options for LaTeX output ---------------------------------------------
151
152 latex_elements = {
153 # The paper size ('letterpaper' or 'a4paper').
154 #
155 # 'papersize': 'letterpaper',
156
157 # The font size ('10pt', '11pt' or '12pt').
158 #
159 # 'pointsize': '10pt',
160
161 # Additional stuff for the LaTeX preamble.
162 #
163 # 'preamble': '',
164
165 # Latex figure (float) alignment
166 #
167 # 'figure_align': 'htbp',
168 }
169
170 # Grouping the document tree into LaTeX files. List of tuples
171 # (source start file, target name, title,
172 # author, documentclass [howto, manual, or own class]).
173 latex_documents = [
174 (master_doc, 'Rucio.tex', u'Rucio Documentation',
175 u'[email protected]', 'manual'),
176 ]
177
178
179 # -- Options for manual page output ---------------------------------------
180
181 # One entry per manual page. List of tuples
182 # (source start file, name, description, authors, manual section).
183 man_pages = [
184 (master_doc, 'rucio', u'Rucio Documentation',
185 [author], 1)
186 ]
187
188
189 # -- Options for Texinfo output -------------------------------------------
190
191 # Grouping the document tree into Texinfo files. List of tuples
192 # (source start file, target name, title, author,
193 # dir menu entry, description, category)
194 texinfo_documents = [
195 (master_doc, 'Rucio', u'Rucio Documentation',
196 author, 'Rucio', 'One line description of project.',
197 'Miscellaneous'),
198 ]
199
[end of doc/source/conf.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/doc/source/conf.py b/doc/source/conf.py
--- a/doc/source/conf.py
+++ b/doc/source/conf.py
@@ -17,9 +17,24 @@
# documentation root, use os.path.abspath to make it absolute, like shown here.
#
import os
-# import sys
+import sys
# sys.path.insert(0, os.path.abspath('.'))
+from mock import Mock as MagicMock
+
+
+class Mock(MagicMock):
+ @classmethod
+ def __getattr__(cls, name):
+ return Mock()
+
+ @classmethod
+ def __getitem__(cls, name):
+ return Mock()
+
+
+MOCK_MODULES = ['pycurl', 'M2Crypto']
+sys.modules.update((mod_name, Mock()) for mod_name in MOCK_MODULES)
# -- General configuration ------------------------------------------------
|
{"golden_diff": "diff --git a/doc/source/conf.py b/doc/source/conf.py\n--- a/doc/source/conf.py\n+++ b/doc/source/conf.py\n@@ -17,9 +17,24 @@\n # documentation root, use os.path.abspath to make it absolute, like shown here.\n #\n import os\n-# import sys\n+import sys\n # sys.path.insert(0, os.path.abspath('.'))\n \n+from mock import Mock as MagicMock\n+\n+\n+class Mock(MagicMock):\n+ @classmethod\n+ def __getattr__(cls, name):\n+ return Mock()\n+\n+ @classmethod\n+ def __getitem__(cls, name):\n+ return Mock()\n+\n+\n+MOCK_MODULES = ['pycurl', 'M2Crypto']\n+sys.modules.update((mod_name, Mock()) for mod_name in MOCK_MODULES)\n \n # -- General configuration ------------------------------------------------\n", "issue": "readthedocs build fails due to pycurl\nMotivation\r\n----------\r\nCurrently the build process on readthedocs fails.\r\n\r\n```\r\nCollecting pycurl>=7.19 (from fts3-rest-API==3.7.1->-r requirements.readthedocs.txt (line 44))\r\n Using cached https://files.pythonhosted.org/packages/e8/e4/0dbb8735407189f00b33d84122b9be52c790c7c3b25286826f4e1bdb7bde/pycurl-7.43.0.2.tar.gz\r\n Complete output from command python setup.py egg_info:\r\n Traceback (most recent call last):\r\n File \"<string>\", line 1, in <module>\r\n File \"/tmp/pip-build-jSh8ZH/pycurl/setup.py\", line 913, in <module>\r\n ext = get_extension(sys.argv, split_extension_source=split_extension_source)\r\n File \"/tmp/pip-build-jSh8ZH/pycurl/setup.py\", line 582, in get_extension\r\n ext_config = ExtensionConfiguration(argv)\r\n File \"/tmp/pip-build-jSh8ZH/pycurl/setup.py\", line 99, in __init__\r\n self.configure()\r\n File \"/tmp/pip-build-jSh8ZH/pycurl/setup.py\", line 227, in configure_unix\r\n raise ConfigurationError(msg)\r\n __main__.ConfigurationError: Could not run curl-config: [Errno 2] No such file or directory\r\n \r\n ----------------------------------------\r\nCommand \"python setup.py egg_info\" failed with error code 1 in /tmp/pip-build-jSh8ZH/pycurl/\r\nYou are using pip version 9.0.1, however version 18.0 is available.\r\nYou should consider upgrading via the 'pip install --upgrade pip' command.\r\n```\r\n\n", "before_files": [{"content": "# -*- coding: utf-8 -*-\n#\n# Rucio documentation build configuration file, created by\n# sphinx-quickstart on Fri Oct 27 14:25:40 2017.\n#\n# This file is execfile()d with the current directory set to its\n# containing dir.\n#\n# Note that not all possible configuration values are present in this\n# autogenerated file.\n#\n# All configuration values have a default; values that are commented out\n# serve to show the default.\n\n# If extensions (or modules to document with autodoc) are in another directory,\n# add these directories to sys.path here. If the directory is relative to the\n# documentation root, use os.path.abspath to make it absolute, like shown here.\n#\nimport os\nimport sys\n# sys.path.insert(0, os.path.abspath('.'))\n\nfrom mock import Mock as MagicMock\n\n\nclass Mock(MagicMock):\n @classmethod\n def __getattr__(cls, name):\n return Mock()\n\n @classmethod\n def __getitem__(cls, name):\n return Mock()\n\n\nMOCK_MODULES = ['pycurl']\nsys.modules.update((mod_name, Mock()) for mod_name in MOCK_MODULES)\n\n# -- General configuration ------------------------------------------------\n\n# If your documentation needs a minimal Sphinx version, state it here.\n#\n# needs_sphinx = '1.0'\n\n# Add any Sphinx extension module names here, as strings. They can be\n# extensions coming with Sphinx (named 'sphinx.ext.*') or your custom\n# ones.\nextensions = ['sphinx.ext.autodoc',\n 'sphinx.ext.doctest',\n 'sphinx.ext.todo',\n 'rucio.common.doc.argparse.ext',\n 'sphinxcontrib.httpdomain',\n 'sphinxcontrib.autohttp.flask',\n 'sphinxcontrib.autohttp.flaskqref']\n\n\non_rtd = os.environ.get('READTHEDOCS', None) == 'True'\n\nif not on_rtd: # only import and set the theme if we're building docs locally\n import sphinx_rtd_theme\n html_theme = 'sphinx_rtd_theme'\n html_theme_path = [sphinx_rtd_theme.get_html_theme_path()]\n\n\n# Add any paths that contain templates here, relative to this directory.\ntemplates_path = ['_templates']\n\n# The suffix(es) of source filenames.\n# You can specify multiple suffix as a list of string:\n#\n# source_suffix = ['.rst', '.md']\nsource_suffix = '.rst'\n\n# The master toctree document.\nmaster_doc = 'index'\n\n# General information about the project.\nproject = u'Rucio'\ncopyright = u'2012-2018 CERN for the benefit of the ATLAS collaboration'\nauthor = u'[email protected]'\n\n# The version info for the project you're documenting, acts as replacement for\n# |version| and |release|, also used in various other places throughout the\n# built documents.\n#\n# The short X.Y version.\nversion = u''\n# The full version, including alpha/beta/rc tags.\nrelease = u'1.2'\n\n# The language for content autogenerated by Sphinx. Refer to documentation\n# for a list of supported languages.\n#\n# This is also used if you do content translation via gettext catalogs.\n# Usually you set \"language\" from the command line for these cases.\nlanguage = None\n\n# List of patterns, relative to source directory, that match files and\n# directories to ignore when looking for source files.\n# This patterns also effect to html_static_path and html_extra_path\nexclude_patterns = []\n\n# The name of the Pygments (syntax highlighting) style to use.\npygments_style = 'sphinx'\n\n# If true, `todo` and `todoList` produce output, else they produce nothing.\ntodo_include_todos = True\n\n\n# -- Options for HTML output ----------------------------------------------\n\n# The theme to use for HTML and HTML Help pages. See the documentation for\n# a list of builtin themes.\n#\n# html_theme = 'alabaster'\n\n# Theme options are theme-specific and customize the look and feel of a theme\n# further. For a list of options available for each theme, see the\n# documentation.\n#\n# html_theme_options = {}\n\n# Add any paths that contain custom static files (such as style sheets) here,\n# relative to this directory. They are copied after the builtin static files,\n# so a file named \"default.css\" will overwrite the builtin \"default.css\".\n# html_static_path = ['_static']\n\n# Custom sidebar templates, must be a dictionary that maps document names\n# to template names.\n#\n# This is required for the alabaster theme\n# refs: http://alabaster.readthedocs.io/en/latest/installation.html#sidebars\n# html_sidebars = {\n# '**': [\n# 'about.html',\n# 'navigation.html',\n# 'relations.html', # needs 'show_related': True theme option to display\n# 'searchbox.html',\n# 'donate.html',\n# ]\n# }\n\n\n# -- Options for HTMLHelp output ------------------------------------------\n\n# Output file base name for HTML help builder.\nhtmlhelp_basename = 'Ruciodoc'\n\n\n# -- Options for LaTeX output ---------------------------------------------\n\nlatex_elements = {\n # The paper size ('letterpaper' or 'a4paper').\n #\n # 'papersize': 'letterpaper',\n\n # The font size ('10pt', '11pt' or '12pt').\n #\n # 'pointsize': '10pt',\n\n # Additional stuff for the LaTeX preamble.\n #\n # 'preamble': '',\n\n # Latex figure (float) alignment\n #\n # 'figure_align': 'htbp',\n}\n\n# Grouping the document tree into LaTeX files. List of tuples\n# (source start file, target name, title,\n# author, documentclass [howto, manual, or own class]).\nlatex_documents = [\n (master_doc, 'Rucio.tex', u'Rucio Documentation',\n u'[email protected]', 'manual'),\n]\n\n\n# -- Options for manual page output ---------------------------------------\n\n# One entry per manual page. List of tuples\n# (source start file, name, description, authors, manual section).\nman_pages = [\n (master_doc, 'rucio', u'Rucio Documentation',\n [author], 1)\n]\n\n\n# -- Options for Texinfo output -------------------------------------------\n\n# Grouping the document tree into Texinfo files. List of tuples\n# (source start file, target name, title, author,\n# dir menu entry, description, category)\ntexinfo_documents = [\n (master_doc, 'Rucio', u'Rucio Documentation',\n author, 'Rucio', 'One line description of project.',\n 'Miscellaneous'),\n]\n", "path": "doc/source/conf.py"}]}
| 2,907 | 182 |
gh_patches_debug_39416
|
rasdani/github-patches
|
git_diff
|
buildbot__buildbot-244
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Don't subclass ThreadPool.
</issue>
<code>
[start of master/contrib/googlecode_atom.py]
1 # GoogleCode Atom Feed Poller
2 # Author: Srivats P. <pstavirs>
3 # Based on Mozilla's HgPoller
4 # http://bonsai.mozilla.org/cvsblame.cgi?file=/mozilla/tools/buildbot/buildbot/changes/Attic/hgpoller.py&revision=1.1.4.2
5 #
6 # Description:
7 # Use this ChangeSource for projects hosted on http://code.google.com/
8 #
9 # This ChangeSource uses the project's commit Atom feed. Depending upon the
10 # frequency of commits, you can tune the polling interval for the feed
11 # (default is 1 hour)
12 #
13 # Parameters:
14 # feedurl (MANDATORY): The Atom feed URL of the GoogleCode repo
15 # pollinterval (OPTIONAL): Polling frequency for the feed (in seconds)
16 #
17 # Example:
18 # To poll the Ostinato project's commit feed every 3 hours, use -
19 # from googlecode_atom import GoogleCodeAtomPoller
20 # poller = GoogleCodeAtomPoller(
21 # feedurl="http://code.google.com/feeds/p/ostinato/hgchanges/basic",
22 # pollinterval=10800)
23 # c['change_source'] = [ poller ]
24 #
25
26 from time import strptime
27 from calendar import timegm
28 from xml.dom import minidom, Node
29
30 from twisted.python import log, failure
31 from twisted.internet import defer, reactor
32 from twisted.internet.task import LoopingCall
33 from twisted.web.client import getPage
34
35 from buildbot.changes import base, changes
36
37 def googleCodePollerForProject(project, vcs, pollinterval=3600):
38 return GoogleCodeAtomPoller(
39 'http://code.google.com/feeds/p/%s/%schanges/basic' % (project, vcs),
40 pollinterval=pollinterval)
41
42
43 class GoogleCodeAtomPoller(base.ChangeSource):
44 """This source will poll a GoogleCode Atom feed for changes and
45 submit them to the change master. Works for both Svn, Git, and Hg
46 repos.
47 TODO: branch processing
48 """
49
50 compare_attrs = ['feedurl', 'pollinterval']
51 parent = None
52 loop = None
53 volatile = ['loop']
54 working = False
55
56 def __init__(self, feedurl, pollinterval=3600):
57 """
58 @type feedurl: string
59 @param feedurl: The Atom feed URL of the GoogleCode repo
60 (e.g. http://code.google.com/feeds/p/ostinato/hgchanges/basic)
61
62 @type pollinterval: int
63 @param pollinterval: The time (in seconds) between queries for
64 changes (default is 1 hour)
65 """
66
67 self.feedurl = feedurl
68 self.branch = None
69 self.pollinterval = pollinterval
70 self.lastChange = None
71 self.loop = LoopingCall(self.poll)
72 self.src = None
73 for word in self.feedurl.split('/'):
74 if word == 'svnchanges':
75 self.src = 'svn'
76 break
77 elif word == 'hgchanges':
78 self.src = 'hg'
79 break
80 elif word == 'gitchanges':
81 self.src = 'git'
82 break
83
84 def startService(self):
85 log.msg("GoogleCodeAtomPoller starting")
86 base.ChangeSource.startService(self)
87 reactor.callLater(0, self.loop.start, self.pollinterval)
88
89 def stopService(self):
90 log.msg("GoogleCodeAtomPoller stoppping")
91 self.loop.stop()
92 return base.ChangeSource.stopService(self)
93
94 def describe(self):
95 return ("Getting changes from the GoogleCode repo changes feed %s" %
96 self._make_url())
97
98 def poll(self):
99 if self.working:
100 log.msg("Not polling because last poll is still working")
101 else:
102 self.working = True
103 d = self._get_changes()
104 d.addCallback(self._process_changes)
105 d.addCallbacks(self._finished_ok, self._finished_failure)
106
107 def _finished_ok(self, res):
108 assert self.working
109 self.working = False
110 log.msg("GoogleCodeAtomPoller poll success")
111
112 return res
113
114 def _finished_failure(self, res):
115 log.msg("GoogleCodeAtomPoller poll failed: %s" % res)
116 assert self.working
117 self.working = False
118 return None
119
120 def _make_url(self):
121 return "%s" % (self.feedurl)
122
123 def _get_changes(self):
124 url = self._make_url()
125 log.msg("GoogleCodeAtomPoller polling %s" % url)
126
127 return getPage(url, timeout=self.pollinterval)
128
129 def _parse_changes(self, query):
130 dom = minidom.parseString(query)
131 entries = dom.getElementsByTagName("entry")
132 changes = []
133 # Entries come in reverse chronological order
134 for i in entries:
135 d = {}
136
137 # revision is the last part of the 'id' url
138 d["revision"] = i.getElementsByTagName(
139 "id")[0].firstChild.data.split('/')[-1]
140 if d["revision"] == self.lastChange:
141 break # no more new changes
142
143 d["when"] = timegm(strptime(
144 i.getElementsByTagName("updated")[0].firstChild.data,
145 "%Y-%m-%dT%H:%M:%SZ"))
146 d["author"] = i.getElementsByTagName(
147 "author")[0].getElementsByTagName("name")[0].firstChild.data
148 # files and commit msg are separated by 2 consecutive <br/>
149 content = i.getElementsByTagName(
150 "content")[0].firstChild.data.split("<br/>\n <br/>")
151 # Remove the action keywords from the file list
152 fl = content[0].replace(
153 u' \xa0\xa0\xa0\xa0Add\xa0\xa0\xa0\xa0', '').replace(
154 u' \xa0\xa0\xa0\xa0Delete\xa0\xa0\xa0\xa0', '').replace(
155 u' \xa0\xa0\xa0\xa0Modify\xa0\xa0\xa0\xa0', '')
156 # Get individual files and remove the 'header'
157 d["files"] = fl.encode("ascii", "replace").split("<br/>")[1:]
158 d["files"] = [f.strip() for f in d["files"]]
159 try:
160 d["comments"] = content[1].encode("ascii", "replace")
161 except:
162 d["comments"] = "No commit message provided"
163
164 changes.append(d)
165
166 changes.reverse() # want them in chronological order
167 return changes
168
169 def _process_changes(self, query):
170 change_list = self._parse_changes(query)
171
172 # Skip calling addChange() if this is the first successful poll.
173 if self.lastChange is not None:
174 for change in change_list:
175 c = changes.Change(revision = change["revision"],
176 who = change["author"],
177 files = change["files"],
178 comments = change["comments"],
179 when = change["when"],
180 branch = self.branch)
181 self.parent.addChange(c, src=self.src)
182 if change_list:
183 self.lastChange = change_list[-1]["revision"]
184
[end of master/contrib/googlecode_atom.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/master/contrib/googlecode_atom.py b/master/contrib/googlecode_atom.py
--- a/master/contrib/googlecode_atom.py
+++ b/master/contrib/googlecode_atom.py
@@ -23,8 +23,7 @@
# c['change_source'] = [ poller ]
#
-from time import strptime
-from calendar import timegm
+import datetime
from xml.dom import minidom, Node
from twisted.python import log, failure
@@ -32,7 +31,8 @@
from twisted.internet.task import LoopingCall
from twisted.web.client import getPage
-from buildbot.changes import base, changes
+from buildbot.changes import base
+
def googleCodePollerForProject(project, vcs, pollinterval=3600):
return GoogleCodeAtomPoller(
@@ -140,9 +140,9 @@
if d["revision"] == self.lastChange:
break # no more new changes
- d["when"] = timegm(strptime(
+ d["when"] = datetime.datetime.strptime(
i.getElementsByTagName("updated")[0].firstChild.data,
- "%Y-%m-%dT%H:%M:%SZ"))
+ "%Y-%m-%dT%H:%M:%SZ")
d["author"] = i.getElementsByTagName(
"author")[0].getElementsByTagName("name")[0].firstChild.data
# files and commit msg are separated by 2 consecutive <br/>
@@ -166,18 +166,22 @@
changes.reverse() # want them in chronological order
return changes
+ @defer.deferredGenerator
def _process_changes(self, query):
change_list = self._parse_changes(query)
# Skip calling addChange() if this is the first successful poll.
if self.lastChange is not None:
for change in change_list:
- c = changes.Change(revision = change["revision"],
- who = change["author"],
- files = change["files"],
- comments = change["comments"],
- when = change["when"],
- branch = self.branch)
- self.parent.addChange(c, src=self.src)
+ d = self.master.addChange(author=change["author"],
+ revision=change["revision"],
+ files=change["files"],
+ comments=change["comments"],
+ when_timestamp=change["when"],
+ branch=self.branch,
+ src=self.src)
+ wfd = defer.waitForDeferred(d)
+ yield wfd
+ results = wfd.getResult()
if change_list:
self.lastChange = change_list[-1]["revision"]
|
{"golden_diff": "diff --git a/master/contrib/googlecode_atom.py b/master/contrib/googlecode_atom.py\n--- a/master/contrib/googlecode_atom.py\n+++ b/master/contrib/googlecode_atom.py\n@@ -23,8 +23,7 @@\n # c['change_source'] = [ poller ]\n #\n \n-from time import strptime\n-from calendar import timegm\n+import datetime\n from xml.dom import minidom, Node\n \n from twisted.python import log, failure\n@@ -32,7 +31,8 @@\n from twisted.internet.task import LoopingCall\n from twisted.web.client import getPage\n \n-from buildbot.changes import base, changes\n+from buildbot.changes import base\n+\n \n def googleCodePollerForProject(project, vcs, pollinterval=3600):\n return GoogleCodeAtomPoller(\n@@ -140,9 +140,9 @@\n if d[\"revision\"] == self.lastChange:\n break # no more new changes\n \n- d[\"when\"] = timegm(strptime(\n+ d[\"when\"] = datetime.datetime.strptime(\n i.getElementsByTagName(\"updated\")[0].firstChild.data,\n- \"%Y-%m-%dT%H:%M:%SZ\"))\n+ \"%Y-%m-%dT%H:%M:%SZ\")\n d[\"author\"] = i.getElementsByTagName(\n \"author\")[0].getElementsByTagName(\"name\")[0].firstChild.data\n # files and commit msg are separated by 2 consecutive <br/>\n@@ -166,18 +166,22 @@\n changes.reverse() # want them in chronological order\n return changes\n \n+ @defer.deferredGenerator\n def _process_changes(self, query):\n change_list = self._parse_changes(query)\n \n # Skip calling addChange() if this is the first successful poll.\n if self.lastChange is not None:\n for change in change_list:\n- c = changes.Change(revision = change[\"revision\"],\n- who = change[\"author\"],\n- files = change[\"files\"],\n- comments = change[\"comments\"],\n- when = change[\"when\"],\n- branch = self.branch)\n- self.parent.addChange(c, src=self.src)\n+ d = self.master.addChange(author=change[\"author\"],\n+ revision=change[\"revision\"],\n+ files=change[\"files\"],\n+ comments=change[\"comments\"],\n+ when_timestamp=change[\"when\"],\n+ branch=self.branch,\n+ src=self.src)\n+ wfd = defer.waitForDeferred(d)\n+ yield wfd\n+ results = wfd.getResult()\n if change_list:\n self.lastChange = change_list[-1][\"revision\"]\n", "issue": "Don't subclass ThreadPool.\n\n", "before_files": [{"content": "# GoogleCode Atom Feed Poller\n# Author: Srivats P. <pstavirs>\n# Based on Mozilla's HgPoller\n# http://bonsai.mozilla.org/cvsblame.cgi?file=/mozilla/tools/buildbot/buildbot/changes/Attic/hgpoller.py&revision=1.1.4.2\n#\n# Description:\n# Use this ChangeSource for projects hosted on http://code.google.com/\n#\n# This ChangeSource uses the project's commit Atom feed. Depending upon the\n# frequency of commits, you can tune the polling interval for the feed\n# (default is 1 hour)\n#\n# Parameters:\n# feedurl (MANDATORY): The Atom feed URL of the GoogleCode repo\n# pollinterval (OPTIONAL): Polling frequency for the feed (in seconds)\n#\n# Example:\n# To poll the Ostinato project's commit feed every 3 hours, use -\n# from googlecode_atom import GoogleCodeAtomPoller\n# poller = GoogleCodeAtomPoller(\n# feedurl=\"http://code.google.com/feeds/p/ostinato/hgchanges/basic\",\n# pollinterval=10800)\n# c['change_source'] = [ poller ]\n#\n\nfrom time import strptime\nfrom calendar import timegm\nfrom xml.dom import minidom, Node\n\nfrom twisted.python import log, failure\nfrom twisted.internet import defer, reactor\nfrom twisted.internet.task import LoopingCall\nfrom twisted.web.client import getPage\n\nfrom buildbot.changes import base, changes\n\ndef googleCodePollerForProject(project, vcs, pollinterval=3600):\n return GoogleCodeAtomPoller(\n 'http://code.google.com/feeds/p/%s/%schanges/basic' % (project, vcs),\n pollinterval=pollinterval)\n\n\nclass GoogleCodeAtomPoller(base.ChangeSource):\n \"\"\"This source will poll a GoogleCode Atom feed for changes and\n submit them to the change master. Works for both Svn, Git, and Hg\n repos.\n TODO: branch processing\n \"\"\"\n\n compare_attrs = ['feedurl', 'pollinterval']\n parent = None\n loop = None\n volatile = ['loop']\n working = False\n\n def __init__(self, feedurl, pollinterval=3600):\n \"\"\"\n @type feedurl: string\n @param feedurl: The Atom feed URL of the GoogleCode repo\n (e.g. http://code.google.com/feeds/p/ostinato/hgchanges/basic)\n\n @type pollinterval: int\n @param pollinterval: The time (in seconds) between queries for\n changes (default is 1 hour)\n \"\"\"\n\n self.feedurl = feedurl\n self.branch = None\n self.pollinterval = pollinterval\n self.lastChange = None\n self.loop = LoopingCall(self.poll)\n self.src = None\n for word in self.feedurl.split('/'):\n if word == 'svnchanges':\n self.src = 'svn'\n break\n elif word == 'hgchanges':\n self.src = 'hg'\n break\n elif word == 'gitchanges':\n self.src = 'git'\n break\n\n def startService(self):\n log.msg(\"GoogleCodeAtomPoller starting\")\n base.ChangeSource.startService(self)\n reactor.callLater(0, self.loop.start, self.pollinterval)\n\n def stopService(self):\n log.msg(\"GoogleCodeAtomPoller stoppping\")\n self.loop.stop()\n return base.ChangeSource.stopService(self)\n\n def describe(self):\n return (\"Getting changes from the GoogleCode repo changes feed %s\" %\n self._make_url())\n\n def poll(self):\n if self.working:\n log.msg(\"Not polling because last poll is still working\")\n else:\n self.working = True\n d = self._get_changes()\n d.addCallback(self._process_changes)\n d.addCallbacks(self._finished_ok, self._finished_failure)\n\n def _finished_ok(self, res):\n assert self.working\n self.working = False\n log.msg(\"GoogleCodeAtomPoller poll success\")\n\n return res\n\n def _finished_failure(self, res):\n log.msg(\"GoogleCodeAtomPoller poll failed: %s\" % res)\n assert self.working\n self.working = False\n return None\n\n def _make_url(self):\n return \"%s\" % (self.feedurl)\n\n def _get_changes(self):\n url = self._make_url()\n log.msg(\"GoogleCodeAtomPoller polling %s\" % url)\n\n return getPage(url, timeout=self.pollinterval)\n\n def _parse_changes(self, query):\n dom = minidom.parseString(query)\n entries = dom.getElementsByTagName(\"entry\")\n changes = []\n # Entries come in reverse chronological order\n for i in entries:\n d = {}\n\n # revision is the last part of the 'id' url\n d[\"revision\"] = i.getElementsByTagName(\n \"id\")[0].firstChild.data.split('/')[-1]\n if d[\"revision\"] == self.lastChange:\n break # no more new changes\n\n d[\"when\"] = timegm(strptime(\n i.getElementsByTagName(\"updated\")[0].firstChild.data,\n \"%Y-%m-%dT%H:%M:%SZ\"))\n d[\"author\"] = i.getElementsByTagName(\n \"author\")[0].getElementsByTagName(\"name\")[0].firstChild.data\n # files and commit msg are separated by 2 consecutive <br/>\n content = i.getElementsByTagName(\n \"content\")[0].firstChild.data.split(\"<br/>\\n <br/>\")\n # Remove the action keywords from the file list\n fl = content[0].replace(\n u' \\xa0\\xa0\\xa0\\xa0Add\\xa0\\xa0\\xa0\\xa0', '').replace(\n u' \\xa0\\xa0\\xa0\\xa0Delete\\xa0\\xa0\\xa0\\xa0', '').replace(\n u' \\xa0\\xa0\\xa0\\xa0Modify\\xa0\\xa0\\xa0\\xa0', '')\n # Get individual files and remove the 'header'\n d[\"files\"] = fl.encode(\"ascii\", \"replace\").split(\"<br/>\")[1:]\n d[\"files\"] = [f.strip() for f in d[\"files\"]]\n try:\n d[\"comments\"] = content[1].encode(\"ascii\", \"replace\")\n except:\n d[\"comments\"] = \"No commit message provided\"\n\n changes.append(d)\n\n changes.reverse() # want them in chronological order\n return changes\n\n def _process_changes(self, query):\n change_list = self._parse_changes(query)\n\n # Skip calling addChange() if this is the first successful poll.\n if self.lastChange is not None:\n for change in change_list:\n c = changes.Change(revision = change[\"revision\"],\n who = change[\"author\"],\n files = change[\"files\"],\n comments = change[\"comments\"],\n when = change[\"when\"],\n branch = self.branch)\n self.parent.addChange(c, src=self.src)\n if change_list:\n self.lastChange = change_list[-1][\"revision\"]\n", "path": "master/contrib/googlecode_atom.py"}]}
| 2,546 | 568 |
gh_patches_debug_4028
|
rasdani/github-patches
|
git_diff
|
diofant__diofant-852
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Project logo
There are few places to put images:
- [x] Top left corner of https://diofant.readthedocs.io/ (see e.g. https://sphinx-rtd-theme.readthedocs.io/)
- [x] favicon.ico.
- [x] logo for pdf logs
- [x] Organization profile on the Github (at least 200x200px)
- [x] ~~repository’s social media preview (640×320px - 1280×640px for best display)~~
</issue>
<code>
[start of docs/conf.py]
1 #
2 # Diofant documentation build configuration file.
3 #
4 # This file is execfile()d with the current directory set to its
5 # containing dir.
6 #
7 # The contents of this file are pickled, so don't put values in the
8 # namespace that aren't pickleable (module imports are okay, they're
9 # removed automatically).
10 #
11
12 import warnings
13
14 import diofant
15
16
17 # Turns numpydoc's section warnings to exceptions, see numpy/numpydoc#58.
18 warnings.simplefilter('error', UserWarning)
19
20 # Add any Sphinx extension module names here, as strings.
21 extensions = ['sphinx.ext.autodoc', 'sphinx.ext.viewcode', 'sphinx.ext.mathjax',
22 'sphinx.ext.graphviz', 'sphinx.ext.intersphinx',
23 'sphinx.ext.extlinks', 'sphinx.ext.napoleon',
24 'sphinxcontrib.bibtex']
25
26 napoleon_google_docstring = False
27 napoleon_use_param = False
28 napoleon_use_rtype = False
29
30 # Sphinx will warn about all references where the target cannot be found.
31 nitpicky = True
32
33 # Glob-style patterns that should be excluded when looking for sources.
34 exclude_patterns = ['README.rst']
35
36 # The document name of the "master" document, that is, the document
37 # that contains the root toctree directive.
38 master_doc = 'index'
39
40 # Project information.
41 project = 'Diofant'
42 copyright = '2006-2018 SymPy Development Team, 2013-2019 Sergey B Kirpichev'
43 version = diofant.__version__
44 release = version
45
46 # The name of default reST role, that is, for text marked up `like this`.
47 default_role = 'math'
48
49 # The theme to use for HTML and HTML Help pages.
50 html_theme = 'sphinx_rtd_theme'
51
52 # The LaTeX engine to build the docs.
53 latex_engine = 'xelatex'
54
55 # If True, the PDF build from the LaTeX files created by Sphinx will use xindy
56 # rather than makeindex.
57 latex_use_xindy = False
58
59 # This value determines how to group the document tree into LaTeX source
60 # files. It must be a list of tuples (startdocname, targetname, title,
61 # author, documentclass, toctree_only),
62 latex_documents = [('index', 'diofant.tex', 'Diofant Documentation',
63 'Diofant Development Team', 'manual', True)]
64
65 # A dictionary that contains LaTeX snippets that override predefined.
66 latex_elements = {
67 'preamble': r'''
68 \setmainfont{DejaVu Serif}
69 \setsansfont{DejaVu Sans}
70 \setmonofont{DejaVu Sans Mono}
71 % redefine \LaTeX to be usable in math mode
72 \expandafter\def\expandafter\LaTeX\expandafter{\expandafter\text\expandafter{\LaTeX}}
73 '''
74 }
75
76 # Add page references after internal references.
77 latex_show_pagerefs = True
78
79 # The output format for Graphviz when building HTML files.
80 graphviz_output_format = 'svg'
81
82 # Contains mapping the locations and names of other projects that
83 # should be linked to in this documentation.
84 intersphinx_mapping = {
85 'python3': ('https://docs.python.org/3/', None),
86 'numpy': ('https://docs.scipy.org/doc/numpy', None),
87 'scipy': ('https://docs.scipy.org/doc/scipy/reference', None),
88 }
89
90 # Dictionary of external sites, mapping unique short alias names to a
91 # base URL and a prefix.
92 extlinks = {
93 'issue': ('https://github.com/diofant/diofant/issues/%s', '#'),
94 'pull': ('https://github.com/diofant/diofant/pull/%s', '#'),
95 'commit': ('https://github.com/diofant/diofant/commit/%s', ''),
96 'sympyissue': ('https://github.com/sympy/sympy/issues/%s', 'sympy/sympy#'),
97 'sympypull': ('https://github.com/sympy/sympy/pull/%s', 'sympy/sympy#'),
98 }
99
100 # The number of times the linkcheck builder will attempt to check a URL
101 # before declaring it broken.
102 linkcheck_retries = 3
103
104 # A list of regular expressions that match URIs that should not be checked.
105 linkcheck_ignore = [r'https://primes.utm.edu/notes/gaps.html',
106 r'https://primes.utm.edu/glossary/xpage/BertrandsPostulate.html',
107 r'https://primes.utm.edu/prove/prove2_3.html',
108 r'https://primes.utm.edu/glossary/xpage/Pseudoprime.html']
109
110 # This value controls if docstring for classes or methods, if not explicitly
111 # set, is inherited form parents.
112 autodoc_inherit_docstrings = False
113
114 # A list of paths that contain custom static files. Relative paths are taken as
115 # relative to the configuration directory. They are copied to the output’s
116 # _static directory.
117 html_static_path = ['_static']
118
119 # Should we show "Created using Sphinx" in the HTML footer?
120 html_show_sphinx = False
121
122 # Theme-specific options.
123 html_theme_options = {
124 'logo_only': True,
125 'display_version': False,
126 }
127
128 mathjax_config = {
129 'CommonHTML': {'linebreaks': {'automatic': True}},
130 'HTML-CSS': {'linebreaks': {'automatic': True}},
131 'SVG': {'linebreaks': {'automatic': True}},
132 }
133
134
135 # https://docs.readthedocs.io/en/latest/guides/adding-custom-css.html
136 def setup(app):
137 app.add_stylesheet('custom.css')
138
[end of docs/conf.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/docs/conf.py b/docs/conf.py
--- a/docs/conf.py
+++ b/docs/conf.py
@@ -119,6 +119,11 @@
# Should we show "Created using Sphinx" in the HTML footer?
html_show_sphinx = False
+# Paths to the logo and favicon.ico, relative to the conf.py's directory.
+html_logo = '_static/logo.svg'
+html_favicon = '_static/favicon.ico'
+latex_logo = '_static/logo.png'
+
# Theme-specific options.
html_theme_options = {
'logo_only': True,
|
{"golden_diff": "diff --git a/docs/conf.py b/docs/conf.py\n--- a/docs/conf.py\n+++ b/docs/conf.py\n@@ -119,6 +119,11 @@\n # Should we show \"Created using Sphinx\" in the HTML footer?\n html_show_sphinx = False\n \n+# Paths to the logo and favicon.ico, relative to the conf.py's directory.\n+html_logo = '_static/logo.svg'\n+html_favicon = '_static/favicon.ico'\n+latex_logo = '_static/logo.png'\n+\n # Theme-specific options.\n html_theme_options = {\n 'logo_only': True,\n", "issue": "Project logo\nThere are few places to put images:\r\n- [x] Top left corner of https://diofant.readthedocs.io/ (see e.g. https://sphinx-rtd-theme.readthedocs.io/)\r\n- [x] favicon.ico.\r\n- [x] logo for pdf logs\r\n- [x] Organization profile on the Github (at least 200x200px)\r\n- [x] ~~repository\u2019s social media preview (640\u00d7320px - 1280\u00d7640px for best display)~~\r\n\n", "before_files": [{"content": "#\n# Diofant documentation build configuration file.\n#\n# This file is execfile()d with the current directory set to its\n# containing dir.\n#\n# The contents of this file are pickled, so don't put values in the\n# namespace that aren't pickleable (module imports are okay, they're\n# removed automatically).\n#\n\nimport warnings\n\nimport diofant\n\n\n# Turns numpydoc's section warnings to exceptions, see numpy/numpydoc#58.\nwarnings.simplefilter('error', UserWarning)\n\n# Add any Sphinx extension module names here, as strings.\nextensions = ['sphinx.ext.autodoc', 'sphinx.ext.viewcode', 'sphinx.ext.mathjax',\n 'sphinx.ext.graphviz', 'sphinx.ext.intersphinx',\n 'sphinx.ext.extlinks', 'sphinx.ext.napoleon',\n 'sphinxcontrib.bibtex']\n\nnapoleon_google_docstring = False\nnapoleon_use_param = False\nnapoleon_use_rtype = False\n\n# Sphinx will warn about all references where the target cannot be found.\nnitpicky = True\n\n# Glob-style patterns that should be excluded when looking for sources.\nexclude_patterns = ['README.rst']\n\n# The document name of the \"master\" document, that is, the document\n# that contains the root toctree directive.\nmaster_doc = 'index'\n\n# Project information.\nproject = 'Diofant'\ncopyright = '2006-2018 SymPy Development Team, 2013-2019 Sergey B Kirpichev'\nversion = diofant.__version__\nrelease = version\n\n# The name of default reST role, that is, for text marked up `like this`.\ndefault_role = 'math'\n\n# The theme to use for HTML and HTML Help pages.\nhtml_theme = 'sphinx_rtd_theme'\n\n# The LaTeX engine to build the docs.\nlatex_engine = 'xelatex'\n\n# If True, the PDF build from the LaTeX files created by Sphinx will use xindy\n# rather than makeindex.\nlatex_use_xindy = False\n\n# This value determines how to group the document tree into LaTeX source\n# files. It must be a list of tuples (startdocname, targetname, title,\n# author, documentclass, toctree_only),\nlatex_documents = [('index', 'diofant.tex', 'Diofant Documentation',\n 'Diofant Development Team', 'manual', True)]\n\n# A dictionary that contains LaTeX snippets that override predefined.\nlatex_elements = {\n 'preamble': r'''\n\\setmainfont{DejaVu Serif}\n\\setsansfont{DejaVu Sans}\n\\setmonofont{DejaVu Sans Mono}\n% redefine \\LaTeX to be usable in math mode\n\\expandafter\\def\\expandafter\\LaTeX\\expandafter{\\expandafter\\text\\expandafter{\\LaTeX}}\n'''\n}\n\n# Add page references after internal references.\nlatex_show_pagerefs = True\n\n# The output format for Graphviz when building HTML files.\ngraphviz_output_format = 'svg'\n\n# Contains mapping the locations and names of other projects that\n# should be linked to in this documentation.\nintersphinx_mapping = {\n 'python3': ('https://docs.python.org/3/', None),\n 'numpy': ('https://docs.scipy.org/doc/numpy', None),\n 'scipy': ('https://docs.scipy.org/doc/scipy/reference', None),\n}\n\n# Dictionary of external sites, mapping unique short alias names to a\n# base URL and a prefix.\nextlinks = {\n 'issue': ('https://github.com/diofant/diofant/issues/%s', '#'),\n 'pull': ('https://github.com/diofant/diofant/pull/%s', '#'),\n 'commit': ('https://github.com/diofant/diofant/commit/%s', ''),\n 'sympyissue': ('https://github.com/sympy/sympy/issues/%s', 'sympy/sympy#'),\n 'sympypull': ('https://github.com/sympy/sympy/pull/%s', 'sympy/sympy#'),\n}\n\n# The number of times the linkcheck builder will attempt to check a URL\n# before declaring it broken.\nlinkcheck_retries = 3\n\n# A list of regular expressions that match URIs that should not be checked.\nlinkcheck_ignore = [r'https://primes.utm.edu/notes/gaps.html',\n r'https://primes.utm.edu/glossary/xpage/BertrandsPostulate.html',\n r'https://primes.utm.edu/prove/prove2_3.html',\n r'https://primes.utm.edu/glossary/xpage/Pseudoprime.html']\n\n# This value controls if docstring for classes or methods, if not explicitly\n# set, is inherited form parents.\nautodoc_inherit_docstrings = False\n\n# A list of paths that contain custom static files. Relative paths are taken as\n# relative to the configuration directory. They are copied to the output\u2019s\n# _static directory.\nhtml_static_path = ['_static']\n\n# Should we show \"Created using Sphinx\" in the HTML footer?\nhtml_show_sphinx = False\n\n# Theme-specific options.\nhtml_theme_options = {\n 'logo_only': True,\n 'display_version': False,\n}\n\nmathjax_config = {\n 'CommonHTML': {'linebreaks': {'automatic': True}},\n 'HTML-CSS': {'linebreaks': {'automatic': True}},\n 'SVG': {'linebreaks': {'automatic': True}},\n}\n\n\n# https://docs.readthedocs.io/en/latest/guides/adding-custom-css.html\ndef setup(app):\n app.add_stylesheet('custom.css')\n", "path": "docs/conf.py"}]}
| 2,201 | 123 |
gh_patches_debug_16223
|
rasdani/github-patches
|
git_diff
|
microsoft__botbuilder-python-1930
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Bump msrest to the 0.6.19 or higher
Is your feature request related to a problem? Please describe.
Old version of msrest is used in botframework components -> https://github.com/microsoft/botbuilder-python/search?q=msrest%3D%3D0.6.10 . This blocks us to use latest versions of the service bus client or event using the new language studio python libraries.
With msrest=0.6.10, we're blocked to using 0.50 service bus package and other packages like event grid.
Describe the solution you'd like
EDITED: Upgrade msrest to the at least 0.6.19 or higher.
Describe alternatives you've considered
No alternatives.
</issue>
<code>
[start of libraries/botframework-connector/setup.py]
1 # Copyright (c) Microsoft Corporation. All rights reserved.
2 # Licensed under the MIT License.
3
4 import os
5 from setuptools import setup
6
7 NAME = "botframework-connector"
8 VERSION = os.environ["packageVersion"] if "packageVersion" in os.environ else "4.15.0"
9 REQUIRES = [
10 "msrest==0.6.10",
11 "requests>=2.23.0,<2.26",
12 "PyJWT>=1.5.3,<2.0.0",
13 "botbuilder-schema==4.15.0",
14 "msal==1.6.0",
15 ]
16
17 root = os.path.abspath(os.path.dirname(__file__))
18
19 with open(os.path.join(root, "README.rst"), encoding="utf-8") as f:
20 long_description = f.read()
21
22 setup(
23 name=NAME,
24 version=VERSION,
25 description="Microsoft Bot Framework Bot Builder SDK for Python.",
26 author="Microsoft",
27 url="https://www.github.com/Microsoft/botbuilder-python",
28 keywords=["BotFrameworkConnector", "bots", "ai", "botframework", "botbuilder"],
29 install_requires=REQUIRES,
30 packages=[
31 "botframework.connector",
32 "botframework.connector.auth",
33 "botframework.connector.async_mixin",
34 "botframework.connector.operations",
35 "botframework.connector.models",
36 "botframework.connector.aio",
37 "botframework.connector.aio.operations_async",
38 "botframework.connector.skills",
39 "botframework.connector.teams",
40 "botframework.connector.teams.operations",
41 "botframework.connector.token_api",
42 "botframework.connector.token_api.aio",
43 "botframework.connector.token_api.aio.operations_async",
44 "botframework.connector.token_api.models",
45 "botframework.connector.token_api.operations",
46 ],
47 include_package_data=True,
48 long_description=long_description,
49 long_description_content_type="text/x-rst",
50 license="MIT",
51 classifiers=[
52 "Programming Language :: Python :: 3.7",
53 "Intended Audience :: Developers",
54 "License :: OSI Approved :: MIT License",
55 "Operating System :: OS Independent",
56 "Development Status :: 5 - Production/Stable",
57 "Topic :: Scientific/Engineering :: Artificial Intelligence",
58 ],
59 )
60
[end of libraries/botframework-connector/setup.py]
[start of libraries/botbuilder-schema/setup.py]
1 # Copyright (c) Microsoft Corporation. All rights reserved.
2 # Licensed under the MIT License.
3
4 import os
5 from setuptools import setup
6
7 NAME = "botbuilder-schema"
8 VERSION = os.environ["packageVersion"] if "packageVersion" in os.environ else "4.15.0"
9 REQUIRES = ["msrest==0.6.10"]
10
11 root = os.path.abspath(os.path.dirname(__file__))
12
13 with open(os.path.join(root, "README.rst"), encoding="utf-8") as f:
14 long_description = f.read()
15
16 setup(
17 name=NAME,
18 version=VERSION,
19 description="BotBuilder Schema",
20 author="Microsoft",
21 url="https://github.com/Microsoft/botbuilder-python",
22 keywords=["BotBuilderSchema", "bots", "ai", "botframework", "botbuilder"],
23 long_description=long_description,
24 long_description_content_type="text/x-rst",
25 license="MIT",
26 install_requires=REQUIRES,
27 packages=["botbuilder.schema", "botbuilder.schema.teams",],
28 include_package_data=True,
29 classifiers=[
30 "Programming Language :: Python :: 3.7",
31 "Intended Audience :: Developers",
32 "License :: OSI Approved :: MIT License",
33 "Operating System :: OS Independent",
34 "Development Status :: 5 - Production/Stable",
35 "Topic :: Scientific/Engineering :: Artificial Intelligence",
36 ],
37 )
38
[end of libraries/botbuilder-schema/setup.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/libraries/botbuilder-schema/setup.py b/libraries/botbuilder-schema/setup.py
--- a/libraries/botbuilder-schema/setup.py
+++ b/libraries/botbuilder-schema/setup.py
@@ -6,7 +6,7 @@
NAME = "botbuilder-schema"
VERSION = os.environ["packageVersion"] if "packageVersion" in os.environ else "4.15.0"
-REQUIRES = ["msrest==0.6.10"]
+REQUIRES = ["msrest==0.6.19"]
root = os.path.abspath(os.path.dirname(__file__))
diff --git a/libraries/botframework-connector/setup.py b/libraries/botframework-connector/setup.py
--- a/libraries/botframework-connector/setup.py
+++ b/libraries/botframework-connector/setup.py
@@ -7,7 +7,7 @@
NAME = "botframework-connector"
VERSION = os.environ["packageVersion"] if "packageVersion" in os.environ else "4.15.0"
REQUIRES = [
- "msrest==0.6.10",
+ "msrest==0.6.19",
"requests>=2.23.0,<2.26",
"PyJWT>=1.5.3,<2.0.0",
"botbuilder-schema==4.15.0",
|
{"golden_diff": "diff --git a/libraries/botbuilder-schema/setup.py b/libraries/botbuilder-schema/setup.py\n--- a/libraries/botbuilder-schema/setup.py\n+++ b/libraries/botbuilder-schema/setup.py\n@@ -6,7 +6,7 @@\n \r\n NAME = \"botbuilder-schema\"\r\n VERSION = os.environ[\"packageVersion\"] if \"packageVersion\" in os.environ else \"4.15.0\"\r\n-REQUIRES = [\"msrest==0.6.10\"]\r\n+REQUIRES = [\"msrest==0.6.19\"]\r\n \r\n root = os.path.abspath(os.path.dirname(__file__))\r\n \r\ndiff --git a/libraries/botframework-connector/setup.py b/libraries/botframework-connector/setup.py\n--- a/libraries/botframework-connector/setup.py\n+++ b/libraries/botframework-connector/setup.py\n@@ -7,7 +7,7 @@\n NAME = \"botframework-connector\"\n VERSION = os.environ[\"packageVersion\"] if \"packageVersion\" in os.environ else \"4.15.0\"\n REQUIRES = [\n- \"msrest==0.6.10\",\n+ \"msrest==0.6.19\",\n \"requests>=2.23.0,<2.26\",\n \"PyJWT>=1.5.3,<2.0.0\",\n \"botbuilder-schema==4.15.0\",\n", "issue": "Bump msrest to the 0.6.19 or higher\nIs your feature request related to a problem? Please describe.\r\nOld version of msrest is used in botframework components -> https://github.com/microsoft/botbuilder-python/search?q=msrest%3D%3D0.6.10 . This blocks us to use latest versions of the service bus client or event using the new language studio python libraries.\r\n\r\nWith msrest=0.6.10, we're blocked to using 0.50 service bus package and other packages like event grid.\r\n\r\nDescribe the solution you'd like\r\nEDITED: Upgrade msrest to the at least 0.6.19 or higher.\r\n\r\nDescribe alternatives you've considered\r\nNo alternatives.\r\n\n", "before_files": [{"content": "# Copyright (c) Microsoft Corporation. All rights reserved.\n# Licensed under the MIT License.\n\nimport os\nfrom setuptools import setup\n\nNAME = \"botframework-connector\"\nVERSION = os.environ[\"packageVersion\"] if \"packageVersion\" in os.environ else \"4.15.0\"\nREQUIRES = [\n \"msrest==0.6.10\",\n \"requests>=2.23.0,<2.26\",\n \"PyJWT>=1.5.3,<2.0.0\",\n \"botbuilder-schema==4.15.0\",\n \"msal==1.6.0\",\n]\n\nroot = os.path.abspath(os.path.dirname(__file__))\n\nwith open(os.path.join(root, \"README.rst\"), encoding=\"utf-8\") as f:\n long_description = f.read()\n\nsetup(\n name=NAME,\n version=VERSION,\n description=\"Microsoft Bot Framework Bot Builder SDK for Python.\",\n author=\"Microsoft\",\n url=\"https://www.github.com/Microsoft/botbuilder-python\",\n keywords=[\"BotFrameworkConnector\", \"bots\", \"ai\", \"botframework\", \"botbuilder\"],\n install_requires=REQUIRES,\n packages=[\n \"botframework.connector\",\n \"botframework.connector.auth\",\n \"botframework.connector.async_mixin\",\n \"botframework.connector.operations\",\n \"botframework.connector.models\",\n \"botframework.connector.aio\",\n \"botframework.connector.aio.operations_async\",\n \"botframework.connector.skills\",\n \"botframework.connector.teams\",\n \"botframework.connector.teams.operations\",\n \"botframework.connector.token_api\",\n \"botframework.connector.token_api.aio\",\n \"botframework.connector.token_api.aio.operations_async\",\n \"botframework.connector.token_api.models\",\n \"botframework.connector.token_api.operations\",\n ],\n include_package_data=True,\n long_description=long_description,\n long_description_content_type=\"text/x-rst\",\n license=\"MIT\",\n classifiers=[\n \"Programming Language :: Python :: 3.7\",\n \"Intended Audience :: Developers\",\n \"License :: OSI Approved :: MIT License\",\n \"Operating System :: OS Independent\",\n \"Development Status :: 5 - Production/Stable\",\n \"Topic :: Scientific/Engineering :: Artificial Intelligence\",\n ],\n)\n", "path": "libraries/botframework-connector/setup.py"}, {"content": "# Copyright (c) Microsoft Corporation. All rights reserved.\r\n# Licensed under the MIT License.\r\n\r\nimport os\r\nfrom setuptools import setup\r\n\r\nNAME = \"botbuilder-schema\"\r\nVERSION = os.environ[\"packageVersion\"] if \"packageVersion\" in os.environ else \"4.15.0\"\r\nREQUIRES = [\"msrest==0.6.10\"]\r\n\r\nroot = os.path.abspath(os.path.dirname(__file__))\r\n\r\nwith open(os.path.join(root, \"README.rst\"), encoding=\"utf-8\") as f:\r\n long_description = f.read()\r\n\r\nsetup(\r\n name=NAME,\r\n version=VERSION,\r\n description=\"BotBuilder Schema\",\r\n author=\"Microsoft\",\r\n url=\"https://github.com/Microsoft/botbuilder-python\",\r\n keywords=[\"BotBuilderSchema\", \"bots\", \"ai\", \"botframework\", \"botbuilder\"],\r\n long_description=long_description,\r\n long_description_content_type=\"text/x-rst\",\r\n license=\"MIT\",\r\n install_requires=REQUIRES,\r\n packages=[\"botbuilder.schema\", \"botbuilder.schema.teams\",],\r\n include_package_data=True,\r\n classifiers=[\r\n \"Programming Language :: Python :: 3.7\",\r\n \"Intended Audience :: Developers\",\r\n \"License :: OSI Approved :: MIT License\",\r\n \"Operating System :: OS Independent\",\r\n \"Development Status :: 5 - Production/Stable\",\r\n \"Topic :: Scientific/Engineering :: Artificial Intelligence\",\r\n ],\r\n)\r\n", "path": "libraries/botbuilder-schema/setup.py"}]}
| 1,672 | 298 |
gh_patches_debug_11520
|
rasdani/github-patches
|
git_diff
|
gratipay__gratipay.com-2999
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Font problem in production
> Font from origin 'https://assets.gratipay.com' has been blocked from loading by Cross-Origin Resource Sharing policy: No 'Access-Control-Allow-Origin' header is present on the requested resource. Origin 'https://gratipay.com' is therefore not allowed access.
</issue>
<code>
[start of gratipay/utils/cache_static.py]
1 """
2 Handles caching of static resources.
3 """
4 from base64 import b64encode
5 from hashlib import md5
6
7 from aspen import Response
8
9
10 ETAGS = {}
11
12
13 def asset_etag(path):
14 if path.endswith('.spt'):
15 return ''
16 if path in ETAGS:
17 h = ETAGS[path]
18 else:
19 with open(path) as f:
20 h = ETAGS[path] = b64encode(md5(f.read()).digest(), '-_').replace('=', '~')
21 return h
22
23
24 # algorithm functions
25
26 def get_etag_for_file(dispatch_result):
27 return {'etag': asset_etag(dispatch_result.match)}
28
29
30 def try_to_serve_304(website, dispatch_result, request, etag):
31 """Try to serve a 304 for static resources.
32 """
33 if not etag:
34 # This is a request for a dynamic resource.
35 return
36
37 qs_etag = request.line.uri.querystring.get('etag')
38 if qs_etag and qs_etag != etag:
39 # Don't serve one version of a file as if it were another.
40 raise Response(410)
41
42 headers_etag = request.headers.get('If-None-Match')
43 if not headers_etag:
44 # This client doesn't want a 304.
45 return
46
47 if headers_etag != etag:
48 # Cache miss, the client sent an old or invalid etag.
49 return
50
51 # Huzzah!
52 # =======
53 # We can serve a 304! :D
54
55 raise Response(304)
56
57
58 def add_caching_to_response(website, response, request=None, etag=None):
59 """Set caching headers for static resources.
60 """
61 if etag is None:
62 return
63 assert request is not None # sanity check
64
65 if response.code not in (200, 304):
66 return
67
68 # https://developers.google.com/speed/docs/best-practices/caching
69 response.headers['Vary'] = 'accept-encoding'
70 response.headers['Etag'] = etag
71
72 if request.line.uri.querystring.get('etag'):
73 # We can cache "indefinitely" when the querystring contains the etag.
74 response.headers['Cache-Control'] = 'public, max-age=31536000'
75 else:
76 # Otherwise we cache for 5 seconds
77 response.headers['Cache-Control'] = 'public, max-age=5'
78
[end of gratipay/utils/cache_static.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/gratipay/utils/cache_static.py b/gratipay/utils/cache_static.py
--- a/gratipay/utils/cache_static.py
+++ b/gratipay/utils/cache_static.py
@@ -68,6 +68,9 @@
# https://developers.google.com/speed/docs/best-practices/caching
response.headers['Vary'] = 'accept-encoding'
response.headers['Etag'] = etag
+ # Set CORS header for https://assets.gratipay.com (see issue #2970)
+ if 'Access-Control-Allow-Origin' not in response.headers:
+ response.headers['Access-Control-Allow-Origin'] = 'https://gratipay.com'
if request.line.uri.querystring.get('etag'):
# We can cache "indefinitely" when the querystring contains the etag.
|
{"golden_diff": "diff --git a/gratipay/utils/cache_static.py b/gratipay/utils/cache_static.py\n--- a/gratipay/utils/cache_static.py\n+++ b/gratipay/utils/cache_static.py\n@@ -68,6 +68,9 @@\n # https://developers.google.com/speed/docs/best-practices/caching\n response.headers['Vary'] = 'accept-encoding'\n response.headers['Etag'] = etag\n+ # Set CORS header for https://assets.gratipay.com (see issue #2970)\n+ if 'Access-Control-Allow-Origin' not in response.headers:\n+ response.headers['Access-Control-Allow-Origin'] = 'https://gratipay.com'\n \n if request.line.uri.querystring.get('etag'):\n # We can cache \"indefinitely\" when the querystring contains the etag.\n", "issue": "Font problem in production\n> Font from origin 'https://assets.gratipay.com' has been blocked from loading by Cross-Origin Resource Sharing policy: No 'Access-Control-Allow-Origin' header is present on the requested resource. Origin 'https://gratipay.com' is therefore not allowed access. \n\n", "before_files": [{"content": "\"\"\"\nHandles caching of static resources.\n\"\"\"\nfrom base64 import b64encode\nfrom hashlib import md5\n\nfrom aspen import Response\n\n\nETAGS = {}\n\n\ndef asset_etag(path):\n if path.endswith('.spt'):\n return ''\n if path in ETAGS:\n h = ETAGS[path]\n else:\n with open(path) as f:\n h = ETAGS[path] = b64encode(md5(f.read()).digest(), '-_').replace('=', '~')\n return h\n\n\n# algorithm functions\n\ndef get_etag_for_file(dispatch_result):\n return {'etag': asset_etag(dispatch_result.match)}\n\n\ndef try_to_serve_304(website, dispatch_result, request, etag):\n \"\"\"Try to serve a 304 for static resources.\n \"\"\"\n if not etag:\n # This is a request for a dynamic resource.\n return\n\n qs_etag = request.line.uri.querystring.get('etag')\n if qs_etag and qs_etag != etag:\n # Don't serve one version of a file as if it were another.\n raise Response(410)\n\n headers_etag = request.headers.get('If-None-Match')\n if not headers_etag:\n # This client doesn't want a 304.\n return\n\n if headers_etag != etag:\n # Cache miss, the client sent an old or invalid etag.\n return\n\n # Huzzah!\n # =======\n # We can serve a 304! :D\n\n raise Response(304)\n\n\ndef add_caching_to_response(website, response, request=None, etag=None):\n \"\"\"Set caching headers for static resources.\n \"\"\"\n if etag is None:\n return\n assert request is not None # sanity check\n\n if response.code not in (200, 304):\n return\n\n # https://developers.google.com/speed/docs/best-practices/caching\n response.headers['Vary'] = 'accept-encoding'\n response.headers['Etag'] = etag\n\n if request.line.uri.querystring.get('etag'):\n # We can cache \"indefinitely\" when the querystring contains the etag.\n response.headers['Cache-Control'] = 'public, max-age=31536000'\n else:\n # Otherwise we cache for 5 seconds\n response.headers['Cache-Control'] = 'public, max-age=5'\n", "path": "gratipay/utils/cache_static.py"}]}
| 1,297 | 180 |
gh_patches_debug_2632
|
rasdani/github-patches
|
git_diff
|
hpcaitech__ColossalAI-5433
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
[tensor] fix some unittests
[tensor] fix some unittests
[tensor] fix some unittests
</issue>
<code>
[start of extensions/inference/inference_ops_cuda.py]
1 from ..cuda_extension import _CudaExtension
2 from ..utils import get_cuda_cc_flag
3
4
5 class InferenceOpsCudaExtension(_CudaExtension):
6 def __init__(self):
7 super().__init__(name="inference_ops_cuda")
8
9 def sources_files(self):
10 ret = [
11 self.csrc_abs_path(fname)
12 for fname in [
13 "cuda/colossal_inference_C_frontend.cpp",
14 "cuda/decode_kv_cache_memcpy_kernel.cu",
15 ]
16 ]
17 return ret
18
19 def include_dirs(self):
20 ret = [self.get_cuda_home_include()]
21 return ret
22
23 def cxx_flags(self):
24 version_dependent_macros = ["-DVERSION_GE_1_1", "-DVERSION_GE_1_3", "-DVERSION_GE_1_5"]
25 return ["-O3"] + version_dependent_macros
26
27 def nvcc_flags(self):
28 extra_cuda_flags = ["-lineinfo"]
29 extra_cuda_flags.extend(get_cuda_cc_flag())
30 return ["-O3", "--use_fast_math"] + extra_cuda_flags
31
[end of extensions/inference/inference_ops_cuda.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/extensions/inference/inference_ops_cuda.py b/extensions/inference/inference_ops_cuda.py
--- a/extensions/inference/inference_ops_cuda.py
+++ b/extensions/inference/inference_ops_cuda.py
@@ -12,6 +12,7 @@
for fname in [
"cuda/colossal_inference_C_frontend.cpp",
"cuda/decode_kv_cache_memcpy_kernel.cu",
+ "cuda/activation_kernel.cu",
]
]
return ret
|
{"golden_diff": "diff --git a/extensions/inference/inference_ops_cuda.py b/extensions/inference/inference_ops_cuda.py\n--- a/extensions/inference/inference_ops_cuda.py\n+++ b/extensions/inference/inference_ops_cuda.py\n@@ -12,6 +12,7 @@\n for fname in [\n \"cuda/colossal_inference_C_frontend.cpp\",\n \"cuda/decode_kv_cache_memcpy_kernel.cu\",\n+ \"cuda/activation_kernel.cu\",\n ]\n ]\n return ret\n", "issue": "[tensor] fix some unittests\n\n[tensor] fix some unittests\n\n[tensor] fix some unittests\n\n", "before_files": [{"content": "from ..cuda_extension import _CudaExtension\nfrom ..utils import get_cuda_cc_flag\n\n\nclass InferenceOpsCudaExtension(_CudaExtension):\n def __init__(self):\n super().__init__(name=\"inference_ops_cuda\")\n\n def sources_files(self):\n ret = [\n self.csrc_abs_path(fname)\n for fname in [\n \"cuda/colossal_inference_C_frontend.cpp\",\n \"cuda/decode_kv_cache_memcpy_kernel.cu\",\n ]\n ]\n return ret\n\n def include_dirs(self):\n ret = [self.get_cuda_home_include()]\n return ret\n\n def cxx_flags(self):\n version_dependent_macros = [\"-DVERSION_GE_1_1\", \"-DVERSION_GE_1_3\", \"-DVERSION_GE_1_5\"]\n return [\"-O3\"] + version_dependent_macros\n\n def nvcc_flags(self):\n extra_cuda_flags = [\"-lineinfo\"]\n extra_cuda_flags.extend(get_cuda_cc_flag())\n return [\"-O3\", \"--use_fast_math\"] + extra_cuda_flags\n", "path": "extensions/inference/inference_ops_cuda.py"}]}
| 847 | 103 |
gh_patches_debug_153
|
rasdani/github-patches
|
git_diff
|
bookwyrm-social__bookwyrm-1018
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Ratings don't federate
**Describe the bug**
I do follow someone on bookwyrm.social from bookwyrm.social and wyrms.de. I have seen on b.s that they rated some books without reviewing them, but those ratings do not appear on w.d. All other posts federate properly (I think).
**Expeceted behaviour**
The rating should show up on connected instances and ideally also be used on those to calculate the average rating of the book.
Here is one example that's not visible from w.d: https://bookwyrm.social/user/tastytea/reviewrating/21469
</issue>
<code>
[start of bookwyrm/activitypub/note.py]
1 """ note serializer and children thereof """
2 from dataclasses import dataclass, field
3 from typing import Dict, List
4 from django.apps import apps
5
6 from .base_activity import ActivityObject, Link
7 from .image import Document
8
9
10 @dataclass(init=False)
11 class Tombstone(ActivityObject):
12 """the placeholder for a deleted status"""
13
14 type: str = "Tombstone"
15
16 def to_model(self, *args, **kwargs): # pylint: disable=unused-argument
17 """this should never really get serialized, just searched for"""
18 model = apps.get_model("bookwyrm.Status")
19 return model.find_existing_by_remote_id(self.id)
20
21
22 @dataclass(init=False)
23 class Note(ActivityObject):
24 """Note activity"""
25
26 published: str
27 attributedTo: str
28 content: str = ""
29 to: List[str] = field(default_factory=lambda: [])
30 cc: List[str] = field(default_factory=lambda: [])
31 replies: Dict = field(default_factory=lambda: {})
32 inReplyTo: str = ""
33 summary: str = ""
34 tag: List[Link] = field(default_factory=lambda: [])
35 attachment: List[Document] = field(default_factory=lambda: [])
36 sensitive: bool = False
37 type: str = "Note"
38
39
40 @dataclass(init=False)
41 class Article(Note):
42 """what's an article except a note with more fields"""
43
44 name: str
45 type: str = "Article"
46
47
48 @dataclass(init=False)
49 class GeneratedNote(Note):
50 """just a re-typed note"""
51
52 type: str = "GeneratedNote"
53
54
55 @dataclass(init=False)
56 class Comment(Note):
57 """like a note but with a book"""
58
59 inReplyToBook: str
60 type: str = "Comment"
61
62
63 @dataclass(init=False)
64 class Quotation(Comment):
65 """a quote and commentary on a book"""
66
67 quote: str
68 type: str = "Quotation"
69
70
71 @dataclass(init=False)
72 class Review(Comment):
73 """a full book review"""
74
75 name: str = None
76 rating: int = None
77 type: str = "Review"
78
79
80 @dataclass(init=False)
81 class Rating(Comment):
82 """just a star rating"""
83
84 rating: int
85 content: str = None
86 type: str = "Rating"
87
[end of bookwyrm/activitypub/note.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/bookwyrm/activitypub/note.py b/bookwyrm/activitypub/note.py
--- a/bookwyrm/activitypub/note.py
+++ b/bookwyrm/activitypub/note.py
@@ -83,4 +83,5 @@
rating: int
content: str = None
+ name: str = None # not used, but the model inherits from Review
type: str = "Rating"
|
{"golden_diff": "diff --git a/bookwyrm/activitypub/note.py b/bookwyrm/activitypub/note.py\n--- a/bookwyrm/activitypub/note.py\n+++ b/bookwyrm/activitypub/note.py\n@@ -83,4 +83,5 @@\n \n rating: int\n content: str = None\n+ name: str = None # not used, but the model inherits from Review\n type: str = \"Rating\"\n", "issue": "Ratings don't federate\n**Describe the bug**\r\nI do follow someone on bookwyrm.social from bookwyrm.social and wyrms.de. I have seen on b.s that they rated some books without reviewing them, but those ratings do not appear on w.d. All other posts federate properly (I think).\r\n\r\n**Expeceted behaviour**\r\nThe rating should show up on connected instances and ideally also be used on those to calculate the average rating of the book.\r\n\r\nHere is one example that's not visible from w.d: https://bookwyrm.social/user/tastytea/reviewrating/21469\n", "before_files": [{"content": "\"\"\" note serializer and children thereof \"\"\"\nfrom dataclasses import dataclass, field\nfrom typing import Dict, List\nfrom django.apps import apps\n\nfrom .base_activity import ActivityObject, Link\nfrom .image import Document\n\n\n@dataclass(init=False)\nclass Tombstone(ActivityObject):\n \"\"\"the placeholder for a deleted status\"\"\"\n\n type: str = \"Tombstone\"\n\n def to_model(self, *args, **kwargs): # pylint: disable=unused-argument\n \"\"\"this should never really get serialized, just searched for\"\"\"\n model = apps.get_model(\"bookwyrm.Status\")\n return model.find_existing_by_remote_id(self.id)\n\n\n@dataclass(init=False)\nclass Note(ActivityObject):\n \"\"\"Note activity\"\"\"\n\n published: str\n attributedTo: str\n content: str = \"\"\n to: List[str] = field(default_factory=lambda: [])\n cc: List[str] = field(default_factory=lambda: [])\n replies: Dict = field(default_factory=lambda: {})\n inReplyTo: str = \"\"\n summary: str = \"\"\n tag: List[Link] = field(default_factory=lambda: [])\n attachment: List[Document] = field(default_factory=lambda: [])\n sensitive: bool = False\n type: str = \"Note\"\n\n\n@dataclass(init=False)\nclass Article(Note):\n \"\"\"what's an article except a note with more fields\"\"\"\n\n name: str\n type: str = \"Article\"\n\n\n@dataclass(init=False)\nclass GeneratedNote(Note):\n \"\"\"just a re-typed note\"\"\"\n\n type: str = \"GeneratedNote\"\n\n\n@dataclass(init=False)\nclass Comment(Note):\n \"\"\"like a note but with a book\"\"\"\n\n inReplyToBook: str\n type: str = \"Comment\"\n\n\n@dataclass(init=False)\nclass Quotation(Comment):\n \"\"\"a quote and commentary on a book\"\"\"\n\n quote: str\n type: str = \"Quotation\"\n\n\n@dataclass(init=False)\nclass Review(Comment):\n \"\"\"a full book review\"\"\"\n\n name: str = None\n rating: int = None\n type: str = \"Review\"\n\n\n@dataclass(init=False)\nclass Rating(Comment):\n \"\"\"just a star rating\"\"\"\n\n rating: int\n content: str = None\n type: str = \"Rating\"\n", "path": "bookwyrm/activitypub/note.py"}]}
| 1,338 | 96 |
gh_patches_debug_591
|
rasdani/github-patches
|
git_diff
|
pex-tool__pex-1140
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Release 2.1.23
On the docket:
+ [x] Upgrade Pex to Pip 20.3.1. #1133
</issue>
<code>
[start of pex/version.py]
1 # Copyright 2015 Pants project contributors (see CONTRIBUTORS.md).
2 # Licensed under the Apache License, Version 2.0 (see LICENSE).
3
4 __version__ = "2.1.22"
5
[end of pex/version.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/pex/version.py b/pex/version.py
--- a/pex/version.py
+++ b/pex/version.py
@@ -1,4 +1,4 @@
# Copyright 2015 Pants project contributors (see CONTRIBUTORS.md).
# Licensed under the Apache License, Version 2.0 (see LICENSE).
-__version__ = "2.1.22"
+__version__ = "2.1.23"
|
{"golden_diff": "diff --git a/pex/version.py b/pex/version.py\n--- a/pex/version.py\n+++ b/pex/version.py\n@@ -1,4 +1,4 @@\n # Copyright 2015 Pants project contributors (see CONTRIBUTORS.md).\n # Licensed under the Apache License, Version 2.0 (see LICENSE).\n \n-__version__ = \"2.1.22\"\n+__version__ = \"2.1.23\"\n", "issue": "Release 2.1.23\nOn the docket:\r\n+ [x] Upgrade Pex to Pip 20.3.1. #1133\r\n\n", "before_files": [{"content": "# Copyright 2015 Pants project contributors (see CONTRIBUTORS.md).\n# Licensed under the Apache License, Version 2.0 (see LICENSE).\n\n__version__ = \"2.1.22\"\n", "path": "pex/version.py"}]}
| 620 | 96 |
gh_patches_debug_13712
|
rasdani/github-patches
|
git_diff
|
chainer__chainer-1312
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
`split_axis` doesn't support empty sections
This code causes a TypeError.
`functions.split_axis(x, [], 0)`
</issue>
<code>
[start of chainer/functions/array/split_axis.py]
1 import collections
2
3 import six
4
5 import chainer
6 from chainer import cuda
7 from chainer import function
8 from chainer.utils import type_check
9
10
11 class SplitAxis(function.Function):
12
13 """Function that splits multiple arrays along the specified axis."""
14
15 def __init__(self, indices_or_sections, axis):
16 if not isinstance(indices_or_sections, (int, collections.Iterable)):
17 raise TypeError('indices_or_sections must be integer or 1-D array')
18 self.indices_or_sections = indices_or_sections
19 self.axis = axis
20
21 def check_type_forward(self, in_types):
22 type_check.expect(in_types.size() == 1)
23 type_check.expect(in_types[0].ndim > self.axis)
24
25 if isinstance(self.indices_or_sections, collections.Iterable):
26 max_index = type_check.Variable(
27 self.indices_or_sections[-1], 'max_index')
28 type_check.expect(in_types[0].shape[self.axis] > max_index)
29 else:
30 sections = type_check.Variable(
31 self.indices_or_sections, 'sections')
32 type_check.expect(in_types[0].shape[self.axis] % sections == 0)
33
34 def forward(self, x):
35 if isinstance(self.indices_or_sections, collections.Iterable):
36 cdimx = x[0].shape[self.axis]
37 ind = list(self.indices_or_sections)
38 ind.append(cdimx)
39 prev_i = 0
40 for i in ind:
41 cdimy = max(0, min(i, cdimx) - prev_i)
42 if cdimy == 0:
43 raise ValueError('Not support if shape contains 0')
44 prev_i = i
45 xp = cuda.get_array_module(*x)
46 return tuple(xp.split(x[0], self.indices_or_sections, self.axis))
47
48 def backward(self, x, gys):
49 xp = cuda.get_array_module(*x)
50 if any(gy is None for gy in gys):
51 gx = xp.zeros_like(x[0])
52 gxs = xp.split(gx, self.indices_or_sections, self.axis)
53 for gxi, gy in six.moves.zip(gxs, gys):
54 if gy is None:
55 continue
56 gxi[:] = gy
57 return gx,
58 else:
59 return xp.concatenate(gys, axis=self.axis),
60
61
62 def split_axis(x, indices_or_sections, axis, force_tuple=False):
63 """Splits given variables along an axis.
64
65 Args:
66 x (tuple of Variables): Variables to be split.
67 indices_or_sections (int or 1-D array): If this argument is an integer,
68 N, the array will be divided into N equal arrays along axis.
69 If it is a 1-D array of sorted integers, it
70 indicates the positions where the array is split.
71 axis (int): Axis that the input array is split along.
72 force_tuple (bool): If ``True``, this method returns a tuple even when
73 the number of outputs is one.
74
75 Returns:
76 tuple or Variable: Tuple of :class:`~chainer.Variable` objects
77 if the number of outputs is more than 1 or
78 :class:`~chainer.Variable` otherwise.
79 When ``force_tuple`` is ``True``, returned value is always a tuple
80 regardless of the number of outputs.
81
82 .. note::
83 This function raises :class:`ValueError` if at least
84 one of the outputs is split to zero-size
85 (i.e. ``axis``-th value of its shape is zero).
86
87 """
88 res = SplitAxis(indices_or_sections, axis)(x)
89 if force_tuple and isinstance(res, chainer.Variable):
90 res = (res,)
91 return res
92
[end of chainer/functions/array/split_axis.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/chainer/functions/array/split_axis.py b/chainer/functions/array/split_axis.py
--- a/chainer/functions/array/split_axis.py
+++ b/chainer/functions/array/split_axis.py
@@ -23,9 +23,10 @@
type_check.expect(in_types[0].ndim > self.axis)
if isinstance(self.indices_or_sections, collections.Iterable):
- max_index = type_check.Variable(
- self.indices_or_sections[-1], 'max_index')
- type_check.expect(in_types[0].shape[self.axis] > max_index)
+ if len(self.indices_or_sections) > 0:
+ max_index = type_check.Variable(
+ self.indices_or_sections[-1], 'max_index')
+ type_check.expect(in_types[0].shape[self.axis] > max_index)
else:
sections = type_check.Variable(
self.indices_or_sections, 'sections')
|
{"golden_diff": "diff --git a/chainer/functions/array/split_axis.py b/chainer/functions/array/split_axis.py\n--- a/chainer/functions/array/split_axis.py\n+++ b/chainer/functions/array/split_axis.py\n@@ -23,9 +23,10 @@\n type_check.expect(in_types[0].ndim > self.axis)\n \n if isinstance(self.indices_or_sections, collections.Iterable):\n- max_index = type_check.Variable(\n- self.indices_or_sections[-1], 'max_index')\n- type_check.expect(in_types[0].shape[self.axis] > max_index)\n+ if len(self.indices_or_sections) > 0:\n+ max_index = type_check.Variable(\n+ self.indices_or_sections[-1], 'max_index')\n+ type_check.expect(in_types[0].shape[self.axis] > max_index)\n else:\n sections = type_check.Variable(\n self.indices_or_sections, 'sections')\n", "issue": "`split_axis` doesn't support empty sections\nThis code causes a TypeError.\n`functions.split_axis(x, [], 0)`\n\n", "before_files": [{"content": "import collections\n\nimport six\n\nimport chainer\nfrom chainer import cuda\nfrom chainer import function\nfrom chainer.utils import type_check\n\n\nclass SplitAxis(function.Function):\n\n \"\"\"Function that splits multiple arrays along the specified axis.\"\"\"\n\n def __init__(self, indices_or_sections, axis):\n if not isinstance(indices_or_sections, (int, collections.Iterable)):\n raise TypeError('indices_or_sections must be integer or 1-D array')\n self.indices_or_sections = indices_or_sections\n self.axis = axis\n\n def check_type_forward(self, in_types):\n type_check.expect(in_types.size() == 1)\n type_check.expect(in_types[0].ndim > self.axis)\n\n if isinstance(self.indices_or_sections, collections.Iterable):\n max_index = type_check.Variable(\n self.indices_or_sections[-1], 'max_index')\n type_check.expect(in_types[0].shape[self.axis] > max_index)\n else:\n sections = type_check.Variable(\n self.indices_or_sections, 'sections')\n type_check.expect(in_types[0].shape[self.axis] % sections == 0)\n\n def forward(self, x):\n if isinstance(self.indices_or_sections, collections.Iterable):\n cdimx = x[0].shape[self.axis]\n ind = list(self.indices_or_sections)\n ind.append(cdimx)\n prev_i = 0\n for i in ind:\n cdimy = max(0, min(i, cdimx) - prev_i)\n if cdimy == 0:\n raise ValueError('Not support if shape contains 0')\n prev_i = i\n xp = cuda.get_array_module(*x)\n return tuple(xp.split(x[0], self.indices_or_sections, self.axis))\n\n def backward(self, x, gys):\n xp = cuda.get_array_module(*x)\n if any(gy is None for gy in gys):\n gx = xp.zeros_like(x[0])\n gxs = xp.split(gx, self.indices_or_sections, self.axis)\n for gxi, gy in six.moves.zip(gxs, gys):\n if gy is None:\n continue\n gxi[:] = gy\n return gx,\n else:\n return xp.concatenate(gys, axis=self.axis),\n\n\ndef split_axis(x, indices_or_sections, axis, force_tuple=False):\n \"\"\"Splits given variables along an axis.\n\n Args:\n x (tuple of Variables): Variables to be split.\n indices_or_sections (int or 1-D array): If this argument is an integer,\n N, the array will be divided into N equal arrays along axis.\n If it is a 1-D array of sorted integers, it\n indicates the positions where the array is split.\n axis (int): Axis that the input array is split along.\n force_tuple (bool): If ``True``, this method returns a tuple even when\n the number of outputs is one.\n\n Returns:\n tuple or Variable: Tuple of :class:`~chainer.Variable` objects\n if the number of outputs is more than 1 or\n :class:`~chainer.Variable` otherwise.\n When ``force_tuple`` is ``True``, returned value is always a tuple\n regardless of the number of outputs.\n\n .. note::\n This function raises :class:`ValueError` if at least\n one of the outputs is split to zero-size\n (i.e. ``axis``-th value of its shape is zero).\n\n \"\"\"\n res = SplitAxis(indices_or_sections, axis)(x)\n if force_tuple and isinstance(res, chainer.Variable):\n res = (res,)\n return res\n", "path": "chainer/functions/array/split_axis.py"}]}
| 1,526 | 197 |
gh_patches_debug_26510
|
rasdani/github-patches
|
git_diff
|
mitmproxy__mitmproxy-2325
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
HEAD method not allowed to access mitm.it
##### Steps to reproduce the problem:
1. Using mitmdump in transparent mode, with on-boarding port on 6969 (did not test on 80)
/mitmdump -T --host --insecure --onboarding-port 6969
2. On iPad + Chrome + Transparent mode, access to http://mitm.it:6969
3. An HEAD /cert/pem is sent, receive an "405 Method Not Allowed"
Any idea of any workaround?
##### Any other comments? What have you tried so far?
Exact packet:
Hypertext Transfer Protocol
HEAD /cert/pem HTTP/1.1\r\n
Host: mitm.it:6969\r\n
Connection: keep-alive\r\n
User-Agent: Mozilla/5.0 (iPad; CPU OS 9_2 like Mac OS X) AppleWebKit/601.1 (KHTML, like Gecko) CriOS/55.0.2883.79 Mobile/13C75 Safari/601.1.46\r\n
Accept-Encoding: gzip, deflate, sdch\r\n
Accept-Language: fr-FR,fr;q=0.8,en-US;q=0.6,en;q=0.4\r\n
\r\n
Answer:
Hypertext Transfer Protocol
HTTP/1.1 405 Method Not Allowed\r\n
Server: TornadoServer/4.4.2\r\n
Content-Length: 87\r\n
Date: Tue, 09 May 2017 13:35:38 GMT\r\n
Content-Type: text/html; charset=UTF-8\r\n
\r\n
##### System information
mitmdump --version
Mitmproxy version: 2.0.1 (release version) Precompiled Binary
Python version: 3.5.2
Platform: Darwin-16.5.0-x86_64-i386-64bit
SSL version: OpenSSL 1.0.2j 26 Sep 2016
Mac version: 10.12.4 ('', '', '') x86_64
HEAD method not allowed to access mitm.it
##### Steps to reproduce the problem:
1. Using mitmdump in transparent mode, with on-boarding port on 6969 (did not test on 80)
/mitmdump -T --host --insecure --onboarding-port 6969
2. On iPad + Chrome + Transparent mode, access to http://mitm.it:6969
3. An HEAD /cert/pem is sent, receive an "405 Method Not Allowed"
Any idea of any workaround?
##### Any other comments? What have you tried so far?
Exact packet:
Hypertext Transfer Protocol
HEAD /cert/pem HTTP/1.1\r\n
Host: mitm.it:6969\r\n
Connection: keep-alive\r\n
User-Agent: Mozilla/5.0 (iPad; CPU OS 9_2 like Mac OS X) AppleWebKit/601.1 (KHTML, like Gecko) CriOS/55.0.2883.79 Mobile/13C75 Safari/601.1.46\r\n
Accept-Encoding: gzip, deflate, sdch\r\n
Accept-Language: fr-FR,fr;q=0.8,en-US;q=0.6,en;q=0.4\r\n
\r\n
Answer:
Hypertext Transfer Protocol
HTTP/1.1 405 Method Not Allowed\r\n
Server: TornadoServer/4.4.2\r\n
Content-Length: 87\r\n
Date: Tue, 09 May 2017 13:35:38 GMT\r\n
Content-Type: text/html; charset=UTF-8\r\n
\r\n
##### System information
mitmdump --version
Mitmproxy version: 2.0.1 (release version) Precompiled Binary
Python version: 3.5.2
Platform: Darwin-16.5.0-x86_64-i386-64bit
SSL version: OpenSSL 1.0.2j 26 Sep 2016
Mac version: 10.12.4 ('', '', '') x86_64
HEAD method not allowed to access mitm.it
##### Steps to reproduce the problem:
1. Using mitmdump in transparent mode, with on-boarding port on 6969 (did not test on 80)
/mitmdump -T --host --insecure --onboarding-port 6969
2. On iPad + Chrome + Transparent mode, access to http://mitm.it:6969
3. An HEAD /cert/pem is sent, receive an "405 Method Not Allowed"
Any idea of any workaround?
##### Any other comments? What have you tried so far?
Exact packet:
Hypertext Transfer Protocol
HEAD /cert/pem HTTP/1.1\r\n
Host: mitm.it:6969\r\n
Connection: keep-alive\r\n
User-Agent: Mozilla/5.0 (iPad; CPU OS 9_2 like Mac OS X) AppleWebKit/601.1 (KHTML, like Gecko) CriOS/55.0.2883.79 Mobile/13C75 Safari/601.1.46\r\n
Accept-Encoding: gzip, deflate, sdch\r\n
Accept-Language: fr-FR,fr;q=0.8,en-US;q=0.6,en;q=0.4\r\n
\r\n
Answer:
Hypertext Transfer Protocol
HTTP/1.1 405 Method Not Allowed\r\n
Server: TornadoServer/4.4.2\r\n
Content-Length: 87\r\n
Date: Tue, 09 May 2017 13:35:38 GMT\r\n
Content-Type: text/html; charset=UTF-8\r\n
\r\n
##### System information
mitmdump --version
Mitmproxy version: 2.0.1 (release version) Precompiled Binary
Python version: 3.5.2
Platform: Darwin-16.5.0-x86_64-i386-64bit
SSL version: OpenSSL 1.0.2j 26 Sep 2016
Mac version: 10.12.4 ('', '', '') x86_64
</issue>
<code>
[start of mitmproxy/addons/onboardingapp/app.py]
1 import os
2
3 import tornado.template
4 import tornado.web
5 import tornado.wsgi
6
7 from mitmproxy.utils import data
8 from mitmproxy.proxy import config
9
10 loader = tornado.template.Loader(data.pkg_data.path("addons/onboardingapp/templates"))
11
12
13 class Adapter(tornado.wsgi.WSGIAdapter):
14 # Tornado doesn't make the WSGI environment available to pages, so this
15 # hideous monkey patch is the easiest way to get to the mitmproxy.master
16 # variable.
17
18 def __init__(self, application):
19 self._application = application
20
21 def application(self, request):
22 request.master = self.environ["mitmproxy.master"]
23 return self._application(request)
24
25 def __call__(self, environ, start_response):
26 self.environ = environ
27 return tornado.wsgi.WSGIAdapter.__call__(
28 self,
29 environ,
30 start_response
31 )
32
33
34 class Index(tornado.web.RequestHandler):
35
36 def get(self):
37 t = loader.load("index.html")
38 self.write(t.generate())
39
40
41 class PEM(tornado.web.RequestHandler):
42
43 @property
44 def filename(self):
45 return config.CONF_BASENAME + "-ca-cert.pem"
46
47 def get(self):
48 p = os.path.join(self.request.master.options.cadir, self.filename)
49 p = os.path.expanduser(p)
50 self.set_header("Content-Type", "application/x-x509-ca-cert")
51 self.set_header(
52 "Content-Disposition",
53 "inline; filename={}".format(
54 self.filename))
55
56 with open(p, "rb") as f:
57 self.write(f.read())
58
59
60 class P12(tornado.web.RequestHandler):
61
62 @property
63 def filename(self):
64 return config.CONF_BASENAME + "-ca-cert.p12"
65
66 def get(self):
67 p = os.path.join(self.request.master.options.cadir, self.filename)
68 p = os.path.expanduser(p)
69 self.set_header("Content-Type", "application/x-pkcs12")
70 self.set_header(
71 "Content-Disposition",
72 "inline; filename={}".format(
73 self.filename))
74
75 with open(p, "rb") as f:
76 self.write(f.read())
77
78
79 application = tornado.web.Application(
80 [
81 (r"/", Index),
82 (r"/cert/pem", PEM),
83 (r"/cert/p12", P12),
84 (
85 r"/static/(.*)",
86 tornado.web.StaticFileHandler,
87 {
88 "path": data.pkg_data.path("addons/onboardingapp/static")
89 }
90 ),
91 ],
92 # debug=True
93 )
94
[end of mitmproxy/addons/onboardingapp/app.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/mitmproxy/addons/onboardingapp/app.py b/mitmproxy/addons/onboardingapp/app.py
--- a/mitmproxy/addons/onboardingapp/app.py
+++ b/mitmproxy/addons/onboardingapp/app.py
@@ -44,6 +44,18 @@
def filename(self):
return config.CONF_BASENAME + "-ca-cert.pem"
+ def head(self):
+ p = os.path.join(self.request.master.options.cadir, self.filename)
+ p = os.path.expanduser(p)
+ content_length = os.path.getsize(p)
+
+ self.set_header("Content-Type", "application/x-x509-ca-cert")
+ self.set_header(
+ "Content-Disposition",
+ "inline; filename={}".format(
+ self.filename))
+ self.set_header("Content-Length", content_length)
+
def get(self):
p = os.path.join(self.request.master.options.cadir, self.filename)
p = os.path.expanduser(p)
@@ -63,6 +75,19 @@
def filename(self):
return config.CONF_BASENAME + "-ca-cert.p12"
+ def head(self):
+ p = os.path.join(self.request.master.options.cadir, self.filename)
+ p = os.path.expanduser(p)
+ content_length = os.path.getsize(p)
+
+ self.set_header("Content-Type", "application/x-pkcs12")
+ self.set_header(
+ "Content-Disposition",
+ "inline; filename={}".format(
+ self.filename))
+
+ self.set_header("Content-Length", content_length)
+
def get(self):
p = os.path.join(self.request.master.options.cadir, self.filename)
p = os.path.expanduser(p)
|
{"golden_diff": "diff --git a/mitmproxy/addons/onboardingapp/app.py b/mitmproxy/addons/onboardingapp/app.py\n--- a/mitmproxy/addons/onboardingapp/app.py\n+++ b/mitmproxy/addons/onboardingapp/app.py\n@@ -44,6 +44,18 @@\n def filename(self):\n return config.CONF_BASENAME + \"-ca-cert.pem\"\n \n+ def head(self):\n+ p = os.path.join(self.request.master.options.cadir, self.filename)\n+ p = os.path.expanduser(p)\n+ content_length = os.path.getsize(p)\n+\n+ self.set_header(\"Content-Type\", \"application/x-x509-ca-cert\")\n+ self.set_header(\n+ \"Content-Disposition\",\n+ \"inline; filename={}\".format(\n+ self.filename))\n+ self.set_header(\"Content-Length\", content_length)\n+\n def get(self):\n p = os.path.join(self.request.master.options.cadir, self.filename)\n p = os.path.expanduser(p)\n@@ -63,6 +75,19 @@\n def filename(self):\n return config.CONF_BASENAME + \"-ca-cert.p12\"\n \n+ def head(self):\n+ p = os.path.join(self.request.master.options.cadir, self.filename)\n+ p = os.path.expanduser(p)\n+ content_length = os.path.getsize(p)\n+\n+ self.set_header(\"Content-Type\", \"application/x-pkcs12\")\n+ self.set_header(\n+ \"Content-Disposition\",\n+ \"inline; filename={}\".format(\n+ self.filename))\n+\n+ self.set_header(\"Content-Length\", content_length)\n+\n def get(self):\n p = os.path.join(self.request.master.options.cadir, self.filename)\n p = os.path.expanduser(p)\n", "issue": "HEAD method not allowed to access mitm.it\n##### Steps to reproduce the problem:\r\n\r\n1. Using mitmdump in transparent mode, with on-boarding port on 6969 (did not test on 80)\r\n/mitmdump -T --host --insecure --onboarding-port 6969\r\n\r\n2. On iPad + Chrome + Transparent mode, access to http://mitm.it:6969\r\n\r\n3. An HEAD /cert/pem is sent, receive an \"405 Method Not Allowed\"\r\n\r\nAny idea of any workaround?\r\n\r\n\r\n##### Any other comments? What have you tried so far?\r\n\r\nExact packet:\r\nHypertext Transfer Protocol\r\n HEAD /cert/pem HTTP/1.1\\r\\n\r\n Host: mitm.it:6969\\r\\n\r\n Connection: keep-alive\\r\\n\r\n User-Agent: Mozilla/5.0 (iPad; CPU OS 9_2 like Mac OS X) AppleWebKit/601.1 (KHTML, like Gecko) CriOS/55.0.2883.79 Mobile/13C75 Safari/601.1.46\\r\\n\r\n Accept-Encoding: gzip, deflate, sdch\\r\\n\r\n Accept-Language: fr-FR,fr;q=0.8,en-US;q=0.6,en;q=0.4\\r\\n\r\n \\r\\n\r\n\r\n\r\n\r\nAnswer:\r\nHypertext Transfer Protocol\r\n HTTP/1.1 405 Method Not Allowed\\r\\n\r\n Server: TornadoServer/4.4.2\\r\\n\r\n Content-Length: 87\\r\\n\r\n Date: Tue, 09 May 2017 13:35:38 GMT\\r\\n\r\n Content-Type: text/html; charset=UTF-8\\r\\n\r\n \\r\\n\r\n\r\n\r\n\r\n##### System information\r\n\r\nmitmdump --version\r\nMitmproxy version: 2.0.1 (release version) Precompiled Binary\r\nPython version: 3.5.2\r\nPlatform: Darwin-16.5.0-x86_64-i386-64bit\r\nSSL version: OpenSSL 1.0.2j 26 Sep 2016\r\nMac version: 10.12.4 ('', '', '') x86_64\r\n\nHEAD method not allowed to access mitm.it\n##### Steps to reproduce the problem:\r\n\r\n1. Using mitmdump in transparent mode, with on-boarding port on 6969 (did not test on 80)\r\n/mitmdump -T --host --insecure --onboarding-port 6969\r\n\r\n2. On iPad + Chrome + Transparent mode, access to http://mitm.it:6969\r\n\r\n3. An HEAD /cert/pem is sent, receive an \"405 Method Not Allowed\"\r\n\r\nAny idea of any workaround?\r\n\r\n\r\n##### Any other comments? What have you tried so far?\r\n\r\nExact packet:\r\nHypertext Transfer Protocol\r\n HEAD /cert/pem HTTP/1.1\\r\\n\r\n Host: mitm.it:6969\\r\\n\r\n Connection: keep-alive\\r\\n\r\n User-Agent: Mozilla/5.0 (iPad; CPU OS 9_2 like Mac OS X) AppleWebKit/601.1 (KHTML, like Gecko) CriOS/55.0.2883.79 Mobile/13C75 Safari/601.1.46\\r\\n\r\n Accept-Encoding: gzip, deflate, sdch\\r\\n\r\n Accept-Language: fr-FR,fr;q=0.8,en-US;q=0.6,en;q=0.4\\r\\n\r\n \\r\\n\r\n\r\n\r\n\r\nAnswer:\r\nHypertext Transfer Protocol\r\n HTTP/1.1 405 Method Not Allowed\\r\\n\r\n Server: TornadoServer/4.4.2\\r\\n\r\n Content-Length: 87\\r\\n\r\n Date: Tue, 09 May 2017 13:35:38 GMT\\r\\n\r\n Content-Type: text/html; charset=UTF-8\\r\\n\r\n \\r\\n\r\n\r\n\r\n\r\n##### System information\r\n\r\nmitmdump --version\r\nMitmproxy version: 2.0.1 (release version) Precompiled Binary\r\nPython version: 3.5.2\r\nPlatform: Darwin-16.5.0-x86_64-i386-64bit\r\nSSL version: OpenSSL 1.0.2j 26 Sep 2016\r\nMac version: 10.12.4 ('', '', '') x86_64\r\n\nHEAD method not allowed to access mitm.it\n##### Steps to reproduce the problem:\r\n\r\n1. Using mitmdump in transparent mode, with on-boarding port on 6969 (did not test on 80)\r\n/mitmdump -T --host --insecure --onboarding-port 6969\r\n\r\n2. On iPad + Chrome + Transparent mode, access to http://mitm.it:6969\r\n\r\n3. An HEAD /cert/pem is sent, receive an \"405 Method Not Allowed\"\r\n\r\nAny idea of any workaround?\r\n\r\n\r\n##### Any other comments? What have you tried so far?\r\n\r\nExact packet:\r\nHypertext Transfer Protocol\r\n HEAD /cert/pem HTTP/1.1\\r\\n\r\n Host: mitm.it:6969\\r\\n\r\n Connection: keep-alive\\r\\n\r\n User-Agent: Mozilla/5.0 (iPad; CPU OS 9_2 like Mac OS X) AppleWebKit/601.1 (KHTML, like Gecko) CriOS/55.0.2883.79 Mobile/13C75 Safari/601.1.46\\r\\n\r\n Accept-Encoding: gzip, deflate, sdch\\r\\n\r\n Accept-Language: fr-FR,fr;q=0.8,en-US;q=0.6,en;q=0.4\\r\\n\r\n \\r\\n\r\n\r\n\r\n\r\nAnswer:\r\nHypertext Transfer Protocol\r\n HTTP/1.1 405 Method Not Allowed\\r\\n\r\n Server: TornadoServer/4.4.2\\r\\n\r\n Content-Length: 87\\r\\n\r\n Date: Tue, 09 May 2017 13:35:38 GMT\\r\\n\r\n Content-Type: text/html; charset=UTF-8\\r\\n\r\n \\r\\n\r\n\r\n\r\n\r\n##### System information\r\n\r\nmitmdump --version\r\nMitmproxy version: 2.0.1 (release version) Precompiled Binary\r\nPython version: 3.5.2\r\nPlatform: Darwin-16.5.0-x86_64-i386-64bit\r\nSSL version: OpenSSL 1.0.2j 26 Sep 2016\r\nMac version: 10.12.4 ('', '', '') x86_64\r\n\n", "before_files": [{"content": "import os\n\nimport tornado.template\nimport tornado.web\nimport tornado.wsgi\n\nfrom mitmproxy.utils import data\nfrom mitmproxy.proxy import config\n\nloader = tornado.template.Loader(data.pkg_data.path(\"addons/onboardingapp/templates\"))\n\n\nclass Adapter(tornado.wsgi.WSGIAdapter):\n # Tornado doesn't make the WSGI environment available to pages, so this\n # hideous monkey patch is the easiest way to get to the mitmproxy.master\n # variable.\n\n def __init__(self, application):\n self._application = application\n\n def application(self, request):\n request.master = self.environ[\"mitmproxy.master\"]\n return self._application(request)\n\n def __call__(self, environ, start_response):\n self.environ = environ\n return tornado.wsgi.WSGIAdapter.__call__(\n self,\n environ,\n start_response\n )\n\n\nclass Index(tornado.web.RequestHandler):\n\n def get(self):\n t = loader.load(\"index.html\")\n self.write(t.generate())\n\n\nclass PEM(tornado.web.RequestHandler):\n\n @property\n def filename(self):\n return config.CONF_BASENAME + \"-ca-cert.pem\"\n\n def get(self):\n p = os.path.join(self.request.master.options.cadir, self.filename)\n p = os.path.expanduser(p)\n self.set_header(\"Content-Type\", \"application/x-x509-ca-cert\")\n self.set_header(\n \"Content-Disposition\",\n \"inline; filename={}\".format(\n self.filename))\n\n with open(p, \"rb\") as f:\n self.write(f.read())\n\n\nclass P12(tornado.web.RequestHandler):\n\n @property\n def filename(self):\n return config.CONF_BASENAME + \"-ca-cert.p12\"\n\n def get(self):\n p = os.path.join(self.request.master.options.cadir, self.filename)\n p = os.path.expanduser(p)\n self.set_header(\"Content-Type\", \"application/x-pkcs12\")\n self.set_header(\n \"Content-Disposition\",\n \"inline; filename={}\".format(\n self.filename))\n\n with open(p, \"rb\") as f:\n self.write(f.read())\n\n\napplication = tornado.web.Application(\n [\n (r\"/\", Index),\n (r\"/cert/pem\", PEM),\n (r\"/cert/p12\", P12),\n (\n r\"/static/(.*)\",\n tornado.web.StaticFileHandler,\n {\n \"path\": data.pkg_data.path(\"addons/onboardingapp/static\")\n }\n ),\n ],\n # debug=True\n)\n", "path": "mitmproxy/addons/onboardingapp/app.py"}]}
| 2,780 | 386 |
gh_patches_debug_35028
|
rasdani/github-patches
|
git_diff
|
strawberry-graphql__strawberry-1071
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Compatibility with pydantic 1.4
I'm trying to use strawberry in a project that has pydantic pinned at 1.4. I chatted with @patrick91 on discord about this, and he thought it would be reasonable to achieve compatibility with this version.
Pydantic appears to only be used in the [strawberry.experimental](https://github.com/strawberry-graphql/strawberry/blob/main/strawberry/experimental/__init__.py) module, which only gets loaded if pydantic is present. One way to solve this for me in particular would be to lazily load strawberry.experimental/pydantic, such that when an older version of pydantic is present, one can still import other packages in strawberry.
Thank you!
</issue>
<code>
[start of strawberry/experimental/pydantic/fields.py]
1 from decimal import Decimal
2 from typing import Optional
3 from uuid import UUID
4
5 import pydantic
6
7 from .exceptions import UnsupportedTypeError
8
9
10 FIELDS_MAP = {
11 pydantic.NoneStr: Optional[str],
12 pydantic.NoneBytes: Optional[bytes],
13 pydantic.StrBytes: None,
14 pydantic.NoneStrBytes: None,
15 pydantic.StrictStr: str,
16 pydantic.ConstrainedBytes: bytes,
17 pydantic.conbytes: bytes,
18 pydantic.ConstrainedList: None,
19 pydantic.conlist: None,
20 pydantic.ConstrainedSet: None,
21 pydantic.conset: None,
22 pydantic.ConstrainedStr: str,
23 pydantic.constr: str,
24 pydantic.EmailStr: str,
25 pydantic.PyObject: None,
26 pydantic.ConstrainedInt: int,
27 pydantic.conint: int,
28 pydantic.PositiveInt: int,
29 pydantic.NegativeInt: int,
30 pydantic.ConstrainedFloat: float,
31 pydantic.confloat: float,
32 pydantic.PositiveFloat: float,
33 pydantic.NegativeFloat: float,
34 pydantic.ConstrainedDecimal: Decimal,
35 pydantic.condecimal: Decimal,
36 pydantic.UUID1: UUID,
37 pydantic.UUID3: UUID,
38 pydantic.UUID4: UUID,
39 pydantic.UUID5: UUID,
40 pydantic.FilePath: None,
41 pydantic.DirectoryPath: None,
42 pydantic.Json: None,
43 pydantic.JsonWrapper: None,
44 pydantic.SecretStr: str,
45 pydantic.SecretBytes: bytes,
46 pydantic.StrictBool: bool,
47 pydantic.StrictInt: int,
48 pydantic.StrictFloat: float,
49 pydantic.PaymentCardNumber: None,
50 pydantic.ByteSize: None,
51 pydantic.AnyUrl: str,
52 pydantic.AnyHttpUrl: str,
53 pydantic.HttpUrl: str,
54 pydantic.PostgresDsn: str,
55 pydantic.RedisDsn: str,
56 }
57
58
59 def get_basic_type(type_):
60 if isinstance(type_, type):
61 if issubclass(type_, pydantic.ConstrainedInt):
62 return int
63 if issubclass(type_, pydantic.ConstrainedStr):
64 return str
65
66 if type_ in FIELDS_MAP:
67 type_ = FIELDS_MAP.get(type_)
68
69 if type_ is None:
70 raise UnsupportedTypeError()
71
72 return type_
73
[end of strawberry/experimental/pydantic/fields.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/strawberry/experimental/pydantic/fields.py b/strawberry/experimental/pydantic/fields.py
--- a/strawberry/experimental/pydantic/fields.py
+++ b/strawberry/experimental/pydantic/fields.py
@@ -7,52 +7,59 @@
from .exceptions import UnsupportedTypeError
+ATTR_TO_TYPE_MAP = {
+ "NoneStr": Optional[str],
+ "NoneBytes": Optional[bytes],
+ "StrBytes": None,
+ "NoneStrBytes": None,
+ "StrictStr": str,
+ "ConstrainedBytes": bytes,
+ "conbytes": bytes,
+ "ConstrainedList": None,
+ "conlist": None,
+ "ConstrainedSet": None,
+ "conset": None,
+ "ConstrainedStr": str,
+ "constr": str,
+ "EmailStr": str,
+ "PyObject": None,
+ "ConstrainedInt": int,
+ "conint": int,
+ "PositiveInt": int,
+ "NegativeInt": int,
+ "ConstrainedFloat": float,
+ "confloat": float,
+ "PositiveFloat": float,
+ "NegativeFloat": float,
+ "ConstrainedDecimal": Decimal,
+ "condecimal": Decimal,
+ "UUID1": UUID,
+ "UUID3": UUID,
+ "UUID4": UUID,
+ "UUID5": UUID,
+ "FilePath": None,
+ "DirectoryPath": None,
+ "Json": None,
+ "JsonWrapper": None,
+ "SecretStr": str,
+ "SecretBytes": bytes,
+ "StrictBool": bool,
+ "StrictInt": int,
+ "StrictFloat": float,
+ "PaymentCardNumber": None,
+ "ByteSize": None,
+ "AnyUrl": str,
+ "AnyHttpUrl": str,
+ "HttpUrl": str,
+ "PostgresDsn": str,
+ "RedisDsn": str,
+}
+
+
FIELDS_MAP = {
- pydantic.NoneStr: Optional[str],
- pydantic.NoneBytes: Optional[bytes],
- pydantic.StrBytes: None,
- pydantic.NoneStrBytes: None,
- pydantic.StrictStr: str,
- pydantic.ConstrainedBytes: bytes,
- pydantic.conbytes: bytes,
- pydantic.ConstrainedList: None,
- pydantic.conlist: None,
- pydantic.ConstrainedSet: None,
- pydantic.conset: None,
- pydantic.ConstrainedStr: str,
- pydantic.constr: str,
- pydantic.EmailStr: str,
- pydantic.PyObject: None,
- pydantic.ConstrainedInt: int,
- pydantic.conint: int,
- pydantic.PositiveInt: int,
- pydantic.NegativeInt: int,
- pydantic.ConstrainedFloat: float,
- pydantic.confloat: float,
- pydantic.PositiveFloat: float,
- pydantic.NegativeFloat: float,
- pydantic.ConstrainedDecimal: Decimal,
- pydantic.condecimal: Decimal,
- pydantic.UUID1: UUID,
- pydantic.UUID3: UUID,
- pydantic.UUID4: UUID,
- pydantic.UUID5: UUID,
- pydantic.FilePath: None,
- pydantic.DirectoryPath: None,
- pydantic.Json: None,
- pydantic.JsonWrapper: None,
- pydantic.SecretStr: str,
- pydantic.SecretBytes: bytes,
- pydantic.StrictBool: bool,
- pydantic.StrictInt: int,
- pydantic.StrictFloat: float,
- pydantic.PaymentCardNumber: None,
- pydantic.ByteSize: None,
- pydantic.AnyUrl: str,
- pydantic.AnyHttpUrl: str,
- pydantic.HttpUrl: str,
- pydantic.PostgresDsn: str,
- pydantic.RedisDsn: str,
+ getattr(pydantic, field_name): type
+ for field_name, type in ATTR_TO_TYPE_MAP.items()
+ if hasattr(pydantic, field_name)
}
|
{"golden_diff": "diff --git a/strawberry/experimental/pydantic/fields.py b/strawberry/experimental/pydantic/fields.py\n--- a/strawberry/experimental/pydantic/fields.py\n+++ b/strawberry/experimental/pydantic/fields.py\n@@ -7,52 +7,59 @@\n from .exceptions import UnsupportedTypeError\n \n \n+ATTR_TO_TYPE_MAP = {\n+ \"NoneStr\": Optional[str],\n+ \"NoneBytes\": Optional[bytes],\n+ \"StrBytes\": None,\n+ \"NoneStrBytes\": None,\n+ \"StrictStr\": str,\n+ \"ConstrainedBytes\": bytes,\n+ \"conbytes\": bytes,\n+ \"ConstrainedList\": None,\n+ \"conlist\": None,\n+ \"ConstrainedSet\": None,\n+ \"conset\": None,\n+ \"ConstrainedStr\": str,\n+ \"constr\": str,\n+ \"EmailStr\": str,\n+ \"PyObject\": None,\n+ \"ConstrainedInt\": int,\n+ \"conint\": int,\n+ \"PositiveInt\": int,\n+ \"NegativeInt\": int,\n+ \"ConstrainedFloat\": float,\n+ \"confloat\": float,\n+ \"PositiveFloat\": float,\n+ \"NegativeFloat\": float,\n+ \"ConstrainedDecimal\": Decimal,\n+ \"condecimal\": Decimal,\n+ \"UUID1\": UUID,\n+ \"UUID3\": UUID,\n+ \"UUID4\": UUID,\n+ \"UUID5\": UUID,\n+ \"FilePath\": None,\n+ \"DirectoryPath\": None,\n+ \"Json\": None,\n+ \"JsonWrapper\": None,\n+ \"SecretStr\": str,\n+ \"SecretBytes\": bytes,\n+ \"StrictBool\": bool,\n+ \"StrictInt\": int,\n+ \"StrictFloat\": float,\n+ \"PaymentCardNumber\": None,\n+ \"ByteSize\": None,\n+ \"AnyUrl\": str,\n+ \"AnyHttpUrl\": str,\n+ \"HttpUrl\": str,\n+ \"PostgresDsn\": str,\n+ \"RedisDsn\": str,\n+}\n+\n+\n FIELDS_MAP = {\n- pydantic.NoneStr: Optional[str],\n- pydantic.NoneBytes: Optional[bytes],\n- pydantic.StrBytes: None,\n- pydantic.NoneStrBytes: None,\n- pydantic.StrictStr: str,\n- pydantic.ConstrainedBytes: bytes,\n- pydantic.conbytes: bytes,\n- pydantic.ConstrainedList: None,\n- pydantic.conlist: None,\n- pydantic.ConstrainedSet: None,\n- pydantic.conset: None,\n- pydantic.ConstrainedStr: str,\n- pydantic.constr: str,\n- pydantic.EmailStr: str,\n- pydantic.PyObject: None,\n- pydantic.ConstrainedInt: int,\n- pydantic.conint: int,\n- pydantic.PositiveInt: int,\n- pydantic.NegativeInt: int,\n- pydantic.ConstrainedFloat: float,\n- pydantic.confloat: float,\n- pydantic.PositiveFloat: float,\n- pydantic.NegativeFloat: float,\n- pydantic.ConstrainedDecimal: Decimal,\n- pydantic.condecimal: Decimal,\n- pydantic.UUID1: UUID,\n- pydantic.UUID3: UUID,\n- pydantic.UUID4: UUID,\n- pydantic.UUID5: UUID,\n- pydantic.FilePath: None,\n- pydantic.DirectoryPath: None,\n- pydantic.Json: None,\n- pydantic.JsonWrapper: None,\n- pydantic.SecretStr: str,\n- pydantic.SecretBytes: bytes,\n- pydantic.StrictBool: bool,\n- pydantic.StrictInt: int,\n- pydantic.StrictFloat: float,\n- pydantic.PaymentCardNumber: None,\n- pydantic.ByteSize: None,\n- pydantic.AnyUrl: str,\n- pydantic.AnyHttpUrl: str,\n- pydantic.HttpUrl: str,\n- pydantic.PostgresDsn: str,\n- pydantic.RedisDsn: str,\n+ getattr(pydantic, field_name): type\n+ for field_name, type in ATTR_TO_TYPE_MAP.items()\n+ if hasattr(pydantic, field_name)\n }\n", "issue": "Compatibility with pydantic 1.4\nI'm trying to use strawberry in a project that has pydantic pinned at 1.4. I chatted with @patrick91 on discord about this, and he thought it would be reasonable to achieve compatibility with this version.\r\n\r\nPydantic appears to only be used in the [strawberry.experimental](https://github.com/strawberry-graphql/strawberry/blob/main/strawberry/experimental/__init__.py) module, which only gets loaded if pydantic is present. One way to solve this for me in particular would be to lazily load strawberry.experimental/pydantic, such that when an older version of pydantic is present, one can still import other packages in strawberry.\r\n\r\nThank you!\n", "before_files": [{"content": "from decimal import Decimal\nfrom typing import Optional\nfrom uuid import UUID\n\nimport pydantic\n\nfrom .exceptions import UnsupportedTypeError\n\n\nFIELDS_MAP = {\n pydantic.NoneStr: Optional[str],\n pydantic.NoneBytes: Optional[bytes],\n pydantic.StrBytes: None,\n pydantic.NoneStrBytes: None,\n pydantic.StrictStr: str,\n pydantic.ConstrainedBytes: bytes,\n pydantic.conbytes: bytes,\n pydantic.ConstrainedList: None,\n pydantic.conlist: None,\n pydantic.ConstrainedSet: None,\n pydantic.conset: None,\n pydantic.ConstrainedStr: str,\n pydantic.constr: str,\n pydantic.EmailStr: str,\n pydantic.PyObject: None,\n pydantic.ConstrainedInt: int,\n pydantic.conint: int,\n pydantic.PositiveInt: int,\n pydantic.NegativeInt: int,\n pydantic.ConstrainedFloat: float,\n pydantic.confloat: float,\n pydantic.PositiveFloat: float,\n pydantic.NegativeFloat: float,\n pydantic.ConstrainedDecimal: Decimal,\n pydantic.condecimal: Decimal,\n pydantic.UUID1: UUID,\n pydantic.UUID3: UUID,\n pydantic.UUID4: UUID,\n pydantic.UUID5: UUID,\n pydantic.FilePath: None,\n pydantic.DirectoryPath: None,\n pydantic.Json: None,\n pydantic.JsonWrapper: None,\n pydantic.SecretStr: str,\n pydantic.SecretBytes: bytes,\n pydantic.StrictBool: bool,\n pydantic.StrictInt: int,\n pydantic.StrictFloat: float,\n pydantic.PaymentCardNumber: None,\n pydantic.ByteSize: None,\n pydantic.AnyUrl: str,\n pydantic.AnyHttpUrl: str,\n pydantic.HttpUrl: str,\n pydantic.PostgresDsn: str,\n pydantic.RedisDsn: str,\n}\n\n\ndef get_basic_type(type_):\n if isinstance(type_, type):\n if issubclass(type_, pydantic.ConstrainedInt):\n return int\n if issubclass(type_, pydantic.ConstrainedStr):\n return str\n\n if type_ in FIELDS_MAP:\n type_ = FIELDS_MAP.get(type_)\n\n if type_ is None:\n raise UnsupportedTypeError()\n\n return type_\n", "path": "strawberry/experimental/pydantic/fields.py"}]}
| 1,388 | 975 |
gh_patches_debug_26101
|
rasdani/github-patches
|
git_diff
|
meltano__meltano-6433
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Disallow calling a `mapper` by name in `meltano run` and alert to use a `mapping` name
The document misses a step for mapper configuration. Property `mapping_name` has to be added to the mappers extras, and its value need to match with the name under mappings. It doesn't seem been documented anywhere.
```yaml
plugins:
mappers:
- name: transform-field
variant: transferwise
pip_url: pipelinewise-transform-field
executable: transform-field
mapping_name: transform-field-hash
mappings:
- name: transform-field-hash
config:
transformations:
- field_id: Name
tap_stream_name: dbo-Test
type: MASK-HIDDEN
```
</issue>
<code>
[start of src/meltano/core/block/parser.py]
1 """Utilities for turning a string list of plugins into a usable list of BlockSet and PluginCommand objects."""
2 from typing import Dict, Generator, List, Optional, Tuple, Union
3
4 import click
5 import structlog
6
7 from meltano.core.plugin import PluginType
8 from meltano.core.plugin.error import PluginNotFoundError
9 from meltano.core.plugin.project_plugin import ProjectPlugin
10 from meltano.core.project_plugins_service import ProjectPluginsService
11 from meltano.core.task_sets_service import TaskSetsService
12
13 from .blockset import BlockSet, BlockSetValidationError
14 from .extract_load import ELBContextBuilder, ExtractLoadBlocks
15 from .plugin_command import PluginCommandBlock, plugin_command_invoker
16 from .singer import CONSUMERS, SingerBlock
17
18
19 def is_command_block(plugin: ProjectPlugin) -> bool:
20 """Check if a plugin is a command block.
21
22 Args:
23 plugin: Plugin to check.
24
25 Returns:
26 True if plugin is a command block.
27 """
28 return plugin.type not in {
29 PluginType.EXTRACTORS,
30 PluginType.LOADERS,
31 PluginType.MAPPERS,
32 }
33
34
35 def validate_block_sets(
36 log: structlog.BoundLogger, blocks: List[Union[BlockSet, PluginCommandBlock]]
37 ) -> bool:
38 """Perform validation of all blocks in a list that implement the BlockSet interface.
39
40 Args:
41 log: Logger to use in the event of a validation error.
42 blocks: A list of blocks.
43
44 Returns:
45 True if all blocks are valid, False otherwise.
46 """
47 for idx, blk in enumerate(blocks):
48 if blk == BlockSet:
49 log.debug("validating ExtractLoadBlock.", set_number=idx)
50 try:
51 blk.validate_set()
52 except Exception as err:
53 log.error("Validation failed.", err=err)
54 return False
55 return True
56
57
58 class BlockParser: # noqa: D101
59 def __init__(
60 self,
61 log: structlog.BoundLogger,
62 project,
63 blocks: List[str],
64 full_refresh: Optional[bool] = False,
65 no_state_update: Optional[bool] = False,
66 force: Optional[bool] = False,
67 ):
68 """
69 Parse a meltano run command invocation into a list of blocks.
70
71 Args:
72 log: Logger to use.
73 project: Project to use.
74 blocks: List of block names to parse.
75 full_refresh: Whether to perform a full refresh (applies to all found sets).
76 no_state_update: Whether to run with or without state updates.
77 force: Whether to force a run if a job is already running (applies to all found sets).
78
79 Raises:
80 ClickException: If a block name is not found.
81 """
82 self.log = log
83 self.project = project
84
85 self._full_refresh = full_refresh
86 self._no_state_update = no_state_update
87 self._force = force
88
89 self._plugins_service = ProjectPluginsService(project)
90 self._plugins: List[ProjectPlugin] = []
91
92 self._commands: Dict[int, str] = {}
93 self._mappings_ref: Dict[int, str] = {}
94
95 task_sets_service: TaskSetsService = TaskSetsService(project)
96
97 blocks = self._expand_jobs(blocks, task_sets_service)
98
99 for idx, name in enumerate(blocks):
100
101 try:
102 parsed_name, command_name = name.split(":")
103 except ValueError:
104 parsed_name = name
105 command_name = None
106
107 plugin = self._find_plugin_or_mapping(parsed_name)
108 if plugin is None:
109 raise click.ClickException(f"Block {name} not found")
110
111 if plugin and task_sets_service.exists(name):
112 raise click.ClickException(
113 f"Ambiguous reference to '{name}' which matches a job name AND a plugin name."
114 )
115
116 if plugin.type == PluginType.MAPPERS:
117 self._mappings_ref[idx] = parsed_name
118
119 self._plugins.append(plugin)
120 if command_name:
121 self._commands[idx] = command_name
122 self.log.debug(
123 "plugin command added for execution",
124 commands=self._commands,
125 command_name=command_name,
126 plugin_name=parsed_name,
127 )
128
129 self.log.debug("found plugin in cli invocation", plugin_name=plugin.name)
130
131 def _expand_jobs(self, blocks: List[str], task_sets: TaskSetsService) -> List[str]:
132 """Expand any jobs present in a list of blocks into their raw block names.
133
134 Example:
135 Given a job named "somejob" which consists of a single task of "tap target":
136 ["somejob", "dbt:run"] -> ["tap", "target", "dbt:run"]
137
138 Args:
139 blocks: List of block names to parse.
140 task_sets: TaskSetsService to use.
141
142 Returns:
143 List of block names with jobs expanded.
144 """
145 expanded_blocks: List[str] = []
146 for name in blocks:
147 if task_sets.exists(name):
148 self.log.debug(
149 "expanding job to tasks",
150 job_name=name,
151 tasks=task_sets.get(name).flat_args,
152 )
153 expanded_blocks.extend(task_sets.get(name).flat_args)
154 else:
155 expanded_blocks.append(name)
156 return expanded_blocks
157
158 def find_blocks(
159 self, offset: int = 0
160 ) -> Generator[Union[BlockSet, PluginCommandBlock], None, None]:
161 """
162 Find all blocks in the invocation.
163
164 Args:
165 offset: Offset to start from.
166
167 Yields:
168 Generator of blocks (either BlockSet or PluginCommandBlock).
169
170 Raises:
171 BlockSetValidationError: If unknown command is found or if a unexpected block sequence is found.
172 """
173 cur = offset
174 while cur < len(self._plugins):
175 plugin = self._plugins[cur]
176 elb, idx = self._find_next_elb_set(cur)
177 if elb:
178 self.log.debug("found ExtractLoadBlocks set", offset=cur)
179 yield elb
180 cur += idx
181 elif is_command_block(plugin):
182 self.log.debug(
183 "found PluginCommand",
184 offset=cur,
185 plugin_type=plugin.type,
186 )
187 yield plugin_command_invoker(
188 self._plugins[cur],
189 self.project,
190 command=self._commands.get(cur),
191 )
192 cur += 1
193 else:
194 raise BlockSetValidationError(
195 f"Unknown command type or bad block sequence at index {cur + 1}, starting block '{plugin.name}'" # noqa: WPS237
196 )
197
198 def _find_plugin_or_mapping(self, name: str) -> Optional[ProjectPlugin]:
199 """Find a plugin by name OR by mapping name.
200
201 Args:
202 name: Name of the plugin or mapping.
203
204 Returns:
205 The actual plugin.
206
207 Raises:
208 ClickException: If mapping name returns multiple matches.
209 """
210 try:
211 return self._plugins_service.find_plugin(name)
212 except PluginNotFoundError:
213 pass
214
215 mapper = None
216 try:
217 mapper = self._plugins_service.find_plugins_by_mapping_name(name)
218 except PluginNotFoundError:
219 pass
220
221 if mapper is None:
222 return None
223
224 if len(mapper) > 1:
225 raise click.ClickException(
226 f"Ambiguous mapping name {name}, found multiple matches."
227 )
228 return mapper[0] if mapper else None
229
230 def _find_next_elb_set( # noqa: WPS231, WPS213
231 self,
232 offset: int = 0,
233 ) -> Tuple[Optional[ExtractLoadBlocks], int]: # noqa: WPS231, WPS213
234 """
235 Search a list of project plugins trying to find an extract ExtractLoad block set.
236
237 Args:
238 offset: Optional starting offset for search.
239
240 Returns:
241 The ExtractLoad object.
242 Offset for remaining plugins.
243
244 Raises:
245 BlockSetValidationError: If the block set is not valid.
246 """
247 blocks: List[SingerBlock] = []
248
249 base_builder = ELBContextBuilder(
250 self.project, self._plugins_service
251 ) # lint work around
252 builder = (
253 base_builder.with_force(self._force)
254 .with_full_refresh(self._full_refresh)
255 .with_no_state_update(self._no_state_update)
256 )
257
258 if self._plugins[offset].type != PluginType.EXTRACTORS:
259 self.log.debug(
260 "next block not extractor",
261 offset=offset,
262 plugin_type=self._plugins[offset].type,
263 )
264 return None, offset
265
266 self.log.debug(
267 "head of set is extractor as expected", block=self._plugins[offset]
268 )
269
270 blocks.append(builder.make_block(self._plugins[offset]))
271
272 for idx, plugin in enumerate(self._plugins[offset + 1 :]): # noqa: E203
273 next_block = idx + 1
274
275 if plugin.type not in CONSUMERS:
276 self.log.debug(
277 "next block not a consumer of output",
278 offset=offset,
279 plugin_type=plugin.type,
280 )
281 return None, offset + next_block
282
283 self.log.debug("found block", block_type=plugin.type, index=next_block)
284
285 if plugin.type == PluginType.MAPPERS:
286 self.log.debug(
287 "found mapper",
288 plugin_type=plugin.type,
289 plugin_name=plugin.name,
290 mapping=self._mappings_ref.get(next_block),
291 idx=next_block,
292 )
293 blocks.append(
294 builder.make_block(
295 plugin,
296 )
297 )
298 elif plugin.type == PluginType.LOADERS:
299 self.log.debug("blocks", offset=offset, idx=next_block)
300 blocks.append(builder.make_block(plugin))
301 elb = ExtractLoadBlocks(builder.context(), blocks)
302 return elb, idx + 2
303 else:
304 self.log.warning(
305 "Found unexpected plugin type for block in middle of block set.",
306 plugin_type=plugin.type,
307 plugin_name=plugin.name,
308 )
309 raise BlockSetValidationError(
310 f"Expected {PluginType.MAPPERS} or {PluginType.LOADERS}."
311 )
312 raise BlockSetValidationError("Found no end in block set!")
313
[end of src/meltano/core/block/parser.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/src/meltano/core/block/parser.py b/src/meltano/core/block/parser.py
--- a/src/meltano/core/block/parser.py
+++ b/src/meltano/core/block/parser.py
@@ -290,11 +290,25 @@
mapping=self._mappings_ref.get(next_block),
idx=next_block,
)
- blocks.append(
- builder.make_block(
- plugin,
+ # Checks to see if the mapper plugin name is the same as the mappings name
+ # If they both match then a validation error is raised because the
+ # meltano run command needs the mappings name to obtain the settings to
+ # pass to the parent mapper plugin. We also want to fail if the user names them
+ # the same to stop errors due to ambiguous commands.
+ if plugin.name == self._mappings_ref.get(next_block):
+ self.log.warning(
+ "Found unexpected mapper plugin name. ",
+ plugin_name=plugin.name,
+ )
+ raise BlockSetValidationError(
+ f"Expected unique mappings name not the mapper plugin name: {plugin.name}."
+ )
+ else:
+ blocks.append(
+ builder.make_block(
+ plugin,
+ )
)
- )
elif plugin.type == PluginType.LOADERS:
self.log.debug("blocks", offset=offset, idx=next_block)
blocks.append(builder.make_block(plugin))
|
{"golden_diff": "diff --git a/src/meltano/core/block/parser.py b/src/meltano/core/block/parser.py\n--- a/src/meltano/core/block/parser.py\n+++ b/src/meltano/core/block/parser.py\n@@ -290,11 +290,25 @@\n mapping=self._mappings_ref.get(next_block),\n idx=next_block,\n )\n- blocks.append(\n- builder.make_block(\n- plugin,\n+ # Checks to see if the mapper plugin name is the same as the mappings name\n+ # If they both match then a validation error is raised because the\n+ # meltano run command needs the mappings name to obtain the settings to\n+ # pass to the parent mapper plugin. We also want to fail if the user names them\n+ # the same to stop errors due to ambiguous commands.\n+ if plugin.name == self._mappings_ref.get(next_block):\n+ self.log.warning(\n+ \"Found unexpected mapper plugin name. \",\n+ plugin_name=plugin.name,\n+ )\n+ raise BlockSetValidationError(\n+ f\"Expected unique mappings name not the mapper plugin name: {plugin.name}.\"\n+ )\n+ else:\n+ blocks.append(\n+ builder.make_block(\n+ plugin,\n+ )\n )\n- )\n elif plugin.type == PluginType.LOADERS:\n self.log.debug(\"blocks\", offset=offset, idx=next_block)\n blocks.append(builder.make_block(plugin))\n", "issue": "Disallow calling a `mapper` by name in `meltano run` and alert to use a `mapping` name\nThe document misses a step for mapper configuration. Property `mapping_name` has to be added to the mappers extras, and its value need to match with the name under mappings. It doesn't seem been documented anywhere.\r\n\r\n```yaml\r\nplugins:\r\n mappers:\r\n - name: transform-field\r\n variant: transferwise\r\n pip_url: pipelinewise-transform-field\r\n executable: transform-field\r\n mapping_name: transform-field-hash\r\n mappings:\r\n - name: transform-field-hash\r\n config:\r\n transformations:\r\n - field_id: Name\r\n tap_stream_name: dbo-Test\r\n type: MASK-HIDDEN\r\n```\n", "before_files": [{"content": "\"\"\"Utilities for turning a string list of plugins into a usable list of BlockSet and PluginCommand objects.\"\"\"\nfrom typing import Dict, Generator, List, Optional, Tuple, Union\n\nimport click\nimport structlog\n\nfrom meltano.core.plugin import PluginType\nfrom meltano.core.plugin.error import PluginNotFoundError\nfrom meltano.core.plugin.project_plugin import ProjectPlugin\nfrom meltano.core.project_plugins_service import ProjectPluginsService\nfrom meltano.core.task_sets_service import TaskSetsService\n\nfrom .blockset import BlockSet, BlockSetValidationError\nfrom .extract_load import ELBContextBuilder, ExtractLoadBlocks\nfrom .plugin_command import PluginCommandBlock, plugin_command_invoker\nfrom .singer import CONSUMERS, SingerBlock\n\n\ndef is_command_block(plugin: ProjectPlugin) -> bool:\n \"\"\"Check if a plugin is a command block.\n\n Args:\n plugin: Plugin to check.\n\n Returns:\n True if plugin is a command block.\n \"\"\"\n return plugin.type not in {\n PluginType.EXTRACTORS,\n PluginType.LOADERS,\n PluginType.MAPPERS,\n }\n\n\ndef validate_block_sets(\n log: structlog.BoundLogger, blocks: List[Union[BlockSet, PluginCommandBlock]]\n) -> bool:\n \"\"\"Perform validation of all blocks in a list that implement the BlockSet interface.\n\n Args:\n log: Logger to use in the event of a validation error.\n blocks: A list of blocks.\n\n Returns:\n True if all blocks are valid, False otherwise.\n \"\"\"\n for idx, blk in enumerate(blocks):\n if blk == BlockSet:\n log.debug(\"validating ExtractLoadBlock.\", set_number=idx)\n try:\n blk.validate_set()\n except Exception as err:\n log.error(\"Validation failed.\", err=err)\n return False\n return True\n\n\nclass BlockParser: # noqa: D101\n def __init__(\n self,\n log: structlog.BoundLogger,\n project,\n blocks: List[str],\n full_refresh: Optional[bool] = False,\n no_state_update: Optional[bool] = False,\n force: Optional[bool] = False,\n ):\n \"\"\"\n Parse a meltano run command invocation into a list of blocks.\n\n Args:\n log: Logger to use.\n project: Project to use.\n blocks: List of block names to parse.\n full_refresh: Whether to perform a full refresh (applies to all found sets).\n no_state_update: Whether to run with or without state updates.\n force: Whether to force a run if a job is already running (applies to all found sets).\n\n Raises:\n ClickException: If a block name is not found.\n \"\"\"\n self.log = log\n self.project = project\n\n self._full_refresh = full_refresh\n self._no_state_update = no_state_update\n self._force = force\n\n self._plugins_service = ProjectPluginsService(project)\n self._plugins: List[ProjectPlugin] = []\n\n self._commands: Dict[int, str] = {}\n self._mappings_ref: Dict[int, str] = {}\n\n task_sets_service: TaskSetsService = TaskSetsService(project)\n\n blocks = self._expand_jobs(blocks, task_sets_service)\n\n for idx, name in enumerate(blocks):\n\n try:\n parsed_name, command_name = name.split(\":\")\n except ValueError:\n parsed_name = name\n command_name = None\n\n plugin = self._find_plugin_or_mapping(parsed_name)\n if plugin is None:\n raise click.ClickException(f\"Block {name} not found\")\n\n if plugin and task_sets_service.exists(name):\n raise click.ClickException(\n f\"Ambiguous reference to '{name}' which matches a job name AND a plugin name.\"\n )\n\n if plugin.type == PluginType.MAPPERS:\n self._mappings_ref[idx] = parsed_name\n\n self._plugins.append(plugin)\n if command_name:\n self._commands[idx] = command_name\n self.log.debug(\n \"plugin command added for execution\",\n commands=self._commands,\n command_name=command_name,\n plugin_name=parsed_name,\n )\n\n self.log.debug(\"found plugin in cli invocation\", plugin_name=plugin.name)\n\n def _expand_jobs(self, blocks: List[str], task_sets: TaskSetsService) -> List[str]:\n \"\"\"Expand any jobs present in a list of blocks into their raw block names.\n\n Example:\n Given a job named \"somejob\" which consists of a single task of \"tap target\":\n [\"somejob\", \"dbt:run\"] -> [\"tap\", \"target\", \"dbt:run\"]\n\n Args:\n blocks: List of block names to parse.\n task_sets: TaskSetsService to use.\n\n Returns:\n List of block names with jobs expanded.\n \"\"\"\n expanded_blocks: List[str] = []\n for name in blocks:\n if task_sets.exists(name):\n self.log.debug(\n \"expanding job to tasks\",\n job_name=name,\n tasks=task_sets.get(name).flat_args,\n )\n expanded_blocks.extend(task_sets.get(name).flat_args)\n else:\n expanded_blocks.append(name)\n return expanded_blocks\n\n def find_blocks(\n self, offset: int = 0\n ) -> Generator[Union[BlockSet, PluginCommandBlock], None, None]:\n \"\"\"\n Find all blocks in the invocation.\n\n Args:\n offset: Offset to start from.\n\n Yields:\n Generator of blocks (either BlockSet or PluginCommandBlock).\n\n Raises:\n BlockSetValidationError: If unknown command is found or if a unexpected block sequence is found.\n \"\"\"\n cur = offset\n while cur < len(self._plugins):\n plugin = self._plugins[cur]\n elb, idx = self._find_next_elb_set(cur)\n if elb:\n self.log.debug(\"found ExtractLoadBlocks set\", offset=cur)\n yield elb\n cur += idx\n elif is_command_block(plugin):\n self.log.debug(\n \"found PluginCommand\",\n offset=cur,\n plugin_type=plugin.type,\n )\n yield plugin_command_invoker(\n self._plugins[cur],\n self.project,\n command=self._commands.get(cur),\n )\n cur += 1\n else:\n raise BlockSetValidationError(\n f\"Unknown command type or bad block sequence at index {cur + 1}, starting block '{plugin.name}'\" # noqa: WPS237\n )\n\n def _find_plugin_or_mapping(self, name: str) -> Optional[ProjectPlugin]:\n \"\"\"Find a plugin by name OR by mapping name.\n\n Args:\n name: Name of the plugin or mapping.\n\n Returns:\n The actual plugin.\n\n Raises:\n ClickException: If mapping name returns multiple matches.\n \"\"\"\n try:\n return self._plugins_service.find_plugin(name)\n except PluginNotFoundError:\n pass\n\n mapper = None\n try:\n mapper = self._plugins_service.find_plugins_by_mapping_name(name)\n except PluginNotFoundError:\n pass\n\n if mapper is None:\n return None\n\n if len(mapper) > 1:\n raise click.ClickException(\n f\"Ambiguous mapping name {name}, found multiple matches.\"\n )\n return mapper[0] if mapper else None\n\n def _find_next_elb_set( # noqa: WPS231, WPS213\n self,\n offset: int = 0,\n ) -> Tuple[Optional[ExtractLoadBlocks], int]: # noqa: WPS231, WPS213\n \"\"\"\n Search a list of project plugins trying to find an extract ExtractLoad block set.\n\n Args:\n offset: Optional starting offset for search.\n\n Returns:\n The ExtractLoad object.\n Offset for remaining plugins.\n\n Raises:\n BlockSetValidationError: If the block set is not valid.\n \"\"\"\n blocks: List[SingerBlock] = []\n\n base_builder = ELBContextBuilder(\n self.project, self._plugins_service\n ) # lint work around\n builder = (\n base_builder.with_force(self._force)\n .with_full_refresh(self._full_refresh)\n .with_no_state_update(self._no_state_update)\n )\n\n if self._plugins[offset].type != PluginType.EXTRACTORS:\n self.log.debug(\n \"next block not extractor\",\n offset=offset,\n plugin_type=self._plugins[offset].type,\n )\n return None, offset\n\n self.log.debug(\n \"head of set is extractor as expected\", block=self._plugins[offset]\n )\n\n blocks.append(builder.make_block(self._plugins[offset]))\n\n for idx, plugin in enumerate(self._plugins[offset + 1 :]): # noqa: E203\n next_block = idx + 1\n\n if plugin.type not in CONSUMERS:\n self.log.debug(\n \"next block not a consumer of output\",\n offset=offset,\n plugin_type=plugin.type,\n )\n return None, offset + next_block\n\n self.log.debug(\"found block\", block_type=plugin.type, index=next_block)\n\n if plugin.type == PluginType.MAPPERS:\n self.log.debug(\n \"found mapper\",\n plugin_type=plugin.type,\n plugin_name=plugin.name,\n mapping=self._mappings_ref.get(next_block),\n idx=next_block,\n )\n blocks.append(\n builder.make_block(\n plugin,\n )\n )\n elif plugin.type == PluginType.LOADERS:\n self.log.debug(\"blocks\", offset=offset, idx=next_block)\n blocks.append(builder.make_block(plugin))\n elb = ExtractLoadBlocks(builder.context(), blocks)\n return elb, idx + 2\n else:\n self.log.warning(\n \"Found unexpected plugin type for block in middle of block set.\",\n plugin_type=plugin.type,\n plugin_name=plugin.name,\n )\n raise BlockSetValidationError(\n f\"Expected {PluginType.MAPPERS} or {PluginType.LOADERS}.\"\n )\n raise BlockSetValidationError(\"Found no end in block set!\")\n", "path": "src/meltano/core/block/parser.py"}]}
| 3,712 | 314 |
gh_patches_debug_9298
|
rasdani/github-patches
|
git_diff
|
joke2k__faker-1607
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
es_ES postalcode is not generating valid codes
* Faker version: 11.3
* OS: Any
When using postcode for es_ES and using it with a field that requires a valid Postal Code, it fails sometimes.
I will assume that there is no logic with postal code generation for Spain.
### Steps to reproduce
Generate postal codes
### Expected behavior
Get a valid Spain postal code
### Actual behavior
Unexpected. Many are wrong
----
I'll dig now into the code. Let's see if I can get some more information and fix it :thinking: Do not expect much from me
</issue>
<code>
[start of faker/providers/address/es_ES/__init__.py]
1 from ..es import Provider as AddressProvider
2
3
4 class Provider(AddressProvider):
5 building_number_formats = ("%", "%#", "%#", "%#", "%##")
6 street_prefixes = (
7 "Plaza",
8 "Calle",
9 "Avenida",
10 "Via",
11 "Vial",
12 "Rambla",
13 "Glorieta",
14 "Urbanización",
15 "Callejón",
16 "Cañada",
17 "Alameda",
18 "Acceso",
19 "C.",
20 "Ronda",
21 "Pasaje",
22 "Cuesta",
23 "Pasadizo",
24 "Paseo",
25 "Camino",
26 )
27 postcode_formats = ("#####",)
28 states = (
29 "Álava",
30 "Albacete",
31 "Alicante",
32 "Almería",
33 "Asturias",
34 "Ávila",
35 "Badajoz",
36 "Baleares",
37 "Barcelona",
38 "Burgos",
39 "Cáceres",
40 "Cádiz",
41 "Cantabria",
42 "Castellón",
43 "Ceuta",
44 "Ciudad",
45 "Córdoba",
46 "Cuenca",
47 "Girona",
48 "Granada",
49 "Guadalajara",
50 "Guipúzcoa",
51 "Huelva",
52 "Huesca",
53 "Jaén",
54 "La Coruña",
55 "La Rioja",
56 "Las Palmas",
57 "León",
58 "Lleida",
59 "Lugo",
60 "Madrid",
61 "Málaga",
62 "Melilla",
63 "Murcia",
64 "Navarra",
65 "Ourense",
66 "Palencia",
67 "Pontevedra",
68 "Salamanca",
69 "Santa Cruz de Tenerife",
70 "Segovia",
71 "Sevilla",
72 "Soria",
73 "Tarragona",
74 "Teruel",
75 "Toledo",
76 "Valencia",
77 "Valladolid",
78 "Vizcaya",
79 "Zamora",
80 "Zaragoza",
81 )
82
83 # Source:
84 # https://administracionelectronica.gob.es/ctt/resources/Soluciones
85 # /238/Descargas/Catalogo-de-Comunidades-Autonomas.xlsx
86 regions = (
87 "Andalucía",
88 "Aragón",
89 "Principado de Asturias",
90 "Illes Balears",
91 "Canarias",
92 "Cantabria",
93 "Castilla y León",
94 "Castilla-La Mancha",
95 "Cataluña",
96 "Comunitat Valenciana",
97 "Extremadura",
98 "Galicia",
99 "Comunidad de Madrid",
100 "Región de Murcia",
101 "Comunidad Foral de Navarra",
102 "País Vasco",
103 "La Rioja",
104 "Ciudad Autónoma de Ceuta",
105 "Ciudad Autónoma de Melilla",
106 )
107
108 city_formats = ("{{state_name}}",)
109
110 street_name_formats = (
111 "{{street_prefix}} {{first_name}} {{last_name}}",
112 "{{street_prefix}} de {{first_name}} {{last_name}}",
113 )
114 street_address_formats = (
115 "{{street_name}} {{building_number}}",
116 "{{street_name}} {{building_number}} {{secondary_address}} ",
117 )
118 address_formats = ("{{street_address}}\n{{city}}, {{postcode}}",)
119 secondary_address_formats = ("Apt. ##", "Piso #", "Puerta #")
120
121 def state_name(self) -> str:
122 return self.random_element(self.states)
123
124 def street_prefix(self) -> str:
125 return self.random_element(self.street_prefixes)
126
127 def secondary_address(self) -> str:
128 return self.numerify(self.random_element(self.secondary_address_formats))
129
130 def administrative_unit(self) -> str:
131 return self.random_element(self.states)
132
133 state = administrative_unit
134
135 def region(self) -> str:
136 return self.random_element(self.regions)
137
138 autonomous_community = region
139
[end of faker/providers/address/es_ES/__init__.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/faker/providers/address/es_ES/__init__.py b/faker/providers/address/es_ES/__init__.py
--- a/faker/providers/address/es_ES/__init__.py
+++ b/faker/providers/address/es_ES/__init__.py
@@ -24,7 +24,6 @@
"Paseo",
"Camino",
)
- postcode_formats = ("#####",)
states = (
"Álava",
"Albacete",
@@ -135,4 +134,7 @@
def region(self) -> str:
return self.random_element(self.regions)
+ def postcode(self) -> str:
+ return str(self.generator.random.randint(1000, 52100)).zfill(5)
+
autonomous_community = region
|
{"golden_diff": "diff --git a/faker/providers/address/es_ES/__init__.py b/faker/providers/address/es_ES/__init__.py\n--- a/faker/providers/address/es_ES/__init__.py\n+++ b/faker/providers/address/es_ES/__init__.py\n@@ -24,7 +24,6 @@\n \"Paseo\",\n \"Camino\",\n )\n- postcode_formats = (\"#####\",)\n states = (\n \"\u00c1lava\",\n \"Albacete\",\n@@ -135,4 +134,7 @@\n def region(self) -> str:\n return self.random_element(self.regions)\n \n+ def postcode(self) -> str:\n+ return str(self.generator.random.randint(1000, 52100)).zfill(5)\n+\n autonomous_community = region\n", "issue": "es_ES postalcode is not generating valid codes\n* Faker version: 11.3\r\n* OS: Any\r\n\r\nWhen using postcode for es_ES and using it with a field that requires a valid Postal Code, it fails sometimes.\r\nI will assume that there is no logic with postal code generation for Spain.\r\n\r\n### Steps to reproduce\r\n\r\nGenerate postal codes\r\n\r\n### Expected behavior\r\n\r\nGet a valid Spain postal code\r\n\r\n### Actual behavior\r\n\r\nUnexpected. Many are wrong\r\n\r\n----\r\n\r\nI'll dig now into the code. Let's see if I can get some more information and fix it :thinking: Do not expect much from me\n", "before_files": [{"content": "from ..es import Provider as AddressProvider\n\n\nclass Provider(AddressProvider):\n building_number_formats = (\"%\", \"%#\", \"%#\", \"%#\", \"%##\")\n street_prefixes = (\n \"Plaza\",\n \"Calle\",\n \"Avenida\",\n \"Via\",\n \"Vial\",\n \"Rambla\",\n \"Glorieta\",\n \"Urbanizaci\u00f3n\",\n \"Callej\u00f3n\",\n \"Ca\u00f1ada\",\n \"Alameda\",\n \"Acceso\",\n \"C.\",\n \"Ronda\",\n \"Pasaje\",\n \"Cuesta\",\n \"Pasadizo\",\n \"Paseo\",\n \"Camino\",\n )\n postcode_formats = (\"#####\",)\n states = (\n \"\u00c1lava\",\n \"Albacete\",\n \"Alicante\",\n \"Almer\u00eda\",\n \"Asturias\",\n \"\u00c1vila\",\n \"Badajoz\",\n \"Baleares\",\n \"Barcelona\",\n \"Burgos\",\n \"C\u00e1ceres\",\n \"C\u00e1diz\",\n \"Cantabria\",\n \"Castell\u00f3n\",\n \"Ceuta\",\n \"Ciudad\",\n \"C\u00f3rdoba\",\n \"Cuenca\",\n \"Girona\",\n \"Granada\",\n \"Guadalajara\",\n \"Guip\u00fazcoa\",\n \"Huelva\",\n \"Huesca\",\n \"Ja\u00e9n\",\n \"La Coru\u00f1a\",\n \"La Rioja\",\n \"Las Palmas\",\n \"Le\u00f3n\",\n \"Lleida\",\n \"Lugo\",\n \"Madrid\",\n \"M\u00e1laga\",\n \"Melilla\",\n \"Murcia\",\n \"Navarra\",\n \"Ourense\",\n \"Palencia\",\n \"Pontevedra\",\n \"Salamanca\",\n \"Santa Cruz de Tenerife\",\n \"Segovia\",\n \"Sevilla\",\n \"Soria\",\n \"Tarragona\",\n \"Teruel\",\n \"Toledo\",\n \"Valencia\",\n \"Valladolid\",\n \"Vizcaya\",\n \"Zamora\",\n \"Zaragoza\",\n )\n\n # Source:\n # https://administracionelectronica.gob.es/ctt/resources/Soluciones\n # /238/Descargas/Catalogo-de-Comunidades-Autonomas.xlsx\n regions = (\n \"Andaluc\u00eda\",\n \"Arag\u00f3n\",\n \"Principado de Asturias\",\n \"Illes Balears\",\n \"Canarias\",\n \"Cantabria\",\n \"Castilla y Le\u00f3n\",\n \"Castilla-La Mancha\",\n \"Catalu\u00f1a\",\n \"Comunitat Valenciana\",\n \"Extremadura\",\n \"Galicia\",\n \"Comunidad de Madrid\",\n \"Regi\u00f3n de Murcia\",\n \"Comunidad Foral de Navarra\",\n \"Pa\u00eds Vasco\",\n \"La Rioja\",\n \"Ciudad Aut\u00f3noma de Ceuta\",\n \"Ciudad Aut\u00f3noma de Melilla\",\n )\n\n city_formats = (\"{{state_name}}\",)\n\n street_name_formats = (\n \"{{street_prefix}} {{first_name}} {{last_name}}\",\n \"{{street_prefix}} de {{first_name}} {{last_name}}\",\n )\n street_address_formats = (\n \"{{street_name}} {{building_number}}\",\n \"{{street_name}} {{building_number}} {{secondary_address}} \",\n )\n address_formats = (\"{{street_address}}\\n{{city}}, {{postcode}}\",)\n secondary_address_formats = (\"Apt. ##\", \"Piso #\", \"Puerta #\")\n\n def state_name(self) -> str:\n return self.random_element(self.states)\n\n def street_prefix(self) -> str:\n return self.random_element(self.street_prefixes)\n\n def secondary_address(self) -> str:\n return self.numerify(self.random_element(self.secondary_address_formats))\n\n def administrative_unit(self) -> str:\n return self.random_element(self.states)\n\n state = administrative_unit\n\n def region(self) -> str:\n return self.random_element(self.regions)\n\n autonomous_community = region\n", "path": "faker/providers/address/es_ES/__init__.py"}]}
| 1,878 | 176 |
gh_patches_debug_3386
|
rasdani/github-patches
|
git_diff
|
svthalia__concrexit-1550
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Event registrations api returns 500
### Describe the bug
`/api/v1/events/{pk}/registrations/` returns a 500
### How to reproduce
1. Register for an event
2. Go to `/api/v1/events/{pk}/registrations/`
3. See a 500
### Expected behaviour
No 500
</issue>
<code>
[start of website/events/api/v1/serializers/event_registrations/list.py]
1 from django.conf import settings
2 from django.templatetags.static import static
3 from rest_framework import serializers
4 from rest_framework.fields import empty
5
6 from events import services
7 from events.exceptions import RegistrationError
8 from events.models import EventRegistration, RegistrationInformationField
9 from payments.api.v1.fields import PaymentTypeField
10 from payments.models import Payment
11 from thaliawebsite.api.services import create_image_thumbnail_dict
12
13
14 class EventRegistrationListSerializer(serializers.ModelSerializer):
15 """Custom registration list serializer."""
16
17 class Meta:
18 model = EventRegistration
19 fields = ("pk", "member", "name", "avatar")
20
21 name = serializers.SerializerMethodField("_name")
22 avatar = serializers.SerializerMethodField("_avatar")
23 member = serializers.SerializerMethodField("_member")
24
25 def _member(self, instance):
26 if instance.member:
27 return instance.member.pk
28 return None
29
30 def _name(self, instance):
31 if instance.member:
32 return instance.member.profile.display_name()
33 return instance.name
34
35 def _avatar(self, instance):
36 placeholder = self.context["request"].build_absolute_uri(
37 static("members/images/default-avatar.jpg")
38 )
39 file = None
40 if instance.member and instance.member.profile.photo:
41 file = instance.member.profile.photo
42 return create_image_thumbnail_dict(
43 self.context["request"], file, placeholder=placeholder, size_large="800x800"
44 )
45
46
47 class EventRegistrationAdminListSerializer(EventRegistrationListSerializer):
48 """Custom registration admin list serializer."""
49
50 class Meta:
51 model = EventRegistration
52 fields = (
53 "pk",
54 "member",
55 "name",
56 "registered_on",
57 "is_cancelled",
58 "is_late_cancellation",
59 "queue_position",
60 "payment",
61 "present",
62 "avatar",
63 )
64
65 registered_on = serializers.DateTimeField(source="date")
66 is_cancelled = serializers.SerializerMethodField("_is_cancelled")
67 is_late_cancellation = serializers.SerializerMethodField("_is_late_cancellation")
68 queue_position = serializers.SerializerMethodField("_queue_position")
69 payment = PaymentTypeField(source="payment.type", choices=Payment.PAYMENT_TYPE)
70
71 def _is_late_cancellation(self, instance):
72 return instance.is_late_cancellation()
73
74 def _queue_position(self, instance):
75 pos = instance.queue_position
76 return pos if pos > 0 else None
77
78 def _is_cancelled(self, instance):
79 return instance.date_cancelled is not None
80
81 def _name(self, instance):
82 if instance.member:
83 return instance.member.get_full_name()
84 return instance.name
85
86
87 class EventRegistrationSerializer(serializers.ModelSerializer):
88 """Registration serializer."""
89
90 information_fields = None
91
92 class Meta:
93 model = EventRegistration
94 fields = (
95 "pk",
96 "member",
97 "name",
98 "photo",
99 "avatar",
100 "registered_on",
101 "is_late_cancellation",
102 "is_cancelled",
103 "queue_position",
104 "fields",
105 "payment",
106 "present",
107 )
108
109 name = serializers.SerializerMethodField("_name")
110 photo = serializers.SerializerMethodField("_photo")
111 avatar = serializers.SerializerMethodField("_avatar")
112 member = serializers.SerializerMethodField("_member")
113 payment = PaymentTypeField(source="payment.type", choices=Payment.PAYMENT_TYPE)
114 registered_on = serializers.DateTimeField(source="date", read_only=True)
115 is_cancelled = serializers.SerializerMethodField("_is_cancelled")
116 is_late_cancellation = serializers.SerializerMethodField("_is_late_cancellation")
117 fields = serializers.HiddenField(default="")
118
119 def _is_late_cancellation(self, instance):
120 val = instance.is_late_cancellation()
121 return False if val is None else val
122
123 def _is_cancelled(self, instance):
124 return instance.date_cancelled is not None
125
126 def _member(self, instance):
127 if instance.member:
128 return instance.member.pk
129 return None
130
131 def _name(self, instance):
132 if instance.member:
133 return instance.member.profile.display_name()
134 return instance.name
135
136 def _photo(self, instance):
137 if instance.member and instance.member.profile.photo:
138 return self.context["request"].build_absolute_uri(
139 f"{settings.MEDIA_URL}{instance.member.profile.photo}"
140 )
141 return self.context["request"].build_absolute_uri(
142 static("members/images/default-avatar.jpg")
143 )
144
145 def _avatar(self, instance):
146 placeholder = self.context["request"].build_absolute_uri(
147 static("members/images/default-avatar.jpg")
148 )
149 file = None
150 if instance.member and instance.member.profile.photo:
151 file = instance.member.profile.photo
152 return create_image_thumbnail_dict(
153 self.context["request"], file, placeholder=placeholder, size_large="800x800"
154 )
155
156 def __init__(self, instance=None, data=empty, **kwargs):
157 super().__init__(instance, data, **kwargs)
158 try:
159 if instance:
160 self.information_fields = services.registration_fields(
161 kwargs["context"]["request"], registration=instance
162 )
163 except RegistrationError:
164 pass
165
166 def get_fields(self):
167 fields = super().get_fields()
168
169 if self.information_fields:
170 for key, field in self.information_fields.items():
171 key = "fields[{}]".format(key)
172 field_type = field["type"]
173
174 if field_type == RegistrationInformationField.BOOLEAN_FIELD:
175 fields[key] = serializers.BooleanField(
176 required=False, write_only=True
177 )
178 elif field_type == RegistrationInformationField.INTEGER_FIELD:
179 fields[key] = serializers.IntegerField(
180 required=field["required"],
181 write_only=True,
182 allow_null=not field["required"],
183 )
184 elif field_type == RegistrationInformationField.TEXT_FIELD:
185 fields[key] = serializers.CharField(
186 required=field["required"],
187 write_only=True,
188 allow_blank=not field["required"],
189 allow_null=not field["required"],
190 )
191
192 fields[key].label = field["label"]
193 fields[key].help_text = field["description"]
194 fields[key].initial = field["value"]
195 fields[key].default = field["value"]
196
197 try:
198 if key in self.information_fields:
199 fields[key].initial = self.validated_data[key]
200 except AssertionError:
201 pass
202
203 return fields
204
205 def to_representation(self, instance):
206 data = super().to_representation(instance)
207 data["fields"] = self.information_fields
208 return data
209
210 def field_values(self):
211 return (
212 (name[7 : len(name) - 1], value)
213 for name, value in self.validated_data.items()
214 if "info_field" in name
215 )
216
[end of website/events/api/v1/serializers/event_registrations/list.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/website/events/api/v1/serializers/event_registrations/list.py b/website/events/api/v1/serializers/event_registrations/list.py
--- a/website/events/api/v1/serializers/event_registrations/list.py
+++ b/website/events/api/v1/serializers/event_registrations/list.py
@@ -73,7 +73,7 @@
def _queue_position(self, instance):
pos = instance.queue_position
- return pos if pos > 0 else None
+ return pos if pos and pos > 0 else None
def _is_cancelled(self, instance):
return instance.date_cancelled is not None
|
{"golden_diff": "diff --git a/website/events/api/v1/serializers/event_registrations/list.py b/website/events/api/v1/serializers/event_registrations/list.py\n--- a/website/events/api/v1/serializers/event_registrations/list.py\n+++ b/website/events/api/v1/serializers/event_registrations/list.py\n@@ -73,7 +73,7 @@\n \n def _queue_position(self, instance):\n pos = instance.queue_position\n- return pos if pos > 0 else None\n+ return pos if pos and pos > 0 else None\n \n def _is_cancelled(self, instance):\n return instance.date_cancelled is not None\n", "issue": "Event registrations api returns 500\n### Describe the bug\r\n`/api/v1/events/{pk}/registrations/` returns a 500\r\n\r\n### How to reproduce\r\n1. Register for an event\r\n2. Go to `/api/v1/events/{pk}/registrations/`\r\n3. See a 500\r\n\r\n### Expected behaviour\r\nNo 500\r\n\n", "before_files": [{"content": "from django.conf import settings\nfrom django.templatetags.static import static\nfrom rest_framework import serializers\nfrom rest_framework.fields import empty\n\nfrom events import services\nfrom events.exceptions import RegistrationError\nfrom events.models import EventRegistration, RegistrationInformationField\nfrom payments.api.v1.fields import PaymentTypeField\nfrom payments.models import Payment\nfrom thaliawebsite.api.services import create_image_thumbnail_dict\n\n\nclass EventRegistrationListSerializer(serializers.ModelSerializer):\n \"\"\"Custom registration list serializer.\"\"\"\n\n class Meta:\n model = EventRegistration\n fields = (\"pk\", \"member\", \"name\", \"avatar\")\n\n name = serializers.SerializerMethodField(\"_name\")\n avatar = serializers.SerializerMethodField(\"_avatar\")\n member = serializers.SerializerMethodField(\"_member\")\n\n def _member(self, instance):\n if instance.member:\n return instance.member.pk\n return None\n\n def _name(self, instance):\n if instance.member:\n return instance.member.profile.display_name()\n return instance.name\n\n def _avatar(self, instance):\n placeholder = self.context[\"request\"].build_absolute_uri(\n static(\"members/images/default-avatar.jpg\")\n )\n file = None\n if instance.member and instance.member.profile.photo:\n file = instance.member.profile.photo\n return create_image_thumbnail_dict(\n self.context[\"request\"], file, placeholder=placeholder, size_large=\"800x800\"\n )\n\n\nclass EventRegistrationAdminListSerializer(EventRegistrationListSerializer):\n \"\"\"Custom registration admin list serializer.\"\"\"\n\n class Meta:\n model = EventRegistration\n fields = (\n \"pk\",\n \"member\",\n \"name\",\n \"registered_on\",\n \"is_cancelled\",\n \"is_late_cancellation\",\n \"queue_position\",\n \"payment\",\n \"present\",\n \"avatar\",\n )\n\n registered_on = serializers.DateTimeField(source=\"date\")\n is_cancelled = serializers.SerializerMethodField(\"_is_cancelled\")\n is_late_cancellation = serializers.SerializerMethodField(\"_is_late_cancellation\")\n queue_position = serializers.SerializerMethodField(\"_queue_position\")\n payment = PaymentTypeField(source=\"payment.type\", choices=Payment.PAYMENT_TYPE)\n\n def _is_late_cancellation(self, instance):\n return instance.is_late_cancellation()\n\n def _queue_position(self, instance):\n pos = instance.queue_position\n return pos if pos > 0 else None\n\n def _is_cancelled(self, instance):\n return instance.date_cancelled is not None\n\n def _name(self, instance):\n if instance.member:\n return instance.member.get_full_name()\n return instance.name\n\n\nclass EventRegistrationSerializer(serializers.ModelSerializer):\n \"\"\"Registration serializer.\"\"\"\n\n information_fields = None\n\n class Meta:\n model = EventRegistration\n fields = (\n \"pk\",\n \"member\",\n \"name\",\n \"photo\",\n \"avatar\",\n \"registered_on\",\n \"is_late_cancellation\",\n \"is_cancelled\",\n \"queue_position\",\n \"fields\",\n \"payment\",\n \"present\",\n )\n\n name = serializers.SerializerMethodField(\"_name\")\n photo = serializers.SerializerMethodField(\"_photo\")\n avatar = serializers.SerializerMethodField(\"_avatar\")\n member = serializers.SerializerMethodField(\"_member\")\n payment = PaymentTypeField(source=\"payment.type\", choices=Payment.PAYMENT_TYPE)\n registered_on = serializers.DateTimeField(source=\"date\", read_only=True)\n is_cancelled = serializers.SerializerMethodField(\"_is_cancelled\")\n is_late_cancellation = serializers.SerializerMethodField(\"_is_late_cancellation\")\n fields = serializers.HiddenField(default=\"\")\n\n def _is_late_cancellation(self, instance):\n val = instance.is_late_cancellation()\n return False if val is None else val\n\n def _is_cancelled(self, instance):\n return instance.date_cancelled is not None\n\n def _member(self, instance):\n if instance.member:\n return instance.member.pk\n return None\n\n def _name(self, instance):\n if instance.member:\n return instance.member.profile.display_name()\n return instance.name\n\n def _photo(self, instance):\n if instance.member and instance.member.profile.photo:\n return self.context[\"request\"].build_absolute_uri(\n f\"{settings.MEDIA_URL}{instance.member.profile.photo}\"\n )\n return self.context[\"request\"].build_absolute_uri(\n static(\"members/images/default-avatar.jpg\")\n )\n\n def _avatar(self, instance):\n placeholder = self.context[\"request\"].build_absolute_uri(\n static(\"members/images/default-avatar.jpg\")\n )\n file = None\n if instance.member and instance.member.profile.photo:\n file = instance.member.profile.photo\n return create_image_thumbnail_dict(\n self.context[\"request\"], file, placeholder=placeholder, size_large=\"800x800\"\n )\n\n def __init__(self, instance=None, data=empty, **kwargs):\n super().__init__(instance, data, **kwargs)\n try:\n if instance:\n self.information_fields = services.registration_fields(\n kwargs[\"context\"][\"request\"], registration=instance\n )\n except RegistrationError:\n pass\n\n def get_fields(self):\n fields = super().get_fields()\n\n if self.information_fields:\n for key, field in self.information_fields.items():\n key = \"fields[{}]\".format(key)\n field_type = field[\"type\"]\n\n if field_type == RegistrationInformationField.BOOLEAN_FIELD:\n fields[key] = serializers.BooleanField(\n required=False, write_only=True\n )\n elif field_type == RegistrationInformationField.INTEGER_FIELD:\n fields[key] = serializers.IntegerField(\n required=field[\"required\"],\n write_only=True,\n allow_null=not field[\"required\"],\n )\n elif field_type == RegistrationInformationField.TEXT_FIELD:\n fields[key] = serializers.CharField(\n required=field[\"required\"],\n write_only=True,\n allow_blank=not field[\"required\"],\n allow_null=not field[\"required\"],\n )\n\n fields[key].label = field[\"label\"]\n fields[key].help_text = field[\"description\"]\n fields[key].initial = field[\"value\"]\n fields[key].default = field[\"value\"]\n\n try:\n if key in self.information_fields:\n fields[key].initial = self.validated_data[key]\n except AssertionError:\n pass\n\n return fields\n\n def to_representation(self, instance):\n data = super().to_representation(instance)\n data[\"fields\"] = self.information_fields\n return data\n\n def field_values(self):\n return (\n (name[7 : len(name) - 1], value)\n for name, value in self.validated_data.items()\n if \"info_field\" in name\n )\n", "path": "website/events/api/v1/serializers/event_registrations/list.py"}]}
| 2,601 | 147 |
gh_patches_debug_23070
|
rasdani/github-patches
|
git_diff
|
pwndbg__pwndbg-1709
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
bugreport command: show more info
The bugreport command should display more information like:
* show osabi
* show architecture
* show endian
Depending on current arch -- note that those are only available if given arch is supported by current GDB, like `gdb-multiarch`:
* show arm
* show mips
Maybe also:
* show charset
* show width
* show height
</issue>
<code>
[start of pwndbg/commands/version.py]
1 """
2 Displays gdb, python and pwndbg versions.
3 """
4
5
6 import argparse
7 import os
8 import platform
9 import re
10 import sys
11 from subprocess import check_call
12 from subprocess import check_output
13 from tempfile import NamedTemporaryFile
14 from urllib.parse import quote
15
16 import gdb
17
18 import pwndbg
19 import pwndbg.commands
20 import pwndbg.ida
21 from pwndbg.color import message
22 from pwndbg.commands import CommandCategory
23
24
25 def _gdb_version():
26 try:
27 return gdb.VERSION # GDB >= 8.1 (or earlier?)
28 except AttributeError:
29 return gdb.execute("show version", to_string=True).split("\n")[0]
30
31
32 def _py_version():
33 return sys.version.replace("\n", " ")
34
35
36 def capstone_version():
37 try:
38 import capstone
39
40 return ".".join(map(str, capstone.cs_version()))
41 except ImportError:
42 return "not found"
43
44
45 def unicorn_version():
46 try:
47 import unicorn
48
49 return unicorn.__version__
50 except ImportError:
51 return "not found"
52
53
54 def all_versions():
55 gdb_str = "Gdb: %s" % _gdb_version()
56 py_str = "Python: %s" % _py_version()
57 pwndbg_str = "Pwndbg: %s" % pwndbg.__version__
58
59 capstone_str = "Capstone: %s" % capstone_version()
60 unicorn_str = "Unicorn: %s" % unicorn_version()
61
62 all_versions = (gdb_str, py_str, pwndbg_str, capstone_str, unicorn_str)
63
64 ida_versions = pwndbg.ida.get_ida_versions()
65
66 if ida_versions is not None:
67 ida_version = "IDA PRO: %s" % ida_versions["ida"]
68 ida_py_ver = "IDA Py: %s" % ida_versions["python"]
69 ida_hr_ver = "Hexrays: %s" % ida_versions["hexrays"]
70 all_versions += (ida_version, ida_py_ver, ida_hr_ver)
71 return all_versions
72
73
74 @pwndbg.commands.ArgparsedCommand(
75 "Displays GDB, Python, and pwndbg versions.", category=CommandCategory.PWNDBG
76 )
77 def version() -> None:
78 """
79 Displays GDB, Python, and pwndbg versions.
80 """
81 print("\n".join(map(message.system, all_versions())))
82
83
84 bugreport_parser = argparse.ArgumentParser(description="Generate a bug report.")
85 bugreport_group = bugreport_parser.add_mutually_exclusive_group()
86 bugreport_group.add_argument(
87 "--run-browser", "-b", action="store_true", help="Open browser on github/issues/new"
88 )
89 bugreport_group.add_argument(
90 "--use-gh", "-g", action="store_true", help="Create issue using Github CLI"
91 )
92
93
94 @pwndbg.commands.ArgparsedCommand(bugreport_parser, category=CommandCategory.PWNDBG)
95 def bugreport(run_browser=False, use_gh=False):
96 ISSUE_TEMPLATE = """
97 <!--
98 Before reporting a new issue, make sure that we do not have any duplicates already open.
99 If there is one it might be good to take part in the discussion there.
100
101 Please make sure you have checked that the issue persists on LATEST pwndbg version.
102
103 Below is a template for BUG REPORTS.
104 Don't include it if this is a FEATURE REQUEST.
105 -->
106
107
108 ### Description
109
110 <!--
111 Briefly describe the problem you are having in a few paragraphs.
112 -->
113
114 ### Steps to reproduce
115
116 <!--
117 What do we have to do to reproduce the problem?
118 If this is connected to particular C/asm code or a binary,
119 please provide the binary or if possible, a smallest C code that reproduces the issue.
120 -->
121
122 Gdb session history:
123 ```
124 {gdb_history}
125 ```
126
127 ### My setup
128
129 <!--
130 Show us your gdb/python/pwndbg/OS/IDA Pro version (depending on your case).
131
132 NOTE: We are currently testing Pwndbg only on Ubuntu installations but it should work fine on other distros as well.
133
134 This can be displayed in pwndbg through `version` command.
135
136 If it is somehow unavailable, use:
137 * `show version` - for gdb
138 * `py import sys; print(sys.version)` - for python
139 * pwndbg version/git commit id
140 -->
141
142 ```
143 {setup}
144 ```"""
145
146 gdb_config = gdb.execute("show configuration", to_string=True).split("\n")
147 all_info = all_versions()
148 os_info = platform.system()
149
150 current_setup = "Platform: %s\n" % platform.platform()
151
152 if os_info.lower() == "linux" and os.path.isfile("/etc/os-release"):
153 with open("/etc/os-release", "r") as os_release:
154 contents = os_release.read()
155 match = re.search('PRETTY_NAME="?([^",\n]+)', contents)
156 if match:
157 os_info = match.group(1)
158
159 current_setup += "OS: %s\n" % os_info
160 current_setup += "\n".join(all_info)
161 current_setup += "\n" + "\n".join(gdb_config)
162
163 # get saved history size (not including current gdb session)
164 gdb_history_file = gdb.execute("show history filename", to_string=True)
165 gdb_history_file = gdb_history_file[
166 gdb_history_file.index('"') + 1 : gdb_history_file.rindex('"')
167 ]
168 gdb_history_len = 0
169 try:
170 with open(gdb_history_file, "r") as f:
171 gdb_history_len = len(f.readlines())
172 except FileNotFoundError:
173 pass
174
175 max_command_no = 0
176 history_commands = gdb.execute("show commands", to_string=True)
177 if history_commands:
178 history_commands = history_commands.split("\n")
179 if len(history_commands) > 1:
180 # The last element of the list is the `show commands` command we
181 # just ran, so we need to get the second to last one
182 last_command = history_commands[-2]
183 max_command_no = int(last_command.split()[0]) - 1
184
185 show_command_size = 10 # 'show command' returns 10 commands
186 gdb_current_session_history = {}
187 current_command_no = gdb_history_len + 1
188
189 while current_command_no <= max_command_no:
190 cmds = gdb.execute(
191 "show commands " + str(current_command_no + (show_command_size // 2) + 1),
192 to_string=True,
193 ).split("\n")[:-1]
194 for cmd in cmds:
195 cmd_no, cmd = cmd.split(maxsplit=1)
196 cmd_no = int(cmd_no)
197 if cmd_no <= gdb_history_len:
198 continue
199 if current_command_no > max_command_no:
200 break
201 gdb_current_session_history[cmd_no] = cmd
202 current_command_no += 1
203
204 gdb_current_session_history = (v for (k, v) in sorted(gdb_current_session_history.items()))
205 gdb_current_session_history = "\n".join(gdb_current_session_history)
206
207 issue_bugreport = ISSUE_TEMPLATE.format(
208 gdb_history=gdb_current_session_history, setup=current_setup
209 )
210 print(issue_bugreport)
211
212 please_please_submit = "Please submit the bugreport generated above at "
213 github_issue_url = "https://github.com/pwndbg/pwndbg/issues/new"
214 github_issue_body = "?body=" + quote(issue_bugreport)
215
216 if use_gh:
217 try:
218 with NamedTemporaryFile("w", delete=True) as f:
219 f.write(issue_bugreport)
220 f.flush()
221 check_call([os.environ.get("EDITOR", "vi"), f.name])
222 check_call(["gh", "issue", "create", "--body-file", f.name])
223 except Exception:
224 print(please_please_submit + github_issue_url)
225 raise
226 elif run_browser:
227 try:
228 check_output(["xdg-open", github_issue_url + github_issue_body])
229 except Exception:
230 print(please_please_submit + github_issue_url)
231 else:
232 print(please_please_submit + github_issue_url)
233
[end of pwndbg/commands/version.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/pwndbg/commands/version.py b/pwndbg/commands/version.py
--- a/pwndbg/commands/version.py
+++ b/pwndbg/commands/version.py
@@ -157,6 +157,40 @@
os_info = match.group(1)
current_setup += "OS: %s\n" % os_info
+
+ # 1. showing osabi
+ osabi_info = platform.uname().version
+ current_setup += "OS ABI: %s\n" % osabi_info
+
+ # 2. showing architecture
+ arch_info = platform.machine()
+ current_setup += "Architecture: %s\n" % arch_info
+
+ # 3. showing endian
+ endian_info = sys.byteorder
+ current_setup += "Endian: %s\n" % endian_info
+
+ # 4. Depending on current arch -- note that those are only available if given arch is supported by current GDB, like gdb-multiarch
+ if arch_info in ["armv7l", "aarch64"]:
+ arm_info = gdb.execute("show arm", to_string=True)
+ current_setup += "ARM: %s\n" % arm_info
+
+ elif arch_info in ["mips", "mips64"]:
+ mips_info = gdb.execute("show mips", to_string=True)
+ current_setup += "MIPS: %s\n" % mips_info
+
+ # 7. showing charset
+ charset_info = sys.getdefaultencoding()
+ current_setup += "Charset: %s\n" % charset_info
+
+ # 8. showing width
+ width_info = os.get_terminal_size().columns
+ current_setup += "Width: %s\n" % width_info
+
+ # 9. showing height
+ height_info = os.get_terminal_size().lines
+ current_setup += "Height: %s\n" % height_info
+
current_setup += "\n".join(all_info)
current_setup += "\n" + "\n".join(gdb_config)
|
{"golden_diff": "diff --git a/pwndbg/commands/version.py b/pwndbg/commands/version.py\n--- a/pwndbg/commands/version.py\n+++ b/pwndbg/commands/version.py\n@@ -157,6 +157,40 @@\n os_info = match.group(1)\n \n current_setup += \"OS: %s\\n\" % os_info\n+\n+ # 1. showing osabi\n+ osabi_info = platform.uname().version\n+ current_setup += \"OS ABI: %s\\n\" % osabi_info\n+\n+ # 2. showing architecture\n+ arch_info = platform.machine()\n+ current_setup += \"Architecture: %s\\n\" % arch_info\n+\n+ # 3. showing endian\n+ endian_info = sys.byteorder\n+ current_setup += \"Endian: %s\\n\" % endian_info\n+\n+ # 4. Depending on current arch -- note that those are only available if given arch is supported by current GDB, like gdb-multiarch\n+ if arch_info in [\"armv7l\", \"aarch64\"]:\n+ arm_info = gdb.execute(\"show arm\", to_string=True)\n+ current_setup += \"ARM: %s\\n\" % arm_info\n+\n+ elif arch_info in [\"mips\", \"mips64\"]:\n+ mips_info = gdb.execute(\"show mips\", to_string=True)\n+ current_setup += \"MIPS: %s\\n\" % mips_info\n+\n+ # 7. showing charset\n+ charset_info = sys.getdefaultencoding()\n+ current_setup += \"Charset: %s\\n\" % charset_info\n+\n+ # 8. showing width\n+ width_info = os.get_terminal_size().columns\n+ current_setup += \"Width: %s\\n\" % width_info\n+\n+ # 9. showing height\n+ height_info = os.get_terminal_size().lines\n+ current_setup += \"Height: %s\\n\" % height_info\n+\n current_setup += \"\\n\".join(all_info)\n current_setup += \"\\n\" + \"\\n\".join(gdb_config)\n", "issue": "bugreport command: show more info\nThe bugreport command should display more information like:\r\n* show osabi\r\n* show architecture\r\n* show endian\r\n\r\nDepending on current arch -- note that those are only available if given arch is supported by current GDB, like `gdb-multiarch`:\r\n* show arm\r\n* show mips\r\n\r\nMaybe also:\r\n* show charset\r\n* show width\r\n* show height\r\n\r\n\n", "before_files": [{"content": "\"\"\"\nDisplays gdb, python and pwndbg versions.\n\"\"\"\n\n\nimport argparse\nimport os\nimport platform\nimport re\nimport sys\nfrom subprocess import check_call\nfrom subprocess import check_output\nfrom tempfile import NamedTemporaryFile\nfrom urllib.parse import quote\n\nimport gdb\n\nimport pwndbg\nimport pwndbg.commands\nimport pwndbg.ida\nfrom pwndbg.color import message\nfrom pwndbg.commands import CommandCategory\n\n\ndef _gdb_version():\n try:\n return gdb.VERSION # GDB >= 8.1 (or earlier?)\n except AttributeError:\n return gdb.execute(\"show version\", to_string=True).split(\"\\n\")[0]\n\n\ndef _py_version():\n return sys.version.replace(\"\\n\", \" \")\n\n\ndef capstone_version():\n try:\n import capstone\n\n return \".\".join(map(str, capstone.cs_version()))\n except ImportError:\n return \"not found\"\n\n\ndef unicorn_version():\n try:\n import unicorn\n\n return unicorn.__version__\n except ImportError:\n return \"not found\"\n\n\ndef all_versions():\n gdb_str = \"Gdb: %s\" % _gdb_version()\n py_str = \"Python: %s\" % _py_version()\n pwndbg_str = \"Pwndbg: %s\" % pwndbg.__version__\n\n capstone_str = \"Capstone: %s\" % capstone_version()\n unicorn_str = \"Unicorn: %s\" % unicorn_version()\n\n all_versions = (gdb_str, py_str, pwndbg_str, capstone_str, unicorn_str)\n\n ida_versions = pwndbg.ida.get_ida_versions()\n\n if ida_versions is not None:\n ida_version = \"IDA PRO: %s\" % ida_versions[\"ida\"]\n ida_py_ver = \"IDA Py: %s\" % ida_versions[\"python\"]\n ida_hr_ver = \"Hexrays: %s\" % ida_versions[\"hexrays\"]\n all_versions += (ida_version, ida_py_ver, ida_hr_ver)\n return all_versions\n\n\[email protected](\n \"Displays GDB, Python, and pwndbg versions.\", category=CommandCategory.PWNDBG\n)\ndef version() -> None:\n \"\"\"\n Displays GDB, Python, and pwndbg versions.\n \"\"\"\n print(\"\\n\".join(map(message.system, all_versions())))\n\n\nbugreport_parser = argparse.ArgumentParser(description=\"Generate a bug report.\")\nbugreport_group = bugreport_parser.add_mutually_exclusive_group()\nbugreport_group.add_argument(\n \"--run-browser\", \"-b\", action=\"store_true\", help=\"Open browser on github/issues/new\"\n)\nbugreport_group.add_argument(\n \"--use-gh\", \"-g\", action=\"store_true\", help=\"Create issue using Github CLI\"\n)\n\n\[email protected](bugreport_parser, category=CommandCategory.PWNDBG)\ndef bugreport(run_browser=False, use_gh=False):\n ISSUE_TEMPLATE = \"\"\"\n<!--\nBefore reporting a new issue, make sure that we do not have any duplicates already open.\nIf there is one it might be good to take part in the discussion there.\n\nPlease make sure you have checked that the issue persists on LATEST pwndbg version.\n\nBelow is a template for BUG REPORTS.\nDon't include it if this is a FEATURE REQUEST.\n-->\n\n\n### Description\n\n<!--\nBriefly describe the problem you are having in a few paragraphs.\n-->\n\n### Steps to reproduce\n\n<!--\nWhat do we have to do to reproduce the problem?\nIf this is connected to particular C/asm code or a binary,\nplease provide the binary or if possible, a smallest C code that reproduces the issue.\n-->\n\nGdb session history:\n```\n{gdb_history}\n```\n\n### My setup\n\n<!--\nShow us your gdb/python/pwndbg/OS/IDA Pro version (depending on your case).\n\nNOTE: We are currently testing Pwndbg only on Ubuntu installations but it should work fine on other distros as well.\n\nThis can be displayed in pwndbg through `version` command.\n\nIf it is somehow unavailable, use:\n* `show version` - for gdb\n* `py import sys; print(sys.version)` - for python\n* pwndbg version/git commit id\n-->\n\n```\n{setup}\n```\"\"\"\n\n gdb_config = gdb.execute(\"show configuration\", to_string=True).split(\"\\n\")\n all_info = all_versions()\n os_info = platform.system()\n\n current_setup = \"Platform: %s\\n\" % platform.platform()\n\n if os_info.lower() == \"linux\" and os.path.isfile(\"/etc/os-release\"):\n with open(\"/etc/os-release\", \"r\") as os_release:\n contents = os_release.read()\n match = re.search('PRETTY_NAME=\"?([^\",\\n]+)', contents)\n if match:\n os_info = match.group(1)\n\n current_setup += \"OS: %s\\n\" % os_info\n current_setup += \"\\n\".join(all_info)\n current_setup += \"\\n\" + \"\\n\".join(gdb_config)\n\n # get saved history size (not including current gdb session)\n gdb_history_file = gdb.execute(\"show history filename\", to_string=True)\n gdb_history_file = gdb_history_file[\n gdb_history_file.index('\"') + 1 : gdb_history_file.rindex('\"')\n ]\n gdb_history_len = 0\n try:\n with open(gdb_history_file, \"r\") as f:\n gdb_history_len = len(f.readlines())\n except FileNotFoundError:\n pass\n\n max_command_no = 0\n history_commands = gdb.execute(\"show commands\", to_string=True)\n if history_commands:\n history_commands = history_commands.split(\"\\n\")\n if len(history_commands) > 1:\n # The last element of the list is the `show commands` command we\n # just ran, so we need to get the second to last one\n last_command = history_commands[-2]\n max_command_no = int(last_command.split()[0]) - 1\n\n show_command_size = 10 # 'show command' returns 10 commands\n gdb_current_session_history = {}\n current_command_no = gdb_history_len + 1\n\n while current_command_no <= max_command_no:\n cmds = gdb.execute(\n \"show commands \" + str(current_command_no + (show_command_size // 2) + 1),\n to_string=True,\n ).split(\"\\n\")[:-1]\n for cmd in cmds:\n cmd_no, cmd = cmd.split(maxsplit=1)\n cmd_no = int(cmd_no)\n if cmd_no <= gdb_history_len:\n continue\n if current_command_no > max_command_no:\n break\n gdb_current_session_history[cmd_no] = cmd\n current_command_no += 1\n\n gdb_current_session_history = (v for (k, v) in sorted(gdb_current_session_history.items()))\n gdb_current_session_history = \"\\n\".join(gdb_current_session_history)\n\n issue_bugreport = ISSUE_TEMPLATE.format(\n gdb_history=gdb_current_session_history, setup=current_setup\n )\n print(issue_bugreport)\n\n please_please_submit = \"Please submit the bugreport generated above at \"\n github_issue_url = \"https://github.com/pwndbg/pwndbg/issues/new\"\n github_issue_body = \"?body=\" + quote(issue_bugreport)\n\n if use_gh:\n try:\n with NamedTemporaryFile(\"w\", delete=True) as f:\n f.write(issue_bugreport)\n f.flush()\n check_call([os.environ.get(\"EDITOR\", \"vi\"), f.name])\n check_call([\"gh\", \"issue\", \"create\", \"--body-file\", f.name])\n except Exception:\n print(please_please_submit + github_issue_url)\n raise\n elif run_browser:\n try:\n check_output([\"xdg-open\", github_issue_url + github_issue_body])\n except Exception:\n print(please_please_submit + github_issue_url)\n else:\n print(please_please_submit + github_issue_url)\n", "path": "pwndbg/commands/version.py"}]}
| 2,960 | 464 |
gh_patches_debug_28461
|
rasdani/github-patches
|
git_diff
|
elastic__apm-agent-python-708
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
database trace show twice
maybe this is not a bug
kibana show every database trace will show twice

but in 4.2.2 version, it show this

i think its related to span? in 4.2.2 agent won't upload span log, please tell me how to resovle it 😢
(forgive my poor English)
**Environment (please complete the following information)**
- OS: [e.g. Linux] linux
- Python version: 3.7.
- Framework and version [e.g. Django 2.1]: Django 3.0
- APM Server version: 7.4.2
- Agent version: 5.3.3
</issue>
<code>
[start of elasticapm/instrumentation/packages/base.py]
1 # BSD 3-Clause License
2 #
3 # Copyright (c) 2019, Elasticsearch BV
4 # All rights reserved.
5 #
6 # Redistribution and use in source and binary forms, with or without
7 # modification, are permitted provided that the following conditions are met:
8 #
9 # * Redistributions of source code must retain the above copyright notice, this
10 # list of conditions and the following disclaimer.
11 #
12 # * Redistributions in binary form must reproduce the above copyright notice,
13 # this list of conditions and the following disclaimer in the documentation
14 # and/or other materials provided with the distribution.
15 #
16 # * Neither the name of the copyright holder nor the names of its
17 # contributors may be used to endorse or promote products derived from
18 # this software without specific prior written permission.
19 #
20 # THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
21 # AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
22 # IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
23 # DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE
24 # FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
25 # DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
26 # SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER
27 # CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
28 # OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
29 # OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
30
31 import functools
32 import os
33
34 from elasticapm.traces import execution_context
35 from elasticapm.utils import wrapt
36 from elasticapm.utils.logging import get_logger
37
38 logger = get_logger("elasticapm.instrument")
39
40
41 class AbstractInstrumentedModule(object):
42 """
43 This class is designed to reduce the amount of code required to
44 instrument library functions using wrapt.
45
46 Instrumentation modules inherit from this class and override pieces as
47 needed. Only `name`, `instrumented_list`, and `call` are required in
48 the inheriting class.
49
50 The `instrument_list` is a list of (module, method) pairs that will be
51 instrumented. The module/method need not be imported -- in fact, because
52 instrumentation modules are all processed during the instrumentation
53 process, lazy imports should be used in order to avoid ImportError
54 exceptions.
55
56 The `instrument()` method will be called for each InstrumentedModule
57 listed in the instrument register (elasticapm.instrumentation.register),
58 and each method in the `instrument_list` will be wrapped (using wrapt)
59 with the `call_if_sampling()` function, which (by default) will either
60 call the wrapped function by itself, or pass it into `call()` to be
61 called if there is a transaction active.
62
63 For simple span-wrapping of instrumented libraries, a very simple
64 InstrumentedModule might look like this::
65
66 from elasticapm.instrumentation.packages.base import AbstractInstrumentedModule
67 from elasticapm.traces import capture_span
68
69 class Jinja2Instrumentation(AbstractInstrumentedModule):
70 name = "jinja2"
71 instrument_list = [("jinja2", "Template.render")]
72 def call(self, module, method, wrapped, instance, args, kwargs):
73 signature = instance.name or instance.filename
74 with capture_span(signature, span_type="template", span_subtype="jinja2", span_action="render"):
75 return wrapped(*args, **kwargs)
76
77 This class can also be used to instrument callables which are expected to
78 create their own transactions (rather than spans within a transaction).
79 In this case, set `creates_transaction = True` next to your `name` and
80 `instrument_list`. This tells the instrumenting code to always wrap the
81 method with `call()`, even if there is no transaction active. It is
82 expected in this case that a new transaction will be created as part of
83 your `call()` method.
84 """
85
86 name = None
87 mutates_unsampled_arguments = False
88 creates_transactions = False
89
90 instrument_list = [
91 # List of (module, method) pairs to instrument. E.g.:
92 # ("requests.sessions", "Session.send"),
93 ]
94
95 def __init__(self):
96 self.originals = {}
97 self.instrumented = False
98
99 assert self.name is not None
100
101 def get_wrapped_name(self, wrapped, instance, fallback_method=None):
102 wrapped_name = []
103 if hasattr(instance, "__class__") and hasattr(instance.__class__, "__name__"):
104 wrapped_name.append(instance.__class__.__name__)
105
106 if hasattr(wrapped, "__name__"):
107 wrapped_name.append(wrapped.__name__)
108 elif fallback_method:
109 attribute = fallback_method.split(".")
110 if len(attribute) == 2:
111 wrapped_name.append(attribute[1])
112
113 return ".".join(wrapped_name)
114
115 def get_instrument_list(self):
116 return self.instrument_list
117
118 def instrument(self):
119 if self.instrumented:
120 return
121
122 skip_env_var = "SKIP_INSTRUMENT_" + str(self.name.upper())
123 if skip_env_var in os.environ:
124 logger.debug("Skipping instrumentation of %s. %s is set.", self.name, skip_env_var)
125 return
126 try:
127 instrument_list = self.get_instrument_list()
128 skipped_modules = set()
129 instrumented_methods = []
130
131 for module, method in instrument_list:
132 try:
133 # Skip modules we already failed to load
134 if module in skipped_modules:
135 continue
136 # We jump through hoop here to get the original
137 # `module`/`method` in the call to `call_if_sampling`
138 parent, attribute, original = wrapt.resolve_path(module, method)
139 self.originals[(module, method)] = original
140 wrapper = wrapt.FunctionWrapper(original, functools.partial(self.call_if_sampling, module, method))
141 wrapt.apply_patch(parent, attribute, wrapper)
142 instrumented_methods.append((module, method))
143 except ImportError:
144 # Could not import module
145 logger.debug("Skipping instrumentation of %s. Module %s not found", self.name, module)
146
147 # Keep track of modules we couldn't load so we don't
148 # try to instrument anything in that module again
149 skipped_modules.add(module)
150 except AttributeError as ex:
151 # Could not find thing in module
152 logger.debug("Skipping instrumentation of %s.%s: %s", module, method, ex)
153 if instrumented_methods:
154 logger.debug("Instrumented %s, %s", self.name, ", ".join(".".join(m) for m in instrumented_methods))
155
156 except ImportError as ex:
157 logger.debug("Skipping instrumentation of %s. %s", self.name, ex)
158 self.instrumented = True
159
160 def uninstrument(self):
161 if not self.instrumented or not self.originals:
162 return
163 uninstrumented_methods = []
164 for module, method in self.get_instrument_list():
165 if (module, method) in self.originals:
166 parent, attribute, wrapper = wrapt.resolve_path(module, method)
167 wrapt.apply_patch(parent, attribute, self.originals[(module, method)])
168 uninstrumented_methods.append((module, method))
169 if uninstrumented_methods:
170 logger.debug("Uninstrumented %s, %s", self.name, ", ".join(".".join(m) for m in uninstrumented_methods))
171 self.instrumented = False
172 self.originals = {}
173
174 def call_if_sampling(self, module, method, wrapped, instance, args, kwargs):
175 """
176 This is the function which will wrap the instrumented method/function.
177
178 By default, will call the instrumented method/function, via `call()`,
179 only if a transaction is active and sampled. This behavior can be
180 overridden by setting `creates_transactions = True` at the class
181 level.
182
183 If `creates_transactions == False` and there's an active transaction
184 with `transaction.is_sampled == False`, then the
185 `mutate_unsampled_call_args()` method is called, and the resulting
186 args and kwargs are passed into the wrapped function directly, not
187 via `call()`. This can e.g. be used to add traceparent headers to the
188 underlying http call for HTTP instrumentations, even if we're not
189 sampling the transaction.
190 """
191 if self.creates_transactions:
192 return self.call(module, method, wrapped, instance, args, kwargs)
193 transaction = execution_context.get_transaction()
194 if not transaction:
195 return wrapped(*args, **kwargs)
196 elif not transaction.is_sampled:
197 args, kwargs = self.mutate_unsampled_call_args(module, method, wrapped, instance, args, kwargs, transaction)
198 return wrapped(*args, **kwargs)
199 else:
200 return self.call(module, method, wrapped, instance, args, kwargs)
201
202 def mutate_unsampled_call_args(self, module, method, wrapped, instance, args, kwargs, transaction):
203 """
204 Method called for unsampled wrapped calls. This can e.g. be used to
205 add traceparent headers to the underlying http call for HTTP
206 instrumentations.
207
208 :param module:
209 :param method:
210 :param wrapped:
211 :param instance:
212 :param args:
213 :param kwargs:
214 :param transaction:
215 :return:
216 """
217 return args, kwargs
218
219 def call(self, module, method, wrapped, instance, args, kwargs):
220 """
221 Wrapped call. This method should gather all necessary data, then call
222 `wrapped` in a `capture_span` context manager.
223
224 Note that by default this wrapper will only be used if a transaction is
225 currently active. If you want the ability to create a transaction in
226 your `call()` method, set `create_transactions = True` at the class
227 level.
228
229 :param module: Name of the wrapped module
230 :param method: Name of the wrapped method/function
231 :param wrapped: the wrapped method/function object
232 :param instance: the wrapped instance
233 :param args: arguments to the wrapped method/function
234 :param kwargs: keyword arguments to the wrapped method/function
235 :return: the result of calling the wrapped method/function
236 """
237 raise NotImplementedError
238
[end of elasticapm/instrumentation/packages/base.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/elasticapm/instrumentation/packages/base.py b/elasticapm/instrumentation/packages/base.py
--- a/elasticapm/instrumentation/packages/base.py
+++ b/elasticapm/instrumentation/packages/base.py
@@ -38,6 +38,11 @@
logger = get_logger("elasticapm.instrument")
+class ElasticAPMFunctionWrapper(wrapt.FunctionWrapper):
+ # used to differentiate between our own function wrappers and 1st/3rd party wrappers
+ pass
+
+
class AbstractInstrumentedModule(object):
"""
This class is designed to reduce the amount of code required to
@@ -136,8 +141,13 @@
# We jump through hoop here to get the original
# `module`/`method` in the call to `call_if_sampling`
parent, attribute, original = wrapt.resolve_path(module, method)
+ if isinstance(original, ElasticAPMFunctionWrapper):
+ logger.debug("%s.%s already instrumented, skipping", module, method)
+ continue
self.originals[(module, method)] = original
- wrapper = wrapt.FunctionWrapper(original, functools.partial(self.call_if_sampling, module, method))
+ wrapper = ElasticAPMFunctionWrapper(
+ original, functools.partial(self.call_if_sampling, module, method)
+ )
wrapt.apply_patch(parent, attribute, wrapper)
instrumented_methods.append((module, method))
except ImportError:
|
{"golden_diff": "diff --git a/elasticapm/instrumentation/packages/base.py b/elasticapm/instrumentation/packages/base.py\n--- a/elasticapm/instrumentation/packages/base.py\n+++ b/elasticapm/instrumentation/packages/base.py\n@@ -38,6 +38,11 @@\n logger = get_logger(\"elasticapm.instrument\")\n \n \n+class ElasticAPMFunctionWrapper(wrapt.FunctionWrapper):\n+ # used to differentiate between our own function wrappers and 1st/3rd party wrappers\n+ pass\n+\n+\n class AbstractInstrumentedModule(object):\n \"\"\"\n This class is designed to reduce the amount of code required to\n@@ -136,8 +141,13 @@\n # We jump through hoop here to get the original\n # `module`/`method` in the call to `call_if_sampling`\n parent, attribute, original = wrapt.resolve_path(module, method)\n+ if isinstance(original, ElasticAPMFunctionWrapper):\n+ logger.debug(\"%s.%s already instrumented, skipping\", module, method)\n+ continue\n self.originals[(module, method)] = original\n- wrapper = wrapt.FunctionWrapper(original, functools.partial(self.call_if_sampling, module, method))\n+ wrapper = ElasticAPMFunctionWrapper(\n+ original, functools.partial(self.call_if_sampling, module, method)\n+ )\n wrapt.apply_patch(parent, attribute, wrapper)\n instrumented_methods.append((module, method))\n except ImportError:\n", "issue": "database trace show twice\nmaybe this is not a bug\r\nkibana show every database trace will show twice\r\n\r\n\r\nbut in 4.2.2 version, it show this \r\n\r\n\r\ni think its related to span? in 4.2.2 agent won't upload span log, please tell me how to resovle it \ud83d\ude22\r\n(forgive my poor English)\r\n\r\n**Environment (please complete the following information)**\r\n- OS: [e.g. Linux] linux\r\n- Python version: 3.7.\r\n- Framework and version [e.g. Django 2.1]: Django 3.0\r\n- APM Server version: 7.4.2\r\n- Agent version: 5.3.3\r\n\n", "before_files": [{"content": "# BSD 3-Clause License\n#\n# Copyright (c) 2019, Elasticsearch BV\n# All rights reserved.\n#\n# Redistribution and use in source and binary forms, with or without\n# modification, are permitted provided that the following conditions are met:\n#\n# * Redistributions of source code must retain the above copyright notice, this\n# list of conditions and the following disclaimer.\n#\n# * Redistributions in binary form must reproduce the above copyright notice,\n# this list of conditions and the following disclaimer in the documentation\n# and/or other materials provided with the distribution.\n#\n# * Neither the name of the copyright holder nor the names of its\n# contributors may be used to endorse or promote products derived from\n# this software without specific prior written permission.\n#\n# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS \"AS IS\"\n# AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE\n# IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE\n# DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE\n# FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL\n# DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR\n# SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER\n# CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,\n# OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE\n# OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.\n\nimport functools\nimport os\n\nfrom elasticapm.traces import execution_context\nfrom elasticapm.utils import wrapt\nfrom elasticapm.utils.logging import get_logger\n\nlogger = get_logger(\"elasticapm.instrument\")\n\n\nclass AbstractInstrumentedModule(object):\n \"\"\"\n This class is designed to reduce the amount of code required to\n instrument library functions using wrapt.\n\n Instrumentation modules inherit from this class and override pieces as\n needed. Only `name`, `instrumented_list`, and `call` are required in\n the inheriting class.\n\n The `instrument_list` is a list of (module, method) pairs that will be\n instrumented. The module/method need not be imported -- in fact, because\n instrumentation modules are all processed during the instrumentation\n process, lazy imports should be used in order to avoid ImportError\n exceptions.\n\n The `instrument()` method will be called for each InstrumentedModule\n listed in the instrument register (elasticapm.instrumentation.register),\n and each method in the `instrument_list` will be wrapped (using wrapt)\n with the `call_if_sampling()` function, which (by default) will either\n call the wrapped function by itself, or pass it into `call()` to be\n called if there is a transaction active.\n\n For simple span-wrapping of instrumented libraries, a very simple\n InstrumentedModule might look like this::\n\n from elasticapm.instrumentation.packages.base import AbstractInstrumentedModule\n from elasticapm.traces import capture_span\n\n class Jinja2Instrumentation(AbstractInstrumentedModule):\n name = \"jinja2\"\n instrument_list = [(\"jinja2\", \"Template.render\")]\n def call(self, module, method, wrapped, instance, args, kwargs):\n signature = instance.name or instance.filename\n with capture_span(signature, span_type=\"template\", span_subtype=\"jinja2\", span_action=\"render\"):\n return wrapped(*args, **kwargs)\n\n This class can also be used to instrument callables which are expected to\n create their own transactions (rather than spans within a transaction).\n In this case, set `creates_transaction = True` next to your `name` and\n `instrument_list`. This tells the instrumenting code to always wrap the\n method with `call()`, even if there is no transaction active. It is\n expected in this case that a new transaction will be created as part of\n your `call()` method.\n \"\"\"\n\n name = None\n mutates_unsampled_arguments = False\n creates_transactions = False\n\n instrument_list = [\n # List of (module, method) pairs to instrument. E.g.:\n # (\"requests.sessions\", \"Session.send\"),\n ]\n\n def __init__(self):\n self.originals = {}\n self.instrumented = False\n\n assert self.name is not None\n\n def get_wrapped_name(self, wrapped, instance, fallback_method=None):\n wrapped_name = []\n if hasattr(instance, \"__class__\") and hasattr(instance.__class__, \"__name__\"):\n wrapped_name.append(instance.__class__.__name__)\n\n if hasattr(wrapped, \"__name__\"):\n wrapped_name.append(wrapped.__name__)\n elif fallback_method:\n attribute = fallback_method.split(\".\")\n if len(attribute) == 2:\n wrapped_name.append(attribute[1])\n\n return \".\".join(wrapped_name)\n\n def get_instrument_list(self):\n return self.instrument_list\n\n def instrument(self):\n if self.instrumented:\n return\n\n skip_env_var = \"SKIP_INSTRUMENT_\" + str(self.name.upper())\n if skip_env_var in os.environ:\n logger.debug(\"Skipping instrumentation of %s. %s is set.\", self.name, skip_env_var)\n return\n try:\n instrument_list = self.get_instrument_list()\n skipped_modules = set()\n instrumented_methods = []\n\n for module, method in instrument_list:\n try:\n # Skip modules we already failed to load\n if module in skipped_modules:\n continue\n # We jump through hoop here to get the original\n # `module`/`method` in the call to `call_if_sampling`\n parent, attribute, original = wrapt.resolve_path(module, method)\n self.originals[(module, method)] = original\n wrapper = wrapt.FunctionWrapper(original, functools.partial(self.call_if_sampling, module, method))\n wrapt.apply_patch(parent, attribute, wrapper)\n instrumented_methods.append((module, method))\n except ImportError:\n # Could not import module\n logger.debug(\"Skipping instrumentation of %s. Module %s not found\", self.name, module)\n\n # Keep track of modules we couldn't load so we don't\n # try to instrument anything in that module again\n skipped_modules.add(module)\n except AttributeError as ex:\n # Could not find thing in module\n logger.debug(\"Skipping instrumentation of %s.%s: %s\", module, method, ex)\n if instrumented_methods:\n logger.debug(\"Instrumented %s, %s\", self.name, \", \".join(\".\".join(m) for m in instrumented_methods))\n\n except ImportError as ex:\n logger.debug(\"Skipping instrumentation of %s. %s\", self.name, ex)\n self.instrumented = True\n\n def uninstrument(self):\n if not self.instrumented or not self.originals:\n return\n uninstrumented_methods = []\n for module, method in self.get_instrument_list():\n if (module, method) in self.originals:\n parent, attribute, wrapper = wrapt.resolve_path(module, method)\n wrapt.apply_patch(parent, attribute, self.originals[(module, method)])\n uninstrumented_methods.append((module, method))\n if uninstrumented_methods:\n logger.debug(\"Uninstrumented %s, %s\", self.name, \", \".join(\".\".join(m) for m in uninstrumented_methods))\n self.instrumented = False\n self.originals = {}\n\n def call_if_sampling(self, module, method, wrapped, instance, args, kwargs):\n \"\"\"\n This is the function which will wrap the instrumented method/function.\n\n By default, will call the instrumented method/function, via `call()`,\n only if a transaction is active and sampled. This behavior can be\n overridden by setting `creates_transactions = True` at the class\n level.\n\n If `creates_transactions == False` and there's an active transaction\n with `transaction.is_sampled == False`, then the\n `mutate_unsampled_call_args()` method is called, and the resulting\n args and kwargs are passed into the wrapped function directly, not\n via `call()`. This can e.g. be used to add traceparent headers to the\n underlying http call for HTTP instrumentations, even if we're not\n sampling the transaction.\n \"\"\"\n if self.creates_transactions:\n return self.call(module, method, wrapped, instance, args, kwargs)\n transaction = execution_context.get_transaction()\n if not transaction:\n return wrapped(*args, **kwargs)\n elif not transaction.is_sampled:\n args, kwargs = self.mutate_unsampled_call_args(module, method, wrapped, instance, args, kwargs, transaction)\n return wrapped(*args, **kwargs)\n else:\n return self.call(module, method, wrapped, instance, args, kwargs)\n\n def mutate_unsampled_call_args(self, module, method, wrapped, instance, args, kwargs, transaction):\n \"\"\"\n Method called for unsampled wrapped calls. This can e.g. be used to\n add traceparent headers to the underlying http call for HTTP\n instrumentations.\n\n :param module:\n :param method:\n :param wrapped:\n :param instance:\n :param args:\n :param kwargs:\n :param transaction:\n :return:\n \"\"\"\n return args, kwargs\n\n def call(self, module, method, wrapped, instance, args, kwargs):\n \"\"\"\n Wrapped call. This method should gather all necessary data, then call\n `wrapped` in a `capture_span` context manager.\n\n Note that by default this wrapper will only be used if a transaction is\n currently active. If you want the ability to create a transaction in\n your `call()` method, set `create_transactions = True` at the class\n level.\n\n :param module: Name of the wrapped module\n :param method: Name of the wrapped method/function\n :param wrapped: the wrapped method/function object\n :param instance: the wrapped instance\n :param args: arguments to the wrapped method/function\n :param kwargs: keyword arguments to the wrapped method/function\n :return: the result of calling the wrapped method/function\n \"\"\"\n raise NotImplementedError\n", "path": "elasticapm/instrumentation/packages/base.py"}]}
| 3,657 | 319 |
gh_patches_debug_14157
|
rasdani/github-patches
|
git_diff
|
jazzband__pip-tools-1159
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
5.1.2 regression: lost dependencies using same zip file with different subdirectories
In `pip-tools` 5.1.2 and later, when I provide two copies of the same zip file URL with different subdirectories, the dependencies of the first are lost:
```
https://github.com/zulip/python-zulip-api/archive/0.7.0.zip#egg=zulip==0.7.0&subdirectory=zulip
https://github.com/zulip/python-zulip-api/archive/0.7.0.zip#egg=zulip-bots==0.7.0&subdirectory=zulip_bots
```
`git bisect` shows that this regression was introduced in commit c0b33e7f2224758524001e755cc338965ca0b2c0 (#1132).
Cc @richafrank
#### Environment Versions
1. OS Type: Ubuntu 18.04
1. Python version: 3.6.9
1. pip version: 20.1.1
1. pip-tools version: 5.2.0
#### Steps to replicate
```sh
cat > z.in <<EOF
https://github.com/zulip/python-zulip-api/archive/0.7.0.zip#egg=zulip==0.7.0&subdirectory=zulip
https://github.com/zulip/python-zulip-api/archive/0.7.0.zip#egg=zulip-bots==0.7.0&subdirectory=zulip_bots
EOF
pip-compile z.in
```
#### Expected result
```
#
# This file is autogenerated by pip-compile
# To update, run:
#
# pip-compile z.in
#
beautifulsoup4==4.9.1 # via zulip-bots
certifi==2020.4.5.1 # via requests
cffi==1.14.0 # via cryptography
chardet==3.0.4 # via requests
cryptography==2.9.2 # via pyopenssl, requests
distro==1.5.0 # via zulip
html2text==2020.1.16 # via zulip-bots
idna==2.9 # via requests
lxml==4.5.1 # via zulip-bots
matrix-client==0.3.2 # via zulip
pycparser==2.20 # via cffi
pyopenssl==19.1.0 # via requests
requests[security]==2.23.0 # via matrix-client, zulip
six==1.15.0 # via cryptography, pyopenssl, zulip
soupsieve==2.0.1 # via beautifulsoup4
urllib3==1.25.9 # via requests
https://github.com/zulip/python-zulip-api/archive/0.7.0.zip#egg=zulip-bots==0.7.0&subdirectory=zulip_bots # via -r z.in
https://github.com/zulip/python-zulip-api/archive/0.7.0.zip#egg=zulip==0.7.0&subdirectory=zulip # via -r z.in, zulip-bots
# The following packages are considered to be unsafe in a requirements file:
# pip
```
#### Actual result
```
#
# This file is autogenerated by pip-compile
# To update, run:
#
# pip-compile z.in
#
beautifulsoup4==4.9.1 # via zulip-bots
html2text==2020.1.16 # via zulip-bots
lxml==4.5.1 # via zulip-bots
soupsieve==2.0.1 # via beautifulsoup4
https://github.com/zulip/python-zulip-api/archive/0.7.0.zip#egg=zulip-bots==0.7.0&subdirectory=zulip_bots # via -r z.in
https://github.com/zulip/python-zulip-api/archive/0.7.0.zip#egg=zulip==0.7.0&subdirectory=zulip # via -r z.in, zulip-bots
# The following packages are considered to be unsafe in a requirements file:
# pip
```
</issue>
<code>
[start of piptools/repositories/base.py]
1 # coding: utf-8
2 from __future__ import absolute_import, division, print_function, unicode_literals
3
4 from abc import ABCMeta, abstractmethod
5 from contextlib import contextmanager
6
7 from six import add_metaclass
8
9
10 @add_metaclass(ABCMeta)
11 class BaseRepository(object):
12 def clear_caches(self):
13 """Should clear any caches used by the implementation."""
14
15 def freshen_build_caches(self):
16 """Should start with fresh build/source caches."""
17
18 @abstractmethod
19 def find_best_match(self, ireq):
20 """
21 Return a Version object that indicates the best match for the given
22 InstallRequirement according to the repository.
23 """
24
25 @abstractmethod
26 def get_dependencies(self, ireq):
27 """
28 Given a pinned, URL, or editable InstallRequirement, returns a set of
29 dependencies (also InstallRequirements, but not necessarily pinned).
30 They indicate the secondary dependencies for the given requirement.
31 """
32
33 @abstractmethod
34 def get_hashes(self, ireq):
35 """
36 Given a pinned InstallRequire, returns a set of hashes that represent
37 all of the files for a given requirement. It is not acceptable for an
38 editable or unpinned requirement to be passed to this function.
39 """
40
41 @abstractmethod
42 @contextmanager
43 def allow_all_wheels(self):
44 """
45 Monkey patches pip.Wheel to allow wheels from all platforms and Python versions.
46 """
47
48 def copy_ireq_dependencies(self, source, dest):
49 """
50 Notifies the repository that `dest` is a copy of `source`, and so it
51 has the same dependencies. Otherwise, once we prepare an ireq to assign
52 it its name, we would lose track of those dependencies on combining
53 that ireq with others.
54 """
55
[end of piptools/repositories/base.py]
[start of piptools/repositories/local.py]
1 # coding: utf-8
2 from __future__ import absolute_import, division, print_function, unicode_literals
3
4 from contextlib import contextmanager
5
6 from pip._internal.utils.hashes import FAVORITE_HASH
7
8 from .._compat import PIP_VERSION
9 from .base import BaseRepository
10
11 from piptools.utils import as_tuple, key_from_ireq, make_install_requirement
12
13
14 def ireq_satisfied_by_existing_pin(ireq, existing_pin):
15 """
16 Return True if the given InstallationRequirement is satisfied by the
17 previously encountered version pin.
18 """
19 version = next(iter(existing_pin.req.specifier)).version
20 return ireq.req.specifier.contains(
21 version, prereleases=existing_pin.req.specifier.prereleases
22 )
23
24
25 class LocalRequirementsRepository(BaseRepository):
26 """
27 The LocalRequirementsRepository proxied the _real_ repository by first
28 checking if a requirement can be satisfied by existing pins (i.e. the
29 result of a previous compile step).
30
31 In effect, if a requirement can be satisfied with a version pinned in the
32 requirements file, we prefer that version over the best match found in
33 PyPI. This keeps updates to the requirements.txt down to a minimum.
34 """
35
36 def __init__(self, existing_pins, proxied_repository):
37 self.repository = proxied_repository
38 self.existing_pins = existing_pins
39
40 @property
41 def options(self):
42 return self.repository.options
43
44 @property
45 def finder(self):
46 return self.repository.finder
47
48 @property
49 def session(self):
50 return self.repository.session
51
52 @property
53 def DEFAULT_INDEX_URL(self):
54 return self.repository.DEFAULT_INDEX_URL
55
56 def clear_caches(self):
57 self.repository.clear_caches()
58
59 def freshen_build_caches(self):
60 self.repository.freshen_build_caches()
61
62 def find_best_match(self, ireq, prereleases=None):
63 key = key_from_ireq(ireq)
64 existing_pin = self.existing_pins.get(key)
65 if existing_pin and ireq_satisfied_by_existing_pin(ireq, existing_pin):
66 project, version, _ = as_tuple(existing_pin)
67 return make_install_requirement(
68 project, version, ireq.extras, constraint=ireq.constraint
69 )
70 else:
71 return self.repository.find_best_match(ireq, prereleases)
72
73 def get_dependencies(self, ireq):
74 return self.repository.get_dependencies(ireq)
75
76 def get_hashes(self, ireq):
77 key = key_from_ireq(ireq)
78 existing_pin = self.existing_pins.get(key)
79 if existing_pin and ireq_satisfied_by_existing_pin(ireq, existing_pin):
80 if PIP_VERSION[:2] <= (20, 0):
81 hashes = existing_pin.options.get("hashes", {})
82 else:
83 hashes = existing_pin.hash_options
84 hexdigests = hashes.get(FAVORITE_HASH)
85 if hexdigests:
86 return {
87 ":".join([FAVORITE_HASH, hexdigest]) for hexdigest in hexdigests
88 }
89 return self.repository.get_hashes(ireq)
90
91 @contextmanager
92 def allow_all_wheels(self):
93 with self.repository.allow_all_wheels():
94 yield
95
[end of piptools/repositories/local.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/piptools/repositories/base.py b/piptools/repositories/base.py
--- a/piptools/repositories/base.py
+++ b/piptools/repositories/base.py
@@ -45,6 +45,7 @@
Monkey patches pip.Wheel to allow wheels from all platforms and Python versions.
"""
+ @abstractmethod
def copy_ireq_dependencies(self, source, dest):
"""
Notifies the repository that `dest` is a copy of `source`, and so it
diff --git a/piptools/repositories/local.py b/piptools/repositories/local.py
--- a/piptools/repositories/local.py
+++ b/piptools/repositories/local.py
@@ -92,3 +92,6 @@
def allow_all_wheels(self):
with self.repository.allow_all_wheels():
yield
+
+ def copy_ireq_dependencies(self, source, dest):
+ self.repository.copy_ireq_dependencies(source, dest)
|
{"golden_diff": "diff --git a/piptools/repositories/base.py b/piptools/repositories/base.py\n--- a/piptools/repositories/base.py\n+++ b/piptools/repositories/base.py\n@@ -45,6 +45,7 @@\n Monkey patches pip.Wheel to allow wheels from all platforms and Python versions.\n \"\"\"\n \n+ @abstractmethod\n def copy_ireq_dependencies(self, source, dest):\n \"\"\"\n Notifies the repository that `dest` is a copy of `source`, and so it\ndiff --git a/piptools/repositories/local.py b/piptools/repositories/local.py\n--- a/piptools/repositories/local.py\n+++ b/piptools/repositories/local.py\n@@ -92,3 +92,6 @@\n def allow_all_wheels(self):\n with self.repository.allow_all_wheels():\n yield\n+\n+ def copy_ireq_dependencies(self, source, dest):\n+ self.repository.copy_ireq_dependencies(source, dest)\n", "issue": "5.1.2 regression: lost dependencies using same zip file with different subdirectories\nIn `pip-tools` 5.1.2 and later, when I provide two copies of the same zip file URL with different subdirectories, the dependencies of the first are lost:\r\n\r\n```\r\nhttps://github.com/zulip/python-zulip-api/archive/0.7.0.zip#egg=zulip==0.7.0&subdirectory=zulip\r\nhttps://github.com/zulip/python-zulip-api/archive/0.7.0.zip#egg=zulip-bots==0.7.0&subdirectory=zulip_bots\r\n```\r\n\r\n`git bisect` shows that this regression was introduced in commit c0b33e7f2224758524001e755cc338965ca0b2c0 (#1132).\r\n\r\nCc @richafrank\r\n\r\n#### Environment Versions\r\n\r\n1. OS Type: Ubuntu 18.04\r\n1. Python version: 3.6.9\r\n1. pip version: 20.1.1\r\n1. pip-tools version: 5.2.0\r\n\r\n#### Steps to replicate\r\n\r\n```sh\r\ncat > z.in <<EOF\r\nhttps://github.com/zulip/python-zulip-api/archive/0.7.0.zip#egg=zulip==0.7.0&subdirectory=zulip\r\nhttps://github.com/zulip/python-zulip-api/archive/0.7.0.zip#egg=zulip-bots==0.7.0&subdirectory=zulip_bots\r\nEOF\r\npip-compile z.in\r\n```\r\n\r\n#### Expected result\r\n\r\n```\r\n#\r\n# This file is autogenerated by pip-compile\r\n# To update, run:\r\n#\r\n# pip-compile z.in\r\n#\r\nbeautifulsoup4==4.9.1 # via zulip-bots\r\ncertifi==2020.4.5.1 # via requests\r\ncffi==1.14.0 # via cryptography\r\nchardet==3.0.4 # via requests\r\ncryptography==2.9.2 # via pyopenssl, requests\r\ndistro==1.5.0 # via zulip\r\nhtml2text==2020.1.16 # via zulip-bots\r\nidna==2.9 # via requests\r\nlxml==4.5.1 # via zulip-bots\r\nmatrix-client==0.3.2 # via zulip\r\npycparser==2.20 # via cffi\r\npyopenssl==19.1.0 # via requests\r\nrequests[security]==2.23.0 # via matrix-client, zulip\r\nsix==1.15.0 # via cryptography, pyopenssl, zulip\r\nsoupsieve==2.0.1 # via beautifulsoup4\r\nurllib3==1.25.9 # via requests\r\nhttps://github.com/zulip/python-zulip-api/archive/0.7.0.zip#egg=zulip-bots==0.7.0&subdirectory=zulip_bots # via -r z.in\r\nhttps://github.com/zulip/python-zulip-api/archive/0.7.0.zip#egg=zulip==0.7.0&subdirectory=zulip # via -r z.in, zulip-bots\r\n\r\n# The following packages are considered to be unsafe in a requirements file:\r\n# pip\r\n```\r\n\r\n#### Actual result\r\n\r\n```\r\n#\r\n# This file is autogenerated by pip-compile\r\n# To update, run:\r\n#\r\n# pip-compile z.in\r\n#\r\nbeautifulsoup4==4.9.1 # via zulip-bots\r\nhtml2text==2020.1.16 # via zulip-bots\r\nlxml==4.5.1 # via zulip-bots\r\nsoupsieve==2.0.1 # via beautifulsoup4\r\nhttps://github.com/zulip/python-zulip-api/archive/0.7.0.zip#egg=zulip-bots==0.7.0&subdirectory=zulip_bots # via -r z.in\r\nhttps://github.com/zulip/python-zulip-api/archive/0.7.0.zip#egg=zulip==0.7.0&subdirectory=zulip # via -r z.in, zulip-bots\r\n\r\n# The following packages are considered to be unsafe in a requirements file:\r\n# pip\r\n```\r\n\n", "before_files": [{"content": "# coding: utf-8\nfrom __future__ import absolute_import, division, print_function, unicode_literals\n\nfrom abc import ABCMeta, abstractmethod\nfrom contextlib import contextmanager\n\nfrom six import add_metaclass\n\n\n@add_metaclass(ABCMeta)\nclass BaseRepository(object):\n def clear_caches(self):\n \"\"\"Should clear any caches used by the implementation.\"\"\"\n\n def freshen_build_caches(self):\n \"\"\"Should start with fresh build/source caches.\"\"\"\n\n @abstractmethod\n def find_best_match(self, ireq):\n \"\"\"\n Return a Version object that indicates the best match for the given\n InstallRequirement according to the repository.\n \"\"\"\n\n @abstractmethod\n def get_dependencies(self, ireq):\n \"\"\"\n Given a pinned, URL, or editable InstallRequirement, returns a set of\n dependencies (also InstallRequirements, but not necessarily pinned).\n They indicate the secondary dependencies for the given requirement.\n \"\"\"\n\n @abstractmethod\n def get_hashes(self, ireq):\n \"\"\"\n Given a pinned InstallRequire, returns a set of hashes that represent\n all of the files for a given requirement. It is not acceptable for an\n editable or unpinned requirement to be passed to this function.\n \"\"\"\n\n @abstractmethod\n @contextmanager\n def allow_all_wheels(self):\n \"\"\"\n Monkey patches pip.Wheel to allow wheels from all platforms and Python versions.\n \"\"\"\n\n def copy_ireq_dependencies(self, source, dest):\n \"\"\"\n Notifies the repository that `dest` is a copy of `source`, and so it\n has the same dependencies. Otherwise, once we prepare an ireq to assign\n it its name, we would lose track of those dependencies on combining\n that ireq with others.\n \"\"\"\n", "path": "piptools/repositories/base.py"}, {"content": "# coding: utf-8\nfrom __future__ import absolute_import, division, print_function, unicode_literals\n\nfrom contextlib import contextmanager\n\nfrom pip._internal.utils.hashes import FAVORITE_HASH\n\nfrom .._compat import PIP_VERSION\nfrom .base import BaseRepository\n\nfrom piptools.utils import as_tuple, key_from_ireq, make_install_requirement\n\n\ndef ireq_satisfied_by_existing_pin(ireq, existing_pin):\n \"\"\"\n Return True if the given InstallationRequirement is satisfied by the\n previously encountered version pin.\n \"\"\"\n version = next(iter(existing_pin.req.specifier)).version\n return ireq.req.specifier.contains(\n version, prereleases=existing_pin.req.specifier.prereleases\n )\n\n\nclass LocalRequirementsRepository(BaseRepository):\n \"\"\"\n The LocalRequirementsRepository proxied the _real_ repository by first\n checking if a requirement can be satisfied by existing pins (i.e. the\n result of a previous compile step).\n\n In effect, if a requirement can be satisfied with a version pinned in the\n requirements file, we prefer that version over the best match found in\n PyPI. This keeps updates to the requirements.txt down to a minimum.\n \"\"\"\n\n def __init__(self, existing_pins, proxied_repository):\n self.repository = proxied_repository\n self.existing_pins = existing_pins\n\n @property\n def options(self):\n return self.repository.options\n\n @property\n def finder(self):\n return self.repository.finder\n\n @property\n def session(self):\n return self.repository.session\n\n @property\n def DEFAULT_INDEX_URL(self):\n return self.repository.DEFAULT_INDEX_URL\n\n def clear_caches(self):\n self.repository.clear_caches()\n\n def freshen_build_caches(self):\n self.repository.freshen_build_caches()\n\n def find_best_match(self, ireq, prereleases=None):\n key = key_from_ireq(ireq)\n existing_pin = self.existing_pins.get(key)\n if existing_pin and ireq_satisfied_by_existing_pin(ireq, existing_pin):\n project, version, _ = as_tuple(existing_pin)\n return make_install_requirement(\n project, version, ireq.extras, constraint=ireq.constraint\n )\n else:\n return self.repository.find_best_match(ireq, prereleases)\n\n def get_dependencies(self, ireq):\n return self.repository.get_dependencies(ireq)\n\n def get_hashes(self, ireq):\n key = key_from_ireq(ireq)\n existing_pin = self.existing_pins.get(key)\n if existing_pin and ireq_satisfied_by_existing_pin(ireq, existing_pin):\n if PIP_VERSION[:2] <= (20, 0):\n hashes = existing_pin.options.get(\"hashes\", {})\n else:\n hashes = existing_pin.hash_options\n hexdigests = hashes.get(FAVORITE_HASH)\n if hexdigests:\n return {\n \":\".join([FAVORITE_HASH, hexdigest]) for hexdigest in hexdigests\n }\n return self.repository.get_hashes(ireq)\n\n @contextmanager\n def allow_all_wheels(self):\n with self.repository.allow_all_wheels():\n yield\n", "path": "piptools/repositories/local.py"}]}
| 2,931 | 211 |
gh_patches_debug_40411
|
rasdani/github-patches
|
git_diff
|
pytorch__audio-1826
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
doc for bg_iterator et al
The [documentation](https://pytorch.org/audio/datasets.html) for torchaudio does not include `bg_iterator` and `disk_cache_iterator` from [code](https://github.com/pytorch/audio/blob/master/torchaudio/datasets/utils.py). Thanks to [comment](https://discuss.pytorch.org/t/audio-dataset-load-large-file-into-memory-in-background/85943) to point this out :)
</issue>
<code>
[start of torchaudio/datasets/__init__.py]
1 from .commonvoice import COMMONVOICE
2 from .librispeech import LIBRISPEECH
3 from .speechcommands import SPEECHCOMMANDS
4 from .utils import bg_iterator, diskcache_iterator
5 from .vctk import VCTK_092
6 from .gtzan import GTZAN
7 from .yesno import YESNO
8 from .ljspeech import LJSPEECH
9 from .cmuarctic import CMUARCTIC
10 from .cmudict import CMUDict
11 from .libritts import LIBRITTS
12 from .tedlium import TEDLIUM
13
14
15 __all__ = [
16 "COMMONVOICE",
17 "LIBRISPEECH",
18 "SPEECHCOMMANDS",
19 "VCTK_092",
20 "YESNO",
21 "LJSPEECH",
22 "GTZAN",
23 "CMUARCTIC",
24 "CMUDict",
25 "LIBRITTS",
26 "diskcache_iterator",
27 "bg_iterator",
28 "TEDLIUM",
29 ]
30
[end of torchaudio/datasets/__init__.py]
[start of torchaudio/datasets/utils.py]
1 import hashlib
2 import logging
3 import os
4 import tarfile
5 import threading
6 import urllib
7 import urllib.request
8 import zipfile
9 from queue import Queue
10 from typing import Any, Iterable, List, Optional
11
12 import torch
13 from torch.utils.data import Dataset
14 from torch.utils.model_zoo import tqdm
15
16 from torchaudio._internal.module_utils import deprecated
17
18
19 def stream_url(url: str,
20 start_byte: Optional[int] = None,
21 block_size: int = 32 * 1024,
22 progress_bar: bool = True) -> Iterable:
23 """Stream url by chunk
24
25 Args:
26 url (str): Url.
27 start_byte (int or None, optional): Start streaming at that point (Default: ``None``).
28 block_size (int, optional): Size of chunks to stream (Default: ``32 * 1024``).
29 progress_bar (bool, optional): Display a progress bar (Default: ``True``).
30 """
31
32 # If we already have the whole file, there is no need to download it again
33 req = urllib.request.Request(url, method="HEAD")
34 with urllib.request.urlopen(req) as response:
35 url_size = int(response.info().get("Content-Length", -1))
36 if url_size == start_byte:
37 return
38
39 req = urllib.request.Request(url)
40 if start_byte:
41 req.headers["Range"] = "bytes={}-".format(start_byte)
42
43 with urllib.request.urlopen(req) as upointer, tqdm(
44 unit="B",
45 unit_scale=True,
46 unit_divisor=1024,
47 total=url_size,
48 disable=not progress_bar,
49 ) as pbar:
50
51 num_bytes = 0
52 while True:
53 chunk = upointer.read(block_size)
54 if not chunk:
55 break
56 yield chunk
57 num_bytes += len(chunk)
58 pbar.update(len(chunk))
59
60
61 def download_url(url: str,
62 download_folder: str,
63 filename: Optional[str] = None,
64 hash_value: Optional[str] = None,
65 hash_type: str = "sha256",
66 progress_bar: bool = True,
67 resume: bool = False) -> None:
68 """Download file to disk.
69
70 Args:
71 url (str): Url.
72 download_folder (str): Folder to download file.
73 filename (str or None, optional): Name of downloaded file. If None, it is inferred from the url
74 (Default: ``None``).
75 hash_value (str or None, optional): Hash for url (Default: ``None``).
76 hash_type (str, optional): Hash type, among "sha256" and "md5" (Default: ``"sha256"``).
77 progress_bar (bool, optional): Display a progress bar (Default: ``True``).
78 resume (bool, optional): Enable resuming download (Default: ``False``).
79 """
80
81 req = urllib.request.Request(url, method="HEAD")
82 req_info = urllib.request.urlopen(req).info()
83
84 # Detect filename
85 filename = filename or req_info.get_filename() or os.path.basename(url)
86 filepath = os.path.join(download_folder, filename)
87 if resume and os.path.exists(filepath):
88 mode = "ab"
89 local_size: Optional[int] = os.path.getsize(filepath)
90
91 elif not resume and os.path.exists(filepath):
92 raise RuntimeError(
93 "{} already exists. Delete the file manually and retry.".format(filepath)
94 )
95 else:
96 mode = "wb"
97 local_size = None
98
99 if hash_value and local_size == int(req_info.get("Content-Length", -1)):
100 with open(filepath, "rb") as file_obj:
101 if validate_file(file_obj, hash_value, hash_type):
102 return
103 raise RuntimeError(
104 "The hash of {} does not match. Delete the file manually and retry.".format(
105 filepath
106 )
107 )
108
109 with open(filepath, mode) as fpointer:
110 for chunk in stream_url(url, start_byte=local_size, progress_bar=progress_bar):
111 fpointer.write(chunk)
112
113 with open(filepath, "rb") as file_obj:
114 if hash_value and not validate_file(file_obj, hash_value, hash_type):
115 raise RuntimeError(
116 "The hash of {} does not match. Delete the file manually and retry.".format(
117 filepath
118 )
119 )
120
121
122 def validate_file(file_obj: Any, hash_value: str, hash_type: str = "sha256") -> bool:
123 """Validate a given file object with its hash.
124
125 Args:
126 file_obj: File object to read from.
127 hash_value (str): Hash for url.
128 hash_type (str, optional): Hash type, among "sha256" and "md5" (Default: ``"sha256"``).
129
130 Returns:
131 bool: return True if its a valid file, else False.
132 """
133
134 if hash_type == "sha256":
135 hash_func = hashlib.sha256()
136 elif hash_type == "md5":
137 hash_func = hashlib.md5()
138 else:
139 raise ValueError
140
141 while True:
142 # Read by chunk to avoid filling memory
143 chunk = file_obj.read(1024 ** 2)
144 if not chunk:
145 break
146 hash_func.update(chunk)
147
148 return hash_func.hexdigest() == hash_value
149
150
151 def extract_archive(from_path: str, to_path: Optional[str] = None, overwrite: bool = False) -> List[str]:
152 """Extract archive.
153 Args:
154 from_path (str): the path of the archive.
155 to_path (str or None, optional): the root path of the extraced files (directory of from_path)
156 (Default: ``None``)
157 overwrite (bool, optional): overwrite existing files (Default: ``False``)
158
159 Returns:
160 list: List of paths to extracted files even if not overwritten.
161
162 Examples:
163 >>> url = 'http://www.quest.dcs.shef.ac.uk/wmt16_files_mmt/validation.tar.gz'
164 >>> from_path = './validation.tar.gz'
165 >>> to_path = './'
166 >>> torchaudio.datasets.utils.download_from_url(url, from_path)
167 >>> torchaudio.datasets.utils.extract_archive(from_path, to_path)
168 """
169
170 if to_path is None:
171 to_path = os.path.dirname(from_path)
172
173 try:
174 with tarfile.open(from_path, "r") as tar:
175 logging.info("Opened tar file {}.".format(from_path))
176 files = []
177 for file_ in tar: # type: Any
178 file_path = os.path.join(to_path, file_.name)
179 if file_.isfile():
180 files.append(file_path)
181 if os.path.exists(file_path):
182 logging.info("{} already extracted.".format(file_path))
183 if not overwrite:
184 continue
185 tar.extract(file_, to_path)
186 return files
187 except tarfile.ReadError:
188 pass
189
190 try:
191 with zipfile.ZipFile(from_path, "r") as zfile:
192 logging.info("Opened zip file {}.".format(from_path))
193 files = zfile.namelist()
194 for file_ in files:
195 file_path = os.path.join(to_path, file_)
196 if os.path.exists(file_path):
197 logging.info("{} already extracted.".format(file_path))
198 if not overwrite:
199 continue
200 zfile.extract(file_, to_path)
201 return files
202 except zipfile.BadZipFile:
203 pass
204
205 raise NotImplementedError("We currently only support tar.gz, tgz, and zip achives.")
206
207
208 class _DiskCache(Dataset):
209 """
210 Wrap a dataset so that, whenever a new item is returned, it is saved to disk.
211 """
212
213 def __init__(self, dataset: Dataset, location: str = ".cached") -> None:
214 self.dataset = dataset
215 self.location = location
216
217 self._id = id(self)
218 self._cache: List = [None] * len(dataset)
219
220 def __getitem__(self, n: int) -> Any:
221 if self._cache[n]:
222 f = self._cache[n]
223 return torch.load(f)
224
225 f = str(self._id) + "-" + str(n)
226 f = os.path.join(self.location, f)
227 item = self.dataset[n]
228
229 self._cache[n] = f
230 os.makedirs(self.location, exist_ok=True)
231 torch.save(item, f)
232
233 return item
234
235 def __len__(self) -> int:
236 return len(self.dataset)
237
238
239 @deprecated('', version='0.11')
240 def diskcache_iterator(dataset: Dataset, location: str = ".cached") -> Dataset:
241 return _DiskCache(dataset, location)
242
243
244 class _ThreadedIterator(threading.Thread):
245 """
246 Prefetch the next queue_length items from iterator in a background thread.
247
248 Example:
249 >> for i in bg_iterator(range(10)):
250 >> print(i)
251 """
252
253 class _End:
254 pass
255
256 def __init__(self, generator: Iterable, maxsize: int) -> None:
257 threading.Thread.__init__(self)
258 self.queue: Queue = Queue(maxsize)
259 self.generator = generator
260 self.daemon = True
261 self.start()
262
263 def run(self) -> None:
264 for item in self.generator:
265 self.queue.put(item)
266 self.queue.put(self._End)
267
268 def __iter__(self) -> Any:
269 return self
270
271 def __next__(self) -> Any:
272 next_item = self.queue.get()
273 if next_item == self._End:
274 raise StopIteration
275 return next_item
276
277 # Required for Python 2.7 compatibility
278 def next(self) -> Any:
279 return self.__next__()
280
281
282 @deprecated('', version='0.11')
283 def bg_iterator(iterable: Iterable, maxsize: int) -> Any:
284 return _ThreadedIterator(iterable, maxsize=maxsize)
285
[end of torchaudio/datasets/utils.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/torchaudio/datasets/__init__.py b/torchaudio/datasets/__init__.py
--- a/torchaudio/datasets/__init__.py
+++ b/torchaudio/datasets/__init__.py
@@ -1,7 +1,6 @@
from .commonvoice import COMMONVOICE
from .librispeech import LIBRISPEECH
from .speechcommands import SPEECHCOMMANDS
-from .utils import bg_iterator, diskcache_iterator
from .vctk import VCTK_092
from .gtzan import GTZAN
from .yesno import YESNO
@@ -23,7 +22,5 @@
"CMUARCTIC",
"CMUDict",
"LIBRITTS",
- "diskcache_iterator",
- "bg_iterator",
"TEDLIUM",
]
diff --git a/torchaudio/datasets/utils.py b/torchaudio/datasets/utils.py
--- a/torchaudio/datasets/utils.py
+++ b/torchaudio/datasets/utils.py
@@ -2,19 +2,13 @@
import logging
import os
import tarfile
-import threading
import urllib
import urllib.request
import zipfile
-from queue import Queue
from typing import Any, Iterable, List, Optional
-import torch
-from torch.utils.data import Dataset
from torch.utils.model_zoo import tqdm
-from torchaudio._internal.module_utils import deprecated
-
def stream_url(url: str,
start_byte: Optional[int] = None,
@@ -203,82 +197,3 @@
pass
raise NotImplementedError("We currently only support tar.gz, tgz, and zip achives.")
-
-
-class _DiskCache(Dataset):
- """
- Wrap a dataset so that, whenever a new item is returned, it is saved to disk.
- """
-
- def __init__(self, dataset: Dataset, location: str = ".cached") -> None:
- self.dataset = dataset
- self.location = location
-
- self._id = id(self)
- self._cache: List = [None] * len(dataset)
-
- def __getitem__(self, n: int) -> Any:
- if self._cache[n]:
- f = self._cache[n]
- return torch.load(f)
-
- f = str(self._id) + "-" + str(n)
- f = os.path.join(self.location, f)
- item = self.dataset[n]
-
- self._cache[n] = f
- os.makedirs(self.location, exist_ok=True)
- torch.save(item, f)
-
- return item
-
- def __len__(self) -> int:
- return len(self.dataset)
-
-
-@deprecated('', version='0.11')
-def diskcache_iterator(dataset: Dataset, location: str = ".cached") -> Dataset:
- return _DiskCache(dataset, location)
-
-
-class _ThreadedIterator(threading.Thread):
- """
- Prefetch the next queue_length items from iterator in a background thread.
-
- Example:
- >> for i in bg_iterator(range(10)):
- >> print(i)
- """
-
- class _End:
- pass
-
- def __init__(self, generator: Iterable, maxsize: int) -> None:
- threading.Thread.__init__(self)
- self.queue: Queue = Queue(maxsize)
- self.generator = generator
- self.daemon = True
- self.start()
-
- def run(self) -> None:
- for item in self.generator:
- self.queue.put(item)
- self.queue.put(self._End)
-
- def __iter__(self) -> Any:
- return self
-
- def __next__(self) -> Any:
- next_item = self.queue.get()
- if next_item == self._End:
- raise StopIteration
- return next_item
-
- # Required for Python 2.7 compatibility
- def next(self) -> Any:
- return self.__next__()
-
-
-@deprecated('', version='0.11')
-def bg_iterator(iterable: Iterable, maxsize: int) -> Any:
- return _ThreadedIterator(iterable, maxsize=maxsize)
|
{"golden_diff": "diff --git a/torchaudio/datasets/__init__.py b/torchaudio/datasets/__init__.py\n--- a/torchaudio/datasets/__init__.py\n+++ b/torchaudio/datasets/__init__.py\n@@ -1,7 +1,6 @@\n from .commonvoice import COMMONVOICE\n from .librispeech import LIBRISPEECH\n from .speechcommands import SPEECHCOMMANDS\n-from .utils import bg_iterator, diskcache_iterator\n from .vctk import VCTK_092\n from .gtzan import GTZAN\n from .yesno import YESNO\n@@ -23,7 +22,5 @@\n \"CMUARCTIC\",\n \"CMUDict\",\n \"LIBRITTS\",\n- \"diskcache_iterator\",\n- \"bg_iterator\",\n \"TEDLIUM\",\n ]\ndiff --git a/torchaudio/datasets/utils.py b/torchaudio/datasets/utils.py\n--- a/torchaudio/datasets/utils.py\n+++ b/torchaudio/datasets/utils.py\n@@ -2,19 +2,13 @@\n import logging\n import os\n import tarfile\n-import threading\n import urllib\n import urllib.request\n import zipfile\n-from queue import Queue\n from typing import Any, Iterable, List, Optional\n \n-import torch\n-from torch.utils.data import Dataset\n from torch.utils.model_zoo import tqdm\n \n-from torchaudio._internal.module_utils import deprecated\n-\n \n def stream_url(url: str,\n start_byte: Optional[int] = None,\n@@ -203,82 +197,3 @@\n pass\n \n raise NotImplementedError(\"We currently only support tar.gz, tgz, and zip achives.\")\n-\n-\n-class _DiskCache(Dataset):\n- \"\"\"\n- Wrap a dataset so that, whenever a new item is returned, it is saved to disk.\n- \"\"\"\n-\n- def __init__(self, dataset: Dataset, location: str = \".cached\") -> None:\n- self.dataset = dataset\n- self.location = location\n-\n- self._id = id(self)\n- self._cache: List = [None] * len(dataset)\n-\n- def __getitem__(self, n: int) -> Any:\n- if self._cache[n]:\n- f = self._cache[n]\n- return torch.load(f)\n-\n- f = str(self._id) + \"-\" + str(n)\n- f = os.path.join(self.location, f)\n- item = self.dataset[n]\n-\n- self._cache[n] = f\n- os.makedirs(self.location, exist_ok=True)\n- torch.save(item, f)\n-\n- return item\n-\n- def __len__(self) -> int:\n- return len(self.dataset)\n-\n-\n-@deprecated('', version='0.11')\n-def diskcache_iterator(dataset: Dataset, location: str = \".cached\") -> Dataset:\n- return _DiskCache(dataset, location)\n-\n-\n-class _ThreadedIterator(threading.Thread):\n- \"\"\"\n- Prefetch the next queue_length items from iterator in a background thread.\n-\n- Example:\n- >> for i in bg_iterator(range(10)):\n- >> print(i)\n- \"\"\"\n-\n- class _End:\n- pass\n-\n- def __init__(self, generator: Iterable, maxsize: int) -> None:\n- threading.Thread.__init__(self)\n- self.queue: Queue = Queue(maxsize)\n- self.generator = generator\n- self.daemon = True\n- self.start()\n-\n- def run(self) -> None:\n- for item in self.generator:\n- self.queue.put(item)\n- self.queue.put(self._End)\n-\n- def __iter__(self) -> Any:\n- return self\n-\n- def __next__(self) -> Any:\n- next_item = self.queue.get()\n- if next_item == self._End:\n- raise StopIteration\n- return next_item\n-\n- # Required for Python 2.7 compatibility\n- def next(self) -> Any:\n- return self.__next__()\n-\n-\n-@deprecated('', version='0.11')\n-def bg_iterator(iterable: Iterable, maxsize: int) -> Any:\n- return _ThreadedIterator(iterable, maxsize=maxsize)\n", "issue": "doc for bg_iterator et al\nThe [documentation](https://pytorch.org/audio/datasets.html) for torchaudio does not include `bg_iterator` and `disk_cache_iterator` from [code](https://github.com/pytorch/audio/blob/master/torchaudio/datasets/utils.py). Thanks to [comment](https://discuss.pytorch.org/t/audio-dataset-load-large-file-into-memory-in-background/85943) to point this out :)\n", "before_files": [{"content": "from .commonvoice import COMMONVOICE\nfrom .librispeech import LIBRISPEECH\nfrom .speechcommands import SPEECHCOMMANDS\nfrom .utils import bg_iterator, diskcache_iterator\nfrom .vctk import VCTK_092\nfrom .gtzan import GTZAN\nfrom .yesno import YESNO\nfrom .ljspeech import LJSPEECH\nfrom .cmuarctic import CMUARCTIC\nfrom .cmudict import CMUDict\nfrom .libritts import LIBRITTS\nfrom .tedlium import TEDLIUM\n\n\n__all__ = [\n \"COMMONVOICE\",\n \"LIBRISPEECH\",\n \"SPEECHCOMMANDS\",\n \"VCTK_092\",\n \"YESNO\",\n \"LJSPEECH\",\n \"GTZAN\",\n \"CMUARCTIC\",\n \"CMUDict\",\n \"LIBRITTS\",\n \"diskcache_iterator\",\n \"bg_iterator\",\n \"TEDLIUM\",\n]\n", "path": "torchaudio/datasets/__init__.py"}, {"content": "import hashlib\nimport logging\nimport os\nimport tarfile\nimport threading\nimport urllib\nimport urllib.request\nimport zipfile\nfrom queue import Queue\nfrom typing import Any, Iterable, List, Optional\n\nimport torch\nfrom torch.utils.data import Dataset\nfrom torch.utils.model_zoo import tqdm\n\nfrom torchaudio._internal.module_utils import deprecated\n\n\ndef stream_url(url: str,\n start_byte: Optional[int] = None,\n block_size: int = 32 * 1024,\n progress_bar: bool = True) -> Iterable:\n \"\"\"Stream url by chunk\n\n Args:\n url (str): Url.\n start_byte (int or None, optional): Start streaming at that point (Default: ``None``).\n block_size (int, optional): Size of chunks to stream (Default: ``32 * 1024``).\n progress_bar (bool, optional): Display a progress bar (Default: ``True``).\n \"\"\"\n\n # If we already have the whole file, there is no need to download it again\n req = urllib.request.Request(url, method=\"HEAD\")\n with urllib.request.urlopen(req) as response:\n url_size = int(response.info().get(\"Content-Length\", -1))\n if url_size == start_byte:\n return\n\n req = urllib.request.Request(url)\n if start_byte:\n req.headers[\"Range\"] = \"bytes={}-\".format(start_byte)\n\n with urllib.request.urlopen(req) as upointer, tqdm(\n unit=\"B\",\n unit_scale=True,\n unit_divisor=1024,\n total=url_size,\n disable=not progress_bar,\n ) as pbar:\n\n num_bytes = 0\n while True:\n chunk = upointer.read(block_size)\n if not chunk:\n break\n yield chunk\n num_bytes += len(chunk)\n pbar.update(len(chunk))\n\n\ndef download_url(url: str,\n download_folder: str,\n filename: Optional[str] = None,\n hash_value: Optional[str] = None,\n hash_type: str = \"sha256\",\n progress_bar: bool = True,\n resume: bool = False) -> None:\n \"\"\"Download file to disk.\n\n Args:\n url (str): Url.\n download_folder (str): Folder to download file.\n filename (str or None, optional): Name of downloaded file. If None, it is inferred from the url\n (Default: ``None``).\n hash_value (str or None, optional): Hash for url (Default: ``None``).\n hash_type (str, optional): Hash type, among \"sha256\" and \"md5\" (Default: ``\"sha256\"``).\n progress_bar (bool, optional): Display a progress bar (Default: ``True``).\n resume (bool, optional): Enable resuming download (Default: ``False``).\n \"\"\"\n\n req = urllib.request.Request(url, method=\"HEAD\")\n req_info = urllib.request.urlopen(req).info()\n\n # Detect filename\n filename = filename or req_info.get_filename() or os.path.basename(url)\n filepath = os.path.join(download_folder, filename)\n if resume and os.path.exists(filepath):\n mode = \"ab\"\n local_size: Optional[int] = os.path.getsize(filepath)\n\n elif not resume and os.path.exists(filepath):\n raise RuntimeError(\n \"{} already exists. Delete the file manually and retry.\".format(filepath)\n )\n else:\n mode = \"wb\"\n local_size = None\n\n if hash_value and local_size == int(req_info.get(\"Content-Length\", -1)):\n with open(filepath, \"rb\") as file_obj:\n if validate_file(file_obj, hash_value, hash_type):\n return\n raise RuntimeError(\n \"The hash of {} does not match. Delete the file manually and retry.\".format(\n filepath\n )\n )\n\n with open(filepath, mode) as fpointer:\n for chunk in stream_url(url, start_byte=local_size, progress_bar=progress_bar):\n fpointer.write(chunk)\n\n with open(filepath, \"rb\") as file_obj:\n if hash_value and not validate_file(file_obj, hash_value, hash_type):\n raise RuntimeError(\n \"The hash of {} does not match. Delete the file manually and retry.\".format(\n filepath\n )\n )\n\n\ndef validate_file(file_obj: Any, hash_value: str, hash_type: str = \"sha256\") -> bool:\n \"\"\"Validate a given file object with its hash.\n\n Args:\n file_obj: File object to read from.\n hash_value (str): Hash for url.\n hash_type (str, optional): Hash type, among \"sha256\" and \"md5\" (Default: ``\"sha256\"``).\n\n Returns:\n bool: return True if its a valid file, else False.\n \"\"\"\n\n if hash_type == \"sha256\":\n hash_func = hashlib.sha256()\n elif hash_type == \"md5\":\n hash_func = hashlib.md5()\n else:\n raise ValueError\n\n while True:\n # Read by chunk to avoid filling memory\n chunk = file_obj.read(1024 ** 2)\n if not chunk:\n break\n hash_func.update(chunk)\n\n return hash_func.hexdigest() == hash_value\n\n\ndef extract_archive(from_path: str, to_path: Optional[str] = None, overwrite: bool = False) -> List[str]:\n \"\"\"Extract archive.\n Args:\n from_path (str): the path of the archive.\n to_path (str or None, optional): the root path of the extraced files (directory of from_path)\n (Default: ``None``)\n overwrite (bool, optional): overwrite existing files (Default: ``False``)\n\n Returns:\n list: List of paths to extracted files even if not overwritten.\n\n Examples:\n >>> url = 'http://www.quest.dcs.shef.ac.uk/wmt16_files_mmt/validation.tar.gz'\n >>> from_path = './validation.tar.gz'\n >>> to_path = './'\n >>> torchaudio.datasets.utils.download_from_url(url, from_path)\n >>> torchaudio.datasets.utils.extract_archive(from_path, to_path)\n \"\"\"\n\n if to_path is None:\n to_path = os.path.dirname(from_path)\n\n try:\n with tarfile.open(from_path, \"r\") as tar:\n logging.info(\"Opened tar file {}.\".format(from_path))\n files = []\n for file_ in tar: # type: Any\n file_path = os.path.join(to_path, file_.name)\n if file_.isfile():\n files.append(file_path)\n if os.path.exists(file_path):\n logging.info(\"{} already extracted.\".format(file_path))\n if not overwrite:\n continue\n tar.extract(file_, to_path)\n return files\n except tarfile.ReadError:\n pass\n\n try:\n with zipfile.ZipFile(from_path, \"r\") as zfile:\n logging.info(\"Opened zip file {}.\".format(from_path))\n files = zfile.namelist()\n for file_ in files:\n file_path = os.path.join(to_path, file_)\n if os.path.exists(file_path):\n logging.info(\"{} already extracted.\".format(file_path))\n if not overwrite:\n continue\n zfile.extract(file_, to_path)\n return files\n except zipfile.BadZipFile:\n pass\n\n raise NotImplementedError(\"We currently only support tar.gz, tgz, and zip achives.\")\n\n\nclass _DiskCache(Dataset):\n \"\"\"\n Wrap a dataset so that, whenever a new item is returned, it is saved to disk.\n \"\"\"\n\n def __init__(self, dataset: Dataset, location: str = \".cached\") -> None:\n self.dataset = dataset\n self.location = location\n\n self._id = id(self)\n self._cache: List = [None] * len(dataset)\n\n def __getitem__(self, n: int) -> Any:\n if self._cache[n]:\n f = self._cache[n]\n return torch.load(f)\n\n f = str(self._id) + \"-\" + str(n)\n f = os.path.join(self.location, f)\n item = self.dataset[n]\n\n self._cache[n] = f\n os.makedirs(self.location, exist_ok=True)\n torch.save(item, f)\n\n return item\n\n def __len__(self) -> int:\n return len(self.dataset)\n\n\n@deprecated('', version='0.11')\ndef diskcache_iterator(dataset: Dataset, location: str = \".cached\") -> Dataset:\n return _DiskCache(dataset, location)\n\n\nclass _ThreadedIterator(threading.Thread):\n \"\"\"\n Prefetch the next queue_length items from iterator in a background thread.\n\n Example:\n >> for i in bg_iterator(range(10)):\n >> print(i)\n \"\"\"\n\n class _End:\n pass\n\n def __init__(self, generator: Iterable, maxsize: int) -> None:\n threading.Thread.__init__(self)\n self.queue: Queue = Queue(maxsize)\n self.generator = generator\n self.daemon = True\n self.start()\n\n def run(self) -> None:\n for item in self.generator:\n self.queue.put(item)\n self.queue.put(self._End)\n\n def __iter__(self) -> Any:\n return self\n\n def __next__(self) -> Any:\n next_item = self.queue.get()\n if next_item == self._End:\n raise StopIteration\n return next_item\n\n # Required for Python 2.7 compatibility\n def next(self) -> Any:\n return self.__next__()\n\n\n@deprecated('', version='0.11')\ndef bg_iterator(iterable: Iterable, maxsize: int) -> Any:\n return _ThreadedIterator(iterable, maxsize=maxsize)\n", "path": "torchaudio/datasets/utils.py"}]}
| 3,818 | 935 |
gh_patches_debug_11009
|
rasdani/github-patches
|
git_diff
|
pyca__cryptography-7895
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Bump BoringSSL and/or OpenSSL in CI
## BoringSSL
[Commit: e2e613c269a6bb3d7c0271150fff48d11fdbbace](https://boringssl.googlesource.com/boringssl/+/e2e613c269a6bb3d7c0271150fff48d11fdbbace)
[Diff](https://boringssl.googlesource.com/boringssl/+/d77fdbff010ee70776036c41155d1b3711ede548..e2e613c269a6bb3d7c0271150fff48d11fdbbace) between the last commit hash merged to this repository and the new commit.
## OpenSSL
[Commit: dc45d4c6faeb53bb68401141d899b9f857bbc51d](https://github.com/openssl/openssl/commit/dc45d4c6faeb53bb68401141d899b9f857bbc51d)
[Diff](https://github.com/openssl/openssl/compare/efec0f4611ee854f2b0b3da0c135e839bf8e7d04...dc45d4c6faeb53bb68401141d899b9f857bbc51d) between the last commit hash merged to this repository and the new commit.
</issue>
<code>
[start of src/_cffi_src/openssl/rsa.py]
1 # This file is dual licensed under the terms of the Apache License, Version
2 # 2.0, and the BSD License. See the LICENSE file in the root of this repository
3 # for complete details.
4
5
6 INCLUDES = """
7 #include <openssl/rsa.h>
8 """
9
10 TYPES = """
11 typedef ... RSA;
12 typedef ... BN_GENCB;
13 static const int RSA_PKCS1_PADDING;
14 static const int RSA_NO_PADDING;
15 static const int RSA_PKCS1_OAEP_PADDING;
16 static const int RSA_PKCS1_PSS_PADDING;
17 static const int RSA_F4;
18 static const int RSA_PSS_SALTLEN_AUTO;
19 """
20
21 FUNCTIONS = """
22 RSA *RSA_new(void);
23 void RSA_free(RSA *);
24 int RSA_generate_key_ex(RSA *, int, BIGNUM *, BN_GENCB *);
25 int RSA_check_key(const RSA *);
26 RSA *RSAPublicKey_dup(RSA *);
27 int RSA_blinding_on(RSA *, BN_CTX *);
28 int RSA_print(BIO *, const RSA *, int);
29
30 int RSA_set0_key(RSA *, BIGNUM *, BIGNUM *, BIGNUM *);
31 int RSA_set0_factors(RSA *, BIGNUM *, BIGNUM *);
32 int RSA_set0_crt_params(RSA *, BIGNUM *, BIGNUM *, BIGNUM *);
33 void RSA_get0_key(const RSA *, const BIGNUM **, const BIGNUM **,
34 const BIGNUM **);
35 void RSA_get0_factors(const RSA *, const BIGNUM **, const BIGNUM **);
36 void RSA_get0_crt_params(const RSA *, const BIGNUM **, const BIGNUM **,
37 const BIGNUM **);
38 int EVP_PKEY_CTX_set_rsa_padding(EVP_PKEY_CTX *, int);
39 int EVP_PKEY_CTX_set_rsa_pss_saltlen(EVP_PKEY_CTX *, int);
40 int EVP_PKEY_CTX_set_rsa_mgf1_md(EVP_PKEY_CTX *, EVP_MD *);
41 int EVP_PKEY_CTX_set0_rsa_oaep_label(EVP_PKEY_CTX *, unsigned char *, int);
42
43 int EVP_PKEY_CTX_set_rsa_oaep_md(EVP_PKEY_CTX *, EVP_MD *);
44 """
45
46 CUSTOMIZATIONS = """
47 // BoringSSL doesn't define this constant, but the value is used for
48 // automatic salt length computation as in OpenSSL and LibreSSL
49 #if !defined(RSA_PSS_SALTLEN_AUTO)
50 #define RSA_PSS_SALTLEN_AUTO -2
51 #endif
52 """
53
[end of src/_cffi_src/openssl/rsa.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/src/_cffi_src/openssl/rsa.py b/src/_cffi_src/openssl/rsa.py
--- a/src/_cffi_src/openssl/rsa.py
+++ b/src/_cffi_src/openssl/rsa.py
@@ -16,6 +16,8 @@
static const int RSA_PKCS1_PSS_PADDING;
static const int RSA_F4;
static const int RSA_PSS_SALTLEN_AUTO;
+
+static const int Cryptography_HAS_IMPLICIT_RSA_REJECTION;
"""
FUNCTIONS = """
@@ -49,4 +51,10 @@
#if !defined(RSA_PSS_SALTLEN_AUTO)
#define RSA_PSS_SALTLEN_AUTO -2
#endif
+
+#if defined(EVP_PKEY_CTRL_RSA_IMPLICIT_REJECTION)
+static const int Cryptography_HAS_IMPLICIT_RSA_REJECTION = 1;
+#else
+static const int Cryptography_HAS_IMPLICIT_RSA_REJECTION = 0;
+#endif
"""
|
{"golden_diff": "diff --git a/src/_cffi_src/openssl/rsa.py b/src/_cffi_src/openssl/rsa.py\n--- a/src/_cffi_src/openssl/rsa.py\n+++ b/src/_cffi_src/openssl/rsa.py\n@@ -16,6 +16,8 @@\n static const int RSA_PKCS1_PSS_PADDING;\n static const int RSA_F4;\n static const int RSA_PSS_SALTLEN_AUTO;\n+\n+static const int Cryptography_HAS_IMPLICIT_RSA_REJECTION;\n \"\"\"\n \n FUNCTIONS = \"\"\"\n@@ -49,4 +51,10 @@\n #if !defined(RSA_PSS_SALTLEN_AUTO)\n #define RSA_PSS_SALTLEN_AUTO -2\n #endif\n+\n+#if defined(EVP_PKEY_CTRL_RSA_IMPLICIT_REJECTION)\n+static const int Cryptography_HAS_IMPLICIT_RSA_REJECTION = 1;\n+#else\n+static const int Cryptography_HAS_IMPLICIT_RSA_REJECTION = 0;\n+#endif\n \"\"\"\n", "issue": "Bump BoringSSL and/or OpenSSL in CI\n## BoringSSL\n[Commit: e2e613c269a6bb3d7c0271150fff48d11fdbbace](https://boringssl.googlesource.com/boringssl/+/e2e613c269a6bb3d7c0271150fff48d11fdbbace)\n\n[Diff](https://boringssl.googlesource.com/boringssl/+/d77fdbff010ee70776036c41155d1b3711ede548..e2e613c269a6bb3d7c0271150fff48d11fdbbace) between the last commit hash merged to this repository and the new commit.\n## OpenSSL\n[Commit: dc45d4c6faeb53bb68401141d899b9f857bbc51d](https://github.com/openssl/openssl/commit/dc45d4c6faeb53bb68401141d899b9f857bbc51d)\n\n[Diff](https://github.com/openssl/openssl/compare/efec0f4611ee854f2b0b3da0c135e839bf8e7d04...dc45d4c6faeb53bb68401141d899b9f857bbc51d) between the last commit hash merged to this repository and the new commit.\n", "before_files": [{"content": "# This file is dual licensed under the terms of the Apache License, Version\n# 2.0, and the BSD License. See the LICENSE file in the root of this repository\n# for complete details.\n\n\nINCLUDES = \"\"\"\n#include <openssl/rsa.h>\n\"\"\"\n\nTYPES = \"\"\"\ntypedef ... RSA;\ntypedef ... BN_GENCB;\nstatic const int RSA_PKCS1_PADDING;\nstatic const int RSA_NO_PADDING;\nstatic const int RSA_PKCS1_OAEP_PADDING;\nstatic const int RSA_PKCS1_PSS_PADDING;\nstatic const int RSA_F4;\nstatic const int RSA_PSS_SALTLEN_AUTO;\n\"\"\"\n\nFUNCTIONS = \"\"\"\nRSA *RSA_new(void);\nvoid RSA_free(RSA *);\nint RSA_generate_key_ex(RSA *, int, BIGNUM *, BN_GENCB *);\nint RSA_check_key(const RSA *);\nRSA *RSAPublicKey_dup(RSA *);\nint RSA_blinding_on(RSA *, BN_CTX *);\nint RSA_print(BIO *, const RSA *, int);\n\nint RSA_set0_key(RSA *, BIGNUM *, BIGNUM *, BIGNUM *);\nint RSA_set0_factors(RSA *, BIGNUM *, BIGNUM *);\nint RSA_set0_crt_params(RSA *, BIGNUM *, BIGNUM *, BIGNUM *);\nvoid RSA_get0_key(const RSA *, const BIGNUM **, const BIGNUM **,\n const BIGNUM **);\nvoid RSA_get0_factors(const RSA *, const BIGNUM **, const BIGNUM **);\nvoid RSA_get0_crt_params(const RSA *, const BIGNUM **, const BIGNUM **,\n const BIGNUM **);\nint EVP_PKEY_CTX_set_rsa_padding(EVP_PKEY_CTX *, int);\nint EVP_PKEY_CTX_set_rsa_pss_saltlen(EVP_PKEY_CTX *, int);\nint EVP_PKEY_CTX_set_rsa_mgf1_md(EVP_PKEY_CTX *, EVP_MD *);\nint EVP_PKEY_CTX_set0_rsa_oaep_label(EVP_PKEY_CTX *, unsigned char *, int);\n\nint EVP_PKEY_CTX_set_rsa_oaep_md(EVP_PKEY_CTX *, EVP_MD *);\n\"\"\"\n\nCUSTOMIZATIONS = \"\"\"\n// BoringSSL doesn't define this constant, but the value is used for\n// automatic salt length computation as in OpenSSL and LibreSSL\n#if !defined(RSA_PSS_SALTLEN_AUTO)\n#define RSA_PSS_SALTLEN_AUTO -2\n#endif\n\"\"\"\n", "path": "src/_cffi_src/openssl/rsa.py"}]}
| 1,532 | 207 |
gh_patches_debug_9169
|
rasdani/github-patches
|
git_diff
|
qutebrowser__qutebrowser-2400
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Disable the automatic "Run :adblock-update" info message on startup
Every time I start qutebrowser I got a message:
> Run :adblock-update to get adblock lists
I've set `host-blocking-enabled = false` in the config, because I have my own hosts file in /etc + I have privoxy.
I couldn't find a way to disable this message. I searched in the config, in the man page, in the help, here through the issues... Am I missing something? :confused:
</issue>
<code>
[start of qutebrowser/browser/adblock.py]
1 # vim: ft=python fileencoding=utf-8 sts=4 sw=4 et:
2
3 # Copyright 2014-2016 Florian Bruhin (The Compiler) <[email protected]>
4 #
5 # This file is part of qutebrowser.
6 #
7 # qutebrowser is free software: you can redistribute it and/or modify
8 # it under the terms of the GNU General Public License as published by
9 # the Free Software Foundation, either version 3 of the License, or
10 # (at your option) any later version.
11 #
12 # qutebrowser is distributed in the hope that it will be useful,
13 # but WITHOUT ANY WARRANTY; without even the implied warranty of
14 # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
15 # GNU General Public License for more details.
16 #
17 # You should have received a copy of the GNU General Public License
18 # along with qutebrowser. If not, see <http://www.gnu.org/licenses/>.
19
20 """Functions related to ad blocking."""
21
22 import io
23 import os.path
24 import functools
25 import posixpath
26 import zipfile
27 import fnmatch
28
29 from qutebrowser.browser import downloads
30 from qutebrowser.config import config
31 from qutebrowser.utils import objreg, standarddir, log, message
32 from qutebrowser.commands import cmdutils
33
34
35 def guess_zip_filename(zf):
36 """Guess which file to use inside a zip file.
37
38 Args:
39 zf: A ZipFile instance.
40 """
41 files = zf.namelist()
42 if len(files) == 1:
43 return files[0]
44 else:
45 for e in files:
46 if posixpath.splitext(e)[0].lower() == 'hosts':
47 return e
48 raise FileNotFoundError("No hosts file found in zip")
49
50
51 def get_fileobj(byte_io):
52 """Get a usable file object to read the hosts file from."""
53 byte_io.seek(0) # rewind downloaded file
54 if zipfile.is_zipfile(byte_io):
55 byte_io.seek(0) # rewind what zipfile.is_zipfile did
56 zf = zipfile.ZipFile(byte_io)
57 filename = guess_zip_filename(zf)
58 byte_io = zf.open(filename, mode='r')
59 else:
60 byte_io.seek(0) # rewind what zipfile.is_zipfile did
61 return io.TextIOWrapper(byte_io, encoding='utf-8')
62
63
64 def is_whitelisted_host(host):
65 """Check if the given host is on the adblock whitelist.
66
67 Args:
68 host: The host of the request as string.
69 """
70 whitelist = config.get('content', 'host-blocking-whitelist')
71 if whitelist is None:
72 return False
73
74 for pattern in whitelist:
75 if fnmatch.fnmatch(host, pattern.lower()):
76 return True
77 return False
78
79
80 class FakeDownload:
81
82 """A download stub to use on_download_finished with local files."""
83
84 def __init__(self, fileobj):
85 self.basename = os.path.basename(fileobj.name)
86 self.fileobj = fileobj
87 self.successful = True
88
89
90 class HostBlocker:
91
92 """Manage blocked hosts based from /etc/hosts-like files.
93
94 Attributes:
95 _blocked_hosts: A set of blocked hosts.
96 _config_blocked_hosts: A set of blocked hosts from ~/.config.
97 _in_progress: The DownloadItems which are currently downloading.
98 _done_count: How many files have been read successfully.
99 _local_hosts_file: The path to the blocked-hosts file.
100 _config_hosts_file: The path to a blocked-hosts in ~/.config
101
102 Class attributes:
103 WHITELISTED: Hosts which never should be blocked.
104 """
105
106 WHITELISTED = ('localhost', 'localhost.localdomain', 'broadcasthost',
107 'local')
108
109 def __init__(self):
110 self._blocked_hosts = set()
111 self._config_blocked_hosts = set()
112 self._in_progress = []
113 self._done_count = 0
114
115 data_dir = standarddir.data()
116 self._local_hosts_file = os.path.join(data_dir, 'blocked-hosts')
117 self.on_config_changed()
118
119 config_dir = standarddir.config()
120 self._config_hosts_file = os.path.join(config_dir, 'blocked-hosts')
121
122 objreg.get('config').changed.connect(self.on_config_changed)
123
124 def is_blocked(self, url):
125 """Check if the given URL (as QUrl) is blocked."""
126 if not config.get('content', 'host-blocking-enabled'):
127 return False
128 host = url.host()
129 return ((host in self._blocked_hosts or
130 host in self._config_blocked_hosts) and
131 not is_whitelisted_host(host))
132
133 def _read_hosts_file(self, filename, target):
134 """Read hosts from the given filename.
135
136 Args:
137 filename: The file to read.
138 target: The set to store the hosts in.
139
140 Return:
141 True if a read was attempted, False otherwise
142 """
143 if not os.path.exists(filename):
144 return False
145
146 try:
147 with open(filename, 'r', encoding='utf-8') as f:
148 for line in f:
149 target.add(line.strip())
150 except OSError:
151 log.misc.exception("Failed to read host blocklist!")
152
153 return True
154
155 def read_hosts(self):
156 """Read hosts from the existing blocked-hosts file."""
157 self._blocked_hosts = set()
158
159 self._read_hosts_file(self._config_hosts_file,
160 self._config_blocked_hosts)
161
162 found = self._read_hosts_file(self._local_hosts_file,
163 self._blocked_hosts)
164
165 if not found:
166 args = objreg.get('args')
167 if (config.get('content', 'host-block-lists') is not None and
168 args.basedir is None):
169 message.info("Run :adblock-update to get adblock lists.")
170
171 @cmdutils.register(instance='host-blocker')
172 def adblock_update(self):
173 """Update the adblock block lists.
174
175 This updates `~/.local/share/qutebrowser/blocked-hosts` with downloaded
176 host lists and re-reads `~/.config/qutebrowser/blocked-hosts`.
177 """
178 self._read_hosts_file(self._config_hosts_file,
179 self._config_blocked_hosts)
180 self._blocked_hosts = set()
181 self._done_count = 0
182 urls = config.get('content', 'host-block-lists')
183 download_manager = objreg.get('qtnetwork-download-manager',
184 scope='window', window='last-focused')
185 if urls is None:
186 return
187 for url in urls:
188 if url.scheme() == 'file':
189 try:
190 fileobj = open(url.path(), 'rb')
191 except OSError as e:
192 message.error("adblock: Error while reading {}: {}".format(
193 url.path(), e.strerror))
194 continue
195 download = FakeDownload(fileobj)
196 self._in_progress.append(download)
197 self.on_download_finished(download)
198 else:
199 fobj = io.BytesIO()
200 fobj.name = 'adblock: ' + url.host()
201 target = downloads.FileObjDownloadTarget(fobj)
202 download = download_manager.get(url, target=target,
203 auto_remove=True)
204 self._in_progress.append(download)
205 download.finished.connect(
206 functools.partial(self.on_download_finished, download))
207
208 def _merge_file(self, byte_io):
209 """Read and merge host files.
210
211 Args:
212 byte_io: The BytesIO object of the completed download.
213
214 Return:
215 A set of the merged hosts.
216 """
217 error_count = 0
218 line_count = 0
219 try:
220 f = get_fileobj(byte_io)
221 except (OSError, UnicodeDecodeError, zipfile.BadZipFile,
222 zipfile.LargeZipFile, LookupError) as e:
223 message.error("adblock: Error while reading {}: {} - {}".format(
224 byte_io.name, e.__class__.__name__, e))
225 return
226 for line in f:
227 line_count += 1
228 # Remove comments
229 try:
230 hash_idx = line.index('#')
231 line = line[:hash_idx]
232 except ValueError:
233 pass
234 line = line.strip()
235 # Skip empty lines
236 if not line:
237 continue
238 parts = line.split()
239 if len(parts) == 1:
240 # "one host per line" format
241 host = parts[0]
242 elif len(parts) == 2:
243 # /etc/hosts format
244 host = parts[1]
245 else:
246 error_count += 1
247 continue
248 if host not in self.WHITELISTED:
249 self._blocked_hosts.add(host)
250 log.misc.debug("{}: read {} lines".format(byte_io.name, line_count))
251 if error_count > 0:
252 message.error("adblock: {} read errors for {}".format(
253 error_count, byte_io.name))
254
255 def on_lists_downloaded(self):
256 """Install block lists after files have been downloaded."""
257 with open(self._local_hosts_file, 'w', encoding='utf-8') as f:
258 for host in sorted(self._blocked_hosts):
259 f.write(host + '\n')
260 message.info("adblock: Read {} hosts from {} sources.".format(
261 len(self._blocked_hosts), self._done_count))
262
263 @config.change_filter('content', 'host-block-lists')
264 def on_config_changed(self):
265 """Update files when the config changed."""
266 urls = config.get('content', 'host-block-lists')
267 if urls is None:
268 try:
269 os.remove(self._local_hosts_file)
270 except FileNotFoundError:
271 pass
272 except OSError as e:
273 log.misc.exception("Failed to delete hosts file: {}".format(e))
274
275 def on_download_finished(self, download):
276 """Check if all downloads are finished and if so, trigger reading.
277
278 Arguments:
279 download: The finished DownloadItem.
280 """
281 self._in_progress.remove(download)
282 if download.successful:
283 self._done_count += 1
284 try:
285 self._merge_file(download.fileobj)
286 finally:
287 download.fileobj.close()
288 if not self._in_progress:
289 try:
290 self.on_lists_downloaded()
291 except OSError:
292 log.misc.exception("Failed to write host block list!")
293
[end of qutebrowser/browser/adblock.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/qutebrowser/browser/adblock.py b/qutebrowser/browser/adblock.py
--- a/qutebrowser/browser/adblock.py
+++ b/qutebrowser/browser/adblock.py
@@ -165,7 +165,8 @@
if not found:
args = objreg.get('args')
if (config.get('content', 'host-block-lists') is not None and
- args.basedir is None):
+ args.basedir is None and
+ config.get('content', 'host-blocking-enabled')):
message.info("Run :adblock-update to get adblock lists.")
@cmdutils.register(instance='host-blocker')
|
{"golden_diff": "diff --git a/qutebrowser/browser/adblock.py b/qutebrowser/browser/adblock.py\n--- a/qutebrowser/browser/adblock.py\n+++ b/qutebrowser/browser/adblock.py\n@@ -165,7 +165,8 @@\n if not found:\n args = objreg.get('args')\n if (config.get('content', 'host-block-lists') is not None and\n- args.basedir is None):\n+ args.basedir is None and\n+ config.get('content', 'host-blocking-enabled')):\n message.info(\"Run :adblock-update to get adblock lists.\")\n \n @cmdutils.register(instance='host-blocker')\n", "issue": "Disable the automatic \"Run :adblock-update\" info message on startup\nEvery time I start qutebrowser I got a message:\r\n\r\n> Run :adblock-update to get adblock lists\r\n\r\nI've set `host-blocking-enabled = false` in the config, because I have my own hosts file in /etc + I have privoxy.\r\n\r\nI couldn't find a way to disable this message. I searched in the config, in the man page, in the help, here through the issues... Am I missing something? :confused: \n", "before_files": [{"content": "# vim: ft=python fileencoding=utf-8 sts=4 sw=4 et:\n\n# Copyright 2014-2016 Florian Bruhin (The Compiler) <[email protected]>\n#\n# This file is part of qutebrowser.\n#\n# qutebrowser is free software: you can redistribute it and/or modify\n# it under the terms of the GNU General Public License as published by\n# the Free Software Foundation, either version 3 of the License, or\n# (at your option) any later version.\n#\n# qutebrowser is distributed in the hope that it will be useful,\n# but WITHOUT ANY WARRANTY; without even the implied warranty of\n# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the\n# GNU General Public License for more details.\n#\n# You should have received a copy of the GNU General Public License\n# along with qutebrowser. If not, see <http://www.gnu.org/licenses/>.\n\n\"\"\"Functions related to ad blocking.\"\"\"\n\nimport io\nimport os.path\nimport functools\nimport posixpath\nimport zipfile\nimport fnmatch\n\nfrom qutebrowser.browser import downloads\nfrom qutebrowser.config import config\nfrom qutebrowser.utils import objreg, standarddir, log, message\nfrom qutebrowser.commands import cmdutils\n\n\ndef guess_zip_filename(zf):\n \"\"\"Guess which file to use inside a zip file.\n\n Args:\n zf: A ZipFile instance.\n \"\"\"\n files = zf.namelist()\n if len(files) == 1:\n return files[0]\n else:\n for e in files:\n if posixpath.splitext(e)[0].lower() == 'hosts':\n return e\n raise FileNotFoundError(\"No hosts file found in zip\")\n\n\ndef get_fileobj(byte_io):\n \"\"\"Get a usable file object to read the hosts file from.\"\"\"\n byte_io.seek(0) # rewind downloaded file\n if zipfile.is_zipfile(byte_io):\n byte_io.seek(0) # rewind what zipfile.is_zipfile did\n zf = zipfile.ZipFile(byte_io)\n filename = guess_zip_filename(zf)\n byte_io = zf.open(filename, mode='r')\n else:\n byte_io.seek(0) # rewind what zipfile.is_zipfile did\n return io.TextIOWrapper(byte_io, encoding='utf-8')\n\n\ndef is_whitelisted_host(host):\n \"\"\"Check if the given host is on the adblock whitelist.\n\n Args:\n host: The host of the request as string.\n \"\"\"\n whitelist = config.get('content', 'host-blocking-whitelist')\n if whitelist is None:\n return False\n\n for pattern in whitelist:\n if fnmatch.fnmatch(host, pattern.lower()):\n return True\n return False\n\n\nclass FakeDownload:\n\n \"\"\"A download stub to use on_download_finished with local files.\"\"\"\n\n def __init__(self, fileobj):\n self.basename = os.path.basename(fileobj.name)\n self.fileobj = fileobj\n self.successful = True\n\n\nclass HostBlocker:\n\n \"\"\"Manage blocked hosts based from /etc/hosts-like files.\n\n Attributes:\n _blocked_hosts: A set of blocked hosts.\n _config_blocked_hosts: A set of blocked hosts from ~/.config.\n _in_progress: The DownloadItems which are currently downloading.\n _done_count: How many files have been read successfully.\n _local_hosts_file: The path to the blocked-hosts file.\n _config_hosts_file: The path to a blocked-hosts in ~/.config\n\n Class attributes:\n WHITELISTED: Hosts which never should be blocked.\n \"\"\"\n\n WHITELISTED = ('localhost', 'localhost.localdomain', 'broadcasthost',\n 'local')\n\n def __init__(self):\n self._blocked_hosts = set()\n self._config_blocked_hosts = set()\n self._in_progress = []\n self._done_count = 0\n\n data_dir = standarddir.data()\n self._local_hosts_file = os.path.join(data_dir, 'blocked-hosts')\n self.on_config_changed()\n\n config_dir = standarddir.config()\n self._config_hosts_file = os.path.join(config_dir, 'blocked-hosts')\n\n objreg.get('config').changed.connect(self.on_config_changed)\n\n def is_blocked(self, url):\n \"\"\"Check if the given URL (as QUrl) is blocked.\"\"\"\n if not config.get('content', 'host-blocking-enabled'):\n return False\n host = url.host()\n return ((host in self._blocked_hosts or\n host in self._config_blocked_hosts) and\n not is_whitelisted_host(host))\n\n def _read_hosts_file(self, filename, target):\n \"\"\"Read hosts from the given filename.\n\n Args:\n filename: The file to read.\n target: The set to store the hosts in.\n\n Return:\n True if a read was attempted, False otherwise\n \"\"\"\n if not os.path.exists(filename):\n return False\n\n try:\n with open(filename, 'r', encoding='utf-8') as f:\n for line in f:\n target.add(line.strip())\n except OSError:\n log.misc.exception(\"Failed to read host blocklist!\")\n\n return True\n\n def read_hosts(self):\n \"\"\"Read hosts from the existing blocked-hosts file.\"\"\"\n self._blocked_hosts = set()\n\n self._read_hosts_file(self._config_hosts_file,\n self._config_blocked_hosts)\n\n found = self._read_hosts_file(self._local_hosts_file,\n self._blocked_hosts)\n\n if not found:\n args = objreg.get('args')\n if (config.get('content', 'host-block-lists') is not None and\n args.basedir is None):\n message.info(\"Run :adblock-update to get adblock lists.\")\n\n @cmdutils.register(instance='host-blocker')\n def adblock_update(self):\n \"\"\"Update the adblock block lists.\n\n This updates `~/.local/share/qutebrowser/blocked-hosts` with downloaded\n host lists and re-reads `~/.config/qutebrowser/blocked-hosts`.\n \"\"\"\n self._read_hosts_file(self._config_hosts_file,\n self._config_blocked_hosts)\n self._blocked_hosts = set()\n self._done_count = 0\n urls = config.get('content', 'host-block-lists')\n download_manager = objreg.get('qtnetwork-download-manager',\n scope='window', window='last-focused')\n if urls is None:\n return\n for url in urls:\n if url.scheme() == 'file':\n try:\n fileobj = open(url.path(), 'rb')\n except OSError as e:\n message.error(\"adblock: Error while reading {}: {}\".format(\n url.path(), e.strerror))\n continue\n download = FakeDownload(fileobj)\n self._in_progress.append(download)\n self.on_download_finished(download)\n else:\n fobj = io.BytesIO()\n fobj.name = 'adblock: ' + url.host()\n target = downloads.FileObjDownloadTarget(fobj)\n download = download_manager.get(url, target=target,\n auto_remove=True)\n self._in_progress.append(download)\n download.finished.connect(\n functools.partial(self.on_download_finished, download))\n\n def _merge_file(self, byte_io):\n \"\"\"Read and merge host files.\n\n Args:\n byte_io: The BytesIO object of the completed download.\n\n Return:\n A set of the merged hosts.\n \"\"\"\n error_count = 0\n line_count = 0\n try:\n f = get_fileobj(byte_io)\n except (OSError, UnicodeDecodeError, zipfile.BadZipFile,\n zipfile.LargeZipFile, LookupError) as e:\n message.error(\"adblock: Error while reading {}: {} - {}\".format(\n byte_io.name, e.__class__.__name__, e))\n return\n for line in f:\n line_count += 1\n # Remove comments\n try:\n hash_idx = line.index('#')\n line = line[:hash_idx]\n except ValueError:\n pass\n line = line.strip()\n # Skip empty lines\n if not line:\n continue\n parts = line.split()\n if len(parts) == 1:\n # \"one host per line\" format\n host = parts[0]\n elif len(parts) == 2:\n # /etc/hosts format\n host = parts[1]\n else:\n error_count += 1\n continue\n if host not in self.WHITELISTED:\n self._blocked_hosts.add(host)\n log.misc.debug(\"{}: read {} lines\".format(byte_io.name, line_count))\n if error_count > 0:\n message.error(\"adblock: {} read errors for {}\".format(\n error_count, byte_io.name))\n\n def on_lists_downloaded(self):\n \"\"\"Install block lists after files have been downloaded.\"\"\"\n with open(self._local_hosts_file, 'w', encoding='utf-8') as f:\n for host in sorted(self._blocked_hosts):\n f.write(host + '\\n')\n message.info(\"adblock: Read {} hosts from {} sources.\".format(\n len(self._blocked_hosts), self._done_count))\n\n @config.change_filter('content', 'host-block-lists')\n def on_config_changed(self):\n \"\"\"Update files when the config changed.\"\"\"\n urls = config.get('content', 'host-block-lists')\n if urls is None:\n try:\n os.remove(self._local_hosts_file)\n except FileNotFoundError:\n pass\n except OSError as e:\n log.misc.exception(\"Failed to delete hosts file: {}\".format(e))\n\n def on_download_finished(self, download):\n \"\"\"Check if all downloads are finished and if so, trigger reading.\n\n Arguments:\n download: The finished DownloadItem.\n \"\"\"\n self._in_progress.remove(download)\n if download.successful:\n self._done_count += 1\n try:\n self._merge_file(download.fileobj)\n finally:\n download.fileobj.close()\n if not self._in_progress:\n try:\n self.on_lists_downloaded()\n except OSError:\n log.misc.exception(\"Failed to write host block list!\")\n", "path": "qutebrowser/browser/adblock.py"}]}
| 3,613 | 145 |
gh_patches_debug_23346
|
rasdani/github-patches
|
git_diff
|
aws-cloudformation__cfn-lint-440
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Invalid PrimaryKey attribute crashes cfn-lint instead of raising lint error.
Hi AWS,
I found a minor bug on cfn-lint version: cfn-lint 0.8.3
An attribute in Resource -> Properties-> PrimaryKey leads to crash instead of a raised lint error
Here is a sample template. If MyInvalidAttr is replaced with a correct attribute eg Name then the template works.
```AWSTemplateFormatVersion: "2010-09-09"
Transform: AWS::Serverless-2016-10-31
Description: >
sam-app
Sample SAM Template for sam-app
Globals:
Function:
Timeout: 3
Resources:
MyTable:
Type: AWS::Serverless::SimpleTable
Properties:
TableName: MyTable
PrimaryKey:
MyInvalidAttr: id
Type: String
```
Below is the error:
```
$ cfn-lint -t template.yaml
Traceback (most recent call last):
File "/Users/emilbryggare/Library/Python/2.7/bin/cfn-lint", line 11, in <module>
sys.exit(main())
File "/Users/emilbryggare/Library/Python/2.7/lib/python/site-packages/cfnlint/__main__.py", line 36, in main
args.regions, args.override_spec))
File "/Users/emilbryggare/Library/Python/2.7/lib/python/site-packages/cfnlint/core.py", line 46, in run_cli
return run_checks(filename, template, rules, regions)
File "/Users/emilbryggare/Library/Python/2.7/lib/python/site-packages/cfnlint/core.py", line 316, in run_checks
matches.extend(runner.transform())
File "/Users/emilbryggare/Library/Python/2.7/lib/python/site-packages/cfnlint/__init__.py", line 894, in transform
matches = transform.transform_template()
File "/Users/emilbryggare/Library/Python/2.7/lib/python/site-packages/cfnlint/transform.py", line 115, in transform_template
sam_translator.translate(sam_template=self._template, parameter_values={}))
File "/Users/emilbryggare/Library/Python/2.7/lib/python/site-packages/samtranslator/translator/translator.py", line 72, in translate
translated = macro.to_cloudformation(**kwargs)
File "/Users/emilbryggare/Library/Python/2.7/lib/python/site-packages/samtranslator/model/sam_resources.py", line 537, in to_cloudformation
dynamodb_resources = self._construct_dynamodb_table()
File "/Users/emilbryggare/Library/Python/2.7/lib/python/site-packages/samtranslator/model/sam_resources.py", line 546, in _construct_dynamodb_table
'AttributeName': self.PrimaryKey['Name'],
KeyError: 'Name'
```
</issue>
<code>
[start of src/cfnlint/transform.py]
1 """
2 Copyright 2018 Amazon.com, Inc. or its affiliates. All Rights Reserved.
3
4 Permission is hereby granted, free of charge, to any person obtaining a copy of this
5 software and associated documentation files (the "Software"), to deal in the Software
6 without restriction, including without limitation the rights to use, copy, modify,
7 merge, publish, distribute, sublicense, and/or sell copies of the Software, and to
8 permit persons to whom the Software is furnished to do so.
9
10 THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED,
11 INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A
12 PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT
13 HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION
14 OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE
15 SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
16 """
17 import os
18 import six
19
20 from samtranslator.parser import parser
21 from samtranslator.translator.translator import Translator
22 from samtranslator.public.exceptions import InvalidDocumentException
23
24 import cfnlint.helpers
25
26
27 class Transform(object):
28 """
29 Application Serverless Module tranform Wrappor. Based on code from AWS SAM CLI:
30 https://github.com/awslabs/aws-sam-cli/blob/develop/samcli/commands/validate/lib/sam_template_validator.py
31 """
32
33 def __init__(self, filename, template, region):
34 """
35 Initialize Transform class
36 """
37 self._filename = filename
38 self._template = template
39 self._region = region
40
41 self._managed_policy_map = self.load_managed_policies()
42 self._sam_parser = parser.Parser()
43
44 def template(self):
45 """Get the template"""
46 return self._template
47
48 def load_managed_policies(self):
49 """
50 Load the ManagedPolicies locally, based on the AWS-CLI:
51 https://github.com/awslabs/aws-sam-cli/blob/develop/samcli/lib/samlib/default_managed_policies.json
52 """
53 return cfnlint.helpers.load_resources('data/Serverless/ManagedPolicies.json')
54
55 def _replace_local_codeuri(self):
56 """
57 Replaces the CodeUri in AWS::Serverless::Function and DefinitionUri in AWS::Serverless::Api to a fake
58 S3 Uri. This is to support running the SAM Translator with valid values for these fields. If this in not done,
59 the template is invalid in the eyes of SAM Translator (the translator does not support local paths)
60 """
61
62 all_resources = self._template.get('Resources', {})
63
64 for _, resource in all_resources.items():
65
66 resource_type = resource.get('Type')
67 resource_dict = resource.get('Properties')
68
69 if resource_type == 'AWS::Serverless::Function':
70
71 Transform._update_to_s3_uri('CodeUri', resource_dict)
72
73 if resource_type == 'AWS::Serverless::Api':
74 if 'DefinitionBody' not in resource_dict:
75 Transform._update_to_s3_uri('DefinitionUri', resource_dict)
76
77 def transform_template(self):
78 """
79 Transform the Template using the Serverless Application Model.
80 """
81 matches = []
82
83 try:
84 sam_translator = Translator(managed_policy_map=self._managed_policy_map,
85 sam_parser=self._sam_parser)
86
87 self._replace_local_codeuri()
88
89 # Tell SAM to use the region we're linting in, this has to be controlled using the default AWS mechanisms, see also:
90 # https://github.com/awslabs/serverless-application-model/blob/master/samtranslator/translator/arn_generator.py
91 os.environ['AWS_DEFAULT_REGION'] = self._region
92
93 # In the Paser class, within the SAM Translator, they log a warning for when the template
94 # does not match the schema. The logger they use is the root logger instead of one scoped to
95 # their module. Currently this does not cause templates to fail, so we will suppress this
96 # by patching the logging.warning method that is used in that class.
97 class WarningSuppressLogger(object):
98 """ Patch the Logger in SAM """
99
100 def __init__(self, obj_to_patch):
101 self.obj_to_patch = obj_to_patch
102
103 def __enter__(self):
104 self.obj_to_patch.warning = self.warning
105
106 def __exit__(self, exc_type, exc_val, exc_tb):
107 self.obj_to_patch.warning = self.obj_to_patch.warning
108
109 def warning(self, message):
110 """ Ignore warnings from SAM """
111 pass
112
113 with WarningSuppressLogger(parser.logging):
114 self._template = cfnlint.helpers.convert_dict(
115 sam_translator.translate(sam_template=self._template, parameter_values={}))
116 except InvalidDocumentException as e:
117 for cause in e.causes:
118 matches.append(cfnlint.Match(
119 1, 1,
120 1, 1,
121 self._filename, cfnlint.TransformError(), cause.message))
122
123 return matches
124
125 @staticmethod
126 def is_s3_uri(uri):
127 """
128 Checks the uri and determines if it is a valid S3 Uri
129 Parameters
130 ----------
131 uri str, required
132 Uri to check
133 Returns
134 -------
135 bool
136 Returns True if the uri given is an S3 uri, otherwise False
137 """
138 return isinstance(uri, six.string_types) and uri.startswith('s3://')
139
140 @staticmethod
141 def _update_to_s3_uri(property_key, resource_property_dict, s3_uri_value='s3://bucket/value'):
142 """
143 Updates the 'property_key' in the 'resource_property_dict' to the value of 's3_uri_value'
144 Note: The function will mutate the resource_property_dict that is pass in
145 Parameters
146 ----------
147 property_key str, required
148 Key in the resource_property_dict
149 resource_property_dict dict, required
150 Property dictionary of a Resource in the template to replace
151 s3_uri_value str, optional
152 Value to update the value of the property_key to
153 """
154 uri_property = resource_property_dict.get(property_key, '.')
155
156 # ignore if dict or already an S3 Uri
157 if isinstance(uri_property, dict) or Transform.is_s3_uri(uri_property):
158 return
159
160 resource_property_dict[property_key] = s3_uri_value
161
[end of src/cfnlint/transform.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/src/cfnlint/transform.py b/src/cfnlint/transform.py
--- a/src/cfnlint/transform.py
+++ b/src/cfnlint/transform.py
@@ -15,14 +15,14 @@
SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
"""
import os
+import logging
import six
-
from samtranslator.parser import parser
from samtranslator.translator.translator import Translator
from samtranslator.public.exceptions import InvalidDocumentException
import cfnlint.helpers
-
+LOGGER = logging.getLogger('cfnlint')
class Transform(object):
"""
@@ -119,6 +119,14 @@
1, 1,
1, 1,
self._filename, cfnlint.TransformError(), cause.message))
+ except Exception as e: # pylint: disable=W0703
+ LOGGER.debug('Error transforming template: %s', str(e))
+ LOGGER.debug('Stack trace: %s', e, exc_info=True)
+ message = 'Error transforming template: {0}'
+ matches.append(cfnlint.Match(
+ 1, 1,
+ 1, 1,
+ self._filename, cfnlint.TransformError(), message.format(str(e))))
return matches
|
{"golden_diff": "diff --git a/src/cfnlint/transform.py b/src/cfnlint/transform.py\n--- a/src/cfnlint/transform.py\n+++ b/src/cfnlint/transform.py\n@@ -15,14 +15,14 @@\n SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.\n \"\"\"\n import os\n+import logging\n import six\n-\n from samtranslator.parser import parser\n from samtranslator.translator.translator import Translator\n from samtranslator.public.exceptions import InvalidDocumentException\n \n import cfnlint.helpers\n-\n+LOGGER = logging.getLogger('cfnlint')\n \n class Transform(object):\n \"\"\"\n@@ -119,6 +119,14 @@\n 1, 1,\n 1, 1,\n self._filename, cfnlint.TransformError(), cause.message))\n+ except Exception as e: # pylint: disable=W0703\n+ LOGGER.debug('Error transforming template: %s', str(e))\n+ LOGGER.debug('Stack trace: %s', e, exc_info=True)\n+ message = 'Error transforming template: {0}'\n+ matches.append(cfnlint.Match(\n+ 1, 1,\n+ 1, 1,\n+ self._filename, cfnlint.TransformError(), message.format(str(e))))\n \n return matches\n", "issue": "Invalid PrimaryKey attribute crashes cfn-lint instead of raising lint error.\nHi AWS,\r\n\r\nI found a minor bug on cfn-lint version: cfn-lint 0.8.3\r\n\r\nAn attribute in Resource -> Properties-> PrimaryKey leads to crash instead of a raised lint error\r\n\r\nHere is a sample template. If MyInvalidAttr is replaced with a correct attribute eg Name then the template works.\r\n```AWSTemplateFormatVersion: \"2010-09-09\"\r\nTransform: AWS::Serverless-2016-10-31\r\nDescription: >\r\n sam-app\r\n\r\n Sample SAM Template for sam-app\r\n\r\nGlobals:\r\n Function:\r\n Timeout: 3\r\n\r\nResources:\r\n MyTable:\r\n Type: AWS::Serverless::SimpleTable\r\n Properties:\r\n TableName: MyTable\r\n PrimaryKey:\r\n MyInvalidAttr: id\r\n Type: String\r\n```\r\nBelow is the error:\r\n```\r\n$ cfn-lint -t template.yaml\r\nTraceback (most recent call last):\r\n File \"/Users/emilbryggare/Library/Python/2.7/bin/cfn-lint\", line 11, in <module>\r\n sys.exit(main())\r\n File \"/Users/emilbryggare/Library/Python/2.7/lib/python/site-packages/cfnlint/__main__.py\", line 36, in main\r\n args.regions, args.override_spec))\r\n File \"/Users/emilbryggare/Library/Python/2.7/lib/python/site-packages/cfnlint/core.py\", line 46, in run_cli\r\n return run_checks(filename, template, rules, regions)\r\n File \"/Users/emilbryggare/Library/Python/2.7/lib/python/site-packages/cfnlint/core.py\", line 316, in run_checks\r\n matches.extend(runner.transform())\r\n File \"/Users/emilbryggare/Library/Python/2.7/lib/python/site-packages/cfnlint/__init__.py\", line 894, in transform\r\n matches = transform.transform_template()\r\n File \"/Users/emilbryggare/Library/Python/2.7/lib/python/site-packages/cfnlint/transform.py\", line 115, in transform_template\r\n sam_translator.translate(sam_template=self._template, parameter_values={}))\r\n File \"/Users/emilbryggare/Library/Python/2.7/lib/python/site-packages/samtranslator/translator/translator.py\", line 72, in translate\r\n translated = macro.to_cloudformation(**kwargs)\r\n File \"/Users/emilbryggare/Library/Python/2.7/lib/python/site-packages/samtranslator/model/sam_resources.py\", line 537, in to_cloudformation\r\n dynamodb_resources = self._construct_dynamodb_table()\r\n File \"/Users/emilbryggare/Library/Python/2.7/lib/python/site-packages/samtranslator/model/sam_resources.py\", line 546, in _construct_dynamodb_table\r\n 'AttributeName': self.PrimaryKey['Name'],\r\nKeyError: 'Name'\r\n```\r\n\n", "before_files": [{"content": "\"\"\"\n Copyright 2018 Amazon.com, Inc. or its affiliates. All Rights Reserved.\n\n Permission is hereby granted, free of charge, to any person obtaining a copy of this\n software and associated documentation files (the \"Software\"), to deal in the Software\n without restriction, including without limitation the rights to use, copy, modify,\n merge, publish, distribute, sublicense, and/or sell copies of the Software, and to\n permit persons to whom the Software is furnished to do so.\n\n THE SOFTWARE IS PROVIDED \"AS IS\", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED,\n INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A\n PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT\n HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION\n OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE\n SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.\n\"\"\"\nimport os\nimport six\n\nfrom samtranslator.parser import parser\nfrom samtranslator.translator.translator import Translator\nfrom samtranslator.public.exceptions import InvalidDocumentException\n\nimport cfnlint.helpers\n\n\nclass Transform(object):\n \"\"\"\n Application Serverless Module tranform Wrappor. Based on code from AWS SAM CLI:\n https://github.com/awslabs/aws-sam-cli/blob/develop/samcli/commands/validate/lib/sam_template_validator.py\n \"\"\"\n\n def __init__(self, filename, template, region):\n \"\"\"\n Initialize Transform class\n \"\"\"\n self._filename = filename\n self._template = template\n self._region = region\n\n self._managed_policy_map = self.load_managed_policies()\n self._sam_parser = parser.Parser()\n\n def template(self):\n \"\"\"Get the template\"\"\"\n return self._template\n\n def load_managed_policies(self):\n \"\"\"\n Load the ManagedPolicies locally, based on the AWS-CLI:\n https://github.com/awslabs/aws-sam-cli/blob/develop/samcli/lib/samlib/default_managed_policies.json\n \"\"\"\n return cfnlint.helpers.load_resources('data/Serverless/ManagedPolicies.json')\n\n def _replace_local_codeuri(self):\n \"\"\"\n Replaces the CodeUri in AWS::Serverless::Function and DefinitionUri in AWS::Serverless::Api to a fake\n S3 Uri. This is to support running the SAM Translator with valid values for these fields. If this in not done,\n the template is invalid in the eyes of SAM Translator (the translator does not support local paths)\n \"\"\"\n\n all_resources = self._template.get('Resources', {})\n\n for _, resource in all_resources.items():\n\n resource_type = resource.get('Type')\n resource_dict = resource.get('Properties')\n\n if resource_type == 'AWS::Serverless::Function':\n\n Transform._update_to_s3_uri('CodeUri', resource_dict)\n\n if resource_type == 'AWS::Serverless::Api':\n if 'DefinitionBody' not in resource_dict:\n Transform._update_to_s3_uri('DefinitionUri', resource_dict)\n\n def transform_template(self):\n \"\"\"\n Transform the Template using the Serverless Application Model.\n \"\"\"\n matches = []\n\n try:\n sam_translator = Translator(managed_policy_map=self._managed_policy_map,\n sam_parser=self._sam_parser)\n\n self._replace_local_codeuri()\n\n # Tell SAM to use the region we're linting in, this has to be controlled using the default AWS mechanisms, see also:\n # https://github.com/awslabs/serverless-application-model/blob/master/samtranslator/translator/arn_generator.py\n os.environ['AWS_DEFAULT_REGION'] = self._region\n\n # In the Paser class, within the SAM Translator, they log a warning for when the template\n # does not match the schema. The logger they use is the root logger instead of one scoped to\n # their module. Currently this does not cause templates to fail, so we will suppress this\n # by patching the logging.warning method that is used in that class.\n class WarningSuppressLogger(object):\n \"\"\" Patch the Logger in SAM \"\"\"\n\n def __init__(self, obj_to_patch):\n self.obj_to_patch = obj_to_patch\n\n def __enter__(self):\n self.obj_to_patch.warning = self.warning\n\n def __exit__(self, exc_type, exc_val, exc_tb):\n self.obj_to_patch.warning = self.obj_to_patch.warning\n\n def warning(self, message):\n \"\"\" Ignore warnings from SAM \"\"\"\n pass\n\n with WarningSuppressLogger(parser.logging):\n self._template = cfnlint.helpers.convert_dict(\n sam_translator.translate(sam_template=self._template, parameter_values={}))\n except InvalidDocumentException as e:\n for cause in e.causes:\n matches.append(cfnlint.Match(\n 1, 1,\n 1, 1,\n self._filename, cfnlint.TransformError(), cause.message))\n\n return matches\n\n @staticmethod\n def is_s3_uri(uri):\n \"\"\"\n Checks the uri and determines if it is a valid S3 Uri\n Parameters\n ----------\n uri str, required\n Uri to check\n Returns\n -------\n bool\n Returns True if the uri given is an S3 uri, otherwise False\n \"\"\"\n return isinstance(uri, six.string_types) and uri.startswith('s3://')\n\n @staticmethod\n def _update_to_s3_uri(property_key, resource_property_dict, s3_uri_value='s3://bucket/value'):\n \"\"\"\n Updates the 'property_key' in the 'resource_property_dict' to the value of 's3_uri_value'\n Note: The function will mutate the resource_property_dict that is pass in\n Parameters\n ----------\n property_key str, required\n Key in the resource_property_dict\n resource_property_dict dict, required\n Property dictionary of a Resource in the template to replace\n s3_uri_value str, optional\n Value to update the value of the property_key to\n \"\"\"\n uri_property = resource_property_dict.get(property_key, '.')\n\n # ignore if dict or already an S3 Uri\n if isinstance(uri_property, dict) or Transform.is_s3_uri(uri_property):\n return\n\n resource_property_dict[property_key] = s3_uri_value\n", "path": "src/cfnlint/transform.py"}]}
| 2,928 | 281 |
gh_patches_debug_37855
|
rasdani/github-patches
|
git_diff
|
tensorflow__addons-2368
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Lookahead might divide weights with mixed_float16
**System information**
- OS Platform and Distribution (e.g., Linux Ubuntu 16.04): Ubuntu 18.04
- TensorFlow version and how it was installed (source or binary): 2.3.0-rc2 (source)
- TensorFlow-Addons version and how it was installed (source or binary): Just lookahead extracted from the source (master)
- Python version: 3.6.9
- Is GPU used? (yes/no): yes
When tf.keras.mixed_precision.experimental.set_policy('mixed_float16') is set, using LookAhead divides the layers weights after a few gradient updates (for example 7 in my case).
After investigating, I think the reason is that when mixed_float16 is used with automatic loss scaling, the loss scale is initialized very high. Thus the first few gradient updates do nothing (see https://github.com/tensorflow/tensorflow/blob/582c8d236cb079023657287c318ff26adb239002/tensorflow/python/keras/mixed_precision/loss_scale_optimizer.py#L714 for how updates are skipped).
As a result, LookAhead ends up with its 'slow vars' initialized to 0 instead of the original layer weight. Indeed the vars are initialized to the correct value only on the first iteration during the gradient update (https://github.com/tensorflow/addons/blob/master/tensorflow_addons/optimizers/lookahead.py#L106).
Thus when the gradient update finally does something, the first variable update does average them with the slow variables (initialized to zero). Which would explain the weight division I'm seeing.
I don't have a short code to replicate the bug, but I think the above explanation and links should highlight well enough the issue with the current way LookAhead's slow variables are initialized.
Lookahead might divide weights with mixed_float16
**System information**
- OS Platform and Distribution (e.g., Linux Ubuntu 16.04): Ubuntu 18.04
- TensorFlow version and how it was installed (source or binary): 2.3.0-rc2 (source)
- TensorFlow-Addons version and how it was installed (source or binary): Just lookahead extracted from the source (master)
- Python version: 3.6.9
- Is GPU used? (yes/no): yes
When tf.keras.mixed_precision.experimental.set_policy('mixed_float16') is set, using LookAhead divides the layers weights after a few gradient updates (for example 7 in my case).
After investigating, I think the reason is that when mixed_float16 is used with automatic loss scaling, the loss scale is initialized very high. Thus the first few gradient updates do nothing (see https://github.com/tensorflow/tensorflow/blob/582c8d236cb079023657287c318ff26adb239002/tensorflow/python/keras/mixed_precision/loss_scale_optimizer.py#L714 for how updates are skipped).
As a result, LookAhead ends up with its 'slow vars' initialized to 0 instead of the original layer weight. Indeed the vars are initialized to the correct value only on the first iteration during the gradient update (https://github.com/tensorflow/addons/blob/master/tensorflow_addons/optimizers/lookahead.py#L106).
Thus when the gradient update finally does something, the first variable update does average them with the slow variables (initialized to zero). Which would explain the weight division I'm seeing.
I don't have a short code to replicate the bug, but I think the above explanation and links should highlight well enough the issue with the current way LookAhead's slow variables are initialized.
</issue>
<code>
[start of tensorflow_addons/optimizers/lookahead.py]
1 # Copyright 2019 The TensorFlow Authors. All Rights Reserved.
2 #
3 # Licensed under the Apache License, Version 2.0 (the "License");
4 # you may not use this file except in compliance with the License.
5 # You may obtain a copy of the License at
6 #
7 # http://www.apache.org/licenses/LICENSE-2.0
8 #
9 # Unless required by applicable law or agreed to in writing, software
10 # distributed under the License is distributed on an "AS IS" BASIS,
11 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12 # See the License for the specific language governing permissions and
13 # limitations under the License.
14 # ==============================================================================
15
16 import tensorflow as tf
17 from tensorflow_addons.utils import types
18
19 from typeguard import typechecked
20
21
22 @tf.keras.utils.register_keras_serializable(package="Addons")
23 class Lookahead(tf.keras.optimizers.Optimizer):
24 """This class allows to extend optimizers with the lookahead mechanism.
25
26 The mechanism is proposed by Michael R. Zhang et.al in the paper
27 [Lookahead Optimizer: k steps forward, 1 step back]
28 (https://arxiv.org/abs/1907.08610v1). The optimizer iteratively updates two
29 sets of weights: the search directions for weights are chosen by the inner
30 optimizer, while the "slow weights" are updated each `k` steps based on the
31 directions of the "fast weights" and the two sets of weights are
32 synchronized. This method improves the learning stability and lowers the
33 variance of its inner optimizer.
34
35 Example of usage:
36
37 ```python
38 opt = tf.keras.optimizers.SGD(learning_rate)
39 opt = tfa.optimizers.Lookahead(opt)
40 ```
41 """
42
43 @typechecked
44 def __init__(
45 self,
46 optimizer: types.Optimizer,
47 sync_period: int = 6,
48 slow_step_size: types.FloatTensorLike = 0.5,
49 name: str = "Lookahead",
50 **kwargs,
51 ):
52 r"""Wrap optimizer with the lookahead mechanism.
53
54 Args:
55 optimizer: The original optimizer that will be used to compute
56 and apply the gradients.
57 sync_period: An integer. The synchronization period of lookahead.
58 Enable lookahead mechanism by setting it with a positive value.
59 slow_step_size: A floating point value.
60 The ratio for updating the slow weights.
61 name: Optional name for the operations created when applying
62 gradients. Defaults to "Lookahead".
63 **kwargs: keyword arguments. Allowed to be {`clipnorm`,
64 `clipvalue`, `lr`, `decay`}. `clipnorm` is clip gradients
65 by norm; `clipvalue` is clip gradients by value, `decay` is
66 included for backward compatibility to allow time inverse
67 decay of learning rate. `lr` is included for backward
68 compatibility, recommended to use `learning_rate` instead.
69 """
70 super().__init__(name, **kwargs)
71
72 if isinstance(optimizer, str):
73 optimizer = tf.keras.optimizers.get(optimizer)
74 if not isinstance(optimizer, tf.keras.optimizers.Optimizer):
75 raise TypeError(
76 "optimizer is not an object of tf.keras.optimizers.Optimizer"
77 )
78
79 self._optimizer = optimizer
80 self._set_hyper("sync_period", sync_period)
81 self._set_hyper("slow_step_size", slow_step_size)
82 self._initialized = False
83 self._track_trackable(self._optimizer, "lh_base_optimizer")
84
85 def _create_slots(self, var_list):
86 self._optimizer._create_slots(
87 var_list=var_list
88 ) # pylint: disable=protected-access
89 for var in var_list:
90 self.add_slot(var, "slow")
91
92 def _create_hypers(self):
93 self._optimizer._create_hypers() # pylint: disable=protected-access
94
95 def _prepare(self, var_list):
96 return self._optimizer._prepare(
97 var_list=var_list
98 ) # pylint: disable=protected-access
99
100 def apply_gradients(self, grads_and_vars, name=None, **kwargs):
101 self._optimizer._iterations = (
102 self.iterations
103 ) # pylint: disable=protected-access
104 return super().apply_gradients(grads_and_vars, name, **kwargs)
105
106 def _init_op(self, var):
107 slow_var = self.get_slot(var, "slow")
108 return slow_var.assign(
109 tf.where(
110 tf.equal(self.iterations, tf.constant(0, dtype=self.iterations.dtype)),
111 var,
112 slow_var,
113 ),
114 use_locking=self._use_locking,
115 )
116
117 def _look_ahead_op(self, var):
118 var_dtype = var.dtype.base_dtype
119 slow_var = self.get_slot(var, "slow")
120 local_step = tf.cast(self.iterations + 1, tf.dtypes.int64)
121 sync_period = self._get_hyper("sync_period", tf.dtypes.int64)
122 slow_step_size = self._get_hyper("slow_step_size", var_dtype)
123 step_back = slow_var + slow_step_size * (var - slow_var)
124 sync_cond = tf.equal(
125 tf.math.floordiv(local_step, sync_period) * sync_period, local_step
126 )
127 with tf.control_dependencies([step_back]):
128 slow_update = slow_var.assign(
129 tf.where(sync_cond, step_back, slow_var), use_locking=self._use_locking
130 )
131 var_update = var.assign(
132 tf.where(sync_cond, step_back, var), use_locking=self._use_locking
133 )
134 return tf.group(slow_update, var_update)
135
136 @property
137 def weights(self):
138 return self._weights + self._optimizer.weights
139
140 def _resource_apply_dense(self, grad, var):
141 init_op = self._init_op(var)
142 with tf.control_dependencies([init_op]):
143 train_op = self._optimizer._resource_apply_dense(
144 grad, var
145 ) # pylint: disable=protected-access
146 with tf.control_dependencies([train_op]):
147 look_ahead_op = self._look_ahead_op(var)
148 return tf.group(init_op, train_op, look_ahead_op)
149
150 def _resource_apply_sparse(self, grad, var, indices):
151 init_op = self._init_op(var)
152 with tf.control_dependencies([init_op]):
153 train_op = self._optimizer._resource_apply_sparse( # pylint: disable=protected-access
154 grad, var, indices
155 )
156 with tf.control_dependencies([train_op]):
157 look_ahead_op = self._look_ahead_op(var)
158 return tf.group(init_op, train_op, look_ahead_op)
159
160 def get_config(self):
161 config = {
162 "optimizer": tf.keras.optimizers.serialize(self._optimizer),
163 "sync_period": self._serialize_hyperparameter("sync_period"),
164 "slow_step_size": self._serialize_hyperparameter("slow_step_size"),
165 }
166 base_config = super().get_config()
167 return {**base_config, **config}
168
169 @property
170 def learning_rate(self):
171 return self._optimizer._get_hyper("learning_rate")
172
173 @learning_rate.setter
174 def learning_rate(self, learning_rate):
175 self._optimizer._set_hyper("learning_rate", learning_rate)
176
177 @property
178 def lr(self):
179 return self.learning_rate
180
181 @lr.setter
182 def lr(self, lr):
183 self.learning_rate = lr
184
185 @classmethod
186 def from_config(cls, config, custom_objects=None):
187 optimizer = tf.keras.optimizers.deserialize(
188 config.pop("optimizer"), custom_objects=custom_objects
189 )
190 return cls(optimizer, **config)
191
[end of tensorflow_addons/optimizers/lookahead.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/tensorflow_addons/optimizers/lookahead.py b/tensorflow_addons/optimizers/lookahead.py
--- a/tensorflow_addons/optimizers/lookahead.py
+++ b/tensorflow_addons/optimizers/lookahead.py
@@ -87,7 +87,7 @@
var_list=var_list
) # pylint: disable=protected-access
for var in var_list:
- self.add_slot(var, "slow")
+ self.add_slot(var, "slow", initializer=var)
def _create_hypers(self):
self._optimizer._create_hypers() # pylint: disable=protected-access
@@ -103,17 +103,6 @@
) # pylint: disable=protected-access
return super().apply_gradients(grads_and_vars, name, **kwargs)
- def _init_op(self, var):
- slow_var = self.get_slot(var, "slow")
- return slow_var.assign(
- tf.where(
- tf.equal(self.iterations, tf.constant(0, dtype=self.iterations.dtype)),
- var,
- slow_var,
- ),
- use_locking=self._use_locking,
- )
-
def _look_ahead_op(self, var):
var_dtype = var.dtype.base_dtype
slow_var = self.get_slot(var, "slow")
@@ -138,24 +127,22 @@
return self._weights + self._optimizer.weights
def _resource_apply_dense(self, grad, var):
- init_op = self._init_op(var)
- with tf.control_dependencies([init_op]):
- train_op = self._optimizer._resource_apply_dense(
- grad, var
- ) # pylint: disable=protected-access
- with tf.control_dependencies([train_op]):
- look_ahead_op = self._look_ahead_op(var)
- return tf.group(init_op, train_op, look_ahead_op)
+ train_op = self._optimizer._resource_apply_dense(
+ grad, var
+ ) # pylint: disable=protected-access
+ with tf.control_dependencies([train_op]):
+ look_ahead_op = self._look_ahead_op(var)
+ return tf.group(train_op, look_ahead_op)
def _resource_apply_sparse(self, grad, var, indices):
- init_op = self._init_op(var)
- with tf.control_dependencies([init_op]):
- train_op = self._optimizer._resource_apply_sparse( # pylint: disable=protected-access
+ train_op = (
+ self._optimizer._resource_apply_sparse( # pylint: disable=protected-access
grad, var, indices
)
- with tf.control_dependencies([train_op]):
- look_ahead_op = self._look_ahead_op(var)
- return tf.group(init_op, train_op, look_ahead_op)
+ )
+ with tf.control_dependencies([train_op]):
+ look_ahead_op = self._look_ahead_op(var)
+ return tf.group(train_op, look_ahead_op)
def get_config(self):
config = {
|
{"golden_diff": "diff --git a/tensorflow_addons/optimizers/lookahead.py b/tensorflow_addons/optimizers/lookahead.py\n--- a/tensorflow_addons/optimizers/lookahead.py\n+++ b/tensorflow_addons/optimizers/lookahead.py\n@@ -87,7 +87,7 @@\n var_list=var_list\n ) # pylint: disable=protected-access\n for var in var_list:\n- self.add_slot(var, \"slow\")\n+ self.add_slot(var, \"slow\", initializer=var)\n \n def _create_hypers(self):\n self._optimizer._create_hypers() # pylint: disable=protected-access\n@@ -103,17 +103,6 @@\n ) # pylint: disable=protected-access\n return super().apply_gradients(grads_and_vars, name, **kwargs)\n \n- def _init_op(self, var):\n- slow_var = self.get_slot(var, \"slow\")\n- return slow_var.assign(\n- tf.where(\n- tf.equal(self.iterations, tf.constant(0, dtype=self.iterations.dtype)),\n- var,\n- slow_var,\n- ),\n- use_locking=self._use_locking,\n- )\n-\n def _look_ahead_op(self, var):\n var_dtype = var.dtype.base_dtype\n slow_var = self.get_slot(var, \"slow\")\n@@ -138,24 +127,22 @@\n return self._weights + self._optimizer.weights\n \n def _resource_apply_dense(self, grad, var):\n- init_op = self._init_op(var)\n- with tf.control_dependencies([init_op]):\n- train_op = self._optimizer._resource_apply_dense(\n- grad, var\n- ) # pylint: disable=protected-access\n- with tf.control_dependencies([train_op]):\n- look_ahead_op = self._look_ahead_op(var)\n- return tf.group(init_op, train_op, look_ahead_op)\n+ train_op = self._optimizer._resource_apply_dense(\n+ grad, var\n+ ) # pylint: disable=protected-access\n+ with tf.control_dependencies([train_op]):\n+ look_ahead_op = self._look_ahead_op(var)\n+ return tf.group(train_op, look_ahead_op)\n \n def _resource_apply_sparse(self, grad, var, indices):\n- init_op = self._init_op(var)\n- with tf.control_dependencies([init_op]):\n- train_op = self._optimizer._resource_apply_sparse( # pylint: disable=protected-access\n+ train_op = (\n+ self._optimizer._resource_apply_sparse( # pylint: disable=protected-access\n grad, var, indices\n )\n- with tf.control_dependencies([train_op]):\n- look_ahead_op = self._look_ahead_op(var)\n- return tf.group(init_op, train_op, look_ahead_op)\n+ )\n+ with tf.control_dependencies([train_op]):\n+ look_ahead_op = self._look_ahead_op(var)\n+ return tf.group(train_op, look_ahead_op)\n \n def get_config(self):\n config = {\n", "issue": "Lookahead might divide weights with mixed_float16\n**System information**\r\n- OS Platform and Distribution (e.g., Linux Ubuntu 16.04): Ubuntu 18.04\r\n- TensorFlow version and how it was installed (source or binary): 2.3.0-rc2 (source)\r\n- TensorFlow-Addons version and how it was installed (source or binary): Just lookahead extracted from the source (master)\r\n- Python version: 3.6.9\r\n- Is GPU used? (yes/no): yes\r\n\r\nWhen tf.keras.mixed_precision.experimental.set_policy('mixed_float16') is set, using LookAhead divides the layers weights after a few gradient updates (for example 7 in my case).\r\n\r\nAfter investigating, I think the reason is that when mixed_float16 is used with automatic loss scaling, the loss scale is initialized very high. Thus the first few gradient updates do nothing (see https://github.com/tensorflow/tensorflow/blob/582c8d236cb079023657287c318ff26adb239002/tensorflow/python/keras/mixed_precision/loss_scale_optimizer.py#L714 for how updates are skipped).\r\n\r\nAs a result, LookAhead ends up with its 'slow vars' initialized to 0 instead of the original layer weight. Indeed the vars are initialized to the correct value only on the first iteration during the gradient update (https://github.com/tensorflow/addons/blob/master/tensorflow_addons/optimizers/lookahead.py#L106).\r\n\r\nThus when the gradient update finally does something, the first variable update does average them with the slow variables (initialized to zero). Which would explain the weight division I'm seeing.\r\n\r\nI don't have a short code to replicate the bug, but I think the above explanation and links should highlight well enough the issue with the current way LookAhead's slow variables are initialized.\r\n\nLookahead might divide weights with mixed_float16\n**System information**\r\n- OS Platform and Distribution (e.g., Linux Ubuntu 16.04): Ubuntu 18.04\r\n- TensorFlow version and how it was installed (source or binary): 2.3.0-rc2 (source)\r\n- TensorFlow-Addons version and how it was installed (source or binary): Just lookahead extracted from the source (master)\r\n- Python version: 3.6.9\r\n- Is GPU used? (yes/no): yes\r\n\r\nWhen tf.keras.mixed_precision.experimental.set_policy('mixed_float16') is set, using LookAhead divides the layers weights after a few gradient updates (for example 7 in my case).\r\n\r\nAfter investigating, I think the reason is that when mixed_float16 is used with automatic loss scaling, the loss scale is initialized very high. Thus the first few gradient updates do nothing (see https://github.com/tensorflow/tensorflow/blob/582c8d236cb079023657287c318ff26adb239002/tensorflow/python/keras/mixed_precision/loss_scale_optimizer.py#L714 for how updates are skipped).\r\n\r\nAs a result, LookAhead ends up with its 'slow vars' initialized to 0 instead of the original layer weight. Indeed the vars are initialized to the correct value only on the first iteration during the gradient update (https://github.com/tensorflow/addons/blob/master/tensorflow_addons/optimizers/lookahead.py#L106).\r\n\r\nThus when the gradient update finally does something, the first variable update does average them with the slow variables (initialized to zero). Which would explain the weight division I'm seeing.\r\n\r\nI don't have a short code to replicate the bug, but I think the above explanation and links should highlight well enough the issue with the current way LookAhead's slow variables are initialized.\r\n\n", "before_files": [{"content": "# Copyright 2019 The TensorFlow Authors. All Rights Reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n# ==============================================================================\n\nimport tensorflow as tf\nfrom tensorflow_addons.utils import types\n\nfrom typeguard import typechecked\n\n\[email protected]_keras_serializable(package=\"Addons\")\nclass Lookahead(tf.keras.optimizers.Optimizer):\n \"\"\"This class allows to extend optimizers with the lookahead mechanism.\n\n The mechanism is proposed by Michael R. Zhang et.al in the paper\n [Lookahead Optimizer: k steps forward, 1 step back]\n (https://arxiv.org/abs/1907.08610v1). The optimizer iteratively updates two\n sets of weights: the search directions for weights are chosen by the inner\n optimizer, while the \"slow weights\" are updated each `k` steps based on the\n directions of the \"fast weights\" and the two sets of weights are\n synchronized. This method improves the learning stability and lowers the\n variance of its inner optimizer.\n\n Example of usage:\n\n ```python\n opt = tf.keras.optimizers.SGD(learning_rate)\n opt = tfa.optimizers.Lookahead(opt)\n ```\n \"\"\"\n\n @typechecked\n def __init__(\n self,\n optimizer: types.Optimizer,\n sync_period: int = 6,\n slow_step_size: types.FloatTensorLike = 0.5,\n name: str = \"Lookahead\",\n **kwargs,\n ):\n r\"\"\"Wrap optimizer with the lookahead mechanism.\n\n Args:\n optimizer: The original optimizer that will be used to compute\n and apply the gradients.\n sync_period: An integer. The synchronization period of lookahead.\n Enable lookahead mechanism by setting it with a positive value.\n slow_step_size: A floating point value.\n The ratio for updating the slow weights.\n name: Optional name for the operations created when applying\n gradients. Defaults to \"Lookahead\".\n **kwargs: keyword arguments. Allowed to be {`clipnorm`,\n `clipvalue`, `lr`, `decay`}. `clipnorm` is clip gradients\n by norm; `clipvalue` is clip gradients by value, `decay` is\n included for backward compatibility to allow time inverse\n decay of learning rate. `lr` is included for backward\n compatibility, recommended to use `learning_rate` instead.\n \"\"\"\n super().__init__(name, **kwargs)\n\n if isinstance(optimizer, str):\n optimizer = tf.keras.optimizers.get(optimizer)\n if not isinstance(optimizer, tf.keras.optimizers.Optimizer):\n raise TypeError(\n \"optimizer is not an object of tf.keras.optimizers.Optimizer\"\n )\n\n self._optimizer = optimizer\n self._set_hyper(\"sync_period\", sync_period)\n self._set_hyper(\"slow_step_size\", slow_step_size)\n self._initialized = False\n self._track_trackable(self._optimizer, \"lh_base_optimizer\")\n\n def _create_slots(self, var_list):\n self._optimizer._create_slots(\n var_list=var_list\n ) # pylint: disable=protected-access\n for var in var_list:\n self.add_slot(var, \"slow\")\n\n def _create_hypers(self):\n self._optimizer._create_hypers() # pylint: disable=protected-access\n\n def _prepare(self, var_list):\n return self._optimizer._prepare(\n var_list=var_list\n ) # pylint: disable=protected-access\n\n def apply_gradients(self, grads_and_vars, name=None, **kwargs):\n self._optimizer._iterations = (\n self.iterations\n ) # pylint: disable=protected-access\n return super().apply_gradients(grads_and_vars, name, **kwargs)\n\n def _init_op(self, var):\n slow_var = self.get_slot(var, \"slow\")\n return slow_var.assign(\n tf.where(\n tf.equal(self.iterations, tf.constant(0, dtype=self.iterations.dtype)),\n var,\n slow_var,\n ),\n use_locking=self._use_locking,\n )\n\n def _look_ahead_op(self, var):\n var_dtype = var.dtype.base_dtype\n slow_var = self.get_slot(var, \"slow\")\n local_step = tf.cast(self.iterations + 1, tf.dtypes.int64)\n sync_period = self._get_hyper(\"sync_period\", tf.dtypes.int64)\n slow_step_size = self._get_hyper(\"slow_step_size\", var_dtype)\n step_back = slow_var + slow_step_size * (var - slow_var)\n sync_cond = tf.equal(\n tf.math.floordiv(local_step, sync_period) * sync_period, local_step\n )\n with tf.control_dependencies([step_back]):\n slow_update = slow_var.assign(\n tf.where(sync_cond, step_back, slow_var), use_locking=self._use_locking\n )\n var_update = var.assign(\n tf.where(sync_cond, step_back, var), use_locking=self._use_locking\n )\n return tf.group(slow_update, var_update)\n\n @property\n def weights(self):\n return self._weights + self._optimizer.weights\n\n def _resource_apply_dense(self, grad, var):\n init_op = self._init_op(var)\n with tf.control_dependencies([init_op]):\n train_op = self._optimizer._resource_apply_dense(\n grad, var\n ) # pylint: disable=protected-access\n with tf.control_dependencies([train_op]):\n look_ahead_op = self._look_ahead_op(var)\n return tf.group(init_op, train_op, look_ahead_op)\n\n def _resource_apply_sparse(self, grad, var, indices):\n init_op = self._init_op(var)\n with tf.control_dependencies([init_op]):\n train_op = self._optimizer._resource_apply_sparse( # pylint: disable=protected-access\n grad, var, indices\n )\n with tf.control_dependencies([train_op]):\n look_ahead_op = self._look_ahead_op(var)\n return tf.group(init_op, train_op, look_ahead_op)\n\n def get_config(self):\n config = {\n \"optimizer\": tf.keras.optimizers.serialize(self._optimizer),\n \"sync_period\": self._serialize_hyperparameter(\"sync_period\"),\n \"slow_step_size\": self._serialize_hyperparameter(\"slow_step_size\"),\n }\n base_config = super().get_config()\n return {**base_config, **config}\n\n @property\n def learning_rate(self):\n return self._optimizer._get_hyper(\"learning_rate\")\n\n @learning_rate.setter\n def learning_rate(self, learning_rate):\n self._optimizer._set_hyper(\"learning_rate\", learning_rate)\n\n @property\n def lr(self):\n return self.learning_rate\n\n @lr.setter\n def lr(self, lr):\n self.learning_rate = lr\n\n @classmethod\n def from_config(cls, config, custom_objects=None):\n optimizer = tf.keras.optimizers.deserialize(\n config.pop(\"optimizer\"), custom_objects=custom_objects\n )\n return cls(optimizer, **config)\n", "path": "tensorflow_addons/optimizers/lookahead.py"}]}
| 3,480 | 692 |
gh_patches_debug_13019
|
rasdani/github-patches
|
git_diff
|
getsentry__sentry-52100
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
SDK Crash Detection: Store Project ID and Event ID
Store project ID and event ID in the SDK crash detection context to find the original SDK crash event, which is only possible with admin Sentry rights.
https://github.com/getsentry/sentry/blob/2c31ee009b44964f78b9e7e8282e602b7ef849b0/src/sentry/utils/sdk_crashes/sdk_crash_detection.py#L40C2-L42
</issue>
<code>
[start of src/sentry/utils/sdk_crashes/sdk_crash_detection.py]
1 from __future__ import annotations
2
3 from typing import Any, Mapping, Optional
4
5 from sentry.eventstore.models import Event
6 from sentry.issues.grouptype import GroupCategory
7 from sentry.utils.safe import get_path, set_path
8 from sentry.utils.sdk_crashes.cocoa_sdk_crash_detector import CocoaSDKCrashDetector
9 from sentry.utils.sdk_crashes.event_stripper import strip_event_data
10 from sentry.utils.sdk_crashes.sdk_crash_detector import SDKCrashDetector
11
12
13 class SDKCrashReporter:
14 def report(self, event_data: Mapping[str, Any], event_project_id: int) -> Event:
15 from sentry.event_manager import EventManager
16
17 manager = EventManager(dict(event_data))
18 manager.normalize()
19 return manager.save(project_id=event_project_id)
20
21
22 class SDKCrashDetection:
23 def __init__(
24 self,
25 sdk_crash_reporter: SDKCrashReporter,
26 sdk_crash_detector: SDKCrashDetector,
27 ):
28 self.sdk_crash_reporter = sdk_crash_reporter
29 self.cocoa_sdk_crash_detector = sdk_crash_detector
30
31 def detect_sdk_crash(self, event: Event, event_project_id: int) -> Optional[Event]:
32 should_detect_sdk_crash = (
33 event.group
34 and event.group.issue_category == GroupCategory.ERROR
35 and event.group.platform == "cocoa"
36 )
37 if not should_detect_sdk_crash:
38 return None
39
40 context = get_path(event.data, "contexts", "sdk_crash_detection")
41 if context is not None and context.get("detected", False):
42 return None
43
44 # Getting the frames and checking if the event is unhandled might different per platform.
45 # We will change this once we implement this for more platforms.
46 is_unhandled = (
47 get_path(event.data, "exception", "values", -1, "mechanism", "handled") is False
48 )
49 if is_unhandled is False:
50 return None
51
52 frames = get_path(event.data, "exception", "values", -1, "stacktrace", "frames")
53 if not frames:
54 return None
55
56 if self.cocoa_sdk_crash_detector.is_sdk_crash(frames):
57 sdk_crash_event_data = strip_event_data(event.data, self.cocoa_sdk_crash_detector)
58
59 set_path(
60 sdk_crash_event_data, "contexts", "sdk_crash_detection", value={"detected": True}
61 )
62
63 return self.sdk_crash_reporter.report(sdk_crash_event_data, event_project_id)
64
65 return None
66
67
68 _crash_reporter = SDKCrashReporter()
69 _cocoa_sdk_crash_detector = CocoaSDKCrashDetector()
70
71 sdk_crash_detection = SDKCrashDetection(_crash_reporter, _cocoa_sdk_crash_detector)
72
[end of src/sentry/utils/sdk_crashes/sdk_crash_detection.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/src/sentry/utils/sdk_crashes/sdk_crash_detection.py b/src/sentry/utils/sdk_crashes/sdk_crash_detection.py
--- a/src/sentry/utils/sdk_crashes/sdk_crash_detection.py
+++ b/src/sentry/utils/sdk_crashes/sdk_crash_detection.py
@@ -57,7 +57,14 @@
sdk_crash_event_data = strip_event_data(event.data, self.cocoa_sdk_crash_detector)
set_path(
- sdk_crash_event_data, "contexts", "sdk_crash_detection", value={"detected": True}
+ sdk_crash_event_data,
+ "contexts",
+ "sdk_crash_detection",
+ value={
+ "detected": True,
+ "original_project_id": event.project.id,
+ "original_event_id": event.event_id,
+ },
)
return self.sdk_crash_reporter.report(sdk_crash_event_data, event_project_id)
|
{"golden_diff": "diff --git a/src/sentry/utils/sdk_crashes/sdk_crash_detection.py b/src/sentry/utils/sdk_crashes/sdk_crash_detection.py\n--- a/src/sentry/utils/sdk_crashes/sdk_crash_detection.py\n+++ b/src/sentry/utils/sdk_crashes/sdk_crash_detection.py\n@@ -57,7 +57,14 @@\n sdk_crash_event_data = strip_event_data(event.data, self.cocoa_sdk_crash_detector)\n \n set_path(\n- sdk_crash_event_data, \"contexts\", \"sdk_crash_detection\", value={\"detected\": True}\n+ sdk_crash_event_data,\n+ \"contexts\",\n+ \"sdk_crash_detection\",\n+ value={\n+ \"detected\": True,\n+ \"original_project_id\": event.project.id,\n+ \"original_event_id\": event.event_id,\n+ },\n )\n \n return self.sdk_crash_reporter.report(sdk_crash_event_data, event_project_id)\n", "issue": "SDK Crash Detection: Store Project ID and Event ID\nStore project ID and event ID in the SDK crash detection context to find the original SDK crash event, which is only possible with admin Sentry rights.\r\n\r\nhttps://github.com/getsentry/sentry/blob/2c31ee009b44964f78b9e7e8282e602b7ef849b0/src/sentry/utils/sdk_crashes/sdk_crash_detection.py#L40C2-L42\n", "before_files": [{"content": "from __future__ import annotations\n\nfrom typing import Any, Mapping, Optional\n\nfrom sentry.eventstore.models import Event\nfrom sentry.issues.grouptype import GroupCategory\nfrom sentry.utils.safe import get_path, set_path\nfrom sentry.utils.sdk_crashes.cocoa_sdk_crash_detector import CocoaSDKCrashDetector\nfrom sentry.utils.sdk_crashes.event_stripper import strip_event_data\nfrom sentry.utils.sdk_crashes.sdk_crash_detector import SDKCrashDetector\n\n\nclass SDKCrashReporter:\n def report(self, event_data: Mapping[str, Any], event_project_id: int) -> Event:\n from sentry.event_manager import EventManager\n\n manager = EventManager(dict(event_data))\n manager.normalize()\n return manager.save(project_id=event_project_id)\n\n\nclass SDKCrashDetection:\n def __init__(\n self,\n sdk_crash_reporter: SDKCrashReporter,\n sdk_crash_detector: SDKCrashDetector,\n ):\n self.sdk_crash_reporter = sdk_crash_reporter\n self.cocoa_sdk_crash_detector = sdk_crash_detector\n\n def detect_sdk_crash(self, event: Event, event_project_id: int) -> Optional[Event]:\n should_detect_sdk_crash = (\n event.group\n and event.group.issue_category == GroupCategory.ERROR\n and event.group.platform == \"cocoa\"\n )\n if not should_detect_sdk_crash:\n return None\n\n context = get_path(event.data, \"contexts\", \"sdk_crash_detection\")\n if context is not None and context.get(\"detected\", False):\n return None\n\n # Getting the frames and checking if the event is unhandled might different per platform.\n # We will change this once we implement this for more platforms.\n is_unhandled = (\n get_path(event.data, \"exception\", \"values\", -1, \"mechanism\", \"handled\") is False\n )\n if is_unhandled is False:\n return None\n\n frames = get_path(event.data, \"exception\", \"values\", -1, \"stacktrace\", \"frames\")\n if not frames:\n return None\n\n if self.cocoa_sdk_crash_detector.is_sdk_crash(frames):\n sdk_crash_event_data = strip_event_data(event.data, self.cocoa_sdk_crash_detector)\n\n set_path(\n sdk_crash_event_data, \"contexts\", \"sdk_crash_detection\", value={\"detected\": True}\n )\n\n return self.sdk_crash_reporter.report(sdk_crash_event_data, event_project_id)\n\n return None\n\n\n_crash_reporter = SDKCrashReporter()\n_cocoa_sdk_crash_detector = CocoaSDKCrashDetector()\n\nsdk_crash_detection = SDKCrashDetection(_crash_reporter, _cocoa_sdk_crash_detector)\n", "path": "src/sentry/utils/sdk_crashes/sdk_crash_detection.py"}]}
| 1,397 | 204 |
gh_patches_debug_12199
|
rasdani/github-patches
|
git_diff
|
Cloud-CV__EvalAI-475
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Participant teams not showing up
The participant teams to which the user has been invited does not show up if that user logs in.
For instance: User A creates a team ABC and invites User B
User B logs in, and goes to the participant teams page.
User B does not see team ABC.
</issue>
<code>
[start of apps/participants/views.py]
1 from rest_framework import permissions, status
2 from rest_framework.decorators import (api_view,
3 authentication_classes,
4 permission_classes,
5 throttle_classes,)
6 from rest_framework.response import Response
7 from rest_framework_expiring_authtoken.authentication import (
8 ExpiringTokenAuthentication,)
9 from rest_framework.throttling import UserRateThrottle
10
11 from accounts.permissions import HasVerifiedEmail
12 from base.utils import paginated_queryset
13 from challenges.models import Challenge
14
15 from .models import (Participant, ParticipantTeam)
16 from .serializers import (InviteParticipantToTeamSerializer,
17 ParticipantTeamSerializer,
18 ChallengeParticipantTeam,
19 ChallengeParticipantTeamList,
20 ChallengeParticipantTeamListSerializer,)
21
22
23 @throttle_classes([UserRateThrottle])
24 @api_view(['GET', 'POST'])
25 @permission_classes((permissions.IsAuthenticated, HasVerifiedEmail))
26 @authentication_classes((ExpiringTokenAuthentication,))
27 def participant_team_list(request):
28
29 if request.method == 'GET':
30 participant_teams = ParticipantTeam.objects.filter(
31 created_by=request.user)
32 paginator, result_page = paginated_queryset(participant_teams, request)
33 serializer = ParticipantTeamSerializer(result_page, many=True)
34 response_data = serializer.data
35 return paginator.get_paginated_response(response_data)
36
37 elif request.method == 'POST':
38 serializer = ParticipantTeamSerializer(data=request.data,
39 context={'request': request})
40 if serializer.is_valid():
41 serializer.save()
42 response_data = serializer.data
43 participant_team = serializer.instance
44 participant = Participant(user=request.user,
45 status=Participant.SELF,
46 team=participant_team)
47 participant.save()
48 return Response(response_data, status=status.HTTP_201_CREATED)
49 return Response(serializer.errors, status=status.HTTP_400_BAD_REQUEST)
50
51
52 @throttle_classes([UserRateThrottle])
53 @api_view(['GET', 'PUT', 'PATCH', 'DELETE'])
54 @permission_classes((permissions.IsAuthenticated, HasVerifiedEmail))
55 @authentication_classes((ExpiringTokenAuthentication,))
56 def participant_team_detail(request, pk):
57
58 try:
59 participant_team = ParticipantTeam.objects.get(pk=pk)
60 except ParticipantTeam.DoesNotExist:
61 response_data = {'error': 'ParticipantTeam does not exist'}
62 return Response(response_data, status=status.HTTP_406_NOT_ACCEPTABLE)
63
64 if request.method == 'GET':
65 serializer = ParticipantTeamSerializer(participant_team)
66 response_data = serializer.data
67 return Response(response_data, status=status.HTTP_200_OK)
68
69 elif request.method in ['PUT', 'PATCH']:
70
71 if request.method == 'PATCH':
72 serializer = ParticipantTeamSerializer(participant_team, data=request.data,
73 context={
74 'request': request},
75 partial=True)
76 else:
77 serializer = ParticipantTeamSerializer(participant_team, data=request.data,
78 context={'request': request})
79 if serializer.is_valid():
80 serializer.save()
81 response_data = serializer.data
82 return Response(response_data, status=status.HTTP_200_OK)
83 else:
84 return Response(serializer.errors, status=status.HTTP_400_BAD_REQUEST)
85
86 elif request.method == 'DELETE':
87 participant_team.delete()
88 return Response(status=status.HTTP_204_NO_CONTENT)
89
90
91 @throttle_classes([UserRateThrottle])
92 @api_view(['POST'])
93 @permission_classes((permissions.IsAuthenticated, HasVerifiedEmail))
94 @authentication_classes((ExpiringTokenAuthentication,))
95 def invite_participant_to_team(request, pk):
96
97 try:
98 participant_team = ParticipantTeam.objects.get(pk=pk)
99 except ParticipantTeam.DoesNotExist:
100 response_data = {'error': 'ParticipantTeam does not exist'}
101 return Response(response_data, status=status.HTTP_406_NOT_ACCEPTABLE)
102
103 serializer = InviteParticipantToTeamSerializer(data=request.data,
104 context={'participant_team': participant_team,
105 'request': request})
106 if serializer.is_valid():
107 serializer.save()
108 response_data = {
109 'message': 'User has been added successfully to the team'}
110 return Response(response_data, status=status.HTTP_202_ACCEPTED)
111 return Response(serializer.errors, status=status.HTTP_400_BAD_REQUEST)
112
113
114 @throttle_classes([UserRateThrottle])
115 @api_view(['DELETE'])
116 @permission_classes((permissions.IsAuthenticated, HasVerifiedEmail))
117 @authentication_classes((ExpiringTokenAuthentication,))
118 def delete_participant_from_team(request, participant_team_pk, participant_pk):
119 """
120 Deletes a participant from a Participant Team
121 """
122 try:
123 participant_team = ParticipantTeam.objects.get(pk=participant_team_pk)
124 except ParticipantTeam.DoesNotExist:
125 response_data = {'error': 'ParticipantTeam does not exist'}
126 return Response(response_data, status=status.HTTP_406_NOT_ACCEPTABLE)
127
128 try:
129 participant = Participant.objects.get(pk=participant_pk)
130 except Participant.DoesNotExist:
131 response_data = {'error': 'Participant does not exist'}
132 return Response(response_data, status=status.HTTP_406_NOT_ACCEPTABLE)
133
134 if participant_team.created_by == request.user:
135
136 if participant.user == request.user: # when the user tries to remove himself
137 response_data = {
138 'error': 'You are not allowed to remove yourself since you are admin. Please delete the team if you want to do so!'} # noqa: ignore=E501
139 return Response(response_data, status=status.HTTP_406_NOT_ACCEPTABLE)
140 else:
141 participant.delete()
142 return Response(status=status.HTTP_204_NO_CONTENT)
143 else:
144 response_data = {
145 'error': 'Sorry, you do not have permissions to remove this participant'}
146 return Response(response_data, status=status.HTTP_401_UNAUTHORIZED)
147
148
149 @throttle_classes([UserRateThrottle])
150 @api_view(['GET', ])
151 @permission_classes((permissions.IsAuthenticated, HasVerifiedEmail))
152 @authentication_classes((ExpiringTokenAuthentication,))
153 def get_teams_and_corresponding_challenges_for_a_participant(request):
154 """
155 Returns list of teams and corresponding challenges for a participant
156 """
157 # first get list of all the participants and teams related to the user
158 participant_objs = Participant.objects.filter(user=request.user).prefetch_related('team')
159
160 challenge_participated_teams = []
161 for participant_obj in participant_objs:
162 participant_team = participant_obj.team
163 try:
164 challenge = Challenge.objects.get(
165 participant_teams=participant_team)
166 except Challenge.DoesNotExist:
167 challenge = None
168
169 challenge_participated_teams.append(ChallengeParticipantTeam(
170 challenge, participant_team))
171
172 serializer = ChallengeParticipantTeamListSerializer(ChallengeParticipantTeamList(challenge_participated_teams))
173 return Response(serializer.data, status=status.HTTP_200_OK)
174
175
176 @throttle_classes([UserRateThrottle])
177 @api_view(['DELETE', ])
178 @permission_classes((permissions.IsAuthenticated, HasVerifiedEmail))
179 @authentication_classes((ExpiringTokenAuthentication,))
180 def remove_self_from_participant_team(request, participant_team_pk):
181 """
182 A user can remove himself from the participant team.
183 """
184 try:
185 ParticipantTeam.objects.get(pk=participant_team_pk)
186 except ParticipantTeam.DoesNotExist:
187 response_data = {'error': 'ParticipantTeam does not exist'}
188 return Response(response_data, status=status.HTTP_406_NOT_ACCEPTABLE)
189
190 try:
191 participant = Participant.objects.get(user=request.user.id, team__pk=participant_team_pk)
192 participant.delete()
193 return Response(status=status.HTTP_204_NO_CONTENT)
194 except:
195 response_data = {'error': 'Sorry, you do not belong to this team.'}
196 return Response(response_data, status=status.HTTP_401_UNAUTHORIZED)
197
[end of apps/participants/views.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/apps/participants/views.py b/apps/participants/views.py
--- a/apps/participants/views.py
+++ b/apps/participants/views.py
@@ -27,8 +27,9 @@
def participant_team_list(request):
if request.method == 'GET':
+ participant_teams_id = Participant.objects.filter(user_id=request.user).values_list('team_id', flat=True)
participant_teams = ParticipantTeam.objects.filter(
- created_by=request.user)
+ id__in=participant_teams_id)
paginator, result_page = paginated_queryset(participant_teams, request)
serializer = ParticipantTeamSerializer(result_page, many=True)
response_data = serializer.data
|
{"golden_diff": "diff --git a/apps/participants/views.py b/apps/participants/views.py\n--- a/apps/participants/views.py\n+++ b/apps/participants/views.py\n@@ -27,8 +27,9 @@\n def participant_team_list(request):\n \n if request.method == 'GET':\n+ participant_teams_id = Participant.objects.filter(user_id=request.user).values_list('team_id', flat=True)\n participant_teams = ParticipantTeam.objects.filter(\n- created_by=request.user)\n+ id__in=participant_teams_id)\n paginator, result_page = paginated_queryset(participant_teams, request)\n serializer = ParticipantTeamSerializer(result_page, many=True)\n response_data = serializer.data\n", "issue": "Participant teams not showing up\nThe participant teams to which the user has been invited does not show up if that user logs in.\r\n\r\nFor instance: User A creates a team ABC and invites User B\r\nUser B logs in, and goes to the participant teams page. \r\nUser B does not see team ABC. \n", "before_files": [{"content": "from rest_framework import permissions, status\nfrom rest_framework.decorators import (api_view,\n authentication_classes,\n permission_classes,\n throttle_classes,)\nfrom rest_framework.response import Response\nfrom rest_framework_expiring_authtoken.authentication import (\n ExpiringTokenAuthentication,)\nfrom rest_framework.throttling import UserRateThrottle\n\nfrom accounts.permissions import HasVerifiedEmail\nfrom base.utils import paginated_queryset\nfrom challenges.models import Challenge\n\nfrom .models import (Participant, ParticipantTeam)\nfrom .serializers import (InviteParticipantToTeamSerializer,\n ParticipantTeamSerializer,\n ChallengeParticipantTeam,\n ChallengeParticipantTeamList,\n ChallengeParticipantTeamListSerializer,)\n\n\n@throttle_classes([UserRateThrottle])\n@api_view(['GET', 'POST'])\n@permission_classes((permissions.IsAuthenticated, HasVerifiedEmail))\n@authentication_classes((ExpiringTokenAuthentication,))\ndef participant_team_list(request):\n\n if request.method == 'GET':\n participant_teams = ParticipantTeam.objects.filter(\n created_by=request.user)\n paginator, result_page = paginated_queryset(participant_teams, request)\n serializer = ParticipantTeamSerializer(result_page, many=True)\n response_data = serializer.data\n return paginator.get_paginated_response(response_data)\n\n elif request.method == 'POST':\n serializer = ParticipantTeamSerializer(data=request.data,\n context={'request': request})\n if serializer.is_valid():\n serializer.save()\n response_data = serializer.data\n participant_team = serializer.instance\n participant = Participant(user=request.user,\n status=Participant.SELF,\n team=participant_team)\n participant.save()\n return Response(response_data, status=status.HTTP_201_CREATED)\n return Response(serializer.errors, status=status.HTTP_400_BAD_REQUEST)\n\n\n@throttle_classes([UserRateThrottle])\n@api_view(['GET', 'PUT', 'PATCH', 'DELETE'])\n@permission_classes((permissions.IsAuthenticated, HasVerifiedEmail))\n@authentication_classes((ExpiringTokenAuthentication,))\ndef participant_team_detail(request, pk):\n\n try:\n participant_team = ParticipantTeam.objects.get(pk=pk)\n except ParticipantTeam.DoesNotExist:\n response_data = {'error': 'ParticipantTeam does not exist'}\n return Response(response_data, status=status.HTTP_406_NOT_ACCEPTABLE)\n\n if request.method == 'GET':\n serializer = ParticipantTeamSerializer(participant_team)\n response_data = serializer.data\n return Response(response_data, status=status.HTTP_200_OK)\n\n elif request.method in ['PUT', 'PATCH']:\n\n if request.method == 'PATCH':\n serializer = ParticipantTeamSerializer(participant_team, data=request.data,\n context={\n 'request': request},\n partial=True)\n else:\n serializer = ParticipantTeamSerializer(participant_team, data=request.data,\n context={'request': request})\n if serializer.is_valid():\n serializer.save()\n response_data = serializer.data\n return Response(response_data, status=status.HTTP_200_OK)\n else:\n return Response(serializer.errors, status=status.HTTP_400_BAD_REQUEST)\n\n elif request.method == 'DELETE':\n participant_team.delete()\n return Response(status=status.HTTP_204_NO_CONTENT)\n\n\n@throttle_classes([UserRateThrottle])\n@api_view(['POST'])\n@permission_classes((permissions.IsAuthenticated, HasVerifiedEmail))\n@authentication_classes((ExpiringTokenAuthentication,))\ndef invite_participant_to_team(request, pk):\n\n try:\n participant_team = ParticipantTeam.objects.get(pk=pk)\n except ParticipantTeam.DoesNotExist:\n response_data = {'error': 'ParticipantTeam does not exist'}\n return Response(response_data, status=status.HTTP_406_NOT_ACCEPTABLE)\n\n serializer = InviteParticipantToTeamSerializer(data=request.data,\n context={'participant_team': participant_team,\n 'request': request})\n if serializer.is_valid():\n serializer.save()\n response_data = {\n 'message': 'User has been added successfully to the team'}\n return Response(response_data, status=status.HTTP_202_ACCEPTED)\n return Response(serializer.errors, status=status.HTTP_400_BAD_REQUEST)\n\n\n@throttle_classes([UserRateThrottle])\n@api_view(['DELETE'])\n@permission_classes((permissions.IsAuthenticated, HasVerifiedEmail))\n@authentication_classes((ExpiringTokenAuthentication,))\ndef delete_participant_from_team(request, participant_team_pk, participant_pk):\n \"\"\"\n Deletes a participant from a Participant Team\n \"\"\"\n try:\n participant_team = ParticipantTeam.objects.get(pk=participant_team_pk)\n except ParticipantTeam.DoesNotExist:\n response_data = {'error': 'ParticipantTeam does not exist'}\n return Response(response_data, status=status.HTTP_406_NOT_ACCEPTABLE)\n\n try:\n participant = Participant.objects.get(pk=participant_pk)\n except Participant.DoesNotExist:\n response_data = {'error': 'Participant does not exist'}\n return Response(response_data, status=status.HTTP_406_NOT_ACCEPTABLE)\n\n if participant_team.created_by == request.user:\n\n if participant.user == request.user: # when the user tries to remove himself\n response_data = {\n 'error': 'You are not allowed to remove yourself since you are admin. Please delete the team if you want to do so!'} # noqa: ignore=E501\n return Response(response_data, status=status.HTTP_406_NOT_ACCEPTABLE)\n else:\n participant.delete()\n return Response(status=status.HTTP_204_NO_CONTENT)\n else:\n response_data = {\n 'error': 'Sorry, you do not have permissions to remove this participant'}\n return Response(response_data, status=status.HTTP_401_UNAUTHORIZED)\n\n\n@throttle_classes([UserRateThrottle])\n@api_view(['GET', ])\n@permission_classes((permissions.IsAuthenticated, HasVerifiedEmail))\n@authentication_classes((ExpiringTokenAuthentication,))\ndef get_teams_and_corresponding_challenges_for_a_participant(request):\n \"\"\"\n Returns list of teams and corresponding challenges for a participant\n \"\"\"\n # first get list of all the participants and teams related to the user\n participant_objs = Participant.objects.filter(user=request.user).prefetch_related('team')\n\n challenge_participated_teams = []\n for participant_obj in participant_objs:\n participant_team = participant_obj.team\n try:\n challenge = Challenge.objects.get(\n participant_teams=participant_team)\n except Challenge.DoesNotExist:\n challenge = None\n\n challenge_participated_teams.append(ChallengeParticipantTeam(\n challenge, participant_team))\n\n serializer = ChallengeParticipantTeamListSerializer(ChallengeParticipantTeamList(challenge_participated_teams))\n return Response(serializer.data, status=status.HTTP_200_OK)\n\n\n@throttle_classes([UserRateThrottle])\n@api_view(['DELETE', ])\n@permission_classes((permissions.IsAuthenticated, HasVerifiedEmail))\n@authentication_classes((ExpiringTokenAuthentication,))\ndef remove_self_from_participant_team(request, participant_team_pk):\n \"\"\"\n A user can remove himself from the participant team.\n \"\"\"\n try:\n ParticipantTeam.objects.get(pk=participant_team_pk)\n except ParticipantTeam.DoesNotExist:\n response_data = {'error': 'ParticipantTeam does not exist'}\n return Response(response_data, status=status.HTTP_406_NOT_ACCEPTABLE)\n\n try:\n participant = Participant.objects.get(user=request.user.id, team__pk=participant_team_pk)\n participant.delete()\n return Response(status=status.HTTP_204_NO_CONTENT)\n except:\n response_data = {'error': 'Sorry, you do not belong to this team.'}\n return Response(response_data, status=status.HTTP_401_UNAUTHORIZED)\n", "path": "apps/participants/views.py"}]}
| 2,659 | 142 |
gh_patches_debug_35563
|
rasdani/github-patches
|
git_diff
|
litestar-org__litestar-784
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Bug: StaticFiles sends files as `content-disposition: 'attachment'` in html-mode
**Describe the bug**
When using `StaticFiles` in html-mode, files are being sent with `content-disposition: 'attachment'`
**To Reproduce**
Create an `html/index.html` file. Run:
```python
from starlite import Starlite, StaticFilesConfig, TestClient
app = Starlite(
static_files_config=[StaticFilesConfig(path="/", directories=["html"], html_mode=True)], route_handlers=[]
)
with TestClient(app=app) as client:
res = client.get("/index.html")
assert not res.headers["content-disposition"].startswith("attachment")
```
Bug: StaticFiles sends files as `content-disposition: 'attachment'` in html-mode
**Describe the bug**
When using `StaticFiles` in html-mode, files are being sent with `content-disposition: 'attachment'`
**To Reproduce**
Create an `html/index.html` file. Run:
```python
from starlite import Starlite, StaticFilesConfig, TestClient
app = Starlite(
static_files_config=[StaticFilesConfig(path="/", directories=["html"], html_mode=True)], route_handlers=[]
)
with TestClient(app=app) as client:
res = client.get("/index.html")
assert not res.headers["content-disposition"].startswith("attachment")
```
</issue>
<code>
[start of starlite/static_files/base.py]
1 from os.path import commonpath, join
2 from typing import TYPE_CHECKING, List, Tuple, Union
3
4 from starlite.enums import ScopeType
5 from starlite.exceptions import MethodNotAllowedException, NotFoundException
6 from starlite.response import FileResponse
7 from starlite.status_codes import HTTP_404_NOT_FOUND
8 from starlite.utils.file import FileSystemAdapter
9
10 if TYPE_CHECKING:
11
12 from starlite.types import Receive, Scope, Send
13 from starlite.types.composite_types import PathType
14 from starlite.types.file_types import FileInfo, FileSystemProtocol
15
16
17 class StaticFiles:
18 __slots__ = ("is_html_mode", "directories", "adapter")
19
20 def __init__(self, is_html_mode: bool, directories: List["PathType"], file_system: "FileSystemProtocol") -> None:
21 """This class is an ASGI App that handles file sending.
22
23 Args:
24 is_html_mode: Flag dictating whether serving html. If true, the default file will be 'index.html'.
25 directories: A list of directories to serve files from.
26 file_system: The file_system spec to use for serving files.
27 """
28 self.adapter = FileSystemAdapter(file_system)
29 self.directories = directories
30 self.is_html_mode = is_html_mode
31
32 async def get_fs_info(
33 self, directories: List["PathType"], file_path: str
34 ) -> Union[Tuple[str, "FileInfo"], Tuple[None, None]]:
35 """Resolves the file path and returns the resolved path and a.
36
37 [stat_result][os.stat_result].
38
39 Args:
40 directories: A list of directory paths.
41 file_path: A file path to resolve
42
43 Returns:
44 A tuple with an optional resolved [Path][anyio.Path] instance and an optional [stat_result][os.stat_result].
45 """
46 for directory in directories:
47 try:
48 joined_path = join(directory, file_path) # noqa: PL118
49 file_info = await self.adapter.info(joined_path)
50 if file_info and commonpath([str(directory), file_info["name"], joined_path]) == str(directory):
51 return joined_path, file_info
52 except FileNotFoundError:
53 continue
54 return None, None
55
56 async def __call__(self, scope: "Scope", receive: "Receive", send: "Send") -> None:
57 if scope["type"] != ScopeType.HTTP or scope["method"] not in {"GET", "HEAD"}:
58 raise MethodNotAllowedException()
59
60 split_path = scope["path"].split("/")
61 filename = split_path[-1]
62 joined_path = join(*split_path) # noqa: PL118
63 resolved_path, fs_info = await self.get_fs_info(directories=self.directories, file_path=joined_path)
64
65 if fs_info and fs_info["type"] == "directory" and self.is_html_mode:
66 filename = "index.html"
67 resolved_path, fs_info = await self.get_fs_info(
68 directories=self.directories, file_path=join(resolved_path or joined_path, filename)
69 )
70
71 if fs_info and fs_info["type"] == "file":
72 await FileResponse(
73 path=resolved_path or joined_path,
74 file_info=fs_info,
75 file_system=self.adapter.file_system,
76 filename=filename,
77 is_head_response=scope["method"] == "HEAD",
78 )(scope, receive, send)
79 return
80
81 if self.is_html_mode:
82 filename = "404.html"
83 resolved_path, fs_info = await self.get_fs_info(directories=self.directories, file_path=filename)
84 if fs_info and fs_info["type"] == "file":
85 await FileResponse(
86 path=resolved_path or joined_path,
87 file_info=fs_info,
88 file_system=self.adapter.file_system,
89 filename=filename,
90 is_head_response=scope["method"] == "HEAD",
91 status_code=HTTP_404_NOT_FOUND,
92 )(scope, receive, send)
93 return
94
95 raise NotFoundException(f"no file or directory match the path {resolved_path or joined_path} was found")
96
[end of starlite/static_files/base.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/starlite/static_files/base.py b/starlite/static_files/base.py
--- a/starlite/static_files/base.py
+++ b/starlite/static_files/base.py
@@ -8,6 +8,7 @@
from starlite.utils.file import FileSystemAdapter
if TYPE_CHECKING:
+ from typing_extensions import Literal
from starlite.types import Receive, Scope, Send
from starlite.types.composite_types import PathType
@@ -61,12 +62,15 @@
filename = split_path[-1]
joined_path = join(*split_path) # noqa: PL118
resolved_path, fs_info = await self.get_fs_info(directories=self.directories, file_path=joined_path)
+ content_disposition_type: "Literal['inline', 'attachment']" = "attachment"
- if fs_info and fs_info["type"] == "directory" and self.is_html_mode:
- filename = "index.html"
- resolved_path, fs_info = await self.get_fs_info(
- directories=self.directories, file_path=join(resolved_path or joined_path, filename)
- )
+ if self.is_html_mode:
+ content_disposition_type = "inline"
+ if fs_info and fs_info["type"] == "directory":
+ filename = "index.html"
+ resolved_path, fs_info = await self.get_fs_info(
+ directories=self.directories, file_path=join(resolved_path or joined_path, filename)
+ )
if fs_info and fs_info["type"] == "file":
await FileResponse(
@@ -75,6 +79,7 @@
file_system=self.adapter.file_system,
filename=filename,
is_head_response=scope["method"] == "HEAD",
+ content_disposition_type=content_disposition_type,
)(scope, receive, send)
return
@@ -89,6 +94,7 @@
filename=filename,
is_head_response=scope["method"] == "HEAD",
status_code=HTTP_404_NOT_FOUND,
+ content_disposition_type=content_disposition_type,
)(scope, receive, send)
return
|
{"golden_diff": "diff --git a/starlite/static_files/base.py b/starlite/static_files/base.py\n--- a/starlite/static_files/base.py\n+++ b/starlite/static_files/base.py\n@@ -8,6 +8,7 @@\n from starlite.utils.file import FileSystemAdapter\n \n if TYPE_CHECKING:\n+ from typing_extensions import Literal\n \n from starlite.types import Receive, Scope, Send\n from starlite.types.composite_types import PathType\n@@ -61,12 +62,15 @@\n filename = split_path[-1]\n joined_path = join(*split_path) # noqa: PL118\n resolved_path, fs_info = await self.get_fs_info(directories=self.directories, file_path=joined_path)\n+ content_disposition_type: \"Literal['inline', 'attachment']\" = \"attachment\"\n \n- if fs_info and fs_info[\"type\"] == \"directory\" and self.is_html_mode:\n- filename = \"index.html\"\n- resolved_path, fs_info = await self.get_fs_info(\n- directories=self.directories, file_path=join(resolved_path or joined_path, filename)\n- )\n+ if self.is_html_mode:\n+ content_disposition_type = \"inline\"\n+ if fs_info and fs_info[\"type\"] == \"directory\":\n+ filename = \"index.html\"\n+ resolved_path, fs_info = await self.get_fs_info(\n+ directories=self.directories, file_path=join(resolved_path or joined_path, filename)\n+ )\n \n if fs_info and fs_info[\"type\"] == \"file\":\n await FileResponse(\n@@ -75,6 +79,7 @@\n file_system=self.adapter.file_system,\n filename=filename,\n is_head_response=scope[\"method\"] == \"HEAD\",\n+ content_disposition_type=content_disposition_type,\n )(scope, receive, send)\n return\n \n@@ -89,6 +94,7 @@\n filename=filename,\n is_head_response=scope[\"method\"] == \"HEAD\",\n status_code=HTTP_404_NOT_FOUND,\n+ content_disposition_type=content_disposition_type,\n )(scope, receive, send)\n return\n", "issue": "Bug: StaticFiles sends files as `content-disposition: 'attachment'` in html-mode\n**Describe the bug**\r\nWhen using `StaticFiles` in html-mode, files are being sent with `content-disposition: 'attachment'`\r\n\r\n**To Reproduce**\r\nCreate an `html/index.html` file. Run:\r\n\r\n```python\r\nfrom starlite import Starlite, StaticFilesConfig, TestClient\r\n\r\napp = Starlite(\r\n static_files_config=[StaticFilesConfig(path=\"/\", directories=[\"html\"], html_mode=True)], route_handlers=[]\r\n)\r\n\r\nwith TestClient(app=app) as client:\r\n res = client.get(\"/index.html\")\r\n assert not res.headers[\"content-disposition\"].startswith(\"attachment\")\r\n```\r\n\nBug: StaticFiles sends files as `content-disposition: 'attachment'` in html-mode\n**Describe the bug**\r\nWhen using `StaticFiles` in html-mode, files are being sent with `content-disposition: 'attachment'`\r\n\r\n**To Reproduce**\r\nCreate an `html/index.html` file. Run:\r\n\r\n```python\r\nfrom starlite import Starlite, StaticFilesConfig, TestClient\r\n\r\napp = Starlite(\r\n static_files_config=[StaticFilesConfig(path=\"/\", directories=[\"html\"], html_mode=True)], route_handlers=[]\r\n)\r\n\r\nwith TestClient(app=app) as client:\r\n res = client.get(\"/index.html\")\r\n assert not res.headers[\"content-disposition\"].startswith(\"attachment\")\r\n```\r\n\n", "before_files": [{"content": "from os.path import commonpath, join\nfrom typing import TYPE_CHECKING, List, Tuple, Union\n\nfrom starlite.enums import ScopeType\nfrom starlite.exceptions import MethodNotAllowedException, NotFoundException\nfrom starlite.response import FileResponse\nfrom starlite.status_codes import HTTP_404_NOT_FOUND\nfrom starlite.utils.file import FileSystemAdapter\n\nif TYPE_CHECKING:\n\n from starlite.types import Receive, Scope, Send\n from starlite.types.composite_types import PathType\n from starlite.types.file_types import FileInfo, FileSystemProtocol\n\n\nclass StaticFiles:\n __slots__ = (\"is_html_mode\", \"directories\", \"adapter\")\n\n def __init__(self, is_html_mode: bool, directories: List[\"PathType\"], file_system: \"FileSystemProtocol\") -> None:\n \"\"\"This class is an ASGI App that handles file sending.\n\n Args:\n is_html_mode: Flag dictating whether serving html. If true, the default file will be 'index.html'.\n directories: A list of directories to serve files from.\n file_system: The file_system spec to use for serving files.\n \"\"\"\n self.adapter = FileSystemAdapter(file_system)\n self.directories = directories\n self.is_html_mode = is_html_mode\n\n async def get_fs_info(\n self, directories: List[\"PathType\"], file_path: str\n ) -> Union[Tuple[str, \"FileInfo\"], Tuple[None, None]]:\n \"\"\"Resolves the file path and returns the resolved path and a.\n\n [stat_result][os.stat_result].\n\n Args:\n directories: A list of directory paths.\n file_path: A file path to resolve\n\n Returns:\n A tuple with an optional resolved [Path][anyio.Path] instance and an optional [stat_result][os.stat_result].\n \"\"\"\n for directory in directories:\n try:\n joined_path = join(directory, file_path) # noqa: PL118\n file_info = await self.adapter.info(joined_path)\n if file_info and commonpath([str(directory), file_info[\"name\"], joined_path]) == str(directory):\n return joined_path, file_info\n except FileNotFoundError:\n continue\n return None, None\n\n async def __call__(self, scope: \"Scope\", receive: \"Receive\", send: \"Send\") -> None:\n if scope[\"type\"] != ScopeType.HTTP or scope[\"method\"] not in {\"GET\", \"HEAD\"}:\n raise MethodNotAllowedException()\n\n split_path = scope[\"path\"].split(\"/\")\n filename = split_path[-1]\n joined_path = join(*split_path) # noqa: PL118\n resolved_path, fs_info = await self.get_fs_info(directories=self.directories, file_path=joined_path)\n\n if fs_info and fs_info[\"type\"] == \"directory\" and self.is_html_mode:\n filename = \"index.html\"\n resolved_path, fs_info = await self.get_fs_info(\n directories=self.directories, file_path=join(resolved_path or joined_path, filename)\n )\n\n if fs_info and fs_info[\"type\"] == \"file\":\n await FileResponse(\n path=resolved_path or joined_path,\n file_info=fs_info,\n file_system=self.adapter.file_system,\n filename=filename,\n is_head_response=scope[\"method\"] == \"HEAD\",\n )(scope, receive, send)\n return\n\n if self.is_html_mode:\n filename = \"404.html\"\n resolved_path, fs_info = await self.get_fs_info(directories=self.directories, file_path=filename)\n if fs_info and fs_info[\"type\"] == \"file\":\n await FileResponse(\n path=resolved_path or joined_path,\n file_info=fs_info,\n file_system=self.adapter.file_system,\n filename=filename,\n is_head_response=scope[\"method\"] == \"HEAD\",\n status_code=HTTP_404_NOT_FOUND,\n )(scope, receive, send)\n return\n\n raise NotFoundException(f\"no file or directory match the path {resolved_path or joined_path} was found\")\n", "path": "starlite/static_files/base.py"}]}
| 1,885 | 468 |
gh_patches_debug_42468
|
rasdani/github-patches
|
git_diff
|
pex-tool__pex-701
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
If sys.executable is not on PATH a pex will re-exec itself forever.
This can happen via:
```
/not/on/PATH/python a.pex
```
If `a.pex` in this example has compatibility constraints that rule out all other interpreters on `PATH` (or there are no intepreters on `PATH`!, c.f.: #53), then, even if sys.executable would be compatible to execute `a.pex` with, it is not considered useable:
https://github.com/pantsbuild/pex/blob/442ce7952674cb8a367ecded7d138661194d4f12/pex/pex_bootstrapper.py#L45-L50
https://github.com/pantsbuild/pex/blob/442ce7952674cb8a367ecded7d138661194d4f12/pex/interpreter.py#L283-L287
And so the pex_bootstrapper re-execs the pex ad infinitum:
https://github.com/pantsbuild/pex/blob/442ce7952674cb8a367ecded7d138661194d4f12/pex/pex_bootstrapper.py#L122-L137
</issue>
<code>
[start of pex/pex_bootstrapper.py]
1 # Copyright 2014 Pants project contributors (see CONTRIBUTORS.md).
2 # Licensed under the Apache License, Version 2.0 (see LICENSE).
3
4 from __future__ import absolute_import, print_function
5
6 import os
7 import sys
8
9 from pex import pex_warnings
10 from pex.common import die
11 from pex.executor import Executor
12 from pex.interpreter import PythonInterpreter
13 from pex.interpreter_constraints import matched_interpreters
14 from pex.tracer import TRACER
15 from pex.variables import ENV
16
17 __all__ = ('bootstrap_pex',)
18
19
20 def find_in_path(target_interpreter):
21 if os.path.exists(target_interpreter):
22 return target_interpreter
23
24 for directory in os.getenv('PATH', '').split(os.pathsep):
25 try_path = os.path.join(directory, target_interpreter)
26 if os.path.exists(try_path):
27 return try_path
28
29
30 def find_compatible_interpreters(pex_python_path, compatibility_constraints):
31 """Find all compatible interpreters on the system within the supplied constraints and use
32 PEX_PYTHON_PATH if it is set. If not, fall back to interpreters on $PATH.
33 """
34 if pex_python_path:
35 interpreters = []
36 for binary in pex_python_path.split(os.pathsep):
37 try:
38 interpreters.append(PythonInterpreter.from_binary(binary))
39 except Executor.ExecutionError:
40 print("Python interpreter %s in PEX_PYTHON_PATH failed to load properly." % binary,
41 file=sys.stderr)
42 if not interpreters:
43 die('PEX_PYTHON_PATH was defined, but no valid interpreters could be identified. Exiting.')
44 else:
45 if not os.getenv('PATH', ''):
46 # no $PATH, use sys.executable
47 interpreters = [PythonInterpreter.get()]
48 else:
49 # get all qualifying interpreters found in $PATH
50 interpreters = PythonInterpreter.all()
51
52 return list(
53 matched_interpreters(interpreters, compatibility_constraints)
54 if compatibility_constraints
55 else interpreters
56 )
57
58
59 def _select_pex_python_interpreter(target_python, compatibility_constraints):
60 target = find_in_path(target_python)
61
62 if not target:
63 die('Failed to find interpreter specified by PEX_PYTHON: %s' % target)
64 if compatibility_constraints:
65 pi = PythonInterpreter.from_binary(target)
66 if not list(matched_interpreters([pi], compatibility_constraints)):
67 die('Interpreter specified by PEX_PYTHON (%s) is not compatible with specified '
68 'interpreter constraints: %s' % (target, str(compatibility_constraints)))
69 if not os.path.exists(target):
70 die('Target interpreter specified by PEX_PYTHON %s does not exist. Exiting.' % target)
71 return target
72
73
74 def _select_interpreter(pex_python_path, compatibility_constraints):
75 compatible_interpreters = find_compatible_interpreters(
76 pex_python_path, compatibility_constraints)
77
78 if not compatible_interpreters:
79 die('Failed to find compatible interpreter for constraints: %s'
80 % str(compatibility_constraints))
81 # TODO: https://github.com/pantsbuild/pex/issues/430
82 target = min(compatible_interpreters).binary
83
84 if os.path.exists(target):
85 return target
86
87
88 def maybe_reexec_pex(compatibility_constraints):
89 """
90 Handle environment overrides for the Python interpreter to use when executing this pex.
91
92 This function supports interpreter filtering based on interpreter constraints stored in PEX-INFO
93 metadata. If PEX_PYTHON is set in a pexrc, it attempts to obtain the binary location of the
94 interpreter specified by PEX_PYTHON. If PEX_PYTHON_PATH is set, it attempts to search the path for
95 a matching interpreter in accordance with the interpreter constraints. If both variables are
96 present in a pexrc, this function gives precedence to PEX_PYTHON_PATH and errors out if no
97 compatible interpreters can be found on said path.
98
99 If neither variable is set, we fall back to plain PEX execution using PATH searching or the
100 currently executing interpreter. If compatibility constraints are used, we match those constraints
101 against these interpreters.
102
103 :param compatibility_constraints: list of requirements-style strings that constrain the
104 Python interpreter to re-exec this pex with.
105 """
106 if os.environ.pop('SHOULD_EXIT_BOOTSTRAP_REEXEC', None):
107 # We've already been here and selected an interpreter. Continue to execution.
108 return
109
110 target = None
111 with TRACER.timed('Selecting runtime interpreter based on pexrc', V=3):
112 if ENV.PEX_PYTHON and not ENV.PEX_PYTHON_PATH:
113 # preserve PEX_PYTHON re-exec for backwards compatibility
114 # TODO: Kill this off completely in favor of PEX_PYTHON_PATH
115 # https://github.com/pantsbuild/pex/issues/431
116 target = _select_pex_python_interpreter(ENV.PEX_PYTHON,
117 compatibility_constraints)
118 elif ENV.PEX_PYTHON_PATH:
119 target = _select_interpreter(ENV.PEX_PYTHON_PATH, compatibility_constraints)
120
121 elif compatibility_constraints:
122 # Apply constraints to target using regular PATH
123 target = _select_interpreter(
124 pex_python_path=None,
125 compatibility_constraints=compatibility_constraints
126 )
127
128 if target and os.path.realpath(target) != os.path.realpath(sys.executable):
129 cmdline = [target] + sys.argv
130 TRACER.log('Re-executing: cmdline="%s", sys.executable="%s", PEX_PYTHON="%s", '
131 'PEX_PYTHON_PATH="%s", COMPATIBILITY_CONSTRAINTS="%s"'
132 % (cmdline, sys.executable, ENV.PEX_PYTHON, ENV.PEX_PYTHON_PATH,
133 compatibility_constraints))
134 ENV.delete('PEX_PYTHON')
135 ENV.delete('PEX_PYTHON_PATH')
136 os.environ['SHOULD_EXIT_BOOTSTRAP_REEXEC'] = '1'
137 os.execve(target, cmdline, ENV.copy())
138
139
140 def _bootstrap(entry_point):
141 from .pex_info import PexInfo
142 pex_info = PexInfo.from_pex(entry_point)
143 pex_warnings.configure_warnings(pex_info)
144
145 from .finders import register_finders
146 register_finders()
147
148 return pex_info
149
150
151 def bootstrap_pex(entry_point):
152 pex_info = _bootstrap(entry_point)
153 maybe_reexec_pex(pex_info.interpreter_constraints)
154
155 from . import pex
156 pex.PEX(entry_point).execute()
157
158
159 # NB: This helper is used by third party libs - namely https://github.com/wickman/lambdex.
160 # TODO(John Sirois): Kill once https://github.com/wickman/lambdex/issues/5 is resolved.
161 def is_compressed(entry_point):
162 from .pex_info import PexInfo
163 return os.path.exists(entry_point) and not os.path.exists(os.path.join(entry_point, PexInfo.PATH))
164
165
166 def bootstrap_pex_env(entry_point):
167 """Bootstrap the current runtime environment using a given pex."""
168 pex_info = _bootstrap(entry_point)
169
170 from .environment import PEXEnvironment
171 PEXEnvironment(entry_point, pex_info).activate()
172
[end of pex/pex_bootstrapper.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/pex/pex_bootstrapper.py b/pex/pex_bootstrapper.py
--- a/pex/pex_bootstrapper.py
+++ b/pex/pex_bootstrapper.py
@@ -11,6 +11,7 @@
from pex.executor import Executor
from pex.interpreter import PythonInterpreter
from pex.interpreter_constraints import matched_interpreters
+from pex.orderedset import OrderedSet
from pex.tracer import TRACER
from pex.variables import ENV
@@ -27,7 +28,7 @@
return try_path
-def find_compatible_interpreters(pex_python_path, compatibility_constraints):
+def find_compatible_interpreters(pex_python_path=None, compatibility_constraints=None):
"""Find all compatible interpreters on the system within the supplied constraints and use
PEX_PYTHON_PATH if it is set. If not, fall back to interpreters on $PATH.
"""
@@ -42,12 +43,12 @@
if not interpreters:
die('PEX_PYTHON_PATH was defined, but no valid interpreters could be identified. Exiting.')
else:
- if not os.getenv('PATH', ''):
- # no $PATH, use sys.executable
- interpreters = [PythonInterpreter.get()]
- else:
- # get all qualifying interpreters found in $PATH
- interpreters = PythonInterpreter.all()
+ # We may have been invoked with a specific interpreter not on the $PATH, make sure our
+ # sys.executable is included as a candidate in this case.
+ interpreters = OrderedSet([PythonInterpreter.get()])
+
+ # Add all qualifying interpreters found in $PATH.
+ interpreters.update(PythonInterpreter.all())
return list(
matched_interpreters(interpreters, compatibility_constraints)
@@ -56,7 +57,7 @@
)
-def _select_pex_python_interpreter(target_python, compatibility_constraints):
+def _select_pex_python_interpreter(target_python, compatibility_constraints=None):
target = find_in_path(target_python)
if not target:
@@ -71,9 +72,9 @@
return target
-def _select_interpreter(pex_python_path, compatibility_constraints):
+def _select_interpreter(pex_python_path=None, compatibility_constraints=None):
compatible_interpreters = find_compatible_interpreters(
- pex_python_path, compatibility_constraints)
+ pex_python_path=pex_python_path, compatibility_constraints=compatibility_constraints)
if not compatible_interpreters:
die('Failed to find compatible interpreter for constraints: %s'
@@ -114,16 +115,14 @@
# TODO: Kill this off completely in favor of PEX_PYTHON_PATH
# https://github.com/pantsbuild/pex/issues/431
target = _select_pex_python_interpreter(ENV.PEX_PYTHON,
- compatibility_constraints)
+ compatibility_constraints=compatibility_constraints)
elif ENV.PEX_PYTHON_PATH:
- target = _select_interpreter(ENV.PEX_PYTHON_PATH, compatibility_constraints)
+ target = _select_interpreter(pex_python_path=ENV.PEX_PYTHON_PATH,
+ compatibility_constraints=compatibility_constraints)
elif compatibility_constraints:
# Apply constraints to target using regular PATH
- target = _select_interpreter(
- pex_python_path=None,
- compatibility_constraints=compatibility_constraints
- )
+ target = _select_interpreter(compatibility_constraints=compatibility_constraints)
if target and os.path.realpath(target) != os.path.realpath(sys.executable):
cmdline = [target] + sys.argv
|
{"golden_diff": "diff --git a/pex/pex_bootstrapper.py b/pex/pex_bootstrapper.py\n--- a/pex/pex_bootstrapper.py\n+++ b/pex/pex_bootstrapper.py\n@@ -11,6 +11,7 @@\n from pex.executor import Executor\n from pex.interpreter import PythonInterpreter\n from pex.interpreter_constraints import matched_interpreters\n+from pex.orderedset import OrderedSet\n from pex.tracer import TRACER\n from pex.variables import ENV\n \n@@ -27,7 +28,7 @@\n return try_path\n \n \n-def find_compatible_interpreters(pex_python_path, compatibility_constraints):\n+def find_compatible_interpreters(pex_python_path=None, compatibility_constraints=None):\n \"\"\"Find all compatible interpreters on the system within the supplied constraints and use\n PEX_PYTHON_PATH if it is set. If not, fall back to interpreters on $PATH.\n \"\"\"\n@@ -42,12 +43,12 @@\n if not interpreters:\n die('PEX_PYTHON_PATH was defined, but no valid interpreters could be identified. Exiting.')\n else:\n- if not os.getenv('PATH', ''):\n- # no $PATH, use sys.executable\n- interpreters = [PythonInterpreter.get()]\n- else:\n- # get all qualifying interpreters found in $PATH\n- interpreters = PythonInterpreter.all()\n+ # We may have been invoked with a specific interpreter not on the $PATH, make sure our\n+ # sys.executable is included as a candidate in this case.\n+ interpreters = OrderedSet([PythonInterpreter.get()])\n+\n+ # Add all qualifying interpreters found in $PATH.\n+ interpreters.update(PythonInterpreter.all())\n \n return list(\n matched_interpreters(interpreters, compatibility_constraints)\n@@ -56,7 +57,7 @@\n )\n \n \n-def _select_pex_python_interpreter(target_python, compatibility_constraints):\n+def _select_pex_python_interpreter(target_python, compatibility_constraints=None):\n target = find_in_path(target_python)\n \n if not target:\n@@ -71,9 +72,9 @@\n return target\n \n \n-def _select_interpreter(pex_python_path, compatibility_constraints):\n+def _select_interpreter(pex_python_path=None, compatibility_constraints=None):\n compatible_interpreters = find_compatible_interpreters(\n- pex_python_path, compatibility_constraints)\n+ pex_python_path=pex_python_path, compatibility_constraints=compatibility_constraints)\n \n if not compatible_interpreters:\n die('Failed to find compatible interpreter for constraints: %s'\n@@ -114,16 +115,14 @@\n # TODO: Kill this off completely in favor of PEX_PYTHON_PATH\n # https://github.com/pantsbuild/pex/issues/431\n target = _select_pex_python_interpreter(ENV.PEX_PYTHON,\n- compatibility_constraints)\n+ compatibility_constraints=compatibility_constraints)\n elif ENV.PEX_PYTHON_PATH:\n- target = _select_interpreter(ENV.PEX_PYTHON_PATH, compatibility_constraints)\n+ target = _select_interpreter(pex_python_path=ENV.PEX_PYTHON_PATH,\n+ compatibility_constraints=compatibility_constraints)\n \n elif compatibility_constraints:\n # Apply constraints to target using regular PATH\n- target = _select_interpreter(\n- pex_python_path=None,\n- compatibility_constraints=compatibility_constraints\n- )\n+ target = _select_interpreter(compatibility_constraints=compatibility_constraints)\n \n if target and os.path.realpath(target) != os.path.realpath(sys.executable):\n cmdline = [target] + sys.argv\n", "issue": "If sys.executable is not on PATH a pex will re-exec itself forever.\nThis can happen via:\r\n```\r\n/not/on/PATH/python a.pex\r\n```\r\n\r\nIf `a.pex` in this example has compatibility constraints that rule out all other interpreters on `PATH` (or there are no intepreters on `PATH`!, c.f.: #53), then, even if sys.executable would be compatible to execute `a.pex` with, it is not considered useable:\r\nhttps://github.com/pantsbuild/pex/blob/442ce7952674cb8a367ecded7d138661194d4f12/pex/pex_bootstrapper.py#L45-L50\r\nhttps://github.com/pantsbuild/pex/blob/442ce7952674cb8a367ecded7d138661194d4f12/pex/interpreter.py#L283-L287\r\n\r\nAnd so the pex_bootstrapper re-execs the pex ad infinitum:\r\nhttps://github.com/pantsbuild/pex/blob/442ce7952674cb8a367ecded7d138661194d4f12/pex/pex_bootstrapper.py#L122-L137\n", "before_files": [{"content": "# Copyright 2014 Pants project contributors (see CONTRIBUTORS.md).\n# Licensed under the Apache License, Version 2.0 (see LICENSE).\n\nfrom __future__ import absolute_import, print_function\n\nimport os\nimport sys\n\nfrom pex import pex_warnings\nfrom pex.common import die\nfrom pex.executor import Executor\nfrom pex.interpreter import PythonInterpreter\nfrom pex.interpreter_constraints import matched_interpreters\nfrom pex.tracer import TRACER\nfrom pex.variables import ENV\n\n__all__ = ('bootstrap_pex',)\n\n\ndef find_in_path(target_interpreter):\n if os.path.exists(target_interpreter):\n return target_interpreter\n\n for directory in os.getenv('PATH', '').split(os.pathsep):\n try_path = os.path.join(directory, target_interpreter)\n if os.path.exists(try_path):\n return try_path\n\n\ndef find_compatible_interpreters(pex_python_path, compatibility_constraints):\n \"\"\"Find all compatible interpreters on the system within the supplied constraints and use\n PEX_PYTHON_PATH if it is set. If not, fall back to interpreters on $PATH.\n \"\"\"\n if pex_python_path:\n interpreters = []\n for binary in pex_python_path.split(os.pathsep):\n try:\n interpreters.append(PythonInterpreter.from_binary(binary))\n except Executor.ExecutionError:\n print(\"Python interpreter %s in PEX_PYTHON_PATH failed to load properly.\" % binary,\n file=sys.stderr)\n if not interpreters:\n die('PEX_PYTHON_PATH was defined, but no valid interpreters could be identified. Exiting.')\n else:\n if not os.getenv('PATH', ''):\n # no $PATH, use sys.executable\n interpreters = [PythonInterpreter.get()]\n else:\n # get all qualifying interpreters found in $PATH\n interpreters = PythonInterpreter.all()\n\n return list(\n matched_interpreters(interpreters, compatibility_constraints)\n if compatibility_constraints\n else interpreters\n )\n\n\ndef _select_pex_python_interpreter(target_python, compatibility_constraints):\n target = find_in_path(target_python)\n\n if not target:\n die('Failed to find interpreter specified by PEX_PYTHON: %s' % target)\n if compatibility_constraints:\n pi = PythonInterpreter.from_binary(target)\n if not list(matched_interpreters([pi], compatibility_constraints)):\n die('Interpreter specified by PEX_PYTHON (%s) is not compatible with specified '\n 'interpreter constraints: %s' % (target, str(compatibility_constraints)))\n if not os.path.exists(target):\n die('Target interpreter specified by PEX_PYTHON %s does not exist. Exiting.' % target)\n return target\n\n\ndef _select_interpreter(pex_python_path, compatibility_constraints):\n compatible_interpreters = find_compatible_interpreters(\n pex_python_path, compatibility_constraints)\n\n if not compatible_interpreters:\n die('Failed to find compatible interpreter for constraints: %s'\n % str(compatibility_constraints))\n # TODO: https://github.com/pantsbuild/pex/issues/430\n target = min(compatible_interpreters).binary\n\n if os.path.exists(target):\n return target\n\n\ndef maybe_reexec_pex(compatibility_constraints):\n \"\"\"\n Handle environment overrides for the Python interpreter to use when executing this pex.\n\n This function supports interpreter filtering based on interpreter constraints stored in PEX-INFO\n metadata. If PEX_PYTHON is set in a pexrc, it attempts to obtain the binary location of the\n interpreter specified by PEX_PYTHON. If PEX_PYTHON_PATH is set, it attempts to search the path for\n a matching interpreter in accordance with the interpreter constraints. If both variables are\n present in a pexrc, this function gives precedence to PEX_PYTHON_PATH and errors out if no\n compatible interpreters can be found on said path.\n\n If neither variable is set, we fall back to plain PEX execution using PATH searching or the\n currently executing interpreter. If compatibility constraints are used, we match those constraints\n against these interpreters.\n\n :param compatibility_constraints: list of requirements-style strings that constrain the\n Python interpreter to re-exec this pex with.\n \"\"\"\n if os.environ.pop('SHOULD_EXIT_BOOTSTRAP_REEXEC', None):\n # We've already been here and selected an interpreter. Continue to execution.\n return\n\n target = None\n with TRACER.timed('Selecting runtime interpreter based on pexrc', V=3):\n if ENV.PEX_PYTHON and not ENV.PEX_PYTHON_PATH:\n # preserve PEX_PYTHON re-exec for backwards compatibility\n # TODO: Kill this off completely in favor of PEX_PYTHON_PATH\n # https://github.com/pantsbuild/pex/issues/431\n target = _select_pex_python_interpreter(ENV.PEX_PYTHON,\n compatibility_constraints)\n elif ENV.PEX_PYTHON_PATH:\n target = _select_interpreter(ENV.PEX_PYTHON_PATH, compatibility_constraints)\n\n elif compatibility_constraints:\n # Apply constraints to target using regular PATH\n target = _select_interpreter(\n pex_python_path=None,\n compatibility_constraints=compatibility_constraints\n )\n\n if target and os.path.realpath(target) != os.path.realpath(sys.executable):\n cmdline = [target] + sys.argv\n TRACER.log('Re-executing: cmdline=\"%s\", sys.executable=\"%s\", PEX_PYTHON=\"%s\", '\n 'PEX_PYTHON_PATH=\"%s\", COMPATIBILITY_CONSTRAINTS=\"%s\"'\n % (cmdline, sys.executable, ENV.PEX_PYTHON, ENV.PEX_PYTHON_PATH,\n compatibility_constraints))\n ENV.delete('PEX_PYTHON')\n ENV.delete('PEX_PYTHON_PATH')\n os.environ['SHOULD_EXIT_BOOTSTRAP_REEXEC'] = '1'\n os.execve(target, cmdline, ENV.copy())\n\n\ndef _bootstrap(entry_point):\n from .pex_info import PexInfo\n pex_info = PexInfo.from_pex(entry_point)\n pex_warnings.configure_warnings(pex_info)\n\n from .finders import register_finders\n register_finders()\n\n return pex_info\n\n\ndef bootstrap_pex(entry_point):\n pex_info = _bootstrap(entry_point)\n maybe_reexec_pex(pex_info.interpreter_constraints)\n\n from . import pex\n pex.PEX(entry_point).execute()\n\n\n# NB: This helper is used by third party libs - namely https://github.com/wickman/lambdex.\n# TODO(John Sirois): Kill once https://github.com/wickman/lambdex/issues/5 is resolved.\ndef is_compressed(entry_point):\n from .pex_info import PexInfo\n return os.path.exists(entry_point) and not os.path.exists(os.path.join(entry_point, PexInfo.PATH))\n\n\ndef bootstrap_pex_env(entry_point):\n \"\"\"Bootstrap the current runtime environment using a given pex.\"\"\"\n pex_info = _bootstrap(entry_point)\n\n from .environment import PEXEnvironment\n PEXEnvironment(entry_point, pex_info).activate()\n", "path": "pex/pex_bootstrapper.py"}]}
| 2,807 | 796 |
gh_patches_debug_27756
|
rasdani/github-patches
|
git_diff
|
scrapy__scrapy-5002
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
refactoring curl_to_request_kwargs to reduce cyclomatic complexity
<!--
Thanks for taking an interest in Scrapy!
If you have a question that starts with "How to...", please see the Scrapy Community page: https://scrapy.org/community/.
The GitHub issue tracker's purpose is to deal with bug reports and feature requests for the project itself.
Keep in mind that by filing an issue, you are expected to comply with Scrapy's Code of Conduct, including treating everyone with respect: https://github.com/scrapy/scrapy/blob/master/CODE_OF_CONDUCT.md
The following is a suggested template to structure your pull request, you can find more guidelines at https://doc.scrapy.org/en/latest/contributing.html#writing-patches and https://doc.scrapy.org/en/latest/contributing.html#submitting-patches
-->
## Summary
After some exploring with cyclomatic complexity tools (lizard), the function was found to have the second highest complexity.
## Motivation
Low complexity allows for higher readability, testability and maintainability.
## Solution
Refactor
## Additional context
N/A
</issue>
<code>
[start of scrapy/utils/curl.py]
1 import argparse
2 import warnings
3 from shlex import split
4 from http.cookies import SimpleCookie
5 from urllib.parse import urlparse
6
7 from w3lib.http import basic_auth_header
8
9
10 class CurlParser(argparse.ArgumentParser):
11 def error(self, message):
12 error_msg = f'There was an error parsing the curl command: {message}'
13 raise ValueError(error_msg)
14
15
16 curl_parser = CurlParser()
17 curl_parser.add_argument('url')
18 curl_parser.add_argument('-H', '--header', dest='headers', action='append')
19 curl_parser.add_argument('-X', '--request', dest='method')
20 curl_parser.add_argument('-d', '--data', '--data-raw', dest='data')
21 curl_parser.add_argument('-u', '--user', dest='auth')
22
23
24 safe_to_ignore_arguments = [
25 ['--compressed'],
26 # `--compressed` argument is not safe to ignore, but it's included here
27 # because the `HttpCompressionMiddleware` is enabled by default
28 ['-s', '--silent'],
29 ['-v', '--verbose'],
30 ['-#', '--progress-bar']
31 ]
32
33 for argument in safe_to_ignore_arguments:
34 curl_parser.add_argument(*argument, action='store_true')
35
36
37 def curl_to_request_kwargs(curl_command, ignore_unknown_options=True):
38 """Convert a cURL command syntax to Request kwargs.
39
40 :param str curl_command: string containing the curl command
41 :param bool ignore_unknown_options: If true, only a warning is emitted when
42 cURL options are unknown. Otherwise
43 raises an error. (default: True)
44 :return: dictionary of Request kwargs
45 """
46
47 curl_args = split(curl_command)
48
49 if curl_args[0] != 'curl':
50 raise ValueError('A curl command must start with "curl"')
51
52 parsed_args, argv = curl_parser.parse_known_args(curl_args[1:])
53
54 if argv:
55 msg = f'Unrecognized options: {", ".join(argv)}'
56 if ignore_unknown_options:
57 warnings.warn(msg)
58 else:
59 raise ValueError(msg)
60
61 url = parsed_args.url
62
63 # curl automatically prepends 'http' if the scheme is missing, but Request
64 # needs the scheme to work
65 parsed_url = urlparse(url)
66 if not parsed_url.scheme:
67 url = 'http://' + url
68
69 method = parsed_args.method or 'GET'
70
71 result = {'method': method.upper(), 'url': url}
72
73 headers = []
74 cookies = {}
75 for header in parsed_args.headers or ():
76 name, val = header.split(':', 1)
77 name = name.strip()
78 val = val.strip()
79 if name.title() == 'Cookie':
80 for name, morsel in SimpleCookie(val).items():
81 cookies[name] = morsel.value
82 else:
83 headers.append((name, val))
84
85 if parsed_args.auth:
86 user, password = parsed_args.auth.split(':', 1)
87 headers.append(('Authorization', basic_auth_header(user, password)))
88
89 if headers:
90 result['headers'] = headers
91 if cookies:
92 result['cookies'] = cookies
93 if parsed_args.data:
94 result['body'] = parsed_args.data
95 if not parsed_args.method:
96 # if the "data" is specified but the "method" is not specified,
97 # the default method is 'POST'
98 result['method'] = 'POST'
99
100 return result
101
[end of scrapy/utils/curl.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/scrapy/utils/curl.py b/scrapy/utils/curl.py
--- a/scrapy/utils/curl.py
+++ b/scrapy/utils/curl.py
@@ -34,6 +34,26 @@
curl_parser.add_argument(*argument, action='store_true')
+def _parse_headers_and_cookies(parsed_args):
+ headers = []
+ cookies = {}
+ for header in parsed_args.headers or ():
+ name, val = header.split(':', 1)
+ name = name.strip()
+ val = val.strip()
+ if name.title() == 'Cookie':
+ for name, morsel in SimpleCookie(val).items():
+ cookies[name] = morsel.value
+ else:
+ headers.append((name, val))
+
+ if parsed_args.auth:
+ user, password = parsed_args.auth.split(':', 1)
+ headers.append(('Authorization', basic_auth_header(user, password)))
+
+ return headers, cookies
+
+
def curl_to_request_kwargs(curl_command, ignore_unknown_options=True):
"""Convert a cURL command syntax to Request kwargs.
@@ -70,21 +90,7 @@
result = {'method': method.upper(), 'url': url}
- headers = []
- cookies = {}
- for header in parsed_args.headers or ():
- name, val = header.split(':', 1)
- name = name.strip()
- val = val.strip()
- if name.title() == 'Cookie':
- for name, morsel in SimpleCookie(val).items():
- cookies[name] = morsel.value
- else:
- headers.append((name, val))
-
- if parsed_args.auth:
- user, password = parsed_args.auth.split(':', 1)
- headers.append(('Authorization', basic_auth_header(user, password)))
+ headers, cookies = _parse_headers_and_cookies(parsed_args)
if headers:
result['headers'] = headers
|
{"golden_diff": "diff --git a/scrapy/utils/curl.py b/scrapy/utils/curl.py\n--- a/scrapy/utils/curl.py\n+++ b/scrapy/utils/curl.py\n@@ -34,6 +34,26 @@\n curl_parser.add_argument(*argument, action='store_true')\n \n \n+def _parse_headers_and_cookies(parsed_args):\n+ headers = []\n+ cookies = {}\n+ for header in parsed_args.headers or ():\n+ name, val = header.split(':', 1)\n+ name = name.strip()\n+ val = val.strip()\n+ if name.title() == 'Cookie':\n+ for name, morsel in SimpleCookie(val).items():\n+ cookies[name] = morsel.value\n+ else:\n+ headers.append((name, val))\n+\n+ if parsed_args.auth:\n+ user, password = parsed_args.auth.split(':', 1)\n+ headers.append(('Authorization', basic_auth_header(user, password)))\n+\n+ return headers, cookies\n+\n+\n def curl_to_request_kwargs(curl_command, ignore_unknown_options=True):\n \"\"\"Convert a cURL command syntax to Request kwargs.\n \n@@ -70,21 +90,7 @@\n \n result = {'method': method.upper(), 'url': url}\n \n- headers = []\n- cookies = {}\n- for header in parsed_args.headers or ():\n- name, val = header.split(':', 1)\n- name = name.strip()\n- val = val.strip()\n- if name.title() == 'Cookie':\n- for name, morsel in SimpleCookie(val).items():\n- cookies[name] = morsel.value\n- else:\n- headers.append((name, val))\n-\n- if parsed_args.auth:\n- user, password = parsed_args.auth.split(':', 1)\n- headers.append(('Authorization', basic_auth_header(user, password)))\n+ headers, cookies = _parse_headers_and_cookies(parsed_args)\n \n if headers:\n result['headers'] = headers\n", "issue": "refactoring curl_to_request_kwargs to reduce cyclomatic complexity\n<!--\r\n\r\nThanks for taking an interest in Scrapy!\r\n\r\nIf you have a question that starts with \"How to...\", please see the Scrapy Community page: https://scrapy.org/community/.\r\nThe GitHub issue tracker's purpose is to deal with bug reports and feature requests for the project itself.\r\n\r\nKeep in mind that by filing an issue, you are expected to comply with Scrapy's Code of Conduct, including treating everyone with respect: https://github.com/scrapy/scrapy/blob/master/CODE_OF_CONDUCT.md\r\n\r\nThe following is a suggested template to structure your pull request, you can find more guidelines at https://doc.scrapy.org/en/latest/contributing.html#writing-patches and https://doc.scrapy.org/en/latest/contributing.html#submitting-patches\r\n\r\n-->\r\n\r\n## Summary\r\n\r\nAfter some exploring with cyclomatic complexity tools (lizard), the function was found to have the second highest complexity. \r\n\r\n## Motivation\r\n\r\nLow complexity allows for higher readability, testability and maintainability. \r\n\r\n## Solution\r\n\r\nRefactor\r\n\r\n## Additional context\r\n\r\nN/A\r\n\n", "before_files": [{"content": "import argparse\nimport warnings\nfrom shlex import split\nfrom http.cookies import SimpleCookie\nfrom urllib.parse import urlparse\n\nfrom w3lib.http import basic_auth_header\n\n\nclass CurlParser(argparse.ArgumentParser):\n def error(self, message):\n error_msg = f'There was an error parsing the curl command: {message}'\n raise ValueError(error_msg)\n\n\ncurl_parser = CurlParser()\ncurl_parser.add_argument('url')\ncurl_parser.add_argument('-H', '--header', dest='headers', action='append')\ncurl_parser.add_argument('-X', '--request', dest='method')\ncurl_parser.add_argument('-d', '--data', '--data-raw', dest='data')\ncurl_parser.add_argument('-u', '--user', dest='auth')\n\n\nsafe_to_ignore_arguments = [\n ['--compressed'],\n # `--compressed` argument is not safe to ignore, but it's included here\n # because the `HttpCompressionMiddleware` is enabled by default\n ['-s', '--silent'],\n ['-v', '--verbose'],\n ['-#', '--progress-bar']\n]\n\nfor argument in safe_to_ignore_arguments:\n curl_parser.add_argument(*argument, action='store_true')\n\n\ndef curl_to_request_kwargs(curl_command, ignore_unknown_options=True):\n \"\"\"Convert a cURL command syntax to Request kwargs.\n\n :param str curl_command: string containing the curl command\n :param bool ignore_unknown_options: If true, only a warning is emitted when\n cURL options are unknown. Otherwise\n raises an error. (default: True)\n :return: dictionary of Request kwargs\n \"\"\"\n\n curl_args = split(curl_command)\n\n if curl_args[0] != 'curl':\n raise ValueError('A curl command must start with \"curl\"')\n\n parsed_args, argv = curl_parser.parse_known_args(curl_args[1:])\n\n if argv:\n msg = f'Unrecognized options: {\", \".join(argv)}'\n if ignore_unknown_options:\n warnings.warn(msg)\n else:\n raise ValueError(msg)\n\n url = parsed_args.url\n\n # curl automatically prepends 'http' if the scheme is missing, but Request\n # needs the scheme to work\n parsed_url = urlparse(url)\n if not parsed_url.scheme:\n url = 'http://' + url\n\n method = parsed_args.method or 'GET'\n\n result = {'method': method.upper(), 'url': url}\n\n headers = []\n cookies = {}\n for header in parsed_args.headers or ():\n name, val = header.split(':', 1)\n name = name.strip()\n val = val.strip()\n if name.title() == 'Cookie':\n for name, morsel in SimpleCookie(val).items():\n cookies[name] = morsel.value\n else:\n headers.append((name, val))\n\n if parsed_args.auth:\n user, password = parsed_args.auth.split(':', 1)\n headers.append(('Authorization', basic_auth_header(user, password)))\n\n if headers:\n result['headers'] = headers\n if cookies:\n result['cookies'] = cookies\n if parsed_args.data:\n result['body'] = parsed_args.data\n if not parsed_args.method:\n # if the \"data\" is specified but the \"method\" is not specified,\n # the default method is 'POST'\n result['method'] = 'POST'\n\n return result\n", "path": "scrapy/utils/curl.py"}]}
| 1,681 | 429 |
gh_patches_debug_802
|
rasdani/github-patches
|
git_diff
|
pyca__cryptography-1599
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Update year in copyright notice for vectors
Refs #1597
</issue>
<code>
[start of vectors/cryptography_vectors/__about__.py]
1 # This file is dual licensed under the terms of the Apache License, Version
2 # 2.0, and the BSD License. See the LICENSE file in the root of this repository
3 # for complete details.
4
5 from __future__ import absolute_import, division, print_function
6
7 __all__ = [
8 "__title__", "__summary__", "__uri__", "__version__", "__author__",
9 "__email__", "__license__", "__copyright__",
10 ]
11
12 __title__ = "cryptography_vectors"
13 __summary__ = "Test vectors for the cryptography package."
14
15 __uri__ = "https://github.com/pyca/cryptography"
16
17 __version__ = "0.8.dev1"
18
19 __author__ = "The cryptography developers"
20 __email__ = "[email protected]"
21
22 __license__ = "BSD or Apache License, Version 2.0"
23 __copyright__ = "Copyright 2013-2014 %s" % __author__
24
[end of vectors/cryptography_vectors/__about__.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/vectors/cryptography_vectors/__about__.py b/vectors/cryptography_vectors/__about__.py
--- a/vectors/cryptography_vectors/__about__.py
+++ b/vectors/cryptography_vectors/__about__.py
@@ -20,4 +20,4 @@
__email__ = "[email protected]"
__license__ = "BSD or Apache License, Version 2.0"
-__copyright__ = "Copyright 2013-2014 %s" % __author__
+__copyright__ = "Copyright 2013-2015 %s" % __author__
|
{"golden_diff": "diff --git a/vectors/cryptography_vectors/__about__.py b/vectors/cryptography_vectors/__about__.py\n--- a/vectors/cryptography_vectors/__about__.py\n+++ b/vectors/cryptography_vectors/__about__.py\n@@ -20,4 +20,4 @@\n __email__ = \"[email protected]\"\n \n __license__ = \"BSD or Apache License, Version 2.0\"\n-__copyright__ = \"Copyright 2013-2014 %s\" % __author__\n+__copyright__ = \"Copyright 2013-2015 %s\" % __author__\n", "issue": "Update year in copyright notice for vectors\nRefs #1597 \n\n", "before_files": [{"content": "# This file is dual licensed under the terms of the Apache License, Version\n# 2.0, and the BSD License. See the LICENSE file in the root of this repository\n# for complete details.\n\nfrom __future__ import absolute_import, division, print_function\n\n__all__ = [\n \"__title__\", \"__summary__\", \"__uri__\", \"__version__\", \"__author__\",\n \"__email__\", \"__license__\", \"__copyright__\",\n]\n\n__title__ = \"cryptography_vectors\"\n__summary__ = \"Test vectors for the cryptography package.\"\n\n__uri__ = \"https://github.com/pyca/cryptography\"\n\n__version__ = \"0.8.dev1\"\n\n__author__ = \"The cryptography developers\"\n__email__ = \"[email protected]\"\n\n__license__ = \"BSD or Apache License, Version 2.0\"\n__copyright__ = \"Copyright 2013-2014 %s\" % __author__\n", "path": "vectors/cryptography_vectors/__about__.py"}]}
| 804 | 137 |
gh_patches_debug_1453
|
rasdani/github-patches
|
git_diff
|
rlworkgroup__garage-971
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
pytest flag --strict-markers requires version 4.5.0
pytest flag `--strict-markers` in https://github.com/rlworkgroup/garage/blob/master/setup.cfg#L79 requires version >= 4.5.0.
See https://docs.pytest.org/en/latest/changelog.html#pytest-4-5-0-2019-05-11
</issue>
<code>
[start of setup.py]
1 """setuptools based setup module."""
2 from setuptools import find_packages
3 from setuptools import setup
4
5 TF_VERSION = '<1.16,>=1.15.0'
6 GYM_VERSION = '==0.12.4'
7
8 # Required dependencies
9 REQUIRED = [
10 # Please keep alphabetized
11 'akro==0.0.6',
12 'cached_property',
13 'click',
14 'cloudpickle',
15 'cma==2.7.0',
16 'dowel==0.0.2',
17 'gym[atari,box2d,classic_control]' + GYM_VERSION,
18 'joblib<0.13,>=0.12',
19 'matplotlib',
20 'numpy>=1.14.5',
21 'psutil',
22 # Pyglet 1.4.0 introduces some api change which breaks some
23 # gym environments
24 # See: https://github.com/openai/gym/issues/1588
25 'pyglet<1.4.0,>=1.3.0',
26 'pyprind',
27 'python-dateutil',
28 'torch==1.3.0',
29 'ray',
30 'scikit-image',
31 'scipy',
32 'tensorflow' + TF_VERSION,
33 'tensorflow-probability',
34 'torchvision==0.4.1'
35 ]
36
37 # Dependencies for optional features
38 EXTRAS = {}
39
40 EXTRAS['mujoco'] = [
41 'mujoco-py<2.1,>=2.0',
42 'gym[all]' + GYM_VERSION,
43 ]
44
45 EXTRAS['dm_control'] = [
46 # dm_control throws an error during install about not being able to
47 # find a build dependency (absl-py). Later pip executes the `install`
48 # command again and the install succeeds because absl-py has been
49 # installed. This is stupid, but harmless.
50 'dm_control @ https://api.github.com/repos/deepmind/dm_control/tarball/7a36377879c57777e5d5b4da5aae2cd2a29b607a', # pylint: disable=line-too-long; # noqa: E501
51 ]
52
53 EXTRAS['all'] = list(set(sum(EXTRAS.values(), [])))
54
55 # dependencies for using gpu, not included in 'all'
56 EXTRAS['gpu'] = ['tensorflow-gpu' + TF_VERSION]
57
58 # Development dependencies (*not* included in 'all')
59 EXTRAS['dev'] = [
60 # Please keep alphabetized
61 'baselines @ https://api.github.com/repos/openai/baselines/tarball/f2729693253c0ef4d4086231d36e0a4307ec1cb3', # pylint: disable=line-too-long; # noqa: E501
62 'flake8',
63 'flake8-docstrings>=1.5.0',
64 'flake8-import-order',
65 'gtimer',
66 'pandas',
67 'pep8-naming==0.7.0',
68 'pre-commit',
69 'pycodestyle>=2.5.0',
70 'pydocstyle>=4.0.0',
71 'pylint>=2.4.3',
72 'pytest>=3.6', # Required for pytest-cov on Python 3.6
73 'pytest-cov',
74 'pytest-xdist',
75 'recommonmark',
76 'rlkit @ git+https://github.com/vitchyr/rlkit/@1d469a509b797ca04a39b8734c1816ca7d108fc8', # pylint: disable=line-too-long; # noqa: E501
77 'seaborn',
78 'sphinx',
79 'sphinx_rtd_theme',
80 'yapf==0.28.0',
81 ]
82
83 with open('README.md') as f:
84 README = f.read()
85
86 # Get the package version dynamically
87 with open('VERSION') as v:
88 VERSION = v.read().strip()
89
90 setup(
91 name='garage',
92 version=VERSION,
93 author='Reinforcement Learning Working Group',
94 description='A toolkit for reproducible reinforcement learning research',
95 url='https://github.com/rlworkgroup/garage',
96 packages=find_packages(where='src'),
97 package_dir={'': 'src'},
98 scripts=['scripts/garage'],
99 python_requires='>=3.5',
100 install_requires=REQUIRED,
101 extras_require=EXTRAS,
102 license='MIT',
103 long_description=README,
104 long_description_content_type='text/markdown',
105 classifiers=[
106 'Development Status :: 4 - Beta',
107 'Intended Audience :: Developers',
108 'Intended Audience :: Education',
109 'Intended Audience :: Science/Research',
110 'License :: OSI Approved :: MIT License',
111 'Programming Language :: Python :: 3.5',
112 'Programming Language :: Python :: 3.6',
113 'Programming Language :: Python :: 3.7',
114 'Programming Language :: Python :: 3 :: Only',
115 'Topic :: Scientific/Engineering :: Artificial Intelligence',
116 'Topic :: Scientific/Engineering :: Mathematics',
117 'Topic :: Software Development :: Libraries',
118 ],
119 )
120
[end of setup.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/setup.py b/setup.py
--- a/setup.py
+++ b/setup.py
@@ -69,7 +69,7 @@
'pycodestyle>=2.5.0',
'pydocstyle>=4.0.0',
'pylint>=2.4.3',
- 'pytest>=3.6', # Required for pytest-cov on Python 3.6
+ 'pytest>=4.5.0', # Required for strict-markers
'pytest-cov',
'pytest-xdist',
'recommonmark',
|
{"golden_diff": "diff --git a/setup.py b/setup.py\n--- a/setup.py\n+++ b/setup.py\n@@ -69,7 +69,7 @@\n 'pycodestyle>=2.5.0',\n 'pydocstyle>=4.0.0',\n 'pylint>=2.4.3',\n- 'pytest>=3.6', # Required for pytest-cov on Python 3.6\n+ 'pytest>=4.5.0', # Required for strict-markers\n 'pytest-cov',\n 'pytest-xdist',\n 'recommonmark',\n", "issue": "pytest flag --strict-markers requires version 4.5.0\npytest flag `--strict-markers` in https://github.com/rlworkgroup/garage/blob/master/setup.cfg#L79 requires version >= 4.5.0. \r\n\r\nSee https://docs.pytest.org/en/latest/changelog.html#pytest-4-5-0-2019-05-11\n", "before_files": [{"content": "\"\"\"setuptools based setup module.\"\"\"\nfrom setuptools import find_packages\nfrom setuptools import setup\n\nTF_VERSION = '<1.16,>=1.15.0'\nGYM_VERSION = '==0.12.4'\n\n# Required dependencies\nREQUIRED = [\n # Please keep alphabetized\n 'akro==0.0.6',\n 'cached_property',\n 'click',\n 'cloudpickle',\n 'cma==2.7.0',\n 'dowel==0.0.2',\n 'gym[atari,box2d,classic_control]' + GYM_VERSION,\n 'joblib<0.13,>=0.12',\n 'matplotlib',\n 'numpy>=1.14.5',\n 'psutil',\n # Pyglet 1.4.0 introduces some api change which breaks some\n # gym environments\n # See: https://github.com/openai/gym/issues/1588\n 'pyglet<1.4.0,>=1.3.0',\n 'pyprind',\n 'python-dateutil',\n 'torch==1.3.0',\n 'ray',\n 'scikit-image',\n 'scipy',\n 'tensorflow' + TF_VERSION,\n 'tensorflow-probability',\n 'torchvision==0.4.1'\n]\n\n# Dependencies for optional features\nEXTRAS = {}\n\nEXTRAS['mujoco'] = [\n 'mujoco-py<2.1,>=2.0',\n 'gym[all]' + GYM_VERSION,\n]\n\nEXTRAS['dm_control'] = [\n # dm_control throws an error during install about not being able to\n # find a build dependency (absl-py). Later pip executes the `install`\n # command again and the install succeeds because absl-py has been\n # installed. This is stupid, but harmless.\n 'dm_control @ https://api.github.com/repos/deepmind/dm_control/tarball/7a36377879c57777e5d5b4da5aae2cd2a29b607a', # pylint: disable=line-too-long; # noqa: E501\n]\n\nEXTRAS['all'] = list(set(sum(EXTRAS.values(), [])))\n\n# dependencies for using gpu, not included in 'all'\nEXTRAS['gpu'] = ['tensorflow-gpu' + TF_VERSION]\n\n# Development dependencies (*not* included in 'all')\nEXTRAS['dev'] = [\n # Please keep alphabetized\n 'baselines @ https://api.github.com/repos/openai/baselines/tarball/f2729693253c0ef4d4086231d36e0a4307ec1cb3', # pylint: disable=line-too-long; # noqa: E501\n 'flake8',\n 'flake8-docstrings>=1.5.0',\n 'flake8-import-order',\n 'gtimer',\n 'pandas',\n 'pep8-naming==0.7.0',\n 'pre-commit',\n 'pycodestyle>=2.5.0',\n 'pydocstyle>=4.0.0',\n 'pylint>=2.4.3',\n 'pytest>=3.6', # Required for pytest-cov on Python 3.6\n 'pytest-cov',\n 'pytest-xdist',\n 'recommonmark',\n 'rlkit @ git+https://github.com/vitchyr/rlkit/@1d469a509b797ca04a39b8734c1816ca7d108fc8', # pylint: disable=line-too-long; # noqa: E501\n 'seaborn',\n 'sphinx',\n 'sphinx_rtd_theme',\n 'yapf==0.28.0',\n]\n\nwith open('README.md') as f:\n README = f.read()\n\n# Get the package version dynamically\nwith open('VERSION') as v:\n VERSION = v.read().strip()\n\nsetup(\n name='garage',\n version=VERSION,\n author='Reinforcement Learning Working Group',\n description='A toolkit for reproducible reinforcement learning research',\n url='https://github.com/rlworkgroup/garage',\n packages=find_packages(where='src'),\n package_dir={'': 'src'},\n scripts=['scripts/garage'],\n python_requires='>=3.5',\n install_requires=REQUIRED,\n extras_require=EXTRAS,\n license='MIT',\n long_description=README,\n long_description_content_type='text/markdown',\n classifiers=[\n 'Development Status :: 4 - Beta',\n 'Intended Audience :: Developers',\n 'Intended Audience :: Education',\n 'Intended Audience :: Science/Research',\n 'License :: OSI Approved :: MIT License',\n 'Programming Language :: Python :: 3.5',\n 'Programming Language :: Python :: 3.6',\n 'Programming Language :: Python :: 3.7',\n 'Programming Language :: Python :: 3 :: Only',\n 'Topic :: Scientific/Engineering :: Artificial Intelligence',\n 'Topic :: Scientific/Engineering :: Mathematics',\n 'Topic :: Software Development :: Libraries',\n ],\n)\n", "path": "setup.py"}]}
| 2,031 | 127 |
gh_patches_debug_20214
|
rasdani/github-patches
|
git_diff
|
getsentry__sentry-python-921
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Not working with older boto version
Hello, we use it in Django==2.1.7 app and this row breaks the app.
https://github.com/getsentry/sentry-python/blob/cc08a6bed116e09db41c712c20ab63eb0a839e41/sentry_sdk/integrations/boto3.py#L36
For versions
boto3==1.7.45
botocore==1.10.84
this throws
`
AttributeError: 'str' object has no attribute 'hyphenize'`
I'm not sure the base of the integrations but I thought they must be enabled in settings, but this part of Boto3Integration is triggered even if we have not enabled it in django settings.
</issue>
<code>
[start of sentry_sdk/integrations/boto3.py]
1 from __future__ import absolute_import
2
3 from sentry_sdk import Hub
4 from sentry_sdk.integrations import Integration, DidNotEnable
5 from sentry_sdk.tracing import Span
6
7 from sentry_sdk._functools import partial
8 from sentry_sdk._types import MYPY
9
10 if MYPY:
11 from typing import Any
12 from typing import Dict
13 from typing import Optional
14 from typing import Type
15
16 try:
17 from botocore.client import BaseClient # type: ignore
18 from botocore.response import StreamingBody # type: ignore
19 from botocore.awsrequest import AWSRequest # type: ignore
20 except ImportError:
21 raise DidNotEnable("botocore is not installed")
22
23
24 class Boto3Integration(Integration):
25 identifier = "boto3"
26
27 @staticmethod
28 def setup_once():
29 # type: () -> None
30 orig_init = BaseClient.__init__
31
32 def sentry_patched_init(self, *args, **kwargs):
33 # type: (Type[BaseClient], *Any, **Any) -> None
34 orig_init(self, *args, **kwargs)
35 meta = self.meta
36 service_id = meta.service_model.service_id.hyphenize()
37 meta.events.register(
38 "request-created",
39 partial(_sentry_request_created, service_id=service_id),
40 )
41 meta.events.register("after-call", _sentry_after_call)
42 meta.events.register("after-call-error", _sentry_after_call_error)
43
44 BaseClient.__init__ = sentry_patched_init
45
46
47 def _sentry_request_created(service_id, request, operation_name, **kwargs):
48 # type: (str, AWSRequest, str, **Any) -> None
49 hub = Hub.current
50 if hub.get_integration(Boto3Integration) is None:
51 return
52
53 description = "aws.%s.%s" % (service_id, operation_name)
54 span = hub.start_span(
55 hub=hub,
56 op="aws.request",
57 description=description,
58 )
59 span.set_tag("aws.service_id", service_id)
60 span.set_tag("aws.operation_name", operation_name)
61 span.set_data("aws.request.url", request.url)
62
63 # We do it in order for subsequent http calls/retries be
64 # attached to this span.
65 span.__enter__()
66
67 # request.context is an open-ended data-structure
68 # where we can add anything useful in request life cycle.
69 request.context["_sentrysdk_span"] = span
70
71
72 def _sentry_after_call(context, parsed, **kwargs):
73 # type: (Dict[str, Any], Dict[str, Any], **Any) -> None
74 span = context.pop("_sentrysdk_span", None) # type: Optional[Span]
75
76 # Span could be absent if the integration is disabled.
77 if span is None:
78 return
79 span.__exit__(None, None, None)
80
81 body = parsed.get("Body")
82 if not isinstance(body, StreamingBody):
83 return
84
85 streaming_span = span.start_child(
86 op="aws.request.stream",
87 description=span.description,
88 )
89
90 orig_read = body.read
91 orig_close = body.close
92
93 def sentry_streaming_body_read(*args, **kwargs):
94 # type: (*Any, **Any) -> bytes
95 try:
96 ret = orig_read(*args, **kwargs)
97 if not ret:
98 streaming_span.finish()
99 return ret
100 except Exception:
101 streaming_span.finish()
102 raise
103
104 body.read = sentry_streaming_body_read
105
106 def sentry_streaming_body_close(*args, **kwargs):
107 # type: (*Any, **Any) -> None
108 streaming_span.finish()
109 orig_close(*args, **kwargs)
110
111 body.close = sentry_streaming_body_close
112
113
114 def _sentry_after_call_error(context, exception, **kwargs):
115 # type: (Dict[str, Any], Type[BaseException], **Any) -> None
116 span = context.pop("_sentrysdk_span", None) # type: Optional[Span]
117
118 # Span could be absent if the integration is disabled.
119 if span is None:
120 return
121 span.__exit__(type(exception), exception, None)
122
[end of sentry_sdk/integrations/boto3.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/sentry_sdk/integrations/boto3.py b/sentry_sdk/integrations/boto3.py
--- a/sentry_sdk/integrations/boto3.py
+++ b/sentry_sdk/integrations/boto3.py
@@ -14,6 +14,7 @@
from typing import Type
try:
+ from botocore import __version__ as BOTOCORE_VERSION # type: ignore
from botocore.client import BaseClient # type: ignore
from botocore.response import StreamingBody # type: ignore
from botocore.awsrequest import AWSRequest # type: ignore
@@ -27,6 +28,14 @@
@staticmethod
def setup_once():
# type: () -> None
+ try:
+ version = tuple(map(int, BOTOCORE_VERSION.split(".")[:3]))
+ except (ValueError, TypeError):
+ raise DidNotEnable(
+ "Unparsable botocore version: {}".format(BOTOCORE_VERSION)
+ )
+ if version < (1, 12):
+ raise DidNotEnable("Botocore 1.12 or newer is required.")
orig_init = BaseClient.__init__
def sentry_patched_init(self, *args, **kwargs):
|
{"golden_diff": "diff --git a/sentry_sdk/integrations/boto3.py b/sentry_sdk/integrations/boto3.py\n--- a/sentry_sdk/integrations/boto3.py\n+++ b/sentry_sdk/integrations/boto3.py\n@@ -14,6 +14,7 @@\n from typing import Type\n \n try:\n+ from botocore import __version__ as BOTOCORE_VERSION # type: ignore\n from botocore.client import BaseClient # type: ignore\n from botocore.response import StreamingBody # type: ignore\n from botocore.awsrequest import AWSRequest # type: ignore\n@@ -27,6 +28,14 @@\n @staticmethod\n def setup_once():\n # type: () -> None\n+ try:\n+ version = tuple(map(int, BOTOCORE_VERSION.split(\".\")[:3]))\n+ except (ValueError, TypeError):\n+ raise DidNotEnable(\n+ \"Unparsable botocore version: {}\".format(BOTOCORE_VERSION)\n+ )\n+ if version < (1, 12):\n+ raise DidNotEnable(\"Botocore 1.12 or newer is required.\")\n orig_init = BaseClient.__init__\n \n def sentry_patched_init(self, *args, **kwargs):\n", "issue": "Not working with older boto version\nHello, we use it in Django==2.1.7 app and this row breaks the app.\r\n\r\nhttps://github.com/getsentry/sentry-python/blob/cc08a6bed116e09db41c712c20ab63eb0a839e41/sentry_sdk/integrations/boto3.py#L36\r\n\r\nFor versions\r\nboto3==1.7.45\r\nbotocore==1.10.84\r\n\r\nthis throws\r\n`\r\nAttributeError: 'str' object has no attribute 'hyphenize'`\r\n\r\nI'm not sure the base of the integrations but I thought they must be enabled in settings, but this part of Boto3Integration is triggered even if we have not enabled it in django settings.\r\n\n", "before_files": [{"content": "from __future__ import absolute_import\n\nfrom sentry_sdk import Hub\nfrom sentry_sdk.integrations import Integration, DidNotEnable\nfrom sentry_sdk.tracing import Span\n\nfrom sentry_sdk._functools import partial\nfrom sentry_sdk._types import MYPY\n\nif MYPY:\n from typing import Any\n from typing import Dict\n from typing import Optional\n from typing import Type\n\ntry:\n from botocore.client import BaseClient # type: ignore\n from botocore.response import StreamingBody # type: ignore\n from botocore.awsrequest import AWSRequest # type: ignore\nexcept ImportError:\n raise DidNotEnable(\"botocore is not installed\")\n\n\nclass Boto3Integration(Integration):\n identifier = \"boto3\"\n\n @staticmethod\n def setup_once():\n # type: () -> None\n orig_init = BaseClient.__init__\n\n def sentry_patched_init(self, *args, **kwargs):\n # type: (Type[BaseClient], *Any, **Any) -> None\n orig_init(self, *args, **kwargs)\n meta = self.meta\n service_id = meta.service_model.service_id.hyphenize()\n meta.events.register(\n \"request-created\",\n partial(_sentry_request_created, service_id=service_id),\n )\n meta.events.register(\"after-call\", _sentry_after_call)\n meta.events.register(\"after-call-error\", _sentry_after_call_error)\n\n BaseClient.__init__ = sentry_patched_init\n\n\ndef _sentry_request_created(service_id, request, operation_name, **kwargs):\n # type: (str, AWSRequest, str, **Any) -> None\n hub = Hub.current\n if hub.get_integration(Boto3Integration) is None:\n return\n\n description = \"aws.%s.%s\" % (service_id, operation_name)\n span = hub.start_span(\n hub=hub,\n op=\"aws.request\",\n description=description,\n )\n span.set_tag(\"aws.service_id\", service_id)\n span.set_tag(\"aws.operation_name\", operation_name)\n span.set_data(\"aws.request.url\", request.url)\n\n # We do it in order for subsequent http calls/retries be\n # attached to this span.\n span.__enter__()\n\n # request.context is an open-ended data-structure\n # where we can add anything useful in request life cycle.\n request.context[\"_sentrysdk_span\"] = span\n\n\ndef _sentry_after_call(context, parsed, **kwargs):\n # type: (Dict[str, Any], Dict[str, Any], **Any) -> None\n span = context.pop(\"_sentrysdk_span\", None) # type: Optional[Span]\n\n # Span could be absent if the integration is disabled.\n if span is None:\n return\n span.__exit__(None, None, None)\n\n body = parsed.get(\"Body\")\n if not isinstance(body, StreamingBody):\n return\n\n streaming_span = span.start_child(\n op=\"aws.request.stream\",\n description=span.description,\n )\n\n orig_read = body.read\n orig_close = body.close\n\n def sentry_streaming_body_read(*args, **kwargs):\n # type: (*Any, **Any) -> bytes\n try:\n ret = orig_read(*args, **kwargs)\n if not ret:\n streaming_span.finish()\n return ret\n except Exception:\n streaming_span.finish()\n raise\n\n body.read = sentry_streaming_body_read\n\n def sentry_streaming_body_close(*args, **kwargs):\n # type: (*Any, **Any) -> None\n streaming_span.finish()\n orig_close(*args, **kwargs)\n\n body.close = sentry_streaming_body_close\n\n\ndef _sentry_after_call_error(context, exception, **kwargs):\n # type: (Dict[str, Any], Type[BaseException], **Any) -> None\n span = context.pop(\"_sentrysdk_span\", None) # type: Optional[Span]\n\n # Span could be absent if the integration is disabled.\n if span is None:\n return\n span.__exit__(type(exception), exception, None)\n", "path": "sentry_sdk/integrations/boto3.py"}]}
| 1,899 | 285 |
gh_patches_debug_12093
|
rasdani/github-patches
|
git_diff
|
AnalogJ__lexicon-374
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
DNSimple crash with team member that does not have a plan
Here's the error I get (I have redacted my account number, domain and auth token)
```sh
$ lexicon dnsimple --auth-token REDACTED list REDACTED A
Traceback (most recent call last):
File "/usr/local/bin/lexicon", line 11, in <module>
sys.exit(main())
File "/usr/local/lib/python2.7/dist-packages/lexicon/cli.py", line 117, in main
results = client.execute()
File "/usr/local/lib/python2.7/dist-packages/lexicon/client.py", line 64, in execute
self.provider.authenticate()
File "/usr/local/lib/python2.7/dist-packages/lexicon/providers/dnsimple.py", line 43, in authenticate
'/{0}/domains'.format(account['id']), query_params={'name_like': self.domain})
File "/usr/local/lib/python2.7/dist-packages/lexicon/providers/base.py", line 87, in _get
return self._request('GET', url, query_params=query_params)
File "/usr/local/lib/python2.7/dist-packages/lexicon/providers/dnsimple.py", line 189, in _request
r.raise_for_status()
File "/usr/lib/python2.7/dist-packages/requests/models.py", line 935, in raise_for_status
raise HTTPError(http_error_msg, response=self)
requests.exceptions.HTTPError: 402 Client Error: Payment Required for url: https://api.dnsimple.com/v2/REDACTED/domains?name_like=REDACTED
```
I believe that this happens because of the way the domain I have is configured. We have a domain configured with a team. There is one account that has a professional plan (with our domain) and we have other users as team members of that account. The only account that has a plan is the one mentioned. Other users have personal accounts by default but they don't have a plan configured at all.
I believe that the code responsible is this: https://github.com/AnalogJ/lexicon/blob/master/lexicon/providers/dnsimple.py#L37-L52
The dnsimple provider iterates through all accounts that the token has access to and tried to see if they have a domain similar to the one specified. When it tries to do this on an account that is not configured it will crash with the http error `402 Client Error: Payment Required for url: ...`.
This is consistent with the documentation from dnsimple: https://developer.dnsimple.com/v2/#response-codes
> `402 Payment Required` - Your account is not subscribed or not in good standing.
I think that it's possible to figure out if an account is not configured by using the `GET /accounts` endpoint. Accounts that are not configured have their `plan_identifier` set to `null`. This does not seem to be documented in the API references tho.
When I do this request I get the following json as output
```json
{
"data": [
{
"created_at": "2018-06-05T19:23:59Z",
"email": "REDACTED",
"id": "REDACTED",
"plan_identifier": "dnsimple-professional",
"updated_at": "2019-01-06T21:16:40Z"
},
{
"created_at": "2018-06-08T19:48:59Z",
"email": "REDACTED",
"id": "REDACTED",
"plan_identifier": null,
"updated_at": "2018-06-08T19:48:59Z"
}
]
}
```
</issue>
<code>
[start of lexicon/providers/dnsimple.py]
1 """Module provider for DNS Simple"""
2 from __future__ import absolute_import
3 import json
4 import logging
5
6 import requests
7 from lexicon.providers.base import Provider as BaseProvider
8
9
10 LOGGER = logging.getLogger(__name__)
11
12 NAMESERVER_DOMAINS = ['dnsimple.com']
13
14
15 def provider_parser(subparser):
16 """Configure provider parser for DNS Simple"""
17 subparser.add_argument(
18 "--auth-token", help="specify api token for authentication")
19 subparser.add_argument(
20 "--auth-username", help="specify email address for authentication")
21 subparser.add_argument(
22 "--auth-password", help="specify password for authentication")
23 subparser.add_argument(
24 "--auth-2fa",
25 help="specify two-factor auth token (OTP) to use with email/password authentication")
26
27
28 class Provider(BaseProvider):
29 """Provider class for DNS Simple"""
30 def __init__(self, config):
31 super(Provider, self).__init__(config)
32 self.domain_id = None
33 self.account_id = None
34 self.api_endpoint = self._get_provider_option(
35 'api_endpoint') or 'https://api.dnsimple.com/v2'
36
37 def _authenticate(self):
38
39 payload = self._get('/accounts')
40
41 if not payload[0]['id']:
42 raise Exception('No account id found')
43
44 for account in payload:
45 dompayload = self._get(
46 '/{0}/domains'.format(account['id']), query_params={'name_like': self.domain})
47 if dompayload and dompayload[0]['id']:
48 self.account_id = account['id']
49 self.domain_id = dompayload[0]['id']
50
51 if not self.account_id:
52 raise Exception('No domain found like {}'.format(self.domain))
53
54 # Create record. If record already exists with the same content, do nothing
55
56 def _create_record(self, rtype, name, content):
57 # check if record already exists
58 existing_records = self._list_records(rtype, name, content)
59 if len(existing_records) == 1:
60 return True
61
62 record = {
63 'type': rtype,
64 'name': self._relative_name(name),
65 'content': content
66 }
67 if self._get_lexicon_option('ttl'):
68 record['ttl'] = self._get_lexicon_option('ttl')
69 if self._get_lexicon_option('priority'):
70 record['priority'] = self._get_lexicon_option('priority')
71 if self._get_provider_option('regions'):
72 record['regions'] = self._get_provider_option('regions')
73
74 payload = self._post(
75 '/{0}/zones/{1}/records'.format(self.account_id, self.domain), record)
76
77 LOGGER.debug('create_record: %s', 'id' in payload)
78 return 'id' in payload
79
80 # List all records. Return an empty list if no records found
81 # type, name and content are used to filter records.
82 # If possible filter during the query, otherwise filter after response is received.
83 def _list_records(self, rtype=None, name=None, content=None):
84 filter_query = {}
85 if rtype:
86 filter_query['type'] = rtype
87 if name:
88 filter_query['name'] = self._relative_name(name)
89 payload = self._get(
90 '/{0}/zones/{1}/records'.format(self.account_id, self.domain),
91 query_params=filter_query)
92
93 records = []
94 for record in payload:
95 processed_record = {
96 'type': record['type'],
97 'name': '{}'.format(
98 self.domain) if record['name'] == "" else '{0}.{1}'.format(
99 record['name'],
100 self.domain),
101 'ttl': record['ttl'],
102 'content': record['content'],
103 'id': record['id']}
104 if record['priority']:
105 processed_record['priority'] = record['priority']
106 records.append(processed_record)
107
108 if content:
109 records = [
110 record for record in records if record['content'] == content]
111
112 LOGGER.debug('list_records: %s', records)
113 return records
114
115 # Create or update a record.
116 def _update_record(self, identifier, rtype=None, name=None, content=None):
117
118 data = {}
119
120 if identifier is None:
121 records = self._list_records(rtype, name, content)
122 identifiers = [record["id"] for record in records]
123 else:
124 identifiers = [identifier]
125
126 if name:
127 data['name'] = self._relative_name(name)
128 if content:
129 data['content'] = content
130 if self._get_lexicon_option('ttl'):
131 data['ttl'] = self._get_lexicon_option('ttl')
132 if self._get_lexicon_option('priority'):
133 data['priority'] = self._get_lexicon_option('priority')
134 if self._get_provider_option('regions'):
135 data['regions'] = self._get_provider_option('regions')
136
137 for one_identifier in identifiers:
138 self._patch('/{0}/zones/{1}/records/{2}'
139 .format(self.account_id, self.domain, one_identifier), data)
140 LOGGER.debug('update_record: %s', one_identifier)
141
142 LOGGER.debug('update_record: %s', True)
143 return True
144
145 # Delete an existing record.
146 # If record does not exist, do nothing.
147 def _delete_record(self, identifier=None, rtype=None, name=None, content=None):
148 delete_record_id = []
149 if not identifier:
150 records = self._list_records(rtype, name, content)
151 delete_record_id = [record['id'] for record in records]
152 else:
153 delete_record_id.append(identifier)
154
155 LOGGER.debug('delete_records: %s', delete_record_id)
156
157 for record_id in delete_record_id:
158 self._delete(
159 '/{0}/zones/{1}/records/{2}'.format(self.account_id, self.domain, record_id))
160
161 # is always True at this point; if a non 2xx response is returned, an error is raised.
162 LOGGER.debug('delete_record: True')
163 return True
164
165 # Helpers
166
167 def _request(self, action='GET', url='/', data=None, query_params=None):
168 if data is None:
169 data = {}
170 if query_params is None:
171 query_params = {}
172 default_headers = {
173 'Accept': 'application/json',
174 'Content-Type': 'application/json'
175 }
176 default_auth = None
177
178 if self._get_provider_option('auth_token'):
179 default_headers['Authorization'] = "Bearer {0}".format(
180 self._get_provider_option('auth_token'))
181 elif (self._get_provider_option('auth_username')
182 and self._get_provider_option('auth_password')):
183 default_auth = (self._get_provider_option(
184 'auth_username'), self._get_provider_option('auth_password'))
185 if self._get_provider_option('auth_2fa'):
186 default_headers['X-Dnsimple-OTP'] = self._get_provider_option(
187 'auth_2fa')
188 else:
189 raise Exception('No valid authentication mechanism found')
190
191 response = requests.request(action, self.api_endpoint + url, params=query_params,
192 data=json.dumps(data),
193 headers=default_headers,
194 auth=default_auth)
195 # if the request fails for any reason, throw an error.
196 response.raise_for_status()
197 if response.text and response.json()['data'] is None:
198 raise Exception('No data returned')
199
200 return response.json()['data'] if response.text else None
201
202 def _patch(self, url='/', data=None, query_params=None):
203 return self._request('PATCH', url, data=data, query_params=query_params)
204
[end of lexicon/providers/dnsimple.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/lexicon/providers/dnsimple.py b/lexicon/providers/dnsimple.py
--- a/lexicon/providers/dnsimple.py
+++ b/lexicon/providers/dnsimple.py
@@ -42,6 +42,13 @@
raise Exception('No account id found')
for account in payload:
+ if account['plan_identifier'] is None:
+ logging.warning(
+ 'Skipping unconfigured account %s (%d). ' \
+ 'To use this account, you must select a plan.',
+ account['email'], account['id'])
+ continue
+
dompayload = self._get(
'/{0}/domains'.format(account['id']), query_params={'name_like': self.domain})
if dompayload and dompayload[0]['id']:
|
{"golden_diff": "diff --git a/lexicon/providers/dnsimple.py b/lexicon/providers/dnsimple.py\n--- a/lexicon/providers/dnsimple.py\n+++ b/lexicon/providers/dnsimple.py\n@@ -42,6 +42,13 @@\n raise Exception('No account id found')\n \n for account in payload:\n+ if account['plan_identifier'] is None:\n+ logging.warning(\n+ 'Skipping unconfigured account %s (%d). ' \\\n+ 'To use this account, you must select a plan.',\n+ account['email'], account['id'])\n+ continue\n+\n dompayload = self._get(\n '/{0}/domains'.format(account['id']), query_params={'name_like': self.domain})\n if dompayload and dompayload[0]['id']:\n", "issue": "DNSimple crash with team member that does not have a plan\nHere's the error I get (I have redacted my account number, domain and auth token)\r\n\r\n```sh\r\n$ lexicon dnsimple --auth-token REDACTED list REDACTED A\r\nTraceback (most recent call last):\r\n File \"/usr/local/bin/lexicon\", line 11, in <module>\r\n sys.exit(main())\r\n File \"/usr/local/lib/python2.7/dist-packages/lexicon/cli.py\", line 117, in main\r\n results = client.execute()\r\n File \"/usr/local/lib/python2.7/dist-packages/lexicon/client.py\", line 64, in execute\r\n self.provider.authenticate()\r\n File \"/usr/local/lib/python2.7/dist-packages/lexicon/providers/dnsimple.py\", line 43, in authenticate\r\n '/{0}/domains'.format(account['id']), query_params={'name_like': self.domain})\r\n File \"/usr/local/lib/python2.7/dist-packages/lexicon/providers/base.py\", line 87, in _get\r\n return self._request('GET', url, query_params=query_params)\r\n File \"/usr/local/lib/python2.7/dist-packages/lexicon/providers/dnsimple.py\", line 189, in _request\r\n r.raise_for_status()\r\n File \"/usr/lib/python2.7/dist-packages/requests/models.py\", line 935, in raise_for_status\r\n raise HTTPError(http_error_msg, response=self)\r\nrequests.exceptions.HTTPError: 402 Client Error: Payment Required for url: https://api.dnsimple.com/v2/REDACTED/domains?name_like=REDACTED\r\n```\r\n\r\nI believe that this happens because of the way the domain I have is configured. We have a domain configured with a team. There is one account that has a professional plan (with our domain) and we have other users as team members of that account. The only account that has a plan is the one mentioned. Other users have personal accounts by default but they don't have a plan configured at all.\r\n\r\nI believe that the code responsible is this: https://github.com/AnalogJ/lexicon/blob/master/lexicon/providers/dnsimple.py#L37-L52\r\n\r\nThe dnsimple provider iterates through all accounts that the token has access to and tried to see if they have a domain similar to the one specified. When it tries to do this on an account that is not configured it will crash with the http error `402 Client Error: Payment Required for url: ...`.\r\n\r\nThis is consistent with the documentation from dnsimple: https://developer.dnsimple.com/v2/#response-codes\r\n\r\n> `402 Payment Required` - Your account is not subscribed or not in good standing.\r\n\r\nI think that it's possible to figure out if an account is not configured by using the `GET /accounts` endpoint. Accounts that are not configured have their `plan_identifier` set to `null`. This does not seem to be documented in the API references tho.\r\n\r\nWhen I do this request I get the following json as output\r\n\r\n```json\r\n{\r\n \"data\": [\r\n {\r\n \"created_at\": \"2018-06-05T19:23:59Z\",\r\n \"email\": \"REDACTED\",\r\n \"id\": \"REDACTED\",\r\n \"plan_identifier\": \"dnsimple-professional\",\r\n \"updated_at\": \"2019-01-06T21:16:40Z\"\r\n },\r\n {\r\n \"created_at\": \"2018-06-08T19:48:59Z\",\r\n \"email\": \"REDACTED\",\r\n \"id\": \"REDACTED\",\r\n \"plan_identifier\": null,\r\n \"updated_at\": \"2018-06-08T19:48:59Z\"\r\n }\r\n ]\r\n}\r\n```\n", "before_files": [{"content": "\"\"\"Module provider for DNS Simple\"\"\"\nfrom __future__ import absolute_import\nimport json\nimport logging\n\nimport requests\nfrom lexicon.providers.base import Provider as BaseProvider\n\n\nLOGGER = logging.getLogger(__name__)\n\nNAMESERVER_DOMAINS = ['dnsimple.com']\n\n\ndef provider_parser(subparser):\n \"\"\"Configure provider parser for DNS Simple\"\"\"\n subparser.add_argument(\n \"--auth-token\", help=\"specify api token for authentication\")\n subparser.add_argument(\n \"--auth-username\", help=\"specify email address for authentication\")\n subparser.add_argument(\n \"--auth-password\", help=\"specify password for authentication\")\n subparser.add_argument(\n \"--auth-2fa\",\n help=\"specify two-factor auth token (OTP) to use with email/password authentication\")\n\n\nclass Provider(BaseProvider):\n \"\"\"Provider class for DNS Simple\"\"\"\n def __init__(self, config):\n super(Provider, self).__init__(config)\n self.domain_id = None\n self.account_id = None\n self.api_endpoint = self._get_provider_option(\n 'api_endpoint') or 'https://api.dnsimple.com/v2'\n\n def _authenticate(self):\n\n payload = self._get('/accounts')\n\n if not payload[0]['id']:\n raise Exception('No account id found')\n\n for account in payload:\n dompayload = self._get(\n '/{0}/domains'.format(account['id']), query_params={'name_like': self.domain})\n if dompayload and dompayload[0]['id']:\n self.account_id = account['id']\n self.domain_id = dompayload[0]['id']\n\n if not self.account_id:\n raise Exception('No domain found like {}'.format(self.domain))\n\n # Create record. If record already exists with the same content, do nothing\n\n def _create_record(self, rtype, name, content):\n # check if record already exists\n existing_records = self._list_records(rtype, name, content)\n if len(existing_records) == 1:\n return True\n\n record = {\n 'type': rtype,\n 'name': self._relative_name(name),\n 'content': content\n }\n if self._get_lexicon_option('ttl'):\n record['ttl'] = self._get_lexicon_option('ttl')\n if self._get_lexicon_option('priority'):\n record['priority'] = self._get_lexicon_option('priority')\n if self._get_provider_option('regions'):\n record['regions'] = self._get_provider_option('regions')\n\n payload = self._post(\n '/{0}/zones/{1}/records'.format(self.account_id, self.domain), record)\n\n LOGGER.debug('create_record: %s', 'id' in payload)\n return 'id' in payload\n\n # List all records. Return an empty list if no records found\n # type, name and content are used to filter records.\n # If possible filter during the query, otherwise filter after response is received.\n def _list_records(self, rtype=None, name=None, content=None):\n filter_query = {}\n if rtype:\n filter_query['type'] = rtype\n if name:\n filter_query['name'] = self._relative_name(name)\n payload = self._get(\n '/{0}/zones/{1}/records'.format(self.account_id, self.domain),\n query_params=filter_query)\n\n records = []\n for record in payload:\n processed_record = {\n 'type': record['type'],\n 'name': '{}'.format(\n self.domain) if record['name'] == \"\" else '{0}.{1}'.format(\n record['name'],\n self.domain),\n 'ttl': record['ttl'],\n 'content': record['content'],\n 'id': record['id']}\n if record['priority']:\n processed_record['priority'] = record['priority']\n records.append(processed_record)\n\n if content:\n records = [\n record for record in records if record['content'] == content]\n\n LOGGER.debug('list_records: %s', records)\n return records\n\n # Create or update a record.\n def _update_record(self, identifier, rtype=None, name=None, content=None):\n\n data = {}\n\n if identifier is None:\n records = self._list_records(rtype, name, content)\n identifiers = [record[\"id\"] for record in records]\n else:\n identifiers = [identifier]\n\n if name:\n data['name'] = self._relative_name(name)\n if content:\n data['content'] = content\n if self._get_lexicon_option('ttl'):\n data['ttl'] = self._get_lexicon_option('ttl')\n if self._get_lexicon_option('priority'):\n data['priority'] = self._get_lexicon_option('priority')\n if self._get_provider_option('regions'):\n data['regions'] = self._get_provider_option('regions')\n\n for one_identifier in identifiers:\n self._patch('/{0}/zones/{1}/records/{2}'\n .format(self.account_id, self.domain, one_identifier), data)\n LOGGER.debug('update_record: %s', one_identifier)\n\n LOGGER.debug('update_record: %s', True)\n return True\n\n # Delete an existing record.\n # If record does not exist, do nothing.\n def _delete_record(self, identifier=None, rtype=None, name=None, content=None):\n delete_record_id = []\n if not identifier:\n records = self._list_records(rtype, name, content)\n delete_record_id = [record['id'] for record in records]\n else:\n delete_record_id.append(identifier)\n\n LOGGER.debug('delete_records: %s', delete_record_id)\n\n for record_id in delete_record_id:\n self._delete(\n '/{0}/zones/{1}/records/{2}'.format(self.account_id, self.domain, record_id))\n\n # is always True at this point; if a non 2xx response is returned, an error is raised.\n LOGGER.debug('delete_record: True')\n return True\n\n # Helpers\n\n def _request(self, action='GET', url='/', data=None, query_params=None):\n if data is None:\n data = {}\n if query_params is None:\n query_params = {}\n default_headers = {\n 'Accept': 'application/json',\n 'Content-Type': 'application/json'\n }\n default_auth = None\n\n if self._get_provider_option('auth_token'):\n default_headers['Authorization'] = \"Bearer {0}\".format(\n self._get_provider_option('auth_token'))\n elif (self._get_provider_option('auth_username')\n and self._get_provider_option('auth_password')):\n default_auth = (self._get_provider_option(\n 'auth_username'), self._get_provider_option('auth_password'))\n if self._get_provider_option('auth_2fa'):\n default_headers['X-Dnsimple-OTP'] = self._get_provider_option(\n 'auth_2fa')\n else:\n raise Exception('No valid authentication mechanism found')\n\n response = requests.request(action, self.api_endpoint + url, params=query_params,\n data=json.dumps(data),\n headers=default_headers,\n auth=default_auth)\n # if the request fails for any reason, throw an error.\n response.raise_for_status()\n if response.text and response.json()['data'] is None:\n raise Exception('No data returned')\n\n return response.json()['data'] if response.text else None\n\n def _patch(self, url='/', data=None, query_params=None):\n return self._request('PATCH', url, data=data, query_params=query_params)\n", "path": "lexicon/providers/dnsimple.py"}]}
| 3,533 | 173 |
gh_patches_debug_9048
|
rasdani/github-patches
|
git_diff
|
frappe__frappe-20401
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
SQL SyntaxError in queries in DocTypes with Table Multiselect fields
### App Versions
```json
{
"frappe": "15.0.0-dev"
}
```
### Route
`List/User Group/List` and `Form/User Group/new-user-group-1`
### Traceback
```python
Syntax error in query:
select distinct
from `tabUser Group`
Traceback (most recent call last):
File "apps/frappe/frappe/app.py", line 53, in application
response = frappe.api.handle()
File "apps/frappe/frappe/api.py", line 53, in handle
return _RESTAPIHandler(call, doctype, name).get_response()
File "apps/frappe/frappe/api.py", line 69, in get_response
return self.handle_method()
File "apps/frappe/frappe/api.py", line 79, in handle_method
return frappe.handler.handle()
File "apps/frappe/frappe/handler.py", line 48, in handle
data = execute_cmd(cmd)
File "apps/frappe/frappe/handler.py", line 86, in execute_cmd
return frappe.call(method, **frappe.form_dict)
File "apps/frappe/frappe/__init__.py", line 1591, in call
return fn(*args, **newargs)
File "apps/frappe/frappe/utils/typing_validations.py", line 33, in wrapper
return func(*args, **kwargs)
File "apps/frappe/frappe/__init__.py", line 796, in wrapper_fn
retval = fn(*args, **get_newargs(fn, kwargs))
File "apps/frappe/frappe/desk/reportview.py", line 58, in get_count
data = execute(**args)[0].get("total_count")
File "apps/frappe/frappe/desk/reportview.py", line 64, in execute
return DatabaseQuery(doctype).execute(*args, **kwargs)
File "apps/frappe/frappe/model/db_query.py", line 202, in execute
result = self.build_and_run()
File "apps/frappe/frappe/model/db_query.py", line 242, in build_and_run
return frappe.db.sql(
File "apps/frappe/frappe/database/database.py", line 222, in sql
self._cursor.execute(query, values)
File "env/lib/python3.10/site-packages/pymysql/cursors.py", line 148, in execute
result = self._query(query)
File "env/lib/python3.10/site-packages/pymysql/cursors.py", line 310, in _query
conn.query(q)
File "env/lib/python3.10/site-packages/pymysql/connections.py", line 548, in query
self._affected_rows = self._read_query_result(unbuffered=unbuffered)
File "env/lib/python3.10/site-packages/pymysql/connections.py", line 775, in _read_query_result
result.read()
File "env/lib/python3.10/site-packages/pymysql/connections.py", line 1156, in read
first_packet = self.connection._read_packet()
File "env/lib/python3.10/site-packages/pymysql/connections.py", line 725, in _read_packet
packet.raise_for_error()
File "env/lib/python3.10/site-packages/pymysql/protocol.py", line 221, in raise_for_error
err.raise_mysql_exception(self._data)
File "env/lib/python3.10/site-packages/pymysql/err.py", line 143, in raise_mysql_exception
raise errorclass(errno, errval)
pymysql.err.ProgrammingError: (1064, "You have an error in your SQL syntax; check the manual that corresponds to your MariaDB server version for the right syntax to use near 'from `tabUser Group`' at line 2")
```
### Request Data
```json
{
"type": "POST",
"args": {
"doctype": "User Group",
"filters": "[]",
"fields": "[]",
"distinct": false
},
"headers": {},
"error_handlers": {},
"url": "/api/method/frappe.desk.reportview.get_count"
}
```
### Response Data
```json
{
"exception": "pymysql.err.ProgrammingError: (1064, \"You have an error in your SQL syntax; check the manual that corresponds to your MariaDB server version for the right syntax to use near 'from `tabUser Group`' at line 2\")",
"_debug_messages": "[\"method:frappe.desk.reportview.get_count\", \"<<<< query\\nSELECT `is_virtual` FROM `tabDocType` WHERE `name`='User Group' ORDER BY `modified` DESC LIMIT 1\\n>>>>\", \"<<<< query\\nselect `share_name`\\n\\t\\t\\tfrom `tabDocShare`\\n\\t\\t\\twhere `tabDocShare`.`read` = 1.0 and `tabDocShare`.`share_doctype` = 'User Group' and (`tabDocShare`.`user` = 'Administrator' or `tabDocShare`.`everyone` = 1.0)\\n\\t\\t\\t\\n\\t\\t\\t order by `tabDocShare`.`modified` DESC\\n>>>>\"]"
}
```
</issue>
<code>
[start of frappe/model/__init__.py]
1 # Copyright (c) 2015, Frappe Technologies Pvt. Ltd. and Contributors
2 # License: MIT. See LICENSE
3
4 # model __init__.py
5 import frappe
6
7 data_fieldtypes = (
8 "Currency",
9 "Int",
10 "Long Int",
11 "Float",
12 "Percent",
13 "Check",
14 "Small Text",
15 "Long Text",
16 "Code",
17 "Text Editor",
18 "Markdown Editor",
19 "HTML Editor",
20 "Date",
21 "Datetime",
22 "Time",
23 "Text",
24 "Data",
25 "Link",
26 "Dynamic Link",
27 "Password",
28 "Select",
29 "Rating",
30 "Read Only",
31 "Attach",
32 "Attach Image",
33 "Signature",
34 "Color",
35 "Barcode",
36 "Geolocation",
37 "Duration",
38 "Icon",
39 "Phone",
40 "Autocomplete",
41 "JSON",
42 )
43
44 float_like_fields = {"Float", "Currency", "Percent"}
45 datetime_fields = {"Datetime", "Date", "Time"}
46
47 attachment_fieldtypes = (
48 "Attach",
49 "Attach Image",
50 )
51
52 no_value_fields = (
53 "Section Break",
54 "Column Break",
55 "Tab Break",
56 "HTML",
57 "Table",
58 "Table MultiSelect",
59 "Button",
60 "Image",
61 "Fold",
62 "Heading",
63 )
64
65 display_fieldtypes = (
66 "Section Break",
67 "Column Break",
68 "Tab Break",
69 "HTML",
70 "Button",
71 "Image",
72 "Fold",
73 "Heading",
74 )
75
76 numeric_fieldtypes = ("Currency", "Int", "Long Int", "Float", "Percent", "Check")
77
78 data_field_options = ("Email", "Name", "Phone", "URL", "Barcode")
79
80 default_fields = (
81 "doctype",
82 "name",
83 "owner",
84 "creation",
85 "modified",
86 "modified_by",
87 "docstatus",
88 "idx",
89 )
90
91 child_table_fields = ("parent", "parentfield", "parenttype")
92
93 optional_fields = ("_user_tags", "_comments", "_assign", "_liked_by", "_seen")
94
95 table_fields = ("Table", "Table MultiSelect")
96
97 core_doctypes_list = (
98 "DefaultValue",
99 "DocType",
100 "DocField",
101 "DocPerm",
102 "DocType Action",
103 "DocType Link",
104 "User",
105 "Role",
106 "Has Role",
107 "Page",
108 "Module Def",
109 "Print Format",
110 "Report",
111 "Customize Form",
112 "Customize Form Field",
113 "Property Setter",
114 "Custom Field",
115 "Client Script",
116 )
117
118 # NOTE: this is being used for dynamic autoincrement in new sites,
119 # removing any of these will require patches.
120 log_types = (
121 "Version",
122 "Error Log",
123 "Scheduled Job Log",
124 "Event Sync Log",
125 "Event Update Log",
126 "Access Log",
127 "View Log",
128 "Activity Log",
129 "Energy Point Log",
130 "Notification Log",
131 "Email Queue",
132 "DocShare",
133 "Document Follow",
134 "Console Log",
135 )
136
137
138 def delete_fields(args_dict, delete=0):
139 """
140 Delete a field.
141 * Deletes record from `tabDocField`
142 * If not single doctype: Drops column from table
143 * If single, deletes record from `tabSingles`
144 args_dict = { dt: [field names] }
145 """
146 import frappe.utils
147
148 for dt in args_dict:
149 fields = args_dict[dt]
150 if not fields:
151 continue
152
153 frappe.db.delete(
154 "DocField",
155 {
156 "parent": dt,
157 "fieldname": ("in", fields),
158 },
159 )
160
161 # Delete the data/column only if delete is specified
162 if not delete:
163 continue
164
165 if frappe.db.get_value("DocType", dt, "issingle"):
166 frappe.db.delete(
167 "Singles",
168 {
169 "doctype": dt,
170 "field": ("in", fields),
171 },
172 )
173 else:
174 existing_fields = frappe.db.describe(dt)
175 existing_fields = existing_fields and [e[0] for e in existing_fields] or []
176 fields_need_to_delete = set(fields) & set(existing_fields)
177 if not fields_need_to_delete:
178 continue
179
180 if frappe.db.db_type == "mariadb":
181 # mariadb implicitly commits before DDL, make it explicit
182 frappe.db.commit()
183
184 query = "ALTER TABLE `tab%s` " % dt + ", ".join(
185 "DROP COLUMN `%s`" % f for f in fields_need_to_delete
186 )
187 frappe.db.sql(query)
188
189 if frappe.db.db_type == "postgres":
190 # commit the results to db
191 frappe.db.commit()
192
193
194 def get_permitted_fields(
195 doctype: str, parenttype: str | None = None, user: str | None = None
196 ) -> list[str]:
197 meta = frappe.get_meta(doctype)
198 valid_columns = meta.get_valid_columns()
199
200 if doctype in core_doctypes_list:
201 return valid_columns
202
203 if permitted_fields := meta.get_permitted_fieldnames(parenttype=parenttype, user=user):
204 meta_fields = meta.default_fields.copy()
205 optional_meta_fields = [x for x in optional_fields if x in valid_columns]
206
207 if meta.istable:
208 meta_fields.extend(child_table_fields)
209
210 return meta_fields + permitted_fields + optional_meta_fields
211
212 return []
213
[end of frappe/model/__init__.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/frappe/model/__init__.py b/frappe/model/__init__.py
--- a/frappe/model/__init__.py
+++ b/frappe/model/__init__.py
@@ -200,6 +200,10 @@
if doctype in core_doctypes_list:
return valid_columns
+ # DocType has only fields of type Table (Table, Table MultiSelect)
+ if set(valid_columns).issubset(default_fields):
+ return valid_columns
+
if permitted_fields := meta.get_permitted_fieldnames(parenttype=parenttype, user=user):
meta_fields = meta.default_fields.copy()
optional_meta_fields = [x for x in optional_fields if x in valid_columns]
|
{"golden_diff": "diff --git a/frappe/model/__init__.py b/frappe/model/__init__.py\n--- a/frappe/model/__init__.py\n+++ b/frappe/model/__init__.py\n@@ -200,6 +200,10 @@\n \tif doctype in core_doctypes_list:\n \t\treturn valid_columns\n \n+\t# DocType has only fields of type Table (Table, Table MultiSelect)\n+\tif set(valid_columns).issubset(default_fields):\n+\t\treturn valid_columns\n+\n \tif permitted_fields := meta.get_permitted_fieldnames(parenttype=parenttype, user=user):\n \t\tmeta_fields = meta.default_fields.copy()\n \t\toptional_meta_fields = [x for x in optional_fields if x in valid_columns]\n", "issue": "SQL SyntaxError in queries in DocTypes with Table Multiselect fields\n### App Versions\r\n```json\r\n{\r\n\t\"frappe\": \"15.0.0-dev\"\r\n}\r\n```\r\n### Route\r\n\r\n`List/User Group/List` and `Form/User Group/new-user-group-1`\r\n\r\n### Traceback\r\n\r\n```python\r\nSyntax error in query:\r\nselect distinct \r\n\t\t\tfrom `tabUser Group` \r\nTraceback (most recent call last):\r\n File \"apps/frappe/frappe/app.py\", line 53, in application\r\n response = frappe.api.handle()\r\n File \"apps/frappe/frappe/api.py\", line 53, in handle\r\n return _RESTAPIHandler(call, doctype, name).get_response()\r\n File \"apps/frappe/frappe/api.py\", line 69, in get_response\r\n return self.handle_method()\r\n File \"apps/frappe/frappe/api.py\", line 79, in handle_method\r\n return frappe.handler.handle()\r\n File \"apps/frappe/frappe/handler.py\", line 48, in handle\r\n data = execute_cmd(cmd)\r\n File \"apps/frappe/frappe/handler.py\", line 86, in execute_cmd\r\n return frappe.call(method, **frappe.form_dict)\r\n File \"apps/frappe/frappe/__init__.py\", line 1591, in call\r\n return fn(*args, **newargs)\r\n File \"apps/frappe/frappe/utils/typing_validations.py\", line 33, in wrapper\r\n return func(*args, **kwargs)\r\n File \"apps/frappe/frappe/__init__.py\", line 796, in wrapper_fn\r\n retval = fn(*args, **get_newargs(fn, kwargs))\r\n File \"apps/frappe/frappe/desk/reportview.py\", line 58, in get_count\r\n data = execute(**args)[0].get(\"total_count\")\r\n File \"apps/frappe/frappe/desk/reportview.py\", line 64, in execute\r\n return DatabaseQuery(doctype).execute(*args, **kwargs)\r\n File \"apps/frappe/frappe/model/db_query.py\", line 202, in execute\r\n result = self.build_and_run()\r\n File \"apps/frappe/frappe/model/db_query.py\", line 242, in build_and_run\r\n return frappe.db.sql(\r\n File \"apps/frappe/frappe/database/database.py\", line 222, in sql\r\n self._cursor.execute(query, values)\r\n File \"env/lib/python3.10/site-packages/pymysql/cursors.py\", line 148, in execute\r\n result = self._query(query)\r\n File \"env/lib/python3.10/site-packages/pymysql/cursors.py\", line 310, in _query\r\n conn.query(q)\r\n File \"env/lib/python3.10/site-packages/pymysql/connections.py\", line 548, in query\r\n self._affected_rows = self._read_query_result(unbuffered=unbuffered)\r\n File \"env/lib/python3.10/site-packages/pymysql/connections.py\", line 775, in _read_query_result\r\n result.read()\r\n File \"env/lib/python3.10/site-packages/pymysql/connections.py\", line 1156, in read\r\n first_packet = self.connection._read_packet()\r\n File \"env/lib/python3.10/site-packages/pymysql/connections.py\", line 725, in _read_packet\r\n packet.raise_for_error()\r\n File \"env/lib/python3.10/site-packages/pymysql/protocol.py\", line 221, in raise_for_error\r\n err.raise_mysql_exception(self._data)\r\n File \"env/lib/python3.10/site-packages/pymysql/err.py\", line 143, in raise_mysql_exception\r\n raise errorclass(errno, errval)\r\npymysql.err.ProgrammingError: (1064, \"You have an error in your SQL syntax; check the manual that corresponds to your MariaDB server version for the right syntax to use near 'from `tabUser Group`' at line 2\")\r\n\r\n```\r\n### Request Data\r\n\r\n```json\r\n{\r\n\t\"type\": \"POST\",\r\n\t\"args\": {\r\n\t\t\"doctype\": \"User Group\",\r\n\t\t\"filters\": \"[]\",\r\n\t\t\"fields\": \"[]\",\r\n\t\t\"distinct\": false\r\n\t},\r\n\t\"headers\": {},\r\n\t\"error_handlers\": {},\r\n\t\"url\": \"/api/method/frappe.desk.reportview.get_count\"\r\n}\r\n```\r\n### Response Data\r\n\r\n```json\r\n{\r\n\t\"exception\": \"pymysql.err.ProgrammingError: (1064, \\\"You have an error in your SQL syntax; check the manual that corresponds to your MariaDB server version for the right syntax to use near 'from `tabUser Group`' at line 2\\\")\",\r\n\t\"_debug_messages\": \"[\\\"method:frappe.desk.reportview.get_count\\\", \\\"<<<< query\\\\nSELECT `is_virtual` FROM `tabDocType` WHERE `name`='User Group' ORDER BY `modified` DESC LIMIT 1\\\\n>>>>\\\", \\\"<<<< query\\\\nselect `share_name`\\\\n\\\\t\\\\t\\\\tfrom `tabDocShare`\\\\n\\\\t\\\\t\\\\twhere `tabDocShare`.`read` = 1.0 and `tabDocShare`.`share_doctype` = 'User Group' and (`tabDocShare`.`user` = 'Administrator' or `tabDocShare`.`everyone` = 1.0)\\\\n\\\\t\\\\t\\\\t\\\\n\\\\t\\\\t\\\\t order by `tabDocShare`.`modified` DESC\\\\n>>>>\\\"]\"\r\n}\r\n```\n", "before_files": [{"content": "# Copyright (c) 2015, Frappe Technologies Pvt. Ltd. and Contributors\n# License: MIT. See LICENSE\n\n# model __init__.py\nimport frappe\n\ndata_fieldtypes = (\n\t\"Currency\",\n\t\"Int\",\n\t\"Long Int\",\n\t\"Float\",\n\t\"Percent\",\n\t\"Check\",\n\t\"Small Text\",\n\t\"Long Text\",\n\t\"Code\",\n\t\"Text Editor\",\n\t\"Markdown Editor\",\n\t\"HTML Editor\",\n\t\"Date\",\n\t\"Datetime\",\n\t\"Time\",\n\t\"Text\",\n\t\"Data\",\n\t\"Link\",\n\t\"Dynamic Link\",\n\t\"Password\",\n\t\"Select\",\n\t\"Rating\",\n\t\"Read Only\",\n\t\"Attach\",\n\t\"Attach Image\",\n\t\"Signature\",\n\t\"Color\",\n\t\"Barcode\",\n\t\"Geolocation\",\n\t\"Duration\",\n\t\"Icon\",\n\t\"Phone\",\n\t\"Autocomplete\",\n\t\"JSON\",\n)\n\nfloat_like_fields = {\"Float\", \"Currency\", \"Percent\"}\ndatetime_fields = {\"Datetime\", \"Date\", \"Time\"}\n\nattachment_fieldtypes = (\n\t\"Attach\",\n\t\"Attach Image\",\n)\n\nno_value_fields = (\n\t\"Section Break\",\n\t\"Column Break\",\n\t\"Tab Break\",\n\t\"HTML\",\n\t\"Table\",\n\t\"Table MultiSelect\",\n\t\"Button\",\n\t\"Image\",\n\t\"Fold\",\n\t\"Heading\",\n)\n\ndisplay_fieldtypes = (\n\t\"Section Break\",\n\t\"Column Break\",\n\t\"Tab Break\",\n\t\"HTML\",\n\t\"Button\",\n\t\"Image\",\n\t\"Fold\",\n\t\"Heading\",\n)\n\nnumeric_fieldtypes = (\"Currency\", \"Int\", \"Long Int\", \"Float\", \"Percent\", \"Check\")\n\ndata_field_options = (\"Email\", \"Name\", \"Phone\", \"URL\", \"Barcode\")\n\ndefault_fields = (\n\t\"doctype\",\n\t\"name\",\n\t\"owner\",\n\t\"creation\",\n\t\"modified\",\n\t\"modified_by\",\n\t\"docstatus\",\n\t\"idx\",\n)\n\nchild_table_fields = (\"parent\", \"parentfield\", \"parenttype\")\n\noptional_fields = (\"_user_tags\", \"_comments\", \"_assign\", \"_liked_by\", \"_seen\")\n\ntable_fields = (\"Table\", \"Table MultiSelect\")\n\ncore_doctypes_list = (\n\t\"DefaultValue\",\n\t\"DocType\",\n\t\"DocField\",\n\t\"DocPerm\",\n\t\"DocType Action\",\n\t\"DocType Link\",\n\t\"User\",\n\t\"Role\",\n\t\"Has Role\",\n\t\"Page\",\n\t\"Module Def\",\n\t\"Print Format\",\n\t\"Report\",\n\t\"Customize Form\",\n\t\"Customize Form Field\",\n\t\"Property Setter\",\n\t\"Custom Field\",\n\t\"Client Script\",\n)\n\n# NOTE: this is being used for dynamic autoincrement in new sites,\n# removing any of these will require patches.\nlog_types = (\n\t\"Version\",\n\t\"Error Log\",\n\t\"Scheduled Job Log\",\n\t\"Event Sync Log\",\n\t\"Event Update Log\",\n\t\"Access Log\",\n\t\"View Log\",\n\t\"Activity Log\",\n\t\"Energy Point Log\",\n\t\"Notification Log\",\n\t\"Email Queue\",\n\t\"DocShare\",\n\t\"Document Follow\",\n\t\"Console Log\",\n)\n\n\ndef delete_fields(args_dict, delete=0):\n\t\"\"\"\n\tDelete a field.\n\t* Deletes record from `tabDocField`\n\t* If not single doctype: Drops column from table\n\t* If single, deletes record from `tabSingles`\n\targs_dict = { dt: [field names] }\n\t\"\"\"\n\timport frappe.utils\n\n\tfor dt in args_dict:\n\t\tfields = args_dict[dt]\n\t\tif not fields:\n\t\t\tcontinue\n\n\t\tfrappe.db.delete(\n\t\t\t\"DocField\",\n\t\t\t{\n\t\t\t\t\"parent\": dt,\n\t\t\t\t\"fieldname\": (\"in\", fields),\n\t\t\t},\n\t\t)\n\n\t\t# Delete the data/column only if delete is specified\n\t\tif not delete:\n\t\t\tcontinue\n\n\t\tif frappe.db.get_value(\"DocType\", dt, \"issingle\"):\n\t\t\tfrappe.db.delete(\n\t\t\t\t\"Singles\",\n\t\t\t\t{\n\t\t\t\t\t\"doctype\": dt,\n\t\t\t\t\t\"field\": (\"in\", fields),\n\t\t\t\t},\n\t\t\t)\n\t\telse:\n\t\t\texisting_fields = frappe.db.describe(dt)\n\t\t\texisting_fields = existing_fields and [e[0] for e in existing_fields] or []\n\t\t\tfields_need_to_delete = set(fields) & set(existing_fields)\n\t\t\tif not fields_need_to_delete:\n\t\t\t\tcontinue\n\n\t\t\tif frappe.db.db_type == \"mariadb\":\n\t\t\t\t# mariadb implicitly commits before DDL, make it explicit\n\t\t\t\tfrappe.db.commit()\n\n\t\t\tquery = \"ALTER TABLE `tab%s` \" % dt + \", \".join(\n\t\t\t\t\"DROP COLUMN `%s`\" % f for f in fields_need_to_delete\n\t\t\t)\n\t\t\tfrappe.db.sql(query)\n\n\t\tif frappe.db.db_type == \"postgres\":\n\t\t\t# commit the results to db\n\t\t\tfrappe.db.commit()\n\n\ndef get_permitted_fields(\n\tdoctype: str, parenttype: str | None = None, user: str | None = None\n) -> list[str]:\n\tmeta = frappe.get_meta(doctype)\n\tvalid_columns = meta.get_valid_columns()\n\n\tif doctype in core_doctypes_list:\n\t\treturn valid_columns\n\n\tif permitted_fields := meta.get_permitted_fieldnames(parenttype=parenttype, user=user):\n\t\tmeta_fields = meta.default_fields.copy()\n\t\toptional_meta_fields = [x for x in optional_fields if x in valid_columns]\n\n\t\tif meta.istable:\n\t\t\tmeta_fields.extend(child_table_fields)\n\n\t\treturn meta_fields + permitted_fields + optional_meta_fields\n\n\treturn []\n", "path": "frappe/model/__init__.py"}]}
| 3,603 | 156 |
gh_patches_debug_30061
|
rasdani/github-patches
|
git_diff
|
Miserlou__Zappa-1993
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Set_Cookie option sets duplicate cookies on AWS Lambda
## Context
I have an API running Python3.7 and Zappa (in a virtualenv).
I am setting 6 cookies by using the option "set_cookie" in flask. It looks something like this:
```
resp = make_response(jsonify({'success':'true', 'message': 'Successfully authenticated!'}), 200)
resp.set_cookie("1", value="1", secure=True, samesite='Lax', domain=".example.com",max_age=3600)
resp.set_cookie("2", value="2", secure=True, samesite='Lax', domain=".example.com",max_age=3600)
resp.set_cookie("3", value="3", secure=True, samesite='Lax', domain=".example.com",max_age=3600)
resp.set_cookie("4", value="4", secure=True, samesite='Lax', domain=".example.com",max_age=3600)
resp.set_cookie("5", value="5", secure=True, samesite='Lax', domain=".example.com",max_age=3600)
resp.set_cookie("6", value="6", secure=True, samesite='Lax', domain=".example.com",max_age=3600)
return resp
```
On localhost testing Flask, this works as expected.
If I deploy the same code to AWS using Zappa, the response header will show 36 "set-cookie" headers. So the formula here is n^2. So if I add 4 cookies using the above method, it will show 16 in the request header.
The browser takes care of duplicate cookies, but the response from the API is still huge because of this issue.
Same thing happens if I use:
`resp.headers.add("set-cookie""1"="1; Domain=.example.com; Max-Age=3600; Secure; Path=/; SameSite=Lax")`
## Expected Behavior
I believe Zappa or something at AWS is at fault here. Expected behaviour is to send 6 "set-cookie" headers and not 36.
## Actual Behavior
Sets n^2 cookies as response.
## Steps to Reproduce
Deploy a Flask route using Zappa which sets the cookies. Use the code above.
## Your Environment
* Zappa version used: 0.48.2
* Operating System and Python version: Ubuntu 18.04, Python3.7
* The output of `pip freeze`: https://pastebin.com/d4QTaTuG
* Your `zappa_settings.py`: https://pastebin.com/d1GK8sbe
</issue>
<code>
[start of zappa/middleware.py]
1 from werkzeug.wsgi import ClosingIterator
2
3
4 def all_casings(input_string):
5 """
6 Permute all casings of a given string.
7
8 A pretty algorithm, via @Amber
9 http://stackoverflow.com/questions/6792803/finding-all-possible-case-permutations-in-python
10 """
11 if not input_string:
12 yield ""
13 else:
14 first = input_string[:1]
15 if first.lower() == first.upper():
16 for sub_casing in all_casings(input_string[1:]):
17 yield first + sub_casing
18 else:
19 for sub_casing in all_casings(input_string[1:]):
20 yield first.lower() + sub_casing
21 yield first.upper() + sub_casing
22
23
24 class ZappaWSGIMiddleware(object):
25 """
26 Middleware functions necessary for a Zappa deployment.
27
28 Most hacks have now been remove except for Set-Cookie permutation.
29 """
30 def __init__(self, application):
31 self.application = application
32
33 def __call__(self, environ, start_response):
34 """
35 We must case-mangle the Set-Cookie header name or AWS will use only a
36 single one of these headers.
37 """
38
39 def encode_response(status, headers, exc_info=None):
40 """
41 Create an APIGW-acceptable version of our cookies.
42
43 We have to use a bizarre hack that turns multiple Set-Cookie headers into
44 their case-permutated format, ex:
45
46 Set-cookie:
47 sEt-cookie:
48 seT-cookie:
49
50 To get around an API Gateway limitation.
51
52 This is weird, but better than our previous hack of creating a Base58-encoded
53 supercookie.
54 """
55
56 # All the non-cookie headers should be sent unharmed.
57
58 # The main app can send 'set-cookie' headers in any casing
59 # Related: https://github.com/Miserlou/Zappa/issues/990
60 new_headers = [header for header in headers
61 if ((type(header[0]) != str) or (header[0].lower() != 'set-cookie'))]
62 cookie_headers = [header for header in headers
63 if ((type(header[0]) == str) and (header[0].lower() == "set-cookie"))]
64 for header, new_name in zip(cookie_headers,
65 all_casings("Set-Cookie")):
66 new_headers.append((new_name, header[1]))
67 return start_response(status, new_headers, exc_info)
68
69 # Call the application with our modifier
70 response = self.application(environ, encode_response)
71
72 # Return the response as a WSGI-safe iterator
73 return ClosingIterator(response)
74
[end of zappa/middleware.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/zappa/middleware.py b/zappa/middleware.py
--- a/zappa/middleware.py
+++ b/zappa/middleware.py
@@ -38,32 +38,17 @@
def encode_response(status, headers, exc_info=None):
"""
- Create an APIGW-acceptable version of our cookies.
-
- We have to use a bizarre hack that turns multiple Set-Cookie headers into
- their case-permutated format, ex:
-
- Set-cookie:
- sEt-cookie:
- seT-cookie:
-
- To get around an API Gateway limitation.
-
- This is weird, but better than our previous hack of creating a Base58-encoded
- supercookie.
+ This makes the 'set-cookie' headers name lowercase,
+ all the non-cookie headers should be sent unharmed.
+ Related: https://github.com/Miserlou/Zappa/issues/1965
"""
- # All the non-cookie headers should be sent unharmed.
-
- # The main app can send 'set-cookie' headers in any casing
- # Related: https://github.com/Miserlou/Zappa/issues/990
new_headers = [header for header in headers
if ((type(header[0]) != str) or (header[0].lower() != 'set-cookie'))]
- cookie_headers = [header for header in headers
+ cookie_headers = [(header[0].lower(), header[1]) for header in headers
if ((type(header[0]) == str) and (header[0].lower() == "set-cookie"))]
- for header, new_name in zip(cookie_headers,
- all_casings("Set-Cookie")):
- new_headers.append((new_name, header[1]))
+ new_headers = new_headers + cookie_headers
+
return start_response(status, new_headers, exc_info)
# Call the application with our modifier
|
{"golden_diff": "diff --git a/zappa/middleware.py b/zappa/middleware.py\n--- a/zappa/middleware.py\n+++ b/zappa/middleware.py\n@@ -38,32 +38,17 @@\n \n def encode_response(status, headers, exc_info=None):\n \"\"\"\n- Create an APIGW-acceptable version of our cookies.\n-\n- We have to use a bizarre hack that turns multiple Set-Cookie headers into\n- their case-permutated format, ex:\n-\n- Set-cookie:\n- sEt-cookie:\n- seT-cookie:\n-\n- To get around an API Gateway limitation.\n-\n- This is weird, but better than our previous hack of creating a Base58-encoded\n- supercookie.\n+ This makes the 'set-cookie' headers name lowercase,\n+ all the non-cookie headers should be sent unharmed.\n+ Related: https://github.com/Miserlou/Zappa/issues/1965\n \"\"\"\n \n- # All the non-cookie headers should be sent unharmed.\n- \n- # The main app can send 'set-cookie' headers in any casing\n- # Related: https://github.com/Miserlou/Zappa/issues/990\n new_headers = [header for header in headers\n if ((type(header[0]) != str) or (header[0].lower() != 'set-cookie'))]\n- cookie_headers = [header for header in headers \n+ cookie_headers = [(header[0].lower(), header[1]) for header in headers\n if ((type(header[0]) == str) and (header[0].lower() == \"set-cookie\"))]\n- for header, new_name in zip(cookie_headers,\n- all_casings(\"Set-Cookie\")):\n- new_headers.append((new_name, header[1]))\n+ new_headers = new_headers + cookie_headers\n+\n return start_response(status, new_headers, exc_info)\n \n # Call the application with our modifier\n", "issue": "Set_Cookie option sets duplicate cookies on AWS Lambda\n## Context\r\nI have an API running Python3.7 and Zappa (in a virtualenv).\r\nI am setting 6 cookies by using the option \"set_cookie\" in flask. It looks something like this:\r\n```\r\nresp = make_response(jsonify({'success':'true', 'message': 'Successfully authenticated!'}), 200)\r\nresp.set_cookie(\"1\", value=\"1\", secure=True, samesite='Lax', domain=\".example.com\",max_age=3600)\r\nresp.set_cookie(\"2\", value=\"2\", secure=True, samesite='Lax', domain=\".example.com\",max_age=3600)\r\nresp.set_cookie(\"3\", value=\"3\", secure=True, samesite='Lax', domain=\".example.com\",max_age=3600)\r\nresp.set_cookie(\"4\", value=\"4\", secure=True, samesite='Lax', domain=\".example.com\",max_age=3600)\r\nresp.set_cookie(\"5\", value=\"5\", secure=True, samesite='Lax', domain=\".example.com\",max_age=3600)\r\nresp.set_cookie(\"6\", value=\"6\", secure=True, samesite='Lax', domain=\".example.com\",max_age=3600)\r\nreturn resp\r\n```\r\n\r\nOn localhost testing Flask, this works as expected.\r\n\r\nIf I deploy the same code to AWS using Zappa, the response header will show 36 \"set-cookie\" headers. So the formula here is n^2. So if I add 4 cookies using the above method, it will show 16 in the request header.\r\n\r\nThe browser takes care of duplicate cookies, but the response from the API is still huge because of this issue.\r\n\r\nSame thing happens if I use: \r\n`resp.headers.add(\"set-cookie\"\"1\"=\"1; Domain=.example.com; Max-Age=3600; Secure; Path=/; SameSite=Lax\")`\r\n\r\n## Expected Behavior\r\nI believe Zappa or something at AWS is at fault here. Expected behaviour is to send 6 \"set-cookie\" headers and not 36.\r\n\r\n## Actual Behavior\r\nSets n^2 cookies as response.\r\n\r\n## Steps to Reproduce\r\nDeploy a Flask route using Zappa which sets the cookies. Use the code above.\r\n\r\n## Your Environment\r\n* Zappa version used: 0.48.2\r\n* Operating System and Python version: Ubuntu 18.04, Python3.7\r\n* The output of `pip freeze`: https://pastebin.com/d4QTaTuG\r\n* Your `zappa_settings.py`: https://pastebin.com/d1GK8sbe\n", "before_files": [{"content": "from werkzeug.wsgi import ClosingIterator\n\n\ndef all_casings(input_string):\n \"\"\"\n Permute all casings of a given string.\n\n A pretty algorithm, via @Amber\n http://stackoverflow.com/questions/6792803/finding-all-possible-case-permutations-in-python\n \"\"\"\n if not input_string:\n yield \"\"\n else:\n first = input_string[:1]\n if first.lower() == first.upper():\n for sub_casing in all_casings(input_string[1:]):\n yield first + sub_casing\n else:\n for sub_casing in all_casings(input_string[1:]):\n yield first.lower() + sub_casing\n yield first.upper() + sub_casing\n\n\nclass ZappaWSGIMiddleware(object):\n \"\"\"\n Middleware functions necessary for a Zappa deployment.\n\n Most hacks have now been remove except for Set-Cookie permutation.\n \"\"\"\n def __init__(self, application):\n self.application = application\n\n def __call__(self, environ, start_response):\n \"\"\"\n We must case-mangle the Set-Cookie header name or AWS will use only a\n single one of these headers.\n \"\"\"\n\n def encode_response(status, headers, exc_info=None):\n \"\"\"\n Create an APIGW-acceptable version of our cookies.\n\n We have to use a bizarre hack that turns multiple Set-Cookie headers into\n their case-permutated format, ex:\n\n Set-cookie:\n sEt-cookie:\n seT-cookie:\n\n To get around an API Gateway limitation.\n\n This is weird, but better than our previous hack of creating a Base58-encoded\n supercookie.\n \"\"\"\n\n # All the non-cookie headers should be sent unharmed.\n \n # The main app can send 'set-cookie' headers in any casing\n # Related: https://github.com/Miserlou/Zappa/issues/990\n new_headers = [header for header in headers\n if ((type(header[0]) != str) or (header[0].lower() != 'set-cookie'))]\n cookie_headers = [header for header in headers \n if ((type(header[0]) == str) and (header[0].lower() == \"set-cookie\"))]\n for header, new_name in zip(cookie_headers,\n all_casings(\"Set-Cookie\")):\n new_headers.append((new_name, header[1]))\n return start_response(status, new_headers, exc_info)\n\n # Call the application with our modifier\n response = self.application(environ, encode_response)\n\n # Return the response as a WSGI-safe iterator\n return ClosingIterator(response)\n", "path": "zappa/middleware.py"}]}
| 1,822 | 430 |
gh_patches_debug_32826
|
rasdani/github-patches
|
git_diff
|
dotkom__onlineweb4-1498
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Don't send additional 500 email if no useful information in it.
After the implementation if #1485 we get an additional email for _all_ 500 errors, even if there is no supplied information. Let's not send an email if there's no useful information in it.
</issue>
<code>
[start of onlineweb4/views.py]
1 # -*- coding: utf-8 -*-
2 import logging
3 from smtplib import SMTPException
4
5 from django.conf import settings
6 from django.contrib import messages
7 from django.core.mail import send_mail
8 from django.shortcuts import redirect, render
9 from onlineweb4.forms import ErrorForm
10
11
12 def server_error(request):
13 log = logging.getLogger(__name__)
14
15 if request.method == 'POST':
16 form = ErrorForm(request.POST)
17 message = request.POST.get('reason', 'Ingen forklaring oppgitt.')
18 try:
19 log.error('%s triggered a 500 server error and provided the following description: %s' % (
20 request.user,
21 message
22 ))
23 send_mail('500error user-report', message,
24 settings.DEFAULT_FROM_EMAIL, [settings.EMAIL_DOTKOM])
25 log.debug('Finished sending error email to %s' % settings.EMAIL_DOTKOM)
26
27 messages.success(request, 'Feilmeldingen din ble sendt til %s' % settings.EMAIL_DOTKOM)
28
29 return redirect('home')
30 except SMTPException:
31 messages.error(request, 'Det oppstod en uventet feil under sending av feilmeldingen')
32 return redirect('home')
33
34 return render(request, '500.html', {'error_form': ErrorForm})
35
[end of onlineweb4/views.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/onlineweb4/views.py b/onlineweb4/views.py
--- a/onlineweb4/views.py
+++ b/onlineweb4/views.py
@@ -1,21 +1,22 @@
# -*- coding: utf-8 -*-
import logging
-from smtplib import SMTPException
from django.conf import settings
from django.contrib import messages
from django.core.mail import send_mail
from django.shortcuts import redirect, render
from onlineweb4.forms import ErrorForm
+from smtplib import SMTPException
def server_error(request):
log = logging.getLogger(__name__)
if request.method == 'POST':
- form = ErrorForm(request.POST)
- message = request.POST.get('reason', 'Ingen forklaring oppgitt.')
- try:
+ message = request.POST.get('reason')
+ if not message:
+ return redirect('home')
+ try:
log.error('%s triggered a 500 server error and provided the following description: %s' % (
request.user,
message
@@ -23,12 +24,9 @@
send_mail('500error user-report', message,
settings.DEFAULT_FROM_EMAIL, [settings.EMAIL_DOTKOM])
log.debug('Finished sending error email to %s' % settings.EMAIL_DOTKOM)
-
messages.success(request, 'Feilmeldingen din ble sendt til %s' % settings.EMAIL_DOTKOM)
-
return redirect('home')
except SMTPException:
messages.error(request, 'Det oppstod en uventet feil under sending av feilmeldingen')
return redirect('home')
-
return render(request, '500.html', {'error_form': ErrorForm})
|
{"golden_diff": "diff --git a/onlineweb4/views.py b/onlineweb4/views.py\n--- a/onlineweb4/views.py\n+++ b/onlineweb4/views.py\n@@ -1,21 +1,22 @@\n # -*- coding: utf-8 -*-\n import logging\n-from smtplib import SMTPException\n \n from django.conf import settings\n from django.contrib import messages\n from django.core.mail import send_mail\n from django.shortcuts import redirect, render\n from onlineweb4.forms import ErrorForm\n+from smtplib import SMTPException\n \n \n def server_error(request):\n log = logging.getLogger(__name__)\n \n if request.method == 'POST':\n- form = ErrorForm(request.POST)\n- message = request.POST.get('reason', 'Ingen forklaring oppgitt.')\n- try: \n+ message = request.POST.get('reason')\n+ if not message:\n+ return redirect('home')\n+ try:\n log.error('%s triggered a 500 server error and provided the following description: %s' % (\n request.user,\n message\n@@ -23,12 +24,9 @@\n send_mail('500error user-report', message,\n settings.DEFAULT_FROM_EMAIL, [settings.EMAIL_DOTKOM])\n log.debug('Finished sending error email to %s' % settings.EMAIL_DOTKOM)\n-\n messages.success(request, 'Feilmeldingen din ble sendt til %s' % settings.EMAIL_DOTKOM)\n-\n return redirect('home')\n except SMTPException:\n messages.error(request, 'Det oppstod en uventet feil under sending av feilmeldingen')\n return redirect('home')\n-\n return render(request, '500.html', {'error_form': ErrorForm})\n", "issue": "Don't send additional 500 email if no useful information in it.\nAfter the implementation if #1485 we get an additional email for _all_ 500 errors, even if there is no supplied information. Let's not send an email if there's no useful information in it.\n\n", "before_files": [{"content": "# -*- coding: utf-8 -*-\nimport logging\nfrom smtplib import SMTPException\n\nfrom django.conf import settings\nfrom django.contrib import messages\nfrom django.core.mail import send_mail\nfrom django.shortcuts import redirect, render\nfrom onlineweb4.forms import ErrorForm\n\n\ndef server_error(request):\n log = logging.getLogger(__name__)\n\n if request.method == 'POST':\n form = ErrorForm(request.POST)\n message = request.POST.get('reason', 'Ingen forklaring oppgitt.')\n try: \n log.error('%s triggered a 500 server error and provided the following description: %s' % (\n request.user,\n message\n ))\n send_mail('500error user-report', message,\n settings.DEFAULT_FROM_EMAIL, [settings.EMAIL_DOTKOM])\n log.debug('Finished sending error email to %s' % settings.EMAIL_DOTKOM)\n\n messages.success(request, 'Feilmeldingen din ble sendt til %s' % settings.EMAIL_DOTKOM)\n\n return redirect('home')\n except SMTPException:\n messages.error(request, 'Det oppstod en uventet feil under sending av feilmeldingen')\n return redirect('home')\n\n return render(request, '500.html', {'error_form': ErrorForm})\n", "path": "onlineweb4/views.py"}]}
| 941 | 377 |
gh_patches_debug_8684
|
rasdani/github-patches
|
git_diff
|
rpm-software-management__dnf-1907
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Bug on almalinux 9
Hi,
I have these messages on almalinux 9/ Rocky Linux 9
[mickael@srvalmalinux ~]$ sudo /usr/bin/dnf-automatic /etc/dnf/automatic.conf --timer
Last metadata expiration check: 2:45:52 ago on Fri Mar 10 14:38:38 2023.
Running transaction check
Transaction check succeeded.
Running transaction test
Transaction test succeeded.
Running transaction
Traceback (most recent call last):
File "/usr/bin/dnf-automatic", line 36, in <module>
sys.exit(dnf.automatic.main.main(sys.argv[1:]))
File "/usr/lib/python3.9/site-packages/dnf/automatic/main.py", line 342, in main
emitters.commit()
File "/usr/lib/python3.9/site-packages/dnf/util.py", line 526, in fn
return list(map(call_what, self))
File "/usr/lib/python3.9/site-packages/dnf/util.py", line 525, in call_what
return method(*args, **kwargs)
File "/usr/lib/python3.9/site-packages/dnf/automatic/emitter.py", line 106, in commit
smtp = smtplib.SMTP(self._conf.email_host, timeout=300)
File "/usr/lib64/python3.9/smtplib.py", line 255, in __init__
(code, msg) = self.connect(host, port)
File "/usr/lib64/python3.9/smtplib.py", line 341, in connect
self.sock = self._get_socket(host, port, self.timeout)
File "/usr/lib64/python3.9/smtplib.py", line 312, in _get_socket
return socket.create_connection((host, port), timeout,
File "/usr/lib64/python3.9/socket.py", line 844, in create_connection
raise err
File "/usr/lib64/python3.9/socket.py", line 832, in create_connection
sock.connect(sa)
ConnectionRefusedError: [Errno 111] Connection refused
</issue>
<code>
[start of dnf/automatic/emitter.py]
1 # emitter.py
2 # Emitters for dnf-automatic.
3 #
4 # Copyright (C) 2014-2016 Red Hat, Inc.
5 #
6 # This copyrighted material is made available to anyone wishing to use,
7 # modify, copy, or redistribute it subject to the terms and conditions of
8 # the GNU General Public License v.2, or (at your option) any later version.
9 # This program is distributed in the hope that it will be useful, but WITHOUT
10 # ANY WARRANTY expressed or implied, including the implied warranties of
11 # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General
12 # Public License for more details. You should have received a copy of the
13 # GNU General Public License along with this program; if not, write to the
14 # Free Software Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA
15 # 02110-1301, USA. Any Red Hat trademarks that are incorporated in the
16 # source code or documentation are not subject to the GNU General Public
17 # License and may only be used or replicated with the express permission of
18 # Red Hat, Inc.
19 #
20
21 from __future__ import absolute_import
22 from __future__ import print_function
23 from __future__ import unicode_literals
24 from dnf.i18n import _
25 import logging
26 import dnf.pycomp
27 import smtplib
28 import email.utils
29 import subprocess
30 import time
31
32 APPLIED = _("The following updates have been applied on '%s':")
33 APPLIED_TIMESTAMP = _("Updates completed at %s")
34 AVAILABLE = _("The following updates are available on '%s':")
35 DOWNLOADED = _("The following updates were downloaded on '%s':")
36
37 logger = logging.getLogger('dnf')
38
39
40 class Emitter(object):
41 def __init__(self, system_name):
42 self._applied = False
43 self._available_msg = None
44 self._downloaded = False
45 self._system_name = system_name
46 self._trans_msg = None
47
48 def _prepare_msg(self):
49 msg = []
50 if self._applied:
51 msg.append(APPLIED % self._system_name)
52 msg.append(self._available_msg)
53 msg.append(APPLIED_TIMESTAMP % time.strftime("%c"))
54 elif self._downloaded:
55 msg.append(DOWNLOADED % self._system_name)
56 msg.append(self._available_msg)
57 elif self._available_msg:
58 msg.append(AVAILABLE % self._system_name)
59 msg.append(self._available_msg)
60 else:
61 return None
62 return '\n'.join(msg)
63
64 def notify_applied(self):
65 assert self._available_msg
66 self._applied = True
67
68 def notify_available(self, msg):
69 self._available_msg = msg
70
71 def notify_downloaded(self):
72 assert self._available_msg
73 self._downloaded = True
74
75
76 class EmailEmitter(Emitter):
77 def __init__(self, system_name, conf):
78 super(EmailEmitter, self).__init__(system_name)
79 self._conf = conf
80
81 def _prepare_msg(self):
82 if self._applied:
83 subj = _("Updates applied on '%s'.") % self._system_name
84 elif self._downloaded:
85 subj = _("Updates downloaded on '%s'.") % self._system_name
86 elif self._available_msg:
87 subj = _("Updates available on '%s'.") % self._system_name
88 else:
89 return None, None
90 return subj, super(EmailEmitter, self)._prepare_msg()
91
92 def commit(self):
93 subj, body = self._prepare_msg()
94 message = dnf.pycomp.email_mime(body)
95 message.set_charset('utf-8')
96 email_from = self._conf.email_from
97 email_to = self._conf.email_to
98 message['Date'] = email.utils.formatdate()
99 message['From'] = email_from
100 message['Subject'] = subj
101 message['To'] = ','.join(email_to)
102 message['Message-ID'] = email.utils.make_msgid()
103
104 # Send the email
105 try:
106 smtp = smtplib.SMTP(self._conf.email_host, timeout=300)
107 smtp.sendmail(email_from, email_to, message.as_string())
108 smtp.close()
109 except smtplib.SMTPException as exc:
110 msg = _("Failed to send an email via '%s': %s") % (
111 self._conf.email_host, exc)
112 logger.error(msg)
113
114
115 class CommandEmitterMixIn(object):
116 """
117 Executes a desired command, and pushes data into its stdin.
118 Both data and command can be formatted according to user preference.
119 For this reason, this class expects a {str:str} dictionary as _prepare_msg
120 return value.
121 Meant for mixing with Emitter classes, as it does not define any names used
122 for formatting on its own.
123 """
124 def commit(self):
125 command_fmt = self._conf.command_format
126 stdin_fmt = self._conf.stdin_format
127 msg = self._prepare_msg()
128 # all strings passed to shell should be quoted to avoid accidental code
129 # execution
130 quoted_msg = dict((key, dnf.pycomp.shlex_quote(val))
131 for key, val in msg.items())
132 command = command_fmt.format(**quoted_msg)
133 stdin_feed = stdin_fmt.format(**msg).encode('utf-8')
134
135 # Execute the command
136 subp = subprocess.Popen(command, shell=True, stdin=subprocess.PIPE)
137 subp.communicate(stdin_feed)
138 subp.stdin.close()
139 if subp.wait() != 0:
140 msg = _("Failed to execute command '%s': returned %d") \
141 % (command, subp.returncode)
142 logger.error(msg)
143
144
145 class CommandEmitter(CommandEmitterMixIn, Emitter):
146 def __init__(self, system_name, conf):
147 super(CommandEmitter, self).__init__(system_name)
148 self._conf = conf
149
150 def _prepare_msg(self):
151 return {'body': super(CommandEmitter, self)._prepare_msg()}
152
153
154 class CommandEmailEmitter(CommandEmitterMixIn, EmailEmitter):
155 def _prepare_msg(self):
156 subject, body = super(CommandEmailEmitter, self)._prepare_msg()
157 return {'subject': subject,
158 'body': body,
159 'email_from': self._conf.email_from,
160 'email_to': ' '.join(self._conf.email_to)}
161
162
163 class StdIoEmitter(Emitter):
164 def commit(self):
165 msg = self._prepare_msg()
166 print(msg)
167
168
169 class MotdEmitter(Emitter):
170 def commit(self):
171 msg = self._prepare_msg()
172 with open('/etc/motd', 'w') as fobj:
173 fobj.write(msg)
174
175
[end of dnf/automatic/emitter.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/dnf/automatic/emitter.py b/dnf/automatic/emitter.py
--- a/dnf/automatic/emitter.py
+++ b/dnf/automatic/emitter.py
@@ -106,7 +106,7 @@
smtp = smtplib.SMTP(self._conf.email_host, timeout=300)
smtp.sendmail(email_from, email_to, message.as_string())
smtp.close()
- except smtplib.SMTPException as exc:
+ except OSError as exc:
msg = _("Failed to send an email via '%s': %s") % (
self._conf.email_host, exc)
logger.error(msg)
|
{"golden_diff": "diff --git a/dnf/automatic/emitter.py b/dnf/automatic/emitter.py\n--- a/dnf/automatic/emitter.py\n+++ b/dnf/automatic/emitter.py\n@@ -106,7 +106,7 @@\n smtp = smtplib.SMTP(self._conf.email_host, timeout=300)\n smtp.sendmail(email_from, email_to, message.as_string())\n smtp.close()\n- except smtplib.SMTPException as exc:\n+ except OSError as exc:\n msg = _(\"Failed to send an email via '%s': %s\") % (\n self._conf.email_host, exc)\n logger.error(msg)\n", "issue": "Bug on almalinux 9\nHi,\r\n\r\nI have these messages on almalinux 9/ Rocky Linux 9\r\n\r\n\r\n[mickael@srvalmalinux ~]$ sudo /usr/bin/dnf-automatic /etc/dnf/automatic.conf --timer\r\nLast metadata expiration check: 2:45:52 ago on Fri Mar 10 14:38:38 2023.\r\nRunning transaction check\r\nTransaction check succeeded.\r\nRunning transaction test\r\nTransaction test succeeded.\r\nRunning transaction\r\nTraceback (most recent call last):\r\n File \"/usr/bin/dnf-automatic\", line 36, in <module>\r\n sys.exit(dnf.automatic.main.main(sys.argv[1:]))\r\n File \"/usr/lib/python3.9/site-packages/dnf/automatic/main.py\", line 342, in main\r\n emitters.commit()\r\n File \"/usr/lib/python3.9/site-packages/dnf/util.py\", line 526, in fn\r\n return list(map(call_what, self))\r\n File \"/usr/lib/python3.9/site-packages/dnf/util.py\", line 525, in call_what\r\n return method(*args, **kwargs)\r\n File \"/usr/lib/python3.9/site-packages/dnf/automatic/emitter.py\", line 106, in commit\r\n smtp = smtplib.SMTP(self._conf.email_host, timeout=300)\r\n File \"/usr/lib64/python3.9/smtplib.py\", line 255, in __init__\r\n (code, msg) = self.connect(host, port)\r\n File \"/usr/lib64/python3.9/smtplib.py\", line 341, in connect\r\n self.sock = self._get_socket(host, port, self.timeout)\r\n File \"/usr/lib64/python3.9/smtplib.py\", line 312, in _get_socket\r\n return socket.create_connection((host, port), timeout,\r\n File \"/usr/lib64/python3.9/socket.py\", line 844, in create_connection\r\n raise err\r\n File \"/usr/lib64/python3.9/socket.py\", line 832, in create_connection\r\n sock.connect(sa)\r\nConnectionRefusedError: [Errno 111] Connection refused\r\n\n", "before_files": [{"content": "# emitter.py\n# Emitters for dnf-automatic.\n#\n# Copyright (C) 2014-2016 Red Hat, Inc.\n#\n# This copyrighted material is made available to anyone wishing to use,\n# modify, copy, or redistribute it subject to the terms and conditions of\n# the GNU General Public License v.2, or (at your option) any later version.\n# This program is distributed in the hope that it will be useful, but WITHOUT\n# ANY WARRANTY expressed or implied, including the implied warranties of\n# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General\n# Public License for more details. You should have received a copy of the\n# GNU General Public License along with this program; if not, write to the\n# Free Software Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA\n# 02110-1301, USA. Any Red Hat trademarks that are incorporated in the\n# source code or documentation are not subject to the GNU General Public\n# License and may only be used or replicated with the express permission of\n# Red Hat, Inc.\n#\n\nfrom __future__ import absolute_import\nfrom __future__ import print_function\nfrom __future__ import unicode_literals\nfrom dnf.i18n import _\nimport logging\nimport dnf.pycomp\nimport smtplib\nimport email.utils\nimport subprocess\nimport time\n\nAPPLIED = _(\"The following updates have been applied on '%s':\")\nAPPLIED_TIMESTAMP = _(\"Updates completed at %s\")\nAVAILABLE = _(\"The following updates are available on '%s':\")\nDOWNLOADED = _(\"The following updates were downloaded on '%s':\")\n\nlogger = logging.getLogger('dnf')\n\n\nclass Emitter(object):\n def __init__(self, system_name):\n self._applied = False\n self._available_msg = None\n self._downloaded = False\n self._system_name = system_name\n self._trans_msg = None\n\n def _prepare_msg(self):\n msg = []\n if self._applied:\n msg.append(APPLIED % self._system_name)\n msg.append(self._available_msg)\n msg.append(APPLIED_TIMESTAMP % time.strftime(\"%c\"))\n elif self._downloaded:\n msg.append(DOWNLOADED % self._system_name)\n msg.append(self._available_msg)\n elif self._available_msg:\n msg.append(AVAILABLE % self._system_name)\n msg.append(self._available_msg)\n else:\n return None\n return '\\n'.join(msg)\n\n def notify_applied(self):\n assert self._available_msg\n self._applied = True\n\n def notify_available(self, msg):\n self._available_msg = msg\n\n def notify_downloaded(self):\n assert self._available_msg\n self._downloaded = True\n\n\nclass EmailEmitter(Emitter):\n def __init__(self, system_name, conf):\n super(EmailEmitter, self).__init__(system_name)\n self._conf = conf\n\n def _prepare_msg(self):\n if self._applied:\n subj = _(\"Updates applied on '%s'.\") % self._system_name\n elif self._downloaded:\n subj = _(\"Updates downloaded on '%s'.\") % self._system_name\n elif self._available_msg:\n subj = _(\"Updates available on '%s'.\") % self._system_name\n else:\n return None, None\n return subj, super(EmailEmitter, self)._prepare_msg()\n\n def commit(self):\n subj, body = self._prepare_msg()\n message = dnf.pycomp.email_mime(body)\n message.set_charset('utf-8')\n email_from = self._conf.email_from\n email_to = self._conf.email_to\n message['Date'] = email.utils.formatdate()\n message['From'] = email_from\n message['Subject'] = subj\n message['To'] = ','.join(email_to)\n message['Message-ID'] = email.utils.make_msgid()\n\n # Send the email\n try:\n smtp = smtplib.SMTP(self._conf.email_host, timeout=300)\n smtp.sendmail(email_from, email_to, message.as_string())\n smtp.close()\n except smtplib.SMTPException as exc:\n msg = _(\"Failed to send an email via '%s': %s\") % (\n self._conf.email_host, exc)\n logger.error(msg)\n\n\nclass CommandEmitterMixIn(object):\n \"\"\"\n Executes a desired command, and pushes data into its stdin.\n Both data and command can be formatted according to user preference.\n For this reason, this class expects a {str:str} dictionary as _prepare_msg\n return value.\n Meant for mixing with Emitter classes, as it does not define any names used\n for formatting on its own.\n \"\"\"\n def commit(self):\n command_fmt = self._conf.command_format\n stdin_fmt = self._conf.stdin_format\n msg = self._prepare_msg()\n # all strings passed to shell should be quoted to avoid accidental code\n # execution\n quoted_msg = dict((key, dnf.pycomp.shlex_quote(val))\n for key, val in msg.items())\n command = command_fmt.format(**quoted_msg)\n stdin_feed = stdin_fmt.format(**msg).encode('utf-8')\n\n # Execute the command\n subp = subprocess.Popen(command, shell=True, stdin=subprocess.PIPE)\n subp.communicate(stdin_feed)\n subp.stdin.close()\n if subp.wait() != 0:\n msg = _(\"Failed to execute command '%s': returned %d\") \\\n % (command, subp.returncode)\n logger.error(msg)\n\n\nclass CommandEmitter(CommandEmitterMixIn, Emitter):\n def __init__(self, system_name, conf):\n super(CommandEmitter, self).__init__(system_name)\n self._conf = conf\n\n def _prepare_msg(self):\n return {'body': super(CommandEmitter, self)._prepare_msg()}\n\n\nclass CommandEmailEmitter(CommandEmitterMixIn, EmailEmitter):\n def _prepare_msg(self):\n subject, body = super(CommandEmailEmitter, self)._prepare_msg()\n return {'subject': subject,\n 'body': body,\n 'email_from': self._conf.email_from,\n 'email_to': ' '.join(self._conf.email_to)}\n\n\nclass StdIoEmitter(Emitter):\n def commit(self):\n msg = self._prepare_msg()\n print(msg)\n\n\nclass MotdEmitter(Emitter):\n def commit(self):\n msg = self._prepare_msg()\n with open('/etc/motd', 'w') as fobj:\n fobj.write(msg)\n\n", "path": "dnf/automatic/emitter.py"}]}
| 2,873 | 141 |
gh_patches_debug_4302
|
rasdani/github-patches
|
git_diff
|
ydataai__ydata-profiling-588
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Nullable types not fully supported
Using pandas >= 1, columns of nullable integer type with low numbers of distinct values appear as "UNSUPPORTED".
This comes down to the check in `is_numeric` on whether there are infinite values in the series.
https://github.com/pandas-profiling/pandas-profiling/blob/216704bc6643ac691f3d1c513c0328c8cb29d088/src/pandas_profiling/model/base.py#L126
The following code gives a TypeError, which triggers the "except" branch in `get_var_type`
```
import pandas as pd
import numpy as np
series = pd.Series([1, None], dtype="Int64")
has_inf = any(np.inf == s or -np.inf == s for s in series)
```
I think that the following would be better:
```
has_inf = series.isin({np.inf, -np.inf}).any()
```
</issue>
<code>
[start of src/pandas_profiling/model/base.py]
1 """Common parts to all other modules, mainly utility functions."""
2 import imghdr
3 import os
4 from enum import Enum, unique
5 from urllib.parse import ParseResult, urlparse
6
7 import numpy as np
8 import pandas as pd
9 from pandas.api.types import is_categorical_dtype
10
11 from pandas_profiling.config import config
12
13
14 @unique
15 class Variable(Enum):
16 """The possible types of variables in the Profiling Report."""
17
18 TYPE_CAT = "CAT"
19 """A categorical variable"""
20
21 TYPE_BOOL = "BOOL"
22 """A boolean variable"""
23
24 TYPE_NUM = "NUM"
25 """A numeric variable"""
26
27 TYPE_DATE = "DATE"
28 """A date variable"""
29
30 TYPE_URL = "URL"
31 """A URL variable"""
32
33 TYPE_COMPLEX = "COMPLEX"
34
35 TYPE_PATH = "PATH"
36 """Absolute path"""
37
38 TYPE_FILE = "FILE"
39 """File (i.e. existing path)"""
40
41 TYPE_IMAGE = "IMAGE"
42 """Images"""
43
44 S_TYPE_UNSUPPORTED = "UNSUPPORTED"
45 """An unsupported variable"""
46
47
48 # Temporary mapping
49 Boolean = Variable.TYPE_BOOL
50 Real = Variable.TYPE_NUM
51 Count = Variable.TYPE_NUM
52 Complex = Variable.TYPE_COMPLEX
53 Date = Variable.TYPE_DATE
54 Categorical = Variable.TYPE_CAT
55 Url = Variable.TYPE_URL
56 AbsolutePath = Variable.TYPE_PATH
57 FilePath = Variable.TYPE_FILE
58 ImagePath = Variable.TYPE_IMAGE
59 Generic = Variable.S_TYPE_UNSUPPORTED
60
61
62 def get_counts(series: pd.Series) -> dict:
63 """Counts the values in a series (with and without NaN, distinct).
64
65 Args:
66 series: Series for which we want to calculate the values.
67
68 Returns:
69 A dictionary with the count values (with and without NaN, distinct).
70 """
71 if is_categorical_dtype(series):
72 series = series.cat.remove_unused_categories()
73
74 value_counts_with_nan = series.value_counts(dropna=False)
75 value_counts_without_nan = (
76 value_counts_with_nan.reset_index().dropna().set_index("index").iloc[:, 0]
77 )
78 distinct_count_with_nan = value_counts_with_nan.count()
79 distinct_count_without_nan = value_counts_without_nan.count()
80
81 return {
82 "value_counts": value_counts_without_nan, # Alias
83 "value_counts_with_nan": value_counts_with_nan,
84 "value_counts_without_nan": value_counts_without_nan,
85 "distinct_count_with_nan": distinct_count_with_nan,
86 "distinct_count_without_nan": distinct_count_without_nan,
87 }
88
89
90 def is_boolean(series: pd.Series, series_description: dict) -> bool:
91 """Is the series boolean type?
92
93 Args:
94 series: Series
95 series_description: Series description
96
97 Returns:
98 True is the series is boolean type in the broad sense (e.g. including yes/no, NaNs allowed).
99 """
100 keys = series_description["value_counts_without_nan"].keys()
101 if pd.api.types.is_bool_dtype(keys):
102 return True
103 elif (
104 1 <= series_description["distinct_count_without_nan"] <= 2
105 and pd.api.types.is_numeric_dtype(series)
106 and series[~series.isnull()].between(0, 1).all()
107 ):
108 return True
109 elif 1 <= series_description["distinct_count_without_nan"] <= 4:
110 unique_values = set([str(value).lower() for value in keys.values])
111 accepted_combinations = [
112 ["y", "n"],
113 ["yes", "no"],
114 ["true", "false"],
115 ["t", "f"],
116 ]
117
118 if len(unique_values) == 2 and any(
119 [unique_values == set(bools) for bools in accepted_combinations]
120 ):
121 return True
122
123 return False
124
125
126 def is_numeric(series: pd.Series, series_description: dict) -> bool:
127 """Is the series numeric type?
128
129 Args:
130 series: Series
131 series_description: Series description
132
133 Returns:
134 True is the series is numeric type (NaNs allowed).
135 """
136 return pd.api.types.is_numeric_dtype(series) and (
137 series_description["distinct_count_without_nan"]
138 >= config["vars"]["num"]["low_categorical_threshold"].get(int)
139 or any(np.inf == s or -np.inf == s for s in series)
140 )
141
142
143 def is_url(series: pd.Series, series_description: dict) -> bool:
144 """Is the series url type?
145
146 Args:
147 series: Series
148 series_description: Series description
149
150 Returns:
151 True is the series is url type (NaNs allowed).
152 """
153
154 def is_url_item(x):
155 return isinstance(x, ParseResult) and all((x.netloc, x.scheme, x.path))
156
157 if series_description["distinct_count_without_nan"] > 0:
158 try:
159 result = series[~series.isnull()].astype(str)
160 return all(is_url_item(urlparse(x)) for x in result)
161 except ValueError:
162 return False
163 else:
164 return False
165
166
167 def is_path(series, series_description) -> bool:
168 """Is the series of the path type (i.e. absolute path)?
169
170 Args:
171 series: Series
172 series_description: Series description
173
174 Returns:
175 True is the series is path type (NaNs allowed).
176 """
177 if series_description["distinct_count_without_nan"] == 0:
178 return False
179
180 try:
181 result = series[~series.isnull()].astype(str)
182 return all(os.path.isabs(p) for p in result)
183 except (ValueError, TypeError):
184 return False
185
186
187 def is_file(series, series_description) -> bool:
188 """Is the series of the type "file" (i.e. existing paths)?
189
190 Args:
191 series: Series
192 series_description: Series description
193
194 Returns:
195 True is the series is of the file type (NaNs allowed).
196 """
197 if series_description["distinct_count_without_nan"] == 0:
198 return False
199
200 try:
201 result = series[~series.isnull()].astype(str)
202 return all(os.path.exists(p) for p in result)
203 except (ValueError, TypeError):
204 return False
205
206
207 def is_image(series, series_description) -> bool:
208 """Is the series of the image type (i.e. "file" with image extensions)?
209
210 Args:
211 series: Series
212 series_description: Series description
213
214 Returns:
215 True is the series is of the image type (NaNs allowed).
216 """
217 if series_description["distinct_count_without_nan"] > 0:
218 try:
219 result = series[~series.isnull()].astype(str)
220 return all(imghdr.what(p) for p in result)
221 except (TypeError, ValueError):
222 return False
223 else:
224 return False
225
226
227 def is_date(series) -> bool:
228 """Is the variable of type datetime? Throws a warning if the series looks like a datetime, but is not typed as
229 datetime64.
230
231 Args:
232 series: Series
233
234 Returns:
235 True if the variable is of type datetime.
236 """
237 is_date_value = pd.api.types.is_datetime64_dtype(series)
238
239 return is_date_value
240
241
242 def get_var_type(series: pd.Series) -> dict:
243 """Get the variable type of a series.
244
245 Args:
246 series: Series for which we want to infer the variable type.
247
248 Returns:
249 The series updated with the variable type included.
250 """
251
252 series_description = {}
253
254 try:
255 series_description = get_counts(series)
256
257 # When the inferred type of the index is just "mixed" probably the types within the series are tuple, dict,
258 # list and so on...
259 if series_description[
260 "value_counts_without_nan"
261 ].index.inferred_type.startswith("mixed"):
262 raise TypeError("Not supported mixed type")
263
264 if series_description["distinct_count_without_nan"] == 0:
265 # Empty
266 var_type = Variable.S_TYPE_UNSUPPORTED
267 elif is_boolean(series, series_description):
268 var_type = Variable.TYPE_BOOL
269 elif is_numeric(series, series_description):
270 var_type = Variable.TYPE_NUM
271 elif is_date(series):
272 var_type = Variable.TYPE_DATE
273 elif is_url(series, series_description):
274 var_type = Variable.TYPE_URL
275 elif is_path(series, series_description):
276 if config["vars"]["file"]["active"].get(bool) and is_file(
277 series, series_description
278 ):
279 if config["vars"]["image"]["active"].get(bool) and is_image(
280 series, series_description
281 ):
282 var_type = Variable.TYPE_IMAGE
283 else:
284 var_type = Variable.TYPE_FILE
285 else:
286 var_type = Variable.TYPE_PATH
287 else:
288 var_type = Variable.TYPE_CAT
289 except TypeError:
290 var_type = Variable.S_TYPE_UNSUPPORTED
291
292 series_description.update({"type": var_type})
293
294 return series_description
295
[end of src/pandas_profiling/model/base.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/src/pandas_profiling/model/base.py b/src/pandas_profiling/model/base.py
--- a/src/pandas_profiling/model/base.py
+++ b/src/pandas_profiling/model/base.py
@@ -136,7 +136,7 @@
return pd.api.types.is_numeric_dtype(series) and (
series_description["distinct_count_without_nan"]
>= config["vars"]["num"]["low_categorical_threshold"].get(int)
- or any(np.inf == s or -np.inf == s for s in series)
+ or series.isin({np.inf, -np.inf}).any()
)
|
{"golden_diff": "diff --git a/src/pandas_profiling/model/base.py b/src/pandas_profiling/model/base.py\n--- a/src/pandas_profiling/model/base.py\n+++ b/src/pandas_profiling/model/base.py\n@@ -136,7 +136,7 @@\n return pd.api.types.is_numeric_dtype(series) and (\n series_description[\"distinct_count_without_nan\"]\n >= config[\"vars\"][\"num\"][\"low_categorical_threshold\"].get(int)\n- or any(np.inf == s or -np.inf == s for s in series)\n+ or series.isin({np.inf, -np.inf}).any()\n )\n", "issue": "Nullable types not fully supported\nUsing pandas >= 1, columns of nullable integer type with low numbers of distinct values appear as \"UNSUPPORTED\".\r\n\r\nThis comes down to the check in `is_numeric` on whether there are infinite values in the series.\r\n\r\nhttps://github.com/pandas-profiling/pandas-profiling/blob/216704bc6643ac691f3d1c513c0328c8cb29d088/src/pandas_profiling/model/base.py#L126\r\n\r\nThe following code gives a TypeError, which triggers the \"except\" branch in `get_var_type`\r\n\r\n```\r\nimport pandas as pd\r\nimport numpy as np\r\n\r\nseries = pd.Series([1, None], dtype=\"Int64\")\r\nhas_inf = any(np.inf == s or -np.inf == s for s in series)\r\n```\r\n\r\nI think that the following would be better:\r\n\r\n```\r\nhas_inf = series.isin({np.inf, -np.inf}).any()\r\n```\r\n\n", "before_files": [{"content": "\"\"\"Common parts to all other modules, mainly utility functions.\"\"\"\nimport imghdr\nimport os\nfrom enum import Enum, unique\nfrom urllib.parse import ParseResult, urlparse\n\nimport numpy as np\nimport pandas as pd\nfrom pandas.api.types import is_categorical_dtype\n\nfrom pandas_profiling.config import config\n\n\n@unique\nclass Variable(Enum):\n \"\"\"The possible types of variables in the Profiling Report.\"\"\"\n\n TYPE_CAT = \"CAT\"\n \"\"\"A categorical variable\"\"\"\n\n TYPE_BOOL = \"BOOL\"\n \"\"\"A boolean variable\"\"\"\n\n TYPE_NUM = \"NUM\"\n \"\"\"A numeric variable\"\"\"\n\n TYPE_DATE = \"DATE\"\n \"\"\"A date variable\"\"\"\n\n TYPE_URL = \"URL\"\n \"\"\"A URL variable\"\"\"\n\n TYPE_COMPLEX = \"COMPLEX\"\n\n TYPE_PATH = \"PATH\"\n \"\"\"Absolute path\"\"\"\n\n TYPE_FILE = \"FILE\"\n \"\"\"File (i.e. existing path)\"\"\"\n\n TYPE_IMAGE = \"IMAGE\"\n \"\"\"Images\"\"\"\n\n S_TYPE_UNSUPPORTED = \"UNSUPPORTED\"\n \"\"\"An unsupported variable\"\"\"\n\n\n# Temporary mapping\nBoolean = Variable.TYPE_BOOL\nReal = Variable.TYPE_NUM\nCount = Variable.TYPE_NUM\nComplex = Variable.TYPE_COMPLEX\nDate = Variable.TYPE_DATE\nCategorical = Variable.TYPE_CAT\nUrl = Variable.TYPE_URL\nAbsolutePath = Variable.TYPE_PATH\nFilePath = Variable.TYPE_FILE\nImagePath = Variable.TYPE_IMAGE\nGeneric = Variable.S_TYPE_UNSUPPORTED\n\n\ndef get_counts(series: pd.Series) -> dict:\n \"\"\"Counts the values in a series (with and without NaN, distinct).\n\n Args:\n series: Series for which we want to calculate the values.\n\n Returns:\n A dictionary with the count values (with and without NaN, distinct).\n \"\"\"\n if is_categorical_dtype(series):\n series = series.cat.remove_unused_categories()\n\n value_counts_with_nan = series.value_counts(dropna=False)\n value_counts_without_nan = (\n value_counts_with_nan.reset_index().dropna().set_index(\"index\").iloc[:, 0]\n )\n distinct_count_with_nan = value_counts_with_nan.count()\n distinct_count_without_nan = value_counts_without_nan.count()\n\n return {\n \"value_counts\": value_counts_without_nan, # Alias\n \"value_counts_with_nan\": value_counts_with_nan,\n \"value_counts_without_nan\": value_counts_without_nan,\n \"distinct_count_with_nan\": distinct_count_with_nan,\n \"distinct_count_without_nan\": distinct_count_without_nan,\n }\n\n\ndef is_boolean(series: pd.Series, series_description: dict) -> bool:\n \"\"\"Is the series boolean type?\n\n Args:\n series: Series\n series_description: Series description\n\n Returns:\n True is the series is boolean type in the broad sense (e.g. including yes/no, NaNs allowed).\n \"\"\"\n keys = series_description[\"value_counts_without_nan\"].keys()\n if pd.api.types.is_bool_dtype(keys):\n return True\n elif (\n 1 <= series_description[\"distinct_count_without_nan\"] <= 2\n and pd.api.types.is_numeric_dtype(series)\n and series[~series.isnull()].between(0, 1).all()\n ):\n return True\n elif 1 <= series_description[\"distinct_count_without_nan\"] <= 4:\n unique_values = set([str(value).lower() for value in keys.values])\n accepted_combinations = [\n [\"y\", \"n\"],\n [\"yes\", \"no\"],\n [\"true\", \"false\"],\n [\"t\", \"f\"],\n ]\n\n if len(unique_values) == 2 and any(\n [unique_values == set(bools) for bools in accepted_combinations]\n ):\n return True\n\n return False\n\n\ndef is_numeric(series: pd.Series, series_description: dict) -> bool:\n \"\"\"Is the series numeric type?\n\n Args:\n series: Series\n series_description: Series description\n\n Returns:\n True is the series is numeric type (NaNs allowed).\n \"\"\"\n return pd.api.types.is_numeric_dtype(series) and (\n series_description[\"distinct_count_without_nan\"]\n >= config[\"vars\"][\"num\"][\"low_categorical_threshold\"].get(int)\n or any(np.inf == s or -np.inf == s for s in series)\n )\n\n\ndef is_url(series: pd.Series, series_description: dict) -> bool:\n \"\"\"Is the series url type?\n\n Args:\n series: Series\n series_description: Series description\n\n Returns:\n True is the series is url type (NaNs allowed).\n \"\"\"\n\n def is_url_item(x):\n return isinstance(x, ParseResult) and all((x.netloc, x.scheme, x.path))\n\n if series_description[\"distinct_count_without_nan\"] > 0:\n try:\n result = series[~series.isnull()].astype(str)\n return all(is_url_item(urlparse(x)) for x in result)\n except ValueError:\n return False\n else:\n return False\n\n\ndef is_path(series, series_description) -> bool:\n \"\"\"Is the series of the path type (i.e. absolute path)?\n\n Args:\n series: Series\n series_description: Series description\n\n Returns:\n True is the series is path type (NaNs allowed).\n \"\"\"\n if series_description[\"distinct_count_without_nan\"] == 0:\n return False\n\n try:\n result = series[~series.isnull()].astype(str)\n return all(os.path.isabs(p) for p in result)\n except (ValueError, TypeError):\n return False\n\n\ndef is_file(series, series_description) -> bool:\n \"\"\"Is the series of the type \"file\" (i.e. existing paths)?\n\n Args:\n series: Series\n series_description: Series description\n\n Returns:\n True is the series is of the file type (NaNs allowed).\n \"\"\"\n if series_description[\"distinct_count_without_nan\"] == 0:\n return False\n\n try:\n result = series[~series.isnull()].astype(str)\n return all(os.path.exists(p) for p in result)\n except (ValueError, TypeError):\n return False\n\n\ndef is_image(series, series_description) -> bool:\n \"\"\"Is the series of the image type (i.e. \"file\" with image extensions)?\n\n Args:\n series: Series\n series_description: Series description\n\n Returns:\n True is the series is of the image type (NaNs allowed).\n \"\"\"\n if series_description[\"distinct_count_without_nan\"] > 0:\n try:\n result = series[~series.isnull()].astype(str)\n return all(imghdr.what(p) for p in result)\n except (TypeError, ValueError):\n return False\n else:\n return False\n\n\ndef is_date(series) -> bool:\n \"\"\"Is the variable of type datetime? Throws a warning if the series looks like a datetime, but is not typed as\n datetime64.\n\n Args:\n series: Series\n\n Returns:\n True if the variable is of type datetime.\n \"\"\"\n is_date_value = pd.api.types.is_datetime64_dtype(series)\n\n return is_date_value\n\n\ndef get_var_type(series: pd.Series) -> dict:\n \"\"\"Get the variable type of a series.\n\n Args:\n series: Series for which we want to infer the variable type.\n\n Returns:\n The series updated with the variable type included.\n \"\"\"\n\n series_description = {}\n\n try:\n series_description = get_counts(series)\n\n # When the inferred type of the index is just \"mixed\" probably the types within the series are tuple, dict,\n # list and so on...\n if series_description[\n \"value_counts_without_nan\"\n ].index.inferred_type.startswith(\"mixed\"):\n raise TypeError(\"Not supported mixed type\")\n\n if series_description[\"distinct_count_without_nan\"] == 0:\n # Empty\n var_type = Variable.S_TYPE_UNSUPPORTED\n elif is_boolean(series, series_description):\n var_type = Variable.TYPE_BOOL\n elif is_numeric(series, series_description):\n var_type = Variable.TYPE_NUM\n elif is_date(series):\n var_type = Variable.TYPE_DATE\n elif is_url(series, series_description):\n var_type = Variable.TYPE_URL\n elif is_path(series, series_description):\n if config[\"vars\"][\"file\"][\"active\"].get(bool) and is_file(\n series, series_description\n ):\n if config[\"vars\"][\"image\"][\"active\"].get(bool) and is_image(\n series, series_description\n ):\n var_type = Variable.TYPE_IMAGE\n else:\n var_type = Variable.TYPE_FILE\n else:\n var_type = Variable.TYPE_PATH\n else:\n var_type = Variable.TYPE_CAT\n except TypeError:\n var_type = Variable.S_TYPE_UNSUPPORTED\n\n series_description.update({\"type\": var_type})\n\n return series_description\n", "path": "src/pandas_profiling/model/base.py"}]}
| 3,421 | 131 |
gh_patches_debug_1048
|
rasdani/github-patches
|
git_diff
|
mindee__doctr-243
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Pb: unitest text_export_size not passing on tf 2.3.1
Unitest text_export_size not OK locally on tf 2.3.1 :
```
def test_export_sizes(test_convert_to_tflite, test_convert_to_fp16, test_quantize_model):
assert sys.getsizeof(test_convert_to_tflite) > sys.getsizeof(test_convert_to_fp16)
> assert sys.getsizeof(test_convert_to_fp16) > sys.getsizeof(test_quantize_model)
E AssertionError: assert 3041 > 3041
```
</issue>
<code>
[start of setup.py]
1 # Copyright (C) 2021, Mindee.
2
3 # This program is licensed under the Apache License version 2.
4 # See LICENSE or go to <https://www.apache.org/licenses/LICENSE-2.0.txt> for full license details.
5
6 """
7 Package installation setup
8 """
9
10 import os
11 from pathlib import Path
12 import subprocess
13
14 from setuptools import find_packages, setup
15
16
17 version = "0.1.2a0"
18 sha = 'Unknown'
19 package_name = 'doctr'
20
21 cwd = Path(__file__).parent.absolute()
22
23 if os.getenv('BUILD_VERSION'):
24 version = os.getenv('BUILD_VERSION')
25 elif sha != 'Unknown':
26 try:
27 sha = subprocess.check_output(['git', 'rev-parse', 'HEAD'], cwd=cwd).decode('ascii').strip()
28 except Exception:
29 pass
30 version += '+' + sha[:7]
31 print(f"Building wheel {package_name}-{version}")
32
33 with open(cwd.joinpath(package_name, 'version.py'), 'w') as f:
34 f.write(f"__version__ = '{version}'\n")
35
36 with open('README.md', 'r') as f:
37 readme = f.read()
38
39 requirements = [
40 "numpy>=1.16.0",
41 "scipy>=1.4.0",
42 "opencv-python>=4.2",
43 "tensorflow>=2.3.0",
44 "PyMuPDF>=1.16.0,<1.18.11",
45 "pyclipper>=1.2.0",
46 "shapely>=1.6.0",
47 "matplotlib>=3.1.0",
48 "mplcursors>=0.3",
49 "rapidfuzz>=1.0.0",
50 "weasyprint>=52.2",
51 ]
52
53 setup(
54 # Metadata
55 name=os.getenv('PKG_INDEX') if os.getenv('PKG_INDEX') else package_name,
56 version=version,
57 author='François-Guillaume Fernandez, Charles Gaillard',
58 author_email='[email protected]',
59 description='Extract valuable text information from your documents',
60 long_description=readme,
61 long_description_content_type="text/markdown",
62 url='https://github.com/mindee/doctr',
63 download_url='https://github.com/mindee/doctr/tags',
64 license='Apache',
65 classifiers=[
66 'Development Status :: 3 - Alpha',
67 'Intended Audience :: Developers',
68 'Intended Audience :: Science/Research',
69 'License :: OSI Approved :: Apache Software License',
70 'Natural Language :: English',
71 'Operating System :: OS Independent',
72 'Programming Language :: Python :: 3',
73 'Programming Language :: Python :: 3.6',
74 'Programming Language :: Python :: 3.7',
75 'Topic :: Scientific/Engineering',
76 'Topic :: Scientific/Engineering :: Artificial Intelligence',
77 'Topic :: Software Development',
78 'Topic :: Software Development :: Libraries',
79 'Topic :: Software Development :: Libraries :: Python Modules',
80 ],
81 keywords=['ocr', 'deep learning', 'tensorflow', 'text detection', 'text recognition'],
82
83 # Package info
84 packages=find_packages(exclude=('test',)),
85 zip_safe=True,
86 python_requires='>=3.6.0',
87 include_package_data=True,
88 install_requires=requirements,
89 package_data={'': ['LICENSE']}
90 )
91
[end of setup.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/setup.py b/setup.py
--- a/setup.py
+++ b/setup.py
@@ -40,7 +40,7 @@
"numpy>=1.16.0",
"scipy>=1.4.0",
"opencv-python>=4.2",
- "tensorflow>=2.3.0",
+ "tensorflow>=2.4.0",
"PyMuPDF>=1.16.0,<1.18.11",
"pyclipper>=1.2.0",
"shapely>=1.6.0",
|
{"golden_diff": "diff --git a/setup.py b/setup.py\n--- a/setup.py\n+++ b/setup.py\n@@ -40,7 +40,7 @@\n \"numpy>=1.16.0\",\n \"scipy>=1.4.0\",\n \"opencv-python>=4.2\",\n- \"tensorflow>=2.3.0\",\n+ \"tensorflow>=2.4.0\",\n \"PyMuPDF>=1.16.0,<1.18.11\",\n \"pyclipper>=1.2.0\",\n \"shapely>=1.6.0\",\n", "issue": "Pb: unitest text_export_size not passing on tf 2.3.1\nUnitest text_export_size not OK locally on tf 2.3.1 : \r\n\r\n```\r\ndef test_export_sizes(test_convert_to_tflite, test_convert_to_fp16, test_quantize_model):\r\n assert sys.getsizeof(test_convert_to_tflite) > sys.getsizeof(test_convert_to_fp16)\r\n> assert sys.getsizeof(test_convert_to_fp16) > sys.getsizeof(test_quantize_model)\r\nE AssertionError: assert 3041 > 3041\r\n\r\n```\n", "before_files": [{"content": "# Copyright (C) 2021, Mindee.\n\n# This program is licensed under the Apache License version 2.\n# See LICENSE or go to <https://www.apache.org/licenses/LICENSE-2.0.txt> for full license details.\n\n\"\"\"\nPackage installation setup\n\"\"\"\n\nimport os\nfrom pathlib import Path\nimport subprocess\n\nfrom setuptools import find_packages, setup\n\n\nversion = \"0.1.2a0\"\nsha = 'Unknown'\npackage_name = 'doctr'\n\ncwd = Path(__file__).parent.absolute()\n\nif os.getenv('BUILD_VERSION'):\n version = os.getenv('BUILD_VERSION')\nelif sha != 'Unknown':\n try:\n sha = subprocess.check_output(['git', 'rev-parse', 'HEAD'], cwd=cwd).decode('ascii').strip()\n except Exception:\n pass\n version += '+' + sha[:7]\nprint(f\"Building wheel {package_name}-{version}\")\n\nwith open(cwd.joinpath(package_name, 'version.py'), 'w') as f:\n f.write(f\"__version__ = '{version}'\\n\")\n\nwith open('README.md', 'r') as f:\n readme = f.read()\n\nrequirements = [\n \"numpy>=1.16.0\",\n \"scipy>=1.4.0\",\n \"opencv-python>=4.2\",\n \"tensorflow>=2.3.0\",\n \"PyMuPDF>=1.16.0,<1.18.11\",\n \"pyclipper>=1.2.0\",\n \"shapely>=1.6.0\",\n \"matplotlib>=3.1.0\",\n \"mplcursors>=0.3\",\n \"rapidfuzz>=1.0.0\",\n \"weasyprint>=52.2\",\n]\n\nsetup(\n # Metadata\n name=os.getenv('PKG_INDEX') if os.getenv('PKG_INDEX') else package_name,\n version=version,\n author='Fran\u00e7ois-Guillaume Fernandez, Charles Gaillard',\n author_email='[email protected]',\n description='Extract valuable text information from your documents',\n long_description=readme,\n long_description_content_type=\"text/markdown\",\n url='https://github.com/mindee/doctr',\n download_url='https://github.com/mindee/doctr/tags',\n license='Apache',\n classifiers=[\n 'Development Status :: 3 - Alpha',\n 'Intended Audience :: Developers',\n 'Intended Audience :: Science/Research',\n 'License :: OSI Approved :: Apache Software License',\n 'Natural Language :: English',\n 'Operating System :: OS Independent',\n 'Programming Language :: Python :: 3',\n 'Programming Language :: Python :: 3.6',\n 'Programming Language :: Python :: 3.7',\n 'Topic :: Scientific/Engineering',\n 'Topic :: Scientific/Engineering :: Artificial Intelligence',\n 'Topic :: Software Development',\n 'Topic :: Software Development :: Libraries',\n 'Topic :: Software Development :: Libraries :: Python Modules',\n ],\n keywords=['ocr', 'deep learning', 'tensorflow', 'text detection', 'text recognition'],\n\n # Package info\n packages=find_packages(exclude=('test',)),\n zip_safe=True,\n python_requires='>=3.6.0',\n include_package_data=True,\n install_requires=requirements,\n package_data={'': ['LICENSE']}\n)\n", "path": "setup.py"}]}
| 1,536 | 130 |
gh_patches_debug_1303
|
rasdani/github-patches
|
git_diff
|
fossasia__open-event-server-9132
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Add the unique ticket code into the downlad CSV file
The CSV download file of the attendee list does not include the numbers on the QR Code. Please add this field "Ticket-ID".
The ticket ID has the following format: 135ccbd7-9b23-4a52-a7fd-326fec1b2c1c
Whereas the order has a format like this: #O1691408152-34896

Expected: The exported CSV should have a table column "Ticket ID" with the ticket ID number that is encoded in the QR code as well.

</issue>
<code>
[start of app/models/ticket_holder.py]
1 import base64
2 from dataclasses import dataclass
3 from datetime import datetime
4 from io import BytesIO
5
6 import qrcode
7 from citext import CIText
8
9 from app.api.helpers.storage import UPLOAD_PATHS, generate_hash
10 from app.models import db
11 from app.models.base import SoftDeletionModel
12
13
14 @dataclass(init=False, unsafe_hash=True)
15 class TicketHolder(SoftDeletionModel):
16 __tablename__ = "ticket_holders"
17
18 id: int = db.Column(db.Integer, primary_key=True)
19 firstname: str = db.Column(db.String)
20 lastname: str = db.Column(db.String)
21 email: str = db.Column(CIText)
22 address: str = db.Column(db.String)
23 city: str = db.Column(db.String)
24 state: str = db.Column(db.String)
25 country: str = db.Column(db.String)
26 job_title: str = db.Column(db.String)
27 phone: str = db.Column(db.String)
28 tax_business_info: str = db.Column(db.String)
29 billing_address: str = db.Column(db.String)
30 home_address: str = db.Column(db.String)
31 shipping_address: str = db.Column(db.String)
32 company: str = db.Column(db.String)
33 work_address: str = db.Column(db.String)
34 work_phone: str = db.Column(db.String)
35 website: str = db.Column(db.String)
36 blog: str = db.Column(db.String)
37 twitter: str = db.Column(db.String)
38 facebook: str = db.Column(db.String)
39 instagram: str = db.Column(db.String)
40 linkedin: str = db.Column(db.String)
41 github: str = db.Column(db.String)
42 gender: str = db.Column(db.String)
43 accept_video_recording: bool = db.Column(db.Boolean)
44 accept_share_details: bool = db.Column(db.Boolean)
45 accept_receive_emails: bool = db.Column(db.Boolean)
46 age_group: str = db.Column(db.String)
47 home_wiki: str = db.Column(db.String)
48 wiki_scholarship: str = db.Column(db.String)
49 birth_date: datetime = db.Column(db.DateTime(timezone=True))
50 pdf_url: str = db.Column(db.String)
51 ticket_id: int = db.Column(
52 db.Integer, db.ForeignKey('tickets.id', ondelete='CASCADE'), nullable=False
53 )
54 order_id: int = db.Column(db.Integer, db.ForeignKey('orders.id', ondelete='CASCADE'))
55 is_checked_in: bool = db.Column(db.Boolean, default=False)
56 is_checked_out: bool = db.Column(db.Boolean, default=False)
57 is_registered: bool = db.Column(db.Boolean, default=False)
58 device_name_checkin: str = db.Column(db.String)
59 checkin_times: str = db.Column(db.String)
60 checkout_times: str = db.Column(db.String)
61 register_times: str = db.Column(db.String)
62 attendee_notes: str = db.Column(db.String)
63 event_id: int = db.Column(
64 db.Integer, db.ForeignKey('events.id', ondelete='CASCADE'), nullable=False
65 )
66 created_at: datetime = db.Column(db.DateTime(timezone=True), default=datetime.utcnow)
67 modified_at: datetime = db.Column(
68 db.DateTime(timezone=True), default=datetime.utcnow, onupdate=datetime.utcnow
69 )
70 complex_field_values: str = db.Column(db.JSON)
71 is_consent_of_refund_policy: bool = db.Column(db.Boolean, default=False)
72 native_language: str = db.Column(db.JSON)
73 fluent_language: str = db.Column(db.JSON)
74 user = db.relationship(
75 'User',
76 foreign_keys=[email],
77 primaryjoin='User.email == TicketHolder.email',
78 viewonly=True,
79 backref='attendees',
80 sync_backref=False,
81 )
82 order = db.relationship('Order', backref='ticket_holders')
83 ticket = db.relationship('Ticket', backref='ticket_holders')
84 is_consent_form_field: bool = db.Column(db.Boolean, default=False)
85 is_consent_form_field_photo: bool = db.Column(db.Boolean, default=False)
86 is_consent_form_field_email: bool = db.Column(db.Boolean, default=False)
87 is_badge_printed: bool = db.Column(db.Boolean, default=False)
88 badge_printed_at: datetime = db.Column(db.DateTime(timezone=True))
89 is_discount_applied: bool = db.Column(db.Boolean, default=False)
90 is_access_code_applied: bool = db.Column(db.Boolean, default=False)
91 tag_id: int = db.Column(db.Integer, db.ForeignKey('tags.id', ondelete='CASCADE'))
92 tag = db.relationship('Tag', backref='ticket_holders')
93
94 @property
95 def name(self):
96 firstname = self.firstname if self.firstname else ''
97 lastname = self.lastname if self.lastname else ''
98 if firstname and lastname:
99 return f'{firstname} {lastname}'
100 else:
101 return ''
102
103 @property
104 def qr_code(self):
105 qr = qrcode.QRCode(
106 version=1,
107 error_correction=qrcode.constants.ERROR_CORRECT_L,
108 box_size=10,
109 border=0,
110 )
111 qr.add_data(self.order.identifier + "-" + str(self.id))
112 qr.make(fit=True)
113 img = qr.make_image()
114
115 buffer = BytesIO()
116 img.save(buffer, format="JPEG")
117 img_str = str(base64.b64encode(buffer.getvalue()), 'utf-8')
118 return img_str
119
120 @property
121 def serialize(self):
122 """Return object data in easily serializable format"""
123 return {
124 'id': self.id,
125 'firstname': self.firstname,
126 'lastname': self.lastname,
127 'email': self.email,
128 'city': self.city,
129 'address': self.address,
130 'state': self.state,
131 'country': self.country,
132 'company': self.company,
133 'taxBusinessInfo': self.tax_business_info,
134 }
135
136 @property
137 def pdf_url_path(self) -> str:
138 key = UPLOAD_PATHS['pdf']['tickets_all'].format(
139 identifier=self.order.identifier, extra_identifier=self.id
140 )
141 return (
142 f'generated/tickets/{key}/{generate_hash(key)}/'
143 + self.order.identifier
144 + '.pdf'
145 )
146
[end of app/models/ticket_holder.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/app/models/ticket_holder.py b/app/models/ticket_holder.py
--- a/app/models/ticket_holder.py
+++ b/app/models/ticket_holder.py
@@ -108,7 +108,7 @@
box_size=10,
border=0,
)
- qr.add_data(self.order.identifier + "-" + str(self.id))
+ qr.add_data(self.order.identifier)
qr.make(fit=True)
img = qr.make_image()
|
{"golden_diff": "diff --git a/app/models/ticket_holder.py b/app/models/ticket_holder.py\n--- a/app/models/ticket_holder.py\n+++ b/app/models/ticket_holder.py\n@@ -108,7 +108,7 @@\n box_size=10,\n border=0,\n )\n- qr.add_data(self.order.identifier + \"-\" + str(self.id))\n+ qr.add_data(self.order.identifier)\n qr.make(fit=True)\n img = qr.make_image()\n", "issue": "Add the unique ticket code into the downlad CSV file\nThe CSV download file of the attendee list does not include the numbers on the QR Code. Please add this field \"Ticket-ID\".\r\n\r\nThe ticket ID has the following format: 135ccbd7-9b23-4a52-a7fd-326fec1b2c1c\r\nWhereas the order has a format like this: #O1691408152-34896\r\n\r\n\r\n\r\nExpected: The exported CSV should have a table column \"Ticket ID\" with the ticket ID number that is encoded in the QR code as well.\r\n\r\n\r\n\n", "before_files": [{"content": "import base64\nfrom dataclasses import dataclass\nfrom datetime import datetime\nfrom io import BytesIO\n\nimport qrcode\nfrom citext import CIText\n\nfrom app.api.helpers.storage import UPLOAD_PATHS, generate_hash\nfrom app.models import db\nfrom app.models.base import SoftDeletionModel\n\n\n@dataclass(init=False, unsafe_hash=True)\nclass TicketHolder(SoftDeletionModel):\n __tablename__ = \"ticket_holders\"\n\n id: int = db.Column(db.Integer, primary_key=True)\n firstname: str = db.Column(db.String)\n lastname: str = db.Column(db.String)\n email: str = db.Column(CIText)\n address: str = db.Column(db.String)\n city: str = db.Column(db.String)\n state: str = db.Column(db.String)\n country: str = db.Column(db.String)\n job_title: str = db.Column(db.String)\n phone: str = db.Column(db.String)\n tax_business_info: str = db.Column(db.String)\n billing_address: str = db.Column(db.String)\n home_address: str = db.Column(db.String)\n shipping_address: str = db.Column(db.String)\n company: str = db.Column(db.String)\n work_address: str = db.Column(db.String)\n work_phone: str = db.Column(db.String)\n website: str = db.Column(db.String)\n blog: str = db.Column(db.String)\n twitter: str = db.Column(db.String)\n facebook: str = db.Column(db.String)\n instagram: str = db.Column(db.String)\n linkedin: str = db.Column(db.String)\n github: str = db.Column(db.String)\n gender: str = db.Column(db.String)\n accept_video_recording: bool = db.Column(db.Boolean)\n accept_share_details: bool = db.Column(db.Boolean)\n accept_receive_emails: bool = db.Column(db.Boolean)\n age_group: str = db.Column(db.String)\n home_wiki: str = db.Column(db.String)\n wiki_scholarship: str = db.Column(db.String)\n birth_date: datetime = db.Column(db.DateTime(timezone=True))\n pdf_url: str = db.Column(db.String)\n ticket_id: int = db.Column(\n db.Integer, db.ForeignKey('tickets.id', ondelete='CASCADE'), nullable=False\n )\n order_id: int = db.Column(db.Integer, db.ForeignKey('orders.id', ondelete='CASCADE'))\n is_checked_in: bool = db.Column(db.Boolean, default=False)\n is_checked_out: bool = db.Column(db.Boolean, default=False)\n is_registered: bool = db.Column(db.Boolean, default=False)\n device_name_checkin: str = db.Column(db.String)\n checkin_times: str = db.Column(db.String)\n checkout_times: str = db.Column(db.String)\n register_times: str = db.Column(db.String)\n attendee_notes: str = db.Column(db.String)\n event_id: int = db.Column(\n db.Integer, db.ForeignKey('events.id', ondelete='CASCADE'), nullable=False\n )\n created_at: datetime = db.Column(db.DateTime(timezone=True), default=datetime.utcnow)\n modified_at: datetime = db.Column(\n db.DateTime(timezone=True), default=datetime.utcnow, onupdate=datetime.utcnow\n )\n complex_field_values: str = db.Column(db.JSON)\n is_consent_of_refund_policy: bool = db.Column(db.Boolean, default=False)\n native_language: str = db.Column(db.JSON)\n fluent_language: str = db.Column(db.JSON)\n user = db.relationship(\n 'User',\n foreign_keys=[email],\n primaryjoin='User.email == TicketHolder.email',\n viewonly=True,\n backref='attendees',\n sync_backref=False,\n )\n order = db.relationship('Order', backref='ticket_holders')\n ticket = db.relationship('Ticket', backref='ticket_holders')\n is_consent_form_field: bool = db.Column(db.Boolean, default=False)\n is_consent_form_field_photo: bool = db.Column(db.Boolean, default=False)\n is_consent_form_field_email: bool = db.Column(db.Boolean, default=False)\n is_badge_printed: bool = db.Column(db.Boolean, default=False)\n badge_printed_at: datetime = db.Column(db.DateTime(timezone=True))\n is_discount_applied: bool = db.Column(db.Boolean, default=False)\n is_access_code_applied: bool = db.Column(db.Boolean, default=False)\n tag_id: int = db.Column(db.Integer, db.ForeignKey('tags.id', ondelete='CASCADE'))\n tag = db.relationship('Tag', backref='ticket_holders')\n\n @property\n def name(self):\n firstname = self.firstname if self.firstname else ''\n lastname = self.lastname if self.lastname else ''\n if firstname and lastname:\n return f'{firstname} {lastname}'\n else:\n return ''\n\n @property\n def qr_code(self):\n qr = qrcode.QRCode(\n version=1,\n error_correction=qrcode.constants.ERROR_CORRECT_L,\n box_size=10,\n border=0,\n )\n qr.add_data(self.order.identifier + \"-\" + str(self.id))\n qr.make(fit=True)\n img = qr.make_image()\n\n buffer = BytesIO()\n img.save(buffer, format=\"JPEG\")\n img_str = str(base64.b64encode(buffer.getvalue()), 'utf-8')\n return img_str\n\n @property\n def serialize(self):\n \"\"\"Return object data in easily serializable format\"\"\"\n return {\n 'id': self.id,\n 'firstname': self.firstname,\n 'lastname': self.lastname,\n 'email': self.email,\n 'city': self.city,\n 'address': self.address,\n 'state': self.state,\n 'country': self.country,\n 'company': self.company,\n 'taxBusinessInfo': self.tax_business_info,\n }\n\n @property\n def pdf_url_path(self) -> str:\n key = UPLOAD_PATHS['pdf']['tickets_all'].format(\n identifier=self.order.identifier, extra_identifier=self.id\n )\n return (\n f'generated/tickets/{key}/{generate_hash(key)}/'\n + self.order.identifier\n + '.pdf'\n )\n", "path": "app/models/ticket_holder.py"}]}
| 2,478 | 102 |
gh_patches_debug_30338
|
rasdani/github-patches
|
git_diff
|
ansible__molecule-3105
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Add support for checking exit codes on shell dependencies
# Issue Type
- Feature request
# Molecule and Ansible details
```
ansible 2.10.5
config file = /Users/jhg03a/<redacted>/ansible.cfg
configured module search path = ['/Users/jhg03a/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules']
ansible python module location = /usr/local/lib/python3.9/site-packages/ansible
executable location = /usr/local/bin/ansible
python version = 3.9.2 (default, Feb 24 2021, 13:26:09) [Clang 12.0.0 (clang-1200.0.32.29)]
molecule 3.2.3 using python 3.9
ansible:2.10.5
delegated:3.2.3 from molecule
docker:0.2.4 from molecule_docker
```
Molecule installation method (one of):
- pip
Ansible installation method (one of):
- pip
# Desired Behavior
Currently it appears that the dependency shell module doesn't take into account the exit code from the command. If something goes wrong in the dependency phase, it's highly likely the rest of the run is going to fail or have inconsistent results.
### Example:
```yaml
dependency:
name: shell
command: 'false'
```
</issue>
<code>
[start of src/molecule/dependency/base.py]
1 # Copyright (c) 2015-2018 Cisco Systems, Inc.
2 #
3 # Permission is hereby granted, free of charge, to any person obtaining a copy
4 # of this software and associated documentation files (the "Software"), to
5 # deal in the Software without restriction, including without limitation the
6 # rights to use, copy, modify, merge, publish, distribute, sublicense, and/or
7 # sell copies of the Software, and to permit persons to whom the Software is
8 # furnished to do so, subject to the following conditions:
9 #
10 # The above copyright notice and this permission notice shall be included in
11 # all copies or substantial portions of the Software.
12 #
13 # THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
14 # IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
15 # FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
16 # AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
17 # LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
18 # FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
19 # DEALINGS IN THE SOFTWARE.
20 """Base Dependency Module."""
21
22 import abc
23 import logging
24 import os
25 import time
26
27 from molecule import constants, util
28
29 LOG = logging.getLogger(__name__)
30
31
32 class Base(object):
33 """Dependency Base Class."""
34
35 __metaclass__ = abc.ABCMeta
36
37 RETRY = 3
38 SLEEP = 3
39 BACKOFF = 3
40
41 def __init__(self, config):
42 """
43 Initialize code for all :ref:`Dependency` classes.
44
45 :param config: An instance of a Molecule config.
46 :returns: None
47 """
48 self._config = config
49
50 def execute_with_retries(self):
51 """Run dependency downloads with retry and timed back-off."""
52 exception = None
53
54 try:
55 # print(555, self._sh_command)
56 util.run_command(self._sh_command, debug=self._config.debug)
57 msg = "Dependency completed successfully."
58 LOG.info(msg)
59 return
60 except Exception:
61 pass
62
63 for counter in range(1, (self.RETRY + 1)):
64 msg = "Retrying dependency ... {}/{} time(s)".format(counter, self.RETRY)
65 LOG.warning(msg)
66
67 msg = "Sleeping {} seconds before retrying ...".format(self.SLEEP)
68 LOG.warning(msg)
69 time.sleep(self.SLEEP)
70 self.SLEEP += self.BACKOFF
71
72 try:
73 util.run_command(self._sh_command, debug=self._config.debug)
74 msg = "Dependency completed successfully."
75 LOG.info(msg)
76 return
77 except Exception as _exception:
78 exception = _exception
79
80 LOG.error(str(exception), self._sh_command)
81 util.sysexit(getattr(exception, "exit_code", constants.RC_UNKNOWN_ERROR))
82
83 @abc.abstractmethod
84 def execute(self): # pragma: no cover
85 """
86 Execute ``cmd`` and returns None.
87
88 :return: None
89 """
90
91 @abc.abstractproperty
92 def default_options(self): # pragma: no cover
93 """
94 Get default CLI arguments provided to ``cmd`` as a dict.
95
96 :return: dict
97 """
98
99 @property
100 def default_env(self): # pragma: no cover
101 """
102 Get default env variables provided to ``cmd`` as a dict.
103
104 :return: dict
105 """
106 env = util.merge_dicts(os.environ, self._config.env)
107 # inject ephemeral_directory on top of path
108 env[self._config.ansible_collections_path] = os.path.join(
109 self._config.scenario.ephemeral_directory, "collections"
110 )
111 return env
112
113 @property
114 def name(self):
115 """
116 Name of the dependency and returns a string.
117
118 :returns: str
119 """
120 return self._config.config["dependency"]["name"]
121
122 @property
123 def enabled(self):
124 return self._config.config["dependency"]["enabled"]
125
126 @property
127 def options(self):
128 return util.merge_dicts(
129 self.default_options, self._config.config["dependency"]["options"]
130 )
131
132 @property
133 def env(self):
134 return util.merge_dicts(
135 self.default_env, self._config.config["dependency"]["env"]
136 )
137
[end of src/molecule/dependency/base.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/src/molecule/dependency/base.py b/src/molecule/dependency/base.py
--- a/src/molecule/dependency/base.py
+++ b/src/molecule/dependency/base.py
@@ -23,8 +23,9 @@
import logging
import os
import time
+from subprocess import CalledProcessError
-from molecule import constants, util
+from molecule import util
LOG = logging.getLogger(__name__)
@@ -53,11 +54,11 @@
try:
# print(555, self._sh_command)
- util.run_command(self._sh_command, debug=self._config.debug)
+ util.run_command(self._sh_command, debug=self._config.debug, check=True)
msg = "Dependency completed successfully."
LOG.info(msg)
return
- except Exception:
+ except CalledProcessError:
pass
for counter in range(1, (self.RETRY + 1)):
@@ -70,15 +71,15 @@
self.SLEEP += self.BACKOFF
try:
- util.run_command(self._sh_command, debug=self._config.debug)
+ util.run_command(self._sh_command, debug=self._config.debug, check=True)
msg = "Dependency completed successfully."
LOG.info(msg)
return
- except Exception as _exception:
+ except CalledProcessError as _exception:
exception = _exception
- LOG.error(str(exception), self._sh_command)
- util.sysexit(getattr(exception, "exit_code", constants.RC_UNKNOWN_ERROR))
+ LOG.error(str(exception))
+ util.sysexit(exception.returncode)
@abc.abstractmethod
def execute(self): # pragma: no cover
|
{"golden_diff": "diff --git a/src/molecule/dependency/base.py b/src/molecule/dependency/base.py\n--- a/src/molecule/dependency/base.py\n+++ b/src/molecule/dependency/base.py\n@@ -23,8 +23,9 @@\n import logging\n import os\n import time\n+from subprocess import CalledProcessError\n \n-from molecule import constants, util\n+from molecule import util\n \n LOG = logging.getLogger(__name__)\n \n@@ -53,11 +54,11 @@\n \n try:\n # print(555, self._sh_command)\n- util.run_command(self._sh_command, debug=self._config.debug)\n+ util.run_command(self._sh_command, debug=self._config.debug, check=True)\n msg = \"Dependency completed successfully.\"\n LOG.info(msg)\n return\n- except Exception:\n+ except CalledProcessError:\n pass\n \n for counter in range(1, (self.RETRY + 1)):\n@@ -70,15 +71,15 @@\n self.SLEEP += self.BACKOFF\n \n try:\n- util.run_command(self._sh_command, debug=self._config.debug)\n+ util.run_command(self._sh_command, debug=self._config.debug, check=True)\n msg = \"Dependency completed successfully.\"\n LOG.info(msg)\n return\n- except Exception as _exception:\n+ except CalledProcessError as _exception:\n exception = _exception\n \n- LOG.error(str(exception), self._sh_command)\n- util.sysexit(getattr(exception, \"exit_code\", constants.RC_UNKNOWN_ERROR))\n+ LOG.error(str(exception))\n+ util.sysexit(exception.returncode)\n \n @abc.abstractmethod\n def execute(self): # pragma: no cover\n", "issue": "Add support for checking exit codes on shell dependencies\n# Issue Type\r\n\r\n- Feature request\r\n\r\n# Molecule and Ansible details\r\n\r\n```\r\nansible 2.10.5\r\n config file = /Users/jhg03a/<redacted>/ansible.cfg\r\n configured module search path = ['/Users/jhg03a/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules']\r\n ansible python module location = /usr/local/lib/python3.9/site-packages/ansible\r\n executable location = /usr/local/bin/ansible\r\n python version = 3.9.2 (default, Feb 24 2021, 13:26:09) [Clang 12.0.0 (clang-1200.0.32.29)]\r\nmolecule 3.2.3 using python 3.9\r\n ansible:2.10.5\r\n delegated:3.2.3 from molecule\r\n docker:0.2.4 from molecule_docker\r\n```\r\n\r\nMolecule installation method (one of):\r\n\r\n- pip\r\n\r\nAnsible installation method (one of):\r\n\r\n- pip\r\n\r\n# Desired Behavior\r\n\r\nCurrently it appears that the dependency shell module doesn't take into account the exit code from the command. If something goes wrong in the dependency phase, it's highly likely the rest of the run is going to fail or have inconsistent results.\r\n\r\n### Example:\r\n```yaml\r\ndependency:\r\n name: shell\r\n command: 'false'\r\n```\r\n\n", "before_files": [{"content": "# Copyright (c) 2015-2018 Cisco Systems, Inc.\n#\n# Permission is hereby granted, free of charge, to any person obtaining a copy\n# of this software and associated documentation files (the \"Software\"), to\n# deal in the Software without restriction, including without limitation the\n# rights to use, copy, modify, merge, publish, distribute, sublicense, and/or\n# sell copies of the Software, and to permit persons to whom the Software is\n# furnished to do so, subject to the following conditions:\n#\n# The above copyright notice and this permission notice shall be included in\n# all copies or substantial portions of the Software.\n#\n# THE SOFTWARE IS PROVIDED \"AS IS\", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR\n# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,\n# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE\n# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER\n# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING\n# FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER\n# DEALINGS IN THE SOFTWARE.\n\"\"\"Base Dependency Module.\"\"\"\n\nimport abc\nimport logging\nimport os\nimport time\n\nfrom molecule import constants, util\n\nLOG = logging.getLogger(__name__)\n\n\nclass Base(object):\n \"\"\"Dependency Base Class.\"\"\"\n\n __metaclass__ = abc.ABCMeta\n\n RETRY = 3\n SLEEP = 3\n BACKOFF = 3\n\n def __init__(self, config):\n \"\"\"\n Initialize code for all :ref:`Dependency` classes.\n\n :param config: An instance of a Molecule config.\n :returns: None\n \"\"\"\n self._config = config\n\n def execute_with_retries(self):\n \"\"\"Run dependency downloads with retry and timed back-off.\"\"\"\n exception = None\n\n try:\n # print(555, self._sh_command)\n util.run_command(self._sh_command, debug=self._config.debug)\n msg = \"Dependency completed successfully.\"\n LOG.info(msg)\n return\n except Exception:\n pass\n\n for counter in range(1, (self.RETRY + 1)):\n msg = \"Retrying dependency ... {}/{} time(s)\".format(counter, self.RETRY)\n LOG.warning(msg)\n\n msg = \"Sleeping {} seconds before retrying ...\".format(self.SLEEP)\n LOG.warning(msg)\n time.sleep(self.SLEEP)\n self.SLEEP += self.BACKOFF\n\n try:\n util.run_command(self._sh_command, debug=self._config.debug)\n msg = \"Dependency completed successfully.\"\n LOG.info(msg)\n return\n except Exception as _exception:\n exception = _exception\n\n LOG.error(str(exception), self._sh_command)\n util.sysexit(getattr(exception, \"exit_code\", constants.RC_UNKNOWN_ERROR))\n\n @abc.abstractmethod\n def execute(self): # pragma: no cover\n \"\"\"\n Execute ``cmd`` and returns None.\n\n :return: None\n \"\"\"\n\n @abc.abstractproperty\n def default_options(self): # pragma: no cover\n \"\"\"\n Get default CLI arguments provided to ``cmd`` as a dict.\n\n :return: dict\n \"\"\"\n\n @property\n def default_env(self): # pragma: no cover\n \"\"\"\n Get default env variables provided to ``cmd`` as a dict.\n\n :return: dict\n \"\"\"\n env = util.merge_dicts(os.environ, self._config.env)\n # inject ephemeral_directory on top of path\n env[self._config.ansible_collections_path] = os.path.join(\n self._config.scenario.ephemeral_directory, \"collections\"\n )\n return env\n\n @property\n def name(self):\n \"\"\"\n Name of the dependency and returns a string.\n\n :returns: str\n \"\"\"\n return self._config.config[\"dependency\"][\"name\"]\n\n @property\n def enabled(self):\n return self._config.config[\"dependency\"][\"enabled\"]\n\n @property\n def options(self):\n return util.merge_dicts(\n self.default_options, self._config.config[\"dependency\"][\"options\"]\n )\n\n @property\n def env(self):\n return util.merge_dicts(\n self.default_env, self._config.config[\"dependency\"][\"env\"]\n )\n", "path": "src/molecule/dependency/base.py"}]}
| 2,107 | 372 |
gh_patches_debug_4099
|
rasdani/github-patches
|
git_diff
|
interlegis__sapl-1234
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Aumenta o tamanho máximo de arquivos de upload para 10MB
Atualmente o tamanho máximo de arquivos de upload é de 5MB. Ao menos uma casa reclamou que este tamanho é insuficiente então vamos dobrar e ver se mais alguém ainda reclama do tamanho de 10MB.
</issue>
<code>
[start of sapl/settings.py]
1 """
2 Django settings for sapl project.
3
4 Generated by 'django-admin startproject' using Django 1.8.2.
5
6 For more information on this file, see
7 https://docs.djangoproject.com/en/1.8/topics/settings/
8
9 For the full list of settings and their values, see
10 https://docs.djangoproject.com/en/1.8/ref/settings/
11
12 Quick-start development settings - unsuitable for production
13 See https://docs.djangoproject.com/en/1.8/howto/deployment/checklist/
14
15 """
16 import logging
17
18 from decouple import config
19 from dj_database_url import parse as db_url
20 from unipath import Path
21
22 from .temp_suppress_crispy_form_warnings import \
23 SUPRESS_CRISPY_FORM_WARNINGS_LOGGING
24
25 BASE_DIR = Path(__file__).ancestor(1)
26 PROJECT_DIR = Path(__file__).ancestor(2)
27
28
29 # SECURITY WARNING: keep the secret key used in production secret!
30 SECRET_KEY = config('SECRET_KEY', default='')
31 # SECURITY WARNING: don't run with debug turned on in production!
32 DEBUG = config('DEBUG', default=False, cast=bool)
33
34 MESSAGE_STORAGE = 'django.contrib.messages.storage.session.SessionStorage'
35
36 ALLOWED_HOSTS = ['*']
37
38 LOGIN_REDIRECT_URL = '/'
39 LOGIN_URL = '/login/?next='
40
41 EMAIL_BACKEND = 'django.core.mail.backends.smtp.EmailBackend'
42
43
44 # SAPL business apps in dependency order
45 SAPL_APPS = (
46 'sapl.base',
47 'sapl.crud',
48 'sapl.parlamentares',
49 'sapl.comissoes',
50 'sapl.materia',
51 'sapl.norma',
52 'sapl.sessao',
53 'sapl.lexml',
54 'sapl.painel',
55 'sapl.protocoloadm',
56 'sapl.redireciona_urls',
57 'sapl.compilacao',
58 'sapl.api',
59
60 'sapl.rules'
61
62 )
63
64 INSTALLED_APPS = (
65 'django_admin_bootstrapped', # must come before django.contrib.admin
66 'django.contrib.admin',
67 'django.contrib.auth',
68 'django.contrib.contenttypes',
69 'django.contrib.sessions',
70 'django.contrib.messages',
71 'django.contrib.staticfiles',
72
73 # more
74 'django_extensions',
75 'djangobower',
76 'bootstrap3', # basically for django_admin_bootstrapped
77 'crispy_forms',
78 'easy_thumbnails',
79 'floppyforms',
80 'haystack',
81 'sass_processor',
82 'rest_framework',
83 'reversion',
84 'whoosh',
85
86 ) + SAPL_APPS
87
88 # FTS = Full Text Search
89 SEARCH_BACKEND = 'haystack.backends.whoosh_backend.WhooshEngine'
90 SEARCH_URL = ('PATH', PROJECT_DIR.child('whoosh'))
91
92 SOLR_URL = config('SOLR_URL', cast=str, default='')
93 if SOLR_URL:
94 SEARCH_BACKEND = 'haystack.backends.solr_backend.SolrEngine'
95 SEARCH_URL = ('URL', config('SOLR_URL', cast=str))
96 # ...or for multicore...
97 # 'URL': 'http://127.0.0.1:8983/solr/mysite',
98
99
100 HAYSTACK_CONNECTIONS = {
101 'default': {
102 'ENGINE': SEARCH_BACKEND,
103 SEARCH_URL[0] : SEARCH_URL[1]
104 },
105 }
106
107
108 if DEBUG:
109 INSTALLED_APPS += ('debug_toolbar', 'rest_framework_docs',)
110
111 MIDDLEWARE_CLASSES = (
112 'reversion.middleware.RevisionMiddleware',
113 'django.contrib.sessions.middleware.SessionMiddleware',
114 'django.middleware.locale.LocaleMiddleware',
115 'django.middleware.common.CommonMiddleware',
116 'django.middleware.csrf.CsrfViewMiddleware',
117 'django.contrib.auth.middleware.AuthenticationMiddleware',
118 'django.contrib.auth.middleware.SessionAuthenticationMiddleware',
119 'django.contrib.messages.middleware.MessageMiddleware',
120 'django.middleware.clickjacking.XFrameOptionsMiddleware',
121 'django.middleware.security.SecurityMiddleware',
122 )
123
124
125 REST_FRAMEWORK = {
126 "UNICODE_JSON": False,
127 "DEFAULT_PARSER_CLASSES": (
128 "rest_framework.parsers.JSONParser",
129 ),
130 "DEFAULT_PERMISSION_CLASSES": (
131 "rest_framework.permissions.IsAuthenticated",
132 "sapl.api.permissions.DjangoModelPermissions",
133 ),
134 "DEFAULT_AUTHENTICATION_CLASSES": (
135 "rest_framework.authentication.SessionAuthentication",
136 ),
137 "DEFAULT_PAGINATION_CLASS": "sapl.api.pagination.StandardPagination",
138 "DEFAULT_FILTER_BACKENDS": (
139 "rest_framework.filters.SearchFilter",
140 "rest_framework.filters.DjangoFilterBackend",
141 ),
142 }
143
144
145 ROOT_URLCONF = 'sapl.urls'
146
147 TEMPLATES = [
148 {
149 'BACKEND': 'django.template.backends.django.DjangoTemplates',
150 'DIRS': ['sapl/templates'],
151 'APP_DIRS': True,
152 'OPTIONS': {
153 'context_processors': [
154 'django.template.context_processors.debug',
155 'django.template.context_processors.request',
156 'django.contrib.auth.context_processors.auth',
157 "django.template.context_processors.media",
158 "django.template.context_processors.static",
159 'django.contrib.messages.context_processors.messages',
160 'sapl.context_processors.parliament_info',
161 ],
162 'debug': DEBUG
163 },
164 },
165 ]
166
167
168 WSGI_APPLICATION = 'sapl.wsgi.application'
169
170 # Database
171 # https://docs.djangoproject.com/en/1.8/ref/settings/#databases
172
173 DATABASES = {
174 'default': config(
175 'DATABASE_URL', default='sqlite://:memory:',
176 cast=db_url,
177 )
178 }
179
180 # troque no caso de reimplementação da classe User conforme
181 # https://docs.djangoproject.com/en/1.9/topics/auth/customizing/#substituting-a-custom-user-model
182 AUTH_USER_MODEL = 'auth.User'
183
184 X_FRAME_OPTIONS = 'ALLOWALL'
185
186 EMAIL_HOST = config('EMAIL_HOST', default='localhost')
187 EMAIL_PORT = config('EMAIL_PORT', cast=int, default=587)
188 EMAIL_HOST_USER = config('EMAIL_HOST_USER', default='')
189 EMAIL_HOST_PASSWORD = config('EMAIL_HOST_PASSWORD', default='')
190 EMAIL_USE_TLS = config('EMAIL_USE_TLS', cast=bool, default=True)
191 EMAIL_SEND_USER = config('EMAIL_SEND_USER', cast=str, default='')
192 DEFAULT_FROM_EMAIL = config('DEFAULT_FROM_EMAIL', cast=str, default='')
193 SERVER_EMAIL = config('SERVER_EMAIL', cast=str, default='')
194
195 MAX_DOC_UPLOAD_SIZE = 5 * 1024 * 1024 # 5MB
196 MAX_IMAGE_UPLOAD_SIZE = 2 * 1024 * 1024 # 2MB
197
198 # Internationalization
199 # https://docs.djangoproject.com/en/1.8/topics/i18n/
200 LANGUAGE_CODE = 'pt-br'
201 LANGUAGES = (
202 ('pt-br', u'Português'),
203 )
204
205 TIME_ZONE = 'America/Sao_Paulo'
206 USE_I18N = True
207 USE_L10N = False
208 USE_TZ = False
209 # DATE_FORMAT = 'N j, Y'
210 DATE_FORMAT = 'd/m/Y'
211 SHORT_DATE_FORMAT = 'd/m/Y'
212 DATETIME_FORMAT = 'd/m/Y H:i:s'
213 SHORT_DATETIME_FORMAT = 'd/m/Y H:i'
214 DATE_INPUT_FORMATS = ('%d/%m/%Y', '%m-%d-%Y', '%Y-%m-%d')
215
216 LOCALE_PATHS = (
217 'locale',
218 )
219
220 # Static files (CSS, JavaScript, Images)
221 # https://docs.djangoproject.com/en/1.8/howto/static-files/
222
223 STATIC_URL = '/static/'
224 STATIC_ROOT = PROJECT_DIR.child("collected_static")
225 STATICFILES_DIRS = (BASE_DIR.child("static"),)
226 STATICFILES_FINDERS = (
227 'django.contrib.staticfiles.finders.FileSystemFinder',
228 'django.contrib.staticfiles.finders.AppDirectoriesFinder',
229 'djangobower.finders.BowerFinder',
230 'sass_processor.finders.CssFinder',
231 )
232
233 MEDIA_ROOT = PROJECT_DIR.child("media")
234 MEDIA_URL = '/media/'
235
236 DAB_FIELD_RENDERER = \
237 'django_admin_bootstrapped.renderers.BootstrapFieldRenderer'
238 CRISPY_TEMPLATE_PACK = 'bootstrap3'
239 CRISPY_ALLOWED_TEMPLATE_PACKS = 'bootstrap3'
240 CRISPY_FAIL_SILENTLY = not DEBUG
241
242 BOWER_COMPONENTS_ROOT = PROJECT_DIR.child("bower")
243 BOWER_INSTALLED_APPS = (
244 'jquery#3.1.1',
245 'bootstrap-sass#3.3.7',
246 'components-font-awesome#4.5.0',
247 'tinymce#4.3.8',
248 'jquery-ui#1.12.1',
249 'jQuery-Mask-Plugin#1.14.0',
250 'jsdiff#2.2.2',
251 'https://github.com/interlegis/drunken-parrot-flat-ui.git',
252 'jquery-query-object#2.2.3',
253 )
254
255 # Additional search paths for SASS files when using the @import statement
256 SASS_PROCESSOR_INCLUDE_DIRS = (BOWER_COMPONENTS_ROOT.child(
257 'bower_components', 'bootstrap-sass', 'assets', 'stylesheets'),
258 )
259
260 # suprime texto de ajuda default do django-filter
261 FILTERS_HELP_TEXT_FILTER = False
262
263
264 # FIXME update cripy-forms and remove this
265 # hack to suppress many annoying warnings from crispy_forms
266 # see sapl.temp_suppress_crispy_form_warnings
267 LOGGING = SUPRESS_CRISPY_FORM_WARNINGS_LOGGING
268
269
270 LOGGING_CONSOLE = config('LOGGING_CONSOLE', default=False, cast=bool)
271 if DEBUG and LOGGING_CONSOLE:
272 # Descomentar linha abaixo fará com que logs aparecam, inclusive SQL
273 # LOGGING['handlers']['console']['level'] = 'DEBUG'
274 LOGGING['loggers']['django']['level'] = 'DEBUG'
275 LOGGING.update({
276 'formatters': {
277 'verbose': {
278 'format': '%(levelname)s %(asctime)s %(pathname)s '
279 '%(funcName)s %(message)s'
280 },
281 'simple': {
282 'format': '%(levelname)s %(message)s'
283 },
284 },
285 })
286 LOGGING['handlers']['console']['formatter'] = 'verbose'
287 LOGGING['loggers'][BASE_DIR.name] = {
288 'handlers': ['console'],
289 'level': 'DEBUG',
290 }
291
292
293 def excepthook(*args):
294 logging.getLogger(BASE_DIR.name).error(
295 'Uncaught exception:', exc_info=args)
296
297 # sys.excepthook = excepthook
298
[end of sapl/settings.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/sapl/settings.py b/sapl/settings.py
--- a/sapl/settings.py
+++ b/sapl/settings.py
@@ -192,7 +192,7 @@
DEFAULT_FROM_EMAIL = config('DEFAULT_FROM_EMAIL', cast=str, default='')
SERVER_EMAIL = config('SERVER_EMAIL', cast=str, default='')
-MAX_DOC_UPLOAD_SIZE = 5 * 1024 * 1024 # 5MB
+MAX_DOC_UPLOAD_SIZE = 10 * 1024 * 1024 # 10MB
MAX_IMAGE_UPLOAD_SIZE = 2 * 1024 * 1024 # 2MB
# Internationalization
|
{"golden_diff": "diff --git a/sapl/settings.py b/sapl/settings.py\n--- a/sapl/settings.py\n+++ b/sapl/settings.py\n@@ -192,7 +192,7 @@\n DEFAULT_FROM_EMAIL = config('DEFAULT_FROM_EMAIL', cast=str, default='')\n SERVER_EMAIL = config('SERVER_EMAIL', cast=str, default='')\n \n-MAX_DOC_UPLOAD_SIZE = 5 * 1024 * 1024 # 5MB\n+MAX_DOC_UPLOAD_SIZE = 10 * 1024 * 1024 # 10MB\n MAX_IMAGE_UPLOAD_SIZE = 2 * 1024 * 1024 # 2MB\n \n # Internationalization\n", "issue": "Aumenta o tamanho m\u00e1ximo de arquivos de upload para 10MB\nAtualmente o tamanho m\u00e1ximo de arquivos de upload \u00e9 de 5MB. Ao menos uma casa reclamou que este tamanho \u00e9 insuficiente ent\u00e3o vamos dobrar e ver se mais algu\u00e9m ainda reclama do tamanho de 10MB.\n", "before_files": [{"content": "\"\"\"\nDjango settings for sapl project.\n\nGenerated by 'django-admin startproject' using Django 1.8.2.\n\nFor more information on this file, see\nhttps://docs.djangoproject.com/en/1.8/topics/settings/\n\nFor the full list of settings and their values, see\nhttps://docs.djangoproject.com/en/1.8/ref/settings/\n\nQuick-start development settings - unsuitable for production\nSee https://docs.djangoproject.com/en/1.8/howto/deployment/checklist/\n\n\"\"\"\nimport logging\n\nfrom decouple import config\nfrom dj_database_url import parse as db_url\nfrom unipath import Path\n\nfrom .temp_suppress_crispy_form_warnings import \\\n SUPRESS_CRISPY_FORM_WARNINGS_LOGGING\n\nBASE_DIR = Path(__file__).ancestor(1)\nPROJECT_DIR = Path(__file__).ancestor(2)\n\n\n# SECURITY WARNING: keep the secret key used in production secret!\nSECRET_KEY = config('SECRET_KEY', default='')\n# SECURITY WARNING: don't run with debug turned on in production!\nDEBUG = config('DEBUG', default=False, cast=bool)\n\nMESSAGE_STORAGE = 'django.contrib.messages.storage.session.SessionStorage'\n\nALLOWED_HOSTS = ['*']\n\nLOGIN_REDIRECT_URL = '/'\nLOGIN_URL = '/login/?next='\n\nEMAIL_BACKEND = 'django.core.mail.backends.smtp.EmailBackend'\n\n\n# SAPL business apps in dependency order\nSAPL_APPS = (\n 'sapl.base',\n 'sapl.crud',\n 'sapl.parlamentares',\n 'sapl.comissoes',\n 'sapl.materia',\n 'sapl.norma',\n 'sapl.sessao',\n 'sapl.lexml',\n 'sapl.painel',\n 'sapl.protocoloadm',\n 'sapl.redireciona_urls',\n 'sapl.compilacao',\n 'sapl.api',\n\n 'sapl.rules'\n\n)\n\nINSTALLED_APPS = (\n 'django_admin_bootstrapped', # must come before django.contrib.admin\n 'django.contrib.admin',\n 'django.contrib.auth',\n 'django.contrib.contenttypes',\n 'django.contrib.sessions',\n 'django.contrib.messages',\n 'django.contrib.staticfiles',\n\n # more\n 'django_extensions',\n 'djangobower',\n 'bootstrap3', # basically for django_admin_bootstrapped\n 'crispy_forms',\n 'easy_thumbnails',\n 'floppyforms',\n 'haystack',\n 'sass_processor',\n 'rest_framework',\n 'reversion',\n 'whoosh',\n\n) + SAPL_APPS\n\n# FTS = Full Text Search\nSEARCH_BACKEND = 'haystack.backends.whoosh_backend.WhooshEngine'\nSEARCH_URL = ('PATH', PROJECT_DIR.child('whoosh'))\n\nSOLR_URL = config('SOLR_URL', cast=str, default='')\nif SOLR_URL:\n SEARCH_BACKEND = 'haystack.backends.solr_backend.SolrEngine'\n SEARCH_URL = ('URL', config('SOLR_URL', cast=str))\n # ...or for multicore...\n # 'URL': 'http://127.0.0.1:8983/solr/mysite',\n\n\nHAYSTACK_CONNECTIONS = {\n 'default': {\n 'ENGINE': SEARCH_BACKEND,\n SEARCH_URL[0] : SEARCH_URL[1]\n },\n}\n\n\nif DEBUG:\n INSTALLED_APPS += ('debug_toolbar', 'rest_framework_docs',)\n\nMIDDLEWARE_CLASSES = (\n 'reversion.middleware.RevisionMiddleware',\n 'django.contrib.sessions.middleware.SessionMiddleware',\n 'django.middleware.locale.LocaleMiddleware',\n 'django.middleware.common.CommonMiddleware',\n 'django.middleware.csrf.CsrfViewMiddleware',\n 'django.contrib.auth.middleware.AuthenticationMiddleware',\n 'django.contrib.auth.middleware.SessionAuthenticationMiddleware',\n 'django.contrib.messages.middleware.MessageMiddleware',\n 'django.middleware.clickjacking.XFrameOptionsMiddleware',\n 'django.middleware.security.SecurityMiddleware',\n)\n\n\nREST_FRAMEWORK = {\n \"UNICODE_JSON\": False,\n \"DEFAULT_PARSER_CLASSES\": (\n \"rest_framework.parsers.JSONParser\",\n ),\n \"DEFAULT_PERMISSION_CLASSES\": (\n \"rest_framework.permissions.IsAuthenticated\",\n \"sapl.api.permissions.DjangoModelPermissions\",\n ),\n \"DEFAULT_AUTHENTICATION_CLASSES\": (\n \"rest_framework.authentication.SessionAuthentication\",\n ),\n \"DEFAULT_PAGINATION_CLASS\": \"sapl.api.pagination.StandardPagination\",\n \"DEFAULT_FILTER_BACKENDS\": (\n \"rest_framework.filters.SearchFilter\",\n \"rest_framework.filters.DjangoFilterBackend\",\n ),\n}\n\n\nROOT_URLCONF = 'sapl.urls'\n\nTEMPLATES = [\n {\n 'BACKEND': 'django.template.backends.django.DjangoTemplates',\n 'DIRS': ['sapl/templates'],\n 'APP_DIRS': True,\n 'OPTIONS': {\n 'context_processors': [\n 'django.template.context_processors.debug',\n 'django.template.context_processors.request',\n 'django.contrib.auth.context_processors.auth',\n \"django.template.context_processors.media\",\n \"django.template.context_processors.static\",\n 'django.contrib.messages.context_processors.messages',\n 'sapl.context_processors.parliament_info',\n ],\n 'debug': DEBUG\n },\n },\n]\n\n\nWSGI_APPLICATION = 'sapl.wsgi.application'\n\n# Database\n# https://docs.djangoproject.com/en/1.8/ref/settings/#databases\n\nDATABASES = {\n 'default': config(\n 'DATABASE_URL', default='sqlite://:memory:',\n cast=db_url,\n )\n}\n\n# troque no caso de reimplementa\u00e7\u00e3o da classe User conforme\n# https://docs.djangoproject.com/en/1.9/topics/auth/customizing/#substituting-a-custom-user-model\nAUTH_USER_MODEL = 'auth.User'\n\nX_FRAME_OPTIONS = 'ALLOWALL'\n\nEMAIL_HOST = config('EMAIL_HOST', default='localhost')\nEMAIL_PORT = config('EMAIL_PORT', cast=int, default=587)\nEMAIL_HOST_USER = config('EMAIL_HOST_USER', default='')\nEMAIL_HOST_PASSWORD = config('EMAIL_HOST_PASSWORD', default='')\nEMAIL_USE_TLS = config('EMAIL_USE_TLS', cast=bool, default=True)\nEMAIL_SEND_USER = config('EMAIL_SEND_USER', cast=str, default='')\nDEFAULT_FROM_EMAIL = config('DEFAULT_FROM_EMAIL', cast=str, default='')\nSERVER_EMAIL = config('SERVER_EMAIL', cast=str, default='')\n\nMAX_DOC_UPLOAD_SIZE = 5 * 1024 * 1024 # 5MB\nMAX_IMAGE_UPLOAD_SIZE = 2 * 1024 * 1024 # 2MB\n\n# Internationalization\n# https://docs.djangoproject.com/en/1.8/topics/i18n/\nLANGUAGE_CODE = 'pt-br'\nLANGUAGES = (\n ('pt-br', u'Portugu\u00eas'),\n)\n\nTIME_ZONE = 'America/Sao_Paulo'\nUSE_I18N = True\nUSE_L10N = False\nUSE_TZ = False\n# DATE_FORMAT = 'N j, Y'\nDATE_FORMAT = 'd/m/Y'\nSHORT_DATE_FORMAT = 'd/m/Y'\nDATETIME_FORMAT = 'd/m/Y H:i:s'\nSHORT_DATETIME_FORMAT = 'd/m/Y H:i'\nDATE_INPUT_FORMATS = ('%d/%m/%Y', '%m-%d-%Y', '%Y-%m-%d')\n\nLOCALE_PATHS = (\n 'locale',\n)\n\n# Static files (CSS, JavaScript, Images)\n# https://docs.djangoproject.com/en/1.8/howto/static-files/\n\nSTATIC_URL = '/static/'\nSTATIC_ROOT = PROJECT_DIR.child(\"collected_static\")\nSTATICFILES_DIRS = (BASE_DIR.child(\"static\"),)\nSTATICFILES_FINDERS = (\n 'django.contrib.staticfiles.finders.FileSystemFinder',\n 'django.contrib.staticfiles.finders.AppDirectoriesFinder',\n 'djangobower.finders.BowerFinder',\n 'sass_processor.finders.CssFinder',\n)\n\nMEDIA_ROOT = PROJECT_DIR.child(\"media\")\nMEDIA_URL = '/media/'\n\nDAB_FIELD_RENDERER = \\\n 'django_admin_bootstrapped.renderers.BootstrapFieldRenderer'\nCRISPY_TEMPLATE_PACK = 'bootstrap3'\nCRISPY_ALLOWED_TEMPLATE_PACKS = 'bootstrap3'\nCRISPY_FAIL_SILENTLY = not DEBUG\n\nBOWER_COMPONENTS_ROOT = PROJECT_DIR.child(\"bower\")\nBOWER_INSTALLED_APPS = (\n 'jquery#3.1.1',\n 'bootstrap-sass#3.3.7',\n 'components-font-awesome#4.5.0',\n 'tinymce#4.3.8',\n 'jquery-ui#1.12.1',\n 'jQuery-Mask-Plugin#1.14.0',\n 'jsdiff#2.2.2',\n 'https://github.com/interlegis/drunken-parrot-flat-ui.git',\n 'jquery-query-object#2.2.3',\n)\n\n# Additional search paths for SASS files when using the @import statement\nSASS_PROCESSOR_INCLUDE_DIRS = (BOWER_COMPONENTS_ROOT.child(\n 'bower_components', 'bootstrap-sass', 'assets', 'stylesheets'),\n)\n\n# suprime texto de ajuda default do django-filter\nFILTERS_HELP_TEXT_FILTER = False\n\n\n# FIXME update cripy-forms and remove this\n# hack to suppress many annoying warnings from crispy_forms\n# see sapl.temp_suppress_crispy_form_warnings\nLOGGING = SUPRESS_CRISPY_FORM_WARNINGS_LOGGING\n\n\nLOGGING_CONSOLE = config('LOGGING_CONSOLE', default=False, cast=bool)\nif DEBUG and LOGGING_CONSOLE:\n # Descomentar linha abaixo far\u00e1 com que logs aparecam, inclusive SQL\n # LOGGING['handlers']['console']['level'] = 'DEBUG'\n LOGGING['loggers']['django']['level'] = 'DEBUG'\n LOGGING.update({\n 'formatters': {\n 'verbose': {\n 'format': '%(levelname)s %(asctime)s %(pathname)s '\n '%(funcName)s %(message)s'\n },\n 'simple': {\n 'format': '%(levelname)s %(message)s'\n },\n },\n })\n LOGGING['handlers']['console']['formatter'] = 'verbose'\n LOGGING['loggers'][BASE_DIR.name] = {\n 'handlers': ['console'],\n 'level': 'DEBUG',\n }\n\n\ndef excepthook(*args):\n logging.getLogger(BASE_DIR.name).error(\n 'Uncaught exception:', exc_info=args)\n\n# sys.excepthook = excepthook\n", "path": "sapl/settings.py"}]}
| 3,592 | 154 |
gh_patches_debug_15732
|
rasdani/github-patches
|
git_diff
|
praw-dev__praw-711
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
LiveContributorRelationship.__call__() returns a list of RedditorList if caller have 'manage' permission
## Issue Description
As the title said. To reproduce:
```
>>> import praw
>>> reddit = praw.Reddit(...)
>>> thread = reddit.live(...) # thread I created
>>> thread.contributor()
[<praw.models.list.redditor.RedditorList object at 0x101865d50>, <praw.models.list.redditor.RedditorList object at 0x100be6f90>]
```
JSON (copied from browser):
```
[
{
"kind": "UserList",
"data": {
"children": [
{
"permissions": [
"all"
],
"id": "t2_ll32z",
"name": "nmtake"
}
]
}
},
{
"kind": "UserList",
"data": {
"children": [ ]
}
}
]
```
It seems that latter UserList reprsents invited users who didn't accept invitation yet: ([source](https://github.com/reddit/reddit-plugin-liveupdate/blob/a7cde1cf917c4118c4cb114c73a92f272e439816/reddit_liveupdate/controllers.py#L515-L540))
## System Information
PRAW Version: 50516a
Python Version: 2.7.13
Operating System: OS X 10.11.6
</issue>
<code>
[start of praw/models/reddit/live.py]
1 """Provide the LiveThread class."""
2 from ...const import API_PATH
3 from ..listing.generator import ListingGenerator
4 from .base import RedditBase
5 from .redditor import Redditor
6
7
8 class LiveThread(RedditBase):
9 """An individual LiveThread object."""
10
11 STR_FIELD = 'id'
12
13 @property
14 def contributor(self):
15 """An instance of :class:`.LiveContributorRelationship`.
16
17 Usage:
18
19 .. code-block:: python
20
21 thread = reddit.live('ukaeu1ik4sw5')
22 for contributor in thread.contributor():
23 print(contributor)
24
25 """
26 if self._contributor is None:
27 self._contributor = LiveContributorRelationship(self)
28 return self._contributor
29
30 def __eq__(self, other):
31 """Return whether the other instance equals the current.
32
33 .. note:: This comparison is case sensitive.
34 """
35 if isinstance(other, str):
36 return other == str(self)
37 return (isinstance(other, self.__class__) and
38 str(self) == str(other))
39
40 def __getitem__(self, update_id):
41 """Return a lazy :class:`.LiveUpdate` instance.
42
43 .. warning:: At this time, accesing lazy attributes, whose value
44 have not loaded, raises ``AttributeError``.
45
46 :param update_id: A live update ID, e.g.,
47 ``'7827987a-c998-11e4-a0b9-22000b6a88d2'``.
48
49 Usage:
50
51 .. code-block:: python
52
53 thread = reddit.live('ukaeu1ik4sw5')
54 update = thread['7827987a-c998-11e4-a0b9-22000b6a88d2']
55 update.thread # LiveThread(id='ukaeu1ik4sw5')
56 update.id # '7827987a-c998-11e4-a0b9-22000b6a88d2'
57 update.author # raise ``AttributeError``
58 """
59 return LiveUpdate(self._reddit, self.id, update_id)
60
61 def __hash__(self):
62 """Return the hash of the current instance."""
63 return hash(self.__class__.__name__) ^ hash(str(self))
64
65 def __init__(self, reddit, id=None, # pylint: disable=redefined-builtin
66 _data=None):
67 """Initialize a lazy :class:`.LiveThread` instance.
68
69 :param reddit: An instance of :class:`.Reddit`.
70 :param id: A live thread ID, e.g., ``'ukaeu1ik4sw5'``
71 """
72 if bool(id) == bool(_data):
73 raise TypeError('Either `id` or `_data` must be provided.')
74 super(LiveThread, self).__init__(reddit, _data)
75 if id:
76 self.id = id # pylint: disable=invalid-name
77 self._contributor = None
78
79 def _info_path(self):
80 return API_PATH['liveabout'].format(id=self.id)
81
82 def updates(self, **generator_kwargs):
83 """Return a :class:`.ListingGenerator` yields :class:`.LiveUpdate` s.
84
85 :param generator_kwargs: keyword arguments passed to
86 :class:`.ListingGenerator` constructor.
87 :returns: A :class:`.ListingGenerator` object which yields
88 :class:`.LiveUpdate` object.
89 """
90 url = API_PATH['live_updates'].format(id=self.id)
91 for update in ListingGenerator(self._reddit, url,
92 **generator_kwargs):
93 update._thread = self
94 yield update
95
96
97 class LiveContributorRelationship(object):
98 """Provide methods to interact with live threads' contributors."""
99
100 def __call__(self):
101 """Return a :class:`.RedditorList` for live threads' contributors.
102
103 Usage:
104
105 .. code-block:: python
106
107 thread = reddit.live('ukaeu1ik4sw5')
108 for contributor in thread.contributor():
109 print(contributor)
110
111 """
112 url = API_PATH['live_contributors'].format(id=self.thread.id)
113 return self.thread._reddit.get(url)
114
115 def __init__(self, thread):
116 """Create a LiveContributorRelationship instance.
117
118 :param thread: An instance of :class:`.LiveThread`.
119
120 """
121 self.thread = thread
122
123
124 class LiveUpdate(RedditBase):
125 """An individual :class:`.LiveUpdate` object."""
126
127 STR_FIELD = 'id'
128
129 @property
130 def thread(self):
131 """Return :class:`.LiveThread` object the update object belongs to."""
132 return self._thread
133
134 def __init__(self, reddit, thread_id=None, update_id=None, _data=None):
135 """Initialize a lazy :class:`.LiveUpdate` instance.
136
137 Either ``thread_id`` and ``update_id``, or ``_data`` must be
138 provided.
139
140 .. warning:: At this time, accesing lazy attributes, whose value
141 have not loaded, raises ``AttributeError``.
142
143 :param reddit: An instance of :class:`.Reddit`.
144 :param thread_id: A live thread ID, e.g., ``'ukaeu1ik4sw5'``.
145 :param update_id: A live update ID, e.g.,
146 ``'7827987a-c998-11e4-a0b9-22000b6a88d2'``.
147
148 Usage:
149
150 .. code-block:: python
151
152 update = LiveUpdate(reddit, 'ukaeu1ik4sw5',
153 '7827987a-c998-11e4-a0b9-22000b6a88d2')
154 update.thread # LiveThread(id='ukaeu1ik4sw5')
155 update.id # '7827987a-c998-11e4-a0b9-22000b6a88d2'
156 update.author # raise ``AttributeError``
157 """
158 if _data is not None:
159 super(LiveUpdate, self).__init__(reddit, _data)
160 self._fetched = True
161 elif thread_id and update_id:
162 super(LiveUpdate, self).__init__(reddit, None)
163 self._thread = LiveThread(self._reddit, thread_id)
164 self.id = update_id # pylint: disable=invalid-name
165 self._fetched = True
166 else:
167 raise TypeError('Either `thread_id` and `update_id`, or '
168 '`_data` must be provided.')
169
170 def __setattr__(self, attribute, value):
171 """Objectify author."""
172 if attribute == 'author':
173 value = Redditor(self._reddit, name=value)
174 super(LiveUpdate, self).__setattr__(attribute, value)
175
[end of praw/models/reddit/live.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/praw/models/reddit/live.py b/praw/models/reddit/live.py
--- a/praw/models/reddit/live.py
+++ b/praw/models/reddit/live.py
@@ -1,6 +1,7 @@
"""Provide the LiveThread class."""
from ...const import API_PATH
from ..listing.generator import ListingGenerator
+from ..list.redditor import RedditorList
from .base import RedditBase
from .redditor import Redditor
@@ -110,7 +111,8 @@
"""
url = API_PATH['live_contributors'].format(id=self.thread.id)
- return self.thread._reddit.get(url)
+ temp = self.thread._reddit.get(url)
+ return temp if isinstance(temp, RedditorList) else temp[0]
def __init__(self, thread):
"""Create a LiveContributorRelationship instance.
|
{"golden_diff": "diff --git a/praw/models/reddit/live.py b/praw/models/reddit/live.py\n--- a/praw/models/reddit/live.py\n+++ b/praw/models/reddit/live.py\n@@ -1,6 +1,7 @@\n \"\"\"Provide the LiveThread class.\"\"\"\n from ...const import API_PATH\n from ..listing.generator import ListingGenerator\n+from ..list.redditor import RedditorList\n from .base import RedditBase\n from .redditor import Redditor\n \n@@ -110,7 +111,8 @@\n \n \"\"\"\n url = API_PATH['live_contributors'].format(id=self.thread.id)\n- return self.thread._reddit.get(url)\n+ temp = self.thread._reddit.get(url)\n+ return temp if isinstance(temp, RedditorList) else temp[0]\n \n def __init__(self, thread):\n \"\"\"Create a LiveContributorRelationship instance.\n", "issue": "LiveContributorRelationship.__call__() returns a list of RedditorList if caller have 'manage' permission\n## Issue Description\r\n\r\nAs the title said. To reproduce:\r\n\r\n```\r\n>>> import praw\r\n>>> reddit = praw.Reddit(...)\r\n>>> thread = reddit.live(...) # thread I created \r\n>>> thread.contributor()\r\n[<praw.models.list.redditor.RedditorList object at 0x101865d50>, <praw.models.list.redditor.RedditorList object at 0x100be6f90>]\r\n```\r\n\r\nJSON (copied from browser):\r\n\r\n```\r\n[\r\n {\r\n \"kind\": \"UserList\",\r\n \"data\": {\r\n \"children\": [\r\n {\r\n \"permissions\": [\r\n \"all\"\r\n ],\r\n \"id\": \"t2_ll32z\",\r\n \"name\": \"nmtake\"\r\n }\r\n ]\r\n }\r\n },\r\n {\r\n \"kind\": \"UserList\",\r\n \"data\": {\r\n \"children\": [ ]\r\n }\r\n }\r\n]\r\n```\r\n\r\nIt seems that latter UserList reprsents invited users who didn't accept invitation yet: ([source](https://github.com/reddit/reddit-plugin-liveupdate/blob/a7cde1cf917c4118c4cb114c73a92f272e439816/reddit_liveupdate/controllers.py#L515-L540))\r\n\r\n## System Information\r\n\r\n PRAW Version: 50516a\r\n Python Version: 2.7.13\r\n Operating System: OS X 10.11.6\n", "before_files": [{"content": "\"\"\"Provide the LiveThread class.\"\"\"\nfrom ...const import API_PATH\nfrom ..listing.generator import ListingGenerator\nfrom .base import RedditBase\nfrom .redditor import Redditor\n\n\nclass LiveThread(RedditBase):\n \"\"\"An individual LiveThread object.\"\"\"\n\n STR_FIELD = 'id'\n\n @property\n def contributor(self):\n \"\"\"An instance of :class:`.LiveContributorRelationship`.\n\n Usage:\n\n .. code-block:: python\n\n thread = reddit.live('ukaeu1ik4sw5')\n for contributor in thread.contributor():\n print(contributor)\n\n \"\"\"\n if self._contributor is None:\n self._contributor = LiveContributorRelationship(self)\n return self._contributor\n\n def __eq__(self, other):\n \"\"\"Return whether the other instance equals the current.\n\n .. note:: This comparison is case sensitive.\n \"\"\"\n if isinstance(other, str):\n return other == str(self)\n return (isinstance(other, self.__class__) and\n str(self) == str(other))\n\n def __getitem__(self, update_id):\n \"\"\"Return a lazy :class:`.LiveUpdate` instance.\n\n .. warning:: At this time, accesing lazy attributes, whose value\n have not loaded, raises ``AttributeError``.\n\n :param update_id: A live update ID, e.g.,\n ``'7827987a-c998-11e4-a0b9-22000b6a88d2'``.\n\n Usage:\n\n .. code-block:: python\n\n thread = reddit.live('ukaeu1ik4sw5')\n update = thread['7827987a-c998-11e4-a0b9-22000b6a88d2']\n update.thread # LiveThread(id='ukaeu1ik4sw5')\n update.id # '7827987a-c998-11e4-a0b9-22000b6a88d2'\n update.author # raise ``AttributeError``\n \"\"\"\n return LiveUpdate(self._reddit, self.id, update_id)\n\n def __hash__(self):\n \"\"\"Return the hash of the current instance.\"\"\"\n return hash(self.__class__.__name__) ^ hash(str(self))\n\n def __init__(self, reddit, id=None, # pylint: disable=redefined-builtin\n _data=None):\n \"\"\"Initialize a lazy :class:`.LiveThread` instance.\n\n :param reddit: An instance of :class:`.Reddit`.\n :param id: A live thread ID, e.g., ``'ukaeu1ik4sw5'``\n \"\"\"\n if bool(id) == bool(_data):\n raise TypeError('Either `id` or `_data` must be provided.')\n super(LiveThread, self).__init__(reddit, _data)\n if id:\n self.id = id # pylint: disable=invalid-name\n self._contributor = None\n\n def _info_path(self):\n return API_PATH['liveabout'].format(id=self.id)\n\n def updates(self, **generator_kwargs):\n \"\"\"Return a :class:`.ListingGenerator` yields :class:`.LiveUpdate` s.\n\n :param generator_kwargs: keyword arguments passed to\n :class:`.ListingGenerator` constructor.\n :returns: A :class:`.ListingGenerator` object which yields\n :class:`.LiveUpdate` object.\n \"\"\"\n url = API_PATH['live_updates'].format(id=self.id)\n for update in ListingGenerator(self._reddit, url,\n **generator_kwargs):\n update._thread = self\n yield update\n\n\nclass LiveContributorRelationship(object):\n \"\"\"Provide methods to interact with live threads' contributors.\"\"\"\n\n def __call__(self):\n \"\"\"Return a :class:`.RedditorList` for live threads' contributors.\n\n Usage:\n\n .. code-block:: python\n\n thread = reddit.live('ukaeu1ik4sw5')\n for contributor in thread.contributor():\n print(contributor)\n\n \"\"\"\n url = API_PATH['live_contributors'].format(id=self.thread.id)\n return self.thread._reddit.get(url)\n\n def __init__(self, thread):\n \"\"\"Create a LiveContributorRelationship instance.\n\n :param thread: An instance of :class:`.LiveThread`.\n\n \"\"\"\n self.thread = thread\n\n\nclass LiveUpdate(RedditBase):\n \"\"\"An individual :class:`.LiveUpdate` object.\"\"\"\n\n STR_FIELD = 'id'\n\n @property\n def thread(self):\n \"\"\"Return :class:`.LiveThread` object the update object belongs to.\"\"\"\n return self._thread\n\n def __init__(self, reddit, thread_id=None, update_id=None, _data=None):\n \"\"\"Initialize a lazy :class:`.LiveUpdate` instance.\n\n Either ``thread_id`` and ``update_id``, or ``_data`` must be\n provided.\n\n .. warning:: At this time, accesing lazy attributes, whose value\n have not loaded, raises ``AttributeError``.\n\n :param reddit: An instance of :class:`.Reddit`.\n :param thread_id: A live thread ID, e.g., ``'ukaeu1ik4sw5'``.\n :param update_id: A live update ID, e.g.,\n ``'7827987a-c998-11e4-a0b9-22000b6a88d2'``.\n\n Usage:\n\n .. code-block:: python\n\n update = LiveUpdate(reddit, 'ukaeu1ik4sw5',\n '7827987a-c998-11e4-a0b9-22000b6a88d2')\n update.thread # LiveThread(id='ukaeu1ik4sw5')\n update.id # '7827987a-c998-11e4-a0b9-22000b6a88d2'\n update.author # raise ``AttributeError``\n \"\"\"\n if _data is not None:\n super(LiveUpdate, self).__init__(reddit, _data)\n self._fetched = True\n elif thread_id and update_id:\n super(LiveUpdate, self).__init__(reddit, None)\n self._thread = LiveThread(self._reddit, thread_id)\n self.id = update_id # pylint: disable=invalid-name\n self._fetched = True\n else:\n raise TypeError('Either `thread_id` and `update_id`, or '\n '`_data` must be provided.')\n\n def __setattr__(self, attribute, value):\n \"\"\"Objectify author.\"\"\"\n if attribute == 'author':\n value = Redditor(self._reddit, name=value)\n super(LiveUpdate, self).__setattr__(attribute, value)\n", "path": "praw/models/reddit/live.py"}]}
| 2,840 | 190 |
gh_patches_debug_32742
|
rasdani/github-patches
|
git_diff
|
freedomofpress__securedrop-3379
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Debian package securedrop-app-code is not getting upgraded
# Bug
## Description
In my prod vm `securedrop-app-code` package is not getting upgraded.
## Steps to Reproduce
Follow the QA steps to fetch the latest packages from apt-test repo.
## Expected Behavior
We should get the 0.7~RC1 packages.
## Actual Behavior
```
$ dpkg -al | grep secure
ii openssh-client 1:6.6p1-2ubuntu2.10 amd64 secure shell (SSH) client, for secure access to remote machines
ii openssh-server 1:6.6p1-2ubuntu2.10 amd64 secure shell (SSH) server, for secure access from remote machines
ii openssh-sftp-server 1:6.6p1-2ubuntu2.10 amd64 secure shell (SSH) sftp server module, for SFTP access from remote machines
ii secure-delete 3.1-6 amd64 tools to wipe files, free disk space, swap and memory
ii securedrop-app-code 0.6 amd64 Packages the SecureDrop application code pip dependencies and apparmor profiles. This package will put the apparmor profiles in enforce mode. This package does use pip to install the pip wheelhouse
ii securedrop-config 0.1.1+0.7.0~rc1 all Establishes baseline system state for running SecureDrop.
ii securedrop-grsec 4.4.115+r1 amd64 Metapackage providing a grsecurity-patched Linux kernel for use
ii securedrop-keyring 0.1.1+0.7.0~rc1 amd64 Provides an apt keyring for SecureDrop-related packages, so the master signing key used for SecureDrop packages can be updated via apt.
ii securedrop-ossec-agent 2.8.2+0.7.0~rc1 amd64 Installs the securedrop pre-configured OSSEC agent
ii ssh-import-id 3.21-0ubuntu1 all securely retrieve an SSH public key and install it locally
```
```
$ sudo apt-get upgrade
Reading package lists... Done
Building dependency tree
Reading state information... Done
Calculating upgrade... Done
The following packages have been kept back:
securedrop-app-code
0 upgraded, 0 newly installed, 0 to remove and 1 not upgraded.
```
</issue>
<code>
[start of securedrop/journalist_app/forms.py]
1 # -*- coding: utf-8 -*-
2
3 from flask_babel import lazy_gettext as gettext
4 from flask_wtf import FlaskForm
5 from flask_wtf.file import FileField, FileAllowed, FileRequired
6 from wtforms import (TextAreaField, TextField, BooleanField, HiddenField,
7 ValidationError)
8 from wtforms.validators import InputRequired, Optional
9
10 from models import Journalist
11
12
13 def otp_secret_validation(form, field):
14 strip_whitespace = field.data.replace(' ', '')
15 if len(strip_whitespace) != 40:
16 raise ValidationError(gettext(
17 'HOTP secrets are 40 characters long - '
18 'you have entered {num_chars}.'.format(
19 num_chars=len(strip_whitespace)
20 )))
21
22
23 def minimum_length_validation(form, field):
24 if len(field.data) < Journalist.MIN_USERNAME_LEN:
25 raise ValidationError(
26 gettext('Field must be at least {min_chars} '
27 'characters long but only got '
28 '{num_chars}.'.format(
29 min_chars=Journalist.MIN_USERNAME_LEN,
30 num_chars=len(field.data))))
31
32
33 class NewUserForm(FlaskForm):
34 username = TextField('username', validators=[
35 InputRequired(message=gettext('This field is required.')),
36 minimum_length_validation
37 ])
38 password = HiddenField('password')
39 is_admin = BooleanField('is_admin')
40 is_hotp = BooleanField('is_hotp')
41 otp_secret = TextField('otp_secret', validators=[
42 otp_secret_validation,
43 Optional()
44 ])
45
46
47 class ReplyForm(FlaskForm):
48 message = TextAreaField(
49 u'Message',
50 id="content-area",
51 validators=[
52 InputRequired(message=gettext(
53 'You cannot send an empty reply.')),
54 ],
55 )
56
57
58 class LogoForm(FlaskForm):
59 logo = FileField(validators=[
60 FileRequired(message=gettext('File required.')),
61 FileAllowed(['jpg', 'png', 'jpeg'],
62 message=gettext("You can only upload JPG/JPEG"
63 " or PNG image files."))
64 ])
65
[end of securedrop/journalist_app/forms.py]
[start of securedrop/journalist_app/admin.py]
1 # -*- coding: utf-8 -*-
2
3 from PIL import Image
4
5 import os
6
7 from flask import (Blueprint, render_template, request, url_for, redirect, g,
8 current_app, flash, abort)
9 from flask_babel import gettext
10 from sqlalchemy.exc import IntegrityError
11 from sqlalchemy.orm.exc import NoResultFound
12
13 from db import db
14 from models import Journalist, InvalidUsernameException, PasswordError
15 from journalist_app.decorators import admin_required
16 from journalist_app.utils import (make_password, commit_account_changes,
17 set_diceware_password, validate_hotp_secret)
18 from journalist_app.forms import LogoForm, NewUserForm
19
20
21 def make_blueprint(config):
22 view = Blueprint('admin', __name__)
23
24 @view.route('/', methods=('GET', 'POST'))
25 @admin_required
26 def index():
27 users = Journalist.query.all()
28 return render_template("admin.html", users=users)
29
30 @view.route('/config', methods=('GET', 'POST'))
31 @admin_required
32 def manage_config():
33 form = LogoForm()
34 if form.validate_on_submit():
35 f = form.logo.data
36 custom_logo_filepath = os.path.join(config.SECUREDROP_ROOT,
37 "static/i/custom_logo.png")
38 try:
39 with Image.open(f) as im:
40 im.thumbnail((500, 450), resample=3)
41 im.save(custom_logo_filepath, "PNG")
42 flash(gettext("Image updated."), "logo-success")
43 except Exception:
44 flash("Unable to process the image file."
45 " Try another one.", "logo-error")
46 finally:
47 return redirect(url_for("admin.manage_config"))
48
49 else:
50 for field, errors in form.errors.items():
51 for error in errors:
52 flash(error, "logo-error")
53 return render_template("config.html", form=form)
54
55 @view.route('/add', methods=('GET', 'POST'))
56 @admin_required
57 def add_user():
58 form = NewUserForm()
59 if form.validate_on_submit():
60 form_valid = True
61 username = request.form['username']
62 password = request.form['password']
63 is_admin = bool(request.form.get('is_admin'))
64
65 try:
66 otp_secret = None
67 if request.form.get('is_hotp', False):
68 otp_secret = request.form.get('otp_secret', '')
69 new_user = Journalist(username=username,
70 password=password,
71 is_admin=is_admin,
72 otp_secret=otp_secret)
73 db.session.add(new_user)
74 db.session.commit()
75 except PasswordError:
76 flash(gettext(
77 'There was an error with the autogenerated password. '
78 'User not created. Please try again.'), 'error')
79 form_valid = False
80 except InvalidUsernameException as e:
81 form_valid = False
82 flash('Invalid username: ' + str(e), "error")
83 except IntegrityError as e:
84 db.session.rollback()
85 form_valid = False
86 if "UNIQUE constraint failed: journalists.username" in str(e):
87 flash(gettext('Username "{user}" already taken.'.format(
88 user=username)), "error")
89 else:
90 flash(gettext("An error occurred saving this user"
91 " to the database."
92 " Please inform your administrator."),
93 "error")
94 current_app.logger.error("Adding user "
95 "'{}' failed: {}".format(
96 username, e))
97
98 if form_valid:
99 return redirect(url_for('admin.new_user_two_factor',
100 uid=new_user.id))
101
102 return render_template("admin_add_user.html",
103 password=make_password(config),
104 form=form)
105
106 @view.route('/2fa', methods=('GET', 'POST'))
107 @admin_required
108 def new_user_two_factor():
109 user = Journalist.query.get(request.args['uid'])
110
111 if request.method == 'POST':
112 token = request.form['token']
113 if user.verify_token(token):
114 flash(gettext(
115 "Token in two-factor authentication "
116 "accepted for user {user}.").format(
117 user=user.username),
118 "notification")
119 return redirect(url_for("admin.index"))
120 else:
121 flash(gettext(
122 "Could not verify token in two-factor authentication."),
123 "error")
124
125 return render_template("admin_new_user_two_factor.html", user=user)
126
127 @view.route('/reset-2fa-totp', methods=['POST'])
128 @admin_required
129 def reset_two_factor_totp():
130 uid = request.form['uid']
131 user = Journalist.query.get(uid)
132 user.is_totp = True
133 user.regenerate_totp_shared_secret()
134 db.session.commit()
135 return redirect(url_for('admin.new_user_two_factor', uid=uid))
136
137 @view.route('/reset-2fa-hotp', methods=['POST'])
138 @admin_required
139 def reset_two_factor_hotp():
140 uid = request.form['uid']
141 otp_secret = request.form.get('otp_secret', None)
142 if otp_secret:
143 user = Journalist.query.get(uid)
144 if not validate_hotp_secret(user, otp_secret):
145 return render_template('admin_edit_hotp_secret.html', uid=uid)
146 db.session.commit()
147 return redirect(url_for('admin.new_user_two_factor', uid=uid))
148 else:
149 return render_template('admin_edit_hotp_secret.html', uid=uid)
150
151 @view.route('/edit/<int:user_id>', methods=('GET', 'POST'))
152 @admin_required
153 def edit_user(user_id):
154 user = Journalist.query.get(user_id)
155
156 if request.method == 'POST':
157 if request.form.get('username', None):
158 new_username = request.form['username']
159
160 try:
161 Journalist.check_username_acceptable(new_username)
162 except InvalidUsernameException as e:
163 flash('Invalid username: ' + str(e), 'error')
164 return redirect(url_for("admin.edit_user",
165 user_id=user_id))
166
167 if new_username == user.username:
168 pass
169 elif Journalist.query.filter_by(
170 username=new_username).one_or_none():
171 flash(gettext(
172 'Username "{user}" already taken.').format(
173 user=new_username),
174 "error")
175 return redirect(url_for("admin.edit_user",
176 user_id=user_id))
177 else:
178 user.username = new_username
179
180 user.is_admin = bool(request.form.get('is_admin'))
181
182 commit_account_changes(user)
183
184 password = make_password(config)
185 return render_template("edit_account.html", user=user,
186 password=password)
187
188 @view.route('/edit/<int:user_id>/new-password', methods=('POST',))
189 @admin_required
190 def set_password(user_id):
191 try:
192 user = Journalist.query.get(user_id)
193 except NoResultFound:
194 abort(404)
195
196 password = request.form.get('password')
197 set_diceware_password(user, password)
198 return redirect(url_for('admin.edit_user', user_id=user_id))
199
200 @view.route('/delete/<int:user_id>', methods=('POST',))
201 @admin_required
202 def delete_user(user_id):
203 user = Journalist.query.get(user_id)
204 if user_id == g.user.id:
205 # Do not flash because the interface already has safe guards.
206 # It can only happen by manually crafting a POST request
207 current_app.logger.error(
208 "Admin {} tried to delete itself".format(g.user.username))
209 abort(403)
210 elif user:
211 db.session.delete(user)
212 db.session.commit()
213 flash(gettext("Deleted user '{user}'").format(
214 user=user.username), "notification")
215 else:
216 current_app.logger.error(
217 "Admin {} tried to delete nonexistent user with pk={}".format(
218 g.user.username, user_id))
219 abort(404)
220
221 return redirect(url_for('admin.index'))
222
223 @view.route('/edit/<int:user_id>/new-password', methods=('POST',))
224 @admin_required
225 def new_password(user_id):
226 try:
227 user = Journalist.query.get(user_id)
228 except NoResultFound:
229 abort(404)
230
231 password = request.form.get('password')
232 set_diceware_password(user, password)
233 return redirect(url_for('admin.edit_user', user_id=user_id))
234
235 @view.route('/ossec-test')
236 @admin_required
237 def ossec_test():
238 current_app.logger.error('This is a test OSSEC alert')
239 flash(gettext('Test alert sent. Please check your email.'),
240 'notification')
241 return redirect(url_for('admin.manage_config'))
242
243 return view
244
[end of securedrop/journalist_app/admin.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/securedrop/journalist_app/admin.py b/securedrop/journalist_app/admin.py
--- a/securedrop/journalist_app/admin.py
+++ b/securedrop/journalist_app/admin.py
@@ -1,7 +1,5 @@
# -*- coding: utf-8 -*-
-from PIL import Image
-
import os
from flask import (Blueprint, render_template, request, url_for, redirect, g,
@@ -33,19 +31,16 @@
form = LogoForm()
if form.validate_on_submit():
f = form.logo.data
- custom_logo_filepath = os.path.join(config.SECUREDROP_ROOT,
- "static/i/custom_logo.png")
+ custom_logo_filepath = os.path.join(current_app.static_folder, 'i',
+ 'custom_logo.png')
try:
- with Image.open(f) as im:
- im.thumbnail((500, 450), resample=3)
- im.save(custom_logo_filepath, "PNG")
+ f.save(custom_logo_filepath)
flash(gettext("Image updated."), "logo-success")
except Exception:
flash("Unable to process the image file."
" Try another one.", "logo-error")
finally:
return redirect(url_for("admin.manage_config"))
-
else:
for field, errors in form.errors.items():
for error in errors:
diff --git a/securedrop/journalist_app/forms.py b/securedrop/journalist_app/forms.py
--- a/securedrop/journalist_app/forms.py
+++ b/securedrop/journalist_app/forms.py
@@ -58,7 +58,6 @@
class LogoForm(FlaskForm):
logo = FileField(validators=[
FileRequired(message=gettext('File required.')),
- FileAllowed(['jpg', 'png', 'jpeg'],
- message=gettext("You can only upload JPG/JPEG"
- " or PNG image files."))
+ FileAllowed(['png'],
+ message=gettext("Upload images only."))
])
|
{"golden_diff": "diff --git a/securedrop/journalist_app/admin.py b/securedrop/journalist_app/admin.py\n--- a/securedrop/journalist_app/admin.py\n+++ b/securedrop/journalist_app/admin.py\n@@ -1,7 +1,5 @@\n # -*- coding: utf-8 -*-\n \n-from PIL import Image\n-\n import os\n \n from flask import (Blueprint, render_template, request, url_for, redirect, g,\n@@ -33,19 +31,16 @@\n form = LogoForm()\n if form.validate_on_submit():\n f = form.logo.data\n- custom_logo_filepath = os.path.join(config.SECUREDROP_ROOT,\n- \"static/i/custom_logo.png\")\n+ custom_logo_filepath = os.path.join(current_app.static_folder, 'i',\n+ 'custom_logo.png')\n try:\n- with Image.open(f) as im:\n- im.thumbnail((500, 450), resample=3)\n- im.save(custom_logo_filepath, \"PNG\")\n+ f.save(custom_logo_filepath)\n flash(gettext(\"Image updated.\"), \"logo-success\")\n except Exception:\n flash(\"Unable to process the image file.\"\n \" Try another one.\", \"logo-error\")\n finally:\n return redirect(url_for(\"admin.manage_config\"))\n-\n else:\n for field, errors in form.errors.items():\n for error in errors:\ndiff --git a/securedrop/journalist_app/forms.py b/securedrop/journalist_app/forms.py\n--- a/securedrop/journalist_app/forms.py\n+++ b/securedrop/journalist_app/forms.py\n@@ -58,7 +58,6 @@\n class LogoForm(FlaskForm):\n logo = FileField(validators=[\n FileRequired(message=gettext('File required.')),\n- FileAllowed(['jpg', 'png', 'jpeg'],\n- message=gettext(\"You can only upload JPG/JPEG\"\n- \" or PNG image files.\"))\n+ FileAllowed(['png'],\n+ message=gettext(\"Upload images only.\"))\n ])\n", "issue": "Debian package securedrop-app-code is not getting upgraded\n# Bug\r\n\r\n## Description\r\n\r\nIn my prod vm `securedrop-app-code` package is not getting upgraded.\r\n\r\n## Steps to Reproduce\r\n\r\nFollow the QA steps to fetch the latest packages from apt-test repo.\r\n\r\n## Expected Behavior\r\n\r\nWe should get the 0.7~RC1 packages.\r\n\r\n## Actual Behavior\r\n\r\n```\r\n$ dpkg -al | grep secure\r\nii openssh-client 1:6.6p1-2ubuntu2.10 amd64 secure shell (SSH) client, for secure access to remote machines\r\nii openssh-server 1:6.6p1-2ubuntu2.10 amd64 secure shell (SSH) server, for secure access from remote machines\r\nii openssh-sftp-server 1:6.6p1-2ubuntu2.10 amd64 secure shell (SSH) sftp server module, for SFTP access from remote machines\r\nii secure-delete 3.1-6 amd64 tools to wipe files, free disk space, swap and memory\r\nii securedrop-app-code 0.6 amd64 Packages the SecureDrop application code pip dependencies and apparmor profiles. This package will put the apparmor profiles in enforce mode. This package does use pip to install the pip wheelhouse\r\nii securedrop-config 0.1.1+0.7.0~rc1 all Establishes baseline system state for running SecureDrop.\r\nii securedrop-grsec 4.4.115+r1 amd64 Metapackage providing a grsecurity-patched Linux kernel for use\r\nii securedrop-keyring 0.1.1+0.7.0~rc1 amd64 Provides an apt keyring for SecureDrop-related packages, so the master signing key used for SecureDrop packages can be updated via apt.\r\nii securedrop-ossec-agent 2.8.2+0.7.0~rc1 amd64 Installs the securedrop pre-configured OSSEC agent\r\nii ssh-import-id 3.21-0ubuntu1 all securely retrieve an SSH public key and install it locally\r\n```\r\n\r\n```\r\n$ sudo apt-get upgrade\r\nReading package lists... Done\r\nBuilding dependency tree \r\nReading state information... Done\r\nCalculating upgrade... Done\r\nThe following packages have been kept back:\r\n securedrop-app-code\r\n0 upgraded, 0 newly installed, 0 to remove and 1 not upgraded.\r\n\r\n```\r\n\n", "before_files": [{"content": "# -*- coding: utf-8 -*-\n\nfrom flask_babel import lazy_gettext as gettext\nfrom flask_wtf import FlaskForm\nfrom flask_wtf.file import FileField, FileAllowed, FileRequired\nfrom wtforms import (TextAreaField, TextField, BooleanField, HiddenField,\n ValidationError)\nfrom wtforms.validators import InputRequired, Optional\n\nfrom models import Journalist\n\n\ndef otp_secret_validation(form, field):\n strip_whitespace = field.data.replace(' ', '')\n if len(strip_whitespace) != 40:\n raise ValidationError(gettext(\n 'HOTP secrets are 40 characters long - '\n 'you have entered {num_chars}.'.format(\n num_chars=len(strip_whitespace)\n )))\n\n\ndef minimum_length_validation(form, field):\n if len(field.data) < Journalist.MIN_USERNAME_LEN:\n raise ValidationError(\n gettext('Field must be at least {min_chars} '\n 'characters long but only got '\n '{num_chars}.'.format(\n min_chars=Journalist.MIN_USERNAME_LEN,\n num_chars=len(field.data))))\n\n\nclass NewUserForm(FlaskForm):\n username = TextField('username', validators=[\n InputRequired(message=gettext('This field is required.')),\n minimum_length_validation\n ])\n password = HiddenField('password')\n is_admin = BooleanField('is_admin')\n is_hotp = BooleanField('is_hotp')\n otp_secret = TextField('otp_secret', validators=[\n otp_secret_validation,\n Optional()\n ])\n\n\nclass ReplyForm(FlaskForm):\n message = TextAreaField(\n u'Message',\n id=\"content-area\",\n validators=[\n InputRequired(message=gettext(\n 'You cannot send an empty reply.')),\n ],\n )\n\n\nclass LogoForm(FlaskForm):\n logo = FileField(validators=[\n FileRequired(message=gettext('File required.')),\n FileAllowed(['jpg', 'png', 'jpeg'],\n message=gettext(\"You can only upload JPG/JPEG\"\n \" or PNG image files.\"))\n ])\n", "path": "securedrop/journalist_app/forms.py"}, {"content": "# -*- coding: utf-8 -*-\n\nfrom PIL import Image\n\nimport os\n\nfrom flask import (Blueprint, render_template, request, url_for, redirect, g,\n current_app, flash, abort)\nfrom flask_babel import gettext\nfrom sqlalchemy.exc import IntegrityError\nfrom sqlalchemy.orm.exc import NoResultFound\n\nfrom db import db\nfrom models import Journalist, InvalidUsernameException, PasswordError\nfrom journalist_app.decorators import admin_required\nfrom journalist_app.utils import (make_password, commit_account_changes,\n set_diceware_password, validate_hotp_secret)\nfrom journalist_app.forms import LogoForm, NewUserForm\n\n\ndef make_blueprint(config):\n view = Blueprint('admin', __name__)\n\n @view.route('/', methods=('GET', 'POST'))\n @admin_required\n def index():\n users = Journalist.query.all()\n return render_template(\"admin.html\", users=users)\n\n @view.route('/config', methods=('GET', 'POST'))\n @admin_required\n def manage_config():\n form = LogoForm()\n if form.validate_on_submit():\n f = form.logo.data\n custom_logo_filepath = os.path.join(config.SECUREDROP_ROOT,\n \"static/i/custom_logo.png\")\n try:\n with Image.open(f) as im:\n im.thumbnail((500, 450), resample=3)\n im.save(custom_logo_filepath, \"PNG\")\n flash(gettext(\"Image updated.\"), \"logo-success\")\n except Exception:\n flash(\"Unable to process the image file.\"\n \" Try another one.\", \"logo-error\")\n finally:\n return redirect(url_for(\"admin.manage_config\"))\n\n else:\n for field, errors in form.errors.items():\n for error in errors:\n flash(error, \"logo-error\")\n return render_template(\"config.html\", form=form)\n\n @view.route('/add', methods=('GET', 'POST'))\n @admin_required\n def add_user():\n form = NewUserForm()\n if form.validate_on_submit():\n form_valid = True\n username = request.form['username']\n password = request.form['password']\n is_admin = bool(request.form.get('is_admin'))\n\n try:\n otp_secret = None\n if request.form.get('is_hotp', False):\n otp_secret = request.form.get('otp_secret', '')\n new_user = Journalist(username=username,\n password=password,\n is_admin=is_admin,\n otp_secret=otp_secret)\n db.session.add(new_user)\n db.session.commit()\n except PasswordError:\n flash(gettext(\n 'There was an error with the autogenerated password. '\n 'User not created. Please try again.'), 'error')\n form_valid = False\n except InvalidUsernameException as e:\n form_valid = False\n flash('Invalid username: ' + str(e), \"error\")\n except IntegrityError as e:\n db.session.rollback()\n form_valid = False\n if \"UNIQUE constraint failed: journalists.username\" in str(e):\n flash(gettext('Username \"{user}\" already taken.'.format(\n user=username)), \"error\")\n else:\n flash(gettext(\"An error occurred saving this user\"\n \" to the database.\"\n \" Please inform your administrator.\"),\n \"error\")\n current_app.logger.error(\"Adding user \"\n \"'{}' failed: {}\".format(\n username, e))\n\n if form_valid:\n return redirect(url_for('admin.new_user_two_factor',\n uid=new_user.id))\n\n return render_template(\"admin_add_user.html\",\n password=make_password(config),\n form=form)\n\n @view.route('/2fa', methods=('GET', 'POST'))\n @admin_required\n def new_user_two_factor():\n user = Journalist.query.get(request.args['uid'])\n\n if request.method == 'POST':\n token = request.form['token']\n if user.verify_token(token):\n flash(gettext(\n \"Token in two-factor authentication \"\n \"accepted for user {user}.\").format(\n user=user.username),\n \"notification\")\n return redirect(url_for(\"admin.index\"))\n else:\n flash(gettext(\n \"Could not verify token in two-factor authentication.\"),\n \"error\")\n\n return render_template(\"admin_new_user_two_factor.html\", user=user)\n\n @view.route('/reset-2fa-totp', methods=['POST'])\n @admin_required\n def reset_two_factor_totp():\n uid = request.form['uid']\n user = Journalist.query.get(uid)\n user.is_totp = True\n user.regenerate_totp_shared_secret()\n db.session.commit()\n return redirect(url_for('admin.new_user_two_factor', uid=uid))\n\n @view.route('/reset-2fa-hotp', methods=['POST'])\n @admin_required\n def reset_two_factor_hotp():\n uid = request.form['uid']\n otp_secret = request.form.get('otp_secret', None)\n if otp_secret:\n user = Journalist.query.get(uid)\n if not validate_hotp_secret(user, otp_secret):\n return render_template('admin_edit_hotp_secret.html', uid=uid)\n db.session.commit()\n return redirect(url_for('admin.new_user_two_factor', uid=uid))\n else:\n return render_template('admin_edit_hotp_secret.html', uid=uid)\n\n @view.route('/edit/<int:user_id>', methods=('GET', 'POST'))\n @admin_required\n def edit_user(user_id):\n user = Journalist.query.get(user_id)\n\n if request.method == 'POST':\n if request.form.get('username', None):\n new_username = request.form['username']\n\n try:\n Journalist.check_username_acceptable(new_username)\n except InvalidUsernameException as e:\n flash('Invalid username: ' + str(e), 'error')\n return redirect(url_for(\"admin.edit_user\",\n user_id=user_id))\n\n if new_username == user.username:\n pass\n elif Journalist.query.filter_by(\n username=new_username).one_or_none():\n flash(gettext(\n 'Username \"{user}\" already taken.').format(\n user=new_username),\n \"error\")\n return redirect(url_for(\"admin.edit_user\",\n user_id=user_id))\n else:\n user.username = new_username\n\n user.is_admin = bool(request.form.get('is_admin'))\n\n commit_account_changes(user)\n\n password = make_password(config)\n return render_template(\"edit_account.html\", user=user,\n password=password)\n\n @view.route('/edit/<int:user_id>/new-password', methods=('POST',))\n @admin_required\n def set_password(user_id):\n try:\n user = Journalist.query.get(user_id)\n except NoResultFound:\n abort(404)\n\n password = request.form.get('password')\n set_diceware_password(user, password)\n return redirect(url_for('admin.edit_user', user_id=user_id))\n\n @view.route('/delete/<int:user_id>', methods=('POST',))\n @admin_required\n def delete_user(user_id):\n user = Journalist.query.get(user_id)\n if user_id == g.user.id:\n # Do not flash because the interface already has safe guards.\n # It can only happen by manually crafting a POST request\n current_app.logger.error(\n \"Admin {} tried to delete itself\".format(g.user.username))\n abort(403)\n elif user:\n db.session.delete(user)\n db.session.commit()\n flash(gettext(\"Deleted user '{user}'\").format(\n user=user.username), \"notification\")\n else:\n current_app.logger.error(\n \"Admin {} tried to delete nonexistent user with pk={}\".format(\n g.user.username, user_id))\n abort(404)\n\n return redirect(url_for('admin.index'))\n\n @view.route('/edit/<int:user_id>/new-password', methods=('POST',))\n @admin_required\n def new_password(user_id):\n try:\n user = Journalist.query.get(user_id)\n except NoResultFound:\n abort(404)\n\n password = request.form.get('password')\n set_diceware_password(user, password)\n return redirect(url_for('admin.edit_user', user_id=user_id))\n\n @view.route('/ossec-test')\n @admin_required\n def ossec_test():\n current_app.logger.error('This is a test OSSEC alert')\n flash(gettext('Test alert sent. Please check your email.'),\n 'notification')\n return redirect(url_for('admin.manage_config'))\n\n return view\n", "path": "securedrop/journalist_app/admin.py"}]}
| 4,094 | 439 |
gh_patches_debug_16475
|
rasdani/github-patches
|
git_diff
|
pyinstaller__pyinstaller-8544
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Could not import `pywintypes` or `win32api` from `win32ctypes.pywin32`
## Description of the issue
Error when running the executable.
Issue is present in 6.6.0 and "latest development version". Issue is not present in 6.5.0.
Output differs between versions at this point:
6.6.0:
```
import 'win32ctypes.core' # <_frozen_importlib_external.SourcelessFileLoader object at 0x000001C6D9E41BB0>
# win32ctypes.core._common not found in PYZ
# win32ctypes.core.ctypes not found in PYZ
# destroy win32ctypes.pywin32.win32api
# destroy win32ctypes.pywin32
# destroy PyInstaller
Could not import `pywintypes` or `win32api` from `win32ctypes.pywin32`.
Please make sure that `pywin32-ctypes` is installed and importable, for example:
pip install pywin32-ctypes
```
6.5.0:
```
# cffi not found in PYZ
# code object from '[...]\\cffi\\__init__.pyc'
# cffi.api not found in PYZ
# code object from '[...]\\cffi\\api.pyc'
# cffi.lock not found in PYZ
# code object from '[...]\\cffi\\lock.pyc'
import 'cffi.lock' # <_frozen_importlib_external.SourcelessFileLoader object at 0x000001F4AEB845C0>
# cffi.error not found in PYZ
# code object from '[...]\\cffi\\error.pyc'
import 'cffi.error' # <_frozen_importlib_external.SourcelessFileLoader object at 0x000001F4AEB846B0>
# cffi.model not found in PYZ
# code object from '[...]\\cffi\\model.pyc'
import 'cffi.model' # <_frozen_importlib_external.SourcelessFileLoader object at 0x000001F4AEB848F0>
import 'cffi.api' # <_frozen_importlib_external.SourcelessFileLoader object at 0x000001F4AEB52330>
import 'cffi' # <_frozen_importlib_external.SourcelessFileLoader object at 0x000001F4AEB51EB0>
import 'win32ctypes.core' # <_frozen_importlib_external.SourcelessFileLoader object at 0x000001F4AEB51B80>
# win32ctypes.core._common not found in PYZ
# win32ctypes.core.cffi not found in PYZ
# code object from '[...]\\win32ctypes\\core\\cffi\\__init__.pyc'
import 'win32ctypes.core.cffi' # <_frozen_importlib_external.SourcelessFileLoader object at 0x000001F4AEB86BA0>
# win32ctypes.core.cffi._common not found in PYZ
# code object from '[...]\\win32ctypes\\core\\cffi\\_common.pyc'
# win32ctypes.core.cffi._util not found in PYZ
# code object from '[...]\\win32ctypes\\core\\cffi\\_util.pyc'
# win32ctypes.core.compat not found in PYZ
# code object from '[...]\\win32ctypes\\core\\compat.pyc'
import 'win32ctypes.core.compat' # <_frozen_importlib_external.SourcelessFileLoader object at 0x000001F4AEB87440>
# _cffi_backend not found in PYZ
# extension module '_cffi_backend' loaded from '[...]\\_cffi_backend.cp312-win_amd64.pyd'
# extension module '_cffi_backend' executed from '[...]\\_cffi_backend.cp312-win_amd64.pyd'
import '_cffi_backend' # <_frozen_importlib_external.ExtensionFileLoader object at 0x000001F4AEB876B0>
# cffi.cparser not found in PYZ
# code object from '[...]\\cffi\\cparser.pyc'
# cffi.commontypes not found in PYZ
# code object from '[...]\\cffi\\commontypes.pyc'
import 'cffi.commontypes' # <_frozen_importlib_external.SourcelessFileLoader object at 0x000001F4AEBA91F0>
```
### Context information (for bug reports)
* 502 INFO: PyInstaller: 6.6.0, contrib hooks: 2024.6
* 503 INFO: Python: 3.12.0
* 541 INFO: Platform: Windows-10-10.0.19045-SP0
</issue>
<code>
[start of PyInstaller/hooks/hook-win32ctypes.core.py]
1 #-----------------------------------------------------------------------------
2 # Copyright (c) 2020-2023, PyInstaller Development Team.
3 #
4 # Distributed under the terms of the GNU General Public License (version 2
5 # or later) with exception for distributing the bootloader.
6 #
7 # The full license is in the file COPYING.txt, distributed with this software.
8 #
9 # SPDX-License-Identifier: (GPL-2.0-or-later WITH Bootloader-exception)
10 #-----------------------------------------------------------------------------
11
12 # TODO: remove this hook during PyInstaller 4.5 release cycle!
13
14 from PyInstaller.utils.hooks import can_import_module, collect_submodules
15
16 # We need to collect submodules from win32ctypes.core.cffi or win32ctypes.core.ctypes for win32ctypes.core to work. The
17 # use of the backend is determined by availability of cffi.
18 if can_import_module('cffi'):
19 hiddenimports = collect_submodules('win32ctypes.core.cffi')
20 else:
21 hiddenimports = collect_submodules('win32ctypes.core.ctypes')
22
[end of PyInstaller/hooks/hook-win32ctypes.core.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/PyInstaller/hooks/hook-win32ctypes.core.py b/PyInstaller/hooks/hook-win32ctypes.core.py
--- a/PyInstaller/hooks/hook-win32ctypes.core.py
+++ b/PyInstaller/hooks/hook-win32ctypes.core.py
@@ -13,9 +13,10 @@
from PyInstaller.utils.hooks import can_import_module, collect_submodules
-# We need to collect submodules from win32ctypes.core.cffi or win32ctypes.core.ctypes for win32ctypes.core to work. The
-# use of the backend is determined by availability of cffi.
+# We need to collect submodules from win32ctypes.core.cffi or win32ctypes.core.ctypes for win32ctypes.core to work.
+# Always collect the `ctypes` backend, and add the `cffi` one if `cffi` is available. Having the `ctypes` backend always
+# available helps in situations when `cffi` is available in the build environment, but is disabled at run-time or not
+# collected (e.g., due to `--exclude cffi`).
+hiddenimports = collect_submodules('win32ctypes.core.ctypes')
if can_import_module('cffi'):
- hiddenimports = collect_submodules('win32ctypes.core.cffi')
-else:
- hiddenimports = collect_submodules('win32ctypes.core.ctypes')
+ hiddenimports += collect_submodules('win32ctypes.core.cffi')
|
{"golden_diff": "diff --git a/PyInstaller/hooks/hook-win32ctypes.core.py b/PyInstaller/hooks/hook-win32ctypes.core.py\n--- a/PyInstaller/hooks/hook-win32ctypes.core.py\n+++ b/PyInstaller/hooks/hook-win32ctypes.core.py\n@@ -13,9 +13,10 @@\n \n from PyInstaller.utils.hooks import can_import_module, collect_submodules\n \n-# We need to collect submodules from win32ctypes.core.cffi or win32ctypes.core.ctypes for win32ctypes.core to work. The\n-# use of the backend is determined by availability of cffi.\n+# We need to collect submodules from win32ctypes.core.cffi or win32ctypes.core.ctypes for win32ctypes.core to work.\n+# Always collect the `ctypes` backend, and add the `cffi` one if `cffi` is available. Having the `ctypes` backend always\n+# available helps in situations when `cffi` is available in the build environment, but is disabled at run-time or not\n+# collected (e.g., due to `--exclude cffi`).\n+hiddenimports = collect_submodules('win32ctypes.core.ctypes')\n if can_import_module('cffi'):\n- hiddenimports = collect_submodules('win32ctypes.core.cffi')\n-else:\n- hiddenimports = collect_submodules('win32ctypes.core.ctypes')\n+ hiddenimports += collect_submodules('win32ctypes.core.cffi')\n", "issue": "Could not import `pywintypes` or `win32api` from `win32ctypes.pywin32`\n## Description of the issue\r\n\r\nError when running the executable.\r\n\r\nIssue is present in 6.6.0 and \"latest development version\". Issue is not present in 6.5.0.\r\n\r\nOutput differs between versions at this point:\r\n\r\n6.6.0:\r\n```\r\nimport 'win32ctypes.core' # <_frozen_importlib_external.SourcelessFileLoader object at 0x000001C6D9E41BB0>\r\n# win32ctypes.core._common not found in PYZ\r\n# win32ctypes.core.ctypes not found in PYZ\r\n# destroy win32ctypes.pywin32.win32api\r\n# destroy win32ctypes.pywin32\r\n# destroy PyInstaller\r\nCould not import `pywintypes` or `win32api` from `win32ctypes.pywin32`.\r\nPlease make sure that `pywin32-ctypes` is installed and importable, for example:\r\n\r\npip install pywin32-ctypes\r\n\r\n```\r\n\r\n6.5.0:\r\n```\r\n# cffi not found in PYZ\r\n# code object from '[...]\\\\cffi\\\\__init__.pyc'\r\n# cffi.api not found in PYZ\r\n# code object from '[...]\\\\cffi\\\\api.pyc'\r\n# cffi.lock not found in PYZ\r\n# code object from '[...]\\\\cffi\\\\lock.pyc'\r\nimport 'cffi.lock' # <_frozen_importlib_external.SourcelessFileLoader object at 0x000001F4AEB845C0>\r\n# cffi.error not found in PYZ\r\n# code object from '[...]\\\\cffi\\\\error.pyc'\r\nimport 'cffi.error' # <_frozen_importlib_external.SourcelessFileLoader object at 0x000001F4AEB846B0>\r\n# cffi.model not found in PYZ\r\n# code object from '[...]\\\\cffi\\\\model.pyc'\r\nimport 'cffi.model' # <_frozen_importlib_external.SourcelessFileLoader object at 0x000001F4AEB848F0>\r\nimport 'cffi.api' # <_frozen_importlib_external.SourcelessFileLoader object at 0x000001F4AEB52330>\r\nimport 'cffi' # <_frozen_importlib_external.SourcelessFileLoader object at 0x000001F4AEB51EB0>\r\nimport 'win32ctypes.core' # <_frozen_importlib_external.SourcelessFileLoader object at 0x000001F4AEB51B80>\r\n# win32ctypes.core._common not found in PYZ\r\n# win32ctypes.core.cffi not found in PYZ\r\n# code object from '[...]\\\\win32ctypes\\\\core\\\\cffi\\\\__init__.pyc'\r\nimport 'win32ctypes.core.cffi' # <_frozen_importlib_external.SourcelessFileLoader object at 0x000001F4AEB86BA0>\r\n# win32ctypes.core.cffi._common not found in PYZ\r\n# code object from '[...]\\\\win32ctypes\\\\core\\\\cffi\\\\_common.pyc'\r\n# win32ctypes.core.cffi._util not found in PYZ\r\n# code object from '[...]\\\\win32ctypes\\\\core\\\\cffi\\\\_util.pyc'\r\n# win32ctypes.core.compat not found in PYZ\r\n# code object from '[...]\\\\win32ctypes\\\\core\\\\compat.pyc'\r\nimport 'win32ctypes.core.compat' # <_frozen_importlib_external.SourcelessFileLoader object at 0x000001F4AEB87440>\r\n# _cffi_backend not found in PYZ\r\n# extension module '_cffi_backend' loaded from '[...]\\\\_cffi_backend.cp312-win_amd64.pyd'\r\n# extension module '_cffi_backend' executed from '[...]\\\\_cffi_backend.cp312-win_amd64.pyd'\r\nimport '_cffi_backend' # <_frozen_importlib_external.ExtensionFileLoader object at 0x000001F4AEB876B0>\r\n# cffi.cparser not found in PYZ\r\n# code object from '[...]\\\\cffi\\\\cparser.pyc'\r\n# cffi.commontypes not found in PYZ\r\n# code object from '[...]\\\\cffi\\\\commontypes.pyc'\r\nimport 'cffi.commontypes' # <_frozen_importlib_external.SourcelessFileLoader object at 0x000001F4AEBA91F0>\r\n```\r\n\r\n\r\n### Context information (for bug reports)\r\n\r\n* 502 INFO: PyInstaller: 6.6.0, contrib hooks: 2024.6\r\n* 503 INFO: Python: 3.12.0\r\n* 541 INFO: Platform: Windows-10-10.0.19045-SP0\r\n\r\n\r\n\n", "before_files": [{"content": "#-----------------------------------------------------------------------------\n# Copyright (c) 2020-2023, PyInstaller Development Team.\n#\n# Distributed under the terms of the GNU General Public License (version 2\n# or later) with exception for distributing the bootloader.\n#\n# The full license is in the file COPYING.txt, distributed with this software.\n#\n# SPDX-License-Identifier: (GPL-2.0-or-later WITH Bootloader-exception)\n#-----------------------------------------------------------------------------\n\n# TODO: remove this hook during PyInstaller 4.5 release cycle!\n\nfrom PyInstaller.utils.hooks import can_import_module, collect_submodules\n\n# We need to collect submodules from win32ctypes.core.cffi or win32ctypes.core.ctypes for win32ctypes.core to work. The\n# use of the backend is determined by availability of cffi.\nif can_import_module('cffi'):\n hiddenimports = collect_submodules('win32ctypes.core.cffi')\nelse:\n hiddenimports = collect_submodules('win32ctypes.core.ctypes')\n", "path": "PyInstaller/hooks/hook-win32ctypes.core.py"}]}
| 1,960 | 338 |
gh_patches_debug_20424
|
rasdani/github-patches
|
git_diff
|
AnalogJ__lexicon-1568
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Misleading error message: ProviderNotAvailableError - This provider (plesk) has required dependencies that are missing. Please install lexicon[plesk] first.
```
(venv) mypc:~/myprojects/dns-lexicon$ pip freeze
certifi==2019.9.11
cffi==1.13.2
chardet==3.0.4
cryptography==2.8
dns-lexicon==3.3.10
future==0.18.2
idna==2.8
pycparser==2.19
pyOpenSSL==19.0.0
PyYAML==5.1.2
requests==2.22.0
requests-file==1.4.3
six==1.13.0
tldextract==2.2.2
urllib3==1.25.7
(venv) mypc:~/myprojects/dns-lexicon$ python main.py
Traceback (most recent call last):
File "main.py", line 111, in <module>
update_server_record(data[0], data[1])
File "main.py", line 62, in update_server_record
plesk_record = list_action(filter_name=server_name, filter_type='A')
File "main.py", line 58, in list_action
return lexicon.client.Client(config).execute()
File "/home/joe/myprojects/dns-lexicon/venv/lib/python3.6/site-packages/lexicon/client.py", line 36, in __init__
self._validate_config()
File "/home/joe/myprojects/dns-lexicon/venv/lib/python3.6/site-packages/lexicon/client.py", line 105, in _validate_config
'Please install lexicon[{0}] first.'.format(provider_name))
lexicon.client.ProviderNotAvailableError: This provider (plesk) has required dependencies that are missing. Please install lexicon[plesk] first.
```
Error is misleading because of this part:
> Please install **lexicon[plesk]** first.
it should be
> Please install **dns-lexicon[plesk]** first.
</issue>
<code>
[start of lexicon/client.py]
1 """Main module of Lexicon. Defines the Client class, that holds all Lexicon logic."""
2 import importlib
3 import logging
4 import os
5 from typing import Dict, List, Optional, Type, Union, cast
6
7 import tldextract # type: ignore
8
9 from lexicon import config as helper_config
10 from lexicon import discovery
11 from lexicon.exceptions import ProviderNotAvailableError
12 from lexicon.providers.base import Provider
13
14
15 class Client(object):
16 """This is the Lexicon client, that will execute all the logic."""
17
18 def __init__(
19 self, config: Optional[Union[helper_config.ConfigResolver, Dict]] = None
20 ):
21 if not config:
22 # If there is not config specified, we load a non-interactive configuration.
23 self.config = helper_config.non_interactive_config_resolver()
24 elif not isinstance(config, helper_config.ConfigResolver):
25 # If config is not a ConfigResolver, we are in a legacy situation.
26 # We protect this part of the Client API.
27 self.config = helper_config.legacy_config_resolver(config)
28 else:
29 self.config = config
30
31 # Validate configuration
32 self._validate_config()
33
34 runtime_config = {}
35
36 # Process domain, strip subdomain
37 try:
38 domain_extractor = tldextract.TLDExtract(
39 cache_dir=_get_tldextract_cache_path(), include_psl_private_domains=True
40 )
41 except TypeError:
42 domain_extractor = tldextract.TLDExtract(
43 cache_file=_get_tldextract_cache_path(), include_psl_private_domains=True # type: ignore
44 )
45 domain_parts = domain_extractor(
46 cast(str, self.config.resolve("lexicon:domain"))
47 )
48 runtime_config["domain"] = f"{domain_parts.domain}.{domain_parts.suffix}"
49
50 delegated = self.config.resolve("lexicon:delegated")
51 if delegated:
52 # handle delegated domain
53 delegated = str(delegated).rstrip(".")
54 initial_domain = str(runtime_config.get("domain"))
55 if delegated != initial_domain:
56 # convert to relative name
57 if delegated.endswith(initial_domain):
58 delegated = delegated[: -len(initial_domain)]
59 delegated = delegated.rstrip(".")
60 # update domain
61 runtime_config["domain"] = f"{delegated}.{initial_domain}"
62
63 self.action = self.config.resolve("lexicon:action")
64 self.provider_name = self.config.resolve(
65 "lexicon:provider_name"
66 ) or self.config.resolve("lexicon:provider")
67
68 if not self.provider_name:
69 raise ValueError("Could not resolve provider name.")
70
71 self.config.add_config_source(helper_config.DictConfigSource(runtime_config), 0)
72
73 provider_module = importlib.import_module(
74 "lexicon.providers." + self.provider_name
75 )
76 provider_class: Type[Provider] = getattr(provider_module, "Provider")
77 self.provider = provider_class(self.config)
78
79 def execute(self) -> Union[bool, List[Dict]]:
80 """Execute provided configuration in class constructor to the DNS records"""
81 self.provider.authenticate()
82 identifier = self.config.resolve("lexicon:identifier")
83 record_type = self.config.resolve("lexicon:type")
84 name = self.config.resolve("lexicon:name")
85 content = self.config.resolve("lexicon:content")
86
87 if self.action == "create":
88 if not record_type or not name or not content:
89 raise ValueError("Missing record_type, name or content parameters.")
90 return self.provider.create_record(record_type, name, content)
91
92 if self.action == "list":
93 return self.provider.list_records(record_type, name, content)
94
95 if self.action == "update":
96 return self.provider.update_record(identifier, record_type, name, content)
97
98 if self.action == "delete":
99 return self.provider.delete_record(identifier, record_type, name, content)
100
101 raise ValueError(f"Invalid action statement: {self.action}")
102
103 def _validate_config(self) -> None:
104 provider_name = self.config.resolve("lexicon:provider_name")
105 if not provider_name:
106 raise AttributeError("provider_name")
107
108 try:
109 available = discovery.find_providers()[provider_name]
110 except KeyError:
111 raise ProviderNotAvailableError(
112 f"This provider ({provider_name}) is not supported by Lexicon."
113 )
114 else:
115 if not available:
116 raise ProviderNotAvailableError(
117 f"This provider ({provider_name}) has required extra dependencies that are missing. "
118 f"Please run `pip install lexicon[{provider_name}]` first before using it."
119 )
120
121 if not self.config.resolve("lexicon:action"):
122 raise AttributeError("action")
123 if not self.config.resolve("lexicon:domain"):
124 raise AttributeError("domain")
125 if not self.config.resolve("lexicon:type"):
126 raise AttributeError("type")
127
128
129 def _get_tldextract_cache_path() -> str:
130 if os.environ.get("TLDEXTRACT_CACHE_FILE"):
131 logging.warning(
132 "TLD_EXTRACT_CACHE_FILE environment variable is deprecated, please use TLDEXTRACT_CACHE_PATH instead."
133 )
134 os.environ["TLDEXTRACT_CACHE_PATH"] = os.environ["TLDEXTRACT_CACHE_FILE"]
135
136 return os.path.expanduser(
137 os.environ.get("TLDEXTRACT_CACHE_PATH", os.path.join("~", ".lexicon_tld_set"))
138 )
139
[end of lexicon/client.py]
[start of lexicon/discovery.py]
1 """
2 This module takes care of finding information about the runtime of Lexicon:
3 * what are the providers installed, and available
4 * what is the version of Lexicon
5 """
6 import pkgutil
7 import re
8 from typing import Dict
9
10 try:
11 from importlib.metadata import Distribution, PackageNotFoundError
12 except ModuleNotFoundError:
13 from importlib_metadata import Distribution, PackageNotFoundError # type: ignore[assignment]
14
15 from lexicon import providers
16
17
18 def find_providers() -> Dict[str, bool]:
19 """Find all providers registered in Lexicon, and their availability"""
20 providers_list = sorted(
21 {
22 modname
23 for (_, modname, _) in pkgutil.iter_modules(providers.__path__) # type: ignore
24 if modname != "base"
25 }
26 )
27
28 try:
29 distribution = Distribution.from_name("dns-lexicon")
30 except PackageNotFoundError:
31 return {provider: True for provider in providers_list}
32 else:
33 return {
34 provider: _resolve_requirements(provider, distribution)
35 for provider in providers_list
36 }
37
38
39 def lexicon_version() -> str:
40 """Retrieve current Lexicon version"""
41 try:
42 return Distribution.from_name("dns-lexicon").version
43 except PackageNotFoundError:
44 return "unknown"
45
46
47 def _resolve_requirements(provider: str, distribution: Distribution) -> bool:
48 requires = distribution.requires
49 if requires is None:
50 raise ValueError("Error while trying finding requirements.")
51
52 requirements = [
53 re.sub(r"^(.*)\s\(.*\)(?:;.*|)$", r"\1", requirement)
54 for requirement in requires
55 if f'extra == "{provider}"' in requirement
56 ]
57
58 if not requirements:
59 # No extra for this provider
60 return True
61
62 for requirement in requirements:
63 try:
64 Distribution.from_name(requirement)
65 except PackageNotFoundError:
66 # At least one extra requirement is not fulfilled
67 return False
68
69 # All extra requirements are fulfilled
70 return True
71
[end of lexicon/discovery.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/lexicon/client.py b/lexicon/client.py
--- a/lexicon/client.py
+++ b/lexicon/client.py
@@ -115,7 +115,7 @@
if not available:
raise ProviderNotAvailableError(
f"This provider ({provider_name}) has required extra dependencies that are missing. "
- f"Please run `pip install lexicon[{provider_name}]` first before using it."
+ f"Please run `pip install dns-lexicon[{provider_name}]` first before using it."
)
if not self.config.resolve("lexicon:action"):
diff --git a/lexicon/discovery.py b/lexicon/discovery.py
--- a/lexicon/discovery.py
+++ b/lexicon/discovery.py
@@ -50,7 +50,7 @@
raise ValueError("Error while trying finding requirements.")
requirements = [
- re.sub(r"^(.*)\s\(.*\)(?:;.*|)$", r"\1", requirement)
+ re.sub(r"^(.*)\s\(.*\)(?:\s*;.*|)$", r"\1", requirement)
for requirement in requires
if f'extra == "{provider}"' in requirement
]
|
{"golden_diff": "diff --git a/lexicon/client.py b/lexicon/client.py\n--- a/lexicon/client.py\n+++ b/lexicon/client.py\n@@ -115,7 +115,7 @@\n if not available:\n raise ProviderNotAvailableError(\n f\"This provider ({provider_name}) has required extra dependencies that are missing. \"\n- f\"Please run `pip install lexicon[{provider_name}]` first before using it.\"\n+ f\"Please run `pip install dns-lexicon[{provider_name}]` first before using it.\"\n )\n \n if not self.config.resolve(\"lexicon:action\"):\ndiff --git a/lexicon/discovery.py b/lexicon/discovery.py\n--- a/lexicon/discovery.py\n+++ b/lexicon/discovery.py\n@@ -50,7 +50,7 @@\n raise ValueError(\"Error while trying finding requirements.\")\n \n requirements = [\n- re.sub(r\"^(.*)\\s\\(.*\\)(?:;.*|)$\", r\"\\1\", requirement)\n+ re.sub(r\"^(.*)\\s\\(.*\\)(?:\\s*;.*|)$\", r\"\\1\", requirement)\n for requirement in requires\n if f'extra == \"{provider}\"' in requirement\n ]\n", "issue": "Misleading error message: ProviderNotAvailableError - This provider (plesk) has required dependencies that are missing. Please install lexicon[plesk] first.\n```\r\n(venv) mypc:~/myprojects/dns-lexicon$ pip freeze\r\ncertifi==2019.9.11\r\ncffi==1.13.2\r\nchardet==3.0.4\r\ncryptography==2.8\r\ndns-lexicon==3.3.10\r\nfuture==0.18.2\r\nidna==2.8\r\npycparser==2.19\r\npyOpenSSL==19.0.0\r\nPyYAML==5.1.2\r\nrequests==2.22.0\r\nrequests-file==1.4.3\r\nsix==1.13.0\r\ntldextract==2.2.2\r\nurllib3==1.25.7\r\n(venv) mypc:~/myprojects/dns-lexicon$ python main.py \r\nTraceback (most recent call last):\r\n File \"main.py\", line 111, in <module>\r\n update_server_record(data[0], data[1])\r\n File \"main.py\", line 62, in update_server_record\r\n plesk_record = list_action(filter_name=server_name, filter_type='A')\r\n File \"main.py\", line 58, in list_action\r\n return lexicon.client.Client(config).execute()\r\n File \"/home/joe/myprojects/dns-lexicon/venv/lib/python3.6/site-packages/lexicon/client.py\", line 36, in __init__\r\n self._validate_config()\r\n File \"/home/joe/myprojects/dns-lexicon/venv/lib/python3.6/site-packages/lexicon/client.py\", line 105, in _validate_config\r\n 'Please install lexicon[{0}] first.'.format(provider_name))\r\nlexicon.client.ProviderNotAvailableError: This provider (plesk) has required dependencies that are missing. Please install lexicon[plesk] first.\r\n\r\n```\r\nError is misleading because of this part: \r\n> Please install **lexicon[plesk]** first.\r\n\r\nit should be \r\n> Please install **dns-lexicon[plesk]** first.\r\n\r\n\r\n\r\n\n", "before_files": [{"content": "\"\"\"Main module of Lexicon. Defines the Client class, that holds all Lexicon logic.\"\"\"\nimport importlib\nimport logging\nimport os\nfrom typing import Dict, List, Optional, Type, Union, cast\n\nimport tldextract # type: ignore\n\nfrom lexicon import config as helper_config\nfrom lexicon import discovery\nfrom lexicon.exceptions import ProviderNotAvailableError\nfrom lexicon.providers.base import Provider\n\n\nclass Client(object):\n \"\"\"This is the Lexicon client, that will execute all the logic.\"\"\"\n\n def __init__(\n self, config: Optional[Union[helper_config.ConfigResolver, Dict]] = None\n ):\n if not config:\n # If there is not config specified, we load a non-interactive configuration.\n self.config = helper_config.non_interactive_config_resolver()\n elif not isinstance(config, helper_config.ConfigResolver):\n # If config is not a ConfigResolver, we are in a legacy situation.\n # We protect this part of the Client API.\n self.config = helper_config.legacy_config_resolver(config)\n else:\n self.config = config\n\n # Validate configuration\n self._validate_config()\n\n runtime_config = {}\n\n # Process domain, strip subdomain\n try:\n domain_extractor = tldextract.TLDExtract(\n cache_dir=_get_tldextract_cache_path(), include_psl_private_domains=True\n )\n except TypeError:\n domain_extractor = tldextract.TLDExtract(\n cache_file=_get_tldextract_cache_path(), include_psl_private_domains=True # type: ignore\n )\n domain_parts = domain_extractor(\n cast(str, self.config.resolve(\"lexicon:domain\"))\n )\n runtime_config[\"domain\"] = f\"{domain_parts.domain}.{domain_parts.suffix}\"\n\n delegated = self.config.resolve(\"lexicon:delegated\")\n if delegated:\n # handle delegated domain\n delegated = str(delegated).rstrip(\".\")\n initial_domain = str(runtime_config.get(\"domain\"))\n if delegated != initial_domain:\n # convert to relative name\n if delegated.endswith(initial_domain):\n delegated = delegated[: -len(initial_domain)]\n delegated = delegated.rstrip(\".\")\n # update domain\n runtime_config[\"domain\"] = f\"{delegated}.{initial_domain}\"\n\n self.action = self.config.resolve(\"lexicon:action\")\n self.provider_name = self.config.resolve(\n \"lexicon:provider_name\"\n ) or self.config.resolve(\"lexicon:provider\")\n\n if not self.provider_name:\n raise ValueError(\"Could not resolve provider name.\")\n\n self.config.add_config_source(helper_config.DictConfigSource(runtime_config), 0)\n\n provider_module = importlib.import_module(\n \"lexicon.providers.\" + self.provider_name\n )\n provider_class: Type[Provider] = getattr(provider_module, \"Provider\")\n self.provider = provider_class(self.config)\n\n def execute(self) -> Union[bool, List[Dict]]:\n \"\"\"Execute provided configuration in class constructor to the DNS records\"\"\"\n self.provider.authenticate()\n identifier = self.config.resolve(\"lexicon:identifier\")\n record_type = self.config.resolve(\"lexicon:type\")\n name = self.config.resolve(\"lexicon:name\")\n content = self.config.resolve(\"lexicon:content\")\n\n if self.action == \"create\":\n if not record_type or not name or not content:\n raise ValueError(\"Missing record_type, name or content parameters.\")\n return self.provider.create_record(record_type, name, content)\n\n if self.action == \"list\":\n return self.provider.list_records(record_type, name, content)\n\n if self.action == \"update\":\n return self.provider.update_record(identifier, record_type, name, content)\n\n if self.action == \"delete\":\n return self.provider.delete_record(identifier, record_type, name, content)\n\n raise ValueError(f\"Invalid action statement: {self.action}\")\n\n def _validate_config(self) -> None:\n provider_name = self.config.resolve(\"lexicon:provider_name\")\n if not provider_name:\n raise AttributeError(\"provider_name\")\n\n try:\n available = discovery.find_providers()[provider_name]\n except KeyError:\n raise ProviderNotAvailableError(\n f\"This provider ({provider_name}) is not supported by Lexicon.\"\n )\n else:\n if not available:\n raise ProviderNotAvailableError(\n f\"This provider ({provider_name}) has required extra dependencies that are missing. \"\n f\"Please run `pip install lexicon[{provider_name}]` first before using it.\"\n )\n\n if not self.config.resolve(\"lexicon:action\"):\n raise AttributeError(\"action\")\n if not self.config.resolve(\"lexicon:domain\"):\n raise AttributeError(\"domain\")\n if not self.config.resolve(\"lexicon:type\"):\n raise AttributeError(\"type\")\n\n\ndef _get_tldextract_cache_path() -> str:\n if os.environ.get(\"TLDEXTRACT_CACHE_FILE\"):\n logging.warning(\n \"TLD_EXTRACT_CACHE_FILE environment variable is deprecated, please use TLDEXTRACT_CACHE_PATH instead.\"\n )\n os.environ[\"TLDEXTRACT_CACHE_PATH\"] = os.environ[\"TLDEXTRACT_CACHE_FILE\"]\n\n return os.path.expanduser(\n os.environ.get(\"TLDEXTRACT_CACHE_PATH\", os.path.join(\"~\", \".lexicon_tld_set\"))\n )\n", "path": "lexicon/client.py"}, {"content": "\"\"\"\nThis module takes care of finding information about the runtime of Lexicon:\n* what are the providers installed, and available\n* what is the version of Lexicon\n\"\"\"\nimport pkgutil\nimport re\nfrom typing import Dict\n\ntry:\n from importlib.metadata import Distribution, PackageNotFoundError\nexcept ModuleNotFoundError:\n from importlib_metadata import Distribution, PackageNotFoundError # type: ignore[assignment]\n\nfrom lexicon import providers\n\n\ndef find_providers() -> Dict[str, bool]:\n \"\"\"Find all providers registered in Lexicon, and their availability\"\"\"\n providers_list = sorted(\n {\n modname\n for (_, modname, _) in pkgutil.iter_modules(providers.__path__) # type: ignore\n if modname != \"base\"\n }\n )\n\n try:\n distribution = Distribution.from_name(\"dns-lexicon\")\n except PackageNotFoundError:\n return {provider: True for provider in providers_list}\n else:\n return {\n provider: _resolve_requirements(provider, distribution)\n for provider in providers_list\n }\n\n\ndef lexicon_version() -> str:\n \"\"\"Retrieve current Lexicon version\"\"\"\n try:\n return Distribution.from_name(\"dns-lexicon\").version\n except PackageNotFoundError:\n return \"unknown\"\n\n\ndef _resolve_requirements(provider: str, distribution: Distribution) -> bool:\n requires = distribution.requires\n if requires is None:\n raise ValueError(\"Error while trying finding requirements.\")\n\n requirements = [\n re.sub(r\"^(.*)\\s\\(.*\\)(?:;.*|)$\", r\"\\1\", requirement)\n for requirement in requires\n if f'extra == \"{provider}\"' in requirement\n ]\n\n if not requirements:\n # No extra for this provider\n return True\n\n for requirement in requirements:\n try:\n Distribution.from_name(requirement)\n except PackageNotFoundError:\n # At least one extra requirement is not fulfilled\n return False\n\n # All extra requirements are fulfilled\n return True\n", "path": "lexicon/discovery.py"}]}
| 3,023 | 273 |
gh_patches_debug_16407
|
rasdani/github-patches
|
git_diff
|
buildbot__buildbot-5729
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
change_hook/poller not working for ReconfigurablePollingChangeSource
In [poller.py](https://github.com/buildbot/buildbot/blob/a0e1d8840e8856ead136a1ad6e2021931355af15/master/buildbot/www/hooks/poller.py#L40), the change sources are filtered like this:
```python
for source in change_svc:
if not isinstance(source, PollingChangeSource):
continue
```
This means that any pollers derived from the super-class `ReconfigurablePollingChangeSource` will not be found. Since [new code is supposed to use `ReconfigurablePollingChangeSource`](https://docs.buildbot.net/current/developer/cls-changesources.html?highlight=reconfigurablepollingchangesource#pollingchangesource), the code should probably read:
```python
for source in change_svc:
if not isinstance(source, ReconfigurablePollingChangeSource):
continue
```
</issue>
<code>
[start of master/buildbot/www/hooks/poller.py]
1 # This file is part of Buildbot. Buildbot is free software: you can
2 # redistribute it and/or modify it under the terms of the GNU General Public
3 # License as published by the Free Software Foundation, version 2.
4 #
5 # This program is distributed in the hope that it will be useful, but WITHOUT
6 # ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS
7 # FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
8 # details.
9 #
10 # You should have received a copy of the GNU General Public License along with
11 # this program; if not, write to the Free Software Foundation, Inc., 51
12 # Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA.
13 #
14 # Copyright Buildbot Team Members
15
16 # This change hook allows GitHub or a hand crafted curl invocation to "knock on
17 # the door" and trigger a change source to poll.
18
19
20 from buildbot.changes.base import PollingChangeSource
21 from buildbot.util import bytes2unicode
22 from buildbot.util import unicode2bytes
23 from buildbot.www.hooks.base import BaseHookHandler
24
25
26 class PollingHandler(BaseHookHandler):
27
28 def getChanges(self, req):
29 change_svc = req.site.master.change_svc
30 poll_all = b"poller" not in req.args
31
32 allow_all = True
33 allowed = []
34 if isinstance(self.options, dict) and b"allowed" in self.options:
35 allow_all = False
36 allowed = self.options[b"allowed"]
37
38 pollers = []
39
40 for source in change_svc:
41 if not isinstance(source, PollingChangeSource):
42 continue
43 if not hasattr(source, "name"):
44 continue
45 if (not poll_all and
46 unicode2bytes(source.name) not in req.args[b'poller']):
47 continue
48 if not allow_all and unicode2bytes(source.name) not in allowed:
49 continue
50 pollers.append(source)
51
52 if not poll_all:
53 missing = (set(req.args[b'poller']) -
54 set(unicode2bytes(s.name) for s in pollers))
55 if missing:
56 raise ValueError("Could not find pollers: {}".format(
57 bytes2unicode(b",".join(missing))))
58
59 for p in pollers:
60 p.force()
61
62 return [], None
63
64
65 poller = PollingHandler
66
[end of master/buildbot/www/hooks/poller.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/master/buildbot/www/hooks/poller.py b/master/buildbot/www/hooks/poller.py
--- a/master/buildbot/www/hooks/poller.py
+++ b/master/buildbot/www/hooks/poller.py
@@ -17,7 +17,7 @@
# the door" and trigger a change source to poll.
-from buildbot.changes.base import PollingChangeSource
+from buildbot.changes.base import ReconfigurablePollingChangeSource
from buildbot.util import bytes2unicode
from buildbot.util import unicode2bytes
from buildbot.www.hooks.base import BaseHookHandler
@@ -38,7 +38,7 @@
pollers = []
for source in change_svc:
- if not isinstance(source, PollingChangeSource):
+ if not isinstance(source, ReconfigurablePollingChangeSource):
continue
if not hasattr(source, "name"):
continue
|
{"golden_diff": "diff --git a/master/buildbot/www/hooks/poller.py b/master/buildbot/www/hooks/poller.py\n--- a/master/buildbot/www/hooks/poller.py\n+++ b/master/buildbot/www/hooks/poller.py\n@@ -17,7 +17,7 @@\n # the door\" and trigger a change source to poll.\n \n \n-from buildbot.changes.base import PollingChangeSource\n+from buildbot.changes.base import ReconfigurablePollingChangeSource\n from buildbot.util import bytes2unicode\n from buildbot.util import unicode2bytes\n from buildbot.www.hooks.base import BaseHookHandler\n@@ -38,7 +38,7 @@\n pollers = []\n \n for source in change_svc:\n- if not isinstance(source, PollingChangeSource):\n+ if not isinstance(source, ReconfigurablePollingChangeSource):\n continue\n if not hasattr(source, \"name\"):\n continue\n", "issue": "change_hook/poller not working for ReconfigurablePollingChangeSource\nIn [poller.py](https://github.com/buildbot/buildbot/blob/a0e1d8840e8856ead136a1ad6e2021931355af15/master/buildbot/www/hooks/poller.py#L40), the change sources are filtered like this:\r\n\r\n```python\r\n for source in change_svc:\r\n if not isinstance(source, PollingChangeSource):\r\n continue\r\n```\r\n\r\nThis means that any pollers derived from the super-class `ReconfigurablePollingChangeSource` will not be found. Since [new code is supposed to use `ReconfigurablePollingChangeSource`](https://docs.buildbot.net/current/developer/cls-changesources.html?highlight=reconfigurablepollingchangesource#pollingchangesource), the code should probably read:\r\n\r\n```python\r\n for source in change_svc:\r\n if not isinstance(source, ReconfigurablePollingChangeSource):\r\n continue\r\n```\r\n\n", "before_files": [{"content": "# This file is part of Buildbot. Buildbot is free software: you can\n# redistribute it and/or modify it under the terms of the GNU General Public\n# License as published by the Free Software Foundation, version 2.\n#\n# This program is distributed in the hope that it will be useful, but WITHOUT\n# ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS\n# FOR A PARTICULAR PURPOSE. See the GNU General Public License for more\n# details.\n#\n# You should have received a copy of the GNU General Public License along with\n# this program; if not, write to the Free Software Foundation, Inc., 51\n# Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA.\n#\n# Copyright Buildbot Team Members\n\n# This change hook allows GitHub or a hand crafted curl invocation to \"knock on\n# the door\" and trigger a change source to poll.\n\n\nfrom buildbot.changes.base import PollingChangeSource\nfrom buildbot.util import bytes2unicode\nfrom buildbot.util import unicode2bytes\nfrom buildbot.www.hooks.base import BaseHookHandler\n\n\nclass PollingHandler(BaseHookHandler):\n\n def getChanges(self, req):\n change_svc = req.site.master.change_svc\n poll_all = b\"poller\" not in req.args\n\n allow_all = True\n allowed = []\n if isinstance(self.options, dict) and b\"allowed\" in self.options:\n allow_all = False\n allowed = self.options[b\"allowed\"]\n\n pollers = []\n\n for source in change_svc:\n if not isinstance(source, PollingChangeSource):\n continue\n if not hasattr(source, \"name\"):\n continue\n if (not poll_all and\n unicode2bytes(source.name) not in req.args[b'poller']):\n continue\n if not allow_all and unicode2bytes(source.name) not in allowed:\n continue\n pollers.append(source)\n\n if not poll_all:\n missing = (set(req.args[b'poller']) -\n set(unicode2bytes(s.name) for s in pollers))\n if missing:\n raise ValueError(\"Could not find pollers: {}\".format(\n bytes2unicode(b\",\".join(missing))))\n\n for p in pollers:\n p.force()\n\n return [], None\n\n\npoller = PollingHandler\n", "path": "master/buildbot/www/hooks/poller.py"}]}
| 1,396 | 194 |
gh_patches_debug_19683
|
rasdani/github-patches
|
git_diff
|
pallets__werkzeug-2413
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
`any` converter should validate value to `url_for`
Expectation:
When using `url_for` for an endpoint with an `<any(foo, bar):arg>` argument in its route, it will raise an error for arguments that are not part of the any statement.
Actual result:
It returns an invalid URL
Example:
```
from flask import Flask, url_for, redirect
app = Flask(__name__)
@app.route("/any/<any(foo, bar):part>")
def url_with_any(part):
return part
@app.route("/")
def index():
return redirect(url_for("url_with_any", part="doesnotexist")) # this will 404
```
Environment:
- Python version: Python 3.10.4
- Flask version: Flask 2.1.1
</issue>
<code>
[start of src/werkzeug/routing/converters.py]
1 import re
2 import typing as t
3 import uuid
4
5 from ..urls import _fast_url_quote
6
7 if t.TYPE_CHECKING:
8 from .map import Map
9
10
11 class ValidationError(ValueError):
12 """Validation error. If a rule converter raises this exception the rule
13 does not match the current URL and the next URL is tried.
14 """
15
16
17 class BaseConverter:
18 """Base class for all converters."""
19
20 regex = "[^/]+"
21 weight = 100
22 part_isolating = True
23
24 def __init__(self, map: "Map", *args: t.Any, **kwargs: t.Any) -> None:
25 self.map = map
26
27 def to_python(self, value: str) -> t.Any:
28 return value
29
30 def to_url(self, value: t.Any) -> str:
31 if isinstance(value, (bytes, bytearray)):
32 return _fast_url_quote(value)
33 return _fast_url_quote(str(value).encode(self.map.charset))
34
35
36 class UnicodeConverter(BaseConverter):
37 """This converter is the default converter and accepts any string but
38 only one path segment. Thus the string can not include a slash.
39
40 This is the default validator.
41
42 Example::
43
44 Rule('/pages/<page>'),
45 Rule('/<string(length=2):lang_code>')
46
47 :param map: the :class:`Map`.
48 :param minlength: the minimum length of the string. Must be greater
49 or equal 1.
50 :param maxlength: the maximum length of the string.
51 :param length: the exact length of the string.
52 """
53
54 part_isolating = True
55
56 def __init__(
57 self,
58 map: "Map",
59 minlength: int = 1,
60 maxlength: t.Optional[int] = None,
61 length: t.Optional[int] = None,
62 ) -> None:
63 super().__init__(map)
64 if length is not None:
65 length_regex = f"{{{int(length)}}}"
66 else:
67 if maxlength is None:
68 maxlength_value = ""
69 else:
70 maxlength_value = str(int(maxlength))
71 length_regex = f"{{{int(minlength)},{maxlength_value}}}"
72 self.regex = f"[^/]{length_regex}"
73
74
75 class AnyConverter(BaseConverter):
76 """Matches one of the items provided. Items can either be Python
77 identifiers or strings::
78
79 Rule('/<any(about, help, imprint, class, "foo,bar"):page_name>')
80
81 :param map: the :class:`Map`.
82 :param items: this function accepts the possible items as positional
83 arguments.
84 """
85
86 part_isolating = True
87
88 def __init__(self, map: "Map", *items: str) -> None:
89 super().__init__(map)
90 self.regex = f"(?:{'|'.join([re.escape(x) for x in items])})"
91
92
93 class PathConverter(BaseConverter):
94 """Like the default :class:`UnicodeConverter`, but it also matches
95 slashes. This is useful for wikis and similar applications::
96
97 Rule('/<path:wikipage>')
98 Rule('/<path:wikipage>/edit')
99
100 :param map: the :class:`Map`.
101 """
102
103 regex = "[^/].*?"
104 weight = 200
105 part_isolating = False
106
107
108 class NumberConverter(BaseConverter):
109 """Baseclass for `IntegerConverter` and `FloatConverter`.
110
111 :internal:
112 """
113
114 weight = 50
115 num_convert: t.Callable = int
116 part_isolating = True
117
118 def __init__(
119 self,
120 map: "Map",
121 fixed_digits: int = 0,
122 min: t.Optional[int] = None,
123 max: t.Optional[int] = None,
124 signed: bool = False,
125 ) -> None:
126 if signed:
127 self.regex = self.signed_regex
128 super().__init__(map)
129 self.fixed_digits = fixed_digits
130 self.min = min
131 self.max = max
132 self.signed = signed
133
134 def to_python(self, value: str) -> t.Any:
135 if self.fixed_digits and len(value) != self.fixed_digits:
136 raise ValidationError()
137 value = self.num_convert(value)
138 if (self.min is not None and value < self.min) or (
139 self.max is not None and value > self.max
140 ):
141 raise ValidationError()
142 return value
143
144 def to_url(self, value: t.Any) -> str:
145 value = str(self.num_convert(value))
146 if self.fixed_digits:
147 value = value.zfill(self.fixed_digits)
148 return value
149
150 @property
151 def signed_regex(self) -> str:
152 return f"-?{self.regex}"
153
154
155 class IntegerConverter(NumberConverter):
156 """This converter only accepts integer values::
157
158 Rule("/page/<int:page>")
159
160 By default it only accepts unsigned, positive values. The ``signed``
161 parameter will enable signed, negative values. ::
162
163 Rule("/page/<int(signed=True):page>")
164
165 :param map: The :class:`Map`.
166 :param fixed_digits: The number of fixed digits in the URL. If you
167 set this to ``4`` for example, the rule will only match if the
168 URL looks like ``/0001/``. The default is variable length.
169 :param min: The minimal value.
170 :param max: The maximal value.
171 :param signed: Allow signed (negative) values.
172
173 .. versionadded:: 0.15
174 The ``signed`` parameter.
175 """
176
177 regex = r"\d+"
178 part_isolating = True
179
180
181 class FloatConverter(NumberConverter):
182 """This converter only accepts floating point values::
183
184 Rule("/probability/<float:probability>")
185
186 By default it only accepts unsigned, positive values. The ``signed``
187 parameter will enable signed, negative values. ::
188
189 Rule("/offset/<float(signed=True):offset>")
190
191 :param map: The :class:`Map`.
192 :param min: The minimal value.
193 :param max: The maximal value.
194 :param signed: Allow signed (negative) values.
195
196 .. versionadded:: 0.15
197 The ``signed`` parameter.
198 """
199
200 regex = r"\d+\.\d+"
201 num_convert = float
202 part_isolating = True
203
204 def __init__(
205 self,
206 map: "Map",
207 min: t.Optional[float] = None,
208 max: t.Optional[float] = None,
209 signed: bool = False,
210 ) -> None:
211 super().__init__(map, min=min, max=max, signed=signed) # type: ignore
212
213
214 class UUIDConverter(BaseConverter):
215 """This converter only accepts UUID strings::
216
217 Rule('/object/<uuid:identifier>')
218
219 .. versionadded:: 0.10
220
221 :param map: the :class:`Map`.
222 """
223
224 regex = (
225 r"[A-Fa-f0-9]{8}-[A-Fa-f0-9]{4}-"
226 r"[A-Fa-f0-9]{4}-[A-Fa-f0-9]{4}-[A-Fa-f0-9]{12}"
227 )
228 part_isolating = True
229
230 def to_python(self, value: str) -> uuid.UUID:
231 return uuid.UUID(value)
232
233 def to_url(self, value: uuid.UUID) -> str:
234 return str(value)
235
236
237 #: the default converter mapping for the map.
238 DEFAULT_CONVERTERS: t.Mapping[str, t.Type[BaseConverter]] = {
239 "default": UnicodeConverter,
240 "string": UnicodeConverter,
241 "any": AnyConverter,
242 "path": PathConverter,
243 "int": IntegerConverter,
244 "float": FloatConverter,
245 "uuid": UUIDConverter,
246 }
247
[end of src/werkzeug/routing/converters.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/src/werkzeug/routing/converters.py b/src/werkzeug/routing/converters.py
--- a/src/werkzeug/routing/converters.py
+++ b/src/werkzeug/routing/converters.py
@@ -81,14 +81,25 @@
:param map: the :class:`Map`.
:param items: this function accepts the possible items as positional
arguments.
+
+ .. versionchanged:: 2.2
+ Value is validated when building a URL.
"""
part_isolating = True
def __init__(self, map: "Map", *items: str) -> None:
super().__init__(map)
+ self.items = set(items)
self.regex = f"(?:{'|'.join([re.escape(x) for x in items])})"
+ def to_url(self, value: t.Any) -> str:
+ if value in self.items:
+ return str(value)
+
+ valid_values = ", ".join(f"'{item}'" for item in sorted(self.items))
+ raise ValueError(f"'{value}' is not one of {valid_values}")
+
class PathConverter(BaseConverter):
"""Like the default :class:`UnicodeConverter`, but it also matches
|
{"golden_diff": "diff --git a/src/werkzeug/routing/converters.py b/src/werkzeug/routing/converters.py\n--- a/src/werkzeug/routing/converters.py\n+++ b/src/werkzeug/routing/converters.py\n@@ -81,14 +81,25 @@\n :param map: the :class:`Map`.\n :param items: this function accepts the possible items as positional\n arguments.\n+\n+ .. versionchanged:: 2.2\n+ Value is validated when building a URL.\n \"\"\"\n \n part_isolating = True\n \n def __init__(self, map: \"Map\", *items: str) -> None:\n super().__init__(map)\n+ self.items = set(items)\n self.regex = f\"(?:{'|'.join([re.escape(x) for x in items])})\"\n \n+ def to_url(self, value: t.Any) -> str:\n+ if value in self.items:\n+ return str(value)\n+\n+ valid_values = \", \".join(f\"'{item}'\" for item in sorted(self.items))\n+ raise ValueError(f\"'{value}' is not one of {valid_values}\")\n+\n \n class PathConverter(BaseConverter):\n \"\"\"Like the default :class:`UnicodeConverter`, but it also matches\n", "issue": "`any` converter should validate value to `url_for`\nExpectation:\r\nWhen using `url_for` for an endpoint with an `<any(foo, bar):arg>` argument in its route, it will raise an error for arguments that are not part of the any statement.\r\n\r\nActual result:\r\nIt returns an invalid URL\r\n\r\n\r\nExample:\r\n```\r\nfrom flask import Flask, url_for, redirect\r\n\r\napp = Flask(__name__)\r\n\r\[email protected](\"/any/<any(foo, bar):part>\")\r\ndef url_with_any(part):\r\n return part\r\n\r\[email protected](\"/\")\r\ndef index():\r\n return redirect(url_for(\"url_with_any\", part=\"doesnotexist\")) # this will 404\r\n\r\n```\r\n\r\nEnvironment:\r\n\r\n- Python version: Python 3.10.4\r\n- Flask version: Flask 2.1.1\r\n\n", "before_files": [{"content": "import re\nimport typing as t\nimport uuid\n\nfrom ..urls import _fast_url_quote\n\nif t.TYPE_CHECKING:\n from .map import Map\n\n\nclass ValidationError(ValueError):\n \"\"\"Validation error. If a rule converter raises this exception the rule\n does not match the current URL and the next URL is tried.\n \"\"\"\n\n\nclass BaseConverter:\n \"\"\"Base class for all converters.\"\"\"\n\n regex = \"[^/]+\"\n weight = 100\n part_isolating = True\n\n def __init__(self, map: \"Map\", *args: t.Any, **kwargs: t.Any) -> None:\n self.map = map\n\n def to_python(self, value: str) -> t.Any:\n return value\n\n def to_url(self, value: t.Any) -> str:\n if isinstance(value, (bytes, bytearray)):\n return _fast_url_quote(value)\n return _fast_url_quote(str(value).encode(self.map.charset))\n\n\nclass UnicodeConverter(BaseConverter):\n \"\"\"This converter is the default converter and accepts any string but\n only one path segment. Thus the string can not include a slash.\n\n This is the default validator.\n\n Example::\n\n Rule('/pages/<page>'),\n Rule('/<string(length=2):lang_code>')\n\n :param map: the :class:`Map`.\n :param minlength: the minimum length of the string. Must be greater\n or equal 1.\n :param maxlength: the maximum length of the string.\n :param length: the exact length of the string.\n \"\"\"\n\n part_isolating = True\n\n def __init__(\n self,\n map: \"Map\",\n minlength: int = 1,\n maxlength: t.Optional[int] = None,\n length: t.Optional[int] = None,\n ) -> None:\n super().__init__(map)\n if length is not None:\n length_regex = f\"{{{int(length)}}}\"\n else:\n if maxlength is None:\n maxlength_value = \"\"\n else:\n maxlength_value = str(int(maxlength))\n length_regex = f\"{{{int(minlength)},{maxlength_value}}}\"\n self.regex = f\"[^/]{length_regex}\"\n\n\nclass AnyConverter(BaseConverter):\n \"\"\"Matches one of the items provided. Items can either be Python\n identifiers or strings::\n\n Rule('/<any(about, help, imprint, class, \"foo,bar\"):page_name>')\n\n :param map: the :class:`Map`.\n :param items: this function accepts the possible items as positional\n arguments.\n \"\"\"\n\n part_isolating = True\n\n def __init__(self, map: \"Map\", *items: str) -> None:\n super().__init__(map)\n self.regex = f\"(?:{'|'.join([re.escape(x) for x in items])})\"\n\n\nclass PathConverter(BaseConverter):\n \"\"\"Like the default :class:`UnicodeConverter`, but it also matches\n slashes. This is useful for wikis and similar applications::\n\n Rule('/<path:wikipage>')\n Rule('/<path:wikipage>/edit')\n\n :param map: the :class:`Map`.\n \"\"\"\n\n regex = \"[^/].*?\"\n weight = 200\n part_isolating = False\n\n\nclass NumberConverter(BaseConverter):\n \"\"\"Baseclass for `IntegerConverter` and `FloatConverter`.\n\n :internal:\n \"\"\"\n\n weight = 50\n num_convert: t.Callable = int\n part_isolating = True\n\n def __init__(\n self,\n map: \"Map\",\n fixed_digits: int = 0,\n min: t.Optional[int] = None,\n max: t.Optional[int] = None,\n signed: bool = False,\n ) -> None:\n if signed:\n self.regex = self.signed_regex\n super().__init__(map)\n self.fixed_digits = fixed_digits\n self.min = min\n self.max = max\n self.signed = signed\n\n def to_python(self, value: str) -> t.Any:\n if self.fixed_digits and len(value) != self.fixed_digits:\n raise ValidationError()\n value = self.num_convert(value)\n if (self.min is not None and value < self.min) or (\n self.max is not None and value > self.max\n ):\n raise ValidationError()\n return value\n\n def to_url(self, value: t.Any) -> str:\n value = str(self.num_convert(value))\n if self.fixed_digits:\n value = value.zfill(self.fixed_digits)\n return value\n\n @property\n def signed_regex(self) -> str:\n return f\"-?{self.regex}\"\n\n\nclass IntegerConverter(NumberConverter):\n \"\"\"This converter only accepts integer values::\n\n Rule(\"/page/<int:page>\")\n\n By default it only accepts unsigned, positive values. The ``signed``\n parameter will enable signed, negative values. ::\n\n Rule(\"/page/<int(signed=True):page>\")\n\n :param map: The :class:`Map`.\n :param fixed_digits: The number of fixed digits in the URL. If you\n set this to ``4`` for example, the rule will only match if the\n URL looks like ``/0001/``. The default is variable length.\n :param min: The minimal value.\n :param max: The maximal value.\n :param signed: Allow signed (negative) values.\n\n .. versionadded:: 0.15\n The ``signed`` parameter.\n \"\"\"\n\n regex = r\"\\d+\"\n part_isolating = True\n\n\nclass FloatConverter(NumberConverter):\n \"\"\"This converter only accepts floating point values::\n\n Rule(\"/probability/<float:probability>\")\n\n By default it only accepts unsigned, positive values. The ``signed``\n parameter will enable signed, negative values. ::\n\n Rule(\"/offset/<float(signed=True):offset>\")\n\n :param map: The :class:`Map`.\n :param min: The minimal value.\n :param max: The maximal value.\n :param signed: Allow signed (negative) values.\n\n .. versionadded:: 0.15\n The ``signed`` parameter.\n \"\"\"\n\n regex = r\"\\d+\\.\\d+\"\n num_convert = float\n part_isolating = True\n\n def __init__(\n self,\n map: \"Map\",\n min: t.Optional[float] = None,\n max: t.Optional[float] = None,\n signed: bool = False,\n ) -> None:\n super().__init__(map, min=min, max=max, signed=signed) # type: ignore\n\n\nclass UUIDConverter(BaseConverter):\n \"\"\"This converter only accepts UUID strings::\n\n Rule('/object/<uuid:identifier>')\n\n .. versionadded:: 0.10\n\n :param map: the :class:`Map`.\n \"\"\"\n\n regex = (\n r\"[A-Fa-f0-9]{8}-[A-Fa-f0-9]{4}-\"\n r\"[A-Fa-f0-9]{4}-[A-Fa-f0-9]{4}-[A-Fa-f0-9]{12}\"\n )\n part_isolating = True\n\n def to_python(self, value: str) -> uuid.UUID:\n return uuid.UUID(value)\n\n def to_url(self, value: uuid.UUID) -> str:\n return str(value)\n\n\n#: the default converter mapping for the map.\nDEFAULT_CONVERTERS: t.Mapping[str, t.Type[BaseConverter]] = {\n \"default\": UnicodeConverter,\n \"string\": UnicodeConverter,\n \"any\": AnyConverter,\n \"path\": PathConverter,\n \"int\": IntegerConverter,\n \"float\": FloatConverter,\n \"uuid\": UUIDConverter,\n}\n", "path": "src/werkzeug/routing/converters.py"}]}
| 3,046 | 277 |
gh_patches_debug_12830
|
rasdani/github-patches
|
git_diff
|
mars-project__mars-82
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
By default use core number as n_parallel for threaded scheduling
Use core number as `n_parallel` for threaded scheduling, currently 1 thread by default.
</issue>
<code>
[start of mars/session.py]
1 #!/usr/bin/env python
2 # -*- coding: utf-8 -*-
3 # Copyright 1999-2018 Alibaba Group Holding Ltd.
4 #
5 # Licensed under the Apache License, Version 2.0 (the "License");
6 # you may not use this file except in compliance with the License.
7 # You may obtain a copy of the License at
8 #
9 # http://www.apache.org/licenses/LICENSE-2.0
10 #
11 # Unless required by applicable law or agreed to in writing, software
12 # distributed under the License is distributed on an "AS IS" BASIS,
13 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
14 # See the License for the specific language governing permissions and
15 # limitations under the License.
16
17 import numpy as np
18
19
20 class LocalSession(object):
21 def __init__(self):
22 from .tensor.execution.core import Executor
23
24 self._executor = Executor()
25 self._endpoint = None
26
27 @property
28 def endpoint(self):
29 return self._endpoint
30
31 @endpoint.setter
32 def endpoint(self, endpoint):
33 if endpoint is not None:
34 raise ValueError('Local session cannot set endpoint')
35 self._endpoint = endpoint
36
37 def run(self, *tensors, **kw):
38 if self._executor is None:
39 raise RuntimeError('Session has closed')
40 return self._executor.execute_tensors(tensors, **kw)
41
42 def decref(self, *keys):
43 self._executor.decref(*keys)
44
45 def __enter__(self):
46 return self
47
48 def __exit__(self, *_):
49 self._executor = None
50
51
52 class Session(object):
53 _default_session = None
54
55 def __init__(self, endpoint=None):
56 if endpoint is not None:
57 if 'http' in endpoint:
58 # connect to web
59 from .web.session import Session as WebSession
60
61 self._sess = WebSession(endpoint)
62 else:
63 # connect to local cluster
64 from .deploy.local.session import LocalClusterSession
65
66 self._sess = LocalClusterSession(endpoint)
67 else:
68 self._sess = LocalSession()
69
70 self._executed_keys = set()
71
72 def run(self, *tensors, **kw):
73 from . import tensor as mt
74
75 ret_list = False
76 if len(tensors) == 1 and isinstance(tensors[0], (tuple, list)):
77 ret_list = True
78 tensors = tensors[0]
79 elif len(tensors) > 1:
80 ret_list = True
81
82 tensors = tuple(mt.tensor(t) for t in tensors)
83 result = self._sess.run(*tensors, **kw)
84 self._executed_keys.update(t.key for t in tensors)
85 for t in tensors:
86 t._execute_session = self
87
88 ret = []
89 for r, t in zip(result, tensors):
90 if r is None:
91 ret.append(r)
92 continue
93 if t.isscalar() and hasattr(r, 'item'):
94 ret.append(np.asscalar(r))
95 else:
96 ret.append(r)
97 if ret_list:
98 return ret
99 return ret[0]
100
101 @property
102 def endpoint(self):
103 return self._sess.endpoint
104
105 @endpoint.setter
106 def endpoint(self, endpoint):
107 self._sess.endpoint = endpoint
108
109 def decref(self, *keys):
110 if hasattr(self._sess, 'decref'):
111 self._sess.decref(*keys)
112
113 def __getattr__(self, attr):
114 try:
115 obj = self._sess.__getattribute__(attr)
116 return obj
117 except AttributeError:
118 raise
119
120 def __enter__(self):
121 self._sess.__enter__()
122 return self
123
124 def __exit__(self, exc_type, exc_val, exc_tb):
125 self._sess.__exit__(exc_type, exc_val, exc_tb)
126
127 close = __exit__
128
129 def as_default(self):
130 Session._default_session = self
131 return self
132
133 @classmethod
134 def default_or_local(cls):
135 if cls._default_session is not None:
136 return cls._default_session
137
138 cls._default_session = Session()
139 return cls._default_session
140
141
142 def new_session(scheduler=None):
143 return Session(scheduler)
144
[end of mars/session.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/mars/session.py b/mars/session.py
--- a/mars/session.py
+++ b/mars/session.py
@@ -16,6 +16,11 @@
import numpy as np
+try:
+ from .resource import cpu_count
+except ImportError:
+ from multiprocessing import cpu_count
+
class LocalSession(object):
def __init__(self):
@@ -37,6 +42,8 @@
def run(self, *tensors, **kw):
if self._executor is None:
raise RuntimeError('Session has closed')
+ if 'n_parallel' not in kw:
+ kw['n_parallel'] = cpu_count()
return self._executor.execute_tensors(tensors, **kw)
def decref(self, *keys):
|
{"golden_diff": "diff --git a/mars/session.py b/mars/session.py\n--- a/mars/session.py\n+++ b/mars/session.py\n@@ -16,6 +16,11 @@\n \n import numpy as np\n \n+try:\n+ from .resource import cpu_count\n+except ImportError:\n+ from multiprocessing import cpu_count\n+\n \n class LocalSession(object):\n def __init__(self):\n@@ -37,6 +42,8 @@\n def run(self, *tensors, **kw):\n if self._executor is None:\n raise RuntimeError('Session has closed')\n+ if 'n_parallel' not in kw:\n+ kw['n_parallel'] = cpu_count()\n return self._executor.execute_tensors(tensors, **kw)\n \n def decref(self, *keys):\n", "issue": "By default use core number as n_parallel for threaded scheduling\nUse core number as `n_parallel` for threaded scheduling, currently 1 thread by default.\n", "before_files": [{"content": "#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n# Copyright 1999-2018 Alibaba Group Holding Ltd.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nimport numpy as np\n\n\nclass LocalSession(object):\n def __init__(self):\n from .tensor.execution.core import Executor\n\n self._executor = Executor()\n self._endpoint = None\n\n @property\n def endpoint(self):\n return self._endpoint\n\n @endpoint.setter\n def endpoint(self, endpoint):\n if endpoint is not None:\n raise ValueError('Local session cannot set endpoint')\n self._endpoint = endpoint\n\n def run(self, *tensors, **kw):\n if self._executor is None:\n raise RuntimeError('Session has closed')\n return self._executor.execute_tensors(tensors, **kw)\n\n def decref(self, *keys):\n self._executor.decref(*keys)\n\n def __enter__(self):\n return self\n\n def __exit__(self, *_):\n self._executor = None\n\n\nclass Session(object):\n _default_session = None\n\n def __init__(self, endpoint=None):\n if endpoint is not None:\n if 'http' in endpoint:\n # connect to web\n from .web.session import Session as WebSession\n\n self._sess = WebSession(endpoint)\n else:\n # connect to local cluster\n from .deploy.local.session import LocalClusterSession\n\n self._sess = LocalClusterSession(endpoint)\n else:\n self._sess = LocalSession()\n\n self._executed_keys = set()\n\n def run(self, *tensors, **kw):\n from . import tensor as mt\n\n ret_list = False\n if len(tensors) == 1 and isinstance(tensors[0], (tuple, list)):\n ret_list = True\n tensors = tensors[0]\n elif len(tensors) > 1:\n ret_list = True\n\n tensors = tuple(mt.tensor(t) for t in tensors)\n result = self._sess.run(*tensors, **kw)\n self._executed_keys.update(t.key for t in tensors)\n for t in tensors:\n t._execute_session = self\n\n ret = []\n for r, t in zip(result, tensors):\n if r is None:\n ret.append(r)\n continue\n if t.isscalar() and hasattr(r, 'item'):\n ret.append(np.asscalar(r))\n else:\n ret.append(r)\n if ret_list:\n return ret\n return ret[0]\n\n @property\n def endpoint(self):\n return self._sess.endpoint\n\n @endpoint.setter\n def endpoint(self, endpoint):\n self._sess.endpoint = endpoint\n\n def decref(self, *keys):\n if hasattr(self._sess, 'decref'):\n self._sess.decref(*keys)\n\n def __getattr__(self, attr):\n try:\n obj = self._sess.__getattribute__(attr)\n return obj\n except AttributeError:\n raise\n\n def __enter__(self):\n self._sess.__enter__()\n return self\n\n def __exit__(self, exc_type, exc_val, exc_tb):\n self._sess.__exit__(exc_type, exc_val, exc_tb)\n\n close = __exit__\n\n def as_default(self):\n Session._default_session = self\n return self\n\n @classmethod\n def default_or_local(cls):\n if cls._default_session is not None:\n return cls._default_session\n\n cls._default_session = Session()\n return cls._default_session\n\n\ndef new_session(scheduler=None):\n return Session(scheduler)\n", "path": "mars/session.py"}]}
| 1,799 | 170 |
gh_patches_debug_15984
|
rasdani/github-patches
|
git_diff
|
OpenMined__PySyft-5397
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Adding loguru compatiblity with pytest caplog
## Description
`caplog` fixture in pytest captures the logging output for testing if appropriate warnings have been raised.
By default pytest uses the standard `logging` module, but since we are using `loguru` appropriate patching needs to be added.
## Additional Context
https://loguru.readthedocs.io/en/stable/resources/migration.html#making-things-work-with-pytest-and-caplog
</issue>
<code>
[start of src/syft/logger.py]
1 # stdlib
2 import os
3 from typing import Any
4 from typing import Callable
5 from typing import NoReturn
6 from typing import TextIO
7 from typing import Union
8
9 # third party
10 from loguru import logger
11
12 LOG_FORMAT = "[{time}][{level}][{module}]][{process.id}] {message}"
13
14 logger.remove()
15 DEFAULT_SINK = "syft_{time}.log"
16
17
18 def remove() -> None:
19 logger.remove()
20
21
22 def add(
23 sink: Union[None, str, os.PathLike, TextIO] = None,
24 level: str = "ERROR",
25 ) -> None:
26 sink = DEFAULT_SINK if sink is None else sink
27 try:
28 logger.add(
29 sink=sink,
30 format=LOG_FORMAT,
31 enqueue=True,
32 colorize=False,
33 diagnose=True,
34 backtrace=True,
35 rotation="10 MB",
36 retention="1 day",
37 level=level,
38 )
39 except BaseException:
40 logger.add(
41 sink=sink,
42 format=LOG_FORMAT,
43 enqueue=True,
44 colorize=False,
45 diagnose=True,
46 backtrace=True,
47 level=level,
48 )
49
50
51 def traceback_and_raise(e: Any, verbose: bool = False) -> NoReturn:
52 try:
53 if verbose:
54 logger.opt(lazy=True).exception(e)
55 else:
56 logger.opt(lazy=True).critical(e)
57 except BaseException as ex:
58 logger.debug("failed to print exception", ex)
59 if not issubclass(type(e), Exception):
60 e = Exception(e)
61 raise e
62
63
64 def create_log_and_print_function(level: str) -> Callable:
65 def log_and_print(*args: Any, **kwargs: Any) -> None:
66 try:
67 method = getattr(logger.opt(lazy=True), level, None)
68 if "print" in kwargs and kwargs["print"] is True:
69 del kwargs["print"]
70 print(*args, **kwargs)
71 if "end" in kwargs:
72 # clean up extra end for printing
73 del kwargs["end"]
74
75 if method is not None:
76 method(*args, **kwargs)
77 else:
78 raise Exception(f"no method {level} on logger")
79 except BaseException as e:
80 msg = f"failed to log exception. {e}"
81 try:
82 logger.debug(msg)
83 except Exception as e:
84 print(f"{msg}. {e}")
85
86 return log_and_print
87
88
89 def traceback(*args: Any, **kwargs: Any) -> None:
90 return create_log_and_print_function(level="exception")(*args, **kwargs)
91
92
93 def critical(*args: Any, **kwargs: Any) -> None:
94 return create_log_and_print_function(level="critical")(*args, **kwargs)
95
96
97 def error(*args: Any, **kwargs: Any) -> None:
98 return create_log_and_print_function(level="error")(*args, **kwargs)
99
100
101 def warning(*args: Any, **kwargs: Any) -> None:
102 return create_log_and_print_function(level="warning")(*args, **kwargs)
103
104
105 def info(*args: Any, **kwargs: Any) -> None:
106 return create_log_and_print_function(level="info")(*args, **kwargs)
107
108
109 def debug(*args: Any, **kwargs: Any) -> None:
110 return create_log_and_print_function(level="debug")(*args, **kwargs)
111
112
113 def trace(*args: Any, **kwargs: Any) -> None:
114 return create_log_and_print_function(level="trace")(*args, **kwargs)
115
[end of src/syft/logger.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/src/syft/logger.py b/src/syft/logger.py
--- a/src/syft/logger.py
+++ b/src/syft/logger.py
@@ -1,4 +1,5 @@
# stdlib
+import logging
import os
from typing import Any
from typing import Callable
@@ -20,7 +21,7 @@
def add(
- sink: Union[None, str, os.PathLike, TextIO] = None,
+ sink: Union[None, str, os.PathLike, TextIO, logging.Handler] = None,
level: str = "ERROR",
) -> None:
sink = DEFAULT_SINK if sink is None else sink
@@ -40,7 +41,6 @@
logger.add(
sink=sink,
format=LOG_FORMAT,
- enqueue=True,
colorize=False,
diagnose=True,
backtrace=True,
|
{"golden_diff": "diff --git a/src/syft/logger.py b/src/syft/logger.py\n--- a/src/syft/logger.py\n+++ b/src/syft/logger.py\n@@ -1,4 +1,5 @@\n # stdlib\n+import logging\n import os\n from typing import Any\n from typing import Callable\n@@ -20,7 +21,7 @@\n \n \n def add(\n- sink: Union[None, str, os.PathLike, TextIO] = None,\n+ sink: Union[None, str, os.PathLike, TextIO, logging.Handler] = None,\n level: str = \"ERROR\",\n ) -> None:\n sink = DEFAULT_SINK if sink is None else sink\n@@ -40,7 +41,6 @@\n logger.add(\n sink=sink,\n format=LOG_FORMAT,\n- enqueue=True,\n colorize=False,\n diagnose=True,\n backtrace=True,\n", "issue": "Adding loguru compatiblity with pytest caplog\n## Description\r\n`caplog` fixture in pytest captures the logging output for testing if appropriate warnings have been raised.\r\n\r\nBy default pytest uses the standard `logging` module, but since we are using `loguru` appropriate patching needs to be added.\r\n\r\n## Additional Context\r\nhttps://loguru.readthedocs.io/en/stable/resources/migration.html#making-things-work-with-pytest-and-caplog\r\n\n", "before_files": [{"content": "# stdlib\nimport os\nfrom typing import Any\nfrom typing import Callable\nfrom typing import NoReturn\nfrom typing import TextIO\nfrom typing import Union\n\n# third party\nfrom loguru import logger\n\nLOG_FORMAT = \"[{time}][{level}][{module}]][{process.id}] {message}\"\n\nlogger.remove()\nDEFAULT_SINK = \"syft_{time}.log\"\n\n\ndef remove() -> None:\n logger.remove()\n\n\ndef add(\n sink: Union[None, str, os.PathLike, TextIO] = None,\n level: str = \"ERROR\",\n) -> None:\n sink = DEFAULT_SINK if sink is None else sink\n try:\n logger.add(\n sink=sink,\n format=LOG_FORMAT,\n enqueue=True,\n colorize=False,\n diagnose=True,\n backtrace=True,\n rotation=\"10 MB\",\n retention=\"1 day\",\n level=level,\n )\n except BaseException:\n logger.add(\n sink=sink,\n format=LOG_FORMAT,\n enqueue=True,\n colorize=False,\n diagnose=True,\n backtrace=True,\n level=level,\n )\n\n\ndef traceback_and_raise(e: Any, verbose: bool = False) -> NoReturn:\n try:\n if verbose:\n logger.opt(lazy=True).exception(e)\n else:\n logger.opt(lazy=True).critical(e)\n except BaseException as ex:\n logger.debug(\"failed to print exception\", ex)\n if not issubclass(type(e), Exception):\n e = Exception(e)\n raise e\n\n\ndef create_log_and_print_function(level: str) -> Callable:\n def log_and_print(*args: Any, **kwargs: Any) -> None:\n try:\n method = getattr(logger.opt(lazy=True), level, None)\n if \"print\" in kwargs and kwargs[\"print\"] is True:\n del kwargs[\"print\"]\n print(*args, **kwargs)\n if \"end\" in kwargs:\n # clean up extra end for printing\n del kwargs[\"end\"]\n\n if method is not None:\n method(*args, **kwargs)\n else:\n raise Exception(f\"no method {level} on logger\")\n except BaseException as e:\n msg = f\"failed to log exception. {e}\"\n try:\n logger.debug(msg)\n except Exception as e:\n print(f\"{msg}. {e}\")\n\n return log_and_print\n\n\ndef traceback(*args: Any, **kwargs: Any) -> None:\n return create_log_and_print_function(level=\"exception\")(*args, **kwargs)\n\n\ndef critical(*args: Any, **kwargs: Any) -> None:\n return create_log_and_print_function(level=\"critical\")(*args, **kwargs)\n\n\ndef error(*args: Any, **kwargs: Any) -> None:\n return create_log_and_print_function(level=\"error\")(*args, **kwargs)\n\n\ndef warning(*args: Any, **kwargs: Any) -> None:\n return create_log_and_print_function(level=\"warning\")(*args, **kwargs)\n\n\ndef info(*args: Any, **kwargs: Any) -> None:\n return create_log_and_print_function(level=\"info\")(*args, **kwargs)\n\n\ndef debug(*args: Any, **kwargs: Any) -> None:\n return create_log_and_print_function(level=\"debug\")(*args, **kwargs)\n\n\ndef trace(*args: Any, **kwargs: Any) -> None:\n return create_log_and_print_function(level=\"trace\")(*args, **kwargs)\n", "path": "src/syft/logger.py"}]}
| 1,627 | 197 |
gh_patches_debug_36091
|
rasdani/github-patches
|
git_diff
|
freedomofpress__securedrop-5894
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
v2 removal on restore does not cover HTTPS services
The logic added in https://github.com/freedomofpress/securedrop/pull/5677 to disable v2 onion services when a backup is restored to a Focal server does not remove config lines for HTTPS services (port 443), potentially resulting in a broken configuration.
</issue>
<code>
[start of install_files/ansible-base/roles/restore/files/disable_v2.py]
1 #!/usr/bin/env python3
2 # To execute on prod:
3 # python3 disable_v2.py /etc/tor/torrc /etc/tor/torrc
4 # To execute for testing locally:
5 # python3 disable_v2.py /etc/tor/torrc /tmp/dumytorrc
6 import sys
7
8
9 def filter_v2(filename):
10 # Read the file
11 with open(filename) as f:
12 data = f.readlines()
13 # We will store the filtered lines to result
14 result = []
15
16 i = 0
17 while i < len(data):
18 line = data[i]
19 if line == "HiddenServiceDir /var/lib/tor/services/source\n":
20 i += 1
21 while data[i].strip() == "":
22 i += 1
23 line = data[i]
24 if line == "HiddenServiceVersion 2\n":
25 i += 1
26 line = data[i]
27 while data[i].strip() == "":
28 i += 1
29 line = data[i]
30 if line == "HiddenServicePort 80 127.0.0.1:80\n":
31 i += 1
32 continue
33 # Now check for journalist
34 if line == "HiddenServiceDir /var/lib/tor/services/journalist\n":
35 i += 1
36 while data[i].strip() == "":
37 i += 1
38 line = data[i]
39 if line == "HiddenServiceVersion 2\n":
40 i += 1
41 line = data[i]
42 while data[i].strip() == "":
43 i += 1
44 line = data[i]
45 if line == "HiddenServicePort 80 127.0.0.1:8080\n":
46 i += 1
47 line = data[i]
48 while data[i].strip() == "":
49 i += 1
50 line = data[i]
51 if line == "HiddenServiceAuthorizeClient stealth journalist\n":
52 i += 1
53 continue
54 # Now the v2 ssh access
55 if line == "HiddenServiceDir /var/lib/tor/services/ssh\n":
56 i += 1
57 while data[i].strip() == "":
58 i += 1
59 line = data[i]
60 if line == "HiddenServiceVersion 2\n":
61 i += 1
62 line = data[i]
63 while data[i].strip() == "":
64 i += 1
65 line = data[i]
66 if line == "HiddenServicePort 22 127.0.0.1:22\n":
67 i += 1
68 line = data[i]
69 while data[i].strip() == "":
70 i += 1
71 line = data[i]
72 if line == "HiddenServiceAuthorizeClient stealth admin\n":
73 i += 1
74 continue
75
76 result.append(line)
77 i += 1
78
79 # Now return the result
80 return result
81
82
83 if __name__ == "__main__":
84 filename = sys.argv[1]
85 outputfilename = sys.argv[2]
86 result = filter_v2(filename)
87 with open(outputfilename, "w") as fobj:
88 for line in result:
89 fobj.write(line)
90
[end of install_files/ansible-base/roles/restore/files/disable_v2.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/install_files/ansible-base/roles/restore/files/disable_v2.py b/install_files/ansible-base/roles/restore/files/disable_v2.py
deleted file mode 100644
--- a/install_files/ansible-base/roles/restore/files/disable_v2.py
+++ /dev/null
@@ -1,89 +0,0 @@
-#!/usr/bin/env python3
-# To execute on prod:
-# python3 disable_v2.py /etc/tor/torrc /etc/tor/torrc
-# To execute for testing locally:
-# python3 disable_v2.py /etc/tor/torrc /tmp/dumytorrc
-import sys
-
-
-def filter_v2(filename):
- # Read the file
- with open(filename) as f:
- data = f.readlines()
- # We will store the filtered lines to result
- result = []
-
- i = 0
- while i < len(data):
- line = data[i]
- if line == "HiddenServiceDir /var/lib/tor/services/source\n":
- i += 1
- while data[i].strip() == "":
- i += 1
- line = data[i]
- if line == "HiddenServiceVersion 2\n":
- i += 1
- line = data[i]
- while data[i].strip() == "":
- i += 1
- line = data[i]
- if line == "HiddenServicePort 80 127.0.0.1:80\n":
- i += 1
- continue
- # Now check for journalist
- if line == "HiddenServiceDir /var/lib/tor/services/journalist\n":
- i += 1
- while data[i].strip() == "":
- i += 1
- line = data[i]
- if line == "HiddenServiceVersion 2\n":
- i += 1
- line = data[i]
- while data[i].strip() == "":
- i += 1
- line = data[i]
- if line == "HiddenServicePort 80 127.0.0.1:8080\n":
- i += 1
- line = data[i]
- while data[i].strip() == "":
- i += 1
- line = data[i]
- if line == "HiddenServiceAuthorizeClient stealth journalist\n":
- i += 1
- continue
- # Now the v2 ssh access
- if line == "HiddenServiceDir /var/lib/tor/services/ssh\n":
- i += 1
- while data[i].strip() == "":
- i += 1
- line = data[i]
- if line == "HiddenServiceVersion 2\n":
- i += 1
- line = data[i]
- while data[i].strip() == "":
- i += 1
- line = data[i]
- if line == "HiddenServicePort 22 127.0.0.1:22\n":
- i += 1
- line = data[i]
- while data[i].strip() == "":
- i += 1
- line = data[i]
- if line == "HiddenServiceAuthorizeClient stealth admin\n":
- i += 1
- continue
-
- result.append(line)
- i += 1
-
- # Now return the result
- return result
-
-
-if __name__ == "__main__":
- filename = sys.argv[1]
- outputfilename = sys.argv[2]
- result = filter_v2(filename)
- with open(outputfilename, "w") as fobj:
- for line in result:
- fobj.write(line)
|
{"golden_diff": "diff --git a/install_files/ansible-base/roles/restore/files/disable_v2.py b/install_files/ansible-base/roles/restore/files/disable_v2.py\ndeleted file mode 100644\n--- a/install_files/ansible-base/roles/restore/files/disable_v2.py\n+++ /dev/null\n@@ -1,89 +0,0 @@\n-#!/usr/bin/env python3\n-# To execute on prod:\n-# python3 disable_v2.py /etc/tor/torrc /etc/tor/torrc\n-# To execute for testing locally:\n-# python3 disable_v2.py /etc/tor/torrc /tmp/dumytorrc\n-import sys\n-\n-\n-def filter_v2(filename):\n- # Read the file\n- with open(filename) as f:\n- data = f.readlines()\n- # We will store the filtered lines to result\n- result = []\n-\n- i = 0\n- while i < len(data):\n- line = data[i]\n- if line == \"HiddenServiceDir /var/lib/tor/services/source\\n\":\n- i += 1\n- while data[i].strip() == \"\":\n- i += 1\n- line = data[i]\n- if line == \"HiddenServiceVersion 2\\n\":\n- i += 1\n- line = data[i]\n- while data[i].strip() == \"\":\n- i += 1\n- line = data[i]\n- if line == \"HiddenServicePort 80 127.0.0.1:80\\n\":\n- i += 1\n- continue\n- # Now check for journalist\n- if line == \"HiddenServiceDir /var/lib/tor/services/journalist\\n\":\n- i += 1\n- while data[i].strip() == \"\":\n- i += 1\n- line = data[i]\n- if line == \"HiddenServiceVersion 2\\n\":\n- i += 1\n- line = data[i]\n- while data[i].strip() == \"\":\n- i += 1\n- line = data[i]\n- if line == \"HiddenServicePort 80 127.0.0.1:8080\\n\":\n- i += 1\n- line = data[i]\n- while data[i].strip() == \"\":\n- i += 1\n- line = data[i]\n- if line == \"HiddenServiceAuthorizeClient stealth journalist\\n\":\n- i += 1\n- continue\n- # Now the v2 ssh access\n- if line == \"HiddenServiceDir /var/lib/tor/services/ssh\\n\":\n- i += 1\n- while data[i].strip() == \"\":\n- i += 1\n- line = data[i]\n- if line == \"HiddenServiceVersion 2\\n\":\n- i += 1\n- line = data[i]\n- while data[i].strip() == \"\":\n- i += 1\n- line = data[i]\n- if line == \"HiddenServicePort 22 127.0.0.1:22\\n\":\n- i += 1\n- line = data[i]\n- while data[i].strip() == \"\":\n- i += 1\n- line = data[i]\n- if line == \"HiddenServiceAuthorizeClient stealth admin\\n\":\n- i += 1\n- continue\n-\n- result.append(line)\n- i += 1\n-\n- # Now return the result\n- return result\n-\n-\n-if __name__ == \"__main__\":\n- filename = sys.argv[1]\n- outputfilename = sys.argv[2]\n- result = filter_v2(filename)\n- with open(outputfilename, \"w\") as fobj:\n- for line in result:\n- fobj.write(line)\n", "issue": "v2 removal on restore does not cover HTTPS services\nThe logic added in https://github.com/freedomofpress/securedrop/pull/5677 to disable v2 onion services when a backup is restored to a Focal server does not remove config lines for HTTPS services (port 443), potentially resulting in a broken configuration.\n", "before_files": [{"content": "#!/usr/bin/env python3\n# To execute on prod:\n# python3 disable_v2.py /etc/tor/torrc /etc/tor/torrc\n# To execute for testing locally:\n# python3 disable_v2.py /etc/tor/torrc /tmp/dumytorrc\nimport sys\n\n\ndef filter_v2(filename):\n # Read the file\n with open(filename) as f:\n data = f.readlines()\n # We will store the filtered lines to result\n result = []\n\n i = 0\n while i < len(data):\n line = data[i]\n if line == \"HiddenServiceDir /var/lib/tor/services/source\\n\":\n i += 1\n while data[i].strip() == \"\":\n i += 1\n line = data[i]\n if line == \"HiddenServiceVersion 2\\n\":\n i += 1\n line = data[i]\n while data[i].strip() == \"\":\n i += 1\n line = data[i]\n if line == \"HiddenServicePort 80 127.0.0.1:80\\n\":\n i += 1\n continue\n # Now check for journalist\n if line == \"HiddenServiceDir /var/lib/tor/services/journalist\\n\":\n i += 1\n while data[i].strip() == \"\":\n i += 1\n line = data[i]\n if line == \"HiddenServiceVersion 2\\n\":\n i += 1\n line = data[i]\n while data[i].strip() == \"\":\n i += 1\n line = data[i]\n if line == \"HiddenServicePort 80 127.0.0.1:8080\\n\":\n i += 1\n line = data[i]\n while data[i].strip() == \"\":\n i += 1\n line = data[i]\n if line == \"HiddenServiceAuthorizeClient stealth journalist\\n\":\n i += 1\n continue\n # Now the v2 ssh access\n if line == \"HiddenServiceDir /var/lib/tor/services/ssh\\n\":\n i += 1\n while data[i].strip() == \"\":\n i += 1\n line = data[i]\n if line == \"HiddenServiceVersion 2\\n\":\n i += 1\n line = data[i]\n while data[i].strip() == \"\":\n i += 1\n line = data[i]\n if line == \"HiddenServicePort 22 127.0.0.1:22\\n\":\n i += 1\n line = data[i]\n while data[i].strip() == \"\":\n i += 1\n line = data[i]\n if line == \"HiddenServiceAuthorizeClient stealth admin\\n\":\n i += 1\n continue\n\n result.append(line)\n i += 1\n\n # Now return the result\n return result\n\n\nif __name__ == \"__main__\":\n filename = sys.argv[1]\n outputfilename = sys.argv[2]\n result = filter_v2(filename)\n with open(outputfilename, \"w\") as fobj:\n for line in result:\n fobj.write(line)\n", "path": "install_files/ansible-base/roles/restore/files/disable_v2.py"}]}
| 1,496 | 856 |
gh_patches_debug_41198
|
rasdani/github-patches
|
git_diff
|
cupy__cupy-6872
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
array_api sort() does not work
### Description
The array_api sort() is copied from the NumPy implementation, but passes through a `kind` keyword which doesn't exist in `cupy.sort`.
### To Reproduce
```py
>>> a = xp.asarray([1, 0, 2])
>>> xp.sort(a)
Traceback (most recent call last):
File "<stdin>", line 1, in <module>
File "/home/aaronmeurer/anaconda3/envs/array-api/lib/python3.10/site-packages/cupy/array_api/_sorting_functions.py", line 45, in sort
res = np.sort(x._array, axis=axis, kind=kind)
TypeError: sort() got an unexpected keyword argument 'kind'
```
### Installation
Conda-Forge (`conda install ...`)
### Environment
```
# Paste the output here
```
### Additional Information
The array API does require support for a `stable` keyword. If `cupy.sort` is always stable, it can just ignore this flag, as `stable=False` is allowed to be stable (see https://data-apis.org/array-api/latest/API_specification/generated/signatures.sorting_functions.sort.html). Otherwise, I would suggest raising NotImplementedError for `stable=True` (which is unfortunately the default).
</issue>
<code>
[start of cupy/array_api/_sorting_functions.py]
1 # mypy: ignore-errors
2
3 from __future__ import annotations
4
5 from ._array_object import Array
6
7 import cupy as np
8
9
10 def argsort(
11 x: Array, /, *, axis: int = -1, descending: bool = False, stable: bool = True
12 ) -> Array:
13 """
14 Array API compatible wrapper for :py:func:`np.argsort <numpy.argsort>`.
15
16 See its docstring for more information.
17 """
18 # Note: this keyword argument is different, and the default is different.
19 kind = "stable" if stable else "quicksort"
20 if not descending:
21 res = np.argsort(x._array, axis=axis, kind=kind)
22 else:
23 # As NumPy has no native descending sort, we imitate it here. Note that
24 # simply flipping the results of np.argsort(x._array, ...) would not
25 # respect the relative order like it would in native descending sorts.
26 res = np.flip(
27 np.argsort(np.flip(x._array, axis=axis), axis=axis, kind=kind),
28 axis=axis,
29 )
30 # Rely on flip()/argsort() to validate axis
31 normalised_axis = axis if axis >= 0 else x.ndim + axis
32 max_i = x.shape[normalised_axis] - 1
33 res = max_i - res
34 return Array._new(res)
35
36
37 def sort(
38 x: Array, /, *, axis: int = -1, descending: bool = False, stable: bool = True
39 ) -> Array:
40 """
41 Array API compatible wrapper for :py:func:`np.sort <numpy.sort>`.
42
43 See its docstring for more information.
44 """
45 # Note: this keyword argument is different, and the default is different.
46 kind = "stable" if stable else "quicksort"
47 res = np.sort(x._array, axis=axis, kind=kind)
48 if descending:
49 res = np.flip(res, axis=axis)
50 return Array._new(res)
51
[end of cupy/array_api/_sorting_functions.py]
[start of cupy/_sorting/sort.py]
1 import cupy
2 import numpy
3
4 from cupy.cuda import thrust
5
6
7 def sort(a, axis=-1):
8 """Returns a sorted copy of an array with a stable sorting algorithm.
9
10 Args:
11 a (cupy.ndarray): Array to be sorted.
12 axis (int or None): Axis along which to sort. Default is -1, which
13 means sort along the last axis. If None is supplied, the array is
14 flattened before sorting.
15
16 Returns:
17 cupy.ndarray: Array of the same type and shape as ``a``.
18
19 .. note::
20 For its implementation reason, ``cupy.sort`` currently does not support
21 ``kind`` and ``order`` parameters that ``numpy.sort`` does
22 support.
23
24 .. seealso:: :func:`numpy.sort`
25
26 """
27 if axis is None:
28 ret = a.flatten()
29 axis = -1
30 else:
31 ret = a.copy()
32 ret.sort(axis=axis)
33 return ret
34
35
36 def lexsort(keys):
37 """Perform an indirect sort using an array of keys.
38
39 Args:
40 keys (cupy.ndarray): ``(k, N)`` array containing ``k`` ``(N,)``-shaped
41 arrays. The ``k`` different "rows" to be sorted. The last row is
42 the primary sort key.
43
44 Returns:
45 cupy.ndarray: Array of indices that sort the keys.
46
47 .. note::
48 For its implementation reason, ``cupy.lexsort`` currently supports only
49 keys with their rank of one or two and does not support ``axis``
50 parameter that ``numpy.lexsort`` supports.
51
52 .. seealso:: :func:`numpy.lexsort`
53
54 """
55
56 # TODO(takagi): Support axis argument.
57
58 if not cupy.cuda.thrust.available:
59 raise RuntimeError('Thrust is needed to use cupy.lexsort. Please '
60 'install CUDA Toolkit with Thrust then reinstall '
61 'CuPy after uninstalling it.')
62
63 if keys.ndim == ():
64 # as numpy.lexsort() raises
65 raise TypeError('need sequence of keys with len > 0 in lexsort')
66
67 if keys.ndim == 1:
68 return 0
69
70 # TODO(takagi): Support ranks of three or more.
71 if keys.ndim > 2:
72 raise NotImplementedError('Keys with the rank of three or more is not '
73 'supported in lexsort')
74
75 # thrust.lexsort() assumes a C-contiguous array
76 if not keys.flags.c_contiguous:
77 keys = keys.copy('C')
78
79 idx_array = cupy.ndarray(keys._shape[1:], dtype=numpy.intp)
80 k = keys._shape[0]
81 n = keys._shape[1]
82 thrust.lexsort(keys.dtype, idx_array.data.ptr, keys.data.ptr, k, n)
83
84 return idx_array
85
86
87 def argsort(a, axis=-1):
88 """Returns the indices that would sort an array with a stable sorting.
89
90 Args:
91 a (cupy.ndarray): Array to sort.
92 axis (int or None): Axis along which to sort. Default is -1, which
93 means sort along the last axis. If None is supplied, the array is
94 flattened before sorting.
95
96 Returns:
97 cupy.ndarray: Array of indices that sort ``a``.
98
99 .. note::
100 For its implementation reason, ``cupy.argsort`` does not support
101 ``kind`` and ``order`` parameters.
102
103 .. seealso:: :func:`numpy.argsort`
104
105 """
106 return a.argsort(axis=axis)
107
108
109 def msort(a):
110 """Returns a copy of an array sorted along the first axis.
111
112 Args:
113 a (cupy.ndarray): Array to be sorted.
114
115 Returns:
116 cupy.ndarray: Array of the same type and shape as ``a``.
117
118 .. note:
119 ``cupy.msort(a)``, the CuPy counterpart of ``numpy.msort(a)``, is
120 equivalent to ``cupy.sort(a, axis=0)``.
121
122 .. seealso:: :func:`numpy.msort`
123
124 """
125
126 return sort(a, axis=0)
127
128
129 def sort_complex(a):
130 """Sort a complex array using the real part first,
131 then the imaginary part.
132
133 Args:
134 a (cupy.ndarray): Array to be sorted.
135
136 Returns:
137 cupy.ndarray: sorted complex array.
138
139 .. seealso:: :func:`numpy.sort_complex`
140
141 """
142 if a.dtype.char in 'bhBHF':
143 a = a.astype('F')
144 else:
145 a = a.astype('D')
146 a.sort()
147 return a
148
149
150 def partition(a, kth, axis=-1):
151 """Returns a partitioned copy of an array.
152
153 Creates a copy of the array whose elements are rearranged such that the
154 value of the element in k-th position would occur in that position in a
155 sorted array. All of the elements before the new k-th element are less
156 than or equal to the elements after the new k-th element.
157
158 Args:
159 a (cupy.ndarray): Array to be sorted.
160 kth (int or sequence of ints): Element index to partition by. If
161 supplied with a sequence of k-th it will partition all elements
162 indexed by k-th of them into their sorted position at once.
163 axis (int or None): Axis along which to sort. Default is -1, which
164 means sort along the last axis. If None is supplied, the array is
165 flattened before sorting.
166
167 Returns:
168 cupy.ndarray: Array of the same type and shape as ``a``.
169
170 .. seealso:: :func:`numpy.partition`
171
172 """
173 if axis is None:
174 ret = a.flatten()
175 axis = -1
176 else:
177 ret = a.copy()
178 ret.partition(kth, axis=axis)
179 return ret
180
181
182 def argpartition(a, kth, axis=-1):
183 """Returns the indices that would partially sort an array.
184
185 Args:
186 a (cupy.ndarray): Array to be sorted.
187 kth (int or sequence of ints): Element index to partition by. If
188 supplied with a sequence of k-th it will partition all elements
189 indexed by k-th of them into their sorted position at once.
190 axis (int or None): Axis along which to sort. Default is -1, which
191 means sort along the last axis. If None is supplied, the array is
192 flattened before sorting.
193
194 Returns:
195 cupy.ndarray: Array of the same type and shape as ``a``.
196
197 .. note::
198 For its implementation reason, `cupy.argpartition` fully sorts the
199 given array as `cupy.argsort` does. It also does not support ``kind``
200 and ``order`` parameters that ``numpy.argpartition`` supports.
201
202 .. seealso:: :func:`numpy.argpartition`
203
204 """
205 return a.argpartition(kth, axis=axis)
206
[end of cupy/_sorting/sort.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/cupy/_sorting/sort.py b/cupy/_sorting/sort.py
--- a/cupy/_sorting/sort.py
+++ b/cupy/_sorting/sort.py
@@ -4,7 +4,7 @@
from cupy.cuda import thrust
-def sort(a, axis=-1):
+def sort(a, axis=-1, kind=None):
"""Returns a sorted copy of an array with a stable sorting algorithm.
Args:
@@ -12,6 +12,8 @@
axis (int or None): Axis along which to sort. Default is -1, which
means sort along the last axis. If None is supplied, the array is
flattened before sorting.
+ kind: Default is `None`, which is equivalent to 'stable'. Unlike in
+ NumPy any other options are not accepted here.
Returns:
cupy.ndarray: Array of the same type and shape as ``a``.
@@ -24,6 +26,8 @@
.. seealso:: :func:`numpy.sort`
"""
+ if kind is not None and kind != 'stable':
+ raise ValueError("kind can only be None or 'stable'")
if axis is None:
ret = a.flatten()
axis = -1
@@ -84,7 +88,7 @@
return idx_array
-def argsort(a, axis=-1):
+def argsort(a, axis=-1, kind=None):
"""Returns the indices that would sort an array with a stable sorting.
Args:
@@ -92,6 +96,8 @@
axis (int or None): Axis along which to sort. Default is -1, which
means sort along the last axis. If None is supplied, the array is
flattened before sorting.
+ kind: Default is `None`, which is equivalent to 'stable'. Unlike in
+ NumPy any other options are not accepted here.
Returns:
cupy.ndarray: Array of indices that sort ``a``.
@@ -103,6 +109,8 @@
.. seealso:: :func:`numpy.argsort`
"""
+ if kind is not None and kind != 'stable':
+ raise ValueError("kind can only be None or 'stable'")
return a.argsort(axis=axis)
diff --git a/cupy/array_api/_sorting_functions.py b/cupy/array_api/_sorting_functions.py
--- a/cupy/array_api/_sorting_functions.py
+++ b/cupy/array_api/_sorting_functions.py
@@ -15,8 +15,9 @@
See its docstring for more information.
"""
- # Note: this keyword argument is different, and the default is different.
- kind = "stable" if stable else "quicksort"
+ # Note: Unlike in NumPy we only support kind={None, 'stable'}, but the standard
+ # does *not* require we need to support unstable sort.
+ kind = None
if not descending:
res = np.argsort(x._array, axis=axis, kind=kind)
else:
@@ -42,8 +43,9 @@
See its docstring for more information.
"""
- # Note: this keyword argument is different, and the default is different.
- kind = "stable" if stable else "quicksort"
+ # Note: Unlike in NumPy we only support kind={None, 'stable'}, but the standard
+ # does *not* require we need to support unstable sort.
+ kind = None
res = np.sort(x._array, axis=axis, kind=kind)
if descending:
res = np.flip(res, axis=axis)
|
{"golden_diff": "diff --git a/cupy/_sorting/sort.py b/cupy/_sorting/sort.py\n--- a/cupy/_sorting/sort.py\n+++ b/cupy/_sorting/sort.py\n@@ -4,7 +4,7 @@\n from cupy.cuda import thrust\n \n \n-def sort(a, axis=-1):\n+def sort(a, axis=-1, kind=None):\n \"\"\"Returns a sorted copy of an array with a stable sorting algorithm.\n \n Args:\n@@ -12,6 +12,8 @@\n axis (int or None): Axis along which to sort. Default is -1, which\n means sort along the last axis. If None is supplied, the array is\n flattened before sorting.\n+ kind: Default is `None`, which is equivalent to 'stable'. Unlike in\n+ NumPy any other options are not accepted here.\n \n Returns:\n cupy.ndarray: Array of the same type and shape as ``a``.\n@@ -24,6 +26,8 @@\n .. seealso:: :func:`numpy.sort`\n \n \"\"\"\n+ if kind is not None and kind != 'stable':\n+ raise ValueError(\"kind can only be None or 'stable'\")\n if axis is None:\n ret = a.flatten()\n axis = -1\n@@ -84,7 +88,7 @@\n return idx_array\n \n \n-def argsort(a, axis=-1):\n+def argsort(a, axis=-1, kind=None):\n \"\"\"Returns the indices that would sort an array with a stable sorting.\n \n Args:\n@@ -92,6 +96,8 @@\n axis (int or None): Axis along which to sort. Default is -1, which\n means sort along the last axis. If None is supplied, the array is\n flattened before sorting.\n+ kind: Default is `None`, which is equivalent to 'stable'. Unlike in\n+ NumPy any other options are not accepted here.\n \n Returns:\n cupy.ndarray: Array of indices that sort ``a``.\n@@ -103,6 +109,8 @@\n .. seealso:: :func:`numpy.argsort`\n \n \"\"\"\n+ if kind is not None and kind != 'stable':\n+ raise ValueError(\"kind can only be None or 'stable'\")\n return a.argsort(axis=axis)\n \n \ndiff --git a/cupy/array_api/_sorting_functions.py b/cupy/array_api/_sorting_functions.py\n--- a/cupy/array_api/_sorting_functions.py\n+++ b/cupy/array_api/_sorting_functions.py\n@@ -15,8 +15,9 @@\n \n See its docstring for more information.\n \"\"\"\n- # Note: this keyword argument is different, and the default is different.\n- kind = \"stable\" if stable else \"quicksort\"\n+ # Note: Unlike in NumPy we only support kind={None, 'stable'}, but the standard\n+ # does *not* require we need to support unstable sort.\n+ kind = None\n if not descending:\n res = np.argsort(x._array, axis=axis, kind=kind)\n else:\n@@ -42,8 +43,9 @@\n \n See its docstring for more information.\n \"\"\"\n- # Note: this keyword argument is different, and the default is different.\n- kind = \"stable\" if stable else \"quicksort\"\n+ # Note: Unlike in NumPy we only support kind={None, 'stable'}, but the standard\n+ # does *not* require we need to support unstable sort.\n+ kind = None\n res = np.sort(x._array, axis=axis, kind=kind)\n if descending:\n res = np.flip(res, axis=axis)\n", "issue": "array_api sort() does not work\n### Description\n\nThe array_api sort() is copied from the NumPy implementation, but passes through a `kind` keyword which doesn't exist in `cupy.sort`.\n\n### To Reproduce\n\n```py\r\n>>> a = xp.asarray([1, 0, 2])\r\n>>> xp.sort(a)\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"/home/aaronmeurer/anaconda3/envs/array-api/lib/python3.10/site-packages/cupy/array_api/_sorting_functions.py\", line 45, in sort\r\n res = np.sort(x._array, axis=axis, kind=kind)\r\nTypeError: sort() got an unexpected keyword argument 'kind'\r\n```\r\n\n\n### Installation\n\nConda-Forge (`conda install ...`)\n\n### Environment\n\n```\r\n# Paste the output here\r\n```\r\n\n\n### Additional Information\n\nThe array API does require support for a `stable` keyword. If `cupy.sort` is always stable, it can just ignore this flag, as `stable=False` is allowed to be stable (see https://data-apis.org/array-api/latest/API_specification/generated/signatures.sorting_functions.sort.html). Otherwise, I would suggest raising NotImplementedError for `stable=True` (which is unfortunately the default). \n", "before_files": [{"content": "# mypy: ignore-errors\n\nfrom __future__ import annotations\n\nfrom ._array_object import Array\n\nimport cupy as np\n\n\ndef argsort(\n x: Array, /, *, axis: int = -1, descending: bool = False, stable: bool = True\n) -> Array:\n \"\"\"\n Array API compatible wrapper for :py:func:`np.argsort <numpy.argsort>`.\n\n See its docstring for more information.\n \"\"\"\n # Note: this keyword argument is different, and the default is different.\n kind = \"stable\" if stable else \"quicksort\"\n if not descending:\n res = np.argsort(x._array, axis=axis, kind=kind)\n else:\n # As NumPy has no native descending sort, we imitate it here. Note that\n # simply flipping the results of np.argsort(x._array, ...) would not\n # respect the relative order like it would in native descending sorts.\n res = np.flip(\n np.argsort(np.flip(x._array, axis=axis), axis=axis, kind=kind),\n axis=axis,\n )\n # Rely on flip()/argsort() to validate axis\n normalised_axis = axis if axis >= 0 else x.ndim + axis\n max_i = x.shape[normalised_axis] - 1\n res = max_i - res\n return Array._new(res)\n\n\ndef sort(\n x: Array, /, *, axis: int = -1, descending: bool = False, stable: bool = True\n) -> Array:\n \"\"\"\n Array API compatible wrapper for :py:func:`np.sort <numpy.sort>`.\n\n See its docstring for more information.\n \"\"\"\n # Note: this keyword argument is different, and the default is different.\n kind = \"stable\" if stable else \"quicksort\"\n res = np.sort(x._array, axis=axis, kind=kind)\n if descending:\n res = np.flip(res, axis=axis)\n return Array._new(res)\n", "path": "cupy/array_api/_sorting_functions.py"}, {"content": "import cupy\nimport numpy\n\nfrom cupy.cuda import thrust\n\n\ndef sort(a, axis=-1):\n \"\"\"Returns a sorted copy of an array with a stable sorting algorithm.\n\n Args:\n a (cupy.ndarray): Array to be sorted.\n axis (int or None): Axis along which to sort. Default is -1, which\n means sort along the last axis. If None is supplied, the array is\n flattened before sorting.\n\n Returns:\n cupy.ndarray: Array of the same type and shape as ``a``.\n\n .. note::\n For its implementation reason, ``cupy.sort`` currently does not support\n ``kind`` and ``order`` parameters that ``numpy.sort`` does\n support.\n\n .. seealso:: :func:`numpy.sort`\n\n \"\"\"\n if axis is None:\n ret = a.flatten()\n axis = -1\n else:\n ret = a.copy()\n ret.sort(axis=axis)\n return ret\n\n\ndef lexsort(keys):\n \"\"\"Perform an indirect sort using an array of keys.\n\n Args:\n keys (cupy.ndarray): ``(k, N)`` array containing ``k`` ``(N,)``-shaped\n arrays. The ``k`` different \"rows\" to be sorted. The last row is\n the primary sort key.\n\n Returns:\n cupy.ndarray: Array of indices that sort the keys.\n\n .. note::\n For its implementation reason, ``cupy.lexsort`` currently supports only\n keys with their rank of one or two and does not support ``axis``\n parameter that ``numpy.lexsort`` supports.\n\n .. seealso:: :func:`numpy.lexsort`\n\n \"\"\"\n\n # TODO(takagi): Support axis argument.\n\n if not cupy.cuda.thrust.available:\n raise RuntimeError('Thrust is needed to use cupy.lexsort. Please '\n 'install CUDA Toolkit with Thrust then reinstall '\n 'CuPy after uninstalling it.')\n\n if keys.ndim == ():\n # as numpy.lexsort() raises\n raise TypeError('need sequence of keys with len > 0 in lexsort')\n\n if keys.ndim == 1:\n return 0\n\n # TODO(takagi): Support ranks of three or more.\n if keys.ndim > 2:\n raise NotImplementedError('Keys with the rank of three or more is not '\n 'supported in lexsort')\n\n # thrust.lexsort() assumes a C-contiguous array\n if not keys.flags.c_contiguous:\n keys = keys.copy('C')\n\n idx_array = cupy.ndarray(keys._shape[1:], dtype=numpy.intp)\n k = keys._shape[0]\n n = keys._shape[1]\n thrust.lexsort(keys.dtype, idx_array.data.ptr, keys.data.ptr, k, n)\n\n return idx_array\n\n\ndef argsort(a, axis=-1):\n \"\"\"Returns the indices that would sort an array with a stable sorting.\n\n Args:\n a (cupy.ndarray): Array to sort.\n axis (int or None): Axis along which to sort. Default is -1, which\n means sort along the last axis. If None is supplied, the array is\n flattened before sorting.\n\n Returns:\n cupy.ndarray: Array of indices that sort ``a``.\n\n .. note::\n For its implementation reason, ``cupy.argsort`` does not support\n ``kind`` and ``order`` parameters.\n\n .. seealso:: :func:`numpy.argsort`\n\n \"\"\"\n return a.argsort(axis=axis)\n\n\ndef msort(a):\n \"\"\"Returns a copy of an array sorted along the first axis.\n\n Args:\n a (cupy.ndarray): Array to be sorted.\n\n Returns:\n cupy.ndarray: Array of the same type and shape as ``a``.\n\n .. note:\n ``cupy.msort(a)``, the CuPy counterpart of ``numpy.msort(a)``, is\n equivalent to ``cupy.sort(a, axis=0)``.\n\n .. seealso:: :func:`numpy.msort`\n\n \"\"\"\n\n return sort(a, axis=0)\n\n\ndef sort_complex(a):\n \"\"\"Sort a complex array using the real part first,\n then the imaginary part.\n\n Args:\n a (cupy.ndarray): Array to be sorted.\n\n Returns:\n cupy.ndarray: sorted complex array.\n\n .. seealso:: :func:`numpy.sort_complex`\n\n \"\"\"\n if a.dtype.char in 'bhBHF':\n a = a.astype('F')\n else:\n a = a.astype('D')\n a.sort()\n return a\n\n\ndef partition(a, kth, axis=-1):\n \"\"\"Returns a partitioned copy of an array.\n\n Creates a copy of the array whose elements are rearranged such that the\n value of the element in k-th position would occur in that position in a\n sorted array. All of the elements before the new k-th element are less\n than or equal to the elements after the new k-th element.\n\n Args:\n a (cupy.ndarray): Array to be sorted.\n kth (int or sequence of ints): Element index to partition by. If\n supplied with a sequence of k-th it will partition all elements\n indexed by k-th of them into their sorted position at once.\n axis (int or None): Axis along which to sort. Default is -1, which\n means sort along the last axis. If None is supplied, the array is\n flattened before sorting.\n\n Returns:\n cupy.ndarray: Array of the same type and shape as ``a``.\n\n .. seealso:: :func:`numpy.partition`\n\n \"\"\"\n if axis is None:\n ret = a.flatten()\n axis = -1\n else:\n ret = a.copy()\n ret.partition(kth, axis=axis)\n return ret\n\n\ndef argpartition(a, kth, axis=-1):\n \"\"\"Returns the indices that would partially sort an array.\n\n Args:\n a (cupy.ndarray): Array to be sorted.\n kth (int or sequence of ints): Element index to partition by. If\n supplied with a sequence of k-th it will partition all elements\n indexed by k-th of them into their sorted position at once.\n axis (int or None): Axis along which to sort. Default is -1, which\n means sort along the last axis. If None is supplied, the array is\n flattened before sorting.\n\n Returns:\n cupy.ndarray: Array of the same type and shape as ``a``.\n\n .. note::\n For its implementation reason, `cupy.argpartition` fully sorts the\n given array as `cupy.argsort` does. It also does not support ``kind``\n and ``order`` parameters that ``numpy.argpartition`` supports.\n\n .. seealso:: :func:`numpy.argpartition`\n\n \"\"\"\n return a.argpartition(kth, axis=axis)\n", "path": "cupy/_sorting/sort.py"}]}
| 3,386 | 801 |
gh_patches_debug_2992
|
rasdani/github-patches
|
git_diff
|
freedomofpress__securedrop-6408
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Test securedrop-admin with Tails 5.0
## Description
https://tails.boum.org/news/test_5.0-beta1/
Tails 5.0 is based on Debian Bullseye, which means it's using a newer Python version (3.9) among plenty of other things.
It's probably worth walking through a full SD install + backup/restore to make sure it works as expected.
</issue>
<code>
[start of admin/bootstrap.py]
1 # -*- mode: python; coding: utf-8 -*-
2 #
3 # Copyright (C) 2013-2018 Freedom of the Press Foundation & al
4 # Copyright (C) 2018 Loic Dachary <[email protected]>
5 #
6 # This program is free software: you can redistribute it and/or modify
7 # it under the terms of the GNU General Public License as published by
8 # the Free Software Foundation, either version 3 of the License, or
9 # (at your option) any later version.
10 #
11 # This program is distributed in the hope that it will be useful,
12 # but WITHOUT ANY WARRANTY; without even the implied warranty of
13 # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
14 # GNU General Public License for more details.
15 #
16 # You should have received a copy of the GNU General Public License
17 # along with this program. If not, see <http://www.gnu.org/licenses/>.
18 #
19
20 import argparse
21 import logging
22 import os
23 import shutil
24 import subprocess
25 import sys
26 from typing import Iterator
27
28 from typing import List
29
30 sdlog = logging.getLogger(__name__)
31
32 DIR = os.path.dirname(os.path.realpath(__file__))
33 VENV_DIR = os.path.join(DIR, ".venv3")
34
35
36 def setup_logger(verbose: bool = False) -> None:
37 """ Configure logging handler """
38 # Set default level on parent
39 sdlog.setLevel(logging.DEBUG)
40 level = logging.DEBUG if verbose else logging.INFO
41
42 stdout = logging.StreamHandler(sys.stdout)
43 stdout.setFormatter(logging.Formatter('%(levelname)s: %(message)s'))
44 stdout.setLevel(level)
45 sdlog.addHandler(stdout)
46
47
48 def run_command(command: List[str]) -> Iterator[bytes]:
49 """
50 Wrapper function to display stdout for running command,
51 similar to how shelling out in a Bash script displays rolling output.
52
53 Yields a list of the stdout from the `command`, and raises a
54 CalledProcessError if `command` returns non-zero.
55 """
56 popen = subprocess.Popen(command,
57 stdout=subprocess.PIPE,
58 stderr=subprocess.STDOUT)
59 if popen.stdout is None:
60 raise EnvironmentError("Could not run command: None stdout")
61 for stdout_line in iter(popen.stdout.readline, b""):
62 yield stdout_line
63 popen.stdout.close()
64 return_code = popen.wait()
65 if return_code:
66 raise subprocess.CalledProcessError(return_code, command)
67
68
69 def is_tails() -> bool:
70 try:
71 id = subprocess.check_output('lsb_release --id --short',
72 shell=True).decode('utf-8').strip()
73 except subprocess.CalledProcessError:
74 return False
75
76 # dirty hack to unreliably detect Tails 4.0~beta2
77 if id == 'Debian':
78 if os.uname()[1] == 'amnesia':
79 id = 'Tails'
80
81 return id == 'Tails'
82
83
84 def clean_up_tails3_venv(virtualenv_dir: str = VENV_DIR) -> None:
85 """
86 Tails 3.x, based on debian stretch uses libpython3.5, whereas Tails 4.x is
87 based on Debian Buster and uses libpython3.7. This means that the Tails 3.x
88 virtualenv will not work under Tails 4.x, and will need to be destroyed and
89 rebuilt. We can detect if the version of libpython is 3.5 in the
90 admin/.venv3/ folder, and delete it if that's the case. This will ensure a
91 smooth upgrade from Tails 3.x to Tails 4.x.
92 """
93 if is_tails():
94 try:
95 dist = subprocess.check_output('lsb_release --codename --short',
96 shell=True).strip()
97 except subprocess.CalledProcessError:
98 return None
99
100 # tails4 is based on buster
101 if dist == b'buster':
102 python_lib_path = os.path.join(virtualenv_dir, "lib/python3.5")
103 if os.path.exists(os.path.join(python_lib_path)):
104 sdlog.info(
105 "Tails 3 Python 3 virtualenv detected. "
106 "Removing it."
107 )
108 shutil.rmtree(virtualenv_dir)
109 sdlog.info("Tails 3 Python 3 virtualenv deleted.")
110
111
112 def checkenv(args: argparse.Namespace) -> None:
113 clean_up_tails3_venv(VENV_DIR)
114 if not os.path.exists(os.path.join(VENV_DIR, "bin/activate")):
115 sdlog.error('Please run "securedrop-admin setup".')
116 sys.exit(1)
117
118
119 def maybe_torify() -> List[str]:
120 if is_tails():
121 return ['torify']
122 else:
123 return []
124
125
126 def install_apt_dependencies(args: argparse.Namespace) -> None:
127 """
128 Install apt dependencies in Tails. In order to install Ansible in
129 a virtualenv, first there are a number of Python prerequisites.
130 """
131 sdlog.info("Installing SecureDrop Admin dependencies")
132 sdlog.info(("You'll be prompted for the temporary Tails admin password,"
133 " which was set on Tails login screen"))
134
135 apt_command = ['sudo', 'su', '-c',
136 "apt-get update && \
137 apt-get -q -o=Dpkg::Use-Pty=0 install -y \
138 python3-virtualenv \
139 python3-yaml \
140 python3-pip \
141 ccontrol \
142 virtualenv \
143 libffi-dev \
144 libssl-dev \
145 libpython3-dev",
146 ]
147
148 try:
149 # Print command results in real-time, to keep Admin apprised
150 # of progress during long-running command.
151 for output_line in run_command(apt_command):
152 print(output_line.decode('utf-8').rstrip())
153 except subprocess.CalledProcessError:
154 # Tails supports apt persistence, which was used by SecureDrop
155 # under Tails 2.x. If updates are being applied, don't try to pile
156 # on with more apt requests.
157 sdlog.error(("Failed to install apt dependencies. Check network"
158 " connection and try again."))
159 raise
160
161
162 def envsetup(args: argparse.Namespace, virtualenv_dir: str = VENV_DIR) -> None:
163 """Installs Admin tooling required for managing SecureDrop. Specifically:
164
165 * updates apt-cache
166 * installs apt packages for Python virtualenv
167 * creates virtualenv
168 * installs pip packages inside virtualenv
169
170 The virtualenv is created within the Persistence volume in Tails, so that
171 Ansible is available to the Admin on subsequent boots without requiring
172 installation of packages again.
173 """
174 # clean up Tails 3.x venv when migrating to Tails 4.x
175 clean_up_tails3_venv(virtualenv_dir)
176
177 # virtualenv doesnt exist? Install dependencies and create
178 if not os.path.exists(virtualenv_dir):
179
180 install_apt_dependencies(args)
181
182 # Technically you can create a virtualenv from within python
183 # but pip can only be run over Tor on Tails, and debugging that
184 # along with instaling a third-party dependency is not worth
185 # the effort here.
186 sdlog.info("Setting up virtualenv")
187 try:
188 sdlog.debug(subprocess.check_output(
189 maybe_torify() + ['virtualenv',
190 '--python=python3',
191 virtualenv_dir
192 ],
193 stderr=subprocess.STDOUT))
194 except subprocess.CalledProcessError as e:
195 sdlog.debug(e.output)
196 sdlog.error(("Unable to create virtualenv. Check network settings"
197 " and try again."))
198 sdlog.debug("Cleaning up virtualenv")
199 if os.path.exists(virtualenv_dir):
200 shutil.rmtree(virtualenv_dir)
201 raise
202 else:
203 sdlog.info("Virtualenv already exists, not creating")
204
205 if args.t:
206 install_pip_dependencies(
207 args,
208 requirements_file='requirements-testinfra.txt',
209 desc="dependencies with verification support"
210 )
211 else:
212 install_pip_dependencies(args)
213
214 if os.path.exists(os.path.join(DIR, 'setup.py')):
215 install_pip_self(args)
216
217 sdlog.info("Finished installing SecureDrop dependencies")
218
219
220 def install_pip_self(args: argparse.Namespace) -> None:
221 pip_install_cmd = [
222 os.path.join(VENV_DIR, 'bin', 'pip3'),
223 'install', '-e', DIR
224 ]
225 try:
226 subprocess.check_output(maybe_torify() + pip_install_cmd,
227 stderr=subprocess.STDOUT)
228 except subprocess.CalledProcessError as e:
229 sdlog.debug(e.output)
230 sdlog.error("Unable to install self, run with -v for more information")
231 raise
232
233
234 def install_pip_dependencies(
235 args: argparse.Namespace,
236 requirements_file: str = "requirements.txt",
237 desc: str = "Python dependencies",
238 ) -> None:
239 """
240 Install Python dependencies via pip into virtualenv.
241 """
242 pip_install_cmd = [
243 os.path.join(VENV_DIR, 'bin', 'pip3'),
244 'install',
245 '--no-deps',
246 '-r', os.path.join(DIR, requirements_file),
247 '--require-hashes',
248 '-U', '--upgrade-strategy', 'only-if-needed',
249 ]
250
251 sdlog.info("Checking {} for securedrop-admin".format(desc))
252 try:
253 pip_output = subprocess.check_output(maybe_torify() + pip_install_cmd,
254 stderr=subprocess.STDOUT)
255 except subprocess.CalledProcessError as e:
256 sdlog.debug(e.output)
257 sdlog.error(("Failed to install {}. Check network"
258 " connection and try again.".format(desc)))
259 raise
260
261 sdlog.debug(pip_output)
262 if "Successfully installed" in str(pip_output):
263 sdlog.info("{} for securedrop-admin upgraded".format(desc))
264 else:
265 sdlog.info("{} for securedrop-admin are up-to-date".format(desc))
266
267
268 def parse_argv(argv: List[str]) -> argparse.Namespace:
269 parser = argparse.ArgumentParser()
270 parser.add_argument('-v', action='store_true', default=False,
271 help="Increase verbosity on output")
272 parser.add_argument('-t', action='store_true', default=False,
273 help="Install additional test dependencies")
274 parser.set_defaults(func=envsetup)
275
276 subparsers = parser.add_subparsers()
277
278 envsetup_parser = subparsers.add_parser(
279 'envsetup',
280 help='Set up the admin virtualenv.'
281 )
282 envsetup_parser.set_defaults(func=envsetup)
283
284 checkenv_parser = subparsers.add_parser(
285 'checkenv',
286 help='Check that the admin virtualenv is properly set up.'
287 )
288 checkenv_parser.set_defaults(func=checkenv)
289
290 return parser.parse_args(argv)
291
292
293 if __name__ == "__main__":
294 args = parse_argv(sys.argv[1:])
295 setup_logger(args.v)
296
297 try:
298 args.func(args)
299 except Exception:
300 sys.exit(1)
301 else:
302 sys.exit(0)
303
[end of admin/bootstrap.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/admin/bootstrap.py b/admin/bootstrap.py
--- a/admin/bootstrap.py
+++ b/admin/bootstrap.py
@@ -138,7 +138,6 @@
python3-virtualenv \
python3-yaml \
python3-pip \
- ccontrol \
virtualenv \
libffi-dev \
libssl-dev \
|
{"golden_diff": "diff --git a/admin/bootstrap.py b/admin/bootstrap.py\n--- a/admin/bootstrap.py\n+++ b/admin/bootstrap.py\n@@ -138,7 +138,6 @@\n python3-virtualenv \\\n python3-yaml \\\n python3-pip \\\n- ccontrol \\\n virtualenv \\\n libffi-dev \\\n libssl-dev \\\n", "issue": "Test securedrop-admin with Tails 5.0\n## Description\r\n\r\nhttps://tails.boum.org/news/test_5.0-beta1/\r\n\r\nTails 5.0 is based on Debian Bullseye, which means it's using a newer Python version (3.9) among plenty of other things.\r\n\r\nIt's probably worth walking through a full SD install + backup/restore to make sure it works as expected.\n", "before_files": [{"content": "# -*- mode: python; coding: utf-8 -*-\n#\n# Copyright (C) 2013-2018 Freedom of the Press Foundation & al\n# Copyright (C) 2018 Loic Dachary <[email protected]>\n#\n# This program is free software: you can redistribute it and/or modify\n# it under the terms of the GNU General Public License as published by\n# the Free Software Foundation, either version 3 of the License, or\n# (at your option) any later version.\n#\n# This program is distributed in the hope that it will be useful,\n# but WITHOUT ANY WARRANTY; without even the implied warranty of\n# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the\n# GNU General Public License for more details.\n#\n# You should have received a copy of the GNU General Public License\n# along with this program. If not, see <http://www.gnu.org/licenses/>.\n#\n\nimport argparse\nimport logging\nimport os\nimport shutil\nimport subprocess\nimport sys\nfrom typing import Iterator\n\nfrom typing import List\n\nsdlog = logging.getLogger(__name__)\n\nDIR = os.path.dirname(os.path.realpath(__file__))\nVENV_DIR = os.path.join(DIR, \".venv3\")\n\n\ndef setup_logger(verbose: bool = False) -> None:\n \"\"\" Configure logging handler \"\"\"\n # Set default level on parent\n sdlog.setLevel(logging.DEBUG)\n level = logging.DEBUG if verbose else logging.INFO\n\n stdout = logging.StreamHandler(sys.stdout)\n stdout.setFormatter(logging.Formatter('%(levelname)s: %(message)s'))\n stdout.setLevel(level)\n sdlog.addHandler(stdout)\n\n\ndef run_command(command: List[str]) -> Iterator[bytes]:\n \"\"\"\n Wrapper function to display stdout for running command,\n similar to how shelling out in a Bash script displays rolling output.\n\n Yields a list of the stdout from the `command`, and raises a\n CalledProcessError if `command` returns non-zero.\n \"\"\"\n popen = subprocess.Popen(command,\n stdout=subprocess.PIPE,\n stderr=subprocess.STDOUT)\n if popen.stdout is None:\n raise EnvironmentError(\"Could not run command: None stdout\")\n for stdout_line in iter(popen.stdout.readline, b\"\"):\n yield stdout_line\n popen.stdout.close()\n return_code = popen.wait()\n if return_code:\n raise subprocess.CalledProcessError(return_code, command)\n\n\ndef is_tails() -> bool:\n try:\n id = subprocess.check_output('lsb_release --id --short',\n shell=True).decode('utf-8').strip()\n except subprocess.CalledProcessError:\n return False\n\n # dirty hack to unreliably detect Tails 4.0~beta2\n if id == 'Debian':\n if os.uname()[1] == 'amnesia':\n id = 'Tails'\n\n return id == 'Tails'\n\n\ndef clean_up_tails3_venv(virtualenv_dir: str = VENV_DIR) -> None:\n \"\"\"\n Tails 3.x, based on debian stretch uses libpython3.5, whereas Tails 4.x is\n based on Debian Buster and uses libpython3.7. This means that the Tails 3.x\n virtualenv will not work under Tails 4.x, and will need to be destroyed and\n rebuilt. We can detect if the version of libpython is 3.5 in the\n admin/.venv3/ folder, and delete it if that's the case. This will ensure a\n smooth upgrade from Tails 3.x to Tails 4.x.\n \"\"\"\n if is_tails():\n try:\n dist = subprocess.check_output('lsb_release --codename --short',\n shell=True).strip()\n except subprocess.CalledProcessError:\n return None\n\n # tails4 is based on buster\n if dist == b'buster':\n python_lib_path = os.path.join(virtualenv_dir, \"lib/python3.5\")\n if os.path.exists(os.path.join(python_lib_path)):\n sdlog.info(\n \"Tails 3 Python 3 virtualenv detected. \"\n \"Removing it.\"\n )\n shutil.rmtree(virtualenv_dir)\n sdlog.info(\"Tails 3 Python 3 virtualenv deleted.\")\n\n\ndef checkenv(args: argparse.Namespace) -> None:\n clean_up_tails3_venv(VENV_DIR)\n if not os.path.exists(os.path.join(VENV_DIR, \"bin/activate\")):\n sdlog.error('Please run \"securedrop-admin setup\".')\n sys.exit(1)\n\n\ndef maybe_torify() -> List[str]:\n if is_tails():\n return ['torify']\n else:\n return []\n\n\ndef install_apt_dependencies(args: argparse.Namespace) -> None:\n \"\"\"\n Install apt dependencies in Tails. In order to install Ansible in\n a virtualenv, first there are a number of Python prerequisites.\n \"\"\"\n sdlog.info(\"Installing SecureDrop Admin dependencies\")\n sdlog.info((\"You'll be prompted for the temporary Tails admin password,\"\n \" which was set on Tails login screen\"))\n\n apt_command = ['sudo', 'su', '-c',\n \"apt-get update && \\\n apt-get -q -o=Dpkg::Use-Pty=0 install -y \\\n python3-virtualenv \\\n python3-yaml \\\n python3-pip \\\n ccontrol \\\n virtualenv \\\n libffi-dev \\\n libssl-dev \\\n libpython3-dev\",\n ]\n\n try:\n # Print command results in real-time, to keep Admin apprised\n # of progress during long-running command.\n for output_line in run_command(apt_command):\n print(output_line.decode('utf-8').rstrip())\n except subprocess.CalledProcessError:\n # Tails supports apt persistence, which was used by SecureDrop\n # under Tails 2.x. If updates are being applied, don't try to pile\n # on with more apt requests.\n sdlog.error((\"Failed to install apt dependencies. Check network\"\n \" connection and try again.\"))\n raise\n\n\ndef envsetup(args: argparse.Namespace, virtualenv_dir: str = VENV_DIR) -> None:\n \"\"\"Installs Admin tooling required for managing SecureDrop. Specifically:\n\n * updates apt-cache\n * installs apt packages for Python virtualenv\n * creates virtualenv\n * installs pip packages inside virtualenv\n\n The virtualenv is created within the Persistence volume in Tails, so that\n Ansible is available to the Admin on subsequent boots without requiring\n installation of packages again.\n \"\"\"\n # clean up Tails 3.x venv when migrating to Tails 4.x\n clean_up_tails3_venv(virtualenv_dir)\n\n # virtualenv doesnt exist? Install dependencies and create\n if not os.path.exists(virtualenv_dir):\n\n install_apt_dependencies(args)\n\n # Technically you can create a virtualenv from within python\n # but pip can only be run over Tor on Tails, and debugging that\n # along with instaling a third-party dependency is not worth\n # the effort here.\n sdlog.info(\"Setting up virtualenv\")\n try:\n sdlog.debug(subprocess.check_output(\n maybe_torify() + ['virtualenv',\n '--python=python3',\n virtualenv_dir\n ],\n stderr=subprocess.STDOUT))\n except subprocess.CalledProcessError as e:\n sdlog.debug(e.output)\n sdlog.error((\"Unable to create virtualenv. Check network settings\"\n \" and try again.\"))\n sdlog.debug(\"Cleaning up virtualenv\")\n if os.path.exists(virtualenv_dir):\n shutil.rmtree(virtualenv_dir)\n raise\n else:\n sdlog.info(\"Virtualenv already exists, not creating\")\n\n if args.t:\n install_pip_dependencies(\n args,\n requirements_file='requirements-testinfra.txt',\n desc=\"dependencies with verification support\"\n )\n else:\n install_pip_dependencies(args)\n\n if os.path.exists(os.path.join(DIR, 'setup.py')):\n install_pip_self(args)\n\n sdlog.info(\"Finished installing SecureDrop dependencies\")\n\n\ndef install_pip_self(args: argparse.Namespace) -> None:\n pip_install_cmd = [\n os.path.join(VENV_DIR, 'bin', 'pip3'),\n 'install', '-e', DIR\n ]\n try:\n subprocess.check_output(maybe_torify() + pip_install_cmd,\n stderr=subprocess.STDOUT)\n except subprocess.CalledProcessError as e:\n sdlog.debug(e.output)\n sdlog.error(\"Unable to install self, run with -v for more information\")\n raise\n\n\ndef install_pip_dependencies(\n args: argparse.Namespace,\n requirements_file: str = \"requirements.txt\",\n desc: str = \"Python dependencies\",\n) -> None:\n \"\"\"\n Install Python dependencies via pip into virtualenv.\n \"\"\"\n pip_install_cmd = [\n os.path.join(VENV_DIR, 'bin', 'pip3'),\n 'install',\n '--no-deps',\n '-r', os.path.join(DIR, requirements_file),\n '--require-hashes',\n '-U', '--upgrade-strategy', 'only-if-needed',\n ]\n\n sdlog.info(\"Checking {} for securedrop-admin\".format(desc))\n try:\n pip_output = subprocess.check_output(maybe_torify() + pip_install_cmd,\n stderr=subprocess.STDOUT)\n except subprocess.CalledProcessError as e:\n sdlog.debug(e.output)\n sdlog.error((\"Failed to install {}. Check network\"\n \" connection and try again.\".format(desc)))\n raise\n\n sdlog.debug(pip_output)\n if \"Successfully installed\" in str(pip_output):\n sdlog.info(\"{} for securedrop-admin upgraded\".format(desc))\n else:\n sdlog.info(\"{} for securedrop-admin are up-to-date\".format(desc))\n\n\ndef parse_argv(argv: List[str]) -> argparse.Namespace:\n parser = argparse.ArgumentParser()\n parser.add_argument('-v', action='store_true', default=False,\n help=\"Increase verbosity on output\")\n parser.add_argument('-t', action='store_true', default=False,\n help=\"Install additional test dependencies\")\n parser.set_defaults(func=envsetup)\n\n subparsers = parser.add_subparsers()\n\n envsetup_parser = subparsers.add_parser(\n 'envsetup',\n help='Set up the admin virtualenv.'\n )\n envsetup_parser.set_defaults(func=envsetup)\n\n checkenv_parser = subparsers.add_parser(\n 'checkenv',\n help='Check that the admin virtualenv is properly set up.'\n )\n checkenv_parser.set_defaults(func=checkenv)\n\n return parser.parse_args(argv)\n\n\nif __name__ == \"__main__\":\n args = parse_argv(sys.argv[1:])\n setup_logger(args.v)\n\n try:\n args.func(args)\n except Exception:\n sys.exit(1)\n else:\n sys.exit(0)\n", "path": "admin/bootstrap.py"}]}
| 3,794 | 76 |
gh_patches_debug_11992
|
rasdani/github-patches
|
git_diff
|
Cloud-CV__EvalAI-674
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Hosts: User cannot see all of his Host Teams
## CURRENT STATE -
`challenge-host-teams` page only shows teams `created_by` the user.
## ISSUE -
On `challenge-host-teams` page, user should be able to see all host teams he is a part of.
## RELATED ISSUE -
#607 Refer to this issue also.
</issue>
<code>
[start of apps/hosts/views.py]
1 from rest_framework import permissions, status
2 from rest_framework.decorators import (api_view,
3 authentication_classes,
4 permission_classes,
5 throttle_classes,)
6 from rest_framework.response import Response
7 from rest_framework_expiring_authtoken.authentication import (ExpiringTokenAuthentication,)
8 from rest_framework.throttling import UserRateThrottle
9
10 from accounts.permissions import HasVerifiedEmail
11 from base.utils import paginated_queryset
12 from .models import (ChallengeHost,
13 ChallengeHostTeam,)
14 from .serializers import (ChallengeHostSerializer,
15 ChallengeHostTeamSerializer,
16 InviteHostToTeamSerializer,
17 HostTeamDetailSerializer,)
18
19
20 @throttle_classes([UserRateThrottle])
21 @api_view(['GET', 'POST'])
22 @permission_classes((permissions.IsAuthenticated, HasVerifiedEmail))
23 @authentication_classes((ExpiringTokenAuthentication,))
24 def challenge_host_team_list(request):
25
26 if request.method == 'GET':
27 challenge_host_teams = ChallengeHostTeam.objects.filter(created_by=request.user)
28 paginator, result_page = paginated_queryset(challenge_host_teams, request)
29 serializer = HostTeamDetailSerializer(result_page, many=True)
30 response_data = serializer.data
31 return paginator.get_paginated_response(response_data)
32
33 elif request.method == 'POST':
34 serializer = ChallengeHostTeamSerializer(data=request.data,
35 context={'request': request})
36 if serializer.is_valid():
37 serializer.save()
38 response_data = serializer.data
39 return Response(response_data, status=status.HTTP_201_CREATED)
40 return Response(serializer.errors, status=status.HTTP_400_BAD_REQUEST)
41
42
43 @throttle_classes([UserRateThrottle])
44 @api_view(['GET', 'PUT', 'PATCH', 'DELETE'])
45 @permission_classes((permissions.IsAuthenticated, HasVerifiedEmail))
46 @authentication_classes((ExpiringTokenAuthentication,))
47 def challenge_host_team_detail(request, pk):
48 try:
49 challenge_host_team = ChallengeHostTeam.objects.get(pk=pk)
50 except ChallengeHostTeam.DoesNotExist:
51 response_data = {'error': 'ChallengeHostTeam does not exist'}
52 return Response(response_data, status=status.HTTP_406_NOT_ACCEPTABLE)
53
54 if request.method == 'GET':
55 serializer = HostTeamDetailSerializer(challenge_host_team)
56 response_data = serializer.data
57 return Response(response_data, status=status.HTTP_200_OK)
58
59 elif request.method in ['PUT', 'PATCH']:
60
61 if request.method == 'PATCH':
62 serializer = ChallengeHostTeamSerializer(challenge_host_team,
63 data=request.data,
64 context={'request': request},
65 partial=True)
66 else:
67 serializer = ChallengeHostTeamSerializer(challenge_host_team,
68 data=request.data,
69 context={'request': request})
70 if serializer.is_valid():
71 serializer.save()
72 response_data = serializer.data
73 return Response(response_data, status=status.HTTP_200_OK)
74 else:
75 return Response(serializer.errors, status=status.HTTP_400_BAD_REQUEST)
76
77 elif request.method == 'DELETE':
78 challenge_host_team.delete()
79 return Response(status=status.HTTP_204_NO_CONTENT)
80
81
82 @throttle_classes([UserRateThrottle])
83 @api_view(['GET', 'POST'])
84 @permission_classes((permissions.IsAuthenticated, HasVerifiedEmail))
85 @authentication_classes((ExpiringTokenAuthentication,))
86 def challenge_host_list(request, challenge_host_team_pk):
87
88 try:
89 challenge_host_team = ChallengeHostTeam.objects.get(pk=challenge_host_team_pk)
90 except ChallengeHostTeam.DoesNotExist:
91 response_data = {'error': 'ChallengeHostTeam does not exist'}
92 return Response(response_data, status=status.HTTP_406_NOT_ACCEPTABLE)
93
94 if request.method == 'GET':
95 challenge_host_status = request.query_params.get('status', None)
96 filter_condition = {
97 'team_name': challenge_host_team,
98 'user': request.user
99 }
100 if challenge_host_status:
101 challenge_host_status = challenge_host_status.split(',')
102 filter_condition.update({'status__in': challenge_host_status})
103
104 challenge_host = ChallengeHost.objects.filter(**filter_condition)
105 paginator, result_page = paginated_queryset(challenge_host, request)
106 serializer = ChallengeHostSerializer(result_page, many=True)
107 response_data = serializer.data
108 return paginator.get_paginated_response(response_data)
109
110 elif request.method == 'POST':
111 serializer = ChallengeHostSerializer(data=request.data,
112 context={'challenge_host_team': challenge_host_team,
113 'request': request})
114 if serializer.is_valid():
115 serializer.save()
116 response_data = serializer.data
117 return Response(response_data, status=status.HTTP_201_CREATED)
118 return Response(serializer.errors, status=status.HTTP_400_BAD_REQUEST)
119
120
121 @throttle_classes([UserRateThrottle])
122 @api_view(['GET', 'PUT', 'PATCH', 'DELETE'])
123 @permission_classes((permissions.IsAuthenticated, HasVerifiedEmail))
124 @authentication_classes((ExpiringTokenAuthentication,))
125 def challenge_host_detail(request, challenge_host_team_pk, pk):
126 try:
127 challenge_host_team = ChallengeHostTeam.objects.get(pk=challenge_host_team_pk)
128 except ChallengeHostTeam.DoesNotExist:
129 response_data = {'error': 'ChallengeHostTeam does not exist'}
130 return Response(response_data, status=status.HTTP_406_NOT_ACCEPTABLE)
131
132 try:
133 challenge_host = ChallengeHost.objects.get(pk=pk)
134 except ChallengeHost.DoesNotExist:
135 response_data = {'error': 'ChallengeHost does not exist'}
136 return Response(response_data, status=status.HTTP_406_NOT_ACCEPTABLE)
137
138 if request.method == 'GET':
139 serializer = ChallengeHostSerializer(challenge_host)
140 response_data = serializer.data
141 return Response(response_data, status=status.HTTP_200_OK)
142
143 elif request.method in ['PUT', 'PATCH']:
144 if request.method == 'PATCH':
145 serializer = ChallengeHostSerializer(challenge_host,
146 data=request.data,
147 context={'challenge_host_team': challenge_host_team,
148 'request': request},
149 partial=True)
150 else:
151 serializer = ChallengeHostSerializer(challenge_host,
152 data=request.data,
153 context={'challenge_host_team': challenge_host_team,
154 'request': request})
155 if serializer.is_valid():
156 serializer.save()
157 response_data = serializer.data
158 return Response(response_data, status=status.HTTP_200_OK)
159 else:
160 return Response(serializer.errors, status=status.HTTP_400_BAD_REQUEST)
161
162 elif request.method == 'DELETE':
163 challenge_host.delete()
164 return Response(status=status.HTTP_204_NO_CONTENT)
165
166
167 @throttle_classes([UserRateThrottle])
168 @api_view(['POST'])
169 @permission_classes((permissions.IsAuthenticated, HasVerifiedEmail))
170 @authentication_classes((ExpiringTokenAuthentication,))
171 def create_challenge_host_team(request):
172
173 serializer = ChallengeHostTeamSerializer(data=request.data,
174 context={'request': request})
175 if serializer.is_valid():
176 serializer.save()
177 response_data = serializer.data
178 challenge_host_team = serializer.instance
179 challenge_host = ChallengeHost(user=request.user,
180 status=ChallengeHost.SELF,
181 permissions=ChallengeHost.ADMIN,
182 team_name=challenge_host_team)
183 challenge_host.save()
184 return Response(response_data, status=status.HTTP_201_CREATED)
185 return Response(serializer.errors, status=status.HTTP_400_BAD_REQUEST)
186
187
188 @throttle_classes([UserRateThrottle, ])
189 @api_view(['DELETE', ])
190 @permission_classes((permissions.IsAuthenticated, HasVerifiedEmail))
191 @authentication_classes((ExpiringTokenAuthentication, ))
192 def remove_self_from_challenge_host_team(request, challenge_host_team_pk):
193 """
194 A user can remove himself from the challenge host team.
195 """
196 try:
197 ChallengeHostTeam.objects.get(pk=challenge_host_team_pk)
198 except ChallengeHostTeam.DoesNotExist:
199 response_data = {'error': 'ChallengeHostTeam does not exist'}
200 return Response(response_data, status=status.HTTP_406_NOT_ACCEPTABLE)
201 try:
202 challenge_host = ChallengeHost.objects.filter(user=request.user.id, team_name__pk=challenge_host_team_pk)
203 challenge_host.delete()
204 return Response(status=status.HTTP_204_NO_CONTENT)
205 except:
206 response_data = {'error': 'Sorry, you do not belong to this team.'}
207 return Response(response_data, status=status.HTTP_401_UNAUTHORIZED)
208
209
210 @throttle_classes([UserRateThrottle])
211 @api_view(['POST'])
212 @permission_classes((permissions.IsAuthenticated, HasVerifiedEmail))
213 @authentication_classes((ExpiringTokenAuthentication,))
214 def invite_host_to_team(request, pk):
215
216 try:
217 challenge_host_team = ChallengeHostTeam.objects.get(pk=pk)
218 except ChallengeHostTeam.DoesNotExist:
219 response_data = {'error': 'ChallengeHostTeam does not exist'}
220 return Response(response_data, status=status.HTTP_406_NOT_ACCEPTABLE)
221
222 serializer = InviteHostToTeamSerializer(data=request.data,
223 context={'challenge_host_team': challenge_host_team,
224 'request': request})
225 if serializer.is_valid():
226 serializer.save()
227 response_data = {
228 'message': 'User has been added successfully to the host team'}
229 return Response(response_data, status=status.HTTP_202_ACCEPTED)
230 return Response(serializer.errors, status=status.HTTP_400_BAD_REQUEST)
231
[end of apps/hosts/views.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/apps/hosts/views.py b/apps/hosts/views.py
--- a/apps/hosts/views.py
+++ b/apps/hosts/views.py
@@ -24,7 +24,8 @@
def challenge_host_team_list(request):
if request.method == 'GET':
- challenge_host_teams = ChallengeHostTeam.objects.filter(created_by=request.user)
+ challenge_host_team_ids = ChallengeHost.objects.filter(user=request.user).values_list('team_name', flat=True)
+ challenge_host_teams = ChallengeHostTeam.objects.filter(id__in=challenge_host_team_ids)
paginator, result_page = paginated_queryset(challenge_host_teams, request)
serializer = HostTeamDetailSerializer(result_page, many=True)
response_data = serializer.data
|
{"golden_diff": "diff --git a/apps/hosts/views.py b/apps/hosts/views.py\n--- a/apps/hosts/views.py\n+++ b/apps/hosts/views.py\n@@ -24,7 +24,8 @@\n def challenge_host_team_list(request):\n \n if request.method == 'GET':\n- challenge_host_teams = ChallengeHostTeam.objects.filter(created_by=request.user)\n+ challenge_host_team_ids = ChallengeHost.objects.filter(user=request.user).values_list('team_name', flat=True)\n+ challenge_host_teams = ChallengeHostTeam.objects.filter(id__in=challenge_host_team_ids)\n paginator, result_page = paginated_queryset(challenge_host_teams, request)\n serializer = HostTeamDetailSerializer(result_page, many=True)\n response_data = serializer.data\n", "issue": "Hosts: User cannot see all of his Host Teams\n## CURRENT STATE - \r\n`challenge-host-teams` page only shows teams `created_by` the user.\r\n\r\n## ISSUE - \r\nOn `challenge-host-teams` page, user should be able to see all host teams he is a part of.\r\n\r\n## RELATED ISSUE -\r\n#607 Refer to this issue also. \r\n\n", "before_files": [{"content": "from rest_framework import permissions, status\nfrom rest_framework.decorators import (api_view,\n authentication_classes,\n permission_classes,\n throttle_classes,)\nfrom rest_framework.response import Response\nfrom rest_framework_expiring_authtoken.authentication import (ExpiringTokenAuthentication,)\nfrom rest_framework.throttling import UserRateThrottle\n\nfrom accounts.permissions import HasVerifiedEmail\nfrom base.utils import paginated_queryset\nfrom .models import (ChallengeHost,\n ChallengeHostTeam,)\nfrom .serializers import (ChallengeHostSerializer,\n ChallengeHostTeamSerializer,\n InviteHostToTeamSerializer,\n HostTeamDetailSerializer,)\n\n\n@throttle_classes([UserRateThrottle])\n@api_view(['GET', 'POST'])\n@permission_classes((permissions.IsAuthenticated, HasVerifiedEmail))\n@authentication_classes((ExpiringTokenAuthentication,))\ndef challenge_host_team_list(request):\n\n if request.method == 'GET':\n challenge_host_teams = ChallengeHostTeam.objects.filter(created_by=request.user)\n paginator, result_page = paginated_queryset(challenge_host_teams, request)\n serializer = HostTeamDetailSerializer(result_page, many=True)\n response_data = serializer.data\n return paginator.get_paginated_response(response_data)\n\n elif request.method == 'POST':\n serializer = ChallengeHostTeamSerializer(data=request.data,\n context={'request': request})\n if serializer.is_valid():\n serializer.save()\n response_data = serializer.data\n return Response(response_data, status=status.HTTP_201_CREATED)\n return Response(serializer.errors, status=status.HTTP_400_BAD_REQUEST)\n\n\n@throttle_classes([UserRateThrottle])\n@api_view(['GET', 'PUT', 'PATCH', 'DELETE'])\n@permission_classes((permissions.IsAuthenticated, HasVerifiedEmail))\n@authentication_classes((ExpiringTokenAuthentication,))\ndef challenge_host_team_detail(request, pk):\n try:\n challenge_host_team = ChallengeHostTeam.objects.get(pk=pk)\n except ChallengeHostTeam.DoesNotExist:\n response_data = {'error': 'ChallengeHostTeam does not exist'}\n return Response(response_data, status=status.HTTP_406_NOT_ACCEPTABLE)\n\n if request.method == 'GET':\n serializer = HostTeamDetailSerializer(challenge_host_team)\n response_data = serializer.data\n return Response(response_data, status=status.HTTP_200_OK)\n\n elif request.method in ['PUT', 'PATCH']:\n\n if request.method == 'PATCH':\n serializer = ChallengeHostTeamSerializer(challenge_host_team,\n data=request.data,\n context={'request': request},\n partial=True)\n else:\n serializer = ChallengeHostTeamSerializer(challenge_host_team,\n data=request.data,\n context={'request': request})\n if serializer.is_valid():\n serializer.save()\n response_data = serializer.data\n return Response(response_data, status=status.HTTP_200_OK)\n else:\n return Response(serializer.errors, status=status.HTTP_400_BAD_REQUEST)\n\n elif request.method == 'DELETE':\n challenge_host_team.delete()\n return Response(status=status.HTTP_204_NO_CONTENT)\n\n\n@throttle_classes([UserRateThrottle])\n@api_view(['GET', 'POST'])\n@permission_classes((permissions.IsAuthenticated, HasVerifiedEmail))\n@authentication_classes((ExpiringTokenAuthentication,))\ndef challenge_host_list(request, challenge_host_team_pk):\n\n try:\n challenge_host_team = ChallengeHostTeam.objects.get(pk=challenge_host_team_pk)\n except ChallengeHostTeam.DoesNotExist:\n response_data = {'error': 'ChallengeHostTeam does not exist'}\n return Response(response_data, status=status.HTTP_406_NOT_ACCEPTABLE)\n\n if request.method == 'GET':\n challenge_host_status = request.query_params.get('status', None)\n filter_condition = {\n 'team_name': challenge_host_team,\n 'user': request.user\n }\n if challenge_host_status:\n challenge_host_status = challenge_host_status.split(',')\n filter_condition.update({'status__in': challenge_host_status})\n\n challenge_host = ChallengeHost.objects.filter(**filter_condition)\n paginator, result_page = paginated_queryset(challenge_host, request)\n serializer = ChallengeHostSerializer(result_page, many=True)\n response_data = serializer.data\n return paginator.get_paginated_response(response_data)\n\n elif request.method == 'POST':\n serializer = ChallengeHostSerializer(data=request.data,\n context={'challenge_host_team': challenge_host_team,\n 'request': request})\n if serializer.is_valid():\n serializer.save()\n response_data = serializer.data\n return Response(response_data, status=status.HTTP_201_CREATED)\n return Response(serializer.errors, status=status.HTTP_400_BAD_REQUEST)\n\n\n@throttle_classes([UserRateThrottle])\n@api_view(['GET', 'PUT', 'PATCH', 'DELETE'])\n@permission_classes((permissions.IsAuthenticated, HasVerifiedEmail))\n@authentication_classes((ExpiringTokenAuthentication,))\ndef challenge_host_detail(request, challenge_host_team_pk, pk):\n try:\n challenge_host_team = ChallengeHostTeam.objects.get(pk=challenge_host_team_pk)\n except ChallengeHostTeam.DoesNotExist:\n response_data = {'error': 'ChallengeHostTeam does not exist'}\n return Response(response_data, status=status.HTTP_406_NOT_ACCEPTABLE)\n\n try:\n challenge_host = ChallengeHost.objects.get(pk=pk)\n except ChallengeHost.DoesNotExist:\n response_data = {'error': 'ChallengeHost does not exist'}\n return Response(response_data, status=status.HTTP_406_NOT_ACCEPTABLE)\n\n if request.method == 'GET':\n serializer = ChallengeHostSerializer(challenge_host)\n response_data = serializer.data\n return Response(response_data, status=status.HTTP_200_OK)\n\n elif request.method in ['PUT', 'PATCH']:\n if request.method == 'PATCH':\n serializer = ChallengeHostSerializer(challenge_host,\n data=request.data,\n context={'challenge_host_team': challenge_host_team,\n 'request': request},\n partial=True)\n else:\n serializer = ChallengeHostSerializer(challenge_host,\n data=request.data,\n context={'challenge_host_team': challenge_host_team,\n 'request': request})\n if serializer.is_valid():\n serializer.save()\n response_data = serializer.data\n return Response(response_data, status=status.HTTP_200_OK)\n else:\n return Response(serializer.errors, status=status.HTTP_400_BAD_REQUEST)\n\n elif request.method == 'DELETE':\n challenge_host.delete()\n return Response(status=status.HTTP_204_NO_CONTENT)\n\n\n@throttle_classes([UserRateThrottle])\n@api_view(['POST'])\n@permission_classes((permissions.IsAuthenticated, HasVerifiedEmail))\n@authentication_classes((ExpiringTokenAuthentication,))\ndef create_challenge_host_team(request):\n\n serializer = ChallengeHostTeamSerializer(data=request.data,\n context={'request': request})\n if serializer.is_valid():\n serializer.save()\n response_data = serializer.data\n challenge_host_team = serializer.instance\n challenge_host = ChallengeHost(user=request.user,\n status=ChallengeHost.SELF,\n permissions=ChallengeHost.ADMIN,\n team_name=challenge_host_team)\n challenge_host.save()\n return Response(response_data, status=status.HTTP_201_CREATED)\n return Response(serializer.errors, status=status.HTTP_400_BAD_REQUEST)\n\n\n@throttle_classes([UserRateThrottle, ])\n@api_view(['DELETE', ])\n@permission_classes((permissions.IsAuthenticated, HasVerifiedEmail))\n@authentication_classes((ExpiringTokenAuthentication, ))\ndef remove_self_from_challenge_host_team(request, challenge_host_team_pk):\n \"\"\"\n A user can remove himself from the challenge host team.\n \"\"\"\n try:\n ChallengeHostTeam.objects.get(pk=challenge_host_team_pk)\n except ChallengeHostTeam.DoesNotExist:\n response_data = {'error': 'ChallengeHostTeam does not exist'}\n return Response(response_data, status=status.HTTP_406_NOT_ACCEPTABLE)\n try:\n challenge_host = ChallengeHost.objects.filter(user=request.user.id, team_name__pk=challenge_host_team_pk)\n challenge_host.delete()\n return Response(status=status.HTTP_204_NO_CONTENT)\n except:\n response_data = {'error': 'Sorry, you do not belong to this team.'}\n return Response(response_data, status=status.HTTP_401_UNAUTHORIZED)\n\n\n@throttle_classes([UserRateThrottle])\n@api_view(['POST'])\n@permission_classes((permissions.IsAuthenticated, HasVerifiedEmail))\n@authentication_classes((ExpiringTokenAuthentication,))\ndef invite_host_to_team(request, pk):\n\n try:\n challenge_host_team = ChallengeHostTeam.objects.get(pk=pk)\n except ChallengeHostTeam.DoesNotExist:\n response_data = {'error': 'ChallengeHostTeam does not exist'}\n return Response(response_data, status=status.HTTP_406_NOT_ACCEPTABLE)\n\n serializer = InviteHostToTeamSerializer(data=request.data,\n context={'challenge_host_team': challenge_host_team,\n 'request': request})\n if serializer.is_valid():\n serializer.save()\n response_data = {\n 'message': 'User has been added successfully to the host team'}\n return Response(response_data, status=status.HTTP_202_ACCEPTED)\n return Response(serializer.errors, status=status.HTTP_400_BAD_REQUEST)\n", "path": "apps/hosts/views.py"}]}
| 3,103 | 156 |
gh_patches_debug_10651
|
rasdani/github-patches
|
git_diff
|
hpcaitech__ColossalAI-5440
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
[tensor] fix some unittests
[tensor] fix some unittests
[tensor] fix some unittests
</issue>
<code>
[start of colossalai/legacy/tensor/tensor_spec.py]
1 from dataclasses import dataclass
2 from typing import Optional
3
4 from colossalai.legacy.tensor.distspec import DistPlacementPattern, _DistSpec
5 from colossalai.legacy.tensor.process_group import ProcessGroup
6
7 from .compute_spec import ComputeSpec
8
9
10 @dataclass
11 class ColoTensorSpec:
12 """ColoTensorSpec
13
14 A data class for specifications of the `ColoTensor`.
15 It contains attributes of `ProcessGroup`, `_DistSpec`, `ComputeSpec`.
16 The latter two attributes are optional. If not set, they are default value is `Replicate()` and `None`.
17 """
18
19 pg: ProcessGroup
20 dist_attr: Optional[_DistSpec] = _DistSpec(DistPlacementPattern.REPLICATE)
21 compute_attr: Optional[ComputeSpec] = None
22
[end of colossalai/legacy/tensor/tensor_spec.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/colossalai/legacy/tensor/tensor_spec.py b/colossalai/legacy/tensor/tensor_spec.py
--- a/colossalai/legacy/tensor/tensor_spec.py
+++ b/colossalai/legacy/tensor/tensor_spec.py
@@ -1,4 +1,4 @@
-from dataclasses import dataclass
+from dataclasses import dataclass, field
from typing import Optional
from colossalai.legacy.tensor.distspec import DistPlacementPattern, _DistSpec
@@ -17,5 +17,5 @@
"""
pg: ProcessGroup
- dist_attr: Optional[_DistSpec] = _DistSpec(DistPlacementPattern.REPLICATE)
+ dist_attr: Optional[_DistSpec] = field(default_factory=lambda: _DistSpec(DistPlacementPattern.REPLICATE))
compute_attr: Optional[ComputeSpec] = None
|
{"golden_diff": "diff --git a/colossalai/legacy/tensor/tensor_spec.py b/colossalai/legacy/tensor/tensor_spec.py\n--- a/colossalai/legacy/tensor/tensor_spec.py\n+++ b/colossalai/legacy/tensor/tensor_spec.py\n@@ -1,4 +1,4 @@\n-from dataclasses import dataclass\n+from dataclasses import dataclass, field\n from typing import Optional\n \n from colossalai.legacy.tensor.distspec import DistPlacementPattern, _DistSpec\n@@ -17,5 +17,5 @@\n \"\"\"\n \n pg: ProcessGroup\n- dist_attr: Optional[_DistSpec] = _DistSpec(DistPlacementPattern.REPLICATE)\n+ dist_attr: Optional[_DistSpec] = field(default_factory=lambda: _DistSpec(DistPlacementPattern.REPLICATE))\n compute_attr: Optional[ComputeSpec] = None\n", "issue": "[tensor] fix some unittests\n\n[tensor] fix some unittests\n\n[tensor] fix some unittests\n\n", "before_files": [{"content": "from dataclasses import dataclass\nfrom typing import Optional\n\nfrom colossalai.legacy.tensor.distspec import DistPlacementPattern, _DistSpec\nfrom colossalai.legacy.tensor.process_group import ProcessGroup\n\nfrom .compute_spec import ComputeSpec\n\n\n@dataclass\nclass ColoTensorSpec:\n \"\"\"ColoTensorSpec\n\n A data class for specifications of the `ColoTensor`.\n It contains attributes of `ProcessGroup`, `_DistSpec`, `ComputeSpec`.\n The latter two attributes are optional. If not set, they are default value is `Replicate()` and `None`.\n \"\"\"\n\n pg: ProcessGroup\n dist_attr: Optional[_DistSpec] = _DistSpec(DistPlacementPattern.REPLICATE)\n compute_attr: Optional[ComputeSpec] = None\n", "path": "colossalai/legacy/tensor/tensor_spec.py"}]}
| 771 | 190 |
gh_patches_debug_12715
|
rasdani/github-patches
|
git_diff
|
open-telemetry__opentelemetry-python-3240
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Export ExponentialBucketHistogramAggregation in opentelemetry.sdk.metrics.view
**Is your feature request related to a problem?**
We want to use the Exponential Histograms features publicly released in version [1.17.0](https://github.com/open-telemetry/opentelemetry-python/blob/main/CHANGELOG.md#version-1170038b0-2023-03-22).
**Describe the solution you'd like**
I'd like to use the public API.
**Describe alternatives you've considered**
One can import it from `opentelemetry.sdk.metrics._internal.aggregation`
**Additional context**
Currently the code in https://github.com/open-telemetry/opentelemetry-python/blob/b6a1b22fa65f41bdefb01d64b76e5e793d039f6d/opentelemetry-sdk/src/opentelemetry/sdk/metrics/view/__init__.py#L25-L33 is not exporting the newly added `ExponentialBucketHistogramAggregation`
Export ExponentialBucketHistogramAggregation in opentelemetry.sdk.metrics.view
**Is your feature request related to a problem?**
We want to use the Exponential Histograms features publicly released in version [1.17.0](https://github.com/open-telemetry/opentelemetry-python/blob/main/CHANGELOG.md#version-1170038b0-2023-03-22).
**Describe the solution you'd like**
I'd like to use the public API.
**Describe alternatives you've considered**
One can import it from `opentelemetry.sdk.metrics._internal.aggregation`
**Additional context**
Currently the code in https://github.com/open-telemetry/opentelemetry-python/blob/b6a1b22fa65f41bdefb01d64b76e5e793d039f6d/opentelemetry-sdk/src/opentelemetry/sdk/metrics/view/__init__.py#L25-L33 is not exporting the newly added `ExponentialBucketHistogramAggregation`
</issue>
<code>
[start of opentelemetry-sdk/src/opentelemetry/sdk/metrics/view/__init__.py]
1 # Copyright The OpenTelemetry Authors
2 #
3 # Licensed under the Apache License, Version 2.0 (the "License");
4 # you may not use this file except in compliance with the License.
5 # You may obtain a copy of the License at
6 #
7 # http://www.apache.org/licenses/LICENSE-2.0
8 #
9 # Unless required by applicable law or agreed to in writing, software
10 # distributed under the License is distributed on an "AS IS" BASIS,
11 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12 # See the License for the specific language governing permissions and
13 # limitations under the License.
14
15 from opentelemetry.sdk.metrics._internal.aggregation import (
16 Aggregation,
17 DefaultAggregation,
18 DropAggregation,
19 ExplicitBucketHistogramAggregation,
20 LastValueAggregation,
21 SumAggregation,
22 )
23 from opentelemetry.sdk.metrics._internal.view import View
24
25 __all__ = [
26 "Aggregation",
27 "DefaultAggregation",
28 "DropAggregation",
29 "ExplicitBucketHistogramAggregation",
30 "LastValueAggregation",
31 "SumAggregation",
32 "View",
33 ]
34
[end of opentelemetry-sdk/src/opentelemetry/sdk/metrics/view/__init__.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/opentelemetry-sdk/src/opentelemetry/sdk/metrics/view/__init__.py b/opentelemetry-sdk/src/opentelemetry/sdk/metrics/view/__init__.py
--- a/opentelemetry-sdk/src/opentelemetry/sdk/metrics/view/__init__.py
+++ b/opentelemetry-sdk/src/opentelemetry/sdk/metrics/view/__init__.py
@@ -17,6 +17,7 @@
DefaultAggregation,
DropAggregation,
ExplicitBucketHistogramAggregation,
+ ExponentialBucketHistogramAggregation,
LastValueAggregation,
SumAggregation,
)
@@ -27,6 +28,7 @@
"DefaultAggregation",
"DropAggregation",
"ExplicitBucketHistogramAggregation",
+ "ExponentialBucketHistogramAggregation",
"LastValueAggregation",
"SumAggregation",
"View",
|
{"golden_diff": "diff --git a/opentelemetry-sdk/src/opentelemetry/sdk/metrics/view/__init__.py b/opentelemetry-sdk/src/opentelemetry/sdk/metrics/view/__init__.py\n--- a/opentelemetry-sdk/src/opentelemetry/sdk/metrics/view/__init__.py\n+++ b/opentelemetry-sdk/src/opentelemetry/sdk/metrics/view/__init__.py\n@@ -17,6 +17,7 @@\n DefaultAggregation,\n DropAggregation,\n ExplicitBucketHistogramAggregation,\n+ ExponentialBucketHistogramAggregation,\n LastValueAggregation,\n SumAggregation,\n )\n@@ -27,6 +28,7 @@\n \"DefaultAggregation\",\n \"DropAggregation\",\n \"ExplicitBucketHistogramAggregation\",\n+ \"ExponentialBucketHistogramAggregation\",\n \"LastValueAggregation\",\n \"SumAggregation\",\n \"View\",\n", "issue": "Export ExponentialBucketHistogramAggregation in opentelemetry.sdk.metrics.view\n**Is your feature request related to a problem?**\r\nWe want to use the Exponential Histograms features publicly released in version [1.17.0](https://github.com/open-telemetry/opentelemetry-python/blob/main/CHANGELOG.md#version-1170038b0-2023-03-22).\r\n\r\n**Describe the solution you'd like**\r\nI'd like to use the public API.\r\n\r\n**Describe alternatives you've considered**\r\nOne can import it from `opentelemetry.sdk.metrics._internal.aggregation`\r\n\r\n**Additional context**\r\nCurrently the code in https://github.com/open-telemetry/opentelemetry-python/blob/b6a1b22fa65f41bdefb01d64b76e5e793d039f6d/opentelemetry-sdk/src/opentelemetry/sdk/metrics/view/__init__.py#L25-L33 is not exporting the newly added `ExponentialBucketHistogramAggregation`\r\n\nExport ExponentialBucketHistogramAggregation in opentelemetry.sdk.metrics.view\n**Is your feature request related to a problem?**\r\nWe want to use the Exponential Histograms features publicly released in version [1.17.0](https://github.com/open-telemetry/opentelemetry-python/blob/main/CHANGELOG.md#version-1170038b0-2023-03-22).\r\n\r\n**Describe the solution you'd like**\r\nI'd like to use the public API.\r\n\r\n**Describe alternatives you've considered**\r\nOne can import it from `opentelemetry.sdk.metrics._internal.aggregation`\r\n\r\n**Additional context**\r\nCurrently the code in https://github.com/open-telemetry/opentelemetry-python/blob/b6a1b22fa65f41bdefb01d64b76e5e793d039f6d/opentelemetry-sdk/src/opentelemetry/sdk/metrics/view/__init__.py#L25-L33 is not exporting the newly added `ExponentialBucketHistogramAggregation`\r\n\n", "before_files": [{"content": "# Copyright The OpenTelemetry Authors\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom opentelemetry.sdk.metrics._internal.aggregation import (\n Aggregation,\n DefaultAggregation,\n DropAggregation,\n ExplicitBucketHistogramAggregation,\n LastValueAggregation,\n SumAggregation,\n)\nfrom opentelemetry.sdk.metrics._internal.view import View\n\n__all__ = [\n \"Aggregation\",\n \"DefaultAggregation\",\n \"DropAggregation\",\n \"ExplicitBucketHistogramAggregation\",\n \"LastValueAggregation\",\n \"SumAggregation\",\n \"View\",\n]\n", "path": "opentelemetry-sdk/src/opentelemetry/sdk/metrics/view/__init__.py"}]}
| 1,304 | 185 |
gh_patches_debug_9367
|
rasdani/github-patches
|
git_diff
|
ytdl-org__youtube-dl-14997
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
[https://www.franceinter.fr] WARNING: unable to extract upload date
---
### Make sure you are using the *latest* version: run `youtube-dl --version` and ensure your version is *2017.12.14*. If it's not, read [this FAQ entry](https://github.com/rg3/youtube-dl/blob/master/README.md#how-do-i-update-youtube-dl) and update. Issues with outdated version will be rejected.
- [x] I've **verified** and **I assure** that I'm running youtube-dl **2017.12.14**
### Before submitting an *issue* make sure you have:
- [x] At least skimmed through the [README](https://github.com/rg3/youtube-dl/blob/master/README.md), **most notably** the [FAQ](https://github.com/rg3/youtube-dl#faq) and [BUGS](https://github.com/rg3/youtube-dl#bugs) sections
- [x] [Searched](https://github.com/rg3/youtube-dl/search?type=Issues) the bugtracker for similar issues including closed ones
### What is the purpose of your *issue*?
- [ ] Bug report (encountered problems with youtube-dl)
- [x] Site support request (request for adding support for a new site)
- [ ] Feature request (request for a new functionality)
- [ ] Question
- [ ] Other
---
```
youtube-dl-mp3 "https://www.franceinter.fr/emissions/les-concerts-d-inter/les-concerts-d-inter-14-decembre-2017"
[FranceInter] les-concerts-d-inter/les-concerts-d-inter-14-decembre-2017: Downloading webpage
WARNING: unable to extract upload date; please report this issue on https://yt-dl.org/bug . Make sure you are using the latest version; see https://yt-dl.org/update on how to update. Be sure to call youtube-dl with the --verbose flag and include its complete output.
```
```
[debug] System config: []
[debug] User config: []
[debug] Custom config: []
[debug] Command-line args: [u'-v']
[debug] Encodings: locale UTF-8, fs UTF-8, out UTF-8, pref UTF-8
[debug] youtube-dl version 2017.12.14
[debug] Python version 2.7.12 - Linux-4.4.0-103-generic-x86_64-with-Ubuntu-16.04-xenial
[debug] exe versions: avconv 2.8.11-0ubuntu0.16.04.1, avprobe 2.8.11-0ubuntu0.16.04.1, ffmpeg 2.8.11-0ubuntu0.16.04.1, ffprobe 2.8.11-0ubuntu0.16.04.1
[debug] Proxy map: {}
```
</issue>
<code>
[start of youtube_dl/extractor/franceinter.py]
1 # coding: utf-8
2 from __future__ import unicode_literals
3
4 from .common import InfoExtractor
5 from ..utils import month_by_name
6
7
8 class FranceInterIE(InfoExtractor):
9 _VALID_URL = r'https?://(?:www\.)?franceinter\.fr/emissions/(?P<id>[^?#]+)'
10
11 _TEST = {
12 'url': 'https://www.franceinter.fr/emissions/affaires-sensibles/affaires-sensibles-07-septembre-2016',
13 'md5': '9e54d7bdb6fdc02a841007f8a975c094',
14 'info_dict': {
15 'id': 'affaires-sensibles/affaires-sensibles-07-septembre-2016',
16 'ext': 'mp3',
17 'title': 'Affaire Cahuzac : le contentieux du compte en Suisse',
18 'description': 'md5:401969c5d318c061f86bda1fa359292b',
19 'upload_date': '20160907',
20 },
21 }
22
23 def _real_extract(self, url):
24 video_id = self._match_id(url)
25
26 webpage = self._download_webpage(url, video_id)
27
28 video_url = self._search_regex(
29 r'(?s)<div[^>]+class=["\']page-diffusion["\'][^>]*>.*?<button[^>]+data-url=(["\'])(?P<url>(?:(?!\1).)+)\1',
30 webpage, 'video url', group='url')
31
32 title = self._og_search_title(webpage)
33 description = self._og_search_description(webpage)
34
35 upload_date_str = self._search_regex(
36 r'class=["\']cover-emission-period["\'][^>]*>[^<]+\s+(\d{1,2}\s+[^\s]+\s+\d{4})<',
37 webpage, 'upload date', fatal=False)
38 if upload_date_str:
39 upload_date_list = upload_date_str.split()
40 upload_date_list.reverse()
41 upload_date_list[1] = '%02d' % (month_by_name(upload_date_list[1], lang='fr') or 0)
42 upload_date_list[2] = '%02d' % int(upload_date_list[2])
43 upload_date = ''.join(upload_date_list)
44 else:
45 upload_date = None
46
47 return {
48 'id': video_id,
49 'title': title,
50 'description': description,
51 'upload_date': upload_date,
52 'formats': [{
53 'url': video_url,
54 'vcodec': 'none',
55 }],
56 }
57
[end of youtube_dl/extractor/franceinter.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/youtube_dl/extractor/franceinter.py b/youtube_dl/extractor/franceinter.py
--- a/youtube_dl/extractor/franceinter.py
+++ b/youtube_dl/extractor/franceinter.py
@@ -33,7 +33,7 @@
description = self._og_search_description(webpage)
upload_date_str = self._search_regex(
- r'class=["\']cover-emission-period["\'][^>]*>[^<]+\s+(\d{1,2}\s+[^\s]+\s+\d{4})<',
+ r'class=["\']\s*cover-emission-period\s*["\'][^>]*>[^<]+\s+(\d{1,2}\s+[^\s]+\s+\d{4})<',
webpage, 'upload date', fatal=False)
if upload_date_str:
upload_date_list = upload_date_str.split()
|
{"golden_diff": "diff --git a/youtube_dl/extractor/franceinter.py b/youtube_dl/extractor/franceinter.py\n--- a/youtube_dl/extractor/franceinter.py\n+++ b/youtube_dl/extractor/franceinter.py\n@@ -33,7 +33,7 @@\n description = self._og_search_description(webpage)\n \n upload_date_str = self._search_regex(\n- r'class=[\"\\']cover-emission-period[\"\\'][^>]*>[^<]+\\s+(\\d{1,2}\\s+[^\\s]+\\s+\\d{4})<',\n+ r'class=[\"\\']\\s*cover-emission-period\\s*[\"\\'][^>]*>[^<]+\\s+(\\d{1,2}\\s+[^\\s]+\\s+\\d{4})<',\n webpage, 'upload date', fatal=False)\n if upload_date_str:\n upload_date_list = upload_date_str.split()\n", "issue": "[https://www.franceinter.fr] WARNING: unable to extract upload date\n---\r\n\r\n### Make sure you are using the *latest* version: run `youtube-dl --version` and ensure your version is *2017.12.14*. If it's not, read [this FAQ entry](https://github.com/rg3/youtube-dl/blob/master/README.md#how-do-i-update-youtube-dl) and update. Issues with outdated version will be rejected.\r\n- [x] I've **verified** and **I assure** that I'm running youtube-dl **2017.12.14**\r\n\r\n### Before submitting an *issue* make sure you have:\r\n- [x] At least skimmed through the [README](https://github.com/rg3/youtube-dl/blob/master/README.md), **most notably** the [FAQ](https://github.com/rg3/youtube-dl#faq) and [BUGS](https://github.com/rg3/youtube-dl#bugs) sections\r\n- [x] [Searched](https://github.com/rg3/youtube-dl/search?type=Issues) the bugtracker for similar issues including closed ones\r\n\r\n### What is the purpose of your *issue*?\r\n- [ ] Bug report (encountered problems with youtube-dl)\r\n- [x] Site support request (request for adding support for a new site)\r\n- [ ] Feature request (request for a new functionality)\r\n- [ ] Question\r\n- [ ] Other\r\n\r\n---\r\n\r\n``` \r\nyoutube-dl-mp3 \"https://www.franceinter.fr/emissions/les-concerts-d-inter/les-concerts-d-inter-14-decembre-2017\"\r\n[FranceInter] les-concerts-d-inter/les-concerts-d-inter-14-decembre-2017: Downloading webpage\r\nWARNING: unable to extract upload date; please report this issue on https://yt-dl.org/bug . Make sure you are using the latest version; see https://yt-dl.org/update on how to update. Be sure to call youtube-dl with the --verbose flag and include its complete output.\r\n```\r\n\r\n```\r\n[debug] System config: []\r\n[debug] User config: []\r\n[debug] Custom config: []\r\n[debug] Command-line args: [u'-v']\r\n[debug] Encodings: locale UTF-8, fs UTF-8, out UTF-8, pref UTF-8\r\n[debug] youtube-dl version 2017.12.14\r\n[debug] Python version 2.7.12 - Linux-4.4.0-103-generic-x86_64-with-Ubuntu-16.04-xenial\r\n[debug] exe versions: avconv 2.8.11-0ubuntu0.16.04.1, avprobe 2.8.11-0ubuntu0.16.04.1, ffmpeg 2.8.11-0ubuntu0.16.04.1, ffprobe 2.8.11-0ubuntu0.16.04.1\r\n[debug] Proxy map: {}\r\n```\n", "before_files": [{"content": "# coding: utf-8\nfrom __future__ import unicode_literals\n\nfrom .common import InfoExtractor\nfrom ..utils import month_by_name\n\n\nclass FranceInterIE(InfoExtractor):\n _VALID_URL = r'https?://(?:www\\.)?franceinter\\.fr/emissions/(?P<id>[^?#]+)'\n\n _TEST = {\n 'url': 'https://www.franceinter.fr/emissions/affaires-sensibles/affaires-sensibles-07-septembre-2016',\n 'md5': '9e54d7bdb6fdc02a841007f8a975c094',\n 'info_dict': {\n 'id': 'affaires-sensibles/affaires-sensibles-07-septembre-2016',\n 'ext': 'mp3',\n 'title': 'Affaire Cahuzac : le contentieux du compte en Suisse',\n 'description': 'md5:401969c5d318c061f86bda1fa359292b',\n 'upload_date': '20160907',\n },\n }\n\n def _real_extract(self, url):\n video_id = self._match_id(url)\n\n webpage = self._download_webpage(url, video_id)\n\n video_url = self._search_regex(\n r'(?s)<div[^>]+class=[\"\\']page-diffusion[\"\\'][^>]*>.*?<button[^>]+data-url=([\"\\'])(?P<url>(?:(?!\\1).)+)\\1',\n webpage, 'video url', group='url')\n\n title = self._og_search_title(webpage)\n description = self._og_search_description(webpage)\n\n upload_date_str = self._search_regex(\n r'class=[\"\\']cover-emission-period[\"\\'][^>]*>[^<]+\\s+(\\d{1,2}\\s+[^\\s]+\\s+\\d{4})<',\n webpage, 'upload date', fatal=False)\n if upload_date_str:\n upload_date_list = upload_date_str.split()\n upload_date_list.reverse()\n upload_date_list[1] = '%02d' % (month_by_name(upload_date_list[1], lang='fr') or 0)\n upload_date_list[2] = '%02d' % int(upload_date_list[2])\n upload_date = ''.join(upload_date_list)\n else:\n upload_date = None\n\n return {\n 'id': video_id,\n 'title': title,\n 'description': description,\n 'upload_date': upload_date,\n 'formats': [{\n 'url': video_url,\n 'vcodec': 'none',\n }],\n }\n", "path": "youtube_dl/extractor/franceinter.py"}]}
| 1,956 | 201 |
gh_patches_debug_7992
|
rasdani/github-patches
|
git_diff
|
lightly-ai__lightly-305
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Fix import of ApiWorkflowClient
# Fix import of ApiWorkflowClient
Currently, the following import statement (from the docs) does not work:
```python
from lightly.api import ApiWorkflowClient
```
TODO:
- [x] We need to fix this by exposing the client in the `__init__.py` file.
- [x] Make sure the other imports in the docs work as well
</issue>
<code>
[start of lightly/active_learning/utils/__init__.py]
[end of lightly/active_learning/utils/__init__.py]
[start of lightly/api/__init__.py]
1 """ The lightly.api module provides access to the Lightly web-app. """
2
3 # Copyright (c) 2020. Lightly AG and its affiliates.
4 # All Rights Reserved
5
6 from lightly.api import routes
7
[end of lightly/api/__init__.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/lightly/active_learning/utils/__init__.py b/lightly/active_learning/utils/__init__.py
--- a/lightly/active_learning/utils/__init__.py
+++ b/lightly/active_learning/utils/__init__.py
@@ -0,0 +1,7 @@
+""" Collection of Utils for Active Learning """
+
+# Copyright (c) 2020. Lightly AG and its affiliates.
+# All Rights Reserved
+
+from lightly.active_learning.utils.bounding_box import BoundingBox
+from lightly.active_learning.utils.object_detection_output import ObjectDetectionOutput
\ No newline at end of file
diff --git a/lightly/api/__init__.py b/lightly/api/__init__.py
--- a/lightly/api/__init__.py
+++ b/lightly/api/__init__.py
@@ -3,4 +3,5 @@
# Copyright (c) 2020. Lightly AG and its affiliates.
# All Rights Reserved
+from lightly.api.api_workflow_client import ApiWorkflowClient
from lightly.api import routes
|
{"golden_diff": "diff --git a/lightly/active_learning/utils/__init__.py b/lightly/active_learning/utils/__init__.py\n--- a/lightly/active_learning/utils/__init__.py\n+++ b/lightly/active_learning/utils/__init__.py\n@@ -0,0 +1,7 @@\n+\"\"\" Collection of Utils for Active Learning \"\"\"\n+\n+# Copyright (c) 2020. Lightly AG and its affiliates.\n+# All Rights Reserved\n+\n+from lightly.active_learning.utils.bounding_box import BoundingBox\n+from lightly.active_learning.utils.object_detection_output import ObjectDetectionOutput\n\\ No newline at end of file\ndiff --git a/lightly/api/__init__.py b/lightly/api/__init__.py\n--- a/lightly/api/__init__.py\n+++ b/lightly/api/__init__.py\n@@ -3,4 +3,5 @@\n # Copyright (c) 2020. Lightly AG and its affiliates.\n # All Rights Reserved\n \n+from lightly.api.api_workflow_client import ApiWorkflowClient\n from lightly.api import routes\n", "issue": "Fix import of ApiWorkflowClient\n# Fix import of ApiWorkflowClient\r\n\r\nCurrently, the following import statement (from the docs) does not work:\r\n```python\r\nfrom lightly.api import ApiWorkflowClient\r\n```\r\n\r\nTODO:\r\n- [x] We need to fix this by exposing the client in the `__init__.py` file. \r\n- [x] Make sure the other imports in the docs work as well\n", "before_files": [{"content": "", "path": "lightly/active_learning/utils/__init__.py"}, {"content": "\"\"\" The lightly.api module provides access to the Lightly web-app. \"\"\"\n\n# Copyright (c) 2020. Lightly AG and its affiliates.\n# All Rights Reserved\n\nfrom lightly.api import routes\n", "path": "lightly/api/__init__.py"}]}
| 700 | 222 |
gh_patches_debug_5392
|
rasdani/github-patches
|
git_diff
|
streamlink__streamlink-1351
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Kanal7 Defective again
Only 2 months later they have changed the design.
Not opening with latest 0.9.0 Release:
[cli][info] Found matching plugin kanal7 for URL http://www.kanal7.com/canli-izle
error: No playable streams found on this URL: http://www.kanal7.com/canli-izle
</issue>
<code>
[start of src/streamlink/plugins/kanal7.py]
1 from __future__ import print_function
2 import re
3
4 from streamlink.plugin import Plugin
5 from streamlink.plugin.api import http
6 from streamlink.plugin.api import useragents
7 from streamlink.plugin.api import validate
8 from streamlink.stream import HLSStream
9
10
11 class Kanal7(Plugin):
12 url_re = re.compile(r"https?://(?:www.)?kanal7.com/canli-izle")
13 iframe_re = re.compile(r'iframe .*?src="(http://[^"]*?)"')
14 stream_re = re.compile(r'src="(http[^"]*?)"')
15
16 @classmethod
17 def can_handle_url(cls, url):
18 return cls.url_re.match(url) is not None
19
20 def find_iframe(self, url):
21 res = http.get(url)
22 # find iframe url
23 iframe = self.iframe_re.search(res.text)
24 iframe_url = iframe and iframe.group(1)
25 if iframe_url:
26 self.logger.debug("Found iframe: {}", iframe_url)
27 return iframe_url
28
29 def _get_streams(self):
30 iframe1 = self.find_iframe(self.url)
31 if iframe1:
32 iframe2 = self.find_iframe(iframe1)
33 if iframe2:
34 ires = http.get(iframe2)
35 stream_m = self.stream_re.search(ires.text)
36 stream_url = stream_m and stream_m.group(1)
37 if stream_url:
38 yield "live", HLSStream(self.session, stream_url, headers={"Referer": iframe2})
39 else:
40 self.logger.error("Could not find second iframe, has the page layout changed?")
41 else:
42 self.logger.error("Could not find iframe, has the page layout changed?")
43
44
45 __plugin__ = Kanal7
46
[end of src/streamlink/plugins/kanal7.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/src/streamlink/plugins/kanal7.py b/src/streamlink/plugins/kanal7.py
--- a/src/streamlink/plugins/kanal7.py
+++ b/src/streamlink/plugins/kanal7.py
@@ -11,7 +11,7 @@
class Kanal7(Plugin):
url_re = re.compile(r"https?://(?:www.)?kanal7.com/canli-izle")
iframe_re = re.compile(r'iframe .*?src="(http://[^"]*?)"')
- stream_re = re.compile(r'src="(http[^"]*?)"')
+ stream_re = re.compile(r'''tp_file\s+=\s+['"](http[^"]*?)['"]''')
@classmethod
def can_handle_url(cls, url):
|
{"golden_diff": "diff --git a/src/streamlink/plugins/kanal7.py b/src/streamlink/plugins/kanal7.py\n--- a/src/streamlink/plugins/kanal7.py\n+++ b/src/streamlink/plugins/kanal7.py\n@@ -11,7 +11,7 @@\n class Kanal7(Plugin):\n url_re = re.compile(r\"https?://(?:www.)?kanal7.com/canli-izle\")\n iframe_re = re.compile(r'iframe .*?src=\"(http://[^\"]*?)\"')\n- stream_re = re.compile(r'src=\"(http[^\"]*?)\"')\n+ stream_re = re.compile(r'''tp_file\\s+=\\s+['\"](http[^\"]*?)['\"]''')\n \n @classmethod\n def can_handle_url(cls, url):\n", "issue": "Kanal7 Defective again\nOnly 2 months later they have changed the design.\r\n\r\nNot opening with latest 0.9.0 Release:\r\n\r\n[cli][info] Found matching plugin kanal7 for URL http://www.kanal7.com/canli-izle\r\nerror: No playable streams found on this URL: http://www.kanal7.com/canli-izle\n", "before_files": [{"content": "from __future__ import print_function\nimport re\n\nfrom streamlink.plugin import Plugin\nfrom streamlink.plugin.api import http\nfrom streamlink.plugin.api import useragents\nfrom streamlink.plugin.api import validate\nfrom streamlink.stream import HLSStream\n\n\nclass Kanal7(Plugin):\n url_re = re.compile(r\"https?://(?:www.)?kanal7.com/canli-izle\")\n iframe_re = re.compile(r'iframe .*?src=\"(http://[^\"]*?)\"')\n stream_re = re.compile(r'src=\"(http[^\"]*?)\"')\n\n @classmethod\n def can_handle_url(cls, url):\n return cls.url_re.match(url) is not None\n\n def find_iframe(self, url):\n res = http.get(url)\n # find iframe url\n iframe = self.iframe_re.search(res.text)\n iframe_url = iframe and iframe.group(1)\n if iframe_url:\n self.logger.debug(\"Found iframe: {}\", iframe_url)\n return iframe_url\n\n def _get_streams(self):\n iframe1 = self.find_iframe(self.url)\n if iframe1:\n iframe2 = self.find_iframe(iframe1)\n if iframe2:\n ires = http.get(iframe2)\n stream_m = self.stream_re.search(ires.text)\n stream_url = stream_m and stream_m.group(1)\n if stream_url:\n yield \"live\", HLSStream(self.session, stream_url, headers={\"Referer\": iframe2})\n else:\n self.logger.error(\"Could not find second iframe, has the page layout changed?\")\n else:\n self.logger.error(\"Could not find iframe, has the page layout changed?\")\n\n\n__plugin__ = Kanal7\n", "path": "src/streamlink/plugins/kanal7.py"}]}
| 1,075 | 172 |
gh_patches_debug_41012
|
rasdani/github-patches
|
git_diff
|
cornellius-gp__gpytorch-1468
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
[Docs] Missing Grid.py documentation
# 📚 Documentation/Examples
** Is there documentation missing? **
The utils section of [GPyTorch documentation](https://gpytorch.readthedocs.io) does not include any information on grid.py, which is referenced [elsewhere in the docs](https://docs.gpytorch.ai/en/stable/kernels.html?highlight=choose_grid_size#gpytorch.kernels.GridKernel.update_grid).
</issue>
<code>
[start of gpytorch/utils/grid.py]
1 #!/usr/bin/env python3
2
3 import math
4 from typing import List, Tuple
5
6 import torch
7
8
9 def scale_to_bounds(x, lower_bound, upper_bound):
10 """
11 Scale the input data so that it lies in between the lower and upper bounds.
12
13 Args:
14 :attr:`x` (Tensor `n` or `b x n`):
15 the input
16 :attr:`lower_bound` (float)
17 :attr:`upper_bound` (float)
18
19 Returns:
20 :obj:`torch.Tensor`
21 """
22 # Scale features so they fit inside grid bounds
23 min_val = x.min()
24 max_val = x.max()
25 diff = max_val - min_val
26 x = (x - min_val) * (0.95 * (upper_bound - lower_bound) / diff) + 0.95 * lower_bound
27 return x
28
29
30 def choose_grid_size(train_inputs, ratio=1.0, kronecker_structure=True):
31 """
32 Given some training inputs, determine a good grid size for KISS-GP.
33
34 Args:
35 :attr:`train_inputs` (Tensor `n` or `n x d` or `b x n x d`):
36 training data
37 :attr:`ratio` (float, optional):
38 Ratio - number of grid points to the amount of data (default: 1.)
39 :attr:`kronecker_structure` (bool, default=True):
40 Whether or not the model will use Kronecker structure in the grid
41 (set to True unless there is an additive or product decomposition in the prior)
42
43 Returns:
44 :obj:`int`
45 """
46 # Scale features so they fit inside grid bounds
47 num_data = train_inputs.numel() if train_inputs.dim() == 1 else train_inputs.size(-2)
48 num_dim = 1 if train_inputs.dim() == 1 else train_inputs.size(-1)
49 if kronecker_structure:
50 return int(ratio * math.pow(num_data, 1.0 / num_dim))
51 else:
52 return ratio * num_data
53
54
55 def convert_legacy_grid(grid: torch.Tensor) -> List[torch.Tensor]:
56 return [grid[:, i] for i in range(grid.size(-1))]
57
58
59 def create_data_from_grid(grid: List[torch.Tensor]) -> torch.Tensor:
60 """
61 Args:
62 :attr:`grid` (List[Tensor])
63 Each Tensor is a 1D set of increments for the grid in that dimension
64 Returns:
65 `grid_data` (Tensor)
66 Returns the set of points on the grid going by column-major order
67 (due to legacy reasons).
68 """
69 if torch.is_tensor(grid):
70 grid = convert_legacy_grid(grid)
71 ndims = len(grid)
72 assert all(axis.dim() == 1 for axis in grid)
73 projections = torch.meshgrid(*grid)
74 grid_tensor = torch.stack(projections, axis=-1)
75 # Note that if we did
76 # grid_data = grid_tensor.reshape(-1, ndims)
77 # instead, we would be iterating through the points of our grid from the
78 # last data dimension to the first data dimension. However, due to legacy
79 # reasons, we need to iterate from the first data dimension to the last data
80 # dimension when creating grid_data
81 grid_data = grid_tensor.permute(*(reversed(range(ndims + 1)))).reshape(ndims, -1).transpose(0, 1)
82 return grid_data
83
84
85 def create_grid(
86 grid_sizes: List[int], grid_bounds: List[Tuple[float, float]], extend: bool = True, device="cpu", dtype=torch.float,
87 ) -> List[torch.Tensor]:
88 """
89 Creates a grid represented by a list of 1D Tensors representing the
90 projections of the grid into each dimension
91
92 If `extend`, we extend the grid by two points past the specified boundary
93 which can be important for getting good grid interpolations
94 """
95 grid = []
96 for i in range(len(grid_bounds)):
97 grid_diff = float(grid_bounds[i][1] - grid_bounds[i][0]) / (grid_sizes[i] - 2)
98 if extend:
99 proj = torch.linspace(
100 grid_bounds[i][0] - grid_diff, grid_bounds[i][1] + grid_diff, grid_sizes[i], device=device, dtype=dtype,
101 )
102 else:
103 proj = torch.linspace(grid_bounds[i][0], grid_bounds[i][1], grid_sizes[i], device=device, dtype=dtype,)
104 grid.append(proj)
105 return grid
106
[end of gpytorch/utils/grid.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/gpytorch/utils/grid.py b/gpytorch/utils/grid.py
--- a/gpytorch/utils/grid.py
+++ b/gpytorch/utils/grid.py
@@ -10,14 +10,12 @@
"""
Scale the input data so that it lies in between the lower and upper bounds.
- Args:
- :attr:`x` (Tensor `n` or `b x n`):
- the input
- :attr:`lower_bound` (float)
- :attr:`upper_bound` (float)
-
- Returns:
- :obj:`torch.Tensor`
+ :param x: the input data
+ :type x: torch.Tensor (... x n x d)
+ :param float lower_bound: lower bound of scaled data
+ :param float upper_bound: upper bound of scaled data
+ :return: scaled data
+ :rtype: torch.Tensor (... x n x d)
"""
# Scale features so they fit inside grid bounds
min_val = x.min()
@@ -31,17 +29,15 @@
"""
Given some training inputs, determine a good grid size for KISS-GP.
- Args:
- :attr:`train_inputs` (Tensor `n` or `n x d` or `b x n x d`):
- training data
- :attr:`ratio` (float, optional):
- Ratio - number of grid points to the amount of data (default: 1.)
- :attr:`kronecker_structure` (bool, default=True):
- Whether or not the model will use Kronecker structure in the grid
- (set to True unless there is an additive or product decomposition in the prior)
-
- Returns:
- :obj:`int`
+ :param x: the input data
+ :type x: torch.Tensor (... x n x d)
+ :param ratio: Amount of grid points per data point (default: 1.)
+ :type ratio: float, optional
+ :param kronecker_structure: Whether or not the model will use Kronecker structure in the grid
+ (set to True unless there is an additive or product decomposition in the prior)
+ :type kronecker_structure: bool, optional
+ :return: Grid size
+ :rtype: int
"""
# Scale features so they fit inside grid bounds
num_data = train_inputs.numel() if train_inputs.dim() == 1 else train_inputs.size(-2)
@@ -58,13 +54,10 @@
def create_data_from_grid(grid: List[torch.Tensor]) -> torch.Tensor:
"""
- Args:
- :attr:`grid` (List[Tensor])
- Each Tensor is a 1D set of increments for the grid in that dimension
- Returns:
- `grid_data` (Tensor)
- Returns the set of points on the grid going by column-major order
- (due to legacy reasons).
+ :param grid: Each Tensor is a 1D set of increments for the grid in that dimension
+ :type grid: List[torch.Tensor]
+ :return: The set of points on the grid going by column-major order
+ :rtype: torch.Tensor
"""
if torch.is_tensor(grid):
grid = convert_legacy_grid(grid)
@@ -90,7 +83,18 @@
projections of the grid into each dimension
If `extend`, we extend the grid by two points past the specified boundary
- which can be important for getting good grid interpolations
+ which can be important for getting good grid interpolations.
+
+ :param grid_sizes: Sizes of each grid dimension
+ :type grid_sizes: List[int]
+ :param grid_bounds: Lower and upper bounds of each grid dimension
+ :type grid_sizes: List[Tuple[float, float]]
+ :param device: target device for output (default: cpu)
+ :type device: torch.device, optional
+ :param dtype: target dtype for output (default: torch.float)
+ :type dtype: torch.dtype, optional
+ :return: Grid points for each dimension. Grid points are stored in a :obj:`torch.Tensor` with shape `grid_sizes[i]`.
+ :rtype: List[torch.Tensor]
"""
grid = []
for i in range(len(grid_bounds)):
|
{"golden_diff": "diff --git a/gpytorch/utils/grid.py b/gpytorch/utils/grid.py\n--- a/gpytorch/utils/grid.py\n+++ b/gpytorch/utils/grid.py\n@@ -10,14 +10,12 @@\n \"\"\"\n Scale the input data so that it lies in between the lower and upper bounds.\n \n- Args:\n- :attr:`x` (Tensor `n` or `b x n`):\n- the input\n- :attr:`lower_bound` (float)\n- :attr:`upper_bound` (float)\n-\n- Returns:\n- :obj:`torch.Tensor`\n+ :param x: the input data\n+ :type x: torch.Tensor (... x n x d)\n+ :param float lower_bound: lower bound of scaled data\n+ :param float upper_bound: upper bound of scaled data\n+ :return: scaled data\n+ :rtype: torch.Tensor (... x n x d)\n \"\"\"\n # Scale features so they fit inside grid bounds\n min_val = x.min()\n@@ -31,17 +29,15 @@\n \"\"\"\n Given some training inputs, determine a good grid size for KISS-GP.\n \n- Args:\n- :attr:`train_inputs` (Tensor `n` or `n x d` or `b x n x d`):\n- training data\n- :attr:`ratio` (float, optional):\n- Ratio - number of grid points to the amount of data (default: 1.)\n- :attr:`kronecker_structure` (bool, default=True):\n- Whether or not the model will use Kronecker structure in the grid\n- (set to True unless there is an additive or product decomposition in the prior)\n-\n- Returns:\n- :obj:`int`\n+ :param x: the input data\n+ :type x: torch.Tensor (... x n x d)\n+ :param ratio: Amount of grid points per data point (default: 1.)\n+ :type ratio: float, optional\n+ :param kronecker_structure: Whether or not the model will use Kronecker structure in the grid\n+ (set to True unless there is an additive or product decomposition in the prior)\n+ :type kronecker_structure: bool, optional\n+ :return: Grid size\n+ :rtype: int\n \"\"\"\n # Scale features so they fit inside grid bounds\n num_data = train_inputs.numel() if train_inputs.dim() == 1 else train_inputs.size(-2)\n@@ -58,13 +54,10 @@\n \n def create_data_from_grid(grid: List[torch.Tensor]) -> torch.Tensor:\n \"\"\"\n- Args:\n- :attr:`grid` (List[Tensor])\n- Each Tensor is a 1D set of increments for the grid in that dimension\n- Returns:\n- `grid_data` (Tensor)\n- Returns the set of points on the grid going by column-major order\n- (due to legacy reasons).\n+ :param grid: Each Tensor is a 1D set of increments for the grid in that dimension\n+ :type grid: List[torch.Tensor]\n+ :return: The set of points on the grid going by column-major order\n+ :rtype: torch.Tensor\n \"\"\"\n if torch.is_tensor(grid):\n grid = convert_legacy_grid(grid)\n@@ -90,7 +83,18 @@\n projections of the grid into each dimension\n \n If `extend`, we extend the grid by two points past the specified boundary\n- which can be important for getting good grid interpolations\n+ which can be important for getting good grid interpolations.\n+\n+ :param grid_sizes: Sizes of each grid dimension\n+ :type grid_sizes: List[int]\n+ :param grid_bounds: Lower and upper bounds of each grid dimension\n+ :type grid_sizes: List[Tuple[float, float]]\n+ :param device: target device for output (default: cpu)\n+ :type device: torch.device, optional\n+ :param dtype: target dtype for output (default: torch.float)\n+ :type dtype: torch.dtype, optional\n+ :return: Grid points for each dimension. Grid points are stored in a :obj:`torch.Tensor` with shape `grid_sizes[i]`.\n+ :rtype: List[torch.Tensor]\n \"\"\"\n grid = []\n for i in range(len(grid_bounds)):\n", "issue": "[Docs] Missing Grid.py documentation\n# \ud83d\udcda Documentation/Examples\r\n\r\n** Is there documentation missing? **\r\nThe utils section of [GPyTorch documentation](https://gpytorch.readthedocs.io) does not include any information on grid.py, which is referenced [elsewhere in the docs](https://docs.gpytorch.ai/en/stable/kernels.html?highlight=choose_grid_size#gpytorch.kernels.GridKernel.update_grid).\r\n\n", "before_files": [{"content": "#!/usr/bin/env python3\n\nimport math\nfrom typing import List, Tuple\n\nimport torch\n\n\ndef scale_to_bounds(x, lower_bound, upper_bound):\n \"\"\"\n Scale the input data so that it lies in between the lower and upper bounds.\n\n Args:\n :attr:`x` (Tensor `n` or `b x n`):\n the input\n :attr:`lower_bound` (float)\n :attr:`upper_bound` (float)\n\n Returns:\n :obj:`torch.Tensor`\n \"\"\"\n # Scale features so they fit inside grid bounds\n min_val = x.min()\n max_val = x.max()\n diff = max_val - min_val\n x = (x - min_val) * (0.95 * (upper_bound - lower_bound) / diff) + 0.95 * lower_bound\n return x\n\n\ndef choose_grid_size(train_inputs, ratio=1.0, kronecker_structure=True):\n \"\"\"\n Given some training inputs, determine a good grid size for KISS-GP.\n\n Args:\n :attr:`train_inputs` (Tensor `n` or `n x d` or `b x n x d`):\n training data\n :attr:`ratio` (float, optional):\n Ratio - number of grid points to the amount of data (default: 1.)\n :attr:`kronecker_structure` (bool, default=True):\n Whether or not the model will use Kronecker structure in the grid\n (set to True unless there is an additive or product decomposition in the prior)\n\n Returns:\n :obj:`int`\n \"\"\"\n # Scale features so they fit inside grid bounds\n num_data = train_inputs.numel() if train_inputs.dim() == 1 else train_inputs.size(-2)\n num_dim = 1 if train_inputs.dim() == 1 else train_inputs.size(-1)\n if kronecker_structure:\n return int(ratio * math.pow(num_data, 1.0 / num_dim))\n else:\n return ratio * num_data\n\n\ndef convert_legacy_grid(grid: torch.Tensor) -> List[torch.Tensor]:\n return [grid[:, i] for i in range(grid.size(-1))]\n\n\ndef create_data_from_grid(grid: List[torch.Tensor]) -> torch.Tensor:\n \"\"\"\n Args:\n :attr:`grid` (List[Tensor])\n Each Tensor is a 1D set of increments for the grid in that dimension\n Returns:\n `grid_data` (Tensor)\n Returns the set of points on the grid going by column-major order\n (due to legacy reasons).\n \"\"\"\n if torch.is_tensor(grid):\n grid = convert_legacy_grid(grid)\n ndims = len(grid)\n assert all(axis.dim() == 1 for axis in grid)\n projections = torch.meshgrid(*grid)\n grid_tensor = torch.stack(projections, axis=-1)\n # Note that if we did\n # grid_data = grid_tensor.reshape(-1, ndims)\n # instead, we would be iterating through the points of our grid from the\n # last data dimension to the first data dimension. However, due to legacy\n # reasons, we need to iterate from the first data dimension to the last data\n # dimension when creating grid_data\n grid_data = grid_tensor.permute(*(reversed(range(ndims + 1)))).reshape(ndims, -1).transpose(0, 1)\n return grid_data\n\n\ndef create_grid(\n grid_sizes: List[int], grid_bounds: List[Tuple[float, float]], extend: bool = True, device=\"cpu\", dtype=torch.float,\n) -> List[torch.Tensor]:\n \"\"\"\n Creates a grid represented by a list of 1D Tensors representing the\n projections of the grid into each dimension\n\n If `extend`, we extend the grid by two points past the specified boundary\n which can be important for getting good grid interpolations\n \"\"\"\n grid = []\n for i in range(len(grid_bounds)):\n grid_diff = float(grid_bounds[i][1] - grid_bounds[i][0]) / (grid_sizes[i] - 2)\n if extend:\n proj = torch.linspace(\n grid_bounds[i][0] - grid_diff, grid_bounds[i][1] + grid_diff, grid_sizes[i], device=device, dtype=dtype,\n )\n else:\n proj = torch.linspace(grid_bounds[i][0], grid_bounds[i][1], grid_sizes[i], device=device, dtype=dtype,)\n grid.append(proj)\n return grid\n", "path": "gpytorch/utils/grid.py"}]}
| 1,815 | 961 |
gh_patches_debug_18841
|
rasdani/github-patches
|
git_diff
|
unionai-oss__pandera-681
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Nullable pandas BooleanDtype should be exposed as `pandera.typing.BOOL`
**Is your feature request related to a problem? Please describe.**
A clear and concise description of what the problem is.
Other nullable types are exposed, eg as `P.INT64`, or `P.STRING`.
`P.BOOL` is missing, even though it is defined.
**Describe the solution you'd like**
A clear and concise description of what you want to happen.
`BOOL` is already defined, so it should be aliased in typing/common.py as
```py
BOOL = pandas_engine.BOOL
```
and added to `GenericDtype` in the same file, then imported again in `typing/__init__.py` to expose it like other generic types.
**Describe alternatives you've considered**
A clear and concise description of any alternative solutions or features you've considered.
We had been using `P.Series[pd.BooleanDtype]`, but this appears not to be prefered in v0.8.0
**Additional context**
Add any other context or screenshots about the feature request here.
</issue>
<code>
[start of pandera/typing/common.py]
1 """Common typing functionality."""
2 # pylint:disable=abstract-method,disable=too-many-ancestors
3
4 import inspect
5 from typing import TYPE_CHECKING, Any, Generic, Type, TypeVar
6
7 import pandas as pd
8 import typing_inspect
9
10 from .. import dtypes
11 from ..engines import numpy_engine, pandas_engine
12
13 Bool = dtypes.Bool #: ``"bool"`` numpy dtype
14 DateTime = dtypes.DateTime #: ``"datetime64[ns]"`` numpy dtype
15 Timedelta = dtypes.Timedelta #: ``"timedelta64[ns]"`` numpy dtype
16 Category = dtypes.Category #: pandas ``"categorical"`` datatype
17 Float = dtypes.Float #: ``"float"`` numpy dtype
18 Float16 = dtypes.Float16 #: ``"float16"`` numpy dtype
19 Float32 = dtypes.Float32 #: ``"float32"`` numpy dtype
20 Float64 = dtypes.Float64 #: ``"float64"`` numpy dtype
21 Int = dtypes.Int #: ``"int"`` numpy dtype
22 Int8 = dtypes.Int8 #: ``"int8"`` numpy dtype
23 Int16 = dtypes.Int16 #: ``"int16"`` numpy dtype
24 Int32 = dtypes.Int32 #: ``"int32"`` numpy dtype
25 Int64 = dtypes.Int64 #: ``"int64"`` numpy dtype
26 UInt8 = dtypes.UInt8 #: ``"uint8"`` numpy dtype
27 UInt16 = dtypes.UInt16 #: ``"uint16"`` numpy dtype
28 UInt32 = dtypes.UInt32 #: ``"uint32"`` numpy dtype
29 UInt64 = dtypes.UInt64 #: ``"uint64"`` numpy dtype
30 INT8 = pandas_engine.INT8 #: ``"Int8"`` pandas dtype:: pandas 0.24.0+
31 INT16 = pandas_engine.INT16 #: ``"Int16"`` pandas dtype: pandas 0.24.0+
32 INT32 = pandas_engine.INT32 #: ``"Int32"`` pandas dtype: pandas 0.24.0+
33 INT64 = pandas_engine.INT64 #: ``"Int64"`` pandas dtype: pandas 0.24.0+
34 UINT8 = pandas_engine.UINT8 #: ``"UInt8"`` pandas dtype:: pandas 0.24.0+
35 UINT16 = pandas_engine.UINT16 #: ``"UInt16"`` pandas dtype: pandas 0.24.0+
36 UINT32 = pandas_engine.UINT32 #: ``"UInt32"`` pandas dtype: pandas 0.24.0+
37 UINT64 = pandas_engine.UINT64 #: ``"UInt64"`` pandas dtype: pandas 0.24.0+
38 Object = numpy_engine.Object #: ``"object"`` numpy dtype
39 String = dtypes.String #: ``"str"`` numpy dtype
40 #: ``"string"`` pandas dtypes: pandas 1.0.0+. For <1.0.0, this enum will
41 #: fall back on the str-as-object-array representation.
42 STRING = pandas_engine.STRING #: ``"str"`` numpy dtype
43
44 GenericDtype = TypeVar( # type: ignore
45 "GenericDtype",
46 bool,
47 int,
48 str,
49 float,
50 pd.core.dtypes.base.ExtensionDtype,
51 Bool,
52 DateTime,
53 Timedelta,
54 Category,
55 Float,
56 Float16,
57 Float32,
58 Float64,
59 Int,
60 Int8,
61 Int16,
62 Int32,
63 Int64,
64 UInt8,
65 UInt16,
66 UInt32,
67 UInt64,
68 INT8,
69 INT16,
70 INT32,
71 INT64,
72 UINT8,
73 UINT16,
74 UINT32,
75 UINT64,
76 Object,
77 String,
78 STRING,
79 covariant=True,
80 )
81 Schema = TypeVar("Schema", bound="SchemaModel") # type: ignore
82
83
84 # pylint:disable=invalid-name
85 if TYPE_CHECKING:
86 T = TypeVar("T") # pragma: no cover
87 else:
88 T = Schema
89
90
91 class DataFrameBase(Generic[T]):
92 # pylint: disable=too-few-public-methods
93 """
94 Pandera Dataframe base class for validating dataframes on
95 initialization.
96 """
97
98 def __setattr__(self, name: str, value: Any) -> None:
99 # pylint: disable=no-member
100 object.__setattr__(self, name, value)
101 if name == "__orig_class__":
102 orig_class = getattr(self, "__orig_class__")
103 class_args = getattr(orig_class, "__args__", None)
104 if any(
105 x.__name__ == "SchemaModel"
106 for x in inspect.getmro(class_args[0])
107 ):
108 schema_model = value.__args__[0]
109
110 # prevent the double validation problem by preventing checks for
111 # dataframes with a defined pandera.schema
112 pandera = getattr(self, "pandera")
113 if (
114 pandera.schema is None
115 or pandera.schema != schema_model.to_schema()
116 ):
117 # pylint: disable=self-cls-assignment
118 self = schema_model.validate(self)
119 pandera.add_schema(schema_model.to_schema())
120
121
122 # pylint:disable=too-few-public-methods
123 class SeriesBase(Generic[GenericDtype]):
124 """Pandera Series base class to use for all pandas-like APIs."""
125
126 def __get__(
127 self, instance: object, owner: Type
128 ) -> str: # pragma: no cover
129 raise AttributeError("Series should resolve to Field-s")
130
131
132 # pylint:disable=too-few-public-methods
133 class IndexBase(Generic[GenericDtype]):
134 """Representation of pandas.Index, only used for type annotation.
135
136 *new in 0.5.0*
137 """
138
139
140 class AnnotationInfo: # pylint:disable=too-few-public-methods
141 """Captures extra information about an annotation.
142
143 Attributes:
144 origin: The non-parameterized generic class.
145 arg: The first generic type (SchemaModel does not support more than
146 1 argument).
147 literal: Whether the annotation is a literal.
148 optional: Whether the annotation is optional.
149 raw_annotation: The raw annotation.
150 metadata: Extra arguments passed to :data:`typing.Annotated`.
151 """
152
153 def __init__(self, raw_annotation: Type) -> None:
154 self._parse_annotation(raw_annotation)
155
156 @property
157 def is_generic_df(self) -> bool:
158 """True if the annotation is a DataFrameBase subclass."""
159 try:
160 if self.origin is None:
161 return False
162 return issubclass(self.origin, DataFrameBase)
163 except TypeError:
164 return False
165
166 def _parse_annotation(self, raw_annotation: Type) -> None:
167 """Parse key information from annotation.
168
169 :param annotation: A subscripted type.
170 :returns: Annotation
171 """
172 self.raw_annotation = raw_annotation
173 self.origin = self.arg = None
174
175 self.optional = typing_inspect.is_optional_type(raw_annotation)
176 if self.optional and typing_inspect.is_union_type(raw_annotation):
177 # Annotated with Optional or Union[..., NoneType]
178 # get_args -> (pandera.typing.Index[str], <class 'NoneType'>)
179 raw_annotation = typing_inspect.get_args(raw_annotation)[0]
180
181 self.origin = typing_inspect.get_origin(raw_annotation)
182 # Replace empty tuple returned from get_args by None
183 args = typing_inspect.get_args(raw_annotation) or None
184 self.arg = args[0] if args else args
185
186 self.metadata = getattr(self.arg, "__metadata__", None)
187 if self.metadata:
188 self.arg = typing_inspect.get_args(self.arg)[0]
189
190 self.literal = typing_inspect.is_literal_type(self.arg)
191 if self.literal:
192 self.arg = typing_inspect.get_args(self.arg)[0]
193
[end of pandera/typing/common.py]
[start of pandera/typing/__init__.py]
1 """Typing module.
2
3 For backwards compatibility, pandas types are exposed to the top-level scope of
4 the typing module.
5 """
6
7 from . import dask, koalas, modin
8 from .common import (
9 INT8,
10 INT16,
11 INT32,
12 INT64,
13 STRING,
14 UINT8,
15 UINT16,
16 UINT32,
17 UINT64,
18 AnnotationInfo,
19 Bool,
20 Category,
21 DateTime,
22 Float,
23 Float16,
24 Float32,
25 Float64,
26 Int,
27 Int8,
28 Int16,
29 Int32,
30 Int64,
31 Object,
32 String,
33 Timedelta,
34 UInt8,
35 UInt16,
36 UInt32,
37 UInt64,
38 )
39 from .pandas import DataFrame, Index, Series
40
41 DATAFRAME_TYPES = {DataFrame}
42 SERIES_TYPES = {Series}
43 INDEX_TYPES = {Index}
44
45 if dask.DASK_INSTALLED:
46 DATAFRAME_TYPES.update({dask.DataFrame})
47 SERIES_TYPES.update({dask.Series})
48 INDEX_TYPES.update({dask.Index})
49
50 if modin.MODIN_INSTALLED:
51 DATAFRAME_TYPES.update({modin.DataFrame})
52 SERIES_TYPES.update({modin.Series})
53 INDEX_TYPES.update({modin.Index})
54
55 if koalas.KOALAS_INSTALLED:
56 DATAFRAME_TYPES.update({koalas.DataFrame})
57 SERIES_TYPES.update({koalas.Series})
58 INDEX_TYPES.update({koalas.Index})
59
[end of pandera/typing/__init__.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/pandera/typing/__init__.py b/pandera/typing/__init__.py
--- a/pandera/typing/__init__.py
+++ b/pandera/typing/__init__.py
@@ -6,6 +6,7 @@
from . import dask, koalas, modin
from .common import (
+ BOOL,
INT8,
INT16,
INT32,
diff --git a/pandera/typing/common.py b/pandera/typing/common.py
--- a/pandera/typing/common.py
+++ b/pandera/typing/common.py
@@ -40,6 +40,7 @@
#: ``"string"`` pandas dtypes: pandas 1.0.0+. For <1.0.0, this enum will
#: fall back on the str-as-object-array representation.
STRING = pandas_engine.STRING #: ``"str"`` numpy dtype
+BOOL = pandas_engine.BOOL #: ``"str"`` numpy dtype
GenericDtype = TypeVar( # type: ignore
"GenericDtype",
@@ -76,6 +77,7 @@
Object,
String,
STRING,
+ BOOL,
covariant=True,
)
Schema = TypeVar("Schema", bound="SchemaModel") # type: ignore
|
{"golden_diff": "diff --git a/pandera/typing/__init__.py b/pandera/typing/__init__.py\n--- a/pandera/typing/__init__.py\n+++ b/pandera/typing/__init__.py\n@@ -6,6 +6,7 @@\n \n from . import dask, koalas, modin\n from .common import (\n+ BOOL,\n INT8,\n INT16,\n INT32,\ndiff --git a/pandera/typing/common.py b/pandera/typing/common.py\n--- a/pandera/typing/common.py\n+++ b/pandera/typing/common.py\n@@ -40,6 +40,7 @@\n #: ``\"string\"`` pandas dtypes: pandas 1.0.0+. For <1.0.0, this enum will\n #: fall back on the str-as-object-array representation.\n STRING = pandas_engine.STRING #: ``\"str\"`` numpy dtype\n+BOOL = pandas_engine.BOOL #: ``\"str\"`` numpy dtype\n \n GenericDtype = TypeVar( # type: ignore\n \"GenericDtype\",\n@@ -76,6 +77,7 @@\n Object,\n String,\n STRING,\n+ BOOL,\n covariant=True,\n )\n Schema = TypeVar(\"Schema\", bound=\"SchemaModel\") # type: ignore\n", "issue": "Nullable pandas BooleanDtype should be exposed as `pandera.typing.BOOL`\n**Is your feature request related to a problem? Please describe.**\r\nA clear and concise description of what the problem is.\r\n\r\nOther nullable types are exposed, eg as `P.INT64`, or `P.STRING`. \r\n\r\n`P.BOOL` is missing, even though it is defined.\r\n\r\n**Describe the solution you'd like**\r\nA clear and concise description of what you want to happen.\r\n\r\n`BOOL` is already defined, so it should be aliased in typing/common.py as \r\n\r\n```py \r\nBOOL = pandas_engine.BOOL\r\n```\r\nand added to `GenericDtype` in the same file, then imported again in `typing/__init__.py` to expose it like other generic types.\r\n\r\n**Describe alternatives you've considered**\r\nA clear and concise description of any alternative solutions or features you've considered.\r\n\r\nWe had been using `P.Series[pd.BooleanDtype]`, but this appears not to be prefered in v0.8.0\r\n\r\n**Additional context**\r\nAdd any other context or screenshots about the feature request here.\r\n\n", "before_files": [{"content": "\"\"\"Common typing functionality.\"\"\"\n# pylint:disable=abstract-method,disable=too-many-ancestors\n\nimport inspect\nfrom typing import TYPE_CHECKING, Any, Generic, Type, TypeVar\n\nimport pandas as pd\nimport typing_inspect\n\nfrom .. import dtypes\nfrom ..engines import numpy_engine, pandas_engine\n\nBool = dtypes.Bool #: ``\"bool\"`` numpy dtype\nDateTime = dtypes.DateTime #: ``\"datetime64[ns]\"`` numpy dtype\nTimedelta = dtypes.Timedelta #: ``\"timedelta64[ns]\"`` numpy dtype\nCategory = dtypes.Category #: pandas ``\"categorical\"`` datatype\nFloat = dtypes.Float #: ``\"float\"`` numpy dtype\nFloat16 = dtypes.Float16 #: ``\"float16\"`` numpy dtype\nFloat32 = dtypes.Float32 #: ``\"float32\"`` numpy dtype\nFloat64 = dtypes.Float64 #: ``\"float64\"`` numpy dtype\nInt = dtypes.Int #: ``\"int\"`` numpy dtype\nInt8 = dtypes.Int8 #: ``\"int8\"`` numpy dtype\nInt16 = dtypes.Int16 #: ``\"int16\"`` numpy dtype\nInt32 = dtypes.Int32 #: ``\"int32\"`` numpy dtype\nInt64 = dtypes.Int64 #: ``\"int64\"`` numpy dtype\nUInt8 = dtypes.UInt8 #: ``\"uint8\"`` numpy dtype\nUInt16 = dtypes.UInt16 #: ``\"uint16\"`` numpy dtype\nUInt32 = dtypes.UInt32 #: ``\"uint32\"`` numpy dtype\nUInt64 = dtypes.UInt64 #: ``\"uint64\"`` numpy dtype\nINT8 = pandas_engine.INT8 #: ``\"Int8\"`` pandas dtype:: pandas 0.24.0+\nINT16 = pandas_engine.INT16 #: ``\"Int16\"`` pandas dtype: pandas 0.24.0+\nINT32 = pandas_engine.INT32 #: ``\"Int32\"`` pandas dtype: pandas 0.24.0+\nINT64 = pandas_engine.INT64 #: ``\"Int64\"`` pandas dtype: pandas 0.24.0+\nUINT8 = pandas_engine.UINT8 #: ``\"UInt8\"`` pandas dtype:: pandas 0.24.0+\nUINT16 = pandas_engine.UINT16 #: ``\"UInt16\"`` pandas dtype: pandas 0.24.0+\nUINT32 = pandas_engine.UINT32 #: ``\"UInt32\"`` pandas dtype: pandas 0.24.0+\nUINT64 = pandas_engine.UINT64 #: ``\"UInt64\"`` pandas dtype: pandas 0.24.0+\nObject = numpy_engine.Object #: ``\"object\"`` numpy dtype\nString = dtypes.String #: ``\"str\"`` numpy dtype\n#: ``\"string\"`` pandas dtypes: pandas 1.0.0+. For <1.0.0, this enum will\n#: fall back on the str-as-object-array representation.\nSTRING = pandas_engine.STRING #: ``\"str\"`` numpy dtype\n\nGenericDtype = TypeVar( # type: ignore\n \"GenericDtype\",\n bool,\n int,\n str,\n float,\n pd.core.dtypes.base.ExtensionDtype,\n Bool,\n DateTime,\n Timedelta,\n Category,\n Float,\n Float16,\n Float32,\n Float64,\n Int,\n Int8,\n Int16,\n Int32,\n Int64,\n UInt8,\n UInt16,\n UInt32,\n UInt64,\n INT8,\n INT16,\n INT32,\n INT64,\n UINT8,\n UINT16,\n UINT32,\n UINT64,\n Object,\n String,\n STRING,\n covariant=True,\n)\nSchema = TypeVar(\"Schema\", bound=\"SchemaModel\") # type: ignore\n\n\n# pylint:disable=invalid-name\nif TYPE_CHECKING:\n T = TypeVar(\"T\") # pragma: no cover\nelse:\n T = Schema\n\n\nclass DataFrameBase(Generic[T]):\n # pylint: disable=too-few-public-methods\n \"\"\"\n Pandera Dataframe base class for validating dataframes on\n initialization.\n \"\"\"\n\n def __setattr__(self, name: str, value: Any) -> None:\n # pylint: disable=no-member\n object.__setattr__(self, name, value)\n if name == \"__orig_class__\":\n orig_class = getattr(self, \"__orig_class__\")\n class_args = getattr(orig_class, \"__args__\", None)\n if any(\n x.__name__ == \"SchemaModel\"\n for x in inspect.getmro(class_args[0])\n ):\n schema_model = value.__args__[0]\n\n # prevent the double validation problem by preventing checks for\n # dataframes with a defined pandera.schema\n pandera = getattr(self, \"pandera\")\n if (\n pandera.schema is None\n or pandera.schema != schema_model.to_schema()\n ):\n # pylint: disable=self-cls-assignment\n self = schema_model.validate(self)\n pandera.add_schema(schema_model.to_schema())\n\n\n# pylint:disable=too-few-public-methods\nclass SeriesBase(Generic[GenericDtype]):\n \"\"\"Pandera Series base class to use for all pandas-like APIs.\"\"\"\n\n def __get__(\n self, instance: object, owner: Type\n ) -> str: # pragma: no cover\n raise AttributeError(\"Series should resolve to Field-s\")\n\n\n# pylint:disable=too-few-public-methods\nclass IndexBase(Generic[GenericDtype]):\n \"\"\"Representation of pandas.Index, only used for type annotation.\n\n *new in 0.5.0*\n \"\"\"\n\n\nclass AnnotationInfo: # pylint:disable=too-few-public-methods\n \"\"\"Captures extra information about an annotation.\n\n Attributes:\n origin: The non-parameterized generic class.\n arg: The first generic type (SchemaModel does not support more than\n 1 argument).\n literal: Whether the annotation is a literal.\n optional: Whether the annotation is optional.\n raw_annotation: The raw annotation.\n metadata: Extra arguments passed to :data:`typing.Annotated`.\n \"\"\"\n\n def __init__(self, raw_annotation: Type) -> None:\n self._parse_annotation(raw_annotation)\n\n @property\n def is_generic_df(self) -> bool:\n \"\"\"True if the annotation is a DataFrameBase subclass.\"\"\"\n try:\n if self.origin is None:\n return False\n return issubclass(self.origin, DataFrameBase)\n except TypeError:\n return False\n\n def _parse_annotation(self, raw_annotation: Type) -> None:\n \"\"\"Parse key information from annotation.\n\n :param annotation: A subscripted type.\n :returns: Annotation\n \"\"\"\n self.raw_annotation = raw_annotation\n self.origin = self.arg = None\n\n self.optional = typing_inspect.is_optional_type(raw_annotation)\n if self.optional and typing_inspect.is_union_type(raw_annotation):\n # Annotated with Optional or Union[..., NoneType]\n # get_args -> (pandera.typing.Index[str], <class 'NoneType'>)\n raw_annotation = typing_inspect.get_args(raw_annotation)[0]\n\n self.origin = typing_inspect.get_origin(raw_annotation)\n # Replace empty tuple returned from get_args by None\n args = typing_inspect.get_args(raw_annotation) or None\n self.arg = args[0] if args else args\n\n self.metadata = getattr(self.arg, \"__metadata__\", None)\n if self.metadata:\n self.arg = typing_inspect.get_args(self.arg)[0]\n\n self.literal = typing_inspect.is_literal_type(self.arg)\n if self.literal:\n self.arg = typing_inspect.get_args(self.arg)[0]\n", "path": "pandera/typing/common.py"}, {"content": "\"\"\"Typing module.\n\nFor backwards compatibility, pandas types are exposed to the top-level scope of\nthe typing module.\n\"\"\"\n\nfrom . import dask, koalas, modin\nfrom .common import (\n INT8,\n INT16,\n INT32,\n INT64,\n STRING,\n UINT8,\n UINT16,\n UINT32,\n UINT64,\n AnnotationInfo,\n Bool,\n Category,\n DateTime,\n Float,\n Float16,\n Float32,\n Float64,\n Int,\n Int8,\n Int16,\n Int32,\n Int64,\n Object,\n String,\n Timedelta,\n UInt8,\n UInt16,\n UInt32,\n UInt64,\n)\nfrom .pandas import DataFrame, Index, Series\n\nDATAFRAME_TYPES = {DataFrame}\nSERIES_TYPES = {Series}\nINDEX_TYPES = {Index}\n\nif dask.DASK_INSTALLED:\n DATAFRAME_TYPES.update({dask.DataFrame})\n SERIES_TYPES.update({dask.Series})\n INDEX_TYPES.update({dask.Index})\n\nif modin.MODIN_INSTALLED:\n DATAFRAME_TYPES.update({modin.DataFrame})\n SERIES_TYPES.update({modin.Series})\n INDEX_TYPES.update({modin.Index})\n\nif koalas.KOALAS_INSTALLED:\n DATAFRAME_TYPES.update({koalas.DataFrame})\n SERIES_TYPES.update({koalas.Series})\n INDEX_TYPES.update({koalas.Index})\n", "path": "pandera/typing/__init__.py"}]}
| 3,444 | 288 |
gh_patches_debug_11978
|
rasdani/github-patches
|
git_diff
|
sublimelsp__LSP-242
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
space wrongly triggers autocompletion
following https://github.com/tomv564/LSP/pull/197#discussion_r151605795
<img width="418" alt="screen shot 2017-11-17 at 12 23 30 am" src="https://user-images.githubusercontent.com/1690993/32930648-f23bb3be-cb2d-11e7-96bf-437671397099.png">
It could be reproduced by typing `y` and `<space>` very quickly or when the language server responds slowly.
We might need to check if the context which triggered the completion is still a prefix of the current cursor.
</issue>
<code>
[start of plugin/completion.py]
1 import sublime
2 import sublime_plugin
3
4 try:
5 from typing import Any, List, Dict, Tuple, Callable, Optional
6 assert Any and List and Dict and Tuple and Callable and Optional
7 except ImportError:
8 pass
9
10 from .core.protocol import Request
11 from .core.settings import settings
12 from .core.logging import debug, exception_log
13 from .core.protocol import CompletionItemKind
14 from .core.clients import client_for_view
15 from .core.configurations import is_supported_syntax
16 from .core.documents import get_document_position, purge_did_change
17
18
19 NO_COMPLETION_SCOPES = 'comment, string'
20 completion_item_kind_names = {v: k for k, v in CompletionItemKind.__dict__.items()}
21
22
23 class CompletionState(object):
24 IDLE = 0
25 REQUESTING = 1
26 APPLYING = 2
27 CANCELLING = 3
28
29
30 resolvable_completion_items = [] # type: List[Any]
31
32
33 def find_completion_item(label: str) -> 'Optional[Any]':
34 matches = list(filter(lambda i: i.get("label") == label, resolvable_completion_items))
35 return matches[0] if matches else None
36
37
38 class CompletionContext(object):
39
40 def __init__(self, begin):
41 self.begin = begin # type: Optional[int]
42 self.end = None # type: Optional[int]
43 self.region = None # type: Optional[sublime.Region]
44 self.committing = False
45
46 def committed_at(self, end):
47 self.end = end
48 self.region = sublime.Region(self.begin, self.end)
49 self.committing = False
50
51
52 current_completion = None # type: Optional[CompletionContext]
53
54
55 def has_resolvable_completions(view):
56 client = client_for_view(view)
57 if client:
58 completionProvider = client.get_capability(
59 'completionProvider')
60 if completionProvider:
61 if completionProvider.get('resolveProvider', False):
62 return True
63 return False
64
65
66 class CompletionSnippetHandler(sublime_plugin.EventListener):
67
68 def on_query_completions(self, view, prefix, locations):
69 global current_completion
70 if settings.resolve_completion_for_snippets and has_resolvable_completions(view):
71 current_completion = CompletionContext(view.sel()[0].begin())
72
73 def on_text_command(self, view, command_name, args):
74 if settings.resolve_completion_for_snippets and current_completion:
75 current_completion.committing = command_name in ('commit_completion', 'insert_best_completion')
76
77 def on_modified(self, view):
78 global current_completion
79
80 if settings.resolve_completion_for_snippets and view.file_name():
81 if current_completion and current_completion.committing:
82 current_completion.committed_at(view.sel()[0].end())
83 inserted = view.substr(current_completion.region)
84 item = find_completion_item(inserted)
85 if item:
86 self.resolve_completion(item, view)
87 else:
88 current_completion = None
89
90 def resolve_completion(self, item, view):
91 client = client_for_view(view)
92 if not client:
93 return
94
95 client.send_request(
96 Request.resolveCompletionItem(item),
97 lambda response: self.handle_resolve_response(response, view))
98
99 def handle_resolve_response(self, response, view):
100 # replace inserted text if a snippet was returned.
101 if current_completion and response.get('insertTextFormat') == 2: # snippet
102 insertText = response.get('insertText')
103 try:
104 sel = view.sel()
105 sel.clear()
106 sel.add(current_completion.region)
107 view.run_command("insert_snippet", {"contents": insertText})
108 except Exception as err:
109 exception_log("Error inserting snippet: " + insertText, err)
110
111
112 last_text_command = None
113
114
115 class CompletionHelper(sublime_plugin.EventListener):
116 def on_text_command(self, view, command_name, args):
117 global last_text_command
118 last_text_command = command_name
119
120
121 class CompletionHandler(sublime_plugin.ViewEventListener):
122 def __init__(self, view):
123 self.view = view
124 self.initialized = False
125 self.enabled = False
126 self.trigger_chars = [] # type: List[str]
127 self.resolve = False
128 self.resolve_details = [] # type: List[Tuple[str, str]]
129 self.state = CompletionState.IDLE
130 self.completions = [] # type: List[Any]
131 self.next_request = None # type: Optional[Tuple[str, List[int]]]
132 self.last_prefix = ""
133 self.last_location = 0
134
135 @classmethod
136 def is_applicable(cls, settings):
137 syntax = settings.get('syntax')
138 if syntax is not None:
139 return is_supported_syntax(syntax)
140 else:
141 return False
142
143 def initialize(self):
144 self.initialized = True
145 client = client_for_view(self.view)
146 if client:
147 completionProvider = client.get_capability(
148 'completionProvider')
149 if completionProvider:
150 self.enabled = True
151 self.trigger_chars = completionProvider.get(
152 'triggerCharacters') or []
153 self.has_resolve_provider = completionProvider.get('resolveProvider', False)
154
155 def is_after_trigger_character(self, location):
156 if location > 0:
157 prev_char = self.view.substr(location - 1)
158 return prev_char in self.trigger_chars
159
160 def is_same_completion(self, prefix, locations):
161 # completion requests from the same location with the same prefix are cached.
162 current_start = locations[0] - len(prefix)
163 last_start = self.last_location - len(self.last_prefix)
164 return prefix.startswith(self.last_prefix) and current_start == last_start
165
166 def on_modified(self):
167 # hide completion when backspacing past last completion.
168 if self.view.sel()[0].begin() < self.last_location:
169 self.last_location = 0
170 self.view.run_command("hide_auto_complete")
171
172 def on_query_completions(self, prefix, locations):
173 if self.view.match_selector(locations[0], NO_COMPLETION_SCOPES):
174 return (
175 [],
176 sublime.INHIBIT_WORD_COMPLETIONS | sublime.INHIBIT_EXPLICIT_COMPLETIONS
177 )
178
179 if not self.initialized:
180 self.initialize()
181
182 if self.enabled:
183 reuse_completion = self.is_same_completion(prefix, locations)
184 if self.state == CompletionState.IDLE:
185 if not reuse_completion:
186 self.last_prefix = prefix
187 self.last_location = locations[0]
188 self.do_request(prefix, locations)
189 self.completions = []
190
191 elif self.state in (CompletionState.REQUESTING, CompletionState.CANCELLING):
192 self.next_request = (prefix, locations)
193 self.state = CompletionState.CANCELLING
194
195 elif self.state == CompletionState.APPLYING:
196 self.state = CompletionState.IDLE
197
198 return (
199 self.completions,
200 0 if not settings.only_show_lsp_completions
201 else sublime.INHIBIT_WORD_COMPLETIONS | sublime.INHIBIT_EXPLICIT_COMPLETIONS
202 )
203
204 def do_request(self, prefix: str, locations: 'List[int]'):
205 self.next_request = None
206 view = self.view
207
208 # don't store client so we can handle restarts
209 client = client_for_view(view)
210 if not client:
211 return
212
213 if settings.complete_all_chars or self.is_after_trigger_character(locations[0]):
214 purge_did_change(view.buffer_id())
215 document_position = get_document_position(view, locations[0])
216 if document_position:
217 client.send_request(
218 Request.complete(document_position),
219 self.handle_response,
220 self.handle_error)
221 self.state = CompletionState.REQUESTING
222
223 def format_completion(self, item: dict) -> 'Tuple[str, str]':
224 # Sublime handles snippets automatically, so we don't have to care about insertTextFormat.
225 label = item["label"]
226 # choose hint based on availability and user preference
227 hint = None
228 if settings.completion_hint_type == "auto":
229 hint = item.get("detail")
230 if not hint:
231 kind = item.get("kind")
232 if kind:
233 hint = completion_item_kind_names[kind]
234 elif settings.completion_hint_type == "detail":
235 hint = item.get("detail")
236 elif settings.completion_hint_type == "kind":
237 kind = item.get("kind")
238 if kind:
239 hint = completion_item_kind_names[kind]
240 # label is an alternative for insertText if insertText not provided
241 insert_text = item.get("insertText") or label
242 if len(insert_text) > 0 and insert_text[0] == '$': # sublime needs leading '$' escaped.
243 insert_text = '\$' + insert_text[1:]
244 # only return label with a hint if available
245 return "\t ".join((label, hint)) if hint else label, insert_text
246
247 def handle_response(self, response: dict):
248 global resolvable_completion_items
249
250 if self.state == CompletionState.REQUESTING:
251 items = response["items"] if isinstance(response,
252 dict) else response
253 if len(items) > 1 and items[0].get("sortText") is not None:
254 # If the first item has a sortText value, assume all of them have a sortText value.
255 items = sorted(items, key=lambda item: item["sortText"])
256 self.completions = list(self.format_completion(item) for item in items)
257
258 if self.has_resolve_provider:
259 resolvable_completion_items = items
260
261 # if insert_best_completion was just ran, undo it before presenting new completions.
262 prev_char = self.view.substr(self.view.sel()[0].begin() - 1)
263 if prev_char.isspace():
264 if last_text_command == "insert_best_completion":
265 self.view.run_command("undo")
266
267 self.state = CompletionState.APPLYING
268 self.view.run_command("hide_auto_complete")
269 self.run_auto_complete()
270 elif self.state == CompletionState.CANCELLING:
271 if self.next_request:
272 prefix, locations = self.next_request
273 self.do_request(prefix, locations)
274 else:
275 debug('Got unexpected response while in state {}'.format(self.state))
276
277 def handle_error(self, error: dict):
278 sublime.status_message('Completion error: ' + str(error.get('message')))
279 self.state = CompletionState.IDLE
280
281 def run_auto_complete(self):
282 self.view.run_command(
283 "auto_complete", {
284 'disable_auto_insert': True,
285 'api_completions_only': settings.only_show_lsp_completions,
286 'next_completion_if_showing': False
287 })
288
[end of plugin/completion.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/plugin/completion.py b/plugin/completion.py
--- a/plugin/completion.py
+++ b/plugin/completion.py
@@ -168,6 +168,10 @@
if self.view.sel()[0].begin() < self.last_location:
self.last_location = 0
self.view.run_command("hide_auto_complete")
+ # cancel current completion if the previous input is an space
+ prev_char = self.view.substr(self.view.sel()[0].begin() - 1)
+ if self.state == CompletionState.REQUESTING and prev_char.isspace():
+ self.state = CompletionState.CANCELLING
def on_query_completions(self, prefix, locations):
if self.view.match_selector(locations[0], NO_COMPLETION_SCOPES):
|
{"golden_diff": "diff --git a/plugin/completion.py b/plugin/completion.py\n--- a/plugin/completion.py\n+++ b/plugin/completion.py\n@@ -168,6 +168,10 @@\n if self.view.sel()[0].begin() < self.last_location:\n self.last_location = 0\n self.view.run_command(\"hide_auto_complete\")\n+ # cancel current completion if the previous input is an space\n+ prev_char = self.view.substr(self.view.sel()[0].begin() - 1)\n+ if self.state == CompletionState.REQUESTING and prev_char.isspace():\n+ self.state = CompletionState.CANCELLING\n \n def on_query_completions(self, prefix, locations):\n if self.view.match_selector(locations[0], NO_COMPLETION_SCOPES):\n", "issue": "space wrongly triggers autocompletion\nfollowing https://github.com/tomv564/LSP/pull/197#discussion_r151605795\r\n\r\n<img width=\"418\" alt=\"screen shot 2017-11-17 at 12 23 30 am\" src=\"https://user-images.githubusercontent.com/1690993/32930648-f23bb3be-cb2d-11e7-96bf-437671397099.png\">\r\n\r\nIt could be reproduced by typing `y` and `<space>` very quickly or when the language server responds slowly.\r\n\r\nWe might need to check if the context which triggered the completion is still a prefix of the current cursor.\n", "before_files": [{"content": "import sublime\nimport sublime_plugin\n\ntry:\n from typing import Any, List, Dict, Tuple, Callable, Optional\n assert Any and List and Dict and Tuple and Callable and Optional\nexcept ImportError:\n pass\n\nfrom .core.protocol import Request\nfrom .core.settings import settings\nfrom .core.logging import debug, exception_log\nfrom .core.protocol import CompletionItemKind\nfrom .core.clients import client_for_view\nfrom .core.configurations import is_supported_syntax\nfrom .core.documents import get_document_position, purge_did_change\n\n\nNO_COMPLETION_SCOPES = 'comment, string'\ncompletion_item_kind_names = {v: k for k, v in CompletionItemKind.__dict__.items()}\n\n\nclass CompletionState(object):\n IDLE = 0\n REQUESTING = 1\n APPLYING = 2\n CANCELLING = 3\n\n\nresolvable_completion_items = [] # type: List[Any]\n\n\ndef find_completion_item(label: str) -> 'Optional[Any]':\n matches = list(filter(lambda i: i.get(\"label\") == label, resolvable_completion_items))\n return matches[0] if matches else None\n\n\nclass CompletionContext(object):\n\n def __init__(self, begin):\n self.begin = begin # type: Optional[int]\n self.end = None # type: Optional[int]\n self.region = None # type: Optional[sublime.Region]\n self.committing = False\n\n def committed_at(self, end):\n self.end = end\n self.region = sublime.Region(self.begin, self.end)\n self.committing = False\n\n\ncurrent_completion = None # type: Optional[CompletionContext]\n\n\ndef has_resolvable_completions(view):\n client = client_for_view(view)\n if client:\n completionProvider = client.get_capability(\n 'completionProvider')\n if completionProvider:\n if completionProvider.get('resolveProvider', False):\n return True\n return False\n\n\nclass CompletionSnippetHandler(sublime_plugin.EventListener):\n\n def on_query_completions(self, view, prefix, locations):\n global current_completion\n if settings.resolve_completion_for_snippets and has_resolvable_completions(view):\n current_completion = CompletionContext(view.sel()[0].begin())\n\n def on_text_command(self, view, command_name, args):\n if settings.resolve_completion_for_snippets and current_completion:\n current_completion.committing = command_name in ('commit_completion', 'insert_best_completion')\n\n def on_modified(self, view):\n global current_completion\n\n if settings.resolve_completion_for_snippets and view.file_name():\n if current_completion and current_completion.committing:\n current_completion.committed_at(view.sel()[0].end())\n inserted = view.substr(current_completion.region)\n item = find_completion_item(inserted)\n if item:\n self.resolve_completion(item, view)\n else:\n current_completion = None\n\n def resolve_completion(self, item, view):\n client = client_for_view(view)\n if not client:\n return\n\n client.send_request(\n Request.resolveCompletionItem(item),\n lambda response: self.handle_resolve_response(response, view))\n\n def handle_resolve_response(self, response, view):\n # replace inserted text if a snippet was returned.\n if current_completion and response.get('insertTextFormat') == 2: # snippet\n insertText = response.get('insertText')\n try:\n sel = view.sel()\n sel.clear()\n sel.add(current_completion.region)\n view.run_command(\"insert_snippet\", {\"contents\": insertText})\n except Exception as err:\n exception_log(\"Error inserting snippet: \" + insertText, err)\n\n\nlast_text_command = None\n\n\nclass CompletionHelper(sublime_plugin.EventListener):\n def on_text_command(self, view, command_name, args):\n global last_text_command\n last_text_command = command_name\n\n\nclass CompletionHandler(sublime_plugin.ViewEventListener):\n def __init__(self, view):\n self.view = view\n self.initialized = False\n self.enabled = False\n self.trigger_chars = [] # type: List[str]\n self.resolve = False\n self.resolve_details = [] # type: List[Tuple[str, str]]\n self.state = CompletionState.IDLE\n self.completions = [] # type: List[Any]\n self.next_request = None # type: Optional[Tuple[str, List[int]]]\n self.last_prefix = \"\"\n self.last_location = 0\n\n @classmethod\n def is_applicable(cls, settings):\n syntax = settings.get('syntax')\n if syntax is not None:\n return is_supported_syntax(syntax)\n else:\n return False\n\n def initialize(self):\n self.initialized = True\n client = client_for_view(self.view)\n if client:\n completionProvider = client.get_capability(\n 'completionProvider')\n if completionProvider:\n self.enabled = True\n self.trigger_chars = completionProvider.get(\n 'triggerCharacters') or []\n self.has_resolve_provider = completionProvider.get('resolveProvider', False)\n\n def is_after_trigger_character(self, location):\n if location > 0:\n prev_char = self.view.substr(location - 1)\n return prev_char in self.trigger_chars\n\n def is_same_completion(self, prefix, locations):\n # completion requests from the same location with the same prefix are cached.\n current_start = locations[0] - len(prefix)\n last_start = self.last_location - len(self.last_prefix)\n return prefix.startswith(self.last_prefix) and current_start == last_start\n\n def on_modified(self):\n # hide completion when backspacing past last completion.\n if self.view.sel()[0].begin() < self.last_location:\n self.last_location = 0\n self.view.run_command(\"hide_auto_complete\")\n\n def on_query_completions(self, prefix, locations):\n if self.view.match_selector(locations[0], NO_COMPLETION_SCOPES):\n return (\n [],\n sublime.INHIBIT_WORD_COMPLETIONS | sublime.INHIBIT_EXPLICIT_COMPLETIONS\n )\n\n if not self.initialized:\n self.initialize()\n\n if self.enabled:\n reuse_completion = self.is_same_completion(prefix, locations)\n if self.state == CompletionState.IDLE:\n if not reuse_completion:\n self.last_prefix = prefix\n self.last_location = locations[0]\n self.do_request(prefix, locations)\n self.completions = []\n\n elif self.state in (CompletionState.REQUESTING, CompletionState.CANCELLING):\n self.next_request = (prefix, locations)\n self.state = CompletionState.CANCELLING\n\n elif self.state == CompletionState.APPLYING:\n self.state = CompletionState.IDLE\n\n return (\n self.completions,\n 0 if not settings.only_show_lsp_completions\n else sublime.INHIBIT_WORD_COMPLETIONS | sublime.INHIBIT_EXPLICIT_COMPLETIONS\n )\n\n def do_request(self, prefix: str, locations: 'List[int]'):\n self.next_request = None\n view = self.view\n\n # don't store client so we can handle restarts\n client = client_for_view(view)\n if not client:\n return\n\n if settings.complete_all_chars or self.is_after_trigger_character(locations[0]):\n purge_did_change(view.buffer_id())\n document_position = get_document_position(view, locations[0])\n if document_position:\n client.send_request(\n Request.complete(document_position),\n self.handle_response,\n self.handle_error)\n self.state = CompletionState.REQUESTING\n\n def format_completion(self, item: dict) -> 'Tuple[str, str]':\n # Sublime handles snippets automatically, so we don't have to care about insertTextFormat.\n label = item[\"label\"]\n # choose hint based on availability and user preference\n hint = None\n if settings.completion_hint_type == \"auto\":\n hint = item.get(\"detail\")\n if not hint:\n kind = item.get(\"kind\")\n if kind:\n hint = completion_item_kind_names[kind]\n elif settings.completion_hint_type == \"detail\":\n hint = item.get(\"detail\")\n elif settings.completion_hint_type == \"kind\":\n kind = item.get(\"kind\")\n if kind:\n hint = completion_item_kind_names[kind]\n # label is an alternative for insertText if insertText not provided\n insert_text = item.get(\"insertText\") or label\n if len(insert_text) > 0 and insert_text[0] == '$': # sublime needs leading '$' escaped.\n insert_text = '\\$' + insert_text[1:]\n # only return label with a hint if available\n return \"\\t \".join((label, hint)) if hint else label, insert_text\n\n def handle_response(self, response: dict):\n global resolvable_completion_items\n\n if self.state == CompletionState.REQUESTING:\n items = response[\"items\"] if isinstance(response,\n dict) else response\n if len(items) > 1 and items[0].get(\"sortText\") is not None:\n # If the first item has a sortText value, assume all of them have a sortText value.\n items = sorted(items, key=lambda item: item[\"sortText\"])\n self.completions = list(self.format_completion(item) for item in items)\n\n if self.has_resolve_provider:\n resolvable_completion_items = items\n\n # if insert_best_completion was just ran, undo it before presenting new completions.\n prev_char = self.view.substr(self.view.sel()[0].begin() - 1)\n if prev_char.isspace():\n if last_text_command == \"insert_best_completion\":\n self.view.run_command(\"undo\")\n\n self.state = CompletionState.APPLYING\n self.view.run_command(\"hide_auto_complete\")\n self.run_auto_complete()\n elif self.state == CompletionState.CANCELLING:\n if self.next_request:\n prefix, locations = self.next_request\n self.do_request(prefix, locations)\n else:\n debug('Got unexpected response while in state {}'.format(self.state))\n\n def handle_error(self, error: dict):\n sublime.status_message('Completion error: ' + str(error.get('message')))\n self.state = CompletionState.IDLE\n\n def run_auto_complete(self):\n self.view.run_command(\n \"auto_complete\", {\n 'disable_auto_insert': True,\n 'api_completions_only': settings.only_show_lsp_completions,\n 'next_completion_if_showing': False\n })\n", "path": "plugin/completion.py"}]}
| 3,725 | 169 |
gh_patches_debug_24180
|
rasdani/github-patches
|
git_diff
|
pypi__warehouse-960
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Default search message
We need to add a message for when a search doesn't match any packages.
</issue>
<code>
[start of warehouse/utils/paginate.py]
1 # Licensed under the Apache License, Version 2.0 (the "License");
2 # you may not use this file except in compliance with the License.
3 # You may obtain a copy of the License at
4 #
5 # http://www.apache.org/licenses/LICENSE-2.0
6 #
7 # Unless required by applicable law or agreed to in writing, software
8 # distributed under the License is distributed on an "AS IS" BASIS,
9 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
10 # See the License for the specific language governing permissions and
11 # limitations under the License.
12
13 from paginate import Page
14
15
16 class _ElasticsearchWrapper:
17
18 def __init__(self, query):
19 self.query = query
20 self.results = None
21
22 def __getitem__(self, range):
23 if self.results is not None:
24 raise RuntimeError("Cannot reslice after having already sliced.")
25 self.results = self.query[range].execute()
26
27 return list(self.results)
28
29 def __len__(self):
30 if self.results is None:
31 raise RuntimeError("Cannot get length until a slice.")
32 return self.results.hits.total
33
34
35 def ElasticsearchPage(*args, **kwargs): # noqa
36 kwargs.setdefault("wrapper_class", _ElasticsearchWrapper)
37 return Page(*args, **kwargs)
38
39
40 def paginate_url_factory(request, query_arg="page"):
41 def make_url(page):
42 query_seq = [
43 (k, v)
44 for k, vs in request.GET.dict_of_lists().items()
45 for v in vs
46 if k != query_arg
47 ]
48 query_seq += [(query_arg, page)]
49 return request.current_route_path(_query=query_seq)
50 return make_url
51
[end of warehouse/utils/paginate.py]
[start of warehouse/views.py]
1 # Licensed under the Apache License, Version 2.0 (the "License");
2 # you may not use this file except in compliance with the License.
3 # You may obtain a copy of the License at
4 #
5 # http://www.apache.org/licenses/LICENSE-2.0
6 #
7 # Unless required by applicable law or agreed to in writing, software
8 # distributed under the License is distributed on an "AS IS" BASIS,
9 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
10 # See the License for the specific language governing permissions and
11 # limitations under the License.
12
13 from pyramid.httpexceptions import (
14 HTTPException, HTTPSeeOther, HTTPMovedPermanently,
15 )
16 from pyramid.view import (
17 notfound_view_config, forbidden_view_config, view_config,
18 )
19 from sqlalchemy import func
20 from sqlalchemy.orm import aliased, joinedload
21
22 from warehouse.accounts import REDIRECT_FIELD_NAME
23 from warehouse.accounts.models import User
24 from warehouse.cache.origin import origin_cache
25 from warehouse.cache.http import cache_control
26 from warehouse.csrf import csrf_exempt
27 from warehouse.packaging.models import Project, Release, File
28 from warehouse.sessions import uses_session
29 from warehouse.utils.row_counter import RowCount
30 from warehouse.utils.paginate import ElasticsearchPage, paginate_url_factory
31
32
33 @view_config(context=HTTPException, decorator=[csrf_exempt])
34 @notfound_view_config(
35 append_slash=HTTPMovedPermanently,
36 decorator=[csrf_exempt],
37 )
38 def httpexception_view(exc, request):
39 return exc
40
41
42 @forbidden_view_config()
43 def forbidden(exc, request):
44 # If the forbidden error is because the user isn't logged in, then we'll
45 # redirect them to the log in page.
46 if request.authenticated_userid is None:
47 url = request.route_url(
48 "accounts.login",
49 _query={REDIRECT_FIELD_NAME: request.path_qs},
50 )
51 return HTTPSeeOther(url)
52
53 # If we've reached here, then the user is logged in and they are genuinely
54 # not allowed to access this page.
55 # TODO: Style the forbidden page.
56 return exc
57
58
59 @view_config(
60 route_name="robots.txt",
61 renderer="robots.txt",
62 decorator=[
63 cache_control(1 * 24 * 60 * 60), # 1 day
64 origin_cache(
65 1 * 24 * 60 * 60, # 1 day
66 stale_while_revalidate=6 * 60 * 60, # 6 hours
67 stale_if_error=1 * 24 * 60 * 60, # 1 day
68 ),
69 ],
70 )
71 def robotstxt(request):
72 request.response.content_type = "text/plain"
73 return {}
74
75
76 @view_config(
77 route_name="index",
78 renderer="index.html",
79 decorator=[
80 origin_cache(
81 1 * 60 * 60, # 1 hour
82 stale_while_revalidate=10 * 60, # 10 minutes
83 stale_if_error=1 * 24 * 60 * 60, # 1 day
84 keys=["all-projects"],
85 ),
86 ]
87 )
88 def index(request):
89 project_names = [
90 r[0] for r in (
91 request.db.query(File.name)
92 .group_by(File.name)
93 .order_by(func.sum(File.downloads).desc())
94 .limit(5)
95 .all())
96 ]
97 release_a = aliased(
98 Release,
99 request.db.query(Release)
100 .distinct(Release.name)
101 .filter(Release.name.in_(project_names))
102 .order_by(Release.name, Release._pypi_ordering.desc())
103 .subquery(),
104 )
105 top_projects = (
106 request.db.query(release_a)
107 .options(joinedload(release_a.project),
108 joinedload(release_a.uploader))
109 .order_by(func.array_idx(project_names, release_a.name))
110 .all()
111 )
112
113 latest_releases = (
114 request.db.query(Release)
115 .options(joinedload(Release.project),
116 joinedload(Release.uploader))
117 .order_by(Release.created.desc())
118 .limit(5)
119 .all()
120 )
121
122 counts = dict(
123 request.db.query(RowCount.table_name, RowCount.count)
124 .filter(
125 RowCount.table_name.in_([
126 Project.__tablename__,
127 Release.__tablename__,
128 File.__tablename__,
129 User.__tablename__,
130 ]))
131 .all()
132 )
133
134 return {
135 "latest_releases": latest_releases,
136 "top_projects": top_projects,
137 "num_projects": counts.get(Project.__tablename__, 0),
138 "num_releases": counts.get(Release.__tablename__, 0),
139 "num_files": counts.get(File.__tablename__, 0),
140 "num_users": counts.get(User.__tablename__, 0),
141 }
142
143
144 @view_config(
145 route_name="search",
146 renderer="search/results.html",
147 decorator=[
148 origin_cache(
149 1 * 60 * 60, # 1 hour
150 stale_while_revalidate=10 * 60, # 10 minutes
151 stale_if_error=1 * 24 * 60 * 60, # 1 day
152 keys=["all-projects"],
153 )
154 ],
155 )
156 def search(request):
157 if request.params.get("q"):
158 query = request.es.query(
159 "multi_match",
160 query=request.params["q"],
161 fields=[
162 "name", "version", "author", "author_email", "maintainer",
163 "maintainer_email", "home_page", "license", "summary",
164 "description", "keywords", "platform", "download_url",
165 ],
166 )
167 else:
168 query = request.es.query()
169
170 page = ElasticsearchPage(
171 query,
172 page=int(request.params.get("page", 1)),
173 url_maker=paginate_url_factory(request),
174 )
175
176 return {"page": page, "term": request.params.get("q")}
177
178
179 @view_config(
180 route_name="includes.current-user-indicator",
181 renderer="includes/current-user-indicator.html",
182 decorator=[uses_session],
183 )
184 def current_user_indicator(request):
185 return {}
186
[end of warehouse/views.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/warehouse/utils/paginate.py b/warehouse/utils/paginate.py
--- a/warehouse/utils/paginate.py
+++ b/warehouse/utils/paginate.py
@@ -18,12 +18,18 @@
def __init__(self, query):
self.query = query
self.results = None
+ self.best_guess = None
def __getitem__(self, range):
if self.results is not None:
raise RuntimeError("Cannot reslice after having already sliced.")
self.results = self.query[range].execute()
+ if hasattr(self.results, "suggest"):
+ suggestion = self.results.suggest.name_suggestion[0]
+ if suggestion.options:
+ self.best_guess = suggestion.options[0]
+
return list(self.results)
def __len__(self):
diff --git a/warehouse/views.py b/warehouse/views.py
--- a/warehouse/views.py
+++ b/warehouse/views.py
@@ -163,6 +163,10 @@
"maintainer_email", "home_page", "license", "summary",
"description", "keywords", "platform", "download_url",
],
+ ).suggest(
+ name="name_suggestion",
+ text=request.params["q"],
+ term={"field": "name"}
)
else:
query = request.es.query()
|
{"golden_diff": "diff --git a/warehouse/utils/paginate.py b/warehouse/utils/paginate.py\n--- a/warehouse/utils/paginate.py\n+++ b/warehouse/utils/paginate.py\n@@ -18,12 +18,18 @@\n def __init__(self, query):\n self.query = query\n self.results = None\n+ self.best_guess = None\n \n def __getitem__(self, range):\n if self.results is not None:\n raise RuntimeError(\"Cannot reslice after having already sliced.\")\n self.results = self.query[range].execute()\n \n+ if hasattr(self.results, \"suggest\"):\n+ suggestion = self.results.suggest.name_suggestion[0]\n+ if suggestion.options:\n+ self.best_guess = suggestion.options[0]\n+\n return list(self.results)\n \n def __len__(self):\ndiff --git a/warehouse/views.py b/warehouse/views.py\n--- a/warehouse/views.py\n+++ b/warehouse/views.py\n@@ -163,6 +163,10 @@\n \"maintainer_email\", \"home_page\", \"license\", \"summary\",\n \"description\", \"keywords\", \"platform\", \"download_url\",\n ],\n+ ).suggest(\n+ name=\"name_suggestion\",\n+ text=request.params[\"q\"],\n+ term={\"field\": \"name\"}\n )\n else:\n query = request.es.query()\n", "issue": "Default search message\nWe need to add a message for when a search doesn't match any packages.\n\n", "before_files": [{"content": "# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom paginate import Page\n\n\nclass _ElasticsearchWrapper:\n\n def __init__(self, query):\n self.query = query\n self.results = None\n\n def __getitem__(self, range):\n if self.results is not None:\n raise RuntimeError(\"Cannot reslice after having already sliced.\")\n self.results = self.query[range].execute()\n\n return list(self.results)\n\n def __len__(self):\n if self.results is None:\n raise RuntimeError(\"Cannot get length until a slice.\")\n return self.results.hits.total\n\n\ndef ElasticsearchPage(*args, **kwargs): # noqa\n kwargs.setdefault(\"wrapper_class\", _ElasticsearchWrapper)\n return Page(*args, **kwargs)\n\n\ndef paginate_url_factory(request, query_arg=\"page\"):\n def make_url(page):\n query_seq = [\n (k, v)\n for k, vs in request.GET.dict_of_lists().items()\n for v in vs\n if k != query_arg\n ]\n query_seq += [(query_arg, page)]\n return request.current_route_path(_query=query_seq)\n return make_url\n", "path": "warehouse/utils/paginate.py"}, {"content": "# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom pyramid.httpexceptions import (\n HTTPException, HTTPSeeOther, HTTPMovedPermanently,\n)\nfrom pyramid.view import (\n notfound_view_config, forbidden_view_config, view_config,\n)\nfrom sqlalchemy import func\nfrom sqlalchemy.orm import aliased, joinedload\n\nfrom warehouse.accounts import REDIRECT_FIELD_NAME\nfrom warehouse.accounts.models import User\nfrom warehouse.cache.origin import origin_cache\nfrom warehouse.cache.http import cache_control\nfrom warehouse.csrf import csrf_exempt\nfrom warehouse.packaging.models import Project, Release, File\nfrom warehouse.sessions import uses_session\nfrom warehouse.utils.row_counter import RowCount\nfrom warehouse.utils.paginate import ElasticsearchPage, paginate_url_factory\n\n\n@view_config(context=HTTPException, decorator=[csrf_exempt])\n@notfound_view_config(\n append_slash=HTTPMovedPermanently,\n decorator=[csrf_exempt],\n)\ndef httpexception_view(exc, request):\n return exc\n\n\n@forbidden_view_config()\ndef forbidden(exc, request):\n # If the forbidden error is because the user isn't logged in, then we'll\n # redirect them to the log in page.\n if request.authenticated_userid is None:\n url = request.route_url(\n \"accounts.login\",\n _query={REDIRECT_FIELD_NAME: request.path_qs},\n )\n return HTTPSeeOther(url)\n\n # If we've reached here, then the user is logged in and they are genuinely\n # not allowed to access this page.\n # TODO: Style the forbidden page.\n return exc\n\n\n@view_config(\n route_name=\"robots.txt\",\n renderer=\"robots.txt\",\n decorator=[\n cache_control(1 * 24 * 60 * 60), # 1 day\n origin_cache(\n 1 * 24 * 60 * 60, # 1 day\n stale_while_revalidate=6 * 60 * 60, # 6 hours\n stale_if_error=1 * 24 * 60 * 60, # 1 day\n ),\n ],\n)\ndef robotstxt(request):\n request.response.content_type = \"text/plain\"\n return {}\n\n\n@view_config(\n route_name=\"index\",\n renderer=\"index.html\",\n decorator=[\n origin_cache(\n 1 * 60 * 60, # 1 hour\n stale_while_revalidate=10 * 60, # 10 minutes\n stale_if_error=1 * 24 * 60 * 60, # 1 day\n keys=[\"all-projects\"],\n ),\n ]\n)\ndef index(request):\n project_names = [\n r[0] for r in (\n request.db.query(File.name)\n .group_by(File.name)\n .order_by(func.sum(File.downloads).desc())\n .limit(5)\n .all())\n ]\n release_a = aliased(\n Release,\n request.db.query(Release)\n .distinct(Release.name)\n .filter(Release.name.in_(project_names))\n .order_by(Release.name, Release._pypi_ordering.desc())\n .subquery(),\n )\n top_projects = (\n request.db.query(release_a)\n .options(joinedload(release_a.project),\n joinedload(release_a.uploader))\n .order_by(func.array_idx(project_names, release_a.name))\n .all()\n )\n\n latest_releases = (\n request.db.query(Release)\n .options(joinedload(Release.project),\n joinedload(Release.uploader))\n .order_by(Release.created.desc())\n .limit(5)\n .all()\n )\n\n counts = dict(\n request.db.query(RowCount.table_name, RowCount.count)\n .filter(\n RowCount.table_name.in_([\n Project.__tablename__,\n Release.__tablename__,\n File.__tablename__,\n User.__tablename__,\n ]))\n .all()\n )\n\n return {\n \"latest_releases\": latest_releases,\n \"top_projects\": top_projects,\n \"num_projects\": counts.get(Project.__tablename__, 0),\n \"num_releases\": counts.get(Release.__tablename__, 0),\n \"num_files\": counts.get(File.__tablename__, 0),\n \"num_users\": counts.get(User.__tablename__, 0),\n }\n\n\n@view_config(\n route_name=\"search\",\n renderer=\"search/results.html\",\n decorator=[\n origin_cache(\n 1 * 60 * 60, # 1 hour\n stale_while_revalidate=10 * 60, # 10 minutes\n stale_if_error=1 * 24 * 60 * 60, # 1 day\n keys=[\"all-projects\"],\n )\n ],\n)\ndef search(request):\n if request.params.get(\"q\"):\n query = request.es.query(\n \"multi_match\",\n query=request.params[\"q\"],\n fields=[\n \"name\", \"version\", \"author\", \"author_email\", \"maintainer\",\n \"maintainer_email\", \"home_page\", \"license\", \"summary\",\n \"description\", \"keywords\", \"platform\", \"download_url\",\n ],\n )\n else:\n query = request.es.query()\n\n page = ElasticsearchPage(\n query,\n page=int(request.params.get(\"page\", 1)),\n url_maker=paginate_url_factory(request),\n )\n\n return {\"page\": page, \"term\": request.params.get(\"q\")}\n\n\n@view_config(\n route_name=\"includes.current-user-indicator\",\n renderer=\"includes/current-user-indicator.html\",\n decorator=[uses_session],\n)\ndef current_user_indicator(request):\n return {}\n", "path": "warehouse/views.py"}]}
| 2,801 | 297 |
gh_patches_debug_63588
|
rasdani/github-patches
|
git_diff
|
docker__docker-py-2687
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Plugin model upgrade: object has no attribute '_reload'
I've already got a fix for this that I'm about to submit. For reference, here are the steps to reproduce in ipython:
```
~$ ipython [venv]
Python 3.7.6 (default, Dec 22 2019, 01:09:06)
Type 'copyright', 'credits' or 'license' for more information
IPython 7.18.1 -- An enhanced Interactive Python. Type '?' for help.
In [1]: import docker
In [2]: c = docker.DockerClient(base_url='unix://var/run/docker.sock')
In [3]: p = c.plugins.get('loki')
In [4]: list(p.upgrade("grafana/loki-docker-driver:latest"))
---------------------------------------------------------------------------
AttributeError Traceback (most recent call last)
<ipython-input-4-e5fc26fa0514> in <module>
----> 1 list(p.upgrade("grafana/loki-docker-driver:latest"))
~/venv/lib/python3.7/site-packages/docker/models/plugins.py in upgrade(self, remote)
120 for d in self.client.api.upgrade_plugin(self.name, remote, privileges):
121 yield d
--> 122 self._reload()
123
124
AttributeError: 'Plugin' object has no attribute '_reload'
```
```
docker==4.3.1
Python 3.7.6
Client: Docker Engine - Community
Cloud integration 0.1.18
Version: 19.03.13
API version: 1.40
Go version: go1.13.15
Git commit: 4484c46d9d
Built: Wed Sep 16 16:58:31 2020
OS/Arch: darwin/amd64
Experimental: true
Server: Docker Engine - Community
Engine:
Version: 19.03.13
API version: 1.40 (minimum version 1.12)
Go version: go1.13.15
Git commit: 4484c46d9d
Built: Wed Sep 16 17:07:04 2020
OS/Arch: linux/amd64
Experimental: true
containerd:
Version: v1.3.7
GitCommit: 8fba4e9a7d01810a393d5d25a3621dc101981175
runc:
Version: 1.0.0-rc10
GitCommit: dc9208a3303feef5b3839f4323d9beb36df0a9dd
docker-init:
Version: 0.18.0
GitCommit: fec3683
```
</issue>
<code>
[start of docker/models/plugins.py]
1 from .. import errors
2 from .resource import Collection, Model
3
4
5 class Plugin(Model):
6 """
7 A plugin on the server.
8 """
9 def __repr__(self):
10 return "<%s: '%s'>" % (self.__class__.__name__, self.name)
11
12 @property
13 def name(self):
14 """
15 The plugin's name.
16 """
17 return self.attrs.get('Name')
18
19 @property
20 def enabled(self):
21 """
22 Whether the plugin is enabled.
23 """
24 return self.attrs.get('Enabled')
25
26 @property
27 def settings(self):
28 """
29 A dictionary representing the plugin's configuration.
30 """
31 return self.attrs.get('Settings')
32
33 def configure(self, options):
34 """
35 Update the plugin's settings.
36
37 Args:
38 options (dict): A key-value mapping of options.
39
40 Raises:
41 :py:class:`docker.errors.APIError`
42 If the server returns an error.
43 """
44 self.client.api.configure_plugin(self.name, options)
45 self.reload()
46
47 def disable(self):
48 """
49 Disable the plugin.
50
51 Raises:
52 :py:class:`docker.errors.APIError`
53 If the server returns an error.
54 """
55
56 self.client.api.disable_plugin(self.name)
57 self.reload()
58
59 def enable(self, timeout=0):
60 """
61 Enable the plugin.
62
63 Args:
64 timeout (int): Timeout in seconds. Default: 0
65
66 Raises:
67 :py:class:`docker.errors.APIError`
68 If the server returns an error.
69 """
70 self.client.api.enable_plugin(self.name, timeout)
71 self.reload()
72
73 def push(self):
74 """
75 Push the plugin to a remote registry.
76
77 Returns:
78 A dict iterator streaming the status of the upload.
79
80 Raises:
81 :py:class:`docker.errors.APIError`
82 If the server returns an error.
83 """
84 return self.client.api.push_plugin(self.name)
85
86 def remove(self, force=False):
87 """
88 Remove the plugin from the server.
89
90 Args:
91 force (bool): Remove even if the plugin is enabled.
92 Default: False
93
94 Raises:
95 :py:class:`docker.errors.APIError`
96 If the server returns an error.
97 """
98 return self.client.api.remove_plugin(self.name, force=force)
99
100 def upgrade(self, remote=None):
101 """
102 Upgrade the plugin.
103
104 Args:
105 remote (string): Remote reference to upgrade to. The
106 ``:latest`` tag is optional and is the default if omitted.
107 Default: this plugin's name.
108
109 Returns:
110 A generator streaming the decoded API logs
111 """
112 if self.enabled:
113 raise errors.DockerError(
114 'Plugin must be disabled before upgrading.'
115 )
116
117 if remote is None:
118 remote = self.name
119 privileges = self.client.api.plugin_privileges(remote)
120 for d in self.client.api.upgrade_plugin(self.name, remote, privileges):
121 yield d
122 self._reload()
123
124
125 class PluginCollection(Collection):
126 model = Plugin
127
128 def create(self, name, plugin_data_dir, gzip=False):
129 """
130 Create a new plugin.
131
132 Args:
133 name (string): The name of the plugin. The ``:latest`` tag is
134 optional, and is the default if omitted.
135 plugin_data_dir (string): Path to the plugin data directory.
136 Plugin data directory must contain the ``config.json``
137 manifest file and the ``rootfs`` directory.
138 gzip (bool): Compress the context using gzip. Default: False
139
140 Returns:
141 (:py:class:`Plugin`): The newly created plugin.
142 """
143 self.client.api.create_plugin(name, plugin_data_dir, gzip)
144 return self.get(name)
145
146 def get(self, name):
147 """
148 Gets a plugin.
149
150 Args:
151 name (str): The name of the plugin.
152
153 Returns:
154 (:py:class:`Plugin`): The plugin.
155
156 Raises:
157 :py:class:`docker.errors.NotFound` If the plugin does not
158 exist.
159 :py:class:`docker.errors.APIError`
160 If the server returns an error.
161 """
162 return self.prepare_model(self.client.api.inspect_plugin(name))
163
164 def install(self, remote_name, local_name=None):
165 """
166 Pull and install a plugin.
167
168 Args:
169 remote_name (string): Remote reference for the plugin to
170 install. The ``:latest`` tag is optional, and is the
171 default if omitted.
172 local_name (string): Local name for the pulled plugin.
173 The ``:latest`` tag is optional, and is the default if
174 omitted. Optional.
175
176 Returns:
177 (:py:class:`Plugin`): The installed plugin
178 Raises:
179 :py:class:`docker.errors.APIError`
180 If the server returns an error.
181 """
182 privileges = self.client.api.plugin_privileges(remote_name)
183 it = self.client.api.pull_plugin(remote_name, privileges, local_name)
184 for data in it:
185 pass
186 return self.get(local_name or remote_name)
187
188 def list(self):
189 """
190 List plugins installed on the server.
191
192 Returns:
193 (list of :py:class:`Plugin`): The plugins.
194
195 Raises:
196 :py:class:`docker.errors.APIError`
197 If the server returns an error.
198 """
199 resp = self.client.api.plugins()
200 return [self.prepare_model(r) for r in resp]
201
[end of docker/models/plugins.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/docker/models/plugins.py b/docker/models/plugins.py
--- a/docker/models/plugins.py
+++ b/docker/models/plugins.py
@@ -119,7 +119,7 @@
privileges = self.client.api.plugin_privileges(remote)
for d in self.client.api.upgrade_plugin(self.name, remote, privileges):
yield d
- self._reload()
+ self.reload()
class PluginCollection(Collection):
|
{"golden_diff": "diff --git a/docker/models/plugins.py b/docker/models/plugins.py\n--- a/docker/models/plugins.py\n+++ b/docker/models/plugins.py\n@@ -119,7 +119,7 @@\n privileges = self.client.api.plugin_privileges(remote)\n for d in self.client.api.upgrade_plugin(self.name, remote, privileges):\n yield d\n- self._reload()\n+ self.reload()\n \n \n class PluginCollection(Collection):\n", "issue": "Plugin model upgrade: object has no attribute '_reload'\nI've already got a fix for this that I'm about to submit. For reference, here are the steps to reproduce in ipython:\r\n```\r\n~$ ipython [venv]\r\nPython 3.7.6 (default, Dec 22 2019, 01:09:06)\r\nType 'copyright', 'credits' or 'license' for more information\r\nIPython 7.18.1 -- An enhanced Interactive Python. Type '?' for help.\r\n\r\nIn [1]: import docker\r\n\r\nIn [2]: c = docker.DockerClient(base_url='unix://var/run/docker.sock')\r\n\r\nIn [3]: p = c.plugins.get('loki')\r\n\r\nIn [4]: list(p.upgrade(\"grafana/loki-docker-driver:latest\"))\r\n---------------------------------------------------------------------------\r\nAttributeError Traceback (most recent call last)\r\n<ipython-input-4-e5fc26fa0514> in <module>\r\n----> 1 list(p.upgrade(\"grafana/loki-docker-driver:latest\"))\r\n\r\n~/venv/lib/python3.7/site-packages/docker/models/plugins.py in upgrade(self, remote)\r\n 120 for d in self.client.api.upgrade_plugin(self.name, remote, privileges):\r\n 121 yield d\r\n--> 122 self._reload()\r\n 123\r\n 124\r\n\r\nAttributeError: 'Plugin' object has no attribute '_reload'\r\n```\r\n\r\n```\r\ndocker==4.3.1\r\nPython 3.7.6\r\nClient: Docker Engine - Community\r\n Cloud integration 0.1.18\r\n Version: 19.03.13\r\n API version: 1.40\r\n Go version: go1.13.15\r\n Git commit: 4484c46d9d\r\n Built: Wed Sep 16 16:58:31 2020\r\n OS/Arch: darwin/amd64\r\n Experimental: true\r\n\r\nServer: Docker Engine - Community\r\n Engine:\r\n Version: 19.03.13\r\n API version: 1.40 (minimum version 1.12)\r\n Go version: go1.13.15\r\n Git commit: 4484c46d9d\r\n Built: Wed Sep 16 17:07:04 2020\r\n OS/Arch: linux/amd64\r\n Experimental: true\r\n containerd:\r\n Version: v1.3.7\r\n GitCommit: 8fba4e9a7d01810a393d5d25a3621dc101981175\r\n runc:\r\n Version: 1.0.0-rc10\r\n GitCommit: dc9208a3303feef5b3839f4323d9beb36df0a9dd\r\n docker-init:\r\n Version: 0.18.0\r\n GitCommit: fec3683\r\n```\n", "before_files": [{"content": "from .. import errors\nfrom .resource import Collection, Model\n\n\nclass Plugin(Model):\n \"\"\"\n A plugin on the server.\n \"\"\"\n def __repr__(self):\n return \"<%s: '%s'>\" % (self.__class__.__name__, self.name)\n\n @property\n def name(self):\n \"\"\"\n The plugin's name.\n \"\"\"\n return self.attrs.get('Name')\n\n @property\n def enabled(self):\n \"\"\"\n Whether the plugin is enabled.\n \"\"\"\n return self.attrs.get('Enabled')\n\n @property\n def settings(self):\n \"\"\"\n A dictionary representing the plugin's configuration.\n \"\"\"\n return self.attrs.get('Settings')\n\n def configure(self, options):\n \"\"\"\n Update the plugin's settings.\n\n Args:\n options (dict): A key-value mapping of options.\n\n Raises:\n :py:class:`docker.errors.APIError`\n If the server returns an error.\n \"\"\"\n self.client.api.configure_plugin(self.name, options)\n self.reload()\n\n def disable(self):\n \"\"\"\n Disable the plugin.\n\n Raises:\n :py:class:`docker.errors.APIError`\n If the server returns an error.\n \"\"\"\n\n self.client.api.disable_plugin(self.name)\n self.reload()\n\n def enable(self, timeout=0):\n \"\"\"\n Enable the plugin.\n\n Args:\n timeout (int): Timeout in seconds. Default: 0\n\n Raises:\n :py:class:`docker.errors.APIError`\n If the server returns an error.\n \"\"\"\n self.client.api.enable_plugin(self.name, timeout)\n self.reload()\n\n def push(self):\n \"\"\"\n Push the plugin to a remote registry.\n\n Returns:\n A dict iterator streaming the status of the upload.\n\n Raises:\n :py:class:`docker.errors.APIError`\n If the server returns an error.\n \"\"\"\n return self.client.api.push_plugin(self.name)\n\n def remove(self, force=False):\n \"\"\"\n Remove the plugin from the server.\n\n Args:\n force (bool): Remove even if the plugin is enabled.\n Default: False\n\n Raises:\n :py:class:`docker.errors.APIError`\n If the server returns an error.\n \"\"\"\n return self.client.api.remove_plugin(self.name, force=force)\n\n def upgrade(self, remote=None):\n \"\"\"\n Upgrade the plugin.\n\n Args:\n remote (string): Remote reference to upgrade to. The\n ``:latest`` tag is optional and is the default if omitted.\n Default: this plugin's name.\n\n Returns:\n A generator streaming the decoded API logs\n \"\"\"\n if self.enabled:\n raise errors.DockerError(\n 'Plugin must be disabled before upgrading.'\n )\n\n if remote is None:\n remote = self.name\n privileges = self.client.api.plugin_privileges(remote)\n for d in self.client.api.upgrade_plugin(self.name, remote, privileges):\n yield d\n self._reload()\n\n\nclass PluginCollection(Collection):\n model = Plugin\n\n def create(self, name, plugin_data_dir, gzip=False):\n \"\"\"\n Create a new plugin.\n\n Args:\n name (string): The name of the plugin. The ``:latest`` tag is\n optional, and is the default if omitted.\n plugin_data_dir (string): Path to the plugin data directory.\n Plugin data directory must contain the ``config.json``\n manifest file and the ``rootfs`` directory.\n gzip (bool): Compress the context using gzip. Default: False\n\n Returns:\n (:py:class:`Plugin`): The newly created plugin.\n \"\"\"\n self.client.api.create_plugin(name, plugin_data_dir, gzip)\n return self.get(name)\n\n def get(self, name):\n \"\"\"\n Gets a plugin.\n\n Args:\n name (str): The name of the plugin.\n\n Returns:\n (:py:class:`Plugin`): The plugin.\n\n Raises:\n :py:class:`docker.errors.NotFound` If the plugin does not\n exist.\n :py:class:`docker.errors.APIError`\n If the server returns an error.\n \"\"\"\n return self.prepare_model(self.client.api.inspect_plugin(name))\n\n def install(self, remote_name, local_name=None):\n \"\"\"\n Pull and install a plugin.\n\n Args:\n remote_name (string): Remote reference for the plugin to\n install. The ``:latest`` tag is optional, and is the\n default if omitted.\n local_name (string): Local name for the pulled plugin.\n The ``:latest`` tag is optional, and is the default if\n omitted. Optional.\n\n Returns:\n (:py:class:`Plugin`): The installed plugin\n Raises:\n :py:class:`docker.errors.APIError`\n If the server returns an error.\n \"\"\"\n privileges = self.client.api.plugin_privileges(remote_name)\n it = self.client.api.pull_plugin(remote_name, privileges, local_name)\n for data in it:\n pass\n return self.get(local_name or remote_name)\n\n def list(self):\n \"\"\"\n List plugins installed on the server.\n\n Returns:\n (list of :py:class:`Plugin`): The plugins.\n\n Raises:\n :py:class:`docker.errors.APIError`\n If the server returns an error.\n \"\"\"\n resp = self.client.api.plugins()\n return [self.prepare_model(r) for r in resp]\n", "path": "docker/models/plugins.py"}]}
| 2,881 | 91 |
gh_patches_debug_49586
|
rasdani/github-patches
|
git_diff
|
quantumlib__Cirq-3358
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
upgrade to numpy 1.19
Discover all the issues / test failures related to numpy 0.19 and upgrade our requirements.txt.
More context: https://github.com/quantumlib/Cirq/pull/3106
</issue>
<code>
[start of cirq/optimizers/decompositions.py]
1 # Copyright 2018 The Cirq Developers
2 #
3 # Licensed under the Apache License, Version 2.0 (the "License");
4 # you may not use this file except in compliance with the License.
5 # You may obtain a copy of the License at
6 #
7 # https://www.apache.org/licenses/LICENSE-2.0
8 #
9 # Unless required by applicable law or agreed to in writing, software
10 # distributed under the License is distributed on an "AS IS" BASIS,
11 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12 # See the License for the specific language governing permissions and
13 # limitations under the License.
14
15 """Utility methods related to optimizing quantum circuits."""
16
17 import math
18 from typing import List, Optional, Tuple, cast
19
20 import numpy as np
21 import sympy
22
23 from cirq import ops, linalg, protocols
24 from cirq.linalg.tolerance import near_zero_mod
25
26
27 def is_negligible_turn(turns: float, tolerance: float) -> bool:
28 if isinstance(turns, sympy.Basic):
29 if not turns.is_constant():
30 return False
31 turns = float(turns)
32 return abs(_signed_mod_1(turns)) <= tolerance
33
34
35 def _signed_mod_1(x: float) -> float:
36 return (x + 0.5) % 1 - 0.5
37
38
39 def single_qubit_matrix_to_pauli_rotations(
40 mat: np.ndarray, atol: float = 0
41 ) -> List[Tuple[ops.Pauli, float]]:
42 """Implements a single-qubit operation with few rotations.
43
44 Args:
45 mat: The 2x2 unitary matrix of the operation to implement.
46 atol: A limit on the amount of absolute error introduced by the
47 construction.
48
49 Returns:
50 A list of (Pauli, half_turns) tuples that, when applied in order,
51 perform the desired operation.
52 """
53
54 def is_clifford_rotation(half_turns):
55 return near_zero_mod(half_turns, 0.5, atol=atol)
56
57 def to_quarter_turns(half_turns):
58 return round(2 * half_turns) % 4
59
60 def is_quarter_turn(half_turns):
61 return (is_clifford_rotation(half_turns) and
62 to_quarter_turns(half_turns) % 2 == 1)
63
64 def is_half_turn(half_turns):
65 return (is_clifford_rotation(half_turns) and
66 to_quarter_turns(half_turns) == 2)
67
68 def is_no_turn(half_turns):
69 return (is_clifford_rotation(half_turns) and
70 to_quarter_turns(half_turns) == 0)
71
72 # Decompose matrix
73 z_rad_before, y_rad, z_rad_after = (
74 linalg.deconstruct_single_qubit_matrix_into_angles(mat))
75 z_ht_before = z_rad_before / np.pi - 0.5
76 m_ht = y_rad / np.pi
77 m_pauli = ops.X # type: ops.Pauli
78 z_ht_after = z_rad_after / np.pi + 0.5
79
80 # Clean up angles
81 if is_clifford_rotation(z_ht_before):
82 if ((is_quarter_turn(z_ht_before) or is_quarter_turn(z_ht_after)) ^
83 (is_half_turn(m_ht) and is_no_turn(z_ht_before-z_ht_after))):
84 z_ht_before += 0.5
85 z_ht_after -= 0.5
86 m_pauli = ops.Y
87 if is_half_turn(z_ht_before) or is_half_turn(z_ht_after):
88 z_ht_before -= 1
89 z_ht_after += 1
90 m_ht = -m_ht
91 if is_no_turn(m_ht):
92 z_ht_before += z_ht_after
93 z_ht_after = 0
94 elif is_half_turn(m_ht):
95 z_ht_after -= z_ht_before
96 z_ht_before = 0
97
98 # Generate operations
99 rotation_list = [(ops.Z, z_ht_before), (m_pauli, m_ht), (ops.Z, z_ht_after)]
100 return [(pauli, ht) for pauli, ht in rotation_list if not is_no_turn(ht)]
101
102
103 def single_qubit_matrix_to_gates(
104 mat: np.ndarray, tolerance: float = 0
105 ) -> List[ops.SingleQubitGate]:
106 """Implements a single-qubit operation with few gates.
107
108 Args:
109 mat: The 2x2 unitary matrix of the operation to implement.
110 tolerance: A limit on the amount of error introduced by the
111 construction.
112
113 Returns:
114 A list of gates that, when applied in order, perform the desired
115 operation.
116 """
117 rotations = single_qubit_matrix_to_pauli_rotations(mat, tolerance)
118 return [cast(ops.SingleQubitGate, pauli)**ht for pauli, ht in rotations]
119
120
121 def single_qubit_op_to_framed_phase_form(
122 mat: np.ndarray) -> Tuple[np.ndarray, complex, complex]:
123 """Decomposes a 2x2 unitary M into U^-1 * diag(1, r) * U * diag(g, g).
124
125 U translates the rotation axis of M to the Z axis.
126 g fixes a global phase factor difference caused by the translation.
127 r's phase is the amount of rotation around M's rotation axis.
128
129 This decomposition can be used to decompose controlled single-qubit
130 rotations into controlled-Z operations bordered by single-qubit operations.
131
132 Args:
133 mat: The qubit operation as a 2x2 unitary matrix.
134
135 Returns:
136 A 2x2 unitary U, the complex relative phase factor r, and the complex
137 global phase factor g. Applying M is equivalent (up to global phase) to
138 applying U, rotating around the Z axis to apply r, then un-applying U.
139 When M is controlled, the control must be rotated around the Z axis to
140 apply g.
141 """
142 vals, vecs = np.linalg.eig(mat)
143 u = np.conj(vecs).T
144 r = vals[1] / vals[0]
145 g = vals[0]
146 return u, r, g
147
148
149 def _deconstruct_single_qubit_matrix_into_gate_turns(
150 mat: np.ndarray) -> Tuple[float, float, float]:
151 """Breaks down a 2x2 unitary into gate parameters.
152
153 Args:
154 mat: The 2x2 unitary matrix to break down.
155
156 Returns:
157 A tuple containing the amount to rotate around an XY axis, the phase of
158 that axis, and the amount to phase around Z. All results will be in
159 fractions of a whole turn, with values canonicalized into the range
160 [-0.5, 0.5).
161 """
162 pre_phase, rotation, post_phase = (
163 linalg.deconstruct_single_qubit_matrix_into_angles(mat))
164
165 # Figure out parameters of the actual gates we will do.
166 tau = 2 * np.pi
167 xy_turn = rotation / tau
168 xy_phase_turn = 0.25 - pre_phase / tau
169 total_z_turn = (post_phase + pre_phase) / tau
170
171 # Normalize turns into the range [-0.5, 0.5).
172 return (_signed_mod_1(xy_turn), _signed_mod_1(xy_phase_turn),
173 _signed_mod_1(total_z_turn))
174
175
176 def single_qubit_matrix_to_phased_x_z(
177 mat: np.ndarray,
178 atol: float = 0
179 ) -> List[ops.SingleQubitGate]:
180 """Implements a single-qubit operation with a PhasedX and Z gate.
181
182 If one of the gates isn't needed, it will be omitted.
183
184 Args:
185 mat: The 2x2 unitary matrix of the operation to implement.
186 atol: A limit on the amount of error introduced by the
187 construction.
188
189 Returns:
190 A list of gates that, when applied in order, perform the desired
191 operation.
192 """
193
194 xy_turn, xy_phase_turn, total_z_turn = (
195 _deconstruct_single_qubit_matrix_into_gate_turns(mat))
196
197 # Build the intended operation out of non-negligible XY and Z rotations.
198 result = [
199 ops.PhasedXPowGate(exponent=2 * xy_turn,
200 phase_exponent=2 * xy_phase_turn),
201 ops.Z**(2 * total_z_turn)
202 ]
203 result = [
204 g for g in result
205 if protocols.trace_distance_bound(g) > atol
206 ]
207
208 # Special case: XY half-turns can absorb Z rotations.
209 if len(result) == 2 and math.isclose(abs(xy_turn), 0.5, abs_tol=atol):
210 return [
211 ops.PhasedXPowGate(phase_exponent=2 * xy_phase_turn + total_z_turn)
212 ]
213
214 return result
215
216
217 def single_qubit_matrix_to_phxz(
218 mat: np.ndarray,
219 atol: float = 0,
220 ) -> Optional[ops.PhasedXZGate]:
221 """Implements a single-qubit operation with a PhasedXZ gate.
222
223 Under the hood, this uses deconstruct_single_qubit_matrix_into_angles which
224 converts the given matrix to a series of three rotations around the Z, Y, Z
225 axes. This is then converted to a phased X rotation followed by a Z, in the
226 form of a single PhasedXZ gate.
227
228 Args:
229 mat: The 2x2 unitary matrix of the operation to implement.
230 atol: A limit on the amount of error introduced by the
231 construction.
232
233 Returns:
234 A PhasedXZ gate that implements the given matrix, or None if it is
235 close to identity (trace distance <= atol).
236 """
237
238 xy_turn, xy_phase_turn, total_z_turn = (
239 _deconstruct_single_qubit_matrix_into_gate_turns(mat))
240
241 # Build the intended operation out of non-negligible XY and Z rotations.
242 g = ops.PhasedXZGate(
243 axis_phase_exponent=2 * xy_phase_turn,
244 x_exponent=2 * xy_turn,
245 z_exponent=2 * total_z_turn,
246 )
247
248 if protocols.trace_distance_bound(g) <= atol:
249 return None
250
251 # Special case: XY half-turns can absorb Z rotations.
252 if math.isclose(abs(xy_turn), 0.5, abs_tol=atol):
253 g = ops.PhasedXZGate(
254 axis_phase_exponent=2 * xy_phase_turn + total_z_turn,
255 x_exponent=1,
256 z_exponent=0,
257 )
258
259 return g
260
[end of cirq/optimizers/decompositions.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/cirq/optimizers/decompositions.py b/cirq/optimizers/decompositions.py
--- a/cirq/optimizers/decompositions.py
+++ b/cirq/optimizers/decompositions.py
@@ -139,7 +139,7 @@
When M is controlled, the control must be rotated around the Z axis to
apply g.
"""
- vals, vecs = np.linalg.eig(mat)
+ vals, vecs = linalg.unitary_eig(mat)
u = np.conj(vecs).T
r = vals[1] / vals[0]
g = vals[0]
|
{"golden_diff": "diff --git a/cirq/optimizers/decompositions.py b/cirq/optimizers/decompositions.py\n--- a/cirq/optimizers/decompositions.py\n+++ b/cirq/optimizers/decompositions.py\n@@ -139,7 +139,7 @@\n When M is controlled, the control must be rotated around the Z axis to\n apply g.\n \"\"\"\n- vals, vecs = np.linalg.eig(mat)\n+ vals, vecs = linalg.unitary_eig(mat)\n u = np.conj(vecs).T\n r = vals[1] / vals[0]\n g = vals[0]\n", "issue": "upgrade to numpy 1.19\nDiscover all the issues / test failures related to numpy 0.19 and upgrade our requirements.txt.\r\nMore context: https://github.com/quantumlib/Cirq/pull/3106\n", "before_files": [{"content": "# Copyright 2018 The Cirq Developers\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# https://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\n\"\"\"Utility methods related to optimizing quantum circuits.\"\"\"\n\nimport math\nfrom typing import List, Optional, Tuple, cast\n\nimport numpy as np\nimport sympy\n\nfrom cirq import ops, linalg, protocols\nfrom cirq.linalg.tolerance import near_zero_mod\n\n\ndef is_negligible_turn(turns: float, tolerance: float) -> bool:\n if isinstance(turns, sympy.Basic):\n if not turns.is_constant():\n return False\n turns = float(turns)\n return abs(_signed_mod_1(turns)) <= tolerance\n\n\ndef _signed_mod_1(x: float) -> float:\n return (x + 0.5) % 1 - 0.5\n\n\ndef single_qubit_matrix_to_pauli_rotations(\n mat: np.ndarray, atol: float = 0\n) -> List[Tuple[ops.Pauli, float]]:\n \"\"\"Implements a single-qubit operation with few rotations.\n\n Args:\n mat: The 2x2 unitary matrix of the operation to implement.\n atol: A limit on the amount of absolute error introduced by the\n construction.\n\n Returns:\n A list of (Pauli, half_turns) tuples that, when applied in order,\n perform the desired operation.\n \"\"\"\n\n def is_clifford_rotation(half_turns):\n return near_zero_mod(half_turns, 0.5, atol=atol)\n\n def to_quarter_turns(half_turns):\n return round(2 * half_turns) % 4\n\n def is_quarter_turn(half_turns):\n return (is_clifford_rotation(half_turns) and\n to_quarter_turns(half_turns) % 2 == 1)\n\n def is_half_turn(half_turns):\n return (is_clifford_rotation(half_turns) and\n to_quarter_turns(half_turns) == 2)\n\n def is_no_turn(half_turns):\n return (is_clifford_rotation(half_turns) and\n to_quarter_turns(half_turns) == 0)\n\n # Decompose matrix\n z_rad_before, y_rad, z_rad_after = (\n linalg.deconstruct_single_qubit_matrix_into_angles(mat))\n z_ht_before = z_rad_before / np.pi - 0.5\n m_ht = y_rad / np.pi\n m_pauli = ops.X # type: ops.Pauli\n z_ht_after = z_rad_after / np.pi + 0.5\n\n # Clean up angles\n if is_clifford_rotation(z_ht_before):\n if ((is_quarter_turn(z_ht_before) or is_quarter_turn(z_ht_after)) ^\n (is_half_turn(m_ht) and is_no_turn(z_ht_before-z_ht_after))):\n z_ht_before += 0.5\n z_ht_after -= 0.5\n m_pauli = ops.Y\n if is_half_turn(z_ht_before) or is_half_turn(z_ht_after):\n z_ht_before -= 1\n z_ht_after += 1\n m_ht = -m_ht\n if is_no_turn(m_ht):\n z_ht_before += z_ht_after\n z_ht_after = 0\n elif is_half_turn(m_ht):\n z_ht_after -= z_ht_before\n z_ht_before = 0\n\n # Generate operations\n rotation_list = [(ops.Z, z_ht_before), (m_pauli, m_ht), (ops.Z, z_ht_after)]\n return [(pauli, ht) for pauli, ht in rotation_list if not is_no_turn(ht)]\n\n\ndef single_qubit_matrix_to_gates(\n mat: np.ndarray, tolerance: float = 0\n) -> List[ops.SingleQubitGate]:\n \"\"\"Implements a single-qubit operation with few gates.\n\n Args:\n mat: The 2x2 unitary matrix of the operation to implement.\n tolerance: A limit on the amount of error introduced by the\n construction.\n\n Returns:\n A list of gates that, when applied in order, perform the desired\n operation.\n \"\"\"\n rotations = single_qubit_matrix_to_pauli_rotations(mat, tolerance)\n return [cast(ops.SingleQubitGate, pauli)**ht for pauli, ht in rotations]\n\n\ndef single_qubit_op_to_framed_phase_form(\n mat: np.ndarray) -> Tuple[np.ndarray, complex, complex]:\n \"\"\"Decomposes a 2x2 unitary M into U^-1 * diag(1, r) * U * diag(g, g).\n\n U translates the rotation axis of M to the Z axis.\n g fixes a global phase factor difference caused by the translation.\n r's phase is the amount of rotation around M's rotation axis.\n\n This decomposition can be used to decompose controlled single-qubit\n rotations into controlled-Z operations bordered by single-qubit operations.\n\n Args:\n mat: The qubit operation as a 2x2 unitary matrix.\n\n Returns:\n A 2x2 unitary U, the complex relative phase factor r, and the complex\n global phase factor g. Applying M is equivalent (up to global phase) to\n applying U, rotating around the Z axis to apply r, then un-applying U.\n When M is controlled, the control must be rotated around the Z axis to\n apply g.\n \"\"\"\n vals, vecs = np.linalg.eig(mat)\n u = np.conj(vecs).T\n r = vals[1] / vals[0]\n g = vals[0]\n return u, r, g\n\n\ndef _deconstruct_single_qubit_matrix_into_gate_turns(\n mat: np.ndarray) -> Tuple[float, float, float]:\n \"\"\"Breaks down a 2x2 unitary into gate parameters.\n\n Args:\n mat: The 2x2 unitary matrix to break down.\n\n Returns:\n A tuple containing the amount to rotate around an XY axis, the phase of\n that axis, and the amount to phase around Z. All results will be in\n fractions of a whole turn, with values canonicalized into the range\n [-0.5, 0.5).\n \"\"\"\n pre_phase, rotation, post_phase = (\n linalg.deconstruct_single_qubit_matrix_into_angles(mat))\n\n # Figure out parameters of the actual gates we will do.\n tau = 2 * np.pi\n xy_turn = rotation / tau\n xy_phase_turn = 0.25 - pre_phase / tau\n total_z_turn = (post_phase + pre_phase) / tau\n\n # Normalize turns into the range [-0.5, 0.5).\n return (_signed_mod_1(xy_turn), _signed_mod_1(xy_phase_turn),\n _signed_mod_1(total_z_turn))\n\n\ndef single_qubit_matrix_to_phased_x_z(\n mat: np.ndarray,\n atol: float = 0\n) -> List[ops.SingleQubitGate]:\n \"\"\"Implements a single-qubit operation with a PhasedX and Z gate.\n\n If one of the gates isn't needed, it will be omitted.\n\n Args:\n mat: The 2x2 unitary matrix of the operation to implement.\n atol: A limit on the amount of error introduced by the\n construction.\n\n Returns:\n A list of gates that, when applied in order, perform the desired\n operation.\n \"\"\"\n\n xy_turn, xy_phase_turn, total_z_turn = (\n _deconstruct_single_qubit_matrix_into_gate_turns(mat))\n\n # Build the intended operation out of non-negligible XY and Z rotations.\n result = [\n ops.PhasedXPowGate(exponent=2 * xy_turn,\n phase_exponent=2 * xy_phase_turn),\n ops.Z**(2 * total_z_turn)\n ]\n result = [\n g for g in result\n if protocols.trace_distance_bound(g) > atol\n ]\n\n # Special case: XY half-turns can absorb Z rotations.\n if len(result) == 2 and math.isclose(abs(xy_turn), 0.5, abs_tol=atol):\n return [\n ops.PhasedXPowGate(phase_exponent=2 * xy_phase_turn + total_z_turn)\n ]\n\n return result\n\n\ndef single_qubit_matrix_to_phxz(\n mat: np.ndarray,\n atol: float = 0,\n) -> Optional[ops.PhasedXZGate]:\n \"\"\"Implements a single-qubit operation with a PhasedXZ gate.\n\n Under the hood, this uses deconstruct_single_qubit_matrix_into_angles which\n converts the given matrix to a series of three rotations around the Z, Y, Z\n axes. This is then converted to a phased X rotation followed by a Z, in the\n form of a single PhasedXZ gate.\n\n Args:\n mat: The 2x2 unitary matrix of the operation to implement.\n atol: A limit on the amount of error introduced by the\n construction.\n\n Returns:\n A PhasedXZ gate that implements the given matrix, or None if it is\n close to identity (trace distance <= atol).\n \"\"\"\n\n xy_turn, xy_phase_turn, total_z_turn = (\n _deconstruct_single_qubit_matrix_into_gate_turns(mat))\n\n # Build the intended operation out of non-negligible XY and Z rotations.\n g = ops.PhasedXZGate(\n axis_phase_exponent=2 * xy_phase_turn,\n x_exponent=2 * xy_turn,\n z_exponent=2 * total_z_turn,\n )\n\n if protocols.trace_distance_bound(g) <= atol:\n return None\n\n # Special case: XY half-turns can absorb Z rotations.\n if math.isclose(abs(xy_turn), 0.5, abs_tol=atol):\n g = ops.PhasedXZGate(\n axis_phase_exponent=2 * xy_phase_turn + total_z_turn,\n x_exponent=1,\n z_exponent=0,\n )\n\n return g\n", "path": "cirq/optimizers/decompositions.py"}]}
| 3,599 | 143 |
gh_patches_debug_27885
|
rasdani/github-patches
|
git_diff
|
pwr-Solaar__Solaar-743
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
authorship of Solaar in setup.py
Daniel Pavel is listed as the sole author of Solaar in setup.py
As far as I can tell, this puts him and his email in several repositories, such as PyPI https://pypi.org/project/solaar/
Who should be put there?
</issue>
<code>
[start of setup.py]
1 #!/usr/bin/env python3
2
3 from glob import glob as _glob
4 try:
5 from setuptools import setup
6 except ImportError:
7 from distutils.core import setup
8
9 autostart_path = '/etc/xdg/autostart'
10
11 import sys
12 backup_path_0 = sys.path[0]
13 sys.path[0] = backup_path_0 + '/lib'
14 #from solaar import NAME, __version__
15 __version__ = '1.0.2-rc1'
16 NAME = 'Solaar'
17
18 sys.path[0] = backup_path_0
19
20 if 'install' in sys.argv:
21 # naively guess where the autostart .desktop file should be installed
22 if '--prefix' in sys.argv or any(x.startswith('--prefix=') for x in sys.argv) or '--home' in sys.argv:
23 autostart_path = 'etc/xdg/autostart'
24 elif '--user' in sys.argv:
25 from os import environ
26 from os import path
27 xdg_config_home = environ.get('XDG_CONFIG_HOME', path.expanduser(path.join('~', '.config')))
28 autostart_path = path.join(xdg_config_home, 'autostart')
29 del environ, path, xdg_config_home
30
31 del sys, backup_path_0
32
33
34 def _data_files():
35 from os.path import dirname as _dirname
36
37 yield 'share/solaar/icons', _glob('share/solaar/icons/solaar*.svg')
38 yield 'share/solaar/icons', _glob('share/solaar/icons/light_*.png')
39 yield 'share/icons/hicolor/scalable/apps', ['share/solaar/icons/solaar.svg']
40
41 for mo in _glob('share/locale/*/LC_MESSAGES/solaar.mo'):
42 yield _dirname(mo), [mo]
43
44 yield 'share/applications', ['share/applications/solaar.desktop']
45 yield autostart_path, ['share/autostart/solaar.desktop']
46
47 del _dirname
48
49
50 setup(name=NAME.lower(),
51 version=__version__,
52 description='Linux devices manager for the Logitech Unifying Receiver.',
53 long_description='''
54 Solaar is a Linux device manager for Logitech's Unifying Receiver peripherals.
55 It is able to pair/unpair devices to the receiver, and for some devices read
56 battery status.
57 '''.strip(),
58 author='Daniel Pavel',
59 author_email='[email protected]',
60 license='GPLv2',
61 url='http://pwr-solaar.github.io/Solaar/',
62 classifiers=[
63 'Development Status :: 4 - Beta',
64 'Environment :: X11 Applications :: GTK',
65 'Environment :: Console',
66 'Intended Audience :: End Users/Desktop',
67 'License :: DFSG approved',
68 'License :: OSI Approved :: GNU General Public License v2 (GPLv2)',
69 'Natural Language :: English',
70 'Programming Language :: Python :: 3 :: Only',
71 'Operating System :: POSIX :: Linux',
72 'Topic :: Utilities',
73 ],
74
75 platforms=['linux'],
76
77 # sudo apt install python-gi python3-gi \
78 # gir1.2-gtk-3.0 gir1.2-notify-0.7 gir1.2-ayatanaappindicator3-0.1
79 # os_requires=['gi.repository.GObject (>= 2.0)', 'gi.repository.Gtk (>= 3.0)'],
80
81 python_requires='>=3.2',
82 install_requires=['pyudev (>= 0.13)', ],
83 package_dir={'': 'lib'},
84 packages=['hidapi', 'logitech_receiver', 'solaar', 'solaar.ui', 'solaar.cli'],
85 data_files=list(_data_files()),
86 scripts=_glob('bin/*'),
87 )
88
[end of setup.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/setup.py b/setup.py
--- a/setup.py
+++ b/setup.py
@@ -8,28 +8,10 @@
autostart_path = '/etc/xdg/autostart'
-import sys
-backup_path_0 = sys.path[0]
-sys.path[0] = backup_path_0 + '/lib'
#from solaar import NAME, __version__
__version__ = '1.0.2-rc1'
NAME = 'Solaar'
-sys.path[0] = backup_path_0
-
-if 'install' in sys.argv:
- # naively guess where the autostart .desktop file should be installed
- if '--prefix' in sys.argv or any(x.startswith('--prefix=') for x in sys.argv) or '--home' in sys.argv:
- autostart_path = 'etc/xdg/autostart'
- elif '--user' in sys.argv:
- from os import environ
- from os import path
- xdg_config_home = environ.get('XDG_CONFIG_HOME', path.expanduser(path.join('~', '.config')))
- autostart_path = path.join(xdg_config_home, 'autostart')
- del environ, path, xdg_config_home
-
-del sys, backup_path_0
-
def _data_files():
from os.path import dirname as _dirname
@@ -43,6 +25,7 @@
yield 'share/applications', ['share/applications/solaar.desktop']
yield autostart_path, ['share/autostart/solaar.desktop']
+ yield '/etc/udev/rules.d', ['rules.d/42-logitech-unify-permissions.rules']
del _dirname
@@ -56,7 +39,6 @@
battery status.
'''.strip(),
author='Daniel Pavel',
- author_email='[email protected]',
license='GPLv2',
url='http://pwr-solaar.github.io/Solaar/',
classifiers=[
|
{"golden_diff": "diff --git a/setup.py b/setup.py\n--- a/setup.py\n+++ b/setup.py\n@@ -8,28 +8,10 @@\n \n autostart_path = '/etc/xdg/autostart'\n \n-import sys\n-backup_path_0 = sys.path[0]\n-sys.path[0] = backup_path_0 + '/lib'\n #from solaar import NAME, __version__\n __version__ = '1.0.2-rc1'\n NAME = 'Solaar'\n \n-sys.path[0] = backup_path_0\n-\n-if 'install' in sys.argv:\n-\t# naively guess where the autostart .desktop file should be installed\n-\tif '--prefix' in sys.argv or any(x.startswith('--prefix=') for x in sys.argv) or '--home' in sys.argv:\n-\t\tautostart_path = 'etc/xdg/autostart'\n-\telif '--user' in sys.argv:\n-\t\tfrom os import environ\n-\t\tfrom os import path\n-\t\txdg_config_home = environ.get('XDG_CONFIG_HOME', path.expanduser(path.join('~', '.config')))\n-\t\tautostart_path = path.join(xdg_config_home, 'autostart')\n-\t\tdel environ, path, xdg_config_home\n-\n-del sys, backup_path_0\n-\n \n def _data_files():\n \tfrom os.path import dirname as _dirname\n@@ -43,6 +25,7 @@\n \n \tyield 'share/applications', ['share/applications/solaar.desktop']\n \tyield autostart_path, ['share/autostart/solaar.desktop']\n+\tyield '/etc/udev/rules.d', ['rules.d/42-logitech-unify-permissions.rules']\n \n \tdel _dirname\n \n@@ -56,7 +39,6 @@\n battery status.\n '''.strip(),\n \t\tauthor='Daniel Pavel',\n-\t\tauthor_email='[email protected]',\n \t\tlicense='GPLv2',\n \t\turl='http://pwr-solaar.github.io/Solaar/',\n \t\tclassifiers=[\n", "issue": "authorship of Solaar in setup.py\nDaniel Pavel is listed as the sole author of Solaar in setup.py \r\n\r\nAs far as I can tell, this puts him and his email in several repositories, such as PyPI https://pypi.org/project/solaar/\r\n\r\nWho should be put there?\n", "before_files": [{"content": "#!/usr/bin/env python3\n\nfrom glob import glob as _glob\ntry:\n from setuptools import setup\nexcept ImportError:\n from distutils.core import setup\n\nautostart_path = '/etc/xdg/autostart'\n\nimport sys\nbackup_path_0 = sys.path[0]\nsys.path[0] = backup_path_0 + '/lib'\n#from solaar import NAME, __version__\n__version__ = '1.0.2-rc1'\nNAME = 'Solaar'\n\nsys.path[0] = backup_path_0\n\nif 'install' in sys.argv:\n\t# naively guess where the autostart .desktop file should be installed\n\tif '--prefix' in sys.argv or any(x.startswith('--prefix=') for x in sys.argv) or '--home' in sys.argv:\n\t\tautostart_path = 'etc/xdg/autostart'\n\telif '--user' in sys.argv:\n\t\tfrom os import environ\n\t\tfrom os import path\n\t\txdg_config_home = environ.get('XDG_CONFIG_HOME', path.expanduser(path.join('~', '.config')))\n\t\tautostart_path = path.join(xdg_config_home, 'autostart')\n\t\tdel environ, path, xdg_config_home\n\ndel sys, backup_path_0\n\n\ndef _data_files():\n\tfrom os.path import dirname as _dirname\n\n\tyield 'share/solaar/icons', _glob('share/solaar/icons/solaar*.svg')\n\tyield 'share/solaar/icons', _glob('share/solaar/icons/light_*.png')\n\tyield 'share/icons/hicolor/scalable/apps', ['share/solaar/icons/solaar.svg']\n\n\tfor mo in _glob('share/locale/*/LC_MESSAGES/solaar.mo'):\n\t\tyield _dirname(mo), [mo]\n\n\tyield 'share/applications', ['share/applications/solaar.desktop']\n\tyield autostart_path, ['share/autostart/solaar.desktop']\n\n\tdel _dirname\n\n\nsetup(name=NAME.lower(),\n\t\tversion=__version__,\n\t\tdescription='Linux devices manager for the Logitech Unifying Receiver.',\n\t\tlong_description='''\nSolaar is a Linux device manager for Logitech's Unifying Receiver peripherals.\nIt is able to pair/unpair devices to the receiver, and for some devices read\nbattery status.\n'''.strip(),\n\t\tauthor='Daniel Pavel',\n\t\tauthor_email='[email protected]',\n\t\tlicense='GPLv2',\n\t\turl='http://pwr-solaar.github.io/Solaar/',\n\t\tclassifiers=[\n\t\t\t'Development Status :: 4 - Beta',\n\t\t\t'Environment :: X11 Applications :: GTK',\n\t\t\t'Environment :: Console',\n\t\t\t'Intended Audience :: End Users/Desktop',\n\t\t\t'License :: DFSG approved',\n\t\t\t'License :: OSI Approved :: GNU General Public License v2 (GPLv2)',\n\t\t\t'Natural Language :: English',\n\t\t\t'Programming Language :: Python :: 3 :: Only',\n\t\t\t'Operating System :: POSIX :: Linux',\n\t\t\t'Topic :: Utilities',\n\t\t\t],\n\n\t\tplatforms=['linux'],\n\n\t\t# sudo apt install python-gi python3-gi \\\n\t\t# gir1.2-gtk-3.0 gir1.2-notify-0.7 gir1.2-ayatanaappindicator3-0.1\n\t\t# os_requires=['gi.repository.GObject (>= 2.0)', 'gi.repository.Gtk (>= 3.0)'],\n\n\t\tpython_requires='>=3.2',\n\t\tinstall_requires=['pyudev (>= 0.13)', ],\n\t\tpackage_dir={'': 'lib'},\n\t\tpackages=['hidapi', 'logitech_receiver', 'solaar', 'solaar.ui', 'solaar.cli'],\n\t\tdata_files=list(_data_files()),\n\t\tscripts=_glob('bin/*'),\n\t)\n", "path": "setup.py"}]}
| 1,605 | 447 |
gh_patches_debug_30750
|
rasdani/github-patches
|
git_diff
|
pyg-team__pytorch_geometric-8164
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
math term missed in dimenet_utils.associated_legendre_polynomials
### 🛠 Proposed Refactor
In order to get the function of 'spherical harmonics' basis used in DimeNet, we need to calc 'Associated Legendre Polynomial' in math. But I just found a term missed in torch_geometric.nn.models.dimenet_utils.associated_legendre_polynomials().
The author of DimeNet use `zero_m_only=True` by default, so there will be no error when running the code by default.
But just a warning for those who also read or use this code, the author of the code (version of pyg: 2.3.1) missed a term when calculating $P_l^l$ which make all results except $P_l^0$ wrong. [original code here](https://github.com/pyg-team/pytorch_geometric/blob/f71ead8ade8a67be23982114cfff649b7d074cfb/torch_geometric/nn/models/dimenet_utils.py#L86)
Take the result of $P_1^1$ as an example:
original code will output -1, but according to eq.14 at https://mathworld.wolfram.com/AssociatedLegendrePolynomial.html, the ground truth result is $-(1-x^2)^{0.5}$.
### Suggest a potential alternative/fix
Beyond adding the missed term, I also add some comment to help understanding.
```python
def associated_legendre_polynomials(k, zero_m_only=True):
'''
helper function to calc Y_l^m
'''
z = sym.symbols('z')
P_l_m = [[0] * (j + 1) for j in range(k)]
P_l_m[0][0] = 1
if k > 0:
P_l_m[1][0] = z
for j in range(2, k):
P_l_m[j][0] = sym.simplify(((2 * j - 1) * z * P_l_m[j - 1][0] -
(j - 1) * P_l_m[j - 2][0]) / j) # use the property of eq.7: https://mathworld.wolfram.com/AssociatedLegendrePolynomial.html
if not zero_m_only:
for i in range(1, k):
P_l_m[i][i] = sym.simplify((1 - 2 * i) * P_l_m[i - 1][i - 1]*(1-z**2)**0.5) # add missed term (*(1-z**2)**0.5) here
if i + 1 < k:
P_l_m[i + 1][i] = sym.simplify(
(2 * i + 1) * z * P_l_m[i][i]) # use the property of eq.11: https://mathworld.wolfram.com/AssociatedLegendrePolynomial.html
for j in range(i + 2, k):
P_l_m[j][i] = sym.simplify(
((2 * j - 1) * z * P_l_m[j - 1][i] -
(i + j - 1) * P_l_m[j - 2][i]) / (j - i)) # use the property of eq.7: https://mathworld.wolfram.com/AssociatedLegendrePolynomial.html
return P_l_m
```
After fixing this tiny bug, all the results are the right
</issue>
<code>
[start of torch_geometric/nn/models/dimenet_utils.py]
1 # Shameless steal from: https://github.com/klicperajo/dimenet
2
3 import numpy as np
4 from scipy import special as sp
5 from scipy.optimize import brentq
6
7 try:
8 import sympy as sym
9 except ImportError:
10 sym = None
11
12
13 def Jn(r, n):
14 return np.sqrt(np.pi / (2 * r)) * sp.jv(n + 0.5, r)
15
16
17 def Jn_zeros(n, k):
18 zerosj = np.zeros((n, k), dtype='float32')
19 zerosj[0] = np.arange(1, k + 1) * np.pi
20 points = np.arange(1, k + n) * np.pi
21 racines = np.zeros(k + n - 1, dtype='float32')
22 for i in range(1, n):
23 for j in range(k + n - 1 - i):
24 foo = brentq(Jn, points[j], points[j + 1], (i, ))
25 racines[j] = foo
26 points = racines
27 zerosj[i][:k] = racines[:k]
28
29 return zerosj
30
31
32 def spherical_bessel_formulas(n):
33 x = sym.symbols('x')
34
35 f = [sym.sin(x) / x]
36 a = sym.sin(x) / x
37 for i in range(1, n):
38 b = sym.diff(a, x) / x
39 f += [sym.simplify(b * (-x)**i)]
40 a = sym.simplify(b)
41 return f
42
43
44 def bessel_basis(n, k):
45 zeros = Jn_zeros(n, k)
46 normalizer = []
47 for order in range(n):
48 normalizer_tmp = []
49 for i in range(k):
50 normalizer_tmp += [0.5 * Jn(zeros[order, i], order + 1)**2]
51 normalizer_tmp = 1 / np.array(normalizer_tmp)**0.5
52 normalizer += [normalizer_tmp]
53
54 f = spherical_bessel_formulas(n)
55 x = sym.symbols('x')
56 bess_basis = []
57 for order in range(n):
58 bess_basis_tmp = []
59 for i in range(k):
60 bess_basis_tmp += [
61 sym.simplify(normalizer[order][i] *
62 f[order].subs(x, zeros[order, i] * x))
63 ]
64 bess_basis += [bess_basis_tmp]
65 return bess_basis
66
67
68 def sph_harm_prefactor(k, m):
69 return ((2 * k + 1) * np.math.factorial(k - abs(m)) /
70 (4 * np.pi * np.math.factorial(k + abs(m))))**0.5
71
72
73 def associated_legendre_polynomials(k, zero_m_only=True):
74 z = sym.symbols('z')
75 P_l_m = [[0] * (j + 1) for j in range(k)]
76
77 P_l_m[0][0] = 1
78 if k > 0:
79 P_l_m[1][0] = z
80
81 for j in range(2, k):
82 P_l_m[j][0] = sym.simplify(((2 * j - 1) * z * P_l_m[j - 1][0] -
83 (j - 1) * P_l_m[j - 2][0]) / j)
84 if not zero_m_only:
85 for i in range(1, k):
86 P_l_m[i][i] = sym.simplify((1 - 2 * i) * P_l_m[i - 1][i - 1])
87 if i + 1 < k:
88 P_l_m[i + 1][i] = sym.simplify(
89 (2 * i + 1) * z * P_l_m[i][i])
90 for j in range(i + 2, k):
91 P_l_m[j][i] = sym.simplify(
92 ((2 * j - 1) * z * P_l_m[j - 1][i] -
93 (i + j - 1) * P_l_m[j - 2][i]) / (j - i))
94
95 return P_l_m
96
97
98 def real_sph_harm(k, zero_m_only=True, spherical_coordinates=True):
99 if not zero_m_only:
100 S_m = [0]
101 C_m = [1]
102 for i in range(1, k):
103 x = sym.symbols('x')
104 y = sym.symbols('y')
105 S_m += [x * S_m[i - 1] + y * C_m[i - 1]]
106 C_m += [x * C_m[i - 1] - y * S_m[i - 1]]
107
108 P_l_m = associated_legendre_polynomials(k, zero_m_only)
109 if spherical_coordinates:
110 theta = sym.symbols('theta')
111 z = sym.symbols('z')
112 for i in range(len(P_l_m)):
113 for j in range(len(P_l_m[i])):
114 if not isinstance(P_l_m[i][j], int):
115 P_l_m[i][j] = P_l_m[i][j].subs(z, sym.cos(theta))
116 if not zero_m_only:
117 phi = sym.symbols('phi')
118 for i in range(len(S_m)):
119 S_m[i] = S_m[i].subs(x,
120 sym.sin(theta) * sym.cos(phi)).subs(
121 y,
122 sym.sin(theta) * sym.sin(phi))
123 for i in range(len(C_m)):
124 C_m[i] = C_m[i].subs(x,
125 sym.sin(theta) * sym.cos(phi)).subs(
126 y,
127 sym.sin(theta) * sym.sin(phi))
128
129 Y_func_l_m = [['0'] * (2 * j + 1) for j in range(k)]
130 for i in range(k):
131 Y_func_l_m[i][0] = sym.simplify(sph_harm_prefactor(i, 0) * P_l_m[i][0])
132
133 if not zero_m_only:
134 for i in range(1, k):
135 for j in range(1, i + 1):
136 Y_func_l_m[i][j] = sym.simplify(
137 2**0.5 * sph_harm_prefactor(i, j) * C_m[j] * P_l_m[i][j])
138 for i in range(1, k):
139 for j in range(1, i + 1):
140 Y_func_l_m[i][-j] = sym.simplify(
141 2**0.5 * sph_harm_prefactor(i, -j) * S_m[j] * P_l_m[i][j])
142
143 return Y_func_l_m
144
[end of torch_geometric/nn/models/dimenet_utils.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/torch_geometric/nn/models/dimenet_utils.py b/torch_geometric/nn/models/dimenet_utils.py
--- a/torch_geometric/nn/models/dimenet_utils.py
+++ b/torch_geometric/nn/models/dimenet_utils.py
@@ -71,6 +71,7 @@
def associated_legendre_polynomials(k, zero_m_only=True):
+ r"""Helper function to calculate Y_l^m."""
z = sym.symbols('z')
P_l_m = [[0] * (j + 1) for j in range(k)]
@@ -79,15 +80,22 @@
P_l_m[1][0] = z
for j in range(2, k):
+ # Use the property of Eq (7) in
+ # https://mathworld.wolfram.com/AssociatedLegendrePolynomial.html:
P_l_m[j][0] = sym.simplify(((2 * j - 1) * z * P_l_m[j - 1][0] -
(j - 1) * P_l_m[j - 2][0]) / j)
if not zero_m_only:
for i in range(1, k):
- P_l_m[i][i] = sym.simplify((1 - 2 * i) * P_l_m[i - 1][i - 1])
+ P_l_m[i][i] = sym.simplify(
+ (1 - 2 * i) * P_l_m[i - 1][i - 1] * (1 - z**2)**0.5)
if i + 1 < k:
+ # Use the property of Eq (11) in
+ # https://mathworld.wolfram.com/AssociatedLegendrePolynomial.html:
P_l_m[i + 1][i] = sym.simplify(
(2 * i + 1) * z * P_l_m[i][i])
for j in range(i + 2, k):
+ # Use the property of Eq (7) in
+ # https://mathworld.wolfram.com/AssociatedLegendrePolynomial.html:
P_l_m[j][i] = sym.simplify(
((2 * j - 1) * z * P_l_m[j - 1][i] -
(i + j - 1) * P_l_m[j - 2][i]) / (j - i))
|
{"golden_diff": "diff --git a/torch_geometric/nn/models/dimenet_utils.py b/torch_geometric/nn/models/dimenet_utils.py\n--- a/torch_geometric/nn/models/dimenet_utils.py\n+++ b/torch_geometric/nn/models/dimenet_utils.py\n@@ -71,6 +71,7 @@\n \n \n def associated_legendre_polynomials(k, zero_m_only=True):\n+ r\"\"\"Helper function to calculate Y_l^m.\"\"\"\n z = sym.symbols('z')\n P_l_m = [[0] * (j + 1) for j in range(k)]\n \n@@ -79,15 +80,22 @@\n P_l_m[1][0] = z\n \n for j in range(2, k):\n+ # Use the property of Eq (7) in\n+ # https://mathworld.wolfram.com/AssociatedLegendrePolynomial.html:\n P_l_m[j][0] = sym.simplify(((2 * j - 1) * z * P_l_m[j - 1][0] -\n (j - 1) * P_l_m[j - 2][0]) / j)\n if not zero_m_only:\n for i in range(1, k):\n- P_l_m[i][i] = sym.simplify((1 - 2 * i) * P_l_m[i - 1][i - 1])\n+ P_l_m[i][i] = sym.simplify(\n+ (1 - 2 * i) * P_l_m[i - 1][i - 1] * (1 - z**2)**0.5)\n if i + 1 < k:\n+ # Use the property of Eq (11) in\n+ # https://mathworld.wolfram.com/AssociatedLegendrePolynomial.html:\n P_l_m[i + 1][i] = sym.simplify(\n (2 * i + 1) * z * P_l_m[i][i])\n for j in range(i + 2, k):\n+ # Use the property of Eq (7) in\n+ # https://mathworld.wolfram.com/AssociatedLegendrePolynomial.html:\n P_l_m[j][i] = sym.simplify(\n ((2 * j - 1) * z * P_l_m[j - 1][i] -\n (i + j - 1) * P_l_m[j - 2][i]) / (j - i))\n", "issue": "math term missed in dimenet_utils.associated_legendre_polynomials\n### \ud83d\udee0 Proposed Refactor\n\nIn order to get the function of 'spherical harmonics' basis used in DimeNet, we need to calc 'Associated Legendre Polynomial' in math. But I just found a term missed in torch_geometric.nn.models.dimenet_utils.associated_legendre_polynomials(). \r\n\r\nThe author of DimeNet use `zero_m_only=True` by default, so there will be no error when running the code by default. \r\n\r\nBut just a warning for those who also read or use this code, the author of the code (version of pyg: 2.3.1) missed a term when calculating $P_l^l$ which make all results except $P_l^0$ wrong. [original code here](https://github.com/pyg-team/pytorch_geometric/blob/f71ead8ade8a67be23982114cfff649b7d074cfb/torch_geometric/nn/models/dimenet_utils.py#L86)\r\n\r\nTake the result of $P_1^1$ as an example:\r\n\r\noriginal code will output -1, but according to eq.14 at https://mathworld.wolfram.com/AssociatedLegendrePolynomial.html, the ground truth result is $-(1-x^2)^{0.5}$. \n\n### Suggest a potential alternative/fix\n\nBeyond adding the missed term, I also add some comment to help understanding.\r\n```python\r\ndef associated_legendre_polynomials(k, zero_m_only=True):\r\n '''\r\n helper function to calc Y_l^m\r\n '''\r\n z = sym.symbols('z')\r\n P_l_m = [[0] * (j + 1) for j in range(k)]\r\n\r\n P_l_m[0][0] = 1\r\n if k > 0:\r\n P_l_m[1][0] = z\r\n\r\n for j in range(2, k):\r\n P_l_m[j][0] = sym.simplify(((2 * j - 1) * z * P_l_m[j - 1][0] -\r\n (j - 1) * P_l_m[j - 2][0]) / j) # use the property of eq.7: https://mathworld.wolfram.com/AssociatedLegendrePolynomial.html\r\n if not zero_m_only:\r\n for i in range(1, k):\r\n P_l_m[i][i] = sym.simplify((1 - 2 * i) * P_l_m[i - 1][i - 1]*(1-z**2)**0.5) # add missed term (*(1-z**2)**0.5) here\r\n if i + 1 < k:\r\n P_l_m[i + 1][i] = sym.simplify(\r\n (2 * i + 1) * z * P_l_m[i][i]) # use the property of eq.11: https://mathworld.wolfram.com/AssociatedLegendrePolynomial.html\r\n for j in range(i + 2, k):\r\n P_l_m[j][i] = sym.simplify(\r\n ((2 * j - 1) * z * P_l_m[j - 1][i] -\r\n (i + j - 1) * P_l_m[j - 2][i]) / (j - i)) # use the property of eq.7: https://mathworld.wolfram.com/AssociatedLegendrePolynomial.html\r\n\r\n return P_l_m\r\n```\r\nAfter fixing this tiny bug, all the results are the right\n", "before_files": [{"content": "# Shameless steal from: https://github.com/klicperajo/dimenet\n\nimport numpy as np\nfrom scipy import special as sp\nfrom scipy.optimize import brentq\n\ntry:\n import sympy as sym\nexcept ImportError:\n sym = None\n\n\ndef Jn(r, n):\n return np.sqrt(np.pi / (2 * r)) * sp.jv(n + 0.5, r)\n\n\ndef Jn_zeros(n, k):\n zerosj = np.zeros((n, k), dtype='float32')\n zerosj[0] = np.arange(1, k + 1) * np.pi\n points = np.arange(1, k + n) * np.pi\n racines = np.zeros(k + n - 1, dtype='float32')\n for i in range(1, n):\n for j in range(k + n - 1 - i):\n foo = brentq(Jn, points[j], points[j + 1], (i, ))\n racines[j] = foo\n points = racines\n zerosj[i][:k] = racines[:k]\n\n return zerosj\n\n\ndef spherical_bessel_formulas(n):\n x = sym.symbols('x')\n\n f = [sym.sin(x) / x]\n a = sym.sin(x) / x\n for i in range(1, n):\n b = sym.diff(a, x) / x\n f += [sym.simplify(b * (-x)**i)]\n a = sym.simplify(b)\n return f\n\n\ndef bessel_basis(n, k):\n zeros = Jn_zeros(n, k)\n normalizer = []\n for order in range(n):\n normalizer_tmp = []\n for i in range(k):\n normalizer_tmp += [0.5 * Jn(zeros[order, i], order + 1)**2]\n normalizer_tmp = 1 / np.array(normalizer_tmp)**0.5\n normalizer += [normalizer_tmp]\n\n f = spherical_bessel_formulas(n)\n x = sym.symbols('x')\n bess_basis = []\n for order in range(n):\n bess_basis_tmp = []\n for i in range(k):\n bess_basis_tmp += [\n sym.simplify(normalizer[order][i] *\n f[order].subs(x, zeros[order, i] * x))\n ]\n bess_basis += [bess_basis_tmp]\n return bess_basis\n\n\ndef sph_harm_prefactor(k, m):\n return ((2 * k + 1) * np.math.factorial(k - abs(m)) /\n (4 * np.pi * np.math.factorial(k + abs(m))))**0.5\n\n\ndef associated_legendre_polynomials(k, zero_m_only=True):\n z = sym.symbols('z')\n P_l_m = [[0] * (j + 1) for j in range(k)]\n\n P_l_m[0][0] = 1\n if k > 0:\n P_l_m[1][0] = z\n\n for j in range(2, k):\n P_l_m[j][0] = sym.simplify(((2 * j - 1) * z * P_l_m[j - 1][0] -\n (j - 1) * P_l_m[j - 2][0]) / j)\n if not zero_m_only:\n for i in range(1, k):\n P_l_m[i][i] = sym.simplify((1 - 2 * i) * P_l_m[i - 1][i - 1])\n if i + 1 < k:\n P_l_m[i + 1][i] = sym.simplify(\n (2 * i + 1) * z * P_l_m[i][i])\n for j in range(i + 2, k):\n P_l_m[j][i] = sym.simplify(\n ((2 * j - 1) * z * P_l_m[j - 1][i] -\n (i + j - 1) * P_l_m[j - 2][i]) / (j - i))\n\n return P_l_m\n\n\ndef real_sph_harm(k, zero_m_only=True, spherical_coordinates=True):\n if not zero_m_only:\n S_m = [0]\n C_m = [1]\n for i in range(1, k):\n x = sym.symbols('x')\n y = sym.symbols('y')\n S_m += [x * S_m[i - 1] + y * C_m[i - 1]]\n C_m += [x * C_m[i - 1] - y * S_m[i - 1]]\n\n P_l_m = associated_legendre_polynomials(k, zero_m_only)\n if spherical_coordinates:\n theta = sym.symbols('theta')\n z = sym.symbols('z')\n for i in range(len(P_l_m)):\n for j in range(len(P_l_m[i])):\n if not isinstance(P_l_m[i][j], int):\n P_l_m[i][j] = P_l_m[i][j].subs(z, sym.cos(theta))\n if not zero_m_only:\n phi = sym.symbols('phi')\n for i in range(len(S_m)):\n S_m[i] = S_m[i].subs(x,\n sym.sin(theta) * sym.cos(phi)).subs(\n y,\n sym.sin(theta) * sym.sin(phi))\n for i in range(len(C_m)):\n C_m[i] = C_m[i].subs(x,\n sym.sin(theta) * sym.cos(phi)).subs(\n y,\n sym.sin(theta) * sym.sin(phi))\n\n Y_func_l_m = [['0'] * (2 * j + 1) for j in range(k)]\n for i in range(k):\n Y_func_l_m[i][0] = sym.simplify(sph_harm_prefactor(i, 0) * P_l_m[i][0])\n\n if not zero_m_only:\n for i in range(1, k):\n for j in range(1, i + 1):\n Y_func_l_m[i][j] = sym.simplify(\n 2**0.5 * sph_harm_prefactor(i, j) * C_m[j] * P_l_m[i][j])\n for i in range(1, k):\n for j in range(1, i + 1):\n Y_func_l_m[i][-j] = sym.simplify(\n 2**0.5 * sph_harm_prefactor(i, -j) * S_m[j] * P_l_m[i][j])\n\n return Y_func_l_m\n", "path": "torch_geometric/nn/models/dimenet_utils.py"}]}
| 3,098 | 538 |
gh_patches_debug_60773
|
rasdani/github-patches
|
git_diff
|
data-for-change__anyway-1848
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Fix walla scraping - see test_scrape_sanity_online_walla
</issue>
<code>
[start of anyway/parsers/rss_sites.py]
1 import requests
2 from bs4 import BeautifulSoup
3 import feedparser
4 from anyway.parsers import timezones
5
6
7 def parse_html_walla(item_rss, html_soup):
8 # For some reason there's html here
9 description = BeautifulSoup(item_rss["summary"], features="lxml").text
10
11 author = html_soup.find("div", class_="author").find("a").get_text()
12 return author, description
13
14
15 def parse_html_ynet(item_rss, html_soup):
16 # This is rather fragile
17 # description_text: "[description] ([author]) [unrelated stuff]"
18 description_text = html_soup.find(id="ArticleBodyComponent").get_text()
19 author = description_text.split("(")[-1].split(")")[0].strip()
20 description = description_text.rsplit("(")[0].strip()
21 return author, description
22
23
24 sites_config = {
25 "ynet": {
26 "rss": "https://www.ynet.co.il:443/Integration/StoryRss1854.xml",
27 "parser": parse_html_ynet,
28 },
29 "walla": {"rss": "https://rss.walla.co.il:443/feed/22", "parser": parse_html_walla},
30 }
31
32
33 def _fetch(url: str) -> str:
34 return requests.get(url).text
35
36
37 def scrape_raw(site_name: str, *, rss_source=None, fetch_html=_fetch):
38 config = sites_config[site_name]
39 if rss_source is None:
40 rss_source = config["rss"]
41 rss_dict = feedparser.parse(rss_source)
42 if rss_dict.get("bozo_exception"):
43 raise rss_dict["bozo_exception"]
44
45 for item_rss in rss_dict["items"]:
46 html_text = fetch_html(item_rss["link"])
47 author, description = config["parser"](item_rss, BeautifulSoup(html_text, "lxml"))
48 yield {
49 "link": item_rss["link"],
50 "date": timezones.from_rss(item_rss["published_parsed"]),
51 "source": site_name,
52 "author": author,
53 "title": item_rss["title"],
54 "description": description,
55 "accident": False,
56 }
57
58
59 def scrape(*args, **kwargs):
60 # lazily load dependencies, so this module will behave like an independent library
61 from anyway.models import NewsFlash
62
63 for dict_item in scrape_raw(*args, **kwargs):
64 yield NewsFlash(**dict_item)
65
[end of anyway/parsers/rss_sites.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/anyway/parsers/rss_sites.py b/anyway/parsers/rss_sites.py
--- a/anyway/parsers/rss_sites.py
+++ b/anyway/parsers/rss_sites.py
@@ -8,7 +8,7 @@
# For some reason there's html here
description = BeautifulSoup(item_rss["summary"], features="lxml").text
- author = html_soup.find("div", class_="author").find("a").get_text()
+ author = html_soup.find("div", class_="author").get_text().strip()
return author, description
|
{"golden_diff": "diff --git a/anyway/parsers/rss_sites.py b/anyway/parsers/rss_sites.py\n--- a/anyway/parsers/rss_sites.py\n+++ b/anyway/parsers/rss_sites.py\n@@ -8,7 +8,7 @@\n # For some reason there's html here\n description = BeautifulSoup(item_rss[\"summary\"], features=\"lxml\").text\n \n- author = html_soup.find(\"div\", class_=\"author\").find(\"a\").get_text()\n+ author = html_soup.find(\"div\", class_=\"author\").get_text().strip()\n return author, description\n", "issue": "Fix walla scraping - see test_scrape_sanity_online_walla\n\n", "before_files": [{"content": "import requests\nfrom bs4 import BeautifulSoup\nimport feedparser\nfrom anyway.parsers import timezones\n\n\ndef parse_html_walla(item_rss, html_soup):\n # For some reason there's html here\n description = BeautifulSoup(item_rss[\"summary\"], features=\"lxml\").text\n\n author = html_soup.find(\"div\", class_=\"author\").find(\"a\").get_text()\n return author, description\n\n\ndef parse_html_ynet(item_rss, html_soup):\n # This is rather fragile\n # description_text: \"[description] ([author]) [unrelated stuff]\"\n description_text = html_soup.find(id=\"ArticleBodyComponent\").get_text()\n author = description_text.split(\"(\")[-1].split(\")\")[0].strip()\n description = description_text.rsplit(\"(\")[0].strip()\n return author, description\n\n\nsites_config = {\n \"ynet\": {\n \"rss\": \"https://www.ynet.co.il:443/Integration/StoryRss1854.xml\",\n \"parser\": parse_html_ynet,\n },\n \"walla\": {\"rss\": \"https://rss.walla.co.il:443/feed/22\", \"parser\": parse_html_walla},\n}\n\n\ndef _fetch(url: str) -> str:\n return requests.get(url).text\n\n\ndef scrape_raw(site_name: str, *, rss_source=None, fetch_html=_fetch):\n config = sites_config[site_name]\n if rss_source is None:\n rss_source = config[\"rss\"]\n rss_dict = feedparser.parse(rss_source)\n if rss_dict.get(\"bozo_exception\"):\n raise rss_dict[\"bozo_exception\"]\n\n for item_rss in rss_dict[\"items\"]:\n html_text = fetch_html(item_rss[\"link\"])\n author, description = config[\"parser\"](item_rss, BeautifulSoup(html_text, \"lxml\"))\n yield {\n \"link\": item_rss[\"link\"],\n \"date\": timezones.from_rss(item_rss[\"published_parsed\"]),\n \"source\": site_name,\n \"author\": author,\n \"title\": item_rss[\"title\"],\n \"description\": description,\n \"accident\": False,\n }\n\n\ndef scrape(*args, **kwargs):\n # lazily load dependencies, so this module will behave like an independent library\n from anyway.models import NewsFlash\n\n for dict_item in scrape_raw(*args, **kwargs):\n yield NewsFlash(**dict_item)\n", "path": "anyway/parsers/rss_sites.py"}]}
| 1,199 | 129 |
gh_patches_debug_25200
|
rasdani/github-patches
|
git_diff
|
internetarchive__openlibrary-6853
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
DeprecationWarning: `urllib.parse.splitquery()` is deprecated as of 3.8
<!-- What problem are we solving? What does the experience look like today? What are the symptoms? -->
### Evidence / Screenshot (if possible)
When we run pytest in our GitHub Actions, we get the following warning
```
=============================== warnings summary ===============================
openlibrary/coverstore/tests/test_doctests.py::test_doctest[openlibrary.coverstore.utils]
openlibrary/coverstore/tests/test_doctests.py::test_doctest[openlibrary.coverstore.utils]
openlibrary/coverstore/tests/test_doctests.py::test_doctest[openlibrary.coverstore.utils]
/home/runner/work/openlibrary/openlibrary/openlibrary/coverstore/utils.py:88:
DeprecationWarning: urllib.parse.splitquery() is deprecated as of 3.8, use urllib.parse.urlparse() instead
base, query = splitquery(URL)
```
https://github.com/internetarchive/openlibrary/blob/master/openlibrary/coverstore/utils.py#L88
### Relevant url?
<!-- `https://openlibrary.org/...` -->
### Steps to Reproduce
<!-- What steps caused you to find the bug? -->
1. Go to ...
2. Do ...
<!-- What actually happened after these steps? What did you expect to happen? -->
* Actual: pytest warning
* Expected: No pytest warning
### Details
- **Logged in (Y/N)?**
- **Browser type/version?**
- **Operating system?**
- **Environment (prod/dev/local)?** prod
<!-- If not sure, put prod -->
### Proposal & Constraints
<!-- What is the proposed solution / implementation? Is there a precedent of this approach succeeding elsewhere? -->
### Related files
<!-- Files related to this issue; this is super useful for new contributors who might want to help! If you're not sure, leave this blank; a maintainer will add them. -->
### Stakeholders
<!-- @ tag stakeholders of this bug -->
</issue>
<code>
[start of openlibrary/coverstore/utils.py]
1 """Utilities for coverstore"""
2
3 import json
4 import mimetypes
5 import os
6 import random
7 import socket
8 import string
9
10 import requests
11 import web
12 from urllib.parse import splitquery, unquote, unquote_plus # type: ignore[attr-defined]
13 from urllib.parse import urlencode as real_urlencode
14
15 from openlibrary.coverstore import config, oldb
16
17 from io import IOBase as file
18
19 socket.setdefaulttimeout(10.0)
20
21
22 def safeint(value, default=None):
23 """
24 >>> safeint('1')
25 1
26 >>> safeint('x')
27 >>> safeint('x', 0)
28 0
29 """
30 try:
31 return int(value)
32 except (TypeError, ValueError):
33 return default
34
35
36 def get_ol_url():
37 return web.rstrips(config.ol_url, "/")
38
39
40 def ol_things(key, value):
41 if oldb.is_supported():
42 return oldb.query(key, value)
43 else:
44 query = {
45 'type': '/type/edition',
46 key: value,
47 'sort': 'last_modified',
48 'limit': 10,
49 }
50 try:
51 d = dict(query=json.dumps(query))
52 result = download(get_ol_url() + '/api/things?' + real_urlencode(d))
53 result = json.loads(result)
54 return result['result']
55 except OSError:
56 import traceback
57
58 traceback.print_exc()
59 return []
60
61
62 def ol_get(olkey):
63 if oldb.is_supported():
64 return oldb.get(olkey)
65 else:
66 try:
67 return json.loads(download(get_ol_url() + olkey + ".json"))
68 except OSError:
69 return None
70
71
72 USER_AGENT = (
73 "Mozilla/5.0 (Compatible; coverstore downloader http://covers.openlibrary.org)"
74 )
75
76
77 def download(url):
78 return requests.get(url, headers={'User-Agent': USER_AGENT}).content
79
80
81 def urldecode(url):
82 """
83 >>> urldecode('http://google.com/search?q=bar&x=y')
84 ('http://google.com/search', {'q': 'bar', 'x': 'y'})
85 >>> urldecode('http://google.com/')
86 ('http://google.com/', {})
87 """
88 base, query = splitquery(url)
89 query = query or ""
90 items = [item.split('=', 1) for item in query.split('&') if '=' in item]
91 d = {unquote(k): unquote_plus(v) for (k, v) in items}
92 return base, d
93
94
95 def changequery(url, **kw):
96 """
97 >>> changequery('http://google.com/search?q=foo', q='bar', x='y')
98 'http://google.com/search?q=bar&x=y'
99 """
100 base, params = urldecode(url)
101 params.update(kw)
102 return base + '?' + real_urlencode(params)
103
104
105 def read_file(path, offset, size, chunk=50 * 1024):
106 """Returns an iterator over file data at specified offset and size.
107
108 >>> len(b"".join(read_file('/dev/urandom', 100, 10000)))
109 10000
110 """
111 with open(path, "rb") as f:
112 f.seek(offset)
113 while size:
114 data = f.read(min(chunk, size))
115 size -= len(data)
116 if data:
117 yield data
118 else:
119 raise OSError("file truncated")
120
121
122 def rm_f(filename):
123 try:
124 os.remove(filename)
125 except OSError:
126 pass
127
128
129 chars = string.ascii_letters + string.digits
130
131
132 def random_string(n):
133 return "".join([random.choice(chars) for i in range(n)])
134
135
136 def urlencode(data):
137 """
138 urlencodes the given data dictionary. If any of the value is a file object, data is multipart encoded.
139
140 @@@ should go into web.browser
141 """
142 multipart = False
143 for v in data.values():
144 if isinstance(v, file):
145 multipart = True
146 break
147
148 if not multipart:
149 return 'application/x-www-form-urlencoded', real_urlencode(data)
150 else:
151 # adopted from http://code.activestate.com/recipes/146306/
152 def get_content_type(filename):
153 return mimetypes.guess_type(filename)[0] or 'application/octet-stream'
154
155 def encode(key, value, out):
156 if isinstance(value, file):
157 out.append('--' + BOUNDARY)
158 out.append(
159 f'Content-Disposition: form-data; name="{key}"; filename="{value.name}"'
160 )
161 out.append('Content-Type: %s' % get_content_type(value.name))
162 out.append('')
163 out.append(value.read())
164 elif isinstance(value, list):
165 for v in value:
166 encode(key, v)
167 else:
168 out.append('--' + BOUNDARY)
169 out.append('Content-Disposition: form-data; name="%s"' % key)
170 out.append('')
171 out.append(value)
172
173 BOUNDARY = "----------ThIs_Is_tHe_bouNdaRY_$"
174 CRLF = '\r\n'
175 out = []
176 for k, v in data.items():
177 encode(k, v, out)
178 body = CRLF.join(out)
179 content_type = 'multipart/form-data; boundary=%s' % BOUNDARY
180 return content_type, body
181
182
183 if __name__ == "__main__":
184 import doctest
185
186 doctest.testmod()
187
[end of openlibrary/coverstore/utils.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/openlibrary/coverstore/utils.py b/openlibrary/coverstore/utils.py
--- a/openlibrary/coverstore/utils.py
+++ b/openlibrary/coverstore/utils.py
@@ -9,7 +9,7 @@
import requests
import web
-from urllib.parse import splitquery, unquote, unquote_plus # type: ignore[attr-defined]
+from urllib.parse import urlsplit, urlunsplit, parse_qsl, unquote, unquote_plus # type: ignore[attr-defined]
from urllib.parse import urlencode as real_urlencode
from openlibrary.coverstore import config, oldb
@@ -78,17 +78,17 @@
return requests.get(url, headers={'User-Agent': USER_AGENT}).content
-def urldecode(url):
+def urldecode(url: str) -> tuple[str, dict[str, str]]:
"""
>>> urldecode('http://google.com/search?q=bar&x=y')
('http://google.com/search', {'q': 'bar', 'x': 'y'})
>>> urldecode('http://google.com/')
('http://google.com/', {})
"""
- base, query = splitquery(url)
- query = query or ""
- items = [item.split('=', 1) for item in query.split('&') if '=' in item]
+ split_url = urlsplit(url)
+ items = parse_qsl(split_url.query)
d = {unquote(k): unquote_plus(v) for (k, v) in items}
+ base = urlunsplit(split_url._replace(query=''))
return base, d
|
{"golden_diff": "diff --git a/openlibrary/coverstore/utils.py b/openlibrary/coverstore/utils.py\n--- a/openlibrary/coverstore/utils.py\n+++ b/openlibrary/coverstore/utils.py\n@@ -9,7 +9,7 @@\n \n import requests\n import web\n-from urllib.parse import splitquery, unquote, unquote_plus # type: ignore[attr-defined]\n+from urllib.parse import urlsplit, urlunsplit, parse_qsl, unquote, unquote_plus # type: ignore[attr-defined]\n from urllib.parse import urlencode as real_urlencode\n \n from openlibrary.coverstore import config, oldb\n@@ -78,17 +78,17 @@\n return requests.get(url, headers={'User-Agent': USER_AGENT}).content\n \n \n-def urldecode(url):\n+def urldecode(url: str) -> tuple[str, dict[str, str]]:\n \"\"\"\n >>> urldecode('http://google.com/search?q=bar&x=y')\n ('http://google.com/search', {'q': 'bar', 'x': 'y'})\n >>> urldecode('http://google.com/')\n ('http://google.com/', {})\n \"\"\"\n- base, query = splitquery(url)\n- query = query or \"\"\n- items = [item.split('=', 1) for item in query.split('&') if '=' in item]\n+ split_url = urlsplit(url)\n+ items = parse_qsl(split_url.query)\n d = {unquote(k): unquote_plus(v) for (k, v) in items}\n+ base = urlunsplit(split_url._replace(query=''))\n return base, d\n", "issue": "DeprecationWarning: `urllib.parse.splitquery()` is deprecated as of 3.8\n<!-- What problem are we solving? What does the experience look like today? What are the symptoms? -->\r\n\r\n\r\n### Evidence / Screenshot (if possible)\r\nWhen we run pytest in our GitHub Actions, we get the following warning\r\n```\r\n=============================== warnings summary ===============================\r\nopenlibrary/coverstore/tests/test_doctests.py::test_doctest[openlibrary.coverstore.utils]\r\nopenlibrary/coverstore/tests/test_doctests.py::test_doctest[openlibrary.coverstore.utils]\r\nopenlibrary/coverstore/tests/test_doctests.py::test_doctest[openlibrary.coverstore.utils]\r\n /home/runner/work/openlibrary/openlibrary/openlibrary/coverstore/utils.py:88:\r\n DeprecationWarning: urllib.parse.splitquery() is deprecated as of 3.8, use urllib.parse.urlparse() instead\r\n base, query = splitquery(URL)\r\n```\r\nhttps://github.com/internetarchive/openlibrary/blob/master/openlibrary/coverstore/utils.py#L88\r\n\r\n### Relevant url?\r\n<!-- `https://openlibrary.org/...` -->\r\n\r\n### Steps to Reproduce\r\n<!-- What steps caused you to find the bug? -->\r\n1. Go to ...\r\n2. Do ...\r\n\r\n<!-- What actually happened after these steps? What did you expect to happen? -->\r\n* Actual: pytest warning\r\n* Expected: No pytest warning\r\n\r\n### Details\r\n\r\n- **Logged in (Y/N)?**\r\n- **Browser type/version?**\r\n- **Operating system?**\r\n- **Environment (prod/dev/local)?** prod\r\n<!-- If not sure, put prod -->\r\n\r\n### Proposal & Constraints\r\n<!-- What is the proposed solution / implementation? Is there a precedent of this approach succeeding elsewhere? -->\r\n\r\n### Related files\r\n<!-- Files related to this issue; this is super useful for new contributors who might want to help! If you're not sure, leave this blank; a maintainer will add them. -->\r\n\r\n### Stakeholders\r\n<!-- @ tag stakeholders of this bug -->\r\n\n", "before_files": [{"content": "\"\"\"Utilities for coverstore\"\"\"\n\nimport json\nimport mimetypes\nimport os\nimport random\nimport socket\nimport string\n\nimport requests\nimport web\nfrom urllib.parse import splitquery, unquote, unquote_plus # type: ignore[attr-defined]\nfrom urllib.parse import urlencode as real_urlencode\n\nfrom openlibrary.coverstore import config, oldb\n\nfrom io import IOBase as file\n\nsocket.setdefaulttimeout(10.0)\n\n\ndef safeint(value, default=None):\n \"\"\"\n >>> safeint('1')\n 1\n >>> safeint('x')\n >>> safeint('x', 0)\n 0\n \"\"\"\n try:\n return int(value)\n except (TypeError, ValueError):\n return default\n\n\ndef get_ol_url():\n return web.rstrips(config.ol_url, \"/\")\n\n\ndef ol_things(key, value):\n if oldb.is_supported():\n return oldb.query(key, value)\n else:\n query = {\n 'type': '/type/edition',\n key: value,\n 'sort': 'last_modified',\n 'limit': 10,\n }\n try:\n d = dict(query=json.dumps(query))\n result = download(get_ol_url() + '/api/things?' + real_urlencode(d))\n result = json.loads(result)\n return result['result']\n except OSError:\n import traceback\n\n traceback.print_exc()\n return []\n\n\ndef ol_get(olkey):\n if oldb.is_supported():\n return oldb.get(olkey)\n else:\n try:\n return json.loads(download(get_ol_url() + olkey + \".json\"))\n except OSError:\n return None\n\n\nUSER_AGENT = (\n \"Mozilla/5.0 (Compatible; coverstore downloader http://covers.openlibrary.org)\"\n)\n\n\ndef download(url):\n return requests.get(url, headers={'User-Agent': USER_AGENT}).content\n\n\ndef urldecode(url):\n \"\"\"\n >>> urldecode('http://google.com/search?q=bar&x=y')\n ('http://google.com/search', {'q': 'bar', 'x': 'y'})\n >>> urldecode('http://google.com/')\n ('http://google.com/', {})\n \"\"\"\n base, query = splitquery(url)\n query = query or \"\"\n items = [item.split('=', 1) for item in query.split('&') if '=' in item]\n d = {unquote(k): unquote_plus(v) for (k, v) in items}\n return base, d\n\n\ndef changequery(url, **kw):\n \"\"\"\n >>> changequery('http://google.com/search?q=foo', q='bar', x='y')\n 'http://google.com/search?q=bar&x=y'\n \"\"\"\n base, params = urldecode(url)\n params.update(kw)\n return base + '?' + real_urlencode(params)\n\n\ndef read_file(path, offset, size, chunk=50 * 1024):\n \"\"\"Returns an iterator over file data at specified offset and size.\n\n >>> len(b\"\".join(read_file('/dev/urandom', 100, 10000)))\n 10000\n \"\"\"\n with open(path, \"rb\") as f:\n f.seek(offset)\n while size:\n data = f.read(min(chunk, size))\n size -= len(data)\n if data:\n yield data\n else:\n raise OSError(\"file truncated\")\n\n\ndef rm_f(filename):\n try:\n os.remove(filename)\n except OSError:\n pass\n\n\nchars = string.ascii_letters + string.digits\n\n\ndef random_string(n):\n return \"\".join([random.choice(chars) for i in range(n)])\n\n\ndef urlencode(data):\n \"\"\"\n urlencodes the given data dictionary. If any of the value is a file object, data is multipart encoded.\n\n @@@ should go into web.browser\n \"\"\"\n multipart = False\n for v in data.values():\n if isinstance(v, file):\n multipart = True\n break\n\n if not multipart:\n return 'application/x-www-form-urlencoded', real_urlencode(data)\n else:\n # adopted from http://code.activestate.com/recipes/146306/\n def get_content_type(filename):\n return mimetypes.guess_type(filename)[0] or 'application/octet-stream'\n\n def encode(key, value, out):\n if isinstance(value, file):\n out.append('--' + BOUNDARY)\n out.append(\n f'Content-Disposition: form-data; name=\"{key}\"; filename=\"{value.name}\"'\n )\n out.append('Content-Type: %s' % get_content_type(value.name))\n out.append('')\n out.append(value.read())\n elif isinstance(value, list):\n for v in value:\n encode(key, v)\n else:\n out.append('--' + BOUNDARY)\n out.append('Content-Disposition: form-data; name=\"%s\"' % key)\n out.append('')\n out.append(value)\n\n BOUNDARY = \"----------ThIs_Is_tHe_bouNdaRY_$\"\n CRLF = '\\r\\n'\n out = []\n for k, v in data.items():\n encode(k, v, out)\n body = CRLF.join(out)\n content_type = 'multipart/form-data; boundary=%s' % BOUNDARY\n return content_type, body\n\n\nif __name__ == \"__main__\":\n import doctest\n\n doctest.testmod()\n", "path": "openlibrary/coverstore/utils.py"}]}
| 2,597 | 344 |
gh_patches_debug_17789
|
rasdani/github-patches
|
git_diff
|
encode__starlette-1018
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Memory usage streaming large responses
We've been running into memory issues when providing very large async generators to a streaming response. We have these generators producing large (larger than memory set) responses in a way that allows us to only keep small chunks in memory at a time. However, it looks like the BaseHTTPMiddleware implementation uses an asyncio queue to store the individual chunks:
https://github.com/encode/starlette/blob/master/starlette/middleware/base.py#L30
This prevents any network backpressure handling -- if the client that is receiving the streaming response is on a slow connection, the queue will happily grow without bound and consume all memory, triggering kernel out-of-memory, when the ideal handling here would be for send to block (yield) when this happens. I believe this would naturally happen if there were no queue here at all, so I am wondering why it needs to be here?
Would a PR to remove the queueing be accepted?
If not, what is the appropriate way to override this to not use a queue? We can write our own, but the use of BaseHTTPMiddleware is hardcoded: https://github.com/encode/starlette/blob/519f5750b5e797bb3d4805fd29657674304ce397/starlette/applications.py#L197, leaving only some fairly hacky approaches to preventing this queueing.
</issue>
<code>
[start of starlette/middleware/base.py]
1 import asyncio
2 import typing
3
4 from starlette.requests import Request
5 from starlette.responses import Response, StreamingResponse
6 from starlette.types import ASGIApp, Receive, Scope, Send
7
8 RequestResponseEndpoint = typing.Callable[[Request], typing.Awaitable[Response]]
9 DispatchFunction = typing.Callable[
10 [Request, RequestResponseEndpoint], typing.Awaitable[Response]
11 ]
12
13
14 class BaseHTTPMiddleware:
15 def __init__(self, app: ASGIApp, dispatch: DispatchFunction = None) -> None:
16 self.app = app
17 self.dispatch_func = self.dispatch if dispatch is None else dispatch
18
19 async def __call__(self, scope: Scope, receive: Receive, send: Send) -> None:
20 if scope["type"] != "http":
21 await self.app(scope, receive, send)
22 return
23
24 request = Request(scope, receive=receive)
25 response = await self.dispatch_func(request, self.call_next)
26 await response(scope, receive, send)
27
28 async def call_next(self, request: Request) -> Response:
29 loop = asyncio.get_event_loop()
30 queue = asyncio.Queue() # type: asyncio.Queue
31
32 scope = request.scope
33 receive = request.receive
34 send = queue.put
35
36 async def coro() -> None:
37 try:
38 await self.app(scope, receive, send)
39 finally:
40 await queue.put(None)
41
42 task = loop.create_task(coro())
43 message = await queue.get()
44 if message is None:
45 task.result()
46 raise RuntimeError("No response returned.")
47 assert message["type"] == "http.response.start"
48
49 async def body_stream() -> typing.AsyncGenerator[bytes, None]:
50 while True:
51 message = await queue.get()
52 if message is None:
53 break
54 assert message["type"] == "http.response.body"
55 yield message.get("body", b"")
56 task.result()
57
58 response = StreamingResponse(
59 status_code=message["status"], content=body_stream()
60 )
61 response.raw_headers = message["headers"]
62 return response
63
64 async def dispatch(
65 self, request: Request, call_next: RequestResponseEndpoint
66 ) -> Response:
67 raise NotImplementedError() # pragma: no cover
68
[end of starlette/middleware/base.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/starlette/middleware/base.py b/starlette/middleware/base.py
--- a/starlette/middleware/base.py
+++ b/starlette/middleware/base.py
@@ -3,7 +3,7 @@
from starlette.requests import Request
from starlette.responses import Response, StreamingResponse
-from starlette.types import ASGIApp, Receive, Scope, Send
+from starlette.types import ASGIApp, Message, Receive, Scope, Send
RequestResponseEndpoint = typing.Callable[[Request], typing.Awaitable[Response]]
DispatchFunction = typing.Callable[
@@ -27,7 +27,7 @@
async def call_next(self, request: Request) -> Response:
loop = asyncio.get_event_loop()
- queue = asyncio.Queue() # type: asyncio.Queue
+ queue: "asyncio.Queue[typing.Optional[Message]]" = asyncio.Queue(maxsize=1)
scope = request.scope
receive = request.receive
|
{"golden_diff": "diff --git a/starlette/middleware/base.py b/starlette/middleware/base.py\n--- a/starlette/middleware/base.py\n+++ b/starlette/middleware/base.py\n@@ -3,7 +3,7 @@\n \n from starlette.requests import Request\n from starlette.responses import Response, StreamingResponse\n-from starlette.types import ASGIApp, Receive, Scope, Send\n+from starlette.types import ASGIApp, Message, Receive, Scope, Send\n \n RequestResponseEndpoint = typing.Callable[[Request], typing.Awaitable[Response]]\n DispatchFunction = typing.Callable[\n@@ -27,7 +27,7 @@\n \n async def call_next(self, request: Request) -> Response:\n loop = asyncio.get_event_loop()\n- queue = asyncio.Queue() # type: asyncio.Queue\n+ queue: \"asyncio.Queue[typing.Optional[Message]]\" = asyncio.Queue(maxsize=1)\n \n scope = request.scope\n receive = request.receive\n", "issue": "Memory usage streaming large responses\nWe've been running into memory issues when providing very large async generators to a streaming response. We have these generators producing large (larger than memory set) responses in a way that allows us to only keep small chunks in memory at a time. However, it looks like the BaseHTTPMiddleware implementation uses an asyncio queue to store the individual chunks:\r\n\r\nhttps://github.com/encode/starlette/blob/master/starlette/middleware/base.py#L30\r\n\r\nThis prevents any network backpressure handling -- if the client that is receiving the streaming response is on a slow connection, the queue will happily grow without bound and consume all memory, triggering kernel out-of-memory, when the ideal handling here would be for send to block (yield) when this happens. I believe this would naturally happen if there were no queue here at all, so I am wondering why it needs to be here?\r\n\r\nWould a PR to remove the queueing be accepted?\r\n\r\nIf not, what is the appropriate way to override this to not use a queue? We can write our own, but the use of BaseHTTPMiddleware is hardcoded: https://github.com/encode/starlette/blob/519f5750b5e797bb3d4805fd29657674304ce397/starlette/applications.py#L197, leaving only some fairly hacky approaches to preventing this queueing.\n", "before_files": [{"content": "import asyncio\nimport typing\n\nfrom starlette.requests import Request\nfrom starlette.responses import Response, StreamingResponse\nfrom starlette.types import ASGIApp, Receive, Scope, Send\n\nRequestResponseEndpoint = typing.Callable[[Request], typing.Awaitable[Response]]\nDispatchFunction = typing.Callable[\n [Request, RequestResponseEndpoint], typing.Awaitable[Response]\n]\n\n\nclass BaseHTTPMiddleware:\n def __init__(self, app: ASGIApp, dispatch: DispatchFunction = None) -> None:\n self.app = app\n self.dispatch_func = self.dispatch if dispatch is None else dispatch\n\n async def __call__(self, scope: Scope, receive: Receive, send: Send) -> None:\n if scope[\"type\"] != \"http\":\n await self.app(scope, receive, send)\n return\n\n request = Request(scope, receive=receive)\n response = await self.dispatch_func(request, self.call_next)\n await response(scope, receive, send)\n\n async def call_next(self, request: Request) -> Response:\n loop = asyncio.get_event_loop()\n queue = asyncio.Queue() # type: asyncio.Queue\n\n scope = request.scope\n receive = request.receive\n send = queue.put\n\n async def coro() -> None:\n try:\n await self.app(scope, receive, send)\n finally:\n await queue.put(None)\n\n task = loop.create_task(coro())\n message = await queue.get()\n if message is None:\n task.result()\n raise RuntimeError(\"No response returned.\")\n assert message[\"type\"] == \"http.response.start\"\n\n async def body_stream() -> typing.AsyncGenerator[bytes, None]:\n while True:\n message = await queue.get()\n if message is None:\n break\n assert message[\"type\"] == \"http.response.body\"\n yield message.get(\"body\", b\"\")\n task.result()\n\n response = StreamingResponse(\n status_code=message[\"status\"], content=body_stream()\n )\n response.raw_headers = message[\"headers\"]\n return response\n\n async def dispatch(\n self, request: Request, call_next: RequestResponseEndpoint\n ) -> Response:\n raise NotImplementedError() # pragma: no cover\n", "path": "starlette/middleware/base.py"}]}
| 1,442 | 206 |
gh_patches_debug_26952
|
rasdani/github-patches
|
git_diff
|
certbot__certbot-2852
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Ensure rename doesn't break third party plugins
There are a lot of third party plugins for Let's Encrypt (listed here: https://letsencrypt.readthedocs.org/en/latest/using.html#plugins)
It's possible that they reference things in such a way that changing the name or package of LE will break their compatibility with the client - we should ensure that they are unbroken
</issue>
<code>
[start of certbot/constants.py]
1 """Certbot constants."""
2 import os
3 import logging
4
5 from acme import challenges
6
7
8 SETUPTOOLS_PLUGINS_ENTRY_POINT = "certbot.plugins"
9 """Setuptools entry point group name for plugins."""
10
11 CLI_DEFAULTS = dict(
12 config_files=[
13 "/etc/letsencrypt/cli.ini",
14 # http://freedesktop.org/wiki/Software/xdg-user-dirs/
15 os.path.join(os.environ.get("XDG_CONFIG_HOME", "~/.config"),
16 "letsencrypt", "cli.ini"),
17 ],
18 verbose_count=-(logging.WARNING / 10),
19 server="https://acme-v01.api.letsencrypt.org/directory",
20 rsa_key_size=2048,
21 rollback_checkpoints=1,
22 config_dir="/etc/letsencrypt",
23 work_dir="/var/lib/letsencrypt",
24 logs_dir="/var/log/letsencrypt",
25 no_verify_ssl=False,
26 http01_port=challenges.HTTP01Response.PORT,
27 tls_sni_01_port=challenges.TLSSNI01Response.PORT,
28
29 auth_cert_path="./cert.pem",
30 auth_chain_path="./chain.pem",
31 strict_permissions=False,
32 )
33 STAGING_URI = "https://acme-staging.api.letsencrypt.org/directory"
34
35 """Defaults for CLI flags and `.IConfig` attributes."""
36
37 RENEWER_DEFAULTS = dict(
38 renewer_enabled="yes",
39 renew_before_expiry="30 days",
40 # This value should ensure that there is never a deployment delay by
41 # default.
42 deploy_before_expiry="99 years",
43 )
44 """Defaults for renewer script."""
45
46
47 ENHANCEMENTS = ["redirect", "http-header", "ocsp-stapling", "spdy"]
48 """List of possible :class:`certbot.interfaces.IInstaller`
49 enhancements.
50
51 List of expected options parameters:
52 - redirect: None
53 - http-header: TODO
54 - ocsp-stapling: TODO
55 - spdy: TODO
56
57 """
58
59 ARCHIVE_DIR = "archive"
60 """Archive directory, relative to `IConfig.config_dir`."""
61
62 CONFIG_DIRS_MODE = 0o755
63 """Directory mode for ``.IConfig.config_dir`` et al."""
64
65 ACCOUNTS_DIR = "accounts"
66 """Directory where all accounts are saved."""
67
68 BACKUP_DIR = "backups"
69 """Directory (relative to `IConfig.work_dir`) where backups are kept."""
70
71 CSR_DIR = "csr"
72 """See `.IConfig.csr_dir`."""
73
74 IN_PROGRESS_DIR = "IN_PROGRESS"
75 """Directory used before a permanent checkpoint is finalized (relative to
76 `IConfig.work_dir`)."""
77
78 KEY_DIR = "keys"
79 """Directory (relative to `IConfig.config_dir`) where keys are saved."""
80
81 LIVE_DIR = "live"
82 """Live directory, relative to `IConfig.config_dir`."""
83
84 TEMP_CHECKPOINT_DIR = "temp_checkpoint"
85 """Temporary checkpoint directory (relative to `IConfig.work_dir`)."""
86
87 RENEWAL_CONFIGS_DIR = "renewal"
88 """Renewal configs directory, relative to `IConfig.config_dir`."""
89
90 RENEWER_CONFIG_FILENAME = "renewer.conf"
91 """Renewer config file name (relative to `IConfig.config_dir`)."""
92
[end of certbot/constants.py]
[start of certbot/plugins/disco.py]
1 """Utilities for plugins discovery and selection."""
2 import collections
3 import logging
4 import pkg_resources
5
6 import zope.interface
7 import zope.interface.verify
8
9 from certbot import constants
10 from certbot import errors
11 from certbot import interfaces
12
13
14 logger = logging.getLogger(__name__)
15
16
17 class PluginEntryPoint(object):
18 """Plugin entry point."""
19
20 PREFIX_FREE_DISTRIBUTIONS = [
21 "certbot",
22 "certbot-apache",
23 "certbot-nginx",
24 ]
25 """Distributions for which prefix will be omitted."""
26
27 # this object is mutable, don't allow it to be hashed!
28 __hash__ = None
29
30 def __init__(self, entry_point):
31 self.name = self.entry_point_to_plugin_name(entry_point)
32 self.plugin_cls = entry_point.load()
33 self.entry_point = entry_point
34 self._initialized = None
35 self._prepared = None
36
37 @classmethod
38 def entry_point_to_plugin_name(cls, entry_point):
39 """Unique plugin name for an ``entry_point``"""
40 if entry_point.dist.key in cls.PREFIX_FREE_DISTRIBUTIONS:
41 return entry_point.name
42 return entry_point.dist.key + ":" + entry_point.name
43
44 @property
45 def description(self):
46 """Description of the plugin."""
47 return self.plugin_cls.description
48
49 @property
50 def description_with_name(self):
51 """Description with name. Handy for UI."""
52 return "{0} ({1})".format(self.description, self.name)
53
54 @property
55 def hidden(self):
56 """Should this plugin be hidden from UI?"""
57 return getattr(self.plugin_cls, "hidden", False)
58
59 def ifaces(self, *ifaces_groups):
60 """Does plugin implements specified interface groups?"""
61 return not ifaces_groups or any(
62 all(iface.implementedBy(self.plugin_cls)
63 for iface in ifaces)
64 for ifaces in ifaces_groups)
65
66 @property
67 def initialized(self):
68 """Has the plugin been initialized already?"""
69 return self._initialized is not None
70
71 def init(self, config=None):
72 """Memoized plugin inititialization."""
73 if not self.initialized:
74 self.entry_point.require() # fetch extras!
75 self._initialized = self.plugin_cls(config, self.name)
76 return self._initialized
77
78 def verify(self, ifaces):
79 """Verify that the plugin conforms to the specified interfaces."""
80 assert self.initialized
81 for iface in ifaces: # zope.interface.providedBy(plugin)
82 try:
83 zope.interface.verify.verifyObject(iface, self.init())
84 except zope.interface.exceptions.BrokenImplementation as error:
85 if iface.implementedBy(self.plugin_cls):
86 logger.debug(
87 "%s implements %s but object does not verify: %s",
88 self.plugin_cls, iface.__name__, error, exc_info=True)
89 return False
90 return True
91
92 @property
93 def prepared(self):
94 """Has the plugin been prepared already?"""
95 if not self.initialized:
96 logger.debug(".prepared called on uninitialized %r", self)
97 return self._prepared is not None
98
99 def prepare(self):
100 """Memoized plugin preparation."""
101 assert self.initialized
102 if self._prepared is None:
103 try:
104 self._initialized.prepare()
105 except errors.MisconfigurationError as error:
106 logger.debug("Misconfigured %r: %s", self, error, exc_info=True)
107 self._prepared = error
108 except errors.NoInstallationError as error:
109 logger.debug(
110 "No installation (%r): %s", self, error, exc_info=True)
111 self._prepared = error
112 except errors.PluginError as error:
113 logger.debug("Other error:(%r): %s", self, error, exc_info=True)
114 self._prepared = error
115 else:
116 self._prepared = True
117 return self._prepared
118
119 @property
120 def misconfigured(self):
121 """Is plugin misconfigured?"""
122 return isinstance(self._prepared, errors.MisconfigurationError)
123
124 @property
125 def problem(self):
126 """Return the Exception raised during plugin setup, or None if all is well"""
127 if isinstance(self._prepared, Exception):
128 return self._prepared
129 return None
130
131 @property
132 def available(self):
133 """Is plugin available, i.e. prepared or misconfigured?"""
134 return self._prepared is True or self.misconfigured
135
136 def __repr__(self):
137 return "PluginEntryPoint#{0}".format(self.name)
138
139 def __str__(self):
140 lines = [
141 "* {0}".format(self.name),
142 "Description: {0}".format(self.plugin_cls.description),
143 "Interfaces: {0}".format(", ".join(
144 iface.__name__ for iface in zope.interface.implementedBy(
145 self.plugin_cls))),
146 "Entry point: {0}".format(self.entry_point),
147 ]
148
149 if self.initialized:
150 lines.append("Initialized: {0}".format(self.init()))
151 if self.prepared:
152 lines.append("Prep: {0}".format(self.prepare()))
153
154 return "\n".join(lines)
155
156
157 class PluginsRegistry(collections.Mapping):
158 """Plugins registry."""
159
160 def __init__(self, plugins):
161 self._plugins = plugins
162
163 @classmethod
164 def find_all(cls):
165 """Find plugins using setuptools entry points."""
166 plugins = {}
167 for entry_point in pkg_resources.iter_entry_points(
168 constants.SETUPTOOLS_PLUGINS_ENTRY_POINT):
169 plugin_ep = PluginEntryPoint(entry_point)
170 assert plugin_ep.name not in plugins, (
171 "PREFIX_FREE_DISTRIBUTIONS messed up")
172 # providedBy | pylint: disable=no-member
173 if interfaces.IPluginFactory.providedBy(plugin_ep.plugin_cls):
174 plugins[plugin_ep.name] = plugin_ep
175 else: # pragma: no cover
176 logger.warning(
177 "%r does not provide IPluginFactory, skipping", plugin_ep)
178 return cls(plugins)
179
180 def __getitem__(self, name):
181 return self._plugins[name]
182
183 def __iter__(self):
184 return iter(self._plugins)
185
186 def __len__(self):
187 return len(self._plugins)
188
189 def init(self, config):
190 """Initialize all plugins in the registry."""
191 return [plugin_ep.init(config) for plugin_ep
192 in self._plugins.itervalues()]
193
194 def filter(self, pred):
195 """Filter plugins based on predicate."""
196 return type(self)(dict((name, plugin_ep) for name, plugin_ep
197 in self._plugins.iteritems() if pred(plugin_ep)))
198
199 def visible(self):
200 """Filter plugins based on visibility."""
201 return self.filter(lambda plugin_ep: not plugin_ep.hidden)
202
203 def ifaces(self, *ifaces_groups):
204 """Filter plugins based on interfaces."""
205 # pylint: disable=star-args
206 return self.filter(lambda p_ep: p_ep.ifaces(*ifaces_groups))
207
208 def verify(self, ifaces):
209 """Filter plugins based on verification."""
210 return self.filter(lambda p_ep: p_ep.verify(ifaces))
211
212 def prepare(self):
213 """Prepare all plugins in the registry."""
214 return [plugin_ep.prepare() for plugin_ep in self._plugins.itervalues()]
215
216 def available(self):
217 """Filter plugins based on availability."""
218 return self.filter(lambda p_ep: p_ep.available)
219 # succefully prepared + misconfigured
220
221 def find_init(self, plugin):
222 """Find an initialized plugin.
223
224 This is particularly useful for finding a name for the plugin
225 (although `.IPluginFactory.__call__` takes ``name`` as one of
226 the arguments, ``IPlugin.name`` is not part of the interface)::
227
228 # plugin is an instance providing IPlugin, initialized
229 # somewhere else in the code
230 plugin_registry.find_init(plugin).name
231
232 Returns ``None`` if ``plugin`` is not found in the registry.
233
234 """
235 # use list instead of set because PluginEntryPoint is not hashable
236 candidates = [plugin_ep for plugin_ep in self._plugins.itervalues()
237 if plugin_ep.initialized and plugin_ep.init() is plugin]
238 assert len(candidates) <= 1
239 if candidates:
240 return candidates[0]
241 else:
242 return None
243
244 def __repr__(self):
245 return "{0}({1})".format(
246 self.__class__.__name__, ','.join(
247 repr(p_ep) for p_ep in self._plugins.itervalues()))
248
249 def __str__(self):
250 if not self._plugins:
251 return "No plugins"
252 return "\n\n".join(str(p_ep) for p_ep in self._plugins.itervalues())
253
[end of certbot/plugins/disco.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/certbot/constants.py b/certbot/constants.py
--- a/certbot/constants.py
+++ b/certbot/constants.py
@@ -8,6 +8,9 @@
SETUPTOOLS_PLUGINS_ENTRY_POINT = "certbot.plugins"
"""Setuptools entry point group name for plugins."""
+OLD_SETUPTOOLS_PLUGINS_ENTRY_POINT = "letsencrypt.plugins"
+"""Plugins Setuptools entry point before rename."""
+
CLI_DEFAULTS = dict(
config_files=[
"/etc/letsencrypt/cli.ini",
diff --git a/certbot/plugins/disco.py b/certbot/plugins/disco.py
--- a/certbot/plugins/disco.py
+++ b/certbot/plugins/disco.py
@@ -1,5 +1,6 @@
"""Utilities for plugins discovery and selection."""
import collections
+import itertools
import logging
import pkg_resources
@@ -164,8 +165,12 @@
def find_all(cls):
"""Find plugins using setuptools entry points."""
plugins = {}
- for entry_point in pkg_resources.iter_entry_points(
- constants.SETUPTOOLS_PLUGINS_ENTRY_POINT):
+ entry_points = itertools.chain(
+ pkg_resources.iter_entry_points(
+ constants.SETUPTOOLS_PLUGINS_ENTRY_POINT),
+ pkg_resources.iter_entry_points(
+ constants.OLD_SETUPTOOLS_PLUGINS_ENTRY_POINT),)
+ for entry_point in entry_points:
plugin_ep = PluginEntryPoint(entry_point)
assert plugin_ep.name not in plugins, (
"PREFIX_FREE_DISTRIBUTIONS messed up")
|
{"golden_diff": "diff --git a/certbot/constants.py b/certbot/constants.py\n--- a/certbot/constants.py\n+++ b/certbot/constants.py\n@@ -8,6 +8,9 @@\n SETUPTOOLS_PLUGINS_ENTRY_POINT = \"certbot.plugins\"\n \"\"\"Setuptools entry point group name for plugins.\"\"\"\n \n+OLD_SETUPTOOLS_PLUGINS_ENTRY_POINT = \"letsencrypt.plugins\"\n+\"\"\"Plugins Setuptools entry point before rename.\"\"\"\n+\n CLI_DEFAULTS = dict(\n config_files=[\n \"/etc/letsencrypt/cli.ini\",\ndiff --git a/certbot/plugins/disco.py b/certbot/plugins/disco.py\n--- a/certbot/plugins/disco.py\n+++ b/certbot/plugins/disco.py\n@@ -1,5 +1,6 @@\n \"\"\"Utilities for plugins discovery and selection.\"\"\"\n import collections\n+import itertools\n import logging\n import pkg_resources\n \n@@ -164,8 +165,12 @@\n def find_all(cls):\n \"\"\"Find plugins using setuptools entry points.\"\"\"\n plugins = {}\n- for entry_point in pkg_resources.iter_entry_points(\n- constants.SETUPTOOLS_PLUGINS_ENTRY_POINT):\n+ entry_points = itertools.chain(\n+ pkg_resources.iter_entry_points(\n+ constants.SETUPTOOLS_PLUGINS_ENTRY_POINT),\n+ pkg_resources.iter_entry_points(\n+ constants.OLD_SETUPTOOLS_PLUGINS_ENTRY_POINT),)\n+ for entry_point in entry_points:\n plugin_ep = PluginEntryPoint(entry_point)\n assert plugin_ep.name not in plugins, (\n \"PREFIX_FREE_DISTRIBUTIONS messed up\")\n", "issue": "Ensure rename doesn't break third party plugins\nThere are a lot of third party plugins for Let's Encrypt (listed here: https://letsencrypt.readthedocs.org/en/latest/using.html#plugins)\nIt's possible that they reference things in such a way that changing the name or package of LE will break their compatibility with the client - we should ensure that they are unbroken\n\n", "before_files": [{"content": "\"\"\"Certbot constants.\"\"\"\nimport os\nimport logging\n\nfrom acme import challenges\n\n\nSETUPTOOLS_PLUGINS_ENTRY_POINT = \"certbot.plugins\"\n\"\"\"Setuptools entry point group name for plugins.\"\"\"\n\nCLI_DEFAULTS = dict(\n config_files=[\n \"/etc/letsencrypt/cli.ini\",\n # http://freedesktop.org/wiki/Software/xdg-user-dirs/\n os.path.join(os.environ.get(\"XDG_CONFIG_HOME\", \"~/.config\"),\n \"letsencrypt\", \"cli.ini\"),\n ],\n verbose_count=-(logging.WARNING / 10),\n server=\"https://acme-v01.api.letsencrypt.org/directory\",\n rsa_key_size=2048,\n rollback_checkpoints=1,\n config_dir=\"/etc/letsencrypt\",\n work_dir=\"/var/lib/letsencrypt\",\n logs_dir=\"/var/log/letsencrypt\",\n no_verify_ssl=False,\n http01_port=challenges.HTTP01Response.PORT,\n tls_sni_01_port=challenges.TLSSNI01Response.PORT,\n\n auth_cert_path=\"./cert.pem\",\n auth_chain_path=\"./chain.pem\",\n strict_permissions=False,\n)\nSTAGING_URI = \"https://acme-staging.api.letsencrypt.org/directory\"\n\n\"\"\"Defaults for CLI flags and `.IConfig` attributes.\"\"\"\n\nRENEWER_DEFAULTS = dict(\n renewer_enabled=\"yes\",\n renew_before_expiry=\"30 days\",\n # This value should ensure that there is never a deployment delay by\n # default.\n deploy_before_expiry=\"99 years\",\n)\n\"\"\"Defaults for renewer script.\"\"\"\n\n\nENHANCEMENTS = [\"redirect\", \"http-header\", \"ocsp-stapling\", \"spdy\"]\n\"\"\"List of possible :class:`certbot.interfaces.IInstaller`\nenhancements.\n\nList of expected options parameters:\n- redirect: None\n- http-header: TODO\n- ocsp-stapling: TODO\n- spdy: TODO\n\n\"\"\"\n\nARCHIVE_DIR = \"archive\"\n\"\"\"Archive directory, relative to `IConfig.config_dir`.\"\"\"\n\nCONFIG_DIRS_MODE = 0o755\n\"\"\"Directory mode for ``.IConfig.config_dir`` et al.\"\"\"\n\nACCOUNTS_DIR = \"accounts\"\n\"\"\"Directory where all accounts are saved.\"\"\"\n\nBACKUP_DIR = \"backups\"\n\"\"\"Directory (relative to `IConfig.work_dir`) where backups are kept.\"\"\"\n\nCSR_DIR = \"csr\"\n\"\"\"See `.IConfig.csr_dir`.\"\"\"\n\nIN_PROGRESS_DIR = \"IN_PROGRESS\"\n\"\"\"Directory used before a permanent checkpoint is finalized (relative to\n`IConfig.work_dir`).\"\"\"\n\nKEY_DIR = \"keys\"\n\"\"\"Directory (relative to `IConfig.config_dir`) where keys are saved.\"\"\"\n\nLIVE_DIR = \"live\"\n\"\"\"Live directory, relative to `IConfig.config_dir`.\"\"\"\n\nTEMP_CHECKPOINT_DIR = \"temp_checkpoint\"\n\"\"\"Temporary checkpoint directory (relative to `IConfig.work_dir`).\"\"\"\n\nRENEWAL_CONFIGS_DIR = \"renewal\"\n\"\"\"Renewal configs directory, relative to `IConfig.config_dir`.\"\"\"\n\nRENEWER_CONFIG_FILENAME = \"renewer.conf\"\n\"\"\"Renewer config file name (relative to `IConfig.config_dir`).\"\"\"\n", "path": "certbot/constants.py"}, {"content": "\"\"\"Utilities for plugins discovery and selection.\"\"\"\nimport collections\nimport logging\nimport pkg_resources\n\nimport zope.interface\nimport zope.interface.verify\n\nfrom certbot import constants\nfrom certbot import errors\nfrom certbot import interfaces\n\n\nlogger = logging.getLogger(__name__)\n\n\nclass PluginEntryPoint(object):\n \"\"\"Plugin entry point.\"\"\"\n\n PREFIX_FREE_DISTRIBUTIONS = [\n \"certbot\",\n \"certbot-apache\",\n \"certbot-nginx\",\n ]\n \"\"\"Distributions for which prefix will be omitted.\"\"\"\n\n # this object is mutable, don't allow it to be hashed!\n __hash__ = None\n\n def __init__(self, entry_point):\n self.name = self.entry_point_to_plugin_name(entry_point)\n self.plugin_cls = entry_point.load()\n self.entry_point = entry_point\n self._initialized = None\n self._prepared = None\n\n @classmethod\n def entry_point_to_plugin_name(cls, entry_point):\n \"\"\"Unique plugin name for an ``entry_point``\"\"\"\n if entry_point.dist.key in cls.PREFIX_FREE_DISTRIBUTIONS:\n return entry_point.name\n return entry_point.dist.key + \":\" + entry_point.name\n\n @property\n def description(self):\n \"\"\"Description of the plugin.\"\"\"\n return self.plugin_cls.description\n\n @property\n def description_with_name(self):\n \"\"\"Description with name. Handy for UI.\"\"\"\n return \"{0} ({1})\".format(self.description, self.name)\n\n @property\n def hidden(self):\n \"\"\"Should this plugin be hidden from UI?\"\"\"\n return getattr(self.plugin_cls, \"hidden\", False)\n\n def ifaces(self, *ifaces_groups):\n \"\"\"Does plugin implements specified interface groups?\"\"\"\n return not ifaces_groups or any(\n all(iface.implementedBy(self.plugin_cls)\n for iface in ifaces)\n for ifaces in ifaces_groups)\n\n @property\n def initialized(self):\n \"\"\"Has the plugin been initialized already?\"\"\"\n return self._initialized is not None\n\n def init(self, config=None):\n \"\"\"Memoized plugin inititialization.\"\"\"\n if not self.initialized:\n self.entry_point.require() # fetch extras!\n self._initialized = self.plugin_cls(config, self.name)\n return self._initialized\n\n def verify(self, ifaces):\n \"\"\"Verify that the plugin conforms to the specified interfaces.\"\"\"\n assert self.initialized\n for iface in ifaces: # zope.interface.providedBy(plugin)\n try:\n zope.interface.verify.verifyObject(iface, self.init())\n except zope.interface.exceptions.BrokenImplementation as error:\n if iface.implementedBy(self.plugin_cls):\n logger.debug(\n \"%s implements %s but object does not verify: %s\",\n self.plugin_cls, iface.__name__, error, exc_info=True)\n return False\n return True\n\n @property\n def prepared(self):\n \"\"\"Has the plugin been prepared already?\"\"\"\n if not self.initialized:\n logger.debug(\".prepared called on uninitialized %r\", self)\n return self._prepared is not None\n\n def prepare(self):\n \"\"\"Memoized plugin preparation.\"\"\"\n assert self.initialized\n if self._prepared is None:\n try:\n self._initialized.prepare()\n except errors.MisconfigurationError as error:\n logger.debug(\"Misconfigured %r: %s\", self, error, exc_info=True)\n self._prepared = error\n except errors.NoInstallationError as error:\n logger.debug(\n \"No installation (%r): %s\", self, error, exc_info=True)\n self._prepared = error\n except errors.PluginError as error:\n logger.debug(\"Other error:(%r): %s\", self, error, exc_info=True)\n self._prepared = error\n else:\n self._prepared = True\n return self._prepared\n\n @property\n def misconfigured(self):\n \"\"\"Is plugin misconfigured?\"\"\"\n return isinstance(self._prepared, errors.MisconfigurationError)\n\n @property\n def problem(self):\n \"\"\"Return the Exception raised during plugin setup, or None if all is well\"\"\"\n if isinstance(self._prepared, Exception):\n return self._prepared\n return None\n\n @property\n def available(self):\n \"\"\"Is plugin available, i.e. prepared or misconfigured?\"\"\"\n return self._prepared is True or self.misconfigured\n\n def __repr__(self):\n return \"PluginEntryPoint#{0}\".format(self.name)\n\n def __str__(self):\n lines = [\n \"* {0}\".format(self.name),\n \"Description: {0}\".format(self.plugin_cls.description),\n \"Interfaces: {0}\".format(\", \".join(\n iface.__name__ for iface in zope.interface.implementedBy(\n self.plugin_cls))),\n \"Entry point: {0}\".format(self.entry_point),\n ]\n\n if self.initialized:\n lines.append(\"Initialized: {0}\".format(self.init()))\n if self.prepared:\n lines.append(\"Prep: {0}\".format(self.prepare()))\n\n return \"\\n\".join(lines)\n\n\nclass PluginsRegistry(collections.Mapping):\n \"\"\"Plugins registry.\"\"\"\n\n def __init__(self, plugins):\n self._plugins = plugins\n\n @classmethod\n def find_all(cls):\n \"\"\"Find plugins using setuptools entry points.\"\"\"\n plugins = {}\n for entry_point in pkg_resources.iter_entry_points(\n constants.SETUPTOOLS_PLUGINS_ENTRY_POINT):\n plugin_ep = PluginEntryPoint(entry_point)\n assert plugin_ep.name not in plugins, (\n \"PREFIX_FREE_DISTRIBUTIONS messed up\")\n # providedBy | pylint: disable=no-member\n if interfaces.IPluginFactory.providedBy(plugin_ep.plugin_cls):\n plugins[plugin_ep.name] = plugin_ep\n else: # pragma: no cover\n logger.warning(\n \"%r does not provide IPluginFactory, skipping\", plugin_ep)\n return cls(plugins)\n\n def __getitem__(self, name):\n return self._plugins[name]\n\n def __iter__(self):\n return iter(self._plugins)\n\n def __len__(self):\n return len(self._plugins)\n\n def init(self, config):\n \"\"\"Initialize all plugins in the registry.\"\"\"\n return [plugin_ep.init(config) for plugin_ep\n in self._plugins.itervalues()]\n\n def filter(self, pred):\n \"\"\"Filter plugins based on predicate.\"\"\"\n return type(self)(dict((name, plugin_ep) for name, plugin_ep\n in self._plugins.iteritems() if pred(plugin_ep)))\n\n def visible(self):\n \"\"\"Filter plugins based on visibility.\"\"\"\n return self.filter(lambda plugin_ep: not plugin_ep.hidden)\n\n def ifaces(self, *ifaces_groups):\n \"\"\"Filter plugins based on interfaces.\"\"\"\n # pylint: disable=star-args\n return self.filter(lambda p_ep: p_ep.ifaces(*ifaces_groups))\n\n def verify(self, ifaces):\n \"\"\"Filter plugins based on verification.\"\"\"\n return self.filter(lambda p_ep: p_ep.verify(ifaces))\n\n def prepare(self):\n \"\"\"Prepare all plugins in the registry.\"\"\"\n return [plugin_ep.prepare() for plugin_ep in self._plugins.itervalues()]\n\n def available(self):\n \"\"\"Filter plugins based on availability.\"\"\"\n return self.filter(lambda p_ep: p_ep.available)\n # succefully prepared + misconfigured\n\n def find_init(self, plugin):\n \"\"\"Find an initialized plugin.\n\n This is particularly useful for finding a name for the plugin\n (although `.IPluginFactory.__call__` takes ``name`` as one of\n the arguments, ``IPlugin.name`` is not part of the interface)::\n\n # plugin is an instance providing IPlugin, initialized\n # somewhere else in the code\n plugin_registry.find_init(plugin).name\n\n Returns ``None`` if ``plugin`` is not found in the registry.\n\n \"\"\"\n # use list instead of set because PluginEntryPoint is not hashable\n candidates = [plugin_ep for plugin_ep in self._plugins.itervalues()\n if plugin_ep.initialized and plugin_ep.init() is plugin]\n assert len(candidates) <= 1\n if candidates:\n return candidates[0]\n else:\n return None\n\n def __repr__(self):\n return \"{0}({1})\".format(\n self.__class__.__name__, ','.join(\n repr(p_ep) for p_ep in self._plugins.itervalues()))\n\n def __str__(self):\n if not self._plugins:\n return \"No plugins\"\n return \"\\n\\n\".join(str(p_ep) for p_ep in self._plugins.itervalues())\n", "path": "certbot/plugins/disco.py"}]}
| 3,986 | 337 |
gh_patches_debug_15866
|
rasdani/github-patches
|
git_diff
|
tornadoweb__tornado-2653
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
tornado.platform.twisted disappeared but did not explain its replacement
Many years' worth of documentation explains that folks can do
```
from tornado.platform.twisted import install
reactor = install()
```
I can see that in https://github.com/tornadoweb/tornado/commit/004de9c301cc4c2dae4d8f2507af1851d0c9763a#diff-77b5a8a33248ef0bcafbc1bb71e9f013 Twisted integration was removed, since we can all depend on the stdlib loop APIs. This is great, but it also breaks a bunch of Jupyter notebooks, tutorials, etc.
Could you be convinced to replace all those sprawling APIs with something like this:
```python3
def install():
from twisted.internet.asyncioreactor import install
install()
from twisted.internet import reactor
reactor.startRunning()
return reactor
```
possibly with a `warnings.warn` explaining that users could just call these APIs directly, if that's the desired end-state?
</issue>
<code>
[start of tornado/platform/twisted.py]
1 # Licensed under the Apache License, Version 2.0 (the "License"); you may
2 # not use this file except in compliance with the License. You may obtain
3 # a copy of the License at
4 #
5 # http://www.apache.org/licenses/LICENSE-2.0
6 #
7 # Unless required by applicable law or agreed to in writing, software
8 # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
9 # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
10 # License for the specific language governing permissions and limitations
11 # under the License.
12 """Bridges between the Twisted package and Tornado.
13 """
14
15 import socket
16 import sys
17
18 import twisted.internet.abstract # type: ignore
19 import twisted.internet.asyncioreactor # type: ignore
20 from twisted.internet.defer import Deferred # type: ignore
21 from twisted.python import failure # type: ignore
22 import twisted.names.cache # type: ignore
23 import twisted.names.client # type: ignore
24 import twisted.names.hosts # type: ignore
25 import twisted.names.resolve # type: ignore
26
27
28 from tornado.concurrent import Future, future_set_exc_info
29 from tornado.escape import utf8
30 from tornado import gen
31 from tornado.netutil import Resolver
32
33 import typing
34
35 if typing.TYPE_CHECKING:
36 from typing import Generator, Any, List, Tuple # noqa: F401
37
38
39 class TwistedResolver(Resolver):
40 """Twisted-based asynchronous resolver.
41
42 This is a non-blocking and non-threaded resolver. It is
43 recommended only when threads cannot be used, since it has
44 limitations compared to the standard ``getaddrinfo``-based
45 `~tornado.netutil.Resolver` and
46 `~tornado.netutil.DefaultExecutorResolver`. Specifically, it returns at
47 most one result, and arguments other than ``host`` and ``family``
48 are ignored. It may fail to resolve when ``family`` is not
49 ``socket.AF_UNSPEC``.
50
51 Requires Twisted 12.1 or newer.
52
53 .. versionchanged:: 5.0
54 The ``io_loop`` argument (deprecated since version 4.1) has been removed.
55 """
56
57 def initialize(self) -> None:
58 # partial copy of twisted.names.client.createResolver, which doesn't
59 # allow for a reactor to be passed in.
60 self.reactor = twisted.internet.asyncioreactor.AsyncioSelectorReactor()
61
62 host_resolver = twisted.names.hosts.Resolver("/etc/hosts")
63 cache_resolver = twisted.names.cache.CacheResolver(reactor=self.reactor)
64 real_resolver = twisted.names.client.Resolver(
65 "/etc/resolv.conf", reactor=self.reactor
66 )
67 self.resolver = twisted.names.resolve.ResolverChain(
68 [host_resolver, cache_resolver, real_resolver]
69 )
70
71 @gen.coroutine
72 def resolve(
73 self, host: str, port: int, family: int = 0
74 ) -> "Generator[Any, Any, List[Tuple[int, Any]]]":
75 # getHostByName doesn't accept IP addresses, so if the input
76 # looks like an IP address just return it immediately.
77 if twisted.internet.abstract.isIPAddress(host):
78 resolved = host
79 resolved_family = socket.AF_INET
80 elif twisted.internet.abstract.isIPv6Address(host):
81 resolved = host
82 resolved_family = socket.AF_INET6
83 else:
84 deferred = self.resolver.getHostByName(utf8(host))
85 fut = Future() # type: Future[Any]
86 deferred.addBoth(fut.set_result)
87 resolved = yield fut
88 if isinstance(resolved, failure.Failure):
89 try:
90 resolved.raiseException()
91 except twisted.names.error.DomainError as e:
92 raise IOError(e)
93 elif twisted.internet.abstract.isIPAddress(resolved):
94 resolved_family = socket.AF_INET
95 elif twisted.internet.abstract.isIPv6Address(resolved):
96 resolved_family = socket.AF_INET6
97 else:
98 resolved_family = socket.AF_UNSPEC
99 if family != socket.AF_UNSPEC and family != resolved_family:
100 raise Exception(
101 "Requested socket family %d but got %d" % (family, resolved_family)
102 )
103 result = [(typing.cast(int, resolved_family), (resolved, port))]
104 return result
105
106
107 if hasattr(gen.convert_yielded, "register"):
108
109 @gen.convert_yielded.register(Deferred) # type: ignore
110 def _(d: Deferred) -> Future:
111 f = Future() # type: Future[Any]
112
113 def errback(failure: failure.Failure) -> None:
114 try:
115 failure.raiseException()
116 # Should never happen, but just in case
117 raise Exception("errback called without error")
118 except:
119 future_set_exc_info(f, sys.exc_info())
120
121 d.addCallbacks(f.set_result, errback)
122 return f
123
[end of tornado/platform/twisted.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/tornado/platform/twisted.py b/tornado/platform/twisted.py
--- a/tornado/platform/twisted.py
+++ b/tornado/platform/twisted.py
@@ -104,6 +104,30 @@
return result
+def install() -> None:
+ """Install ``AsyncioSelectorReactor`` as the default Twisted reactor.
+
+ .. deprecated:: 5.1
+
+ This function is provided for backwards compatibility; code
+ that does not require compatibility with older versions of
+ Tornado should use
+ ``twisted.internet.asyncioreactor.install()`` directly.
+
+ .. versionchanged:: 6.0.3
+
+ In Tornado 5.x and before, this function installed a reactor
+ based on the Tornado ``IOLoop``. When that reactor
+ implementation was removed in Tornado 6.0.0, this function was
+ removed as well. It was restored in Tornado 6.0.3 using the
+ ``asyncio`` reactor instead.
+
+ """
+ from twisted.internet.asyncioreactor import install
+
+ install()
+
+
if hasattr(gen.convert_yielded, "register"):
@gen.convert_yielded.register(Deferred) # type: ignore
|
{"golden_diff": "diff --git a/tornado/platform/twisted.py b/tornado/platform/twisted.py\n--- a/tornado/platform/twisted.py\n+++ b/tornado/platform/twisted.py\n@@ -104,6 +104,30 @@\n return result\n \n \n+def install() -> None:\n+ \"\"\"Install ``AsyncioSelectorReactor`` as the default Twisted reactor.\n+\n+ .. deprecated:: 5.1\n+\n+ This function is provided for backwards compatibility; code\n+ that does not require compatibility with older versions of\n+ Tornado should use\n+ ``twisted.internet.asyncioreactor.install()`` directly.\n+\n+ .. versionchanged:: 6.0.3\n+\n+ In Tornado 5.x and before, this function installed a reactor\n+ based on the Tornado ``IOLoop``. When that reactor\n+ implementation was removed in Tornado 6.0.0, this function was\n+ removed as well. It was restored in Tornado 6.0.3 using the\n+ ``asyncio`` reactor instead.\n+\n+ \"\"\"\n+ from twisted.internet.asyncioreactor import install\n+\n+ install()\n+\n+\n if hasattr(gen.convert_yielded, \"register\"):\n \n @gen.convert_yielded.register(Deferred) # type: ignore\n", "issue": "tornado.platform.twisted disappeared but did not explain its replacement\nMany years' worth of documentation explains that folks can do\r\n\r\n```\r\nfrom tornado.platform.twisted import install\r\nreactor = install()\r\n```\r\n\r\nI can see that in https://github.com/tornadoweb/tornado/commit/004de9c301cc4c2dae4d8f2507af1851d0c9763a#diff-77b5a8a33248ef0bcafbc1bb71e9f013 Twisted integration was removed, since we can all depend on the stdlib loop APIs. This is great, but it also breaks a bunch of Jupyter notebooks, tutorials, etc.\r\n\r\nCould you be convinced to replace all those sprawling APIs with something like this:\r\n\r\n```python3\r\ndef install():\r\n from twisted.internet.asyncioreactor import install\r\n install()\r\n from twisted.internet import reactor\r\n reactor.startRunning()\r\n return reactor\r\n```\r\n\r\npossibly with a `warnings.warn` explaining that users could just call these APIs directly, if that's the desired end-state?\n", "before_files": [{"content": "# Licensed under the Apache License, Version 2.0 (the \"License\"); you may\n# not use this file except in compliance with the License. You may obtain\n# a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS, WITHOUT\n# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the\n# License for the specific language governing permissions and limitations\n# under the License.\n\"\"\"Bridges between the Twisted package and Tornado.\n\"\"\"\n\nimport socket\nimport sys\n\nimport twisted.internet.abstract # type: ignore\nimport twisted.internet.asyncioreactor # type: ignore\nfrom twisted.internet.defer import Deferred # type: ignore\nfrom twisted.python import failure # type: ignore\nimport twisted.names.cache # type: ignore\nimport twisted.names.client # type: ignore\nimport twisted.names.hosts # type: ignore\nimport twisted.names.resolve # type: ignore\n\n\nfrom tornado.concurrent import Future, future_set_exc_info\nfrom tornado.escape import utf8\nfrom tornado import gen\nfrom tornado.netutil import Resolver\n\nimport typing\n\nif typing.TYPE_CHECKING:\n from typing import Generator, Any, List, Tuple # noqa: F401\n\n\nclass TwistedResolver(Resolver):\n \"\"\"Twisted-based asynchronous resolver.\n\n This is a non-blocking and non-threaded resolver. It is\n recommended only when threads cannot be used, since it has\n limitations compared to the standard ``getaddrinfo``-based\n `~tornado.netutil.Resolver` and\n `~tornado.netutil.DefaultExecutorResolver`. Specifically, it returns at\n most one result, and arguments other than ``host`` and ``family``\n are ignored. It may fail to resolve when ``family`` is not\n ``socket.AF_UNSPEC``.\n\n Requires Twisted 12.1 or newer.\n\n .. versionchanged:: 5.0\n The ``io_loop`` argument (deprecated since version 4.1) has been removed.\n \"\"\"\n\n def initialize(self) -> None:\n # partial copy of twisted.names.client.createResolver, which doesn't\n # allow for a reactor to be passed in.\n self.reactor = twisted.internet.asyncioreactor.AsyncioSelectorReactor()\n\n host_resolver = twisted.names.hosts.Resolver(\"/etc/hosts\")\n cache_resolver = twisted.names.cache.CacheResolver(reactor=self.reactor)\n real_resolver = twisted.names.client.Resolver(\n \"/etc/resolv.conf\", reactor=self.reactor\n )\n self.resolver = twisted.names.resolve.ResolverChain(\n [host_resolver, cache_resolver, real_resolver]\n )\n\n @gen.coroutine\n def resolve(\n self, host: str, port: int, family: int = 0\n ) -> \"Generator[Any, Any, List[Tuple[int, Any]]]\":\n # getHostByName doesn't accept IP addresses, so if the input\n # looks like an IP address just return it immediately.\n if twisted.internet.abstract.isIPAddress(host):\n resolved = host\n resolved_family = socket.AF_INET\n elif twisted.internet.abstract.isIPv6Address(host):\n resolved = host\n resolved_family = socket.AF_INET6\n else:\n deferred = self.resolver.getHostByName(utf8(host))\n fut = Future() # type: Future[Any]\n deferred.addBoth(fut.set_result)\n resolved = yield fut\n if isinstance(resolved, failure.Failure):\n try:\n resolved.raiseException()\n except twisted.names.error.DomainError as e:\n raise IOError(e)\n elif twisted.internet.abstract.isIPAddress(resolved):\n resolved_family = socket.AF_INET\n elif twisted.internet.abstract.isIPv6Address(resolved):\n resolved_family = socket.AF_INET6\n else:\n resolved_family = socket.AF_UNSPEC\n if family != socket.AF_UNSPEC and family != resolved_family:\n raise Exception(\n \"Requested socket family %d but got %d\" % (family, resolved_family)\n )\n result = [(typing.cast(int, resolved_family), (resolved, port))]\n return result\n\n\nif hasattr(gen.convert_yielded, \"register\"):\n\n @gen.convert_yielded.register(Deferred) # type: ignore\n def _(d: Deferred) -> Future:\n f = Future() # type: Future[Any]\n\n def errback(failure: failure.Failure) -> None:\n try:\n failure.raiseException()\n # Should never happen, but just in case\n raise Exception(\"errback called without error\")\n except:\n future_set_exc_info(f, sys.exc_info())\n\n d.addCallbacks(f.set_result, errback)\n return f\n", "path": "tornado/platform/twisted.py"}]}
| 2,073 | 289 |
gh_patches_debug_6078
|
rasdani/github-patches
|
git_diff
|
nerfstudio-project__nerfstudio-1094
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Training crashes if all cameras are deleted while in "Render View"
Steps to replicate:
1. Start train and open viewer
2. Add camera to scene
3. Switch to "RENDER VIEW"
4. Delete camera
Resulting error:
```
File "/projects/pyrad/scripts/train.py", line 234, in main
launch(
File "/projects/pyrad/scripts/train.py", line 173, in launch
main_func(local_rank=0, world_size=world_size, config=config)
File "/projects/pyrad/scripts/train.py", line 88, in train_loop
trainer.train()
File "/projects/pyrad/nerfstudio/engine/trainer.py", line 166, in train
self._update_viewer_state(step)
File "/projects/pyrad/nerfstudio/utils/decorators.py", line 58, in wrapper
ret = func(self, *args, **kwargs)
File "/projects/pyrad/nerfstudio/engine/trainer.py", line 229, in _update_viewer_state
self.viewer_state.update_scene(self, step, self.pipeline.model, num_rays_per_batch)
File "/projects/pyrad/nerfstudio/utils/decorators.py", line 82, in wrapper
ret = func(*args, **kwargs)
File "/projects/pyrad/nerfstudio/viewer/server/viewer_utils.py", line 398, in update_scene
self._render_image_in_viewer(camera_object, graph, is_training)
File "/projects/pyrad/nerfstudio/utils/decorators.py", line 82, in wrapper
ret = func(*args, **kwargs)
File "/projects/pyrad/nerfstudio/utils/profiler.py", line 43, in wrapper
ret = func(*args, **kwargs)
File "/projects/pyrad/nerfstudio/viewer/server/viewer_utils.py", line 745, in _render_image_in_viewer
intrinsics_matrix, camera_to_world_h = get_intrinsics_matrix_and_camera_to_world_h(
File "/projects/pyrad/nerfstudio/viewer/server/utils.py", line 77, in get_intrinsics_matrix_and_camera_to_world_h
focal_length = three_js_perspective_camera_focal_length(fov, image_height)
File "/projects/pyrad/nerfstudio/viewer/server/utils.py", line 58, in three_js_perspective_camera_focal_length
focal_length = pp_h / np.tan(fov * (np.pi / 180.0) / 2.0)
TypeError: unsupported operand type(s) for *: 'NoneType' and 'float'
```
</issue>
<code>
[start of nerfstudio/viewer/server/utils.py]
1 # Copyright 2022 The Nerfstudio Team. All rights reserved.
2 #
3 # Licensed under the Apache License, Version 2.0 (the "License");
4 # you may not use this file except in compliance with the License.
5 # You may obtain a copy of the License at
6 #
7 # http://www.apache.org/licenses/LICENSE-2.0
8 #
9 # Unless required by applicable law or agreed to in writing, software
10 # distributed under the License is distributed on an "AS IS" BASIS,
11 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12 # See the License for the specific language governing permissions and
13 # limitations under the License.
14
15 """Generic utility functions
16 """
17
18 import sys
19 from typing import Any, Callable, Dict, List, Optional, Tuple
20
21 import numpy as np
22 import torch
23 import zmq
24 from aiortc import RTCPeerConnection
25 from aiortc.rtcrtpsender import RTCRtpSender
26
27
28 def get_chunks(
29 lst: List[float], num_chunks: Optional[int] = None, size_of_chunk: Optional[int] = None
30 ) -> List[List[float]]:
31 """Returns list of n elements, constaining a sublist.
32
33 Args:
34 lst: List to be chunked up
35 num_chunks: number of chunks to split list into
36 size_of_chunk: size of each chunk
37 """
38 if num_chunks:
39 assert not size_of_chunk
40 size = len(lst) // num_chunks
41 if size_of_chunk:
42 assert not num_chunks
43 size = size_of_chunk
44 chunks = []
45 for i in range(0, len(lst), size):
46 chunks.append(lst[i : i + size])
47 return chunks
48
49
50 def three_js_perspective_camera_focal_length(fov: float, image_height: int):
51 """Returns the focal length of a three.js perspective camera.
52
53 Args:
54 fov: the field of view of the camera in degrees.
55 image_height: the height of the image in pixels.
56 """
57 pp_h = image_height / 2.0
58 focal_length = pp_h / np.tan(fov * (np.pi / 180.0) / 2.0)
59 return focal_length
60
61
62 def get_intrinsics_matrix_and_camera_to_world_h(
63 camera_object: Dict[str, Any], image_height: int
64 ) -> Tuple[torch.Tensor, torch.Tensor]:
65 """Returns the camera intrinsics matrix and the camera to world homogeneous matrix.
66
67 Args:
68 camera_object: a Camera object.
69 image_size: the size of the image (height, width)
70 """
71 # intrinsics
72 fov = camera_object["fov"]
73 aspect = camera_object["aspect"]
74 image_width = aspect * image_height
75 pp_w = image_width / 2.0
76 pp_h = image_height / 2.0
77 focal_length = three_js_perspective_camera_focal_length(fov, image_height)
78 intrinsics_matrix = torch.tensor([[focal_length, 0, pp_w], [0, focal_length, pp_h], [0, 0, 1]]).float()
79
80 # extrinsics
81 camera_to_world_h = torch.tensor(get_chunks(camera_object["matrix"], size_of_chunk=4)).T.float()
82 camera_to_world_h = torch.stack(
83 [
84 camera_to_world_h[0, :],
85 camera_to_world_h[2, :],
86 camera_to_world_h[1, :],
87 camera_to_world_h[3, :],
88 ],
89 dim=0,
90 )
91
92 return intrinsics_matrix, camera_to_world_h
93
94
95 def find_available_port(func: Callable, default_port: int, max_attempts: int = 1000, **kwargs) -> None:
96 """Finds and attempts to connect to a port
97
98 Args:
99 func: function used on connecting to port
100 default_port: the default port
101 max_attempts: max number of attempts to try connection. Defaults to MAX_ATTEMPTS.
102 """
103 for i in range(max_attempts):
104 port = default_port + i
105 try:
106 return func(port, **kwargs), port
107 except (OSError, zmq.error.ZMQError):
108 print(f"Port: {port:d} in use, trying another...", file=sys.stderr)
109 except Exception as e:
110 print(type(e))
111 raise
112 raise (
113 Exception(f"Could not find an available port in the range: [{default_port:d}, {max_attempts + default_port:d})")
114 )
115
116
117 def force_codec(pc: RTCPeerConnection, sender: RTCRtpSender, forced_codec: str) -> None:
118 """Sets the codec preferences on a connection between sender and reciever
119
120 Args:
121 pc: peer connection point
122 sender: sender that will send to connection point
123 forced_codec: codec to set
124 """
125 kind = forced_codec.split("/")[0]
126 codecs = RTCRtpSender.getCapabilities(kind).codecs
127 transceiver = next(t for t in pc.getTransceivers() if t.sender == sender)
128 transceiver.setCodecPreferences([codec for codec in codecs if codec.mimeType == forced_codec])
129
[end of nerfstudio/viewer/server/utils.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/nerfstudio/viewer/server/utils.py b/nerfstudio/viewer/server/utils.py
--- a/nerfstudio/viewer/server/utils.py
+++ b/nerfstudio/viewer/server/utils.py
@@ -54,6 +54,9 @@
fov: the field of view of the camera in degrees.
image_height: the height of the image in pixels.
"""
+ if fov is None:
+ print("Warning: fov is None, using default value")
+ return 50
pp_h = image_height / 2.0
focal_length = pp_h / np.tan(fov * (np.pi / 180.0) / 2.0)
return focal_length
|
{"golden_diff": "diff --git a/nerfstudio/viewer/server/utils.py b/nerfstudio/viewer/server/utils.py\n--- a/nerfstudio/viewer/server/utils.py\n+++ b/nerfstudio/viewer/server/utils.py\n@@ -54,6 +54,9 @@\n fov: the field of view of the camera in degrees.\n image_height: the height of the image in pixels.\n \"\"\"\n+ if fov is None:\n+ print(\"Warning: fov is None, using default value\")\n+ return 50\n pp_h = image_height / 2.0\n focal_length = pp_h / np.tan(fov * (np.pi / 180.0) / 2.0)\n return focal_length\n", "issue": "Training crashes if all cameras are deleted while in \"Render View\"\nSteps to replicate:\r\n1. Start train and open viewer\r\n2. Add camera to scene\r\n3. Switch to \"RENDER VIEW\"\r\n4. Delete camera\r\n\r\nResulting error:\r\n```\r\n File \"/projects/pyrad/scripts/train.py\", line 234, in main\r\n launch(\r\n File \"/projects/pyrad/scripts/train.py\", line 173, in launch\r\n main_func(local_rank=0, world_size=world_size, config=config)\r\n File \"/projects/pyrad/scripts/train.py\", line 88, in train_loop\r\n trainer.train()\r\n File \"/projects/pyrad/nerfstudio/engine/trainer.py\", line 166, in train\r\n self._update_viewer_state(step)\r\n File \"/projects/pyrad/nerfstudio/utils/decorators.py\", line 58, in wrapper\r\n ret = func(self, *args, **kwargs)\r\n File \"/projects/pyrad/nerfstudio/engine/trainer.py\", line 229, in _update_viewer_state\r\n self.viewer_state.update_scene(self, step, self.pipeline.model, num_rays_per_batch)\r\n File \"/projects/pyrad/nerfstudio/utils/decorators.py\", line 82, in wrapper\r\n ret = func(*args, **kwargs)\r\n File \"/projects/pyrad/nerfstudio/viewer/server/viewer_utils.py\", line 398, in update_scene\r\n self._render_image_in_viewer(camera_object, graph, is_training)\r\n File \"/projects/pyrad/nerfstudio/utils/decorators.py\", line 82, in wrapper\r\n ret = func(*args, **kwargs)\r\n File \"/projects/pyrad/nerfstudio/utils/profiler.py\", line 43, in wrapper\r\n ret = func(*args, **kwargs)\r\n File \"/projects/pyrad/nerfstudio/viewer/server/viewer_utils.py\", line 745, in _render_image_in_viewer\r\n intrinsics_matrix, camera_to_world_h = get_intrinsics_matrix_and_camera_to_world_h(\r\n File \"/projects/pyrad/nerfstudio/viewer/server/utils.py\", line 77, in get_intrinsics_matrix_and_camera_to_world_h\r\n focal_length = three_js_perspective_camera_focal_length(fov, image_height)\r\n File \"/projects/pyrad/nerfstudio/viewer/server/utils.py\", line 58, in three_js_perspective_camera_focal_length\r\n focal_length = pp_h / np.tan(fov * (np.pi / 180.0) / 2.0)\r\nTypeError: unsupported operand type(s) for *: 'NoneType' and 'float'\r\n```\n", "before_files": [{"content": "# Copyright 2022 The Nerfstudio Team. All rights reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\n\"\"\"Generic utility functions\n\"\"\"\n\nimport sys\nfrom typing import Any, Callable, Dict, List, Optional, Tuple\n\nimport numpy as np\nimport torch\nimport zmq\nfrom aiortc import RTCPeerConnection\nfrom aiortc.rtcrtpsender import RTCRtpSender\n\n\ndef get_chunks(\n lst: List[float], num_chunks: Optional[int] = None, size_of_chunk: Optional[int] = None\n) -> List[List[float]]:\n \"\"\"Returns list of n elements, constaining a sublist.\n\n Args:\n lst: List to be chunked up\n num_chunks: number of chunks to split list into\n size_of_chunk: size of each chunk\n \"\"\"\n if num_chunks:\n assert not size_of_chunk\n size = len(lst) // num_chunks\n if size_of_chunk:\n assert not num_chunks\n size = size_of_chunk\n chunks = []\n for i in range(0, len(lst), size):\n chunks.append(lst[i : i + size])\n return chunks\n\n\ndef three_js_perspective_camera_focal_length(fov: float, image_height: int):\n \"\"\"Returns the focal length of a three.js perspective camera.\n\n Args:\n fov: the field of view of the camera in degrees.\n image_height: the height of the image in pixels.\n \"\"\"\n pp_h = image_height / 2.0\n focal_length = pp_h / np.tan(fov * (np.pi / 180.0) / 2.0)\n return focal_length\n\n\ndef get_intrinsics_matrix_and_camera_to_world_h(\n camera_object: Dict[str, Any], image_height: int\n) -> Tuple[torch.Tensor, torch.Tensor]:\n \"\"\"Returns the camera intrinsics matrix and the camera to world homogeneous matrix.\n\n Args:\n camera_object: a Camera object.\n image_size: the size of the image (height, width)\n \"\"\"\n # intrinsics\n fov = camera_object[\"fov\"]\n aspect = camera_object[\"aspect\"]\n image_width = aspect * image_height\n pp_w = image_width / 2.0\n pp_h = image_height / 2.0\n focal_length = three_js_perspective_camera_focal_length(fov, image_height)\n intrinsics_matrix = torch.tensor([[focal_length, 0, pp_w], [0, focal_length, pp_h], [0, 0, 1]]).float()\n\n # extrinsics\n camera_to_world_h = torch.tensor(get_chunks(camera_object[\"matrix\"], size_of_chunk=4)).T.float()\n camera_to_world_h = torch.stack(\n [\n camera_to_world_h[0, :],\n camera_to_world_h[2, :],\n camera_to_world_h[1, :],\n camera_to_world_h[3, :],\n ],\n dim=0,\n )\n\n return intrinsics_matrix, camera_to_world_h\n\n\ndef find_available_port(func: Callable, default_port: int, max_attempts: int = 1000, **kwargs) -> None:\n \"\"\"Finds and attempts to connect to a port\n\n Args:\n func: function used on connecting to port\n default_port: the default port\n max_attempts: max number of attempts to try connection. Defaults to MAX_ATTEMPTS.\n \"\"\"\n for i in range(max_attempts):\n port = default_port + i\n try:\n return func(port, **kwargs), port\n except (OSError, zmq.error.ZMQError):\n print(f\"Port: {port:d} in use, trying another...\", file=sys.stderr)\n except Exception as e:\n print(type(e))\n raise\n raise (\n Exception(f\"Could not find an available port in the range: [{default_port:d}, {max_attempts + default_port:d})\")\n )\n\n\ndef force_codec(pc: RTCPeerConnection, sender: RTCRtpSender, forced_codec: str) -> None:\n \"\"\"Sets the codec preferences on a connection between sender and reciever\n\n Args:\n pc: peer connection point\n sender: sender that will send to connection point\n forced_codec: codec to set\n \"\"\"\n kind = forced_codec.split(\"/\")[0]\n codecs = RTCRtpSender.getCapabilities(kind).codecs\n transceiver = next(t for t in pc.getTransceivers() if t.sender == sender)\n transceiver.setCodecPreferences([codec for codec in codecs if codec.mimeType == forced_codec])\n", "path": "nerfstudio/viewer/server/utils.py"}]}
| 2,493 | 162 |
gh_patches_debug_28489
|
rasdani/github-patches
|
git_diff
|
sql-machine-learning__elasticdl-1214
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Remove manually create arguments in master/worker
In ElasticDL's master and worker. We manually create the arguments collection for them. It is a little hacky as we are adding more and more arguments.
After #1210 is merged, we can also update these two places where we manually add arguments.
</issue>
<code>
[start of elasticdl/python/master/main.py]
1 import os
2 import time
3 from concurrent import futures
4
5 import grpc
6 from kubernetes.client import V1EnvVar
7
8 from elasticdl.proto import elasticdl_pb2_grpc
9 from elasticdl.python.common.args import parse_envs, parse_master_args
10 from elasticdl.python.common.constants import (
11 GRPC,
12 JobType,
13 WorkerManagerStatus,
14 )
15 from elasticdl.python.common.data_reader import RecordIODataReader
16 from elasticdl.python.common.k8s_tensorboard_client import TensorBoardClient
17 from elasticdl.python.common.log_util import get_logger
18 from elasticdl.python.common.model_helper import (
19 find_layer,
20 get_module_file_path,
21 load_model_from_module,
22 load_module,
23 )
24 from elasticdl.python.elasticdl.layers.embedding import Embedding
25 from elasticdl.python.master.checkpoint_service import CheckpointService
26 from elasticdl.python.master.embedding_service import EmbeddingService
27 from elasticdl.python.master.evaluation_service import EvaluationService
28 from elasticdl.python.master.k8s_worker_manager import WorkerManager
29 from elasticdl.python.master.servicer import MasterServicer
30 from elasticdl.python.master.task_dispatcher import _TaskDispatcher
31 from elasticdl.python.master.tensorboard_service import TensorboardService
32
33
34 def _make_task_dispatcher(
35 training_data_dir,
36 evaluation_data_dir,
37 prediction_data_dir,
38 records_per_task,
39 num_epochs,
40 ):
41 # TODO: Support any subclasses of `AbstractDataReader`
42 # and support passing specified parameters to the constructor
43 prediction_f_records = RecordIODataReader(
44 data_dir=prediction_data_dir
45 ).create_shards()
46
47 return _TaskDispatcher(
48 RecordIODataReader(data_dir=training_data_dir).create_shards(),
49 RecordIODataReader(data_dir=evaluation_data_dir).create_shards(),
50 prediction_f_records,
51 records_per_task,
52 # Only generate prediction tasks for 1 epoch
53 1 if prediction_f_records else num_epochs,
54 )
55
56
57 def main():
58 args = parse_master_args()
59 logger = get_logger("master", level=args.log_level.upper())
60
61 # Master addr
62 master_ip = os.getenv("MY_POD_IP", "localhost")
63 master_addr = "%s:%d" % (master_ip, args.port)
64
65 # Start TensorBoard service if requested
66 if args.tensorboard_log_dir:
67 logger.info(
68 "Starting TensorBoard service with log directory %s",
69 args.tensorboard_log_dir,
70 )
71 # Start TensorBoard CLI
72 tb_service = TensorboardService(args.tensorboard_log_dir, master_ip)
73 tb_service.start()
74 else:
75 tb_service = None
76
77 # Start task queue
78 logger.debug(
79 "Starting task queue with training data directory %s, "
80 "evaluation data directory %s, "
81 "and prediction data directory %s",
82 args.training_data_dir,
83 args.evaluation_data_dir,
84 args.prediction_data_dir,
85 )
86 task_d = _make_task_dispatcher(
87 args.training_data_dir,
88 args.evaluation_data_dir,
89 args.prediction_data_dir,
90 args.records_per_task,
91 args.num_epochs,
92 )
93 model_module = load_module(
94 get_module_file_path(args.model_zoo, args.model_def)
95 ).__dict__
96 model_inst = load_model_from_module(
97 args.model_def, model_module, args.model_params
98 )
99 optimizer = model_module[args.optimizer]()
100
101 if all(
102 (
103 args.training_data_dir,
104 args.evaluation_data_dir,
105 args.evaluation_throttle_secs or args.evaluation_steps,
106 )
107 ):
108 job_type = JobType.TRAINING_WITH_EVALUATION
109 elif all(
110 (
111 args.evaluation_data_dir,
112 not args.training_data_dir,
113 not args.prediction_data_dir,
114 )
115 ):
116 job_type = JobType.EVALUATION_ONLY
117 elif all(
118 (
119 args.prediction_data_dir,
120 not args.evaluation_data_dir,
121 not args.training_data_dir,
122 )
123 ):
124 job_type = JobType.PREDICTION_ONLY
125 else:
126 job_type = JobType.TRAINING_ONLY
127
128 # Initialize checkpoint service
129 if args.checkpoint_steps or job_type == JobType.TRAINING_WITH_EVALUATION:
130 logger.info("Starting checkpoint service")
131 checkpoint_service = CheckpointService(
132 args.checkpoint_dir,
133 args.checkpoint_steps,
134 args.keep_checkpoint_max,
135 job_type == JobType.TRAINING_WITH_EVALUATION,
136 )
137 else:
138 checkpoint_service = None
139
140 # Initialize evaluation service
141 evaluation_service = None
142 if (
143 job_type == JobType.TRAINING_WITH_EVALUATION
144 or job_type == JobType.EVALUATION_ONLY
145 ):
146 logger.info(
147 "Starting evaluation service with throttle seconds %d "
148 " and evaluation steps %d",
149 args.evaluation_throttle_secs,
150 args.evaluation_steps,
151 )
152 evaluation_service = EvaluationService(
153 checkpoint_service,
154 tb_service,
155 task_d,
156 args.evaluation_start_delay_secs,
157 args.evaluation_throttle_secs,
158 args.evaluation_steps,
159 job_type == JobType.EVALUATION_ONLY,
160 )
161 evaluation_service.start()
162 task_d.set_evaluation_service(evaluation_service)
163
164 embedding_service_endpoint = None
165 embedding_dims = {}
166 # Search for embedding layers in the model,
167 # if found, initialize embedding service
168 layers = find_layer(model_inst, Embedding)
169 if layers:
170 embedding_service = EmbeddingService()
171 embedding_service_endpoint = embedding_service.start_embedding_service(
172 job_name=args.job_name,
173 image_name=args.worker_image,
174 namespace=args.namespace,
175 resource_request=args.master_resource_request,
176 resource_limit=args.master_resource_limit,
177 pod_priority=args.worker_pod_priority,
178 volume=args.volume,
179 image_pull_policy=args.image_pull_policy,
180 restart_policy=args.restart_policy,
181 cluster_spec=args.cluster_spec,
182 )
183 logger.info(
184 "Embedding service start succeeded. The endpoint is %s."
185 % str(embedding_service_endpoint)
186 )
187 embedding_dims = dict(
188 [(layer.name, layer.output_dim) for layer in layers]
189 )
190
191 # The master service
192 logger.info("Starting master service")
193 server = grpc.server(
194 futures.ThreadPoolExecutor(max_workers=64),
195 options=[
196 ("grpc.max_send_message_length", GRPC.MAX_SEND_MESSAGE_LENGTH),
197 (
198 "grpc.max_receive_message_length",
199 GRPC.MAX_RECEIVE_MESSAGE_LENGTH,
200 ),
201 ],
202 )
203 master_servicer = MasterServicer(
204 args.grads_to_wait,
205 args.minibatch_size,
206 optimizer,
207 task_d,
208 init_var=model_inst.trainable_variables if model_inst.built else [],
209 embedding_dims=embedding_dims,
210 checkpoint_filename_for_init=args.checkpoint_filename_for_init,
211 checkpoint_service=checkpoint_service,
212 evaluation_service=evaluation_service,
213 embedding_service_endpoint=embedding_service_endpoint,
214 lr_staleness_modulation=args.lr_staleness_modulation,
215 use_async=args.use_async,
216 )
217 elasticdl_pb2_grpc.add_MasterServicer_to_server(master_servicer, server)
218 server.add_insecure_port("[::]:{}".format(args.port))
219 server.start()
220 logger.info("Server started at port: %d", args.port)
221
222 worker_manager = None
223 if args.num_workers:
224 assert args.worker_image, "Worker image cannot be empty"
225
226 worker_command = ["python"]
227 worker_args = [
228 "-m",
229 "elasticdl.python.worker.main",
230 "--model_zoo",
231 args.model_zoo,
232 "--master_addr",
233 master_addr,
234 "--log_level",
235 args.log_level,
236 "--dataset_fn",
237 args.dataset_fn,
238 "--loss",
239 args.loss,
240 "--optimizer",
241 args.optimizer,
242 "--eval_metrics_fn",
243 args.eval_metrics_fn,
244 "--model_def",
245 args.model_def,
246 "--job_type",
247 job_type,
248 "--minibatch_size",
249 str(args.minibatch_size),
250 "--embedding_service_endpoint",
251 str(embedding_service_endpoint),
252 "--get_model_steps",
253 str(args.get_model_steps),
254 ]
255
256 env_dict = parse_envs(args.envs)
257 env = []
258 for key in env_dict:
259 env.append(V1EnvVar(name=key, value=env_dict[key]))
260
261 worker_manager = WorkerManager(
262 task_d,
263 job_name=args.job_name,
264 image_name=args.worker_image,
265 command=worker_command,
266 args=worker_args,
267 namespace=args.namespace,
268 num_workers=args.num_workers,
269 worker_resource_request=args.worker_resource_request,
270 worker_resource_limit=args.worker_resource_limit,
271 pod_priority=args.worker_pod_priority,
272 volume=args.volume,
273 image_pull_policy=args.image_pull_policy,
274 restart_policy=args.restart_policy,
275 cluster_spec=args.cluster_spec,
276 envs=env,
277 )
278 worker_manager.update_status(WorkerManagerStatus.PENDING)
279 logger.info("Launching %d workers", args.num_workers)
280 worker_manager.start_workers()
281 worker_manager.update_status(WorkerManagerStatus.RUNNING)
282
283 # Start TensorBoard k8s Service if requested
284 if tb_service:
285 TensorBoardClient(
286 job_name=args.job_name,
287 image_name=args.worker_image,
288 namespace=args.namespace,
289 ).start_tensorboard_service()
290
291 try:
292 while True:
293 if task_d.finished():
294 if worker_manager:
295 worker_manager.update_status(WorkerManagerStatus.FINISHED)
296 if args.output:
297 master_servicer.save_latest_checkpoint(args.output)
298 break
299 time.sleep(30)
300 except KeyboardInterrupt:
301 logger.warning("Server stopping")
302
303 if evaluation_service:
304 logger.info("Stopping evaluation service")
305 evaluation_service.stop()
306
307 logger.info("Stopping RPC server")
308 server.stop(0)
309
310 # Keep TensorBoard running when all the tasks are finished
311 if tb_service:
312 logger.info(
313 "All tasks finished. Keeping TensorBoard service running..."
314 )
315 while True:
316 if tb_service.is_active():
317 time.sleep(10)
318 else:
319 logger.warning(
320 "Unable to keep TensorBoard running. "
321 "It has already terminated"
322 )
323 break
324 logger.info("Master stopped")
325
326
327 if __name__ == "__main__":
328 main()
329
[end of elasticdl/python/master/main.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/elasticdl/python/master/main.py b/elasticdl/python/master/main.py
--- a/elasticdl/python/master/main.py
+++ b/elasticdl/python/master/main.py
@@ -6,7 +6,11 @@
from kubernetes.client import V1EnvVar
from elasticdl.proto import elasticdl_pb2_grpc
-from elasticdl.python.common.args import parse_envs, parse_master_args
+from elasticdl.python.common.args import (
+ build_arguments_from_parsed_result,
+ parse_envs,
+ parse_master_args,
+)
from elasticdl.python.common.constants import (
GRPC,
JobType,
@@ -227,31 +231,14 @@
worker_args = [
"-m",
"elasticdl.python.worker.main",
- "--model_zoo",
- args.model_zoo,
"--master_addr",
master_addr,
- "--log_level",
- args.log_level,
- "--dataset_fn",
- args.dataset_fn,
- "--loss",
- args.loss,
- "--optimizer",
- args.optimizer,
- "--eval_metrics_fn",
- args.eval_metrics_fn,
- "--model_def",
- args.model_def,
"--job_type",
job_type,
- "--minibatch_size",
- str(args.minibatch_size),
"--embedding_service_endpoint",
str(embedding_service_endpoint),
- "--get_model_steps",
- str(args.get_model_steps),
]
+ worker_args.extend(build_arguments_from_parsed_result(args))
env_dict = parse_envs(args.envs)
env = []
|
{"golden_diff": "diff --git a/elasticdl/python/master/main.py b/elasticdl/python/master/main.py\n--- a/elasticdl/python/master/main.py\n+++ b/elasticdl/python/master/main.py\n@@ -6,7 +6,11 @@\n from kubernetes.client import V1EnvVar\n \n from elasticdl.proto import elasticdl_pb2_grpc\n-from elasticdl.python.common.args import parse_envs, parse_master_args\n+from elasticdl.python.common.args import (\n+ build_arguments_from_parsed_result,\n+ parse_envs,\n+ parse_master_args,\n+)\n from elasticdl.python.common.constants import (\n GRPC,\n JobType,\n@@ -227,31 +231,14 @@\n worker_args = [\n \"-m\",\n \"elasticdl.python.worker.main\",\n- \"--model_zoo\",\n- args.model_zoo,\n \"--master_addr\",\n master_addr,\n- \"--log_level\",\n- args.log_level,\n- \"--dataset_fn\",\n- args.dataset_fn,\n- \"--loss\",\n- args.loss,\n- \"--optimizer\",\n- args.optimizer,\n- \"--eval_metrics_fn\",\n- args.eval_metrics_fn,\n- \"--model_def\",\n- args.model_def,\n \"--job_type\",\n job_type,\n- \"--minibatch_size\",\n- str(args.minibatch_size),\n \"--embedding_service_endpoint\",\n str(embedding_service_endpoint),\n- \"--get_model_steps\",\n- str(args.get_model_steps),\n ]\n+ worker_args.extend(build_arguments_from_parsed_result(args))\n \n env_dict = parse_envs(args.envs)\n env = []\n", "issue": "Remove manually create arguments in master/worker\nIn ElasticDL's master and worker. We manually create the arguments collection for them. It is a little hacky as we are adding more and more arguments.\r\nAfter #1210 is merged, we can also update these two places where we manually add arguments.\n", "before_files": [{"content": "import os\nimport time\nfrom concurrent import futures\n\nimport grpc\nfrom kubernetes.client import V1EnvVar\n\nfrom elasticdl.proto import elasticdl_pb2_grpc\nfrom elasticdl.python.common.args import parse_envs, parse_master_args\nfrom elasticdl.python.common.constants import (\n GRPC,\n JobType,\n WorkerManagerStatus,\n)\nfrom elasticdl.python.common.data_reader import RecordIODataReader\nfrom elasticdl.python.common.k8s_tensorboard_client import TensorBoardClient\nfrom elasticdl.python.common.log_util import get_logger\nfrom elasticdl.python.common.model_helper import (\n find_layer,\n get_module_file_path,\n load_model_from_module,\n load_module,\n)\nfrom elasticdl.python.elasticdl.layers.embedding import Embedding\nfrom elasticdl.python.master.checkpoint_service import CheckpointService\nfrom elasticdl.python.master.embedding_service import EmbeddingService\nfrom elasticdl.python.master.evaluation_service import EvaluationService\nfrom elasticdl.python.master.k8s_worker_manager import WorkerManager\nfrom elasticdl.python.master.servicer import MasterServicer\nfrom elasticdl.python.master.task_dispatcher import _TaskDispatcher\nfrom elasticdl.python.master.tensorboard_service import TensorboardService\n\n\ndef _make_task_dispatcher(\n training_data_dir,\n evaluation_data_dir,\n prediction_data_dir,\n records_per_task,\n num_epochs,\n):\n # TODO: Support any subclasses of `AbstractDataReader`\n # and support passing specified parameters to the constructor\n prediction_f_records = RecordIODataReader(\n data_dir=prediction_data_dir\n ).create_shards()\n\n return _TaskDispatcher(\n RecordIODataReader(data_dir=training_data_dir).create_shards(),\n RecordIODataReader(data_dir=evaluation_data_dir).create_shards(),\n prediction_f_records,\n records_per_task,\n # Only generate prediction tasks for 1 epoch\n 1 if prediction_f_records else num_epochs,\n )\n\n\ndef main():\n args = parse_master_args()\n logger = get_logger(\"master\", level=args.log_level.upper())\n\n # Master addr\n master_ip = os.getenv(\"MY_POD_IP\", \"localhost\")\n master_addr = \"%s:%d\" % (master_ip, args.port)\n\n # Start TensorBoard service if requested\n if args.tensorboard_log_dir:\n logger.info(\n \"Starting TensorBoard service with log directory %s\",\n args.tensorboard_log_dir,\n )\n # Start TensorBoard CLI\n tb_service = TensorboardService(args.tensorboard_log_dir, master_ip)\n tb_service.start()\n else:\n tb_service = None\n\n # Start task queue\n logger.debug(\n \"Starting task queue with training data directory %s, \"\n \"evaluation data directory %s, \"\n \"and prediction data directory %s\",\n args.training_data_dir,\n args.evaluation_data_dir,\n args.prediction_data_dir,\n )\n task_d = _make_task_dispatcher(\n args.training_data_dir,\n args.evaluation_data_dir,\n args.prediction_data_dir,\n args.records_per_task,\n args.num_epochs,\n )\n model_module = load_module(\n get_module_file_path(args.model_zoo, args.model_def)\n ).__dict__\n model_inst = load_model_from_module(\n args.model_def, model_module, args.model_params\n )\n optimizer = model_module[args.optimizer]()\n\n if all(\n (\n args.training_data_dir,\n args.evaluation_data_dir,\n args.evaluation_throttle_secs or args.evaluation_steps,\n )\n ):\n job_type = JobType.TRAINING_WITH_EVALUATION\n elif all(\n (\n args.evaluation_data_dir,\n not args.training_data_dir,\n not args.prediction_data_dir,\n )\n ):\n job_type = JobType.EVALUATION_ONLY\n elif all(\n (\n args.prediction_data_dir,\n not args.evaluation_data_dir,\n not args.training_data_dir,\n )\n ):\n job_type = JobType.PREDICTION_ONLY\n else:\n job_type = JobType.TRAINING_ONLY\n\n # Initialize checkpoint service\n if args.checkpoint_steps or job_type == JobType.TRAINING_WITH_EVALUATION:\n logger.info(\"Starting checkpoint service\")\n checkpoint_service = CheckpointService(\n args.checkpoint_dir,\n args.checkpoint_steps,\n args.keep_checkpoint_max,\n job_type == JobType.TRAINING_WITH_EVALUATION,\n )\n else:\n checkpoint_service = None\n\n # Initialize evaluation service\n evaluation_service = None\n if (\n job_type == JobType.TRAINING_WITH_EVALUATION\n or job_type == JobType.EVALUATION_ONLY\n ):\n logger.info(\n \"Starting evaluation service with throttle seconds %d \"\n \" and evaluation steps %d\",\n args.evaluation_throttle_secs,\n args.evaluation_steps,\n )\n evaluation_service = EvaluationService(\n checkpoint_service,\n tb_service,\n task_d,\n args.evaluation_start_delay_secs,\n args.evaluation_throttle_secs,\n args.evaluation_steps,\n job_type == JobType.EVALUATION_ONLY,\n )\n evaluation_service.start()\n task_d.set_evaluation_service(evaluation_service)\n\n embedding_service_endpoint = None\n embedding_dims = {}\n # Search for embedding layers in the model,\n # if found, initialize embedding service\n layers = find_layer(model_inst, Embedding)\n if layers:\n embedding_service = EmbeddingService()\n embedding_service_endpoint = embedding_service.start_embedding_service(\n job_name=args.job_name,\n image_name=args.worker_image,\n namespace=args.namespace,\n resource_request=args.master_resource_request,\n resource_limit=args.master_resource_limit,\n pod_priority=args.worker_pod_priority,\n volume=args.volume,\n image_pull_policy=args.image_pull_policy,\n restart_policy=args.restart_policy,\n cluster_spec=args.cluster_spec,\n )\n logger.info(\n \"Embedding service start succeeded. The endpoint is %s.\"\n % str(embedding_service_endpoint)\n )\n embedding_dims = dict(\n [(layer.name, layer.output_dim) for layer in layers]\n )\n\n # The master service\n logger.info(\"Starting master service\")\n server = grpc.server(\n futures.ThreadPoolExecutor(max_workers=64),\n options=[\n (\"grpc.max_send_message_length\", GRPC.MAX_SEND_MESSAGE_LENGTH),\n (\n \"grpc.max_receive_message_length\",\n GRPC.MAX_RECEIVE_MESSAGE_LENGTH,\n ),\n ],\n )\n master_servicer = MasterServicer(\n args.grads_to_wait,\n args.minibatch_size,\n optimizer,\n task_d,\n init_var=model_inst.trainable_variables if model_inst.built else [],\n embedding_dims=embedding_dims,\n checkpoint_filename_for_init=args.checkpoint_filename_for_init,\n checkpoint_service=checkpoint_service,\n evaluation_service=evaluation_service,\n embedding_service_endpoint=embedding_service_endpoint,\n lr_staleness_modulation=args.lr_staleness_modulation,\n use_async=args.use_async,\n )\n elasticdl_pb2_grpc.add_MasterServicer_to_server(master_servicer, server)\n server.add_insecure_port(\"[::]:{}\".format(args.port))\n server.start()\n logger.info(\"Server started at port: %d\", args.port)\n\n worker_manager = None\n if args.num_workers:\n assert args.worker_image, \"Worker image cannot be empty\"\n\n worker_command = [\"python\"]\n worker_args = [\n \"-m\",\n \"elasticdl.python.worker.main\",\n \"--model_zoo\",\n args.model_zoo,\n \"--master_addr\",\n master_addr,\n \"--log_level\",\n args.log_level,\n \"--dataset_fn\",\n args.dataset_fn,\n \"--loss\",\n args.loss,\n \"--optimizer\",\n args.optimizer,\n \"--eval_metrics_fn\",\n args.eval_metrics_fn,\n \"--model_def\",\n args.model_def,\n \"--job_type\",\n job_type,\n \"--minibatch_size\",\n str(args.minibatch_size),\n \"--embedding_service_endpoint\",\n str(embedding_service_endpoint),\n \"--get_model_steps\",\n str(args.get_model_steps),\n ]\n\n env_dict = parse_envs(args.envs)\n env = []\n for key in env_dict:\n env.append(V1EnvVar(name=key, value=env_dict[key]))\n\n worker_manager = WorkerManager(\n task_d,\n job_name=args.job_name,\n image_name=args.worker_image,\n command=worker_command,\n args=worker_args,\n namespace=args.namespace,\n num_workers=args.num_workers,\n worker_resource_request=args.worker_resource_request,\n worker_resource_limit=args.worker_resource_limit,\n pod_priority=args.worker_pod_priority,\n volume=args.volume,\n image_pull_policy=args.image_pull_policy,\n restart_policy=args.restart_policy,\n cluster_spec=args.cluster_spec,\n envs=env,\n )\n worker_manager.update_status(WorkerManagerStatus.PENDING)\n logger.info(\"Launching %d workers\", args.num_workers)\n worker_manager.start_workers()\n worker_manager.update_status(WorkerManagerStatus.RUNNING)\n\n # Start TensorBoard k8s Service if requested\n if tb_service:\n TensorBoardClient(\n job_name=args.job_name,\n image_name=args.worker_image,\n namespace=args.namespace,\n ).start_tensorboard_service()\n\n try:\n while True:\n if task_d.finished():\n if worker_manager:\n worker_manager.update_status(WorkerManagerStatus.FINISHED)\n if args.output:\n master_servicer.save_latest_checkpoint(args.output)\n break\n time.sleep(30)\n except KeyboardInterrupt:\n logger.warning(\"Server stopping\")\n\n if evaluation_service:\n logger.info(\"Stopping evaluation service\")\n evaluation_service.stop()\n\n logger.info(\"Stopping RPC server\")\n server.stop(0)\n\n # Keep TensorBoard running when all the tasks are finished\n if tb_service:\n logger.info(\n \"All tasks finished. Keeping TensorBoard service running...\"\n )\n while True:\n if tb_service.is_active():\n time.sleep(10)\n else:\n logger.warning(\n \"Unable to keep TensorBoard running. \"\n \"It has already terminated\"\n )\n break\n logger.info(\"Master stopped\")\n\n\nif __name__ == \"__main__\":\n main()\n", "path": "elasticdl/python/master/main.py"}]}
| 3,633 | 345 |
gh_patches_debug_335
|
rasdani/github-patches
|
git_diff
|
pymodbus-dev__pymodbus-1395
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
pip show pymodbus, misses information.
```
pymodbus) pymodbus % pip show pymodbus
Name: pymodbus
Version: 3.1.x
Summary: A fully featured modbus protocol stack in python
Home-page: https://github.com/pymodbus-dev/pymodbus/
Author: attr: pymodbus.__author__
Author-email:
License: BSD-3-Clause
Location: /Users/jan/repos/pymodbus
Editable project location: /Users/jan/repos/pymodbus
Requires: setuptools
Required-by:
```
Normally it gets the information from setup.cfg, but for some reason it does not work with "pip show".
</issue>
<code>
[start of pymodbus/__init__.py]
1 """Pymodbus: Modbus Protocol Implementation.
2
3 Released under the the BSD license
4 """
5
6 from logging import WARNING
7
8 import pymodbus.version as __version
9 from pymodbus.logging import Log
10
11
12 __version__ = __version.version.short()
13 __author__ = "Galen Collins"
14 __maintainer__ = "dhoomakethu, janiversen"
15
16
17 def pymodbus_apply_logging_config(level=WARNING):
18 """Apply basic logging configuration used by default by Pymodbus maintainers.
19
20 Please call this function to format logging appropriately when opening issues.
21 """
22 Log.apply_logging_config(level)
23
[end of pymodbus/__init__.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/pymodbus/__init__.py b/pymodbus/__init__.py
--- a/pymodbus/__init__.py
+++ b/pymodbus/__init__.py
@@ -10,7 +10,7 @@
__version__ = __version.version.short()
-__author__ = "Galen Collins"
+__author__ = "Galen Collins, Jan Iversen"
__maintainer__ = "dhoomakethu, janiversen"
|
{"golden_diff": "diff --git a/pymodbus/__init__.py b/pymodbus/__init__.py\n--- a/pymodbus/__init__.py\n+++ b/pymodbus/__init__.py\n@@ -10,7 +10,7 @@\n \n \n __version__ = __version.version.short()\n-__author__ = \"Galen Collins\"\n+__author__ = \"Galen Collins, Jan Iversen\"\n __maintainer__ = \"dhoomakethu, janiversen\"\n", "issue": "pip show pymodbus, misses information.\n```\r\npymodbus) pymodbus % pip show pymodbus\r\n\r\nName: pymodbus\r\nVersion: 3.1.x\r\nSummary: A fully featured modbus protocol stack in python\r\nHome-page: https://github.com/pymodbus-dev/pymodbus/\r\nAuthor: attr: pymodbus.__author__\r\nAuthor-email: \r\nLicense: BSD-3-Clause\r\nLocation: /Users/jan/repos/pymodbus\r\nEditable project location: /Users/jan/repos/pymodbus\r\nRequires: setuptools\r\nRequired-by: \r\n```\r\nNormally it gets the information from setup.cfg, but for some reason it does not work with \"pip show\".\n", "before_files": [{"content": "\"\"\"Pymodbus: Modbus Protocol Implementation.\n\nReleased under the the BSD license\n\"\"\"\n\nfrom logging import WARNING\n\nimport pymodbus.version as __version\nfrom pymodbus.logging import Log\n\n\n__version__ = __version.version.short()\n__author__ = \"Galen Collins\"\n__maintainer__ = \"dhoomakethu, janiversen\"\n\n\ndef pymodbus_apply_logging_config(level=WARNING):\n \"\"\"Apply basic logging configuration used by default by Pymodbus maintainers.\n\n Please call this function to format logging appropriately when opening issues.\n \"\"\"\n Log.apply_logging_config(level)\n", "path": "pymodbus/__init__.py"}]}
| 858 | 107 |
gh_patches_debug_17036
|
rasdani/github-patches
|
git_diff
|
HybirdCorp__creme_crm-234
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
management command 'geolocation --stat' failed
really not a blocking bug, but seems very easy to fix:
```
self.sysout(f'{Town.objects.count()} town(s) in database.')
TypeError: sysout() missing 1 required positional argument: 'visible'
```
there is just a missing argument in self.sysout()
</issue>
<code>
[start of creme/geolocation/management/commands/geolocation.py]
1 # -*- coding: utf-8 -*-
2
3 ################################################################################
4 # Creme is a free/open-source Customer Relationship Management software
5 # Copyright (C) 2015-2021 Hybird
6 #
7 # This program is free software: you can redistribute it and/or modify
8 # it under the terms of the GNU Affero General Public License as published by
9 # the Free Software Foundation, either version 3 of the License, or
10 # (at your option) any later version.
11 #
12 # This program is distributed in the hope that it will be useful,
13 # but WITHOUT ANY WARRANTY; without even the implied warranty of
14 # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
15 # GNU Affero General Public License for more details.
16 #
17 # You should have received a copy of the GNU Affero General Public License
18 # along with this program. If not, see <http://www.gnu.org/licenses/>.
19 ################################################################################
20
21 import csv
22 import io
23 import logging
24 from functools import partial
25 from urllib.request import urlopen
26 from zipfile import ZipFile
27
28 from django.conf import settings
29 from django.core.management.base import BaseCommand
30 from django.db import transaction
31 from django.template.defaultfilters import slugify
32
33 from creme.creme_core.utils.chunktools import iter_as_chunk
34 from creme.creme_core.utils.collections import OrderedSet
35 from creme.creme_core.utils.url import parse_path
36 from creme.persons import get_address_model
37
38 from ...models import GeoAddress, Town
39
40 logger = logging.getLogger(__name__)
41
42
43 class CSVPopulatorError(Exception):
44 pass
45
46
47 class CSVPopulator:
48 class ProtocolError(CSVPopulatorError):
49 pass
50
51 class ReadError(CSVPopulatorError):
52 pass
53
54 class ParseError(CSVPopulatorError):
55 pass
56
57 class Context:
58 def __init__(self, defaults):
59 self.line = 1
60 self.defaults = defaults
61
62 def __init__(self, columns, defaults=None, chunksize=50):
63 """Constructor.
64 @param columns: Names of columns to extract from CSV file.
65 Raises an error if a column is neither in file nor in defaults.
66 @param defaults: dict of default values.
67 @param chunksize: Number of lines in same transaction.
68 By default sqlite supports 999 entries for each transaction,
69 so use 999/fields as max chunksize value.
70 """
71 self.columns = columns
72 self.defaults = defaults or {}
73 self.chunksize = chunksize
74
75 def _get_source_file(self, url_info):
76 if url_info.scheme in {'file', ''}:
77 self.info(f'Reading database from {url_info.geturl()}...')
78 return open(url_info.path, 'rb')
79 elif url_info.scheme in {'http', 'https'}:
80 self.info(f'Downloading database from {url_info.geturl()}...')
81 return urlopen(url_info.geturl())
82 else:
83 raise self.ProtocolError(
84 f'Unable to open CSV data from {url_info.geturl()} : '
85 f'unsupported protocol.'
86 )
87
88 def _mapper(self, header):
89 columns = self.columns
90 defaults = self.defaults
91
92 column_keys = OrderedSet(h.lower() for h in columns) # TODO: OrderedFrozenSet
93 row_keys = frozenset(k.lower() for k in header)
94
95 missings = []
96 constants = {}
97 indices = [(key, index) for index, key in enumerate(header) if key in column_keys]
98
99 for key in column_keys:
100 if key not in row_keys:
101 try:
102 constants[key] = defaults[key]
103 except KeyError:
104 missings.append(key)
105
106 if missings:
107 raise self.ParseError(
108 f"Following columns are missing and haven't got any default value : {missings}"
109 )
110
111 def _aux(row):
112 data = {key: row[index] or defaults.get(key) for key, index in indices}
113 data.update(constants)
114 return data
115
116 return _aux
117
118 def create(self, row, context):
119 raise NotImplementedError
120
121 def save(self, entries, context):
122 raise NotImplementedError
123
124 def pre(self, rows, context):
125 pass
126
127 def post(self, entries, context):
128 pass
129
130 def line_error(self, e, row, context):
131 pass
132
133 def chunk_error(self, e, rows, context):
134 pass
135
136 def info(self, message):
137 logger.info(message)
138
139 def populate(self, source):
140 if isinstance(source, str):
141 try:
142 url_info = parse_path(source)
143
144 with self._get_source_file(url_info) as bytes_input:
145 if url_info.path.endswith('.zip'):
146 archive = ZipFile(
147 bytes_input
148 if bytes_input.seekable() else
149 io.BytesIO(bytes_input.read())
150 )
151
152 with archive.open(archive.namelist()[0]) as zipped_bytes_input:
153 self._populate_from_bytes(zipped_bytes_input)
154 else:
155 self._populate_from_bytes(bytes_input)
156 except CSVPopulatorError:
157 raise
158 except Exception as e:
159 raise self.ReadError(f'Unable to open CSV data from {source} : {e}') from e
160 elif hasattr(source, '__iter__'):
161 self._populate_from_lines(iter(source))
162 else:
163 raise ValueError('The source must be a path or an iterable.')
164
165 def _populate_from_bytes(self, bytes_input):
166 with io.TextIOWrapper(bytes_input) as wrapped_bytes_input:
167 self._populate_from_lines(csv.reader(wrapped_bytes_input))
168
169 def _populate_from_lines(self, lines):
170 mapper = self._mapper(next(lines))
171 context = self.Context(self.defaults)
172
173 for rows in iter_as_chunk(lines, self.chunksize):
174 entries = []
175
176 if mapper:
177 rows = [mapper(row) for row in rows]
178
179 try:
180 self.pre(rows, context)
181
182 for row in rows:
183 try:
184 entries.extend(self.create(row, context))
185 except Exception as e:
186 self.line_error(e, row, context)
187
188 context.line += 1
189
190 self.save(entries, context)
191 self.post(entries, context)
192 except Exception as e:
193 self.chunk_error(e, rows, context)
194
195 def sync(self, model, entries, build_pk):
196 created = []
197 updated = []
198
199 for t in entries:
200 pk = build_pk(t)
201
202 if not pk:
203 created.append(t)
204 else:
205 t.pk = pk
206 updated.append(t)
207
208 with transaction.atomic():
209 model.objects.bulk_create(created)
210
211 # TODO: bulk_update() ?
212 for entry in updated:
213 entry.save(force_update=True)
214
215
216 class CSVTownPopulator(CSVPopulator):
217 def __init__(self, defaults=None, chunksize=100):
218 super().__init__(['title', 'zipcode', 'latitude', 'longitude', 'country'],
219 defaults=defaults, chunksize=chunksize,
220 )
221
222 def line_error(self, e, row, context):
223 logger.error(' invalid data (line %d) : %s', context.line, e)
224
225 def chunk_error(self, e, rows, context):
226 logger.error(' invalid data chunk : %s', e)
227
228 def create(self, row, context):
229 zipcodes = row['zipcode'].split('-')
230
231 name = row['title']
232 latitude = row['latitude']
233 longitude = row['longitude']
234
235 slug = slugify(name)
236 country = row['country']
237
238 build_town = partial(Town, country=country)
239
240 return [build_town(name=name,
241 slug=slug,
242 zipcode=zipcode,
243 latitude=latitude,
244 longitude=longitude,
245 ) for zipcode in zipcodes
246 ]
247
248 def save(self, entries, context):
249 get_existing_pk_by_zipcode = dict(
250 Town.objects.filter(
251 zipcode__in=(t.zipcode for t in entries),
252 slug__in=(t.slug for t in entries),
253 ).values_list('zipcode', 'pk')
254 ).get
255
256 self.sync(
257 Town, entries,
258 lambda town: get_existing_pk_by_zipcode(town.zipcode),
259 )
260
261
262 class Command(BaseCommand):
263 def add_arguments(self, parser):
264 add_argument = parser.add_argument
265 add_argument(
266 '-p', '--populate', action='store_true', dest='populate',
267 help='Populate addresses', default=False,
268 )
269 add_argument(
270 '-s', '--stat', action='store_true', dest='stats',
271 help='Display geolocation database stats', default=False,
272 )
273 add_argument(
274 '-i', '--import', action='store_true', dest='import', default=False,
275 help='Import towns configured in GEOLOCATION_TOWNS setting',
276 )
277
278 def sysout(self, message, visible):
279 if visible:
280 self.stdout.write(message)
281
282 def syserr(self, message):
283 self.stderr.write(message)
284
285 def populate_addresses(self, verbosity=0):
286 self.sysout('Populate geolocation information of addresses...', verbosity > 0)
287 GeoAddress.populate_geoaddresses(get_address_model().objects.exclude(zipcode='', city=''))
288
289 def import_town_database(self, url, defaults):
290 try:
291 CSVTownPopulator(defaults=defaults).populate(url)
292 except Exception as e:
293 self.syserr(str(e))
294
295 def import_town_all(self, verbosity=0):
296 self.sysout('Importing Towns database...', verbosity > 0)
297
298 for url, defaults in settings.GEOLOCATION_TOWNS:
299 self.sysout(url, verbosity > 1)
300 self.import_town_database(url, defaults)
301
302 def print_stats(self):
303 self.sysout(f'{Town.objects.count()} town(s) in database.')
304
305 def handle(self, *args, **options):
306 populate = options.get('populate')
307 stats = options.get('stats')
308 imports = options.get('import')
309 verbosity = options.get('verbosity')
310
311 if stats:
312 self.print_stats()
313
314 if imports:
315 self.import_town_all(verbosity)
316
317 if populate:
318 self.populate_addresses(verbosity)
319
[end of creme/geolocation/management/commands/geolocation.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/creme/geolocation/management/commands/geolocation.py b/creme/geolocation/management/commands/geolocation.py
--- a/creme/geolocation/management/commands/geolocation.py
+++ b/creme/geolocation/management/commands/geolocation.py
@@ -299,8 +299,8 @@
self.sysout(url, verbosity > 1)
self.import_town_database(url, defaults)
- def print_stats(self):
- self.sysout(f'{Town.objects.count()} town(s) in database.')
+ def print_stats(self, verbosity=0):
+ self.sysout(f'{Town.objects.count()} town(s) in database.', verbosity > 0)
def handle(self, *args, **options):
populate = options.get('populate')
@@ -309,7 +309,7 @@
verbosity = options.get('verbosity')
if stats:
- self.print_stats()
+ self.print_stats(verbosity)
if imports:
self.import_town_all(verbosity)
|
{"golden_diff": "diff --git a/creme/geolocation/management/commands/geolocation.py b/creme/geolocation/management/commands/geolocation.py\n--- a/creme/geolocation/management/commands/geolocation.py\n+++ b/creme/geolocation/management/commands/geolocation.py\n@@ -299,8 +299,8 @@\n self.sysout(url, verbosity > 1)\n self.import_town_database(url, defaults)\n \n- def print_stats(self):\n- self.sysout(f'{Town.objects.count()} town(s) in database.')\n+ def print_stats(self, verbosity=0):\n+ self.sysout(f'{Town.objects.count()} town(s) in database.', verbosity > 0)\n \n def handle(self, *args, **options):\n populate = options.get('populate')\n@@ -309,7 +309,7 @@\n verbosity = options.get('verbosity')\n \n if stats:\n- self.print_stats()\n+ self.print_stats(verbosity)\n \n if imports:\n self.import_town_all(verbosity)\n", "issue": "management command 'geolocation --stat' failed\nreally not a blocking bug, but seems very easy to fix:\r\n```\r\n self.sysout(f'{Town.objects.count()} town(s) in database.')\r\nTypeError: sysout() missing 1 required positional argument: 'visible'\r\n```\r\n\r\nthere is just a missing argument in self.sysout()\n", "before_files": [{"content": "# -*- coding: utf-8 -*-\n\n################################################################################\n# Creme is a free/open-source Customer Relationship Management software\n# Copyright (C) 2015-2021 Hybird\n#\n# This program is free software: you can redistribute it and/or modify\n# it under the terms of the GNU Affero General Public License as published by\n# the Free Software Foundation, either version 3 of the License, or\n# (at your option) any later version.\n#\n# This program is distributed in the hope that it will be useful,\n# but WITHOUT ANY WARRANTY; without even the implied warranty of\n# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the\n# GNU Affero General Public License for more details.\n#\n# You should have received a copy of the GNU Affero General Public License\n# along with this program. If not, see <http://www.gnu.org/licenses/>.\n################################################################################\n\nimport csv\nimport io\nimport logging\nfrom functools import partial\nfrom urllib.request import urlopen\nfrom zipfile import ZipFile\n\nfrom django.conf import settings\nfrom django.core.management.base import BaseCommand\nfrom django.db import transaction\nfrom django.template.defaultfilters import slugify\n\nfrom creme.creme_core.utils.chunktools import iter_as_chunk\nfrom creme.creme_core.utils.collections import OrderedSet\nfrom creme.creme_core.utils.url import parse_path\nfrom creme.persons import get_address_model\n\nfrom ...models import GeoAddress, Town\n\nlogger = logging.getLogger(__name__)\n\n\nclass CSVPopulatorError(Exception):\n pass\n\n\nclass CSVPopulator:\n class ProtocolError(CSVPopulatorError):\n pass\n\n class ReadError(CSVPopulatorError):\n pass\n\n class ParseError(CSVPopulatorError):\n pass\n\n class Context:\n def __init__(self, defaults):\n self.line = 1\n self.defaults = defaults\n\n def __init__(self, columns, defaults=None, chunksize=50):\n \"\"\"Constructor.\n @param columns: Names of columns to extract from CSV file.\n Raises an error if a column is neither in file nor in defaults.\n @param defaults: dict of default values.\n @param chunksize: Number of lines in same transaction.\n By default sqlite supports 999 entries for each transaction,\n so use 999/fields as max chunksize value.\n \"\"\"\n self.columns = columns\n self.defaults = defaults or {}\n self.chunksize = chunksize\n\n def _get_source_file(self, url_info):\n if url_info.scheme in {'file', ''}:\n self.info(f'Reading database from {url_info.geturl()}...')\n return open(url_info.path, 'rb')\n elif url_info.scheme in {'http', 'https'}:\n self.info(f'Downloading database from {url_info.geturl()}...')\n return urlopen(url_info.geturl())\n else:\n raise self.ProtocolError(\n f'Unable to open CSV data from {url_info.geturl()} : '\n f'unsupported protocol.'\n )\n\n def _mapper(self, header):\n columns = self.columns\n defaults = self.defaults\n\n column_keys = OrderedSet(h.lower() for h in columns) # TODO: OrderedFrozenSet\n row_keys = frozenset(k.lower() for k in header)\n\n missings = []\n constants = {}\n indices = [(key, index) for index, key in enumerate(header) if key in column_keys]\n\n for key in column_keys:\n if key not in row_keys:\n try:\n constants[key] = defaults[key]\n except KeyError:\n missings.append(key)\n\n if missings:\n raise self.ParseError(\n f\"Following columns are missing and haven't got any default value : {missings}\"\n )\n\n def _aux(row):\n data = {key: row[index] or defaults.get(key) for key, index in indices}\n data.update(constants)\n return data\n\n return _aux\n\n def create(self, row, context):\n raise NotImplementedError\n\n def save(self, entries, context):\n raise NotImplementedError\n\n def pre(self, rows, context):\n pass\n\n def post(self, entries, context):\n pass\n\n def line_error(self, e, row, context):\n pass\n\n def chunk_error(self, e, rows, context):\n pass\n\n def info(self, message):\n logger.info(message)\n\n def populate(self, source):\n if isinstance(source, str):\n try:\n url_info = parse_path(source)\n\n with self._get_source_file(url_info) as bytes_input:\n if url_info.path.endswith('.zip'):\n archive = ZipFile(\n bytes_input\n if bytes_input.seekable() else\n io.BytesIO(bytes_input.read())\n )\n\n with archive.open(archive.namelist()[0]) as zipped_bytes_input:\n self._populate_from_bytes(zipped_bytes_input)\n else:\n self._populate_from_bytes(bytes_input)\n except CSVPopulatorError:\n raise\n except Exception as e:\n raise self.ReadError(f'Unable to open CSV data from {source} : {e}') from e\n elif hasattr(source, '__iter__'):\n self._populate_from_lines(iter(source))\n else:\n raise ValueError('The source must be a path or an iterable.')\n\n def _populate_from_bytes(self, bytes_input):\n with io.TextIOWrapper(bytes_input) as wrapped_bytes_input:\n self._populate_from_lines(csv.reader(wrapped_bytes_input))\n\n def _populate_from_lines(self, lines):\n mapper = self._mapper(next(lines))\n context = self.Context(self.defaults)\n\n for rows in iter_as_chunk(lines, self.chunksize):\n entries = []\n\n if mapper:\n rows = [mapper(row) for row in rows]\n\n try:\n self.pre(rows, context)\n\n for row in rows:\n try:\n entries.extend(self.create(row, context))\n except Exception as e:\n self.line_error(e, row, context)\n\n context.line += 1\n\n self.save(entries, context)\n self.post(entries, context)\n except Exception as e:\n self.chunk_error(e, rows, context)\n\n def sync(self, model, entries, build_pk):\n created = []\n updated = []\n\n for t in entries:\n pk = build_pk(t)\n\n if not pk:\n created.append(t)\n else:\n t.pk = pk\n updated.append(t)\n\n with transaction.atomic():\n model.objects.bulk_create(created)\n\n # TODO: bulk_update() ?\n for entry in updated:\n entry.save(force_update=True)\n\n\nclass CSVTownPopulator(CSVPopulator):\n def __init__(self, defaults=None, chunksize=100):\n super().__init__(['title', 'zipcode', 'latitude', 'longitude', 'country'],\n defaults=defaults, chunksize=chunksize,\n )\n\n def line_error(self, e, row, context):\n logger.error(' invalid data (line %d) : %s', context.line, e)\n\n def chunk_error(self, e, rows, context):\n logger.error(' invalid data chunk : %s', e)\n\n def create(self, row, context):\n zipcodes = row['zipcode'].split('-')\n\n name = row['title']\n latitude = row['latitude']\n longitude = row['longitude']\n\n slug = slugify(name)\n country = row['country']\n\n build_town = partial(Town, country=country)\n\n return [build_town(name=name,\n slug=slug,\n zipcode=zipcode,\n latitude=latitude,\n longitude=longitude,\n ) for zipcode in zipcodes\n ]\n\n def save(self, entries, context):\n get_existing_pk_by_zipcode = dict(\n Town.objects.filter(\n zipcode__in=(t.zipcode for t in entries),\n slug__in=(t.slug for t in entries),\n ).values_list('zipcode', 'pk')\n ).get\n\n self.sync(\n Town, entries,\n lambda town: get_existing_pk_by_zipcode(town.zipcode),\n )\n\n\nclass Command(BaseCommand):\n def add_arguments(self, parser):\n add_argument = parser.add_argument\n add_argument(\n '-p', '--populate', action='store_true', dest='populate',\n help='Populate addresses', default=False,\n )\n add_argument(\n '-s', '--stat', action='store_true', dest='stats',\n help='Display geolocation database stats', default=False,\n )\n add_argument(\n '-i', '--import', action='store_true', dest='import', default=False,\n help='Import towns configured in GEOLOCATION_TOWNS setting',\n )\n\n def sysout(self, message, visible):\n if visible:\n self.stdout.write(message)\n\n def syserr(self, message):\n self.stderr.write(message)\n\n def populate_addresses(self, verbosity=0):\n self.sysout('Populate geolocation information of addresses...', verbosity > 0)\n GeoAddress.populate_geoaddresses(get_address_model().objects.exclude(zipcode='', city=''))\n\n def import_town_database(self, url, defaults):\n try:\n CSVTownPopulator(defaults=defaults).populate(url)\n except Exception as e:\n self.syserr(str(e))\n\n def import_town_all(self, verbosity=0):\n self.sysout('Importing Towns database...', verbosity > 0)\n\n for url, defaults in settings.GEOLOCATION_TOWNS:\n self.sysout(url, verbosity > 1)\n self.import_town_database(url, defaults)\n\n def print_stats(self):\n self.sysout(f'{Town.objects.count()} town(s) in database.')\n\n def handle(self, *args, **options):\n populate = options.get('populate')\n stats = options.get('stats')\n imports = options.get('import')\n verbosity = options.get('verbosity')\n\n if stats:\n self.print_stats()\n\n if imports:\n self.import_town_all(verbosity)\n\n if populate:\n self.populate_addresses(verbosity)\n", "path": "creme/geolocation/management/commands/geolocation.py"}]}
| 3,660 | 228 |
gh_patches_debug_26733
|
rasdani/github-patches
|
git_diff
|
pytorch__vision-2696
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Reading PNG/JPG images into a torch::tensor and saving a torch::tensor to PNG/JPG in C++ without OpenCV
## 🚀 Feature
After integrating Siv3D with Libtorch (https://github.com/QuantScientist/Siv3DTorch) I am now trying to read and write images from and to Siv3D **in C++, not Python**. The way it works is:
## Motivation
In C++ I need to do the following:
1. An image is read from disk (usually using OpenCV which is easy but I am trying to avoid)
2. The image is converted to torch::tensor
3. A DL model is applied on the tensor
4. A tensor is returned from the model
5. The tensor is converted to an image for display/saving purposes.
This is one example where they used stb_image to this, avoiding the use of OpenCV.
https://github.com/prabhuomkar/pytorch-cpp/blob/master/utils/image_io/src/image_io.cpp
## Pitch
## Alternatives
For reference this is the OpenCV to Libtorch conversion utils which I use, I would like something very similiar:
```
at::Tensor matToTensor(cv::Mat frame, int h, int w, int c) {
cv::cvtColor(frame, frame, CV_BGR2RGB);
frame.convertTo(frame, CV_32FC3, 1.0f / 255.0f);
auto input_tensor = torch::from_blob(frame.data, {1, h, w, c});
input_tensor = input_tensor.permute({0, 3, 1, 2});
torch::DeviceType device_type = torch::kCPU;
// if (torch::cuda::is_available()) {
device_type = torch::kCUDA;
// }
input_tensor = input_tensor.to(device_type);
return input_tensor;
}
cv::Mat tensorToOpenCv(at::Tensor out_tensor, int h, int w, int c) {
out_tensor = out_tensor.squeeze().detach().permute({1, 2, 0});
out_tensor = out_tensor.mul(255).clamp(0, 255).to(torch::kU8);
out_tensor = out_tensor.to(torch::kCPU);
cv::Mat resultImg(h, w, CV_8UC3);
// cv::Mat resultImg(h, w, CV_8UC1);
std::memcpy((void *) resultImg.data, out_tensor.data_ptr(), sizeof(torch::kU8) * out_tensor.numel());
return resultImg;
}
```
## Additional context
I found this:"https://github.com/pytorch/vision/blob/5e4a9f6d1a2bf85137f4826dbf76e4f25986f878/torchvision/csrc/cpu/image/readpng_cpu.cpp
however, could not get any useful method out of it.
Thanks,
</issue>
<code>
[start of torchvision/io/image.py]
1 import torch
2
3 import os
4 import os.path as osp
5 import importlib.machinery
6
7 _HAS_IMAGE_OPT = False
8
9 try:
10 lib_dir = osp.join(osp.dirname(__file__), "..")
11
12 loader_details = (
13 importlib.machinery.ExtensionFileLoader,
14 importlib.machinery.EXTENSION_SUFFIXES
15 )
16
17 extfinder = importlib.machinery.FileFinder(lib_dir, loader_details) # type: ignore[arg-type]
18 ext_specs = extfinder.find_spec("image")
19 if ext_specs is not None:
20 torch.ops.load_library(ext_specs.origin)
21 _HAS_IMAGE_OPT = True
22 except (ImportError, OSError):
23 pass
24
25
26 def decode_png(input: torch.Tensor) -> torch.Tensor:
27 """
28 Decodes a PNG image into a 3 dimensional RGB Tensor.
29 The values of the output tensor are uint8 between 0 and 255.
30
31 Arguments:
32 input (Tensor[1]): a one dimensional int8 tensor containing
33 the raw bytes of the PNG image.
34
35 Returns:
36 output (Tensor[3, image_height, image_width])
37 """
38 if not isinstance(input, torch.Tensor) or input.numel() == 0 or input.ndim != 1: # type: ignore[attr-defined]
39 raise ValueError("Expected a non empty 1-dimensional tensor.")
40
41 if not input.dtype == torch.uint8:
42 raise ValueError("Expected a torch.uint8 tensor.")
43 output = torch.ops.image.decode_png(input)
44 return output
45
46
47 def read_png(path: str) -> torch.Tensor:
48 """
49 Reads a PNG image into a 3 dimensional RGB Tensor.
50 The values of the output tensor are uint8 between 0 and 255.
51
52 Arguments:
53 path (str): path of the PNG image.
54
55 Returns:
56 output (Tensor[3, image_height, image_width])
57 """
58 if not os.path.isfile(path):
59 raise ValueError("Expected a valid file path.")
60
61 size = os.path.getsize(path)
62 if size == 0:
63 raise ValueError("Expected a non empty file.")
64 data = torch.from_file(path, dtype=torch.uint8, size=size)
65 return decode_png(data)
66
67
68 def decode_jpeg(input: torch.Tensor) -> torch.Tensor:
69 """
70 Decodes a JPEG image into a 3 dimensional RGB Tensor.
71 The values of the output tensor are uint8 between 0 and 255.
72 Arguments:
73 input (Tensor[1]): a one dimensional int8 tensor containing
74 the raw bytes of the JPEG image.
75 Returns:
76 output (Tensor[3, image_height, image_width])
77 """
78 if not isinstance(input, torch.Tensor) or len(input) == 0 or input.ndim != 1: # type: ignore[attr-defined]
79 raise ValueError("Expected a non empty 1-dimensional tensor.")
80
81 if not input.dtype == torch.uint8:
82 raise ValueError("Expected a torch.uint8 tensor.")
83
84 output = torch.ops.image.decode_jpeg(input)
85 return output
86
87
88 def read_jpeg(path: str) -> torch.Tensor:
89 """
90 Reads a JPEG image into a 3 dimensional RGB Tensor.
91 The values of the output tensor are uint8 between 0 and 255.
92 Arguments:
93 path (str): path of the JPEG image.
94 Returns:
95 output (Tensor[3, image_height, image_width])
96 """
97 if not os.path.isfile(path):
98 raise ValueError("Expected a valid file path.")
99
100 size = os.path.getsize(path)
101 if size == 0:
102 raise ValueError("Expected a non empty file.")
103 data = torch.from_file(path, dtype=torch.uint8, size=size)
104 return decode_jpeg(data)
105
[end of torchvision/io/image.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/torchvision/io/image.py b/torchvision/io/image.py
--- a/torchvision/io/image.py
+++ b/torchvision/io/image.py
@@ -102,3 +102,42 @@
raise ValueError("Expected a non empty file.")
data = torch.from_file(path, dtype=torch.uint8, size=size)
return decode_jpeg(data)
+
+
+def encode_jpeg(input: torch.Tensor, quality: int = 75) -> torch.Tensor:
+ """
+ Takes an input tensor in CHW layout (or HW in the case of grayscale images)
+ and returns a buffer with the contents of its corresponding JPEG file.
+ Arguments:
+ input (Tensor[channels, image_height, image_width]): int8 image tensor
+ of `c` channels, where `c` must be 1 or 3.
+ quality (int): Quality of the resulting JPEG file, it must be a number
+ between 1 and 100. Default: 75
+ Returns
+ output (Tensor[1]): A one dimensional int8 tensor that contains the raw
+ bytes of the JPEG file.
+ """
+ if quality < 1 or quality > 100:
+ raise ValueError('Image quality should be a positive number '
+ 'between 1 and 100')
+
+ output = torch.ops.image.encode_jpeg(input, quality)
+ return output
+
+
+def write_jpeg(input: torch.Tensor, filename: str, quality: int = 75):
+ """
+ Takes an input tensor in CHW layout (or HW in the case of grayscale images)
+ and saves it in a JPEG file.
+ Arguments:
+ input (Tensor[channels, image_height, image_width]): int8 image tensor
+ of `c` channels, where `c` must be 1 or 3.
+ filename (str): Path to save the image.
+ quality (int): Quality of the resulting JPEG file, it must be a number
+ between 1 and 100. Default: 75
+ """
+ if quality < 1 or quality > 100:
+ raise ValueError('Image quality should be a positive number '
+ 'between 1 and 100')
+
+ torch.ops.image.write_jpeg(input, filename, quality)
|
{"golden_diff": "diff --git a/torchvision/io/image.py b/torchvision/io/image.py\n--- a/torchvision/io/image.py\n+++ b/torchvision/io/image.py\n@@ -102,3 +102,42 @@\n raise ValueError(\"Expected a non empty file.\")\n data = torch.from_file(path, dtype=torch.uint8, size=size)\n return decode_jpeg(data)\n+\n+\n+def encode_jpeg(input: torch.Tensor, quality: int = 75) -> torch.Tensor:\n+ \"\"\"\n+ Takes an input tensor in CHW layout (or HW in the case of grayscale images)\n+ and returns a buffer with the contents of its corresponding JPEG file.\n+ Arguments:\n+ input (Tensor[channels, image_height, image_width]): int8 image tensor\n+ of `c` channels, where `c` must be 1 or 3.\n+ quality (int): Quality of the resulting JPEG file, it must be a number\n+ between 1 and 100. Default: 75\n+ Returns\n+ output (Tensor[1]): A one dimensional int8 tensor that contains the raw\n+ bytes of the JPEG file.\n+ \"\"\"\n+ if quality < 1 or quality > 100:\n+ raise ValueError('Image quality should be a positive number '\n+ 'between 1 and 100')\n+\n+ output = torch.ops.image.encode_jpeg(input, quality)\n+ return output\n+\n+\n+def write_jpeg(input: torch.Tensor, filename: str, quality: int = 75):\n+ \"\"\"\n+ Takes an input tensor in CHW layout (or HW in the case of grayscale images)\n+ and saves it in a JPEG file.\n+ Arguments:\n+ input (Tensor[channels, image_height, image_width]): int8 image tensor\n+ of `c` channels, where `c` must be 1 or 3.\n+ filename (str): Path to save the image.\n+ quality (int): Quality of the resulting JPEG file, it must be a number\n+ between 1 and 100. Default: 75\n+ \"\"\"\n+ if quality < 1 or quality > 100:\n+ raise ValueError('Image quality should be a positive number '\n+ 'between 1 and 100')\n+\n+ torch.ops.image.write_jpeg(input, filename, quality)\n", "issue": "Reading PNG/JPG images into a torch::tensor and saving a torch::tensor to PNG/JPG in C++ without OpenCV \n## \ud83d\ude80 Feature\r\n\r\nAfter integrating Siv3D with Libtorch (https://github.com/QuantScientist/Siv3DTorch) I am now trying to read and write images from and to Siv3D **in C++, not Python**. The way it works is:\r\n\r\n## Motivation\r\nIn C++ I need to do the following:\r\n1. An image is read from disk (usually using OpenCV which is easy but I am trying to avoid)\r\n2. The image is converted to torch::tensor\r\n3. A DL model is applied on the tensor\r\n4. A tensor is returned from the model\r\n5. The tensor is converted to an image for display/saving purposes.\r\n\r\nThis is one example where they used stb_image to this, avoiding the use of OpenCV.\r\nhttps://github.com/prabhuomkar/pytorch-cpp/blob/master/utils/image_io/src/image_io.cpp\r\n\r\n## Pitch\r\n\r\n## Alternatives\r\n\r\nFor reference this is the OpenCV to Libtorch conversion utils which I use, I would like something very similiar:\r\n```\r\nat::Tensor matToTensor(cv::Mat frame, int h, int w, int c) {\r\n cv::cvtColor(frame, frame, CV_BGR2RGB);\r\n frame.convertTo(frame, CV_32FC3, 1.0f / 255.0f);\r\n auto input_tensor = torch::from_blob(frame.data, {1, h, w, c});\r\n input_tensor = input_tensor.permute({0, 3, 1, 2});\r\n\r\n torch::DeviceType device_type = torch::kCPU;\r\n// if (torch::cuda::is_available()) {\r\n device_type = torch::kCUDA;\r\n// }\r\n input_tensor = input_tensor.to(device_type);\r\n return input_tensor;\r\n}\r\n\r\ncv::Mat tensorToOpenCv(at::Tensor out_tensor, int h, int w, int c) {\r\n out_tensor = out_tensor.squeeze().detach().permute({1, 2, 0});\r\n out_tensor = out_tensor.mul(255).clamp(0, 255).to(torch::kU8);\r\n out_tensor = out_tensor.to(torch::kCPU);\r\n cv::Mat resultImg(h, w, CV_8UC3);\r\n // cv::Mat resultImg(h, w, CV_8UC1);\r\n std::memcpy((void *) resultImg.data, out_tensor.data_ptr(), sizeof(torch::kU8) * out_tensor.numel());\r\n return resultImg;\r\n}\r\n\r\n```\r\n## Additional context\r\n\r\nI found this:\"https://github.com/pytorch/vision/blob/5e4a9f6d1a2bf85137f4826dbf76e4f25986f878/torchvision/csrc/cpu/image/readpng_cpu.cpp\r\nhowever, could not get any useful method out of it. \r\n\r\nThanks, \r\n\r\n\r\n\r\n\n", "before_files": [{"content": "import torch\n\nimport os\nimport os.path as osp\nimport importlib.machinery\n\n_HAS_IMAGE_OPT = False\n\ntry:\n lib_dir = osp.join(osp.dirname(__file__), \"..\")\n\n loader_details = (\n importlib.machinery.ExtensionFileLoader,\n importlib.machinery.EXTENSION_SUFFIXES\n )\n\n extfinder = importlib.machinery.FileFinder(lib_dir, loader_details) # type: ignore[arg-type]\n ext_specs = extfinder.find_spec(\"image\")\n if ext_specs is not None:\n torch.ops.load_library(ext_specs.origin)\n _HAS_IMAGE_OPT = True\nexcept (ImportError, OSError):\n pass\n\n\ndef decode_png(input: torch.Tensor) -> torch.Tensor:\n \"\"\"\n Decodes a PNG image into a 3 dimensional RGB Tensor.\n The values of the output tensor are uint8 between 0 and 255.\n\n Arguments:\n input (Tensor[1]): a one dimensional int8 tensor containing\n the raw bytes of the PNG image.\n\n Returns:\n output (Tensor[3, image_height, image_width])\n \"\"\"\n if not isinstance(input, torch.Tensor) or input.numel() == 0 or input.ndim != 1: # type: ignore[attr-defined]\n raise ValueError(\"Expected a non empty 1-dimensional tensor.\")\n\n if not input.dtype == torch.uint8:\n raise ValueError(\"Expected a torch.uint8 tensor.\")\n output = torch.ops.image.decode_png(input)\n return output\n\n\ndef read_png(path: str) -> torch.Tensor:\n \"\"\"\n Reads a PNG image into a 3 dimensional RGB Tensor.\n The values of the output tensor are uint8 between 0 and 255.\n\n Arguments:\n path (str): path of the PNG image.\n\n Returns:\n output (Tensor[3, image_height, image_width])\n \"\"\"\n if not os.path.isfile(path):\n raise ValueError(\"Expected a valid file path.\")\n\n size = os.path.getsize(path)\n if size == 0:\n raise ValueError(\"Expected a non empty file.\")\n data = torch.from_file(path, dtype=torch.uint8, size=size)\n return decode_png(data)\n\n\ndef decode_jpeg(input: torch.Tensor) -> torch.Tensor:\n \"\"\"\n Decodes a JPEG image into a 3 dimensional RGB Tensor.\n The values of the output tensor are uint8 between 0 and 255.\n Arguments:\n input (Tensor[1]): a one dimensional int8 tensor containing\n the raw bytes of the JPEG image.\n Returns:\n output (Tensor[3, image_height, image_width])\n \"\"\"\n if not isinstance(input, torch.Tensor) or len(input) == 0 or input.ndim != 1: # type: ignore[attr-defined]\n raise ValueError(\"Expected a non empty 1-dimensional tensor.\")\n\n if not input.dtype == torch.uint8:\n raise ValueError(\"Expected a torch.uint8 tensor.\")\n\n output = torch.ops.image.decode_jpeg(input)\n return output\n\n\ndef read_jpeg(path: str) -> torch.Tensor:\n \"\"\"\n Reads a JPEG image into a 3 dimensional RGB Tensor.\n The values of the output tensor are uint8 between 0 and 255.\n Arguments:\n path (str): path of the JPEG image.\n Returns:\n output (Tensor[3, image_height, image_width])\n \"\"\"\n if not os.path.isfile(path):\n raise ValueError(\"Expected a valid file path.\")\n\n size = os.path.getsize(path)\n if size == 0:\n raise ValueError(\"Expected a non empty file.\")\n data = torch.from_file(path, dtype=torch.uint8, size=size)\n return decode_jpeg(data)\n", "path": "torchvision/io/image.py"}]}
| 2,178 | 525 |
gh_patches_debug_33263
|
rasdani/github-patches
|
git_diff
|
kserve__kserve-2684
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
GPL License Violation in the kserve python package
The version of the `kserve` package that is currently on PyPI (version `0.10`) violates the GPL license because it depends on [`table-logger`](https://github.com/AleksTk/table-logger), distributed under GPLv2 (you'll see that the library is now MIT, the author updated the license just a few days ago, but hasn't released a new version with the new license yet). No GPLv2 packages should be imported given that `kserve` has an Apache 2 license.
This was recently fixed by this PR https://github.com/kserve/kserve/pull/2673, which accidentally resolved the issue by replacing `table-logger` with `tabulate` (MIT License)
cc @yuzisun @cliveseldon @jinchihe @ellistarn
Is it possible to quickly release a patch release `0.10.1` to include the above patch and make sure `kserve` is compliant with the Apache license? As it stands, any distribution and vendor using `kserve` is liable for a license violation.
</issue>
<code>
[start of python/kserve/kserve/api/watch.py]
1 # Copyright 2021 The KServe Authors.
2 #
3 # Licensed under the Apache License, Version 2.0 (the "License");
4 # you may not use this file except in compliance with the License.
5 # You may obtain a copy of the License at
6 #
7 # http://www.apache.org/licenses/LICENSE-2.0
8 #
9 # Unless required by applicable law or agreed to in writing, software
10 # distributed under the License is distributed on an "AS IS" BASIS,
11 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12 # See the License for the specific language governing permissions and
13 # limitations under the License.
14
15 import time
16 from kubernetes import client
17 from kubernetes import watch as k8s_watch
18 from table_logger import TableLogger
19
20 from ..constants import constants
21 from ..utils import utils
22
23
24 def isvc_watch(name=None, namespace=None, timeout_seconds=600, generation=0):
25 """Watch the created or patched InferenceService in the specified namespace"""
26
27 if namespace is None:
28 namespace = utils.get_default_target_namespace()
29
30 tbl = TableLogger(
31 columns='NAME,READY,PREV,LATEST,URL',
32 colwidth={'NAME': 20, 'READY': 10, 'PREV': 25, 'LATEST': 25, 'URL': 65},
33 border=False)
34
35 stream = k8s_watch.Watch().stream(
36 client.CustomObjectsApi().list_namespaced_custom_object,
37 constants.KSERVE_GROUP,
38 constants.KSERVE_V1BETA1_VERSION,
39 namespace,
40 constants.KSERVE_PLURAL,
41 timeout_seconds=timeout_seconds)
42
43 for event in stream:
44 isvc = event['object']
45 isvc_name = isvc['metadata']['name']
46 if name and name != isvc_name:
47 continue
48 else:
49 status = 'Unknown'
50 if isvc.get('status', ''):
51 url = isvc['status'].get('url', '')
52 traffic = isvc['status'].get('components', {}).get(
53 'predictor', {}).get('traffic', [])
54 traffic_percent = 100
55 if constants.OBSERVED_GENERATION in isvc['status']:
56 observed_generation = isvc['status'][constants.OBSERVED_GENERATION]
57 for t in traffic:
58 if t["latestRevision"]:
59 traffic_percent = t["percent"]
60
61 if generation != 0 and observed_generation != generation:
62 continue
63 for condition in isvc['status'].get('conditions', {}):
64 if condition.get('type', '') == 'Ready':
65 status = condition.get('status', 'Unknown')
66 tbl(isvc_name, status, 100-traffic_percent, traffic_percent, url)
67 if status == 'True':
68 break
69
70 else:
71 tbl(isvc_name, status, '', '', '')
72 # Sleep 2 to avoid status section is not generated within a very short time.
73 time.sleep(2)
74 continue
75
[end of python/kserve/kserve/api/watch.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/python/kserve/kserve/api/watch.py b/python/kserve/kserve/api/watch.py
--- a/python/kserve/kserve/api/watch.py
+++ b/python/kserve/kserve/api/watch.py
@@ -13,9 +13,10 @@
# limitations under the License.
import time
+
from kubernetes import client
from kubernetes import watch as k8s_watch
-from table_logger import TableLogger
+from tabulate import tabulate
from ..constants import constants
from ..utils import utils
@@ -27,10 +28,8 @@
if namespace is None:
namespace = utils.get_default_target_namespace()
- tbl = TableLogger(
- columns='NAME,READY,PREV,LATEST,URL',
- colwidth={'NAME': 20, 'READY': 10, 'PREV': 25, 'LATEST': 25, 'URL': 65},
- border=False)
+ headers = ['NAME', 'READY', 'PREV', 'LATEST', 'URL']
+ table_fmt = 'plain'
stream = k8s_watch.Watch().stream(
client.CustomObjectsApi().list_namespaced_custom_object,
@@ -63,12 +62,13 @@
for condition in isvc['status'].get('conditions', {}):
if condition.get('type', '') == 'Ready':
status = condition.get('status', 'Unknown')
- tbl(isvc_name, status, 100-traffic_percent, traffic_percent, url)
+ print(tabulate([[isvc_name, status, 100 - traffic_percent, traffic_percent, url]],
+ headers=headers, tablefmt=table_fmt))
if status == 'True':
break
else:
- tbl(isvc_name, status, '', '', '')
+ print(tabulate([[isvc_name, status, '', '', '']], headers=headers, tablefmt=table_fmt))
# Sleep 2 to avoid status section is not generated within a very short time.
time.sleep(2)
continue
|
{"golden_diff": "diff --git a/python/kserve/kserve/api/watch.py b/python/kserve/kserve/api/watch.py\n--- a/python/kserve/kserve/api/watch.py\n+++ b/python/kserve/kserve/api/watch.py\n@@ -13,9 +13,10 @@\n # limitations under the License.\n \n import time\n+\n from kubernetes import client\n from kubernetes import watch as k8s_watch\n-from table_logger import TableLogger\n+from tabulate import tabulate\n \n from ..constants import constants\n from ..utils import utils\n@@ -27,10 +28,8 @@\n if namespace is None:\n namespace = utils.get_default_target_namespace()\n \n- tbl = TableLogger(\n- columns='NAME,READY,PREV,LATEST,URL',\n- colwidth={'NAME': 20, 'READY': 10, 'PREV': 25, 'LATEST': 25, 'URL': 65},\n- border=False)\n+ headers = ['NAME', 'READY', 'PREV', 'LATEST', 'URL']\n+ table_fmt = 'plain'\n \n stream = k8s_watch.Watch().stream(\n client.CustomObjectsApi().list_namespaced_custom_object,\n@@ -63,12 +62,13 @@\n for condition in isvc['status'].get('conditions', {}):\n if condition.get('type', '') == 'Ready':\n status = condition.get('status', 'Unknown')\n- tbl(isvc_name, status, 100-traffic_percent, traffic_percent, url)\n+ print(tabulate([[isvc_name, status, 100 - traffic_percent, traffic_percent, url]],\n+ headers=headers, tablefmt=table_fmt))\n if status == 'True':\n break\n \n else:\n- tbl(isvc_name, status, '', '', '')\n+ print(tabulate([[isvc_name, status, '', '', '']], headers=headers, tablefmt=table_fmt))\n # Sleep 2 to avoid status section is not generated within a very short time.\n time.sleep(2)\n continue\n", "issue": "GPL License Violation in the kserve python package\nThe version of the `kserve` package that is currently on PyPI (version `0.10`) violates the GPL license because it depends on [`table-logger`](https://github.com/AleksTk/table-logger), distributed under GPLv2 (you'll see that the library is now MIT, the author updated the license just a few days ago, but hasn't released a new version with the new license yet). No GPLv2 packages should be imported given that `kserve` has an Apache 2 license.\r\n\r\n\r\nThis was recently fixed by this PR https://github.com/kserve/kserve/pull/2673, which accidentally resolved the issue by replacing `table-logger` with `tabulate` (MIT License)\r\n\r\ncc @yuzisun @cliveseldon @jinchihe @ellistarn \r\n\r\nIs it possible to quickly release a patch release `0.10.1` to include the above patch and make sure `kserve` is compliant with the Apache license? As it stands, any distribution and vendor using `kserve` is liable for a license violation.\r\n\r\n\r\n\r\n\n", "before_files": [{"content": "# Copyright 2021 The KServe Authors.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nimport time\nfrom kubernetes import client\nfrom kubernetes import watch as k8s_watch\nfrom table_logger import TableLogger\n\nfrom ..constants import constants\nfrom ..utils import utils\n\n\ndef isvc_watch(name=None, namespace=None, timeout_seconds=600, generation=0):\n \"\"\"Watch the created or patched InferenceService in the specified namespace\"\"\"\n\n if namespace is None:\n namespace = utils.get_default_target_namespace()\n\n tbl = TableLogger(\n columns='NAME,READY,PREV,LATEST,URL',\n colwidth={'NAME': 20, 'READY': 10, 'PREV': 25, 'LATEST': 25, 'URL': 65},\n border=False)\n\n stream = k8s_watch.Watch().stream(\n client.CustomObjectsApi().list_namespaced_custom_object,\n constants.KSERVE_GROUP,\n constants.KSERVE_V1BETA1_VERSION,\n namespace,\n constants.KSERVE_PLURAL,\n timeout_seconds=timeout_seconds)\n\n for event in stream:\n isvc = event['object']\n isvc_name = isvc['metadata']['name']\n if name and name != isvc_name:\n continue\n else:\n status = 'Unknown'\n if isvc.get('status', ''):\n url = isvc['status'].get('url', '')\n traffic = isvc['status'].get('components', {}).get(\n 'predictor', {}).get('traffic', [])\n traffic_percent = 100\n if constants.OBSERVED_GENERATION in isvc['status']:\n observed_generation = isvc['status'][constants.OBSERVED_GENERATION]\n for t in traffic:\n if t[\"latestRevision\"]:\n traffic_percent = t[\"percent\"]\n\n if generation != 0 and observed_generation != generation:\n continue\n for condition in isvc['status'].get('conditions', {}):\n if condition.get('type', '') == 'Ready':\n status = condition.get('status', 'Unknown')\n tbl(isvc_name, status, 100-traffic_percent, traffic_percent, url)\n if status == 'True':\n break\n\n else:\n tbl(isvc_name, status, '', '', '')\n # Sleep 2 to avoid status section is not generated within a very short time.\n time.sleep(2)\n continue\n", "path": "python/kserve/kserve/api/watch.py"}]}
| 1,562 | 450 |
gh_patches_debug_13710
|
rasdani/github-patches
|
git_diff
|
openfun__marsha-2238
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Properly clause websocket connection on video/webinar deletion
## Bug Report
**Problematic Behavior**
When a video/webinar is deleted, a websocket connection is still open. The dashboard is unmounted by react so it tries to close it but an error is thrown because on closing the connection, the weboscket tries to fetch the attached live session in database but this one does not exists anymore.
**Expected behavior/code**
When a video/webinar is deleted, the websocket should be properly closed.
**Steps to Reproduce**
1. Create a webinar
2. Delete it
**Additional context/Screenshots**
https://gip-fun-mooc.sentry.io/share/issue/668071c3736a446a890af2b99df1c6c6/
</issue>
<code>
[start of src/backend/marsha/websocket/consumers/video.py]
1 """Video consumer module"""
2 from urllib.parse import parse_qs
3
4 from django.db.models import Q
5
6 from channels.db import database_sync_to_async
7 from channels.generic.websocket import AsyncJsonWebsocketConsumer
8
9 from marsha.core.models import (
10 ADMINISTRATOR,
11 INSTRUCTOR,
12 SharedLiveMedia,
13 Thumbnail,
14 TimedTextTrack,
15 Video,
16 )
17 from marsha.core.permissions import IsTokenAdmin, IsTokenInstructor
18 from marsha.core.services import live_session as LiveSessionServices
19 from marsha.core.simple_jwt.tokens import ResourceAccessToken, UserAccessToken
20 from marsha.websocket import defaults
21
22
23 class VideoConsumer(AsyncJsonWebsocketConsumer):
24 """Video consumer."""
25
26 room_group_name = None
27 is_connected = False
28
29 def __get_video_id(self):
30 return self.scope["url_route"]["kwargs"]["video_id"]
31
32 async def _check_permissions(self):
33 """
34 Check if the user has the required permissions.
35
36 Raises:
37 ConnectionRefusedError: if the user does not have the required permissions.
38 """
39 token = self.scope["token"]
40 if token is None:
41 raise ConnectionRefusedError()
42
43 # Check permissions, MUST be the same as in the `retrieve` method
44 # of the Video API view set.
45
46 if isinstance(token, ResourceAccessToken):
47 # With LTI: anyone with a valid token for the video can access
48 if token.payload.get("resource_id") != self.__get_video_id():
49 raise ConnectionRefusedError()
50
51 elif isinstance(token, UserAccessToken):
52 # With standalone site, only playlist admin or organization admin can access
53 if not await self._user_has_playlist_or_organization_admin_role(
54 token.payload.get("user_id")
55 ):
56 raise ConnectionRefusedError()
57
58 else:
59 raise RuntimeError("This should not happen")
60
61 async def connect(self):
62 """
63 Manage connection to this consumer.
64 During handshake it is not possible to close the websocket with a specific
65 code. To do that we must accept first the connection and then close it with the code we
66 want
67 """
68 try:
69 await self._check_permissions()
70 await self._check_video_exists()
71 if not await self._is_admin():
72 live_session = await self.retrieve_live_session()
73 await self.update_live_session_with_channel_name(
74 live_session=live_session
75 )
76 except ConnectionRefusedError:
77 await self.accept()
78 return await self.close(code=4003)
79
80 self.room_group_name = await self._get_room_name()
81 # Join room group
82 await self.channel_layer.group_add(self.room_group_name, self.channel_name)
83
84 await self.accept()
85 self.is_connected = True
86
87 async def _check_video_exists(self):
88 """Close the room if the video does not exists."""
89 if not await self._video_exists():
90 raise ConnectionRefusedError()
91
92 @database_sync_to_async
93 def _video_exists(self):
94 """Return if a video exists in database or not."""
95 return Video.objects.filter(pk=self.__get_video_id()).exists()
96
97 @database_sync_to_async
98 def _user_has_playlist_or_organization_admin_role(self, user_id):
99 """Return if the user belongs to the video playlist admin or organization admin."""
100 return Video.objects.filter(
101 Q(pk=self.__get_video_id())
102 & (
103 Q(
104 playlist__user_accesses__user_id=user_id,
105 playlist__user_accesses__role__in=[ADMINISTRATOR, INSTRUCTOR],
106 )
107 | Q(
108 playlist__organization__user_accesses__user_id=user_id,
109 playlist__organization__user_accesses__role=ADMINISTRATOR,
110 )
111 )
112 ).exists()
113
114 @database_sync_to_async
115 def retrieve_live_session(self):
116 """Guess a live_session from the token and create it id not present."""
117 token = self.scope["token"]
118 if LiveSessionServices.is_lti_token(token):
119 live_session, _ = LiveSessionServices.get_livesession_from_lti(token)
120 else:
121 query_string = parse_qs(self.scope["query_string"])
122 if b"anonymous_id" not in query_string:
123 raise ConnectionRefusedError()
124 live_session, _ = LiveSessionServices.get_livesession_from_anonymous_id(
125 anonymous_id=query_string[b"anonymous_id"][0].decode("utf-8"),
126 video_id=token.payload["resource_id"],
127 )
128
129 return live_session
130
131 @database_sync_to_async
132 def update_live_session_with_channel_name(self, live_session):
133 """Update the live_session with the current channel_name."""
134 live_session.channel_name = self.channel_name
135 live_session.save()
136
137 @database_sync_to_async
138 def reset_live_session(self, live_session):
139 """Reset to None the live_session channel_name."""
140 live_session.channel_name = None
141 live_session.save()
142
143 async def _is_admin(self):
144 """Check if the connected user has admin permissions."""
145 token = self.scope["token"]
146 if not token:
147 return False
148
149 if isinstance(token, ResourceAccessToken):
150 return IsTokenInstructor().check_role(token) or IsTokenAdmin().check_role(
151 token
152 )
153
154 if isinstance(token, UserAccessToken):
155 return await self._user_has_playlist_or_organization_admin_role(
156 token.payload.get("user_id"),
157 )
158
159 raise RuntimeError("Should not be called please check the code.", type(token))
160
161 async def _get_room_name(self):
162 """Generate the room name the user is connected on depending its permissions."""
163 if await self._is_admin():
164 return defaults.VIDEO_ADMIN_ROOM_NAME.format(video_id=self.__get_video_id())
165
166 return defaults.VIDEO_ROOM_NAME.format(video_id=self.__get_video_id())
167
168 # pylint: disable=unused-argument
169 async def disconnect(self, code):
170 """Manage disconnection to this consumer."""
171 # If connection was previously aborted nothing to do
172 if not self.is_connected:
173 return
174
175 # Leave room group
176 await self.channel_layer.group_discard(self.room_group_name, self.channel_name)
177 if not await self._is_admin():
178 live_session = await self.retrieve_live_session()
179 await self.reset_live_session(live_session)
180
181 async def video_updated(self, event):
182 """Listener for the video_updated event."""
183 message = {"type": Video.RESOURCE_NAME, "resource": event["video"]}
184 await self.send_json(message)
185
186 async def thumbnail_updated(self, event):
187 """Listener for the thumbnail updated event."""
188 message = {"type": Thumbnail.RESOURCE_NAME, "resource": event["thumbnail"]}
189 await self.send_json(message)
190
191 async def timed_text_track_updated(self, event):
192 """Listener for the timed text track updated event."""
193 message = {
194 "type": TimedTextTrack.RESOURCE_NAME,
195 "resource": event["timed_text_track"],
196 }
197 await self.send_json(message)
198
199 async def shared_live_media_updated(self, event):
200 """Listener for the shared_live_media updated event."""
201 message = {
202 "type": SharedLiveMedia.RESOURCE_NAME,
203 "resource": event["shared_live_media"],
204 }
205 await self.send_json(message)
206
[end of src/backend/marsha/websocket/consumers/video.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/src/backend/marsha/websocket/consumers/video.py b/src/backend/marsha/websocket/consumers/video.py
--- a/src/backend/marsha/websocket/consumers/video.py
+++ b/src/backend/marsha/websocket/consumers/video.py
@@ -175,8 +175,12 @@
# Leave room group
await self.channel_layer.group_discard(self.room_group_name, self.channel_name)
if not await self._is_admin():
- live_session = await self.retrieve_live_session()
- await self.reset_live_session(live_session)
+ try:
+ live_session = await self.retrieve_live_session()
+ await self.reset_live_session(live_session)
+ except ConnectionRefusedError:
+ # No live session found, nothing to do
+ pass
async def video_updated(self, event):
"""Listener for the video_updated event."""
|
{"golden_diff": "diff --git a/src/backend/marsha/websocket/consumers/video.py b/src/backend/marsha/websocket/consumers/video.py\n--- a/src/backend/marsha/websocket/consumers/video.py\n+++ b/src/backend/marsha/websocket/consumers/video.py\n@@ -175,8 +175,12 @@\n # Leave room group\n await self.channel_layer.group_discard(self.room_group_name, self.channel_name)\n if not await self._is_admin():\n- live_session = await self.retrieve_live_session()\n- await self.reset_live_session(live_session)\n+ try:\n+ live_session = await self.retrieve_live_session()\n+ await self.reset_live_session(live_session)\n+ except ConnectionRefusedError:\n+ # No live session found, nothing to do\n+ pass\n \n async def video_updated(self, event):\n \"\"\"Listener for the video_updated event.\"\"\"\n", "issue": "Properly clause websocket connection on video/webinar deletion\n## Bug Report\r\n\r\n**Problematic Behavior**\r\n\r\nWhen a video/webinar is deleted, a websocket connection is still open. The dashboard is unmounted by react so it tries to close it but an error is thrown because on closing the connection, the weboscket tries to fetch the attached live session in database but this one does not exists anymore.\r\n\r\n**Expected behavior/code**\r\n\r\nWhen a video/webinar is deleted, the websocket should be properly closed.\r\n\r\n**Steps to Reproduce**\r\n1. Create a webinar\r\n2. Delete it\r\n\r\n\r\n**Additional context/Screenshots**\r\n\r\nhttps://gip-fun-mooc.sentry.io/share/issue/668071c3736a446a890af2b99df1c6c6/\r\n\n", "before_files": [{"content": "\"\"\"Video consumer module\"\"\"\nfrom urllib.parse import parse_qs\n\nfrom django.db.models import Q\n\nfrom channels.db import database_sync_to_async\nfrom channels.generic.websocket import AsyncJsonWebsocketConsumer\n\nfrom marsha.core.models import (\n ADMINISTRATOR,\n INSTRUCTOR,\n SharedLiveMedia,\n Thumbnail,\n TimedTextTrack,\n Video,\n)\nfrom marsha.core.permissions import IsTokenAdmin, IsTokenInstructor\nfrom marsha.core.services import live_session as LiveSessionServices\nfrom marsha.core.simple_jwt.tokens import ResourceAccessToken, UserAccessToken\nfrom marsha.websocket import defaults\n\n\nclass VideoConsumer(AsyncJsonWebsocketConsumer):\n \"\"\"Video consumer.\"\"\"\n\n room_group_name = None\n is_connected = False\n\n def __get_video_id(self):\n return self.scope[\"url_route\"][\"kwargs\"][\"video_id\"]\n\n async def _check_permissions(self):\n \"\"\"\n Check if the user has the required permissions.\n\n Raises:\n ConnectionRefusedError: if the user does not have the required permissions.\n \"\"\"\n token = self.scope[\"token\"]\n if token is None:\n raise ConnectionRefusedError()\n\n # Check permissions, MUST be the same as in the `retrieve` method\n # of the Video API view set.\n\n if isinstance(token, ResourceAccessToken):\n # With LTI: anyone with a valid token for the video can access\n if token.payload.get(\"resource_id\") != self.__get_video_id():\n raise ConnectionRefusedError()\n\n elif isinstance(token, UserAccessToken):\n # With standalone site, only playlist admin or organization admin can access\n if not await self._user_has_playlist_or_organization_admin_role(\n token.payload.get(\"user_id\")\n ):\n raise ConnectionRefusedError()\n\n else:\n raise RuntimeError(\"This should not happen\")\n\n async def connect(self):\n \"\"\"\n Manage connection to this consumer.\n During handshake it is not possible to close the websocket with a specific\n code. To do that we must accept first the connection and then close it with the code we\n want\n \"\"\"\n try:\n await self._check_permissions()\n await self._check_video_exists()\n if not await self._is_admin():\n live_session = await self.retrieve_live_session()\n await self.update_live_session_with_channel_name(\n live_session=live_session\n )\n except ConnectionRefusedError:\n await self.accept()\n return await self.close(code=4003)\n\n self.room_group_name = await self._get_room_name()\n # Join room group\n await self.channel_layer.group_add(self.room_group_name, self.channel_name)\n\n await self.accept()\n self.is_connected = True\n\n async def _check_video_exists(self):\n \"\"\"Close the room if the video does not exists.\"\"\"\n if not await self._video_exists():\n raise ConnectionRefusedError()\n\n @database_sync_to_async\n def _video_exists(self):\n \"\"\"Return if a video exists in database or not.\"\"\"\n return Video.objects.filter(pk=self.__get_video_id()).exists()\n\n @database_sync_to_async\n def _user_has_playlist_or_organization_admin_role(self, user_id):\n \"\"\"Return if the user belongs to the video playlist admin or organization admin.\"\"\"\n return Video.objects.filter(\n Q(pk=self.__get_video_id())\n & (\n Q(\n playlist__user_accesses__user_id=user_id,\n playlist__user_accesses__role__in=[ADMINISTRATOR, INSTRUCTOR],\n )\n | Q(\n playlist__organization__user_accesses__user_id=user_id,\n playlist__organization__user_accesses__role=ADMINISTRATOR,\n )\n )\n ).exists()\n\n @database_sync_to_async\n def retrieve_live_session(self):\n \"\"\"Guess a live_session from the token and create it id not present.\"\"\"\n token = self.scope[\"token\"]\n if LiveSessionServices.is_lti_token(token):\n live_session, _ = LiveSessionServices.get_livesession_from_lti(token)\n else:\n query_string = parse_qs(self.scope[\"query_string\"])\n if b\"anonymous_id\" not in query_string:\n raise ConnectionRefusedError()\n live_session, _ = LiveSessionServices.get_livesession_from_anonymous_id(\n anonymous_id=query_string[b\"anonymous_id\"][0].decode(\"utf-8\"),\n video_id=token.payload[\"resource_id\"],\n )\n\n return live_session\n\n @database_sync_to_async\n def update_live_session_with_channel_name(self, live_session):\n \"\"\"Update the live_session with the current channel_name.\"\"\"\n live_session.channel_name = self.channel_name\n live_session.save()\n\n @database_sync_to_async\n def reset_live_session(self, live_session):\n \"\"\"Reset to None the live_session channel_name.\"\"\"\n live_session.channel_name = None\n live_session.save()\n\n async def _is_admin(self):\n \"\"\"Check if the connected user has admin permissions.\"\"\"\n token = self.scope[\"token\"]\n if not token:\n return False\n\n if isinstance(token, ResourceAccessToken):\n return IsTokenInstructor().check_role(token) or IsTokenAdmin().check_role(\n token\n )\n\n if isinstance(token, UserAccessToken):\n return await self._user_has_playlist_or_organization_admin_role(\n token.payload.get(\"user_id\"),\n )\n\n raise RuntimeError(\"Should not be called please check the code.\", type(token))\n\n async def _get_room_name(self):\n \"\"\"Generate the room name the user is connected on depending its permissions.\"\"\"\n if await self._is_admin():\n return defaults.VIDEO_ADMIN_ROOM_NAME.format(video_id=self.__get_video_id())\n\n return defaults.VIDEO_ROOM_NAME.format(video_id=self.__get_video_id())\n\n # pylint: disable=unused-argument\n async def disconnect(self, code):\n \"\"\"Manage disconnection to this consumer.\"\"\"\n # If connection was previously aborted nothing to do\n if not self.is_connected:\n return\n\n # Leave room group\n await self.channel_layer.group_discard(self.room_group_name, self.channel_name)\n if not await self._is_admin():\n live_session = await self.retrieve_live_session()\n await self.reset_live_session(live_session)\n\n async def video_updated(self, event):\n \"\"\"Listener for the video_updated event.\"\"\"\n message = {\"type\": Video.RESOURCE_NAME, \"resource\": event[\"video\"]}\n await self.send_json(message)\n\n async def thumbnail_updated(self, event):\n \"\"\"Listener for the thumbnail updated event.\"\"\"\n message = {\"type\": Thumbnail.RESOURCE_NAME, \"resource\": event[\"thumbnail\"]}\n await self.send_json(message)\n\n async def timed_text_track_updated(self, event):\n \"\"\"Listener for the timed text track updated event.\"\"\"\n message = {\n \"type\": TimedTextTrack.RESOURCE_NAME,\n \"resource\": event[\"timed_text_track\"],\n }\n await self.send_json(message)\n\n async def shared_live_media_updated(self, event):\n \"\"\"Listener for the shared_live_media updated event.\"\"\"\n message = {\n \"type\": SharedLiveMedia.RESOURCE_NAME,\n \"resource\": event[\"shared_live_media\"],\n }\n await self.send_json(message)\n", "path": "src/backend/marsha/websocket/consumers/video.py"}]}
| 2,758 | 197 |
gh_patches_debug_27026
|
rasdani/github-patches
|
git_diff
|
DataDog__integrations-core-619
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
postfix integration should not require sudo to root
Reading the source code to integrations-core/postfix/check.py I note that it does a sudo to root to run the find command.
This is noted in the docs / comments :
> WARNING: the user that dd-agent runs as must have sudo access for the 'find' command
> --
> | sudo access is not required when running dd-agent as root (not recommended)
> |
> | example /etc/sudoers entry:
> | dd-agent ALL=(ALL) NOPASSWD:/usr/bin/find /var/spool/postfix* -type f
root should not be required here - postfix user should be sufficient. That would be combined with a '-u postfix' on line 64's sudo command to allow this to work.
This is a concern because find has a -exec parameter and your command list has a wildcard in it - this could be used to run arbitrary commands as root if the dd-agent user is compromised.
</issue>
<code>
[start of postfix/check.py]
1 # (C) Datadog, Inc. 2013-2016
2 # (C) Josiah C Webb <[email protected]> 2013
3 # All rights reserved
4 # Licensed under Simplified BSD License (see LICENSE)
5
6 # stdlib
7 import os
8
9 # project
10 from checks import AgentCheck
11 from utils.subprocess_output import get_subprocess_output
12
13 class PostfixCheck(AgentCheck):
14 """This check provides metrics on the number of messages in a given postfix queue
15
16 WARNING: the user that dd-agent runs as must have sudo access for the 'find' command
17 sudo access is not required when running dd-agent as root (not recommended)
18
19 example /etc/sudoers entry:
20 dd-agent ALL=(ALL) NOPASSWD:/usr/bin/find /var/spool/postfix* -type f
21
22 YAML config options:
23 "directory" - the value of 'postconf -h queue_directory'
24 "queues" - the postfix mail queues you would like to get message count totals for
25 """
26 def check(self, instance):
27 config = self._get_config(instance)
28
29 directory = config['directory']
30 queues = config['queues']
31 tags = config['tags']
32
33 self._get_queue_count(directory, queues, tags)
34
35 def _get_config(self, instance):
36 directory = instance.get('directory', None)
37 queues = instance.get('queues', None)
38 tags = instance.get('tags', [])
39 if not queues or not directory:
40 raise Exception('missing required yaml config entry')
41
42 instance_config = {
43 'directory': directory,
44 'queues': queues,
45 'tags': tags,
46 }
47
48 return instance_config
49
50 def _get_queue_count(self, directory, queues, tags):
51 for queue in queues:
52 queue_path = os.path.join(directory, queue)
53 if not os.path.exists(queue_path):
54 raise Exception('%s does not exist' % queue_path)
55
56 count = 0
57 if os.geteuid() == 0:
58 # dd-agent is running as root (not recommended)
59 count = sum(len(files) for root, dirs, files in os.walk(queue_path))
60 else:
61 # can dd-agent user run sudo?
62 test_sudo = os.system('setsid sudo -l < /dev/null')
63 if test_sudo == 0:
64 output, _, _ = get_subprocess_output(['sudo', 'find', queue_path, '-type', 'f'], self.log, False)
65 count = len(output.splitlines())
66 else:
67 raise Exception('The dd-agent user does not have sudo access')
68
69 # emit an individually tagged metric
70 self.gauge('postfix.queue.size', count, tags=tags + ['queue:%s' % queue, 'instance:%s' % os.path.basename(directory)])
71
72 # these can be retrieved in a single graph statement
73 # for example:
74 # sum:postfix.queue.size{instance:postfix-2,queue:incoming,host:hostname.domain.tld}
75
[end of postfix/check.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/postfix/check.py b/postfix/check.py
--- a/postfix/check.py
+++ b/postfix/check.py
@@ -17,7 +17,9 @@
sudo access is not required when running dd-agent as root (not recommended)
example /etc/sudoers entry:
- dd-agent ALL=(ALL) NOPASSWD:/usr/bin/find /var/spool/postfix* -type f
+ dd-agent ALL=(postfix) NOPASSWD:/usr/bin/find /var/spool/postfix/incoming -type f
+ dd-agent ALL=(postfix) NOPASSWD:/usr/bin/find /var/spool/postfix/active -type f
+ dd-agent ALL=(postfix) NOPASSWD:/usr/bin/find /var/spool/postfix/deferred -type f
YAML config options:
"directory" - the value of 'postconf -h queue_directory'
@@ -61,7 +63,9 @@
# can dd-agent user run sudo?
test_sudo = os.system('setsid sudo -l < /dev/null')
if test_sudo == 0:
- output, _, _ = get_subprocess_output(['sudo', 'find', queue_path, '-type', 'f'], self.log, False)
+ # default to `root` for backward compatibility
+ postfix_user = self.init_config.get('postfix_user', 'root')
+ output, _, _ = get_subprocess_output(['sudo', '-u', postfix_user, 'find', queue_path, '-type', 'f'], self.log, False)
count = len(output.splitlines())
else:
raise Exception('The dd-agent user does not have sudo access')
|
{"golden_diff": "diff --git a/postfix/check.py b/postfix/check.py\n--- a/postfix/check.py\n+++ b/postfix/check.py\n@@ -17,7 +17,9 @@\n sudo access is not required when running dd-agent as root (not recommended)\n \n example /etc/sudoers entry:\n- dd-agent ALL=(ALL) NOPASSWD:/usr/bin/find /var/spool/postfix* -type f\n+ dd-agent ALL=(postfix) NOPASSWD:/usr/bin/find /var/spool/postfix/incoming -type f\n+ dd-agent ALL=(postfix) NOPASSWD:/usr/bin/find /var/spool/postfix/active -type f\n+ dd-agent ALL=(postfix) NOPASSWD:/usr/bin/find /var/spool/postfix/deferred -type f\n \n YAML config options:\n \"directory\" - the value of 'postconf -h queue_directory'\n@@ -61,7 +63,9 @@\n # can dd-agent user run sudo?\n test_sudo = os.system('setsid sudo -l < /dev/null')\n if test_sudo == 0:\n- output, _, _ = get_subprocess_output(['sudo', 'find', queue_path, '-type', 'f'], self.log, False)\n+ # default to `root` for backward compatibility\n+ postfix_user = self.init_config.get('postfix_user', 'root')\n+ output, _, _ = get_subprocess_output(['sudo', '-u', postfix_user, 'find', queue_path, '-type', 'f'], self.log, False)\n count = len(output.splitlines())\n else:\n raise Exception('The dd-agent user does not have sudo access')\n", "issue": "postfix integration should not require sudo to root\nReading the source code to integrations-core/postfix/check.py I note that it does a sudo to root to run the find command.\r\n\r\nThis is noted in the docs / comments :\r\n\r\n> WARNING: the user that dd-agent runs as must have sudo access for the 'find' command\r\n> --\r\n> \u00a0 | sudo access is not required when running dd-agent as root (not recommended)\r\n> \u00a0 | \u00a0\r\n> \u00a0 | example /etc/sudoers entry:\r\n> \u00a0 | dd-agent ALL=(ALL) NOPASSWD:/usr/bin/find /var/spool/postfix* -type f\r\n\r\nroot should not be required here - postfix user should be sufficient. That would be combined with a '-u postfix' on line 64's sudo command to allow this to work.\r\n\r\nThis is a concern because find has a -exec parameter and your command list has a wildcard in it - this could be used to run arbitrary commands as root if the dd-agent user is compromised.\r\n\n", "before_files": [{"content": "# (C) Datadog, Inc. 2013-2016\n# (C) Josiah C Webb <[email protected]> 2013\n# All rights reserved\n# Licensed under Simplified BSD License (see LICENSE)\n\n# stdlib\nimport os\n\n# project\nfrom checks import AgentCheck\nfrom utils.subprocess_output import get_subprocess_output\n\nclass PostfixCheck(AgentCheck):\n \"\"\"This check provides metrics on the number of messages in a given postfix queue\n\n WARNING: the user that dd-agent runs as must have sudo access for the 'find' command\n sudo access is not required when running dd-agent as root (not recommended)\n\n example /etc/sudoers entry:\n dd-agent ALL=(ALL) NOPASSWD:/usr/bin/find /var/spool/postfix* -type f\n\n YAML config options:\n \"directory\" - the value of 'postconf -h queue_directory'\n \"queues\" - the postfix mail queues you would like to get message count totals for\n \"\"\"\n def check(self, instance):\n config = self._get_config(instance)\n\n directory = config['directory']\n queues = config['queues']\n tags = config['tags']\n\n self._get_queue_count(directory, queues, tags)\n\n def _get_config(self, instance):\n directory = instance.get('directory', None)\n queues = instance.get('queues', None)\n tags = instance.get('tags', [])\n if not queues or not directory:\n raise Exception('missing required yaml config entry')\n\n instance_config = {\n 'directory': directory,\n 'queues': queues,\n 'tags': tags,\n }\n\n return instance_config\n\n def _get_queue_count(self, directory, queues, tags):\n for queue in queues:\n queue_path = os.path.join(directory, queue)\n if not os.path.exists(queue_path):\n raise Exception('%s does not exist' % queue_path)\n\n count = 0\n if os.geteuid() == 0:\n # dd-agent is running as root (not recommended)\n count = sum(len(files) for root, dirs, files in os.walk(queue_path))\n else:\n # can dd-agent user run sudo?\n test_sudo = os.system('setsid sudo -l < /dev/null')\n if test_sudo == 0:\n output, _, _ = get_subprocess_output(['sudo', 'find', queue_path, '-type', 'f'], self.log, False)\n count = len(output.splitlines())\n else:\n raise Exception('The dd-agent user does not have sudo access')\n\n # emit an individually tagged metric\n self.gauge('postfix.queue.size', count, tags=tags + ['queue:%s' % queue, 'instance:%s' % os.path.basename(directory)])\n\n # these can be retrieved in a single graph statement\n # for example:\n # sum:postfix.queue.size{instance:postfix-2,queue:incoming,host:hostname.domain.tld}\n", "path": "postfix/check.py"}]}
| 1,538 | 368 |
gh_patches_debug_21267
|
rasdani/github-patches
|
git_diff
|
conan-io__conan-8625
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Remove removal of "sh" from path for MinGW in CMake
New conan.tools.cmake CMake helper contains:
```python
if is_windows_mingw:
with tools.remove_from_path("sh"):
self._conanfile.run(command)
```
This is no longer necessary, can be replaced by ``set(DCMAKE_SH="CMAKE_SH-NOTFOUND")``. Please remove it and add a red/green test for it.
</issue>
<code>
[start of conan/tools/cmake/cmake.py]
1 import os
2 import platform
3
4 from conan.tools.cmake.base import CMakeToolchainBase
5 from conan.tools.cmake.utils import get_generator, is_multi_configuration
6 from conan.tools.microsoft.msbuild import msbuild_verbosity_cmd_line_arg
7 from conans.client import tools
8 from conans.client.build import join_arguments
9 from conans.client.tools.files import chdir
10 from conans.client.tools.oss import cpu_count, args_to_string
11 from conans.errors import ConanException
12 from conans.model.version import Version
13 from conans.util.conan_v2_mode import conan_v2_error
14 from conans.util.files import mkdir
15
16
17 def _validate_recipe(conanfile):
18 forbidden_generators = ["cmake", "cmake_multi"]
19 if any(it in conanfile.generators for it in forbidden_generators):
20 raise ConanException("Usage of toolchain is only supported with 'cmake_find_package'"
21 " or 'cmake_find_package_multi' generators")
22
23
24 def _cmake_cmd_line_args(conanfile, generator, parallel):
25 args = []
26 compiler_version = conanfile.settings.get_safe("compiler.version")
27 if generator and parallel:
28 if ("Makefiles" in generator or "Ninja" in generator) and "NMake" not in generator:
29 args.append("-j%i" % cpu_count(conanfile.output))
30 elif "Visual Studio" in generator and compiler_version and Version(compiler_version) >= "10":
31 # Parallel for building projects in the solution
32 args.append("/m:%i" % cpu_count(output=conanfile.output))
33
34 if generator and "Visual Studio" in generator:
35 verbosity = msbuild_verbosity_cmd_line_arg(conanfile)
36 if verbosity:
37 args.append(verbosity)
38
39 return args
40
41
42 class CMake(object):
43 """ CMake helper to use together with the toolchain feature. It implements a very simple
44 wrapper to call the cmake executable, but without passing compile flags, preprocessor
45 definitions... all that is set by the toolchain. Only the generator and the CMAKE_TOOLCHAIN_FILE
46 are passed to the command line, plus the ``--config Release`` for builds in multi-config
47 """
48
49 def __init__(self, conanfile, generator=None, build_folder=None, parallel=True):
50 _validate_recipe(conanfile)
51
52 # assert generator is None, "'generator' is handled by the toolchain"
53 self._generator = generator or get_generator(conanfile)
54 self._is_multiconfiguration = is_multi_configuration(self._generator)
55
56 # Store a reference to useful data
57 self._conanfile = conanfile
58 self._parallel = parallel
59
60 self._build_folder = build_folder
61 self._cmake_program = "cmake" # Path to CMake should be handled by environment
62
63 def configure(self, source_folder=None):
64 # TODO: environment?
65 if not self._conanfile.should_configure:
66 return
67
68 source = self._conanfile.source_folder
69 if source_folder:
70 source = os.path.join(self._conanfile.source_folder, source_folder)
71
72 build_folder = self._conanfile.build_folder
73 if self._build_folder:
74 build_folder = os.path.join(self._conanfile.build_folder, self._build_folder)
75
76 mkdir(build_folder)
77 arg_list = '-DCMAKE_TOOLCHAIN_FILE="{}" -DCMAKE_INSTALL_PREFIX="{}" "{}"'.format(
78 CMakeToolchainBase.filename,
79 self._conanfile.package_folder.replace("\\", "/"),
80 source)
81
82 generator = '-G "{}" '.format(self._generator) if self._generator else ""
83 command = "%s %s%s" % (self._cmake_program, generator, arg_list)
84
85 is_windows_mingw = platform.system() == "Windows" and self._generator == "MinGW Makefiles"
86 self._conanfile.output.info("CMake command: %s" % command)
87 with chdir(build_folder):
88 if is_windows_mingw:
89 with tools.remove_from_path("sh"):
90 self._conanfile.run(command)
91 else:
92 self._conanfile.run(command)
93
94 def _build(self, build_type=None, target=None):
95 bf = self._conanfile.build_folder
96 if self._build_folder:
97 bf = os.path.join(self._conanfile.build_folder, self._build_folder)
98
99 if build_type and not self._is_multiconfiguration:
100 self._conanfile.output.error("Don't specify 'build_type' at build time for "
101 "single-config build systems")
102
103 bt = build_type or self._conanfile.settings.get_safe("build_type")
104 conan_v2_error("build_type setting should be defined.", not bt)
105
106 if bt and self._is_multiconfiguration:
107 build_config = "--config %s" % bt
108 else:
109 build_config = ""
110
111 args = []
112 if target is not None:
113 args = ["--target", target]
114
115 cmd_line_args = _cmake_cmd_line_args(self._conanfile, self._generator, self._parallel)
116 if cmd_line_args:
117 args += ['--'] + cmd_line_args
118
119 arg_list = [args_to_string([bf]), build_config, args_to_string(args)]
120 command = "%s --build %s" % (self._cmake_program, join_arguments(arg_list))
121 self._conanfile.output.info("CMake command: %s" % command)
122 self._conanfile.run(command)
123
124 def build(self, build_type=None, target=None):
125 if not self._conanfile.should_build:
126 return
127 self._build(build_type, target)
128
129 def install(self, build_type=None):
130 if not self._conanfile.should_install:
131 return
132 mkdir(self._conanfile.package_folder)
133 self._build(build_type=build_type, target="install")
134
135 def test(self, build_type=None, target=None, output_on_failure=False):
136 if not self._conanfile.should_test:
137 return
138 if not target:
139 target = "RUN_TESTS" if self._is_multiconfiguration else "test"
140
141 env = {'CTEST_OUTPUT_ON_FAILURE': '1' if output_on_failure else '0'}
142 if self._parallel:
143 env['CTEST_PARALLEL_LEVEL'] = str(cpu_count(self._conanfile.output))
144 with tools.environment_append(env):
145 self._build(build_type=build_type, target=target)
146
[end of conan/tools/cmake/cmake.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/conan/tools/cmake/cmake.py b/conan/tools/cmake/cmake.py
--- a/conan/tools/cmake/cmake.py
+++ b/conan/tools/cmake/cmake.py
@@ -79,17 +79,15 @@
self._conanfile.package_folder.replace("\\", "/"),
source)
+ if platform.system() == "Windows" and self._generator == "MinGW Makefiles":
+ arg_list += ' -DCMAKE_SH="CMAKE_SH-NOTFOUND"'
+
generator = '-G "{}" '.format(self._generator) if self._generator else ""
command = "%s %s%s" % (self._cmake_program, generator, arg_list)
- is_windows_mingw = platform.system() == "Windows" and self._generator == "MinGW Makefiles"
self._conanfile.output.info("CMake command: %s" % command)
with chdir(build_folder):
- if is_windows_mingw:
- with tools.remove_from_path("sh"):
- self._conanfile.run(command)
- else:
- self._conanfile.run(command)
+ self._conanfile.run(command)
def _build(self, build_type=None, target=None):
bf = self._conanfile.build_folder
|
{"golden_diff": "diff --git a/conan/tools/cmake/cmake.py b/conan/tools/cmake/cmake.py\n--- a/conan/tools/cmake/cmake.py\n+++ b/conan/tools/cmake/cmake.py\n@@ -79,17 +79,15 @@\n self._conanfile.package_folder.replace(\"\\\\\", \"/\"),\n source)\n \n+ if platform.system() == \"Windows\" and self._generator == \"MinGW Makefiles\":\n+ arg_list += ' -DCMAKE_SH=\"CMAKE_SH-NOTFOUND\"'\n+\n generator = '-G \"{}\" '.format(self._generator) if self._generator else \"\"\n command = \"%s %s%s\" % (self._cmake_program, generator, arg_list)\n \n- is_windows_mingw = platform.system() == \"Windows\" and self._generator == \"MinGW Makefiles\"\n self._conanfile.output.info(\"CMake command: %s\" % command)\n with chdir(build_folder):\n- if is_windows_mingw:\n- with tools.remove_from_path(\"sh\"):\n- self._conanfile.run(command)\n- else:\n- self._conanfile.run(command)\n+ self._conanfile.run(command)\n \n def _build(self, build_type=None, target=None):\n bf = self._conanfile.build_folder\n", "issue": "Remove removal of \"sh\" from path for MinGW in CMake\nNew conan.tools.cmake CMake helper contains:\r\n\r\n```python\r\n if is_windows_mingw:\r\n with tools.remove_from_path(\"sh\"):\r\n self._conanfile.run(command)\r\n```\r\n\r\nThis is no longer necessary, can be replaced by ``set(DCMAKE_SH=\"CMAKE_SH-NOTFOUND\")``. Please remove it and add a red/green test for it.\n", "before_files": [{"content": "import os\nimport platform\n\nfrom conan.tools.cmake.base import CMakeToolchainBase\nfrom conan.tools.cmake.utils import get_generator, is_multi_configuration\nfrom conan.tools.microsoft.msbuild import msbuild_verbosity_cmd_line_arg\nfrom conans.client import tools\nfrom conans.client.build import join_arguments\nfrom conans.client.tools.files import chdir\nfrom conans.client.tools.oss import cpu_count, args_to_string\nfrom conans.errors import ConanException\nfrom conans.model.version import Version\nfrom conans.util.conan_v2_mode import conan_v2_error\nfrom conans.util.files import mkdir\n\n\ndef _validate_recipe(conanfile):\n forbidden_generators = [\"cmake\", \"cmake_multi\"]\n if any(it in conanfile.generators for it in forbidden_generators):\n raise ConanException(\"Usage of toolchain is only supported with 'cmake_find_package'\"\n \" or 'cmake_find_package_multi' generators\")\n\n\ndef _cmake_cmd_line_args(conanfile, generator, parallel):\n args = []\n compiler_version = conanfile.settings.get_safe(\"compiler.version\")\n if generator and parallel:\n if (\"Makefiles\" in generator or \"Ninja\" in generator) and \"NMake\" not in generator:\n args.append(\"-j%i\" % cpu_count(conanfile.output))\n elif \"Visual Studio\" in generator and compiler_version and Version(compiler_version) >= \"10\":\n # Parallel for building projects in the solution\n args.append(\"/m:%i\" % cpu_count(output=conanfile.output))\n\n if generator and \"Visual Studio\" in generator:\n verbosity = msbuild_verbosity_cmd_line_arg(conanfile)\n if verbosity:\n args.append(verbosity)\n\n return args\n\n\nclass CMake(object):\n \"\"\" CMake helper to use together with the toolchain feature. It implements a very simple\n wrapper to call the cmake executable, but without passing compile flags, preprocessor\n definitions... all that is set by the toolchain. Only the generator and the CMAKE_TOOLCHAIN_FILE\n are passed to the command line, plus the ``--config Release`` for builds in multi-config\n \"\"\"\n\n def __init__(self, conanfile, generator=None, build_folder=None, parallel=True):\n _validate_recipe(conanfile)\n\n # assert generator is None, \"'generator' is handled by the toolchain\"\n self._generator = generator or get_generator(conanfile)\n self._is_multiconfiguration = is_multi_configuration(self._generator)\n\n # Store a reference to useful data\n self._conanfile = conanfile\n self._parallel = parallel\n\n self._build_folder = build_folder\n self._cmake_program = \"cmake\" # Path to CMake should be handled by environment\n\n def configure(self, source_folder=None):\n # TODO: environment?\n if not self._conanfile.should_configure:\n return\n\n source = self._conanfile.source_folder\n if source_folder:\n source = os.path.join(self._conanfile.source_folder, source_folder)\n\n build_folder = self._conanfile.build_folder\n if self._build_folder:\n build_folder = os.path.join(self._conanfile.build_folder, self._build_folder)\n\n mkdir(build_folder)\n arg_list = '-DCMAKE_TOOLCHAIN_FILE=\"{}\" -DCMAKE_INSTALL_PREFIX=\"{}\" \"{}\"'.format(\n CMakeToolchainBase.filename,\n self._conanfile.package_folder.replace(\"\\\\\", \"/\"),\n source)\n\n generator = '-G \"{}\" '.format(self._generator) if self._generator else \"\"\n command = \"%s %s%s\" % (self._cmake_program, generator, arg_list)\n\n is_windows_mingw = platform.system() == \"Windows\" and self._generator == \"MinGW Makefiles\"\n self._conanfile.output.info(\"CMake command: %s\" % command)\n with chdir(build_folder):\n if is_windows_mingw:\n with tools.remove_from_path(\"sh\"):\n self._conanfile.run(command)\n else:\n self._conanfile.run(command)\n\n def _build(self, build_type=None, target=None):\n bf = self._conanfile.build_folder\n if self._build_folder:\n bf = os.path.join(self._conanfile.build_folder, self._build_folder)\n\n if build_type and not self._is_multiconfiguration:\n self._conanfile.output.error(\"Don't specify 'build_type' at build time for \"\n \"single-config build systems\")\n\n bt = build_type or self._conanfile.settings.get_safe(\"build_type\")\n conan_v2_error(\"build_type setting should be defined.\", not bt)\n\n if bt and self._is_multiconfiguration:\n build_config = \"--config %s\" % bt\n else:\n build_config = \"\"\n\n args = []\n if target is not None:\n args = [\"--target\", target]\n\n cmd_line_args = _cmake_cmd_line_args(self._conanfile, self._generator, self._parallel)\n if cmd_line_args:\n args += ['--'] + cmd_line_args\n\n arg_list = [args_to_string([bf]), build_config, args_to_string(args)]\n command = \"%s --build %s\" % (self._cmake_program, join_arguments(arg_list))\n self._conanfile.output.info(\"CMake command: %s\" % command)\n self._conanfile.run(command)\n\n def build(self, build_type=None, target=None):\n if not self._conanfile.should_build:\n return\n self._build(build_type, target)\n\n def install(self, build_type=None):\n if not self._conanfile.should_install:\n return\n mkdir(self._conanfile.package_folder)\n self._build(build_type=build_type, target=\"install\")\n\n def test(self, build_type=None, target=None, output_on_failure=False):\n if not self._conanfile.should_test:\n return\n if not target:\n target = \"RUN_TESTS\" if self._is_multiconfiguration else \"test\"\n\n env = {'CTEST_OUTPUT_ON_FAILURE': '1' if output_on_failure else '0'}\n if self._parallel:\n env['CTEST_PARALLEL_LEVEL'] = str(cpu_count(self._conanfile.output))\n with tools.environment_append(env):\n self._build(build_type=build_type, target=target)\n", "path": "conan/tools/cmake/cmake.py"}]}
| 2,369 | 287 |
gh_patches_debug_15930
|
rasdani/github-patches
|
git_diff
|
conan-io__conan-4591
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
[question][suggestion] Listing profile should be recursive
Hello,
I use multiple profiles, and those are organized in subdirectories:
+ `~/.conan/profiles/application/x64_gcc6_app1`
+ `~/.conan/profiles/application/x64_msvc_app1`
+ `~/.conan/profiles/compilers/x64_gcc6`
+ `~/.conan/profiles/compilers/x64_msvc`
The "applications" profile include other profiles etc. This works pretty well, so I assume
using subdirectory in profiles is supported and is not a problem.
However, the `conan profile list` command does not list profiles contained in subdirectories.
I believe it should recursively search for profile files rather than only list the files available directly in the `~/.conan/profiles` directory.
I'm wondering if there is a particular reason why the search is limited to the `~/.conan/profiles` directory and if you'd be open to changing this behavior.
</issue>
<code>
[start of conans/client/cmd/profile.py]
1 import os
2
3 from conans.client.conf.detect import detect_defaults_settings
4 from conans.client.profile_loader import get_profile_path, read_profile
5 from conans.errors import ConanException
6 from conans.model.options import OptionsValues
7 from conans.model.profile import Profile
8 from conans.unicode import get_cwd
9 from conans.util.files import save
10
11
12 def _get_profile_keys(key):
13 # settings.compiler.version => settings, compiler.version
14 tmp = key.split(".")
15 first_key = tmp[0]
16 rest_key = ".".join(tmp[1:]) if len(tmp) > 1 else None
17 if first_key not in ("build_requires", "settings", "options", "env"):
18 raise ConanException("Invalid specified key: %s" % key)
19
20 return first_key, rest_key
21
22
23 def cmd_profile_list(cache_profiles_path, output):
24 folder = cache_profiles_path
25 if os.path.exists(folder):
26 return [name for name in os.listdir(folder)
27 if not os.path.isdir(os.path.join(folder, name))]
28 else:
29 output.info("No profiles defined")
30 return []
31
32
33 def cmd_profile_create(profile_name, cache_profiles_path, output, detect=False):
34 profile_path = get_profile_path(profile_name, cache_profiles_path, get_cwd(),
35 exists=False)
36 if os.path.exists(profile_path):
37 raise ConanException("Profile already exists")
38
39 profile = Profile()
40 if detect:
41 settings = detect_defaults_settings(output)
42 for name, value in settings:
43 profile.settings[name] = value
44
45 contents = profile.dumps()
46 save(profile_path, contents)
47
48 if detect:
49 output.info("Profile created with detected settings: %s" % profile_path)
50 else:
51 output.info("Empty profile created: %s" % profile_path)
52 return profile_path
53
54
55 def cmd_profile_update(profile_name, key, value, cache_profiles_path):
56 first_key, rest_key = _get_profile_keys(key)
57
58 profile, _ = read_profile(profile_name, get_cwd(), cache_profiles_path)
59 if first_key == "settings":
60 profile.settings[rest_key] = value
61 elif first_key == "options":
62 tmp = OptionsValues([(rest_key, value)])
63 profile.options.update(tmp)
64 elif first_key == "env":
65 profile.env_values.update_replace(rest_key, value)
66 elif first_key == "build_requires":
67 raise ConanException("Edit the profile manually to change the build_requires")
68
69 contents = profile.dumps()
70 profile_path = get_profile_path(profile_name, cache_profiles_path, get_cwd())
71 save(profile_path, contents)
72
73
74 def cmd_profile_get(profile_name, key, cache_profiles_path):
75 first_key, rest_key = _get_profile_keys(key)
76 profile, _ = read_profile(profile_name, get_cwd(), cache_profiles_path)
77 try:
78 if first_key == "settings":
79 return profile.settings[rest_key]
80 elif first_key == "options":
81 return dict(profile.options.as_list())[rest_key]
82 elif first_key == "env":
83 package = None
84 var = rest_key
85 if ":" in rest_key:
86 package, var = rest_key.split(":")
87 return profile.env_values.data[package][var]
88 elif first_key == "build_requires":
89 raise ConanException("List the profile manually to see the build_requires")
90 except KeyError:
91 raise ConanException("Key not found: '%s'" % key)
92
93
94 def cmd_profile_delete_key(profile_name, key, cache_profiles_path):
95 first_key, rest_key = _get_profile_keys(key)
96 profile, _ = read_profile(profile_name, get_cwd(), cache_profiles_path)
97
98 try:
99 package, name = rest_key.split(":")
100 except ValueError:
101 package = None
102 name = rest_key
103
104 try:
105 if first_key == "settings":
106 del profile.settings[rest_key]
107 elif first_key == "options":
108 profile.options.remove(name, package)
109 elif first_key == "env":
110 profile.env_values.remove(name, package)
111 elif first_key == "build_requires":
112 raise ConanException("Edit the profile manually to delete a build_require")
113 except KeyError:
114 raise ConanException("Profile key '%s' doesn't exist" % key)
115
116 contents = profile.dumps()
117 profile_path = get_profile_path(profile_name, cache_profiles_path, get_cwd())
118 save(profile_path, contents)
119
[end of conans/client/cmd/profile.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/conans/client/cmd/profile.py b/conans/client/cmd/profile.py
--- a/conans/client/cmd/profile.py
+++ b/conans/client/cmd/profile.py
@@ -21,13 +21,18 @@
def cmd_profile_list(cache_profiles_path, output):
- folder = cache_profiles_path
- if os.path.exists(folder):
- return [name for name in os.listdir(folder)
- if not os.path.isdir(os.path.join(folder, name))]
- else:
+ profiles = []
+ if os.path.exists(cache_profiles_path):
+ for current_directory, _, files in os.walk(cache_profiles_path, followlinks=True):
+ for filename in files:
+ rel_path = os.path.relpath(os.path.join(current_directory, filename),
+ cache_profiles_path)
+ profiles.append(rel_path)
+
+ if not profiles:
output.info("No profiles defined")
- return []
+ profiles.sort()
+ return profiles
def cmd_profile_create(profile_name, cache_profiles_path, output, detect=False):
|
{"golden_diff": "diff --git a/conans/client/cmd/profile.py b/conans/client/cmd/profile.py\n--- a/conans/client/cmd/profile.py\n+++ b/conans/client/cmd/profile.py\n@@ -21,13 +21,18 @@\n \n \n def cmd_profile_list(cache_profiles_path, output):\n- folder = cache_profiles_path\n- if os.path.exists(folder):\n- return [name for name in os.listdir(folder)\n- if not os.path.isdir(os.path.join(folder, name))]\n- else:\n+ profiles = []\n+ if os.path.exists(cache_profiles_path):\n+ for current_directory, _, files in os.walk(cache_profiles_path, followlinks=True):\n+ for filename in files:\n+ rel_path = os.path.relpath(os.path.join(current_directory, filename),\n+ cache_profiles_path)\n+ profiles.append(rel_path)\n+\n+ if not profiles:\n output.info(\"No profiles defined\")\n- return []\n+ profiles.sort()\n+ return profiles\n \n \n def cmd_profile_create(profile_name, cache_profiles_path, output, detect=False):\n", "issue": "[question][suggestion] Listing profile should be recursive\nHello,\r\n\r\nI use multiple profiles, and those are organized in subdirectories:\r\n + `~/.conan/profiles/application/x64_gcc6_app1`\r\n + `~/.conan/profiles/application/x64_msvc_app1`\r\n + `~/.conan/profiles/compilers/x64_gcc6`\r\n + `~/.conan/profiles/compilers/x64_msvc`\r\n\r\nThe \"applications\" profile include other profiles etc. This works pretty well, so I assume \r\nusing subdirectory in profiles is supported and is not a problem.\r\n\r\nHowever, the `conan profile list` command does not list profiles contained in subdirectories.\r\nI believe it should recursively search for profile files rather than only list the files available directly in the `~/.conan/profiles` directory.\r\n\r\nI'm wondering if there is a particular reason why the search is limited to the `~/.conan/profiles` directory and if you'd be open to changing this behavior.\r\n\n", "before_files": [{"content": "import os\n\nfrom conans.client.conf.detect import detect_defaults_settings\nfrom conans.client.profile_loader import get_profile_path, read_profile\nfrom conans.errors import ConanException\nfrom conans.model.options import OptionsValues\nfrom conans.model.profile import Profile\nfrom conans.unicode import get_cwd\nfrom conans.util.files import save\n\n\ndef _get_profile_keys(key):\n # settings.compiler.version => settings, compiler.version\n tmp = key.split(\".\")\n first_key = tmp[0]\n rest_key = \".\".join(tmp[1:]) if len(tmp) > 1 else None\n if first_key not in (\"build_requires\", \"settings\", \"options\", \"env\"):\n raise ConanException(\"Invalid specified key: %s\" % key)\n\n return first_key, rest_key\n\n\ndef cmd_profile_list(cache_profiles_path, output):\n folder = cache_profiles_path\n if os.path.exists(folder):\n return [name for name in os.listdir(folder)\n if not os.path.isdir(os.path.join(folder, name))]\n else:\n output.info(\"No profiles defined\")\n return []\n\n\ndef cmd_profile_create(profile_name, cache_profiles_path, output, detect=False):\n profile_path = get_profile_path(profile_name, cache_profiles_path, get_cwd(),\n exists=False)\n if os.path.exists(profile_path):\n raise ConanException(\"Profile already exists\")\n\n profile = Profile()\n if detect:\n settings = detect_defaults_settings(output)\n for name, value in settings:\n profile.settings[name] = value\n\n contents = profile.dumps()\n save(profile_path, contents)\n\n if detect:\n output.info(\"Profile created with detected settings: %s\" % profile_path)\n else:\n output.info(\"Empty profile created: %s\" % profile_path)\n return profile_path\n\n\ndef cmd_profile_update(profile_name, key, value, cache_profiles_path):\n first_key, rest_key = _get_profile_keys(key)\n\n profile, _ = read_profile(profile_name, get_cwd(), cache_profiles_path)\n if first_key == \"settings\":\n profile.settings[rest_key] = value\n elif first_key == \"options\":\n tmp = OptionsValues([(rest_key, value)])\n profile.options.update(tmp)\n elif first_key == \"env\":\n profile.env_values.update_replace(rest_key, value)\n elif first_key == \"build_requires\":\n raise ConanException(\"Edit the profile manually to change the build_requires\")\n\n contents = profile.dumps()\n profile_path = get_profile_path(profile_name, cache_profiles_path, get_cwd())\n save(profile_path, contents)\n\n\ndef cmd_profile_get(profile_name, key, cache_profiles_path):\n first_key, rest_key = _get_profile_keys(key)\n profile, _ = read_profile(profile_name, get_cwd(), cache_profiles_path)\n try:\n if first_key == \"settings\":\n return profile.settings[rest_key]\n elif first_key == \"options\":\n return dict(profile.options.as_list())[rest_key]\n elif first_key == \"env\":\n package = None\n var = rest_key\n if \":\" in rest_key:\n package, var = rest_key.split(\":\")\n return profile.env_values.data[package][var]\n elif first_key == \"build_requires\":\n raise ConanException(\"List the profile manually to see the build_requires\")\n except KeyError:\n raise ConanException(\"Key not found: '%s'\" % key)\n\n\ndef cmd_profile_delete_key(profile_name, key, cache_profiles_path):\n first_key, rest_key = _get_profile_keys(key)\n profile, _ = read_profile(profile_name, get_cwd(), cache_profiles_path)\n\n try:\n package, name = rest_key.split(\":\")\n except ValueError:\n package = None\n name = rest_key\n\n try:\n if first_key == \"settings\":\n del profile.settings[rest_key]\n elif first_key == \"options\":\n profile.options.remove(name, package)\n elif first_key == \"env\":\n profile.env_values.remove(name, package)\n elif first_key == \"build_requires\":\n raise ConanException(\"Edit the profile manually to delete a build_require\")\n except KeyError:\n raise ConanException(\"Profile key '%s' doesn't exist\" % key)\n\n contents = profile.dumps()\n profile_path = get_profile_path(profile_name, cache_profiles_path, get_cwd())\n save(profile_path, contents)\n", "path": "conans/client/cmd/profile.py"}]}
| 1,938 | 224 |
gh_patches_debug_24330
|
rasdani/github-patches
|
git_diff
|
cookiecutter__cookiecutter-1626
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Boolean parameters parsed as String
* Cookiecutter version: 2.0
* Template project url: https://github.com/cookiecutter/cookiecutter-django
* Python version: 3.8
* Operating System: Linux
### Description:
We want to be able to ask the user for true/false questions.
The example template shows the current usage of boolean parameters, in {{cookiecutter.project_slug}}/.github/dependabot.yml:
`{%- if cookiecutter.use_docker == 'y' %}`
This usage is rather problematic, as it is requires the user to enter exactly 'y'. If the user enters "yes" or "true" it won't be valid.
We want to add the ability to specify boolean parameters as "true" JSON booleans and naturally prompt for them.
Thus the usage of booleans in the template's cookiecutter.json would look like:
`"use_docker": false`
Instead of the current usage, which is:
`"use_docker": "n"`
Currently this example boolean which is specified in the cookiecutter.json file is parsed as a string and thus checking for it would require a string comparison, which raises and exact same problem as the user would have to explicitly enter "true" or "false" instead of a boolean.
This would also simplify the usage of the boolean parameters in files, so the above shown dependabot.yml would change to:
`{%- if cookiecutter.use_docker %}`
</issue>
<code>
[start of cookiecutter/prompt.py]
1 """Functions for prompting the user for project info."""
2 import functools
3 import json
4 from collections import OrderedDict
5
6 import click
7 from jinja2.exceptions import UndefinedError
8
9 from cookiecutter.environment import StrictEnvironment
10 from cookiecutter.exceptions import UndefinedVariableInTemplate
11
12
13 def read_user_variable(var_name, default_value):
14 """Prompt user for variable and return the entered value or given default.
15
16 :param str var_name: Variable of the context to query the user
17 :param default_value: Value that will be returned if no input happens
18 """
19 # Please see https://click.palletsprojects.com/en/7.x/api/#click.prompt
20 return click.prompt(var_name, default=default_value)
21
22
23 def read_user_yes_no(question, default_value):
24 """Prompt the user to reply with 'yes' or 'no' (or equivalent values).
25
26 Note:
27 Possible choices are 'true', '1', 'yes', 'y' or 'false', '0', 'no', 'n'
28
29 :param str question: Question to the user
30 :param default_value: Value that will be returned if no input happens
31 """
32 # Please see https://click.palletsprojects.com/en/7.x/api/#click.prompt
33 return click.prompt(question, default=default_value, type=click.BOOL)
34
35
36 def read_repo_password(question):
37 """Prompt the user to enter a password.
38
39 :param str question: Question to the user
40 """
41 # Please see https://click.palletsprojects.com/en/7.x/api/#click.prompt
42 return click.prompt(question, hide_input=True)
43
44
45 def read_user_choice(var_name, options):
46 """Prompt the user to choose from several options for the given variable.
47
48 The first item will be returned if no input happens.
49
50 :param str var_name: Variable as specified in the context
51 :param list options: Sequence of options that are available to select from
52 :return: Exactly one item of ``options`` that has been chosen by the user
53 """
54 # Please see https://click.palletsprojects.com/en/7.x/api/#click.prompt
55 if not isinstance(options, list):
56 raise TypeError
57
58 if not options:
59 raise ValueError
60
61 choice_map = OrderedDict((f'{i}', value) for i, value in enumerate(options, 1))
62 choices = choice_map.keys()
63 default = '1'
64
65 choice_lines = ['{} - {}'.format(*c) for c in choice_map.items()]
66 prompt = '\n'.join(
67 (
68 f'Select {var_name}:',
69 '\n'.join(choice_lines),
70 'Choose from {}'.format(', '.join(choices)),
71 )
72 )
73
74 user_choice = click.prompt(
75 prompt, type=click.Choice(choices), default=default, show_choices=False
76 )
77 return choice_map[user_choice]
78
79
80 DEFAULT_DISPLAY = 'default'
81
82
83 def process_json(user_value, default_value=None):
84 """Load user-supplied value as a JSON dict.
85
86 :param str user_value: User-supplied value to load as a JSON dict
87 """
88 if user_value == DEFAULT_DISPLAY:
89 # Return the given default w/o any processing
90 return default_value
91
92 try:
93 user_dict = json.loads(user_value, object_pairs_hook=OrderedDict)
94 except Exception:
95 # Leave it up to click to ask the user again
96 raise click.UsageError('Unable to decode to JSON.')
97
98 if not isinstance(user_dict, dict):
99 # Leave it up to click to ask the user again
100 raise click.UsageError('Requires JSON dict.')
101
102 return user_dict
103
104
105 def read_user_dict(var_name, default_value):
106 """Prompt the user to provide a dictionary of data.
107
108 :param str var_name: Variable as specified in the context
109 :param default_value: Value that will be returned if no input is provided
110 :return: A Python dictionary to use in the context.
111 """
112 # Please see https://click.palletsprojects.com/en/7.x/api/#click.prompt
113 if not isinstance(default_value, dict):
114 raise TypeError
115
116 user_value = click.prompt(
117 var_name,
118 default=DEFAULT_DISPLAY,
119 type=click.STRING,
120 value_proc=functools.partial(process_json, default_value=default_value),
121 )
122
123 if click.__version__.startswith("7.") and user_value == DEFAULT_DISPLAY:
124 # click 7.x does not invoke value_proc on the default value.
125 return default_value # pragma: no cover
126 return user_value
127
128
129 def render_variable(env, raw, cookiecutter_dict):
130 """Render the next variable to be displayed in the user prompt.
131
132 Inside the prompting taken from the cookiecutter.json file, this renders
133 the next variable. For example, if a project_name is "Peanut Butter
134 Cookie", the repo_name could be be rendered with:
135
136 `{{ cookiecutter.project_name.replace(" ", "_") }}`.
137
138 This is then presented to the user as the default.
139
140 :param Environment env: A Jinja2 Environment object.
141 :param raw: The next value to be prompted for by the user.
142 :param dict cookiecutter_dict: The current context as it's gradually
143 being populated with variables.
144 :return: The rendered value for the default variable.
145 """
146 if raw is None:
147 return None
148 elif isinstance(raw, dict):
149 return {
150 render_variable(env, k, cookiecutter_dict): render_variable(
151 env, v, cookiecutter_dict
152 )
153 for k, v in raw.items()
154 }
155 elif isinstance(raw, list):
156 return [render_variable(env, v, cookiecutter_dict) for v in raw]
157 elif not isinstance(raw, str):
158 raw = str(raw)
159
160 template = env.from_string(raw)
161
162 rendered_template = template.render(cookiecutter=cookiecutter_dict)
163 return rendered_template
164
165
166 def prompt_choice_for_config(cookiecutter_dict, env, key, options, no_input):
167 """Prompt user with a set of options to choose from.
168
169 :param no_input: Do not prompt for user input and return the first available option.
170 """
171 rendered_options = [render_variable(env, raw, cookiecutter_dict) for raw in options]
172 if no_input:
173 return rendered_options[0]
174 return read_user_choice(key, rendered_options)
175
176
177 def prompt_for_config(context, no_input=False):
178 """Prompt user to enter a new config.
179
180 :param dict context: Source for field names and sample values.
181 :param no_input: Do not prompt for user input and use only values from context.
182 """
183 cookiecutter_dict = OrderedDict([])
184 env = StrictEnvironment(context=context)
185
186 # First pass: Handle simple and raw variables, plus choices.
187 # These must be done first because the dictionaries keys and
188 # values might refer to them.
189 for key, raw in context['cookiecutter'].items():
190 if key.startswith('_') and not key.startswith('__'):
191 cookiecutter_dict[key] = raw
192 continue
193 elif key.startswith('__'):
194 cookiecutter_dict[key] = render_variable(env, raw, cookiecutter_dict)
195 continue
196
197 try:
198 if isinstance(raw, list):
199 # We are dealing with a choice variable
200 val = prompt_choice_for_config(
201 cookiecutter_dict, env, key, raw, no_input
202 )
203 cookiecutter_dict[key] = val
204 elif not isinstance(raw, dict):
205 # We are dealing with a regular variable
206 val = render_variable(env, raw, cookiecutter_dict)
207
208 if not no_input:
209 val = read_user_variable(key, val)
210
211 cookiecutter_dict[key] = val
212 except UndefinedError as err:
213 msg = f"Unable to render variable '{key}'"
214 raise UndefinedVariableInTemplate(msg, err, context)
215
216 # Second pass; handle the dictionaries.
217 for key, raw in context['cookiecutter'].items():
218 # Skip private type dicts not ot be rendered.
219 if key.startswith('_') and not key.startswith('__'):
220 continue
221
222 try:
223 if isinstance(raw, dict):
224 # We are dealing with a dict variable
225 val = render_variable(env, raw, cookiecutter_dict)
226
227 if not no_input and not key.startswith('__'):
228 val = read_user_dict(key, val)
229
230 cookiecutter_dict[key] = val
231 except UndefinedError as err:
232 msg = f"Unable to render variable '{key}'"
233 raise UndefinedVariableInTemplate(msg, err, context)
234
235 return cookiecutter_dict
236
[end of cookiecutter/prompt.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/cookiecutter/prompt.py b/cookiecutter/prompt.py
--- a/cookiecutter/prompt.py
+++ b/cookiecutter/prompt.py
@@ -143,8 +143,8 @@
being populated with variables.
:return: The rendered value for the default variable.
"""
- if raw is None:
- return None
+ if raw is None or isinstance(raw, bool):
+ return raw
elif isinstance(raw, dict):
return {
render_variable(env, k, cookiecutter_dict): render_variable(
@@ -201,6 +201,14 @@
cookiecutter_dict, env, key, raw, no_input
)
cookiecutter_dict[key] = val
+ elif isinstance(raw, bool):
+ # We are dealing with a boolean variable
+ if no_input:
+ cookiecutter_dict[key] = render_variable(
+ env, raw, cookiecutter_dict
+ )
+ else:
+ cookiecutter_dict[key] = read_user_yes_no(key, raw)
elif not isinstance(raw, dict):
# We are dealing with a regular variable
val = render_variable(env, raw, cookiecutter_dict)
|
{"golden_diff": "diff --git a/cookiecutter/prompt.py b/cookiecutter/prompt.py\n--- a/cookiecutter/prompt.py\n+++ b/cookiecutter/prompt.py\n@@ -143,8 +143,8 @@\n being populated with variables.\n :return: The rendered value for the default variable.\n \"\"\"\n- if raw is None:\n- return None\n+ if raw is None or isinstance(raw, bool):\n+ return raw\n elif isinstance(raw, dict):\n return {\n render_variable(env, k, cookiecutter_dict): render_variable(\n@@ -201,6 +201,14 @@\n cookiecutter_dict, env, key, raw, no_input\n )\n cookiecutter_dict[key] = val\n+ elif isinstance(raw, bool):\n+ # We are dealing with a boolean variable\n+ if no_input:\n+ cookiecutter_dict[key] = render_variable(\n+ env, raw, cookiecutter_dict\n+ )\n+ else:\n+ cookiecutter_dict[key] = read_user_yes_no(key, raw)\n elif not isinstance(raw, dict):\n # We are dealing with a regular variable\n val = render_variable(env, raw, cookiecutter_dict)\n", "issue": "Boolean parameters parsed as String\n* Cookiecutter version: 2.0\r\n* Template project url: https://github.com/cookiecutter/cookiecutter-django\r\n* Python version: 3.8\r\n* Operating System: Linux\r\n\r\n### Description:\r\n\r\nWe want to be able to ask the user for true/false questions.\r\nThe example template shows the current usage of boolean parameters, in {{cookiecutter.project_slug}}/.github/dependabot.yml:\r\n`{%- if cookiecutter.use_docker == 'y' %}` \r\nThis usage is rather problematic, as it is requires the user to enter exactly 'y'. If the user enters \"yes\" or \"true\" it won't be valid.\r\n\r\nWe want to add the ability to specify boolean parameters as \"true\" JSON booleans and naturally prompt for them.\r\nThus the usage of booleans in the template's cookiecutter.json would look like:\r\n`\"use_docker\": false`\r\nInstead of the current usage, which is:\r\n `\"use_docker\": \"n\"`\r\n\r\nCurrently this example boolean which is specified in the cookiecutter.json file is parsed as a string and thus checking for it would require a string comparison, which raises and exact same problem as the user would have to explicitly enter \"true\" or \"false\" instead of a boolean.\r\n\r\nThis would also simplify the usage of the boolean parameters in files, so the above shown dependabot.yml would change to:\r\n`{%- if cookiecutter.use_docker %}` \n", "before_files": [{"content": "\"\"\"Functions for prompting the user for project info.\"\"\"\nimport functools\nimport json\nfrom collections import OrderedDict\n\nimport click\nfrom jinja2.exceptions import UndefinedError\n\nfrom cookiecutter.environment import StrictEnvironment\nfrom cookiecutter.exceptions import UndefinedVariableInTemplate\n\n\ndef read_user_variable(var_name, default_value):\n \"\"\"Prompt user for variable and return the entered value or given default.\n\n :param str var_name: Variable of the context to query the user\n :param default_value: Value that will be returned if no input happens\n \"\"\"\n # Please see https://click.palletsprojects.com/en/7.x/api/#click.prompt\n return click.prompt(var_name, default=default_value)\n\n\ndef read_user_yes_no(question, default_value):\n \"\"\"Prompt the user to reply with 'yes' or 'no' (or equivalent values).\n\n Note:\n Possible choices are 'true', '1', 'yes', 'y' or 'false', '0', 'no', 'n'\n\n :param str question: Question to the user\n :param default_value: Value that will be returned if no input happens\n \"\"\"\n # Please see https://click.palletsprojects.com/en/7.x/api/#click.prompt\n return click.prompt(question, default=default_value, type=click.BOOL)\n\n\ndef read_repo_password(question):\n \"\"\"Prompt the user to enter a password.\n\n :param str question: Question to the user\n \"\"\"\n # Please see https://click.palletsprojects.com/en/7.x/api/#click.prompt\n return click.prompt(question, hide_input=True)\n\n\ndef read_user_choice(var_name, options):\n \"\"\"Prompt the user to choose from several options for the given variable.\n\n The first item will be returned if no input happens.\n\n :param str var_name: Variable as specified in the context\n :param list options: Sequence of options that are available to select from\n :return: Exactly one item of ``options`` that has been chosen by the user\n \"\"\"\n # Please see https://click.palletsprojects.com/en/7.x/api/#click.prompt\n if not isinstance(options, list):\n raise TypeError\n\n if not options:\n raise ValueError\n\n choice_map = OrderedDict((f'{i}', value) for i, value in enumerate(options, 1))\n choices = choice_map.keys()\n default = '1'\n\n choice_lines = ['{} - {}'.format(*c) for c in choice_map.items()]\n prompt = '\\n'.join(\n (\n f'Select {var_name}:',\n '\\n'.join(choice_lines),\n 'Choose from {}'.format(', '.join(choices)),\n )\n )\n\n user_choice = click.prompt(\n prompt, type=click.Choice(choices), default=default, show_choices=False\n )\n return choice_map[user_choice]\n\n\nDEFAULT_DISPLAY = 'default'\n\n\ndef process_json(user_value, default_value=None):\n \"\"\"Load user-supplied value as a JSON dict.\n\n :param str user_value: User-supplied value to load as a JSON dict\n \"\"\"\n if user_value == DEFAULT_DISPLAY:\n # Return the given default w/o any processing\n return default_value\n\n try:\n user_dict = json.loads(user_value, object_pairs_hook=OrderedDict)\n except Exception:\n # Leave it up to click to ask the user again\n raise click.UsageError('Unable to decode to JSON.')\n\n if not isinstance(user_dict, dict):\n # Leave it up to click to ask the user again\n raise click.UsageError('Requires JSON dict.')\n\n return user_dict\n\n\ndef read_user_dict(var_name, default_value):\n \"\"\"Prompt the user to provide a dictionary of data.\n\n :param str var_name: Variable as specified in the context\n :param default_value: Value that will be returned if no input is provided\n :return: A Python dictionary to use in the context.\n \"\"\"\n # Please see https://click.palletsprojects.com/en/7.x/api/#click.prompt\n if not isinstance(default_value, dict):\n raise TypeError\n\n user_value = click.prompt(\n var_name,\n default=DEFAULT_DISPLAY,\n type=click.STRING,\n value_proc=functools.partial(process_json, default_value=default_value),\n )\n\n if click.__version__.startswith(\"7.\") and user_value == DEFAULT_DISPLAY:\n # click 7.x does not invoke value_proc on the default value.\n return default_value # pragma: no cover\n return user_value\n\n\ndef render_variable(env, raw, cookiecutter_dict):\n \"\"\"Render the next variable to be displayed in the user prompt.\n\n Inside the prompting taken from the cookiecutter.json file, this renders\n the next variable. For example, if a project_name is \"Peanut Butter\n Cookie\", the repo_name could be be rendered with:\n\n `{{ cookiecutter.project_name.replace(\" \", \"_\") }}`.\n\n This is then presented to the user as the default.\n\n :param Environment env: A Jinja2 Environment object.\n :param raw: The next value to be prompted for by the user.\n :param dict cookiecutter_dict: The current context as it's gradually\n being populated with variables.\n :return: The rendered value for the default variable.\n \"\"\"\n if raw is None:\n return None\n elif isinstance(raw, dict):\n return {\n render_variable(env, k, cookiecutter_dict): render_variable(\n env, v, cookiecutter_dict\n )\n for k, v in raw.items()\n }\n elif isinstance(raw, list):\n return [render_variable(env, v, cookiecutter_dict) for v in raw]\n elif not isinstance(raw, str):\n raw = str(raw)\n\n template = env.from_string(raw)\n\n rendered_template = template.render(cookiecutter=cookiecutter_dict)\n return rendered_template\n\n\ndef prompt_choice_for_config(cookiecutter_dict, env, key, options, no_input):\n \"\"\"Prompt user with a set of options to choose from.\n\n :param no_input: Do not prompt for user input and return the first available option.\n \"\"\"\n rendered_options = [render_variable(env, raw, cookiecutter_dict) for raw in options]\n if no_input:\n return rendered_options[0]\n return read_user_choice(key, rendered_options)\n\n\ndef prompt_for_config(context, no_input=False):\n \"\"\"Prompt user to enter a new config.\n\n :param dict context: Source for field names and sample values.\n :param no_input: Do not prompt for user input and use only values from context.\n \"\"\"\n cookiecutter_dict = OrderedDict([])\n env = StrictEnvironment(context=context)\n\n # First pass: Handle simple and raw variables, plus choices.\n # These must be done first because the dictionaries keys and\n # values might refer to them.\n for key, raw in context['cookiecutter'].items():\n if key.startswith('_') and not key.startswith('__'):\n cookiecutter_dict[key] = raw\n continue\n elif key.startswith('__'):\n cookiecutter_dict[key] = render_variable(env, raw, cookiecutter_dict)\n continue\n\n try:\n if isinstance(raw, list):\n # We are dealing with a choice variable\n val = prompt_choice_for_config(\n cookiecutter_dict, env, key, raw, no_input\n )\n cookiecutter_dict[key] = val\n elif not isinstance(raw, dict):\n # We are dealing with a regular variable\n val = render_variable(env, raw, cookiecutter_dict)\n\n if not no_input:\n val = read_user_variable(key, val)\n\n cookiecutter_dict[key] = val\n except UndefinedError as err:\n msg = f\"Unable to render variable '{key}'\"\n raise UndefinedVariableInTemplate(msg, err, context)\n\n # Second pass; handle the dictionaries.\n for key, raw in context['cookiecutter'].items():\n # Skip private type dicts not ot be rendered.\n if key.startswith('_') and not key.startswith('__'):\n continue\n\n try:\n if isinstance(raw, dict):\n # We are dealing with a dict variable\n val = render_variable(env, raw, cookiecutter_dict)\n\n if not no_input and not key.startswith('__'):\n val = read_user_dict(key, val)\n\n cookiecutter_dict[key] = val\n except UndefinedError as err:\n msg = f\"Unable to render variable '{key}'\"\n raise UndefinedVariableInTemplate(msg, err, context)\n\n return cookiecutter_dict\n", "path": "cookiecutter/prompt.py"}]}
| 3,301 | 274 |
gh_patches_debug_60854
|
rasdani/github-patches
|
git_diff
|
airctic__icevision-441
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Add icedata to icevision.all
## 🚀 Feature
Currently to train a dataset available with icedata the following two lines are necessary:
```python
import icedata
from icevision.all import *
```
Because icedata already depends on icevision, icevision cannot depend on icedata. **But** I guess we can add icedata as a soft dependency to `.all`, we just have to be sure not to use `icedata` internally in icevision.
</issue>
<code>
[start of icevision/all.py]
1 from icevision.imports import *
2 from icevision import *
3
[end of icevision/all.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/icevision/all.py b/icevision/all.py
--- a/icevision/all.py
+++ b/icevision/all.py
@@ -1,2 +1,9 @@
from icevision.imports import *
from icevision import *
+
+# soft import icedata
+try:
+ import icedata
+except ModuleNotFoundError as e:
+ if str(e) != f"No module named 'icedata'":
+ raise e
|
{"golden_diff": "diff --git a/icevision/all.py b/icevision/all.py\n--- a/icevision/all.py\n+++ b/icevision/all.py\n@@ -1,2 +1,9 @@\n from icevision.imports import *\n from icevision import *\n+\n+# soft import icedata\n+try:\n+ import icedata\n+except ModuleNotFoundError as e:\n+ if str(e) != f\"No module named 'icedata'\":\n+ raise e\n", "issue": "Add icedata to icevision.all\n## \ud83d\ude80 Feature\r\nCurrently to train a dataset available with icedata the following two lines are necessary:\r\n```python\r\nimport icedata\r\nfrom icevision.all import *\r\n```\r\n\r\nBecause icedata already depends on icevision, icevision cannot depend on icedata. **But** I guess we can add icedata as a soft dependency to `.all`, we just have to be sure not to use `icedata` internally in icevision.\n", "before_files": [{"content": "from icevision.imports import *\nfrom icevision import *\n", "path": "icevision/all.py"}]}
| 652 | 100 |
gh_patches_debug_5156
|
rasdani/github-patches
|
git_diff
|
DataBiosphere__toil-2834
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Example script duplicated
`examples/hello.py` and `src/toil/test/docs/scripts/tutorial_arguments.py` are duplicate scripts with the same contents.
┆Issue is synchronized with this [Jira Task](https://ucsc-cgl.atlassian.net/browse/TOIL-443)
┆Issue Number: TOIL-443
</issue>
<code>
[start of examples/hello.py]
1 from toil.common import Toil
2 from toil.job import Job
3
4 class HelloWorld(Job):
5 def __init__(self, message):
6 Job.__init__(self, memory="1G", cores=2, disk="2G")
7 self.message = message
8
9 def run(self, fileStore):
10 return "Hello, world!, here's a message: %s" % self.message
11
12 if __name__=="__main__":
13 parser = Job.Runner.getDefaultArgumentParser()
14 options = parser.parse_args()
15
16 hello_job = HelloWorld("Woot")
17
18 with Toil(options) as toil:
19 print(toil.start(hello_job))
20
[end of examples/hello.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/examples/hello.py b/examples/hello.py
deleted file mode 100644
--- a/examples/hello.py
+++ /dev/null
@@ -1,19 +0,0 @@
-from toil.common import Toil
-from toil.job import Job
-
-class HelloWorld(Job):
- def __init__(self, message):
- Job.__init__(self, memory="1G", cores=2, disk="2G")
- self.message = message
-
- def run(self, fileStore):
- return "Hello, world!, here's a message: %s" % self.message
-
-if __name__=="__main__":
- parser = Job.Runner.getDefaultArgumentParser()
- options = parser.parse_args()
-
- hello_job = HelloWorld("Woot")
-
- with Toil(options) as toil:
- print(toil.start(hello_job))
|
{"golden_diff": "diff --git a/examples/hello.py b/examples/hello.py\ndeleted file mode 100644\n--- a/examples/hello.py\n+++ /dev/null\n@@ -1,19 +0,0 @@\n-from toil.common import Toil\n-from toil.job import Job\n-\n-class HelloWorld(Job):\n- def __init__(self, message):\n- Job.__init__(self, memory=\"1G\", cores=2, disk=\"2G\")\n- self.message = message\n-\n- def run(self, fileStore):\n- return \"Hello, world!, here's a message: %s\" % self.message\n-\n-if __name__==\"__main__\":\n- parser = Job.Runner.getDefaultArgumentParser()\n- options = parser.parse_args()\n-\n- hello_job = HelloWorld(\"Woot\")\n-\n- with Toil(options) as toil:\n- print(toil.start(hello_job))\n", "issue": "Example script duplicated\n`examples/hello.py` and `src/toil/test/docs/scripts/tutorial_arguments.py` are duplicate scripts with the same contents.\n\n\u2506Issue is synchronized with this [Jira Task](https://ucsc-cgl.atlassian.net/browse/TOIL-443)\n\u2506Issue Number: TOIL-443\n\n", "before_files": [{"content": "from toil.common import Toil\nfrom toil.job import Job\n\nclass HelloWorld(Job):\n def __init__(self, message):\n Job.__init__(self, memory=\"1G\", cores=2, disk=\"2G\")\n self.message = message\n\n def run(self, fileStore):\n return \"Hello, world!, here's a message: %s\" % self.message\n\nif __name__==\"__main__\":\n parser = Job.Runner.getDefaultArgumentParser()\n options = parser.parse_args()\n\n hello_job = HelloWorld(\"Woot\")\n\n with Toil(options) as toil:\n print(toil.start(hello_job))\n", "path": "examples/hello.py"}]}
| 777 | 199 |
gh_patches_debug_18244
|
rasdani/github-patches
|
git_diff
|
interlegis__sapl-1606
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Exclusão Tramitação - status de tramitação
Ao incluir uma tramitação à matéria, em que esta tramitação tenha seu indicador de tramitação definido nas tabelas auxiliares como fim, a matéria em questão passa de "em tramitação - sim para não" corretamente. Se essa tramitação por ventura for excluída, não seria o caso de alterar novamente a matéria de "em tramitação - não para sim" ?
</issue>
<code>
[start of sapl/materia/receivers.py]
1 from django.dispatch import receiver
2
3 from sapl.materia.signals import tramitacao_signal
4 from sapl.utils import get_base_url
5
6 from .email_utils import do_envia_email_tramitacao
7
8
9 @receiver(tramitacao_signal)
10 def handle_tramitacao_signal(sender, **kwargs):
11 tramitacao = kwargs.get("post")
12 request = kwargs.get("request")
13 materia = tramitacao.materia
14
15 do_envia_email_tramitacao(
16 get_base_url(request),
17 materia,
18 tramitacao.status,
19 tramitacao.unidade_tramitacao_destino)
20
[end of sapl/materia/receivers.py]
[start of sapl/materia/apps.py]
1 from django import apps
2 from django.utils.translation import ugettext_lazy as _
3
4
5 class AppConfig(apps.AppConfig):
6 name = 'sapl.materia'
7 label = 'materia'
8 verbose_name = _('Matéria')
9
[end of sapl/materia/apps.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/sapl/materia/apps.py b/sapl/materia/apps.py
--- a/sapl/materia/apps.py
+++ b/sapl/materia/apps.py
@@ -6,3 +6,6 @@
name = 'sapl.materia'
label = 'materia'
verbose_name = _('Matéria')
+
+ def ready(self):
+ from . import receivers
\ No newline at end of file
diff --git a/sapl/materia/receivers.py b/sapl/materia/receivers.py
--- a/sapl/materia/receivers.py
+++ b/sapl/materia/receivers.py
@@ -1,5 +1,7 @@
+from django.db.models.signals import post_delete, post_save
from django.dispatch import receiver
+from sapl.materia.models import Tramitacao
from sapl.materia.signals import tramitacao_signal
from sapl.utils import get_base_url
@@ -17,3 +19,11 @@
materia,
tramitacao.status,
tramitacao.unidade_tramitacao_destino)
+
+
+@receiver(post_delete, sender=Tramitacao)
+def status_tramitacao_materia(sender, instance, **kwargs):
+ if instance.turno == 'F':
+ materia = instance.materia
+ materia.em_tramitacao = True
+ materia.save()
|
{"golden_diff": "diff --git a/sapl/materia/apps.py b/sapl/materia/apps.py\n--- a/sapl/materia/apps.py\n+++ b/sapl/materia/apps.py\n@@ -6,3 +6,6 @@\n name = 'sapl.materia'\n label = 'materia'\n verbose_name = _('Mat\u00e9ria')\n+\n+ def ready(self):\n+ from . import receivers\n\\ No newline at end of file\ndiff --git a/sapl/materia/receivers.py b/sapl/materia/receivers.py\n--- a/sapl/materia/receivers.py\n+++ b/sapl/materia/receivers.py\n@@ -1,5 +1,7 @@\n+from django.db.models.signals import post_delete, post_save\n from django.dispatch import receiver\n \n+from sapl.materia.models import Tramitacao\n from sapl.materia.signals import tramitacao_signal\n from sapl.utils import get_base_url\n \n@@ -17,3 +19,11 @@\n materia,\n tramitacao.status,\n tramitacao.unidade_tramitacao_destino)\n+\n+\n+@receiver(post_delete, sender=Tramitacao)\n+def status_tramitacao_materia(sender, instance, **kwargs):\n+ if instance.turno == 'F':\n+ materia = instance.materia\n+ materia.em_tramitacao = True\n+ materia.save()\n", "issue": "Exclus\u00e3o Tramita\u00e7\u00e3o - status de tramita\u00e7\u00e3o\nAo incluir uma tramita\u00e7\u00e3o \u00e0 mat\u00e9ria, em que esta tramita\u00e7\u00e3o tenha seu indicador de tramita\u00e7\u00e3o definido nas tabelas auxiliares como fim, a mat\u00e9ria em quest\u00e3o passa de \"em tramita\u00e7\u00e3o - sim para n\u00e3o\" corretamente. Se essa tramita\u00e7\u00e3o por ventura for exclu\u00edda, n\u00e3o seria o caso de alterar novamente a mat\u00e9ria de \"em tramita\u00e7\u00e3o - n\u00e3o para sim\" ?\n", "before_files": [{"content": "from django.dispatch import receiver\n\nfrom sapl.materia.signals import tramitacao_signal\nfrom sapl.utils import get_base_url\n\nfrom .email_utils import do_envia_email_tramitacao\n\n\n@receiver(tramitacao_signal)\ndef handle_tramitacao_signal(sender, **kwargs):\n tramitacao = kwargs.get(\"post\")\n request = kwargs.get(\"request\")\n materia = tramitacao.materia\n\n do_envia_email_tramitacao(\n get_base_url(request),\n materia,\n tramitacao.status,\n tramitacao.unidade_tramitacao_destino)\n", "path": "sapl/materia/receivers.py"}, {"content": "from django import apps\nfrom django.utils.translation import ugettext_lazy as _\n\n\nclass AppConfig(apps.AppConfig):\n name = 'sapl.materia'\n label = 'materia'\n verbose_name = _('Mat\u00e9ria')\n", "path": "sapl/materia/apps.py"}]}
| 885 | 300 |
gh_patches_debug_34214
|
rasdani/github-patches
|
git_diff
|
deepchecks__deepchecks-740
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
[FEAT] Add force permutation option in UnusedFeatures check
`UnusedFeatures` looks for the feature importance attr if possible, which is bias in some [sklearn models](https://explained.ai/rf-importance/).
I suggest adding an option to override this behaviour. This is already an option in `.utils.features.calculate_feature_importance`.
</issue>
<code>
[start of deepchecks/checks/methodology/unused_features.py]
1 # ----------------------------------------------------------------------------
2 # Copyright (C) 2021 Deepchecks (https://www.deepchecks.com)
3 #
4 # This file is part of Deepchecks.
5 # Deepchecks is distributed under the terms of the GNU Affero General
6 # Public License (version 3 or later).
7 # You should have received a copy of the GNU Affero General Public License
8 # along with Deepchecks. If not, see <http://www.gnu.org/licenses/>.
9 # ----------------------------------------------------------------------------
10 #
11 """The UnusedFeatures check module."""
12 from typing import Tuple
13
14 import plotly.graph_objects as go
15 import numpy as np
16 import pandas as pd
17 from sklearn.base import TransformerMixin
18 from sklearn.compose import ColumnTransformer
19 from sklearn.decomposition import PCA
20 from sklearn.impute import SimpleImputer
21 from sklearn.pipeline import Pipeline
22 from sklearn.preprocessing import RobustScaler, OrdinalEncoder
23
24 from deepchecks.base.check_context import CheckRunContext
25 from deepchecks import Dataset, CheckResult, TrainTestBaseCheck, ConditionResult, ConditionCategory
26
27
28 __all__ = ['UnusedFeatures']
29
30
31 class UnusedFeatures(TrainTestBaseCheck):
32 """Detect features that are nearly unused by the model.
33
34 The check uses feature importance (either internally computed in appropriate models or calculated by permutation
35 feature importance) to detect features that are not used by the model. From this list, the check sorts the features
36 by their variance (as calculated by a PCA transformation). High variance unused features may be containing
37 information that is ignored by the model.
38
39 Parameters
40 ----------
41 feature_importance_threshold : float , default: 0.2
42 A cutoff value for the feature importance, measured by the ratio of
43 each features' feature importance to the mean feature importance. Features with lower importance
44 are not shown in the check display.
45 feature_variance_threshold : float , default: 0.4
46 A cutoff value for the feature variance, measured by the ratio of
47 each features' feature variance to the mean feature variance. Unused features with lower variance
48 are not shown in the check display.
49 n_top_fi_to_show : int , default: 5
50 The max number of important features to show in the check display.
51 n_top_unused_to_show : int , default: 15
52 The max number of unused features to show in the check display, from among
53 unused features that have higher variance then is defined by feature_variance_threshold.
54 random_state : int , default: 42
55 The random state to use for permutation feature importance and PCA.
56 """
57
58 def __init__(self, feature_importance_threshold: float = 0.2, feature_variance_threshold: float = 0.4,
59 n_top_fi_to_show: int = 5, n_top_unused_to_show: int = 15, random_state: int = 42):
60 super().__init__()
61 self.feature_importance_threshold = feature_importance_threshold
62 self.feature_variance_threshold = feature_variance_threshold
63 self.n_top_fi_to_show = n_top_fi_to_show
64 self.n_top_unused_to_show = n_top_unused_to_show
65 self.random_state = random_state
66
67 def run_logic(self, context: CheckRunContext) -> CheckResult:
68 """Run check.
69
70 Returns
71 -------
72 CheckResult
73 value is a dataframe with metrics as indexes, and scores per training and test in the columns.
74 display data is a bar graph of the metrics for training and test data.
75
76 Raises
77 ------
78 DeepchecksValueError
79 If neither train dataset nor test dataset exist, or either of the dataset objects are
80 not a Dataset instance with a label.
81 """
82 if context.have_test():
83 dataset = context.test
84 else:
85 dataset = context.train
86
87 feature_importance = context.features_importance
88 features = context.features
89
90 # Calculate normalized variance per feature based on PCA decomposition
91 pre_pca_transformer, var_col_order = naive_encoder(dataset)
92 pca_trans = PCA(n_components=len(dataset.features) // 2, random_state=self.random_state)
93 n_samples = min(10000, dataset.n_samples)
94 pca_trans.fit(pre_pca_transformer.fit_transform(
95 dataset.data[features].sample(n_samples, random_state=self.random_state)
96 ))
97
98 feature_normed_variance = pd.Series(np.abs(pca_trans.components_).sum(axis=0), index=var_col_order)
99 feature_normed_variance = feature_normed_variance / feature_normed_variance.sum()
100
101 feature_df = pd.concat([feature_importance, feature_normed_variance], axis=1)
102 feature_df.columns = ['Feature Importance', 'Feature Variance']
103 feature_df.sort_values(by='Feature Importance', ascending=False, inplace=True)
104
105 # For feature importance and variance, calculate their "ratio to average" per feature. The ratio to average
106 # is, for example, the amount of feature importance a feature has, divided by the the amount he would have
107 # if all features where equally important (which is basically 1 / n_of_features).
108 feature_ratio_to_avg_df = feature_df / (1 / len(feature_importance))
109
110 # Find last viable feature (not unused). All features from there on are sorted by variance
111 last_important_feature_index = sum(
112 feature_ratio_to_avg_df['Feature Importance'] > self.feature_importance_threshold
113 ) - 1
114
115 unviable_feature_df = feature_df.iloc[(last_important_feature_index + 1):]
116 # Only display if there are features considered unimportant
117 if not unviable_feature_df.empty:
118 unviable_feature_df.sort_values(by='Feature Variance', ascending=False, inplace=True)
119 unviable_feature_ratio_to_avg_df = unviable_feature_df / (1 / len(feature_df))
120 last_variable_feature_index = sum(
121 unviable_feature_ratio_to_avg_df['Feature Variance'] > self.feature_variance_threshold
122 )
123
124 # limit display to n_top_to_show params
125 display_feature_df = pd.concat(
126 [feature_df.iloc[:(last_important_feature_index + 1)].head(self.n_top_fi_to_show),
127 unviable_feature_df.iloc[:last_variable_feature_index].head(self.n_top_unused_to_show)],
128 axis=0)
129
130 fig = go.Figure()
131 fig.add_trace(go.Bar(
132 y=display_feature_df.index,
133 x=display_feature_df['Feature Importance'].multiply(100).values.flatten(),
134 name='Feature Importance %',
135 marker_color='indianred',
136 orientation='h'
137 ))
138 fig.add_trace(go.Bar(
139 y=display_feature_df.index,
140 x=display_feature_df['Feature Variance'].multiply(100).values.flatten(),
141 name='Feature Variance %',
142 marker_color='lightsalmon',
143 orientation='h'
144 ))
145
146 fig.update_yaxes(autorange='reversed')
147 fig.update_layout(title_text='Unused features compared to top important features',
148 width=800, height=500)
149
150 last_important_feature_index_to_plot = min(last_important_feature_index, self.n_top_fi_to_show - 1)
151
152 if last_important_feature_index_to_plot < len(display_feature_df) - 1:
153 last_important_feature_line_loc = last_important_feature_index_to_plot + 0.5
154 fig.add_hline(y=last_important_feature_line_loc, line_width=2, line_dash='dash', line_color='green',
155 annotation_text='Last shown significant feature')
156
157 # display only if high variance features exist (as set by self.feature_variance_threshold)
158 if not last_variable_feature_index:
159 display_list = []
160 else:
161 display_list = [
162 'Features above the line are a sample of the most important features, while the features '
163 'below the line are the unused features with highest variance, as defined by check'
164 ' parameters', fig]
165
166 else:
167 display_list = []
168
169 return_value = {
170 'used features': feature_df.index[:(last_important_feature_index + 1)].values.tolist(),
171 'unused features': {
172 'high variance': (
173 [] if unviable_feature_df.empty
174 else unviable_feature_df.index[:last_variable_feature_index].values.tolist()
175 ),
176 'low variance': (
177 [] if unviable_feature_df.empty
178 else unviable_feature_df.index[last_variable_feature_index:].values.tolist()
179 )
180 }}
181
182 return CheckResult(return_value, header='Unused Features', display=display_list)
183
184 def add_condition_number_of_high_variance_unused_features_not_greater_than(
185 self, max_high_variance_unused_features: int = 5):
186 """Add condition - require number of high variance unused features to be not greater than a given number.
187
188 Parameters
189 ----------
190 max_high_variance_unused_features : int , default: 5
191 Maximum allowed number of high variance unused features.
192 """
193 def max_high_variance_unused_features_condition(result: dict) -> ConditionResult:
194 high_var_features = result['unused features']['high variance']
195 if len(high_var_features) > max_high_variance_unused_features:
196 return ConditionResult(
197 False,
198 f'Found number of unused high variance features above threshold: {high_var_features}',
199 category=ConditionCategory.WARN)
200 else:
201 return ConditionResult(True)
202
203 return self.add_condition(f'Number of high variance unused features is not greater than'
204 f' {max_high_variance_unused_features}',
205 max_high_variance_unused_features_condition)
206
207
208 def naive_encoder(dataset: Dataset) -> Tuple[TransformerMixin, list]:
209 """Create a naive encoder for categorical and numerical features.
210
211 The encoder handles nans for all features and uses label encoder for categorical features. Then, all features are
212 scaled using RobustScaler.
213
214 Parameters
215 ----------
216 dataset : Dataset
217 The dataset to encode.
218
219 Returns
220 -------
221 Tuple[TransformerMixin, list]
222 A transformer object, a list of columns returned
223 """
224 numeric_features = [col for col in dataset.features if col not in dataset.cat_features]
225
226 return ColumnTransformer(
227 transformers=[
228 ('num', Pipeline([
229 ('nan_handling', SimpleImputer()),
230 ('norm', RobustScaler())
231 ]),
232 numeric_features),
233 ('cat',
234 Pipeline([
235 ('nan_handling', SimpleImputer(strategy='most_frequent')),
236 ('encode', OrdinalEncoder(handle_unknown='use_encoded_value', unknown_value=-1)),
237 ('norm', RobustScaler())
238 ]),
239 dataset.cat_features)
240 ]
241 ), numeric_features + dataset.cat_features
242
[end of deepchecks/checks/methodology/unused_features.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/deepchecks/checks/methodology/unused_features.py b/deepchecks/checks/methodology/unused_features.py
--- a/deepchecks/checks/methodology/unused_features.py
+++ b/deepchecks/checks/methodology/unused_features.py
@@ -23,6 +23,7 @@
from deepchecks.base.check_context import CheckRunContext
from deepchecks import Dataset, CheckResult, TrainTestBaseCheck, ConditionResult, ConditionCategory
+from deepchecks.utils.typing import BasicModel
__all__ = ['UnusedFeatures']
@@ -64,8 +65,26 @@
self.n_top_unused_to_show = n_top_unused_to_show
self.random_state = random_state
- def run_logic(self, context: CheckRunContext) -> CheckResult:
- """Run check.
+ def run(self,
+ train_dataset: Dataset,
+ test_dataset: Dataset,
+ model: BasicModel = None,
+ feature_importance_force_permutation: bool = False,
+ feature_importance_timeout: int = None) -> CheckResult:
+ """Run the check.
+
+ Parameters
+ ----------
+ train_dataset : Dataset
+ dataset representing data an estimator was fitted on
+ test_dataset : Dataset
+ dataset representing data an estimator predicts on
+ model : BasicModel
+ A scikit-learn-compatible fitted estimator instance
+ feature_importance_force_permutation : bool , default: False
+ force calculation of permutation features importance
+ feature_importance_timeout : int , default: None
+ timeout in second for the permutation features importance calculation
Returns
-------
@@ -79,6 +98,13 @@
If neither train dataset nor test dataset exist, or either of the dataset objects are
not a Dataset instance with a label.
"""
+ c = CheckRunContext(train_dataset, test_dataset, model,
+ feature_importance_force_permutation=feature_importance_force_permutation,
+ feature_importance_timeout=feature_importance_timeout)
+ return self.run_logic(c)
+
+ def run_logic(self, context: CheckRunContext) -> CheckResult:
+ """Run check."""
if context.have_test():
dataset = context.test
else:
|
{"golden_diff": "diff --git a/deepchecks/checks/methodology/unused_features.py b/deepchecks/checks/methodology/unused_features.py\n--- a/deepchecks/checks/methodology/unused_features.py\n+++ b/deepchecks/checks/methodology/unused_features.py\n@@ -23,6 +23,7 @@\n \n from deepchecks.base.check_context import CheckRunContext\n from deepchecks import Dataset, CheckResult, TrainTestBaseCheck, ConditionResult, ConditionCategory\n+from deepchecks.utils.typing import BasicModel\n \n \n __all__ = ['UnusedFeatures']\n@@ -64,8 +65,26 @@\n self.n_top_unused_to_show = n_top_unused_to_show\n self.random_state = random_state\n \n- def run_logic(self, context: CheckRunContext) -> CheckResult:\n- \"\"\"Run check.\n+ def run(self,\n+ train_dataset: Dataset,\n+ test_dataset: Dataset,\n+ model: BasicModel = None,\n+ feature_importance_force_permutation: bool = False,\n+ feature_importance_timeout: int = None) -> CheckResult:\n+ \"\"\"Run the check.\n+\n+ Parameters\n+ ----------\n+ train_dataset : Dataset\n+ dataset representing data an estimator was fitted on\n+ test_dataset : Dataset\n+ dataset representing data an estimator predicts on\n+ model : BasicModel\n+ A scikit-learn-compatible fitted estimator instance\n+ feature_importance_force_permutation : bool , default: False\n+ force calculation of permutation features importance\n+ feature_importance_timeout : int , default: None\n+ timeout in second for the permutation features importance calculation\n \n Returns\n -------\n@@ -79,6 +98,13 @@\n If neither train dataset nor test dataset exist, or either of the dataset objects are\n not a Dataset instance with a label.\n \"\"\"\n+ c = CheckRunContext(train_dataset, test_dataset, model,\n+ feature_importance_force_permutation=feature_importance_force_permutation,\n+ feature_importance_timeout=feature_importance_timeout)\n+ return self.run_logic(c)\n+\n+ def run_logic(self, context: CheckRunContext) -> CheckResult:\n+ \"\"\"Run check.\"\"\"\n if context.have_test():\n dataset = context.test\n else:\n", "issue": "[FEAT] Add force permutation option in UnusedFeatures check\n`UnusedFeatures` looks for the feature importance attr if possible, which is bias in some [sklearn models](https://explained.ai/rf-importance/).\r\nI suggest adding an option to override this behaviour. This is already an option in `.utils.features.calculate_feature_importance`.\r\n\r\n\r\n\n", "before_files": [{"content": "# ----------------------------------------------------------------------------\n# Copyright (C) 2021 Deepchecks (https://www.deepchecks.com)\n#\n# This file is part of Deepchecks.\n# Deepchecks is distributed under the terms of the GNU Affero General\n# Public License (version 3 or later).\n# You should have received a copy of the GNU Affero General Public License\n# along with Deepchecks. If not, see <http://www.gnu.org/licenses/>.\n# ----------------------------------------------------------------------------\n#\n\"\"\"The UnusedFeatures check module.\"\"\"\nfrom typing import Tuple\n\nimport plotly.graph_objects as go\nimport numpy as np\nimport pandas as pd\nfrom sklearn.base import TransformerMixin\nfrom sklearn.compose import ColumnTransformer\nfrom sklearn.decomposition import PCA\nfrom sklearn.impute import SimpleImputer\nfrom sklearn.pipeline import Pipeline\nfrom sklearn.preprocessing import RobustScaler, OrdinalEncoder\n\nfrom deepchecks.base.check_context import CheckRunContext\nfrom deepchecks import Dataset, CheckResult, TrainTestBaseCheck, ConditionResult, ConditionCategory\n\n\n__all__ = ['UnusedFeatures']\n\n\nclass UnusedFeatures(TrainTestBaseCheck):\n \"\"\"Detect features that are nearly unused by the model.\n\n The check uses feature importance (either internally computed in appropriate models or calculated by permutation\n feature importance) to detect features that are not used by the model. From this list, the check sorts the features\n by their variance (as calculated by a PCA transformation). High variance unused features may be containing\n information that is ignored by the model.\n\n Parameters\n ----------\n feature_importance_threshold : float , default: 0.2\n A cutoff value for the feature importance, measured by the ratio of\n each features' feature importance to the mean feature importance. Features with lower importance\n are not shown in the check display.\n feature_variance_threshold : float , default: 0.4\n A cutoff value for the feature variance, measured by the ratio of\n each features' feature variance to the mean feature variance. Unused features with lower variance\n are not shown in the check display.\n n_top_fi_to_show : int , default: 5\n The max number of important features to show in the check display.\n n_top_unused_to_show : int , default: 15\n The max number of unused features to show in the check display, from among\n unused features that have higher variance then is defined by feature_variance_threshold.\n random_state : int , default: 42\n The random state to use for permutation feature importance and PCA.\n \"\"\"\n\n def __init__(self, feature_importance_threshold: float = 0.2, feature_variance_threshold: float = 0.4,\n n_top_fi_to_show: int = 5, n_top_unused_to_show: int = 15, random_state: int = 42):\n super().__init__()\n self.feature_importance_threshold = feature_importance_threshold\n self.feature_variance_threshold = feature_variance_threshold\n self.n_top_fi_to_show = n_top_fi_to_show\n self.n_top_unused_to_show = n_top_unused_to_show\n self.random_state = random_state\n\n def run_logic(self, context: CheckRunContext) -> CheckResult:\n \"\"\"Run check.\n\n Returns\n -------\n CheckResult\n value is a dataframe with metrics as indexes, and scores per training and test in the columns.\n display data is a bar graph of the metrics for training and test data.\n\n Raises\n ------\n DeepchecksValueError\n If neither train dataset nor test dataset exist, or either of the dataset objects are\n not a Dataset instance with a label.\n \"\"\"\n if context.have_test():\n dataset = context.test\n else:\n dataset = context.train\n\n feature_importance = context.features_importance\n features = context.features\n\n # Calculate normalized variance per feature based on PCA decomposition\n pre_pca_transformer, var_col_order = naive_encoder(dataset)\n pca_trans = PCA(n_components=len(dataset.features) // 2, random_state=self.random_state)\n n_samples = min(10000, dataset.n_samples)\n pca_trans.fit(pre_pca_transformer.fit_transform(\n dataset.data[features].sample(n_samples, random_state=self.random_state)\n ))\n\n feature_normed_variance = pd.Series(np.abs(pca_trans.components_).sum(axis=0), index=var_col_order)\n feature_normed_variance = feature_normed_variance / feature_normed_variance.sum()\n\n feature_df = pd.concat([feature_importance, feature_normed_variance], axis=1)\n feature_df.columns = ['Feature Importance', 'Feature Variance']\n feature_df.sort_values(by='Feature Importance', ascending=False, inplace=True)\n\n # For feature importance and variance, calculate their \"ratio to average\" per feature. The ratio to average\n # is, for example, the amount of feature importance a feature has, divided by the the amount he would have\n # if all features where equally important (which is basically 1 / n_of_features).\n feature_ratio_to_avg_df = feature_df / (1 / len(feature_importance))\n\n # Find last viable feature (not unused). All features from there on are sorted by variance\n last_important_feature_index = sum(\n feature_ratio_to_avg_df['Feature Importance'] > self.feature_importance_threshold\n ) - 1\n\n unviable_feature_df = feature_df.iloc[(last_important_feature_index + 1):]\n # Only display if there are features considered unimportant\n if not unviable_feature_df.empty:\n unviable_feature_df.sort_values(by='Feature Variance', ascending=False, inplace=True)\n unviable_feature_ratio_to_avg_df = unviable_feature_df / (1 / len(feature_df))\n last_variable_feature_index = sum(\n unviable_feature_ratio_to_avg_df['Feature Variance'] > self.feature_variance_threshold\n )\n\n # limit display to n_top_to_show params\n display_feature_df = pd.concat(\n [feature_df.iloc[:(last_important_feature_index + 1)].head(self.n_top_fi_to_show),\n unviable_feature_df.iloc[:last_variable_feature_index].head(self.n_top_unused_to_show)],\n axis=0)\n\n fig = go.Figure()\n fig.add_trace(go.Bar(\n y=display_feature_df.index,\n x=display_feature_df['Feature Importance'].multiply(100).values.flatten(),\n name='Feature Importance %',\n marker_color='indianred',\n orientation='h'\n ))\n fig.add_trace(go.Bar(\n y=display_feature_df.index,\n x=display_feature_df['Feature Variance'].multiply(100).values.flatten(),\n name='Feature Variance %',\n marker_color='lightsalmon',\n orientation='h'\n ))\n\n fig.update_yaxes(autorange='reversed')\n fig.update_layout(title_text='Unused features compared to top important features',\n width=800, height=500)\n\n last_important_feature_index_to_plot = min(last_important_feature_index, self.n_top_fi_to_show - 1)\n\n if last_important_feature_index_to_plot < len(display_feature_df) - 1:\n last_important_feature_line_loc = last_important_feature_index_to_plot + 0.5\n fig.add_hline(y=last_important_feature_line_loc, line_width=2, line_dash='dash', line_color='green',\n annotation_text='Last shown significant feature')\n\n # display only if high variance features exist (as set by self.feature_variance_threshold)\n if not last_variable_feature_index:\n display_list = []\n else:\n display_list = [\n 'Features above the line are a sample of the most important features, while the features '\n 'below the line are the unused features with highest variance, as defined by check'\n ' parameters', fig]\n\n else:\n display_list = []\n\n return_value = {\n 'used features': feature_df.index[:(last_important_feature_index + 1)].values.tolist(),\n 'unused features': {\n 'high variance': (\n [] if unviable_feature_df.empty\n else unviable_feature_df.index[:last_variable_feature_index].values.tolist()\n ),\n 'low variance': (\n [] if unviable_feature_df.empty\n else unviable_feature_df.index[last_variable_feature_index:].values.tolist()\n )\n }}\n\n return CheckResult(return_value, header='Unused Features', display=display_list)\n\n def add_condition_number_of_high_variance_unused_features_not_greater_than(\n self, max_high_variance_unused_features: int = 5):\n \"\"\"Add condition - require number of high variance unused features to be not greater than a given number.\n\n Parameters\n ----------\n max_high_variance_unused_features : int , default: 5\n Maximum allowed number of high variance unused features.\n \"\"\"\n def max_high_variance_unused_features_condition(result: dict) -> ConditionResult:\n high_var_features = result['unused features']['high variance']\n if len(high_var_features) > max_high_variance_unused_features:\n return ConditionResult(\n False,\n f'Found number of unused high variance features above threshold: {high_var_features}',\n category=ConditionCategory.WARN)\n else:\n return ConditionResult(True)\n\n return self.add_condition(f'Number of high variance unused features is not greater than'\n f' {max_high_variance_unused_features}',\n max_high_variance_unused_features_condition)\n\n\ndef naive_encoder(dataset: Dataset) -> Tuple[TransformerMixin, list]:\n \"\"\"Create a naive encoder for categorical and numerical features.\n\n The encoder handles nans for all features and uses label encoder for categorical features. Then, all features are\n scaled using RobustScaler.\n\n Parameters\n ----------\n dataset : Dataset\n The dataset to encode.\n\n Returns\n -------\n Tuple[TransformerMixin, list]\n A transformer object, a list of columns returned\n \"\"\"\n numeric_features = [col for col in dataset.features if col not in dataset.cat_features]\n\n return ColumnTransformer(\n transformers=[\n ('num', Pipeline([\n ('nan_handling', SimpleImputer()),\n ('norm', RobustScaler())\n ]),\n numeric_features),\n ('cat',\n Pipeline([\n ('nan_handling', SimpleImputer(strategy='most_frequent')),\n ('encode', OrdinalEncoder(handle_unknown='use_encoded_value', unknown_value=-1)),\n ('norm', RobustScaler())\n ]),\n dataset.cat_features)\n ]\n ), numeric_features + dataset.cat_features\n", "path": "deepchecks/checks/methodology/unused_features.py"}]}
| 3,480 | 490 |
gh_patches_debug_4716
|
rasdani/github-patches
|
git_diff
|
crytic__slither-2331
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
[Bug]: High level call does not always have function
### Describe the issue:
I have two somehow similar examples, in the first one I have the function on the high level call in function safeAdd. In the next example I edited the array size to be constant instead of a literal, and I got None instead of a function object
### Code example to reproduce the issue:
```solidity
library SafeMath {
uint256 private constant twelve = 12;
struct A {uint256 a;}
function add(A[twelve] storage z) internal { }
}
contract MathContract {
uint256 private constant twelve = 12;
using SafeMath for SafeMath.A[12];
SafeMath.A[12] public z;
function safeAdd() public {
z.add();
}
}
```
```solidity
library SafeMath {
uint256 private constant twelve = 12;
struct A {uint256 a;}
function add(A[twelve] storage z) internal { }
}
contract MathContract {
uint256 private constant twelve = 12;
using SafeMath for SafeMath.A[twelve];
SafeMath.A[twelve] public z;
function safeAdd() public {
z.add();
}
}
```
### Version:
0.10.0
### Relevant log output:
```shell
>>> from slither import Slither
>>> math = Slither('a.sol').contracts[1]
>>> math.name
'MathContract'
>>> f = math.functions[0]
>>> f.name
'safeAdd'
>>> f.nodes
[<slither.core.cfg.node.Node object at 0x7f5460aa1e50>, <slither.core.cfg.node.No
de object at 0x7f5460aa2090>]
>>> f.nodes[1]
<slither.core.cfg.node.Node object at 0x7f5460aa2090>
>>> f.nodes[1].irs
[<slither.slithir.operations.library_call.LibraryCall object at 0x7f5460a748d0>]
>>> f.nodes[1].irs[0].function
<slither.core.declarations.function_contract.FunctionContract object at 0x7f5460a
9e090>
>>> f.nodes[1].irs[0].function.name
'add'
----------------------------------------------------------------------------------
>>> from slither import Slither
>>> math = Slither('a.sol').contracts[1]
>>> math.name
'MathContract'
>>> f = math.functions[0]
>>> f.name
'safeAdd'
>>> f.nodes
[<slither.core.cfg.node.Node object at 0x7f9d6379db10>, <slither.core.cfg.node.No
de object at 0x7f9d63a47850>]
>>> f.nodes[1]
<slither.core.cfg.node.Node object at 0x7f9d63a47850>
>>> f.nodes[1].irs
[<slither.slithir.operations.high_level_call.HighLevelCall object at 0x7f9d63a376
90>]
>>> f.nodes[1].irs[0].function
>>> print(f.nodes[1].irs[0].function)
None
```
</issue>
<code>
[start of slither/core/solidity_types/array_type.py]
1 from typing import Union, Optional, Tuple, Any, TYPE_CHECKING
2
3 from slither.core.expressions.expression import Expression
4 from slither.core.expressions.literal import Literal
5 from slither.core.solidity_types.elementary_type import ElementaryType
6 from slither.core.solidity_types.type import Type
7 from slither.visitors.expression.constants_folding import ConstantFolding
8
9 if TYPE_CHECKING:
10 from slither.core.expressions.binary_operation import BinaryOperation
11 from slither.core.expressions.identifier import Identifier
12
13
14 class ArrayType(Type):
15 def __init__(
16 self,
17 t: Type,
18 length: Optional[Union["Identifier", Literal, "BinaryOperation", int]],
19 ) -> None:
20 assert isinstance(t, Type)
21 if length:
22 if isinstance(length, int):
23 length = Literal(length, ElementaryType("uint256"))
24
25 super().__init__()
26 self._type: Type = t
27 assert length is None or isinstance(length, Expression)
28 self._length: Optional[Expression] = length
29
30 if length:
31 if not isinstance(length, Literal):
32 cf = ConstantFolding(length, "uint256")
33 length = cf.result()
34 self._length_value: Optional[Literal] = length
35 else:
36 self._length_value = None
37
38 @property
39 def type(self) -> Type:
40 return self._type
41
42 @property
43 def is_dynamic(self) -> bool:
44 return self.length is None
45
46 @property
47 def length(self) -> Optional[Expression]:
48 return self._length
49
50 @property
51 def length_value(self) -> Optional[Literal]:
52 return self._length_value
53
54 @property
55 def is_fixed_array(self) -> bool:
56 return bool(self.length)
57
58 @property
59 def is_dynamic_array(self) -> bool:
60 return not self.is_fixed_array
61
62 @property
63 def storage_size(self) -> Tuple[int, bool]:
64 if self._length_value:
65 elem_size, _ = self._type.storage_size
66 return elem_size * int(str(self._length_value)), True
67 return 32, True
68
69 def __str__(self) -> str:
70 if self._length:
71 return str(self._type) + f"[{str(self._length_value)}]"
72 return str(self._type) + "[]"
73
74 def __eq__(self, other: Any) -> bool:
75 if not isinstance(other, ArrayType):
76 return False
77 return self._type == other.type and self.length == other.length
78
79 def __hash__(self) -> int:
80 return hash(str(self))
81
[end of slither/core/solidity_types/array_type.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/slither/core/solidity_types/array_type.py b/slither/core/solidity_types/array_type.py
--- a/slither/core/solidity_types/array_type.py
+++ b/slither/core/solidity_types/array_type.py
@@ -74,7 +74,7 @@
def __eq__(self, other: Any) -> bool:
if not isinstance(other, ArrayType):
return False
- return self._type == other.type and self.length == other.length
+ return self._type == other.type and self._length_value == other.length_value
def __hash__(self) -> int:
return hash(str(self))
|
{"golden_diff": "diff --git a/slither/core/solidity_types/array_type.py b/slither/core/solidity_types/array_type.py\n--- a/slither/core/solidity_types/array_type.py\n+++ b/slither/core/solidity_types/array_type.py\n@@ -74,7 +74,7 @@\n def __eq__(self, other: Any) -> bool:\n if not isinstance(other, ArrayType):\n return False\n- return self._type == other.type and self.length == other.length\n+ return self._type == other.type and self._length_value == other.length_value\n \n def __hash__(self) -> int:\n return hash(str(self))\n", "issue": "[Bug]: High level call does not always have function\n### Describe the issue:\r\n\r\nI have two somehow similar examples, in the first one I have the function on the high level call in function safeAdd. In the next example I edited the array size to be constant instead of a literal, and I got None instead of a function object\r\n\r\n### Code example to reproduce the issue:\r\n\r\n```solidity\r\nlibrary SafeMath {\r\n uint256 private constant twelve = 12; \r\n struct A {uint256 a;}\r\n function add(A[twelve] storage z) internal { }\r\n}\r\n\r\ncontract MathContract {\r\n uint256 private constant twelve = 12; \r\n using SafeMath for SafeMath.A[12];\r\n SafeMath.A[12] public z;\r\n function safeAdd() public {\r\n z.add();\r\n }\r\n}\r\n```\r\n```solidity\r\nlibrary SafeMath {\r\n uint256 private constant twelve = 12; \r\n struct A {uint256 a;}\r\n function add(A[twelve] storage z) internal { }\r\n}\r\n\r\ncontract MathContract {\r\n uint256 private constant twelve = 12; \r\n using SafeMath for SafeMath.A[twelve];\r\n SafeMath.A[twelve] public z;\r\n function safeAdd() public {\r\n z.add();\r\n }\r\n}\r\n```\r\n\r\n### Version:\r\n\r\n0.10.0\r\n\r\n### Relevant log output:\r\n\r\n```shell\r\n>>> from slither import Slither\r\n>>> math = Slither('a.sol').contracts[1]\r\n>>> math.name\r\n'MathContract'\r\n>>> f = math.functions[0]\r\n>>> f.name\r\n'safeAdd'\r\n>>> f.nodes\r\n[<slither.core.cfg.node.Node object at 0x7f5460aa1e50>, <slither.core.cfg.node.No\r\nde object at 0x7f5460aa2090>]\r\n>>> f.nodes[1]\r\n<slither.core.cfg.node.Node object at 0x7f5460aa2090>\r\n>>> f.nodes[1].irs\r\n[<slither.slithir.operations.library_call.LibraryCall object at 0x7f5460a748d0>]\r\n>>> f.nodes[1].irs[0].function\r\n<slither.core.declarations.function_contract.FunctionContract object at 0x7f5460a\r\n9e090>\r\n>>> f.nodes[1].irs[0].function.name\r\n'add'\r\n----------------------------------------------------------------------------------\r\n>>> from slither import Slither\r\n>>> math = Slither('a.sol').contracts[1]\r\n>>> math.name\r\n'MathContract'\r\n>>> f = math.functions[0]\r\n>>> f.name\r\n'safeAdd'\r\n>>> f.nodes\r\n[<slither.core.cfg.node.Node object at 0x7f9d6379db10>, <slither.core.cfg.node.No\r\nde object at 0x7f9d63a47850>]\r\n>>> f.nodes[1]\r\n<slither.core.cfg.node.Node object at 0x7f9d63a47850>\r\n>>> f.nodes[1].irs\r\n[<slither.slithir.operations.high_level_call.HighLevelCall object at 0x7f9d63a376\r\n90>]\r\n>>> f.nodes[1].irs[0].function\r\n>>> print(f.nodes[1].irs[0].function)\r\nNone\r\n```\r\n\n", "before_files": [{"content": "from typing import Union, Optional, Tuple, Any, TYPE_CHECKING\n\nfrom slither.core.expressions.expression import Expression\nfrom slither.core.expressions.literal import Literal\nfrom slither.core.solidity_types.elementary_type import ElementaryType\nfrom slither.core.solidity_types.type import Type\nfrom slither.visitors.expression.constants_folding import ConstantFolding\n\nif TYPE_CHECKING:\n from slither.core.expressions.binary_operation import BinaryOperation\n from slither.core.expressions.identifier import Identifier\n\n\nclass ArrayType(Type):\n def __init__(\n self,\n t: Type,\n length: Optional[Union[\"Identifier\", Literal, \"BinaryOperation\", int]],\n ) -> None:\n assert isinstance(t, Type)\n if length:\n if isinstance(length, int):\n length = Literal(length, ElementaryType(\"uint256\"))\n\n super().__init__()\n self._type: Type = t\n assert length is None or isinstance(length, Expression)\n self._length: Optional[Expression] = length\n\n if length:\n if not isinstance(length, Literal):\n cf = ConstantFolding(length, \"uint256\")\n length = cf.result()\n self._length_value: Optional[Literal] = length\n else:\n self._length_value = None\n\n @property\n def type(self) -> Type:\n return self._type\n\n @property\n def is_dynamic(self) -> bool:\n return self.length is None\n\n @property\n def length(self) -> Optional[Expression]:\n return self._length\n\n @property\n def length_value(self) -> Optional[Literal]:\n return self._length_value\n\n @property\n def is_fixed_array(self) -> bool:\n return bool(self.length)\n\n @property\n def is_dynamic_array(self) -> bool:\n return not self.is_fixed_array\n\n @property\n def storage_size(self) -> Tuple[int, bool]:\n if self._length_value:\n elem_size, _ = self._type.storage_size\n return elem_size * int(str(self._length_value)), True\n return 32, True\n\n def __str__(self) -> str:\n if self._length:\n return str(self._type) + f\"[{str(self._length_value)}]\"\n return str(self._type) + \"[]\"\n\n def __eq__(self, other: Any) -> bool:\n if not isinstance(other, ArrayType):\n return False\n return self._type == other.type and self.length == other.length\n\n def __hash__(self) -> int:\n return hash(str(self))\n", "path": "slither/core/solidity_types/array_type.py"}]}
| 2,003 | 142 |
gh_patches_debug_1930
|
rasdani/github-patches
|
git_diff
|
CTFd__CTFd-1800
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Invalid model identifier
https://github.com/CTFd/CTFd/blob/master/CTFd/themes/core/templates/scoreboard.html#L26
This should change depending on the mode of the CTF
</issue>
<code>
[start of CTFd/plugins/__init__.py]
1 import glob
2 import importlib
3 import os
4 from collections import namedtuple
5
6 from flask import current_app as app
7 from flask import send_file, send_from_directory, url_for
8
9 from CTFd.utils.config.pages import get_pages
10 from CTFd.utils.decorators import admins_only as admins_only_wrapper
11 from CTFd.utils.plugins import override_template as utils_override_template
12 from CTFd.utils.plugins import (
13 register_admin_script as utils_register_admin_plugin_script,
14 )
15 from CTFd.utils.plugins import (
16 register_admin_stylesheet as utils_register_admin_plugin_stylesheet,
17 )
18 from CTFd.utils.plugins import register_script as utils_register_plugin_script
19 from CTFd.utils.plugins import register_stylesheet as utils_register_plugin_stylesheet
20
21 Menu = namedtuple("Menu", ["title", "route"])
22
23
24 def register_plugin_assets_directory(app, base_path, admins_only=False, endpoint=None):
25 """
26 Registers a directory to serve assets
27
28 :param app: A CTFd application
29 :param string base_path: The path to the directory
30 :param boolean admins_only: Whether or not the assets served out of the directory should be accessible to the public
31 :return:
32 """
33 base_path = base_path.strip("/")
34 if endpoint is None:
35 endpoint = base_path.replace("/", ".")
36
37 def assets_handler(path):
38 return send_from_directory(base_path, path)
39
40 rule = "/" + base_path + "/<path:path>"
41 app.add_url_rule(rule=rule, endpoint=endpoint, view_func=assets_handler)
42
43
44 def register_plugin_asset(app, asset_path, admins_only=False, endpoint=None):
45 """
46 Registers an file path to be served by CTFd
47
48 :param app: A CTFd application
49 :param string asset_path: The path to the asset file
50 :param boolean admins_only: Whether or not this file should be accessible to the public
51 :return:
52 """
53 asset_path = asset_path.strip("/")
54 if endpoint is None:
55 endpoint = asset_path.replace("/", ".")
56
57 def asset_handler():
58 return send_file(asset_path)
59
60 if admins_only:
61 asset_handler = admins_only_wrapper(asset_handler)
62 rule = "/" + asset_path
63 app.add_url_rule(rule=rule, endpoint=endpoint, view_func=asset_handler)
64
65
66 def override_template(*args, **kwargs):
67 """
68 Overrides a template with the provided html content.
69
70 e.g. override_template('scoreboard.html', '<h1>scores</h1>')
71 """
72 utils_override_template(*args, **kwargs)
73
74
75 def register_plugin_script(*args, **kwargs):
76 """
77 Adds a given script to the base.html template which all pages inherit from
78 """
79 utils_register_plugin_script(*args, **kwargs)
80
81
82 def register_plugin_stylesheet(*args, **kwargs):
83 """
84 Adds a given stylesheet to the base.html template which all pages inherit from.
85 """
86 utils_register_plugin_stylesheet(*args, **kwargs)
87
88
89 def register_admin_plugin_script(*args, **kwargs):
90 """
91 Adds a given script to the base.html of the admin theme which all admin pages inherit from
92 :param args:
93 :param kwargs:
94 :return:
95 """
96 utils_register_admin_plugin_script(*args, **kwargs)
97
98
99 def register_admin_plugin_stylesheet(*args, **kwargs):
100 """
101 Adds a given stylesheet to the base.html of the admin theme which all admin pages inherit from
102 :param args:
103 :param kwargs:
104 :return:
105 """
106 utils_register_admin_plugin_stylesheet(*args, **kwargs)
107
108
109 def register_admin_plugin_menu_bar(title, route):
110 """
111 Registers links on the Admin Panel menubar/navbar
112
113 :param name: A string that is shown on the navbar HTML
114 :param route: A string that is the href used by the link
115 :return:
116 """
117 am = Menu(title=title, route=route)
118 app.admin_plugin_menu_bar.append(am)
119
120
121 def get_admin_plugin_menu_bar():
122 """
123 Access the list used to store the plugin menu bar
124
125 :return: Returns a list of Menu namedtuples. They have name, and route attributes.
126 """
127 return app.admin_plugin_menu_bar
128
129
130 def register_user_page_menu_bar(title, route):
131 """
132 Registers links on the User side menubar/navbar
133
134 :param name: A string that is shown on the navbar HTML
135 :param route: A string that is the href used by the link
136 :return:
137 """
138 p = Menu(title=title, route=route)
139 app.plugin_menu_bar.append(p)
140
141
142 def get_user_page_menu_bar():
143 """
144 Access the list used to store the user page menu bar
145
146 :return: Returns a list of Menu namedtuples. They have name, and route attributes.
147 """
148 pages = []
149 for p in get_pages() + app.plugin_menu_bar:
150 if p.route.startswith("http"):
151 route = p.route
152 else:
153 route = url_for("views.static_html", route=p.route)
154 print(route)
155 pages.append(Menu(title=p.title, route=route))
156 return pages
157
158
159 def bypass_csrf_protection(f):
160 """
161 Decorator that allows a route to bypass the need for a CSRF nonce on POST requests.
162
163 This should be considered beta and may change in future versions.
164
165 :param f: A function that needs to bypass CSRF protection
166 :return: Returns a function with the _bypass_csrf attribute set which tells CTFd to not require CSRF protection.
167 """
168 f._bypass_csrf = True
169 return f
170
171
172 def get_plugin_names():
173 modules = sorted(glob.glob(app.plugins_dir + "/*"))
174 blacklist = {"__pycache__"}
175 plugins = []
176 for module in modules:
177 module_name = os.path.basename(module)
178 if os.path.isdir(module) and module_name not in blacklist:
179 plugins.append(module_name)
180 return plugins
181
182
183 def init_plugins(app):
184 """
185 Searches for the load function in modules in the CTFd/plugins folder. This function is called with the current CTFd
186 app as a parameter. This allows CTFd plugins to modify CTFd's behavior.
187
188 :param app: A CTFd application
189 :return:
190 """
191 app.admin_plugin_scripts = []
192 app.admin_plugin_stylesheets = []
193 app.plugin_scripts = []
194 app.plugin_stylesheets = []
195
196 app.admin_plugin_menu_bar = []
197 app.plugin_menu_bar = []
198 app.plugins_dir = os.path.dirname(__file__)
199
200 if app.config.get("SAFE_MODE", False) is False:
201 for plugin in get_plugin_names():
202 module = "." + plugin
203 module = importlib.import_module(module, package="CTFd.plugins")
204 module.load(app)
205 print(" * Loaded module, %s" % module)
206
207 app.jinja_env.globals.update(get_admin_plugin_menu_bar=get_admin_plugin_menu_bar)
208 app.jinja_env.globals.update(get_user_page_menu_bar=get_user_page_menu_bar)
209
[end of CTFd/plugins/__init__.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/CTFd/plugins/__init__.py b/CTFd/plugins/__init__.py
--- a/CTFd/plugins/__init__.py
+++ b/CTFd/plugins/__init__.py
@@ -151,7 +151,6 @@
route = p.route
else:
route = url_for("views.static_html", route=p.route)
- print(route)
pages.append(Menu(title=p.title, route=route))
return pages
|
{"golden_diff": "diff --git a/CTFd/plugins/__init__.py b/CTFd/plugins/__init__.py\n--- a/CTFd/plugins/__init__.py\n+++ b/CTFd/plugins/__init__.py\n@@ -151,7 +151,6 @@\n route = p.route\n else:\n route = url_for(\"views.static_html\", route=p.route)\n- print(route)\n pages.append(Menu(title=p.title, route=route))\n return pages\n", "issue": "Invalid model identifier\nhttps://github.com/CTFd/CTFd/blob/master/CTFd/themes/core/templates/scoreboard.html#L26\r\n\r\nThis should change depending on the mode of the CTF\n", "before_files": [{"content": "import glob\nimport importlib\nimport os\nfrom collections import namedtuple\n\nfrom flask import current_app as app\nfrom flask import send_file, send_from_directory, url_for\n\nfrom CTFd.utils.config.pages import get_pages\nfrom CTFd.utils.decorators import admins_only as admins_only_wrapper\nfrom CTFd.utils.plugins import override_template as utils_override_template\nfrom CTFd.utils.plugins import (\n register_admin_script as utils_register_admin_plugin_script,\n)\nfrom CTFd.utils.plugins import (\n register_admin_stylesheet as utils_register_admin_plugin_stylesheet,\n)\nfrom CTFd.utils.plugins import register_script as utils_register_plugin_script\nfrom CTFd.utils.plugins import register_stylesheet as utils_register_plugin_stylesheet\n\nMenu = namedtuple(\"Menu\", [\"title\", \"route\"])\n\n\ndef register_plugin_assets_directory(app, base_path, admins_only=False, endpoint=None):\n \"\"\"\n Registers a directory to serve assets\n\n :param app: A CTFd application\n :param string base_path: The path to the directory\n :param boolean admins_only: Whether or not the assets served out of the directory should be accessible to the public\n :return:\n \"\"\"\n base_path = base_path.strip(\"/\")\n if endpoint is None:\n endpoint = base_path.replace(\"/\", \".\")\n\n def assets_handler(path):\n return send_from_directory(base_path, path)\n\n rule = \"/\" + base_path + \"/<path:path>\"\n app.add_url_rule(rule=rule, endpoint=endpoint, view_func=assets_handler)\n\n\ndef register_plugin_asset(app, asset_path, admins_only=False, endpoint=None):\n \"\"\"\n Registers an file path to be served by CTFd\n\n :param app: A CTFd application\n :param string asset_path: The path to the asset file\n :param boolean admins_only: Whether or not this file should be accessible to the public\n :return:\n \"\"\"\n asset_path = asset_path.strip(\"/\")\n if endpoint is None:\n endpoint = asset_path.replace(\"/\", \".\")\n\n def asset_handler():\n return send_file(asset_path)\n\n if admins_only:\n asset_handler = admins_only_wrapper(asset_handler)\n rule = \"/\" + asset_path\n app.add_url_rule(rule=rule, endpoint=endpoint, view_func=asset_handler)\n\n\ndef override_template(*args, **kwargs):\n \"\"\"\n Overrides a template with the provided html content.\n\n e.g. override_template('scoreboard.html', '<h1>scores</h1>')\n \"\"\"\n utils_override_template(*args, **kwargs)\n\n\ndef register_plugin_script(*args, **kwargs):\n \"\"\"\n Adds a given script to the base.html template which all pages inherit from\n \"\"\"\n utils_register_plugin_script(*args, **kwargs)\n\n\ndef register_plugin_stylesheet(*args, **kwargs):\n \"\"\"\n Adds a given stylesheet to the base.html template which all pages inherit from.\n \"\"\"\n utils_register_plugin_stylesheet(*args, **kwargs)\n\n\ndef register_admin_plugin_script(*args, **kwargs):\n \"\"\"\n Adds a given script to the base.html of the admin theme which all admin pages inherit from\n :param args:\n :param kwargs:\n :return:\n \"\"\"\n utils_register_admin_plugin_script(*args, **kwargs)\n\n\ndef register_admin_plugin_stylesheet(*args, **kwargs):\n \"\"\"\n Adds a given stylesheet to the base.html of the admin theme which all admin pages inherit from\n :param args:\n :param kwargs:\n :return:\n \"\"\"\n utils_register_admin_plugin_stylesheet(*args, **kwargs)\n\n\ndef register_admin_plugin_menu_bar(title, route):\n \"\"\"\n Registers links on the Admin Panel menubar/navbar\n\n :param name: A string that is shown on the navbar HTML\n :param route: A string that is the href used by the link\n :return:\n \"\"\"\n am = Menu(title=title, route=route)\n app.admin_plugin_menu_bar.append(am)\n\n\ndef get_admin_plugin_menu_bar():\n \"\"\"\n Access the list used to store the plugin menu bar\n\n :return: Returns a list of Menu namedtuples. They have name, and route attributes.\n \"\"\"\n return app.admin_plugin_menu_bar\n\n\ndef register_user_page_menu_bar(title, route):\n \"\"\"\n Registers links on the User side menubar/navbar\n\n :param name: A string that is shown on the navbar HTML\n :param route: A string that is the href used by the link\n :return:\n \"\"\"\n p = Menu(title=title, route=route)\n app.plugin_menu_bar.append(p)\n\n\ndef get_user_page_menu_bar():\n \"\"\"\n Access the list used to store the user page menu bar\n\n :return: Returns a list of Menu namedtuples. They have name, and route attributes.\n \"\"\"\n pages = []\n for p in get_pages() + app.plugin_menu_bar:\n if p.route.startswith(\"http\"):\n route = p.route\n else:\n route = url_for(\"views.static_html\", route=p.route)\n print(route)\n pages.append(Menu(title=p.title, route=route))\n return pages\n\n\ndef bypass_csrf_protection(f):\n \"\"\"\n Decorator that allows a route to bypass the need for a CSRF nonce on POST requests.\n\n This should be considered beta and may change in future versions.\n\n :param f: A function that needs to bypass CSRF protection\n :return: Returns a function with the _bypass_csrf attribute set which tells CTFd to not require CSRF protection.\n \"\"\"\n f._bypass_csrf = True\n return f\n\n\ndef get_plugin_names():\n modules = sorted(glob.glob(app.plugins_dir + \"/*\"))\n blacklist = {\"__pycache__\"}\n plugins = []\n for module in modules:\n module_name = os.path.basename(module)\n if os.path.isdir(module) and module_name not in blacklist:\n plugins.append(module_name)\n return plugins\n\n\ndef init_plugins(app):\n \"\"\"\n Searches for the load function in modules in the CTFd/plugins folder. This function is called with the current CTFd\n app as a parameter. This allows CTFd plugins to modify CTFd's behavior.\n\n :param app: A CTFd application\n :return:\n \"\"\"\n app.admin_plugin_scripts = []\n app.admin_plugin_stylesheets = []\n app.plugin_scripts = []\n app.plugin_stylesheets = []\n\n app.admin_plugin_menu_bar = []\n app.plugin_menu_bar = []\n app.plugins_dir = os.path.dirname(__file__)\n\n if app.config.get(\"SAFE_MODE\", False) is False:\n for plugin in get_plugin_names():\n module = \".\" + plugin\n module = importlib.import_module(module, package=\"CTFd.plugins\")\n module.load(app)\n print(\" * Loaded module, %s\" % module)\n\n app.jinja_env.globals.update(get_admin_plugin_menu_bar=get_admin_plugin_menu_bar)\n app.jinja_env.globals.update(get_user_page_menu_bar=get_user_page_menu_bar)\n", "path": "CTFd/plugins/__init__.py"}]}
| 2,594 | 102 |
gh_patches_debug_20362
|
rasdani/github-patches
|
git_diff
|
elastic__apm-agent-python-1593
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Performance Bug in the varmap function
**Describe the bug**: ...
When python code throws an error, and one of the local variables contains a very large list of lists of lists, it takes the elastic python apm client a very long time to process through the list. In our case it was taking over 60 seconds for the code to complete.
**To Reproduce**
To simply replicate the issue, I: (see code snipit in the elastic forum, link below)
- created a simple script to create a list of list of lists of integers.
- in one method I had my "with elasticAPMTransaction:" statement.
- under this I created the list (200 integers) of list (300 integers) of list (40 integers) aka 24M items total.
- started to iterate through the outer list and picked an arbitrary spot to through an exception. This essentially mimicked the method where the real code threw an exception.
- wrapped this whole thing in a profiler and evaluated the output.
This would take 60 seconds +/- to complete.
Seems like the issue is in the varmap function in elasticapm\utils_init_.py but there maybe lots of ways to solve this issue.
**Environment (please complete the following information)**
- OS: [e.g. Linux]
- Python version: 3.x
- Framework and version [e.g. Django 2.1]:
- APM Server version: N/A
- Agent version: 6.7.2
**Additional context**
Created a case here https://discuss.elastic.co/t/python-client-apm-very-slow-when-local-list-variable-is-very-large/309935. See this for additional details.
</issue>
<code>
[start of elasticapm/utils/__init__.py]
1 # BSD 3-Clause License
2 #
3 # Copyright (c) 2012, the Sentry Team, see AUTHORS for more details
4 # Copyright (c) 2019, Elasticsearch BV
5 # All rights reserved.
6 #
7 # Redistribution and use in source and binary forms, with or without
8 # modification, are permitted provided that the following conditions are met:
9 #
10 # * Redistributions of source code must retain the above copyright notice, this
11 # list of conditions and the following disclaimer.
12 #
13 # * Redistributions in binary form must reproduce the above copyright notice,
14 # this list of conditions and the following disclaimer in the documentation
15 # and/or other materials provided with the distribution.
16 #
17 # * Neither the name of the copyright holder nor the names of its
18 # contributors may be used to endorse or promote products derived from
19 # this software without specific prior written permission.
20 #
21 # THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
22 # AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
23 # IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
24 # DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE
25 # FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
26 # DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
27 # SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER
28 # CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
29 # OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
30
31 import base64
32 import os
33 import re
34 import urllib.parse
35 from functools import partial
36 from types import FunctionType
37 from typing import Pattern
38
39 from elasticapm.conf import constants
40 from elasticapm.utils import encoding
41
42 try:
43 from functools import partialmethod
44
45 partial_types = (partial, partialmethod)
46 except ImportError:
47 # Python 2
48 partial_types = (partial,)
49
50
51 default_ports = {"https": 443, "http": 80, "postgresql": 5432, "mysql": 3306, "mssql": 1433}
52
53
54 def varmap(func, var, context=None, name=None, **kwargs):
55 """
56 Executes ``func(key_name, value)`` on all values,
57 recursively discovering dict and list scoped
58 values.
59 """
60 if context is None:
61 context = set()
62 objid = id(var)
63 if objid in context:
64 return func(name, "<...>", **kwargs)
65 context.add(objid)
66 if isinstance(var, dict):
67 # iterate over a copy of the dictionary to avoid "dictionary changed size during iteration" issues
68 ret = func(name, dict((k, varmap(func, v, context, k, **kwargs)) for k, v in var.copy().items()), **kwargs)
69 elif isinstance(var, (list, tuple)):
70 ret = func(name, [varmap(func, f, context, name, **kwargs) for f in var], **kwargs)
71 else:
72 ret = func(name, var, **kwargs)
73 context.remove(objid)
74 return ret
75
76
77 def get_name_from_func(func: FunctionType) -> str:
78 # partials don't have `__module__` or `__name__`, so we use the values from the "inner" function
79 if isinstance(func, partial_types):
80 return "partial({})".format(get_name_from_func(func.func))
81 elif hasattr(func, "_partialmethod") and hasattr(func._partialmethod, "func"):
82 return "partial({})".format(get_name_from_func(func._partialmethod.func))
83
84 module = func.__module__
85
86 if hasattr(func, "view_class"):
87 view_name = func.view_class.__name__
88 elif hasattr(func, "__name__"):
89 view_name = func.__name__
90 else: # Fall back if there's no __name__
91 view_name = func.__class__.__name__
92
93 return "{0}.{1}".format(module, view_name)
94
95
96 def build_name_with_http_method_prefix(name, request):
97 return " ".join((request.method, name)) if name else name
98
99
100 def is_master_process() -> bool:
101 # currently only recognizes uwsgi master process
102 try:
103 import uwsgi
104
105 return os.getpid() == uwsgi.masterpid()
106 except ImportError:
107 return False
108
109
110 def get_url_dict(url: str) -> dict:
111 parse_result = urllib.parse.urlparse(url)
112
113 url_dict = {
114 "full": encoding.keyword_field(url),
115 "protocol": parse_result.scheme + ":",
116 "hostname": encoding.keyword_field(parse_result.hostname),
117 "pathname": encoding.keyword_field(parse_result.path),
118 }
119
120 port = None if parse_result.port is None else str(parse_result.port)
121
122 if port:
123 url_dict["port"] = port
124 if parse_result.query:
125 url_dict["search"] = encoding.keyword_field("?" + parse_result.query)
126 return url_dict
127
128
129 def sanitize_url(url: str) -> str:
130 if "@" not in url:
131 return url
132 parts = urllib.parse.urlparse(url)
133 return url.replace("%s:%s" % (parts.username, parts.password), "%s:%s" % (parts.username, constants.MASK))
134
135
136 def get_host_from_url(url: str) -> str:
137 parsed_url = urllib.parse.urlparse(url)
138 host = parsed_url.hostname or " "
139
140 if parsed_url.port and default_ports.get(parsed_url.scheme) != parsed_url.port:
141 host += ":" + str(parsed_url.port)
142
143 return host
144
145
146 def url_to_destination_resource(url: str) -> str:
147 parts = urllib.parse.urlsplit(url)
148 hostname = parts.hostname if parts.hostname else ""
149 # preserve brackets for IPv6 URLs
150 if "://[" in url:
151 hostname = "[%s]" % hostname
152 try:
153 port = parts.port
154 except ValueError:
155 # Malformed port, just use None rather than raising an exception
156 port = None
157 default_port = default_ports.get(parts.scheme, None)
158 name = "%s://%s" % (parts.scheme, hostname)
159 resource = hostname
160 if not port and parts.scheme in default_ports:
161 port = default_ports[parts.scheme]
162 if port:
163 if port != default_port:
164 name += ":%d" % port
165 resource += ":%d" % port
166 return resource
167
168
169 def read_pem_file(file_obj) -> bytes:
170 cert = b""
171 for line in file_obj:
172 if line.startswith(b"-----BEGIN CERTIFICATE-----"):
173 break
174 # scan until we find the first END CERTIFICATE marker
175 for line in file_obj:
176 if line.startswith(b"-----END CERTIFICATE-----"):
177 break
178 cert += line.strip()
179 return base64.b64decode(cert)
180
181
182 def starmatch_to_regex(pattern: str) -> Pattern:
183 options = re.DOTALL
184 # check if we are case sensitive
185 if pattern.startswith("(?-i)"):
186 pattern = pattern[5:]
187 else:
188 options |= re.IGNORECASE
189 i, n = 0, len(pattern)
190 res = []
191 while i < n:
192 c = pattern[i]
193 i = i + 1
194 if c == "*":
195 res.append(".*")
196 else:
197 res.append(re.escape(c))
198 return re.compile(r"(?:%s)\Z" % "".join(res), options)
199
200
201 def nested_key(d: dict, *args):
202 """
203 Traverses a dictionary for nested keys. Returns `None` if the at any point
204 in the traversal a key cannot be found.
205
206 Example:
207
208 >>> from elasticapm.utils import nested_key
209 >>> d = {"a": {"b": {"c": 0}}}
210 >>> nested_key(d, "a", "b", "c")
211 0
212 >>> nested_key(d, "a", "b", "d")
213 None
214 """
215 for arg in args:
216 try:
217 d = d[arg]
218 except (TypeError, KeyError):
219 d = None
220 break
221 return d
222
[end of elasticapm/utils/__init__.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/elasticapm/utils/__init__.py b/elasticapm/utils/__init__.py
--- a/elasticapm/utils/__init__.py
+++ b/elasticapm/utils/__init__.py
@@ -64,10 +64,14 @@
return func(name, "<...>", **kwargs)
context.add(objid)
if isinstance(var, dict):
+ # Apply func() before recursion, so that `shorten()` doesn't have to iterate over all the trimmed values
+ ret = func(name, var, **kwargs)
# iterate over a copy of the dictionary to avoid "dictionary changed size during iteration" issues
- ret = func(name, dict((k, varmap(func, v, context, k, **kwargs)) for k, v in var.copy().items()), **kwargs)
+ ret = dict((k, varmap(func, v, context, k, **kwargs)) for k, v in ret.copy().items())
elif isinstance(var, (list, tuple)):
- ret = func(name, [varmap(func, f, context, name, **kwargs) for f in var], **kwargs)
+ # Apply func() before recursion, so that `shorten()` doesn't have to iterate over all the trimmed values
+ ret = func(name, var, **kwargs)
+ ret = [varmap(func, f, context, name, **kwargs) for f in ret]
else:
ret = func(name, var, **kwargs)
context.remove(objid)
|
{"golden_diff": "diff --git a/elasticapm/utils/__init__.py b/elasticapm/utils/__init__.py\n--- a/elasticapm/utils/__init__.py\n+++ b/elasticapm/utils/__init__.py\n@@ -64,10 +64,14 @@\n return func(name, \"<...>\", **kwargs)\n context.add(objid)\n if isinstance(var, dict):\n+ # Apply func() before recursion, so that `shorten()` doesn't have to iterate over all the trimmed values\n+ ret = func(name, var, **kwargs)\n # iterate over a copy of the dictionary to avoid \"dictionary changed size during iteration\" issues\n- ret = func(name, dict((k, varmap(func, v, context, k, **kwargs)) for k, v in var.copy().items()), **kwargs)\n+ ret = dict((k, varmap(func, v, context, k, **kwargs)) for k, v in ret.copy().items())\n elif isinstance(var, (list, tuple)):\n- ret = func(name, [varmap(func, f, context, name, **kwargs) for f in var], **kwargs)\n+ # Apply func() before recursion, so that `shorten()` doesn't have to iterate over all the trimmed values\n+ ret = func(name, var, **kwargs)\n+ ret = [varmap(func, f, context, name, **kwargs) for f in ret]\n else:\n ret = func(name, var, **kwargs)\n context.remove(objid)\n", "issue": "Performance Bug in the varmap function\n**Describe the bug**: ...\r\nWhen python code throws an error, and one of the local variables contains a very large list of lists of lists, it takes the elastic python apm client a very long time to process through the list. In our case it was taking over 60 seconds for the code to complete.\r\n\r\n\r\n**To Reproduce**\r\nTo simply replicate the issue, I: (see code snipit in the elastic forum, link below)\r\n\r\n- created a simple script to create a list of list of lists of integers.\r\n- in one method I had my \"with elasticAPMTransaction:\" statement.\r\n- under this I created the list (200 integers) of list (300 integers) of list (40 integers) aka 24M items total.\r\n- started to iterate through the outer list and picked an arbitrary spot to through an exception. This essentially mimicked the method where the real code threw an exception.\r\n- wrapped this whole thing in a profiler and evaluated the output.\r\nThis would take 60 seconds +/- to complete.\r\n\r\nSeems like the issue is in the varmap function in elasticapm\\utils_init_.py but there maybe lots of ways to solve this issue.\r\n\r\n**Environment (please complete the following information)**\r\n- OS: [e.g. Linux]\r\n- Python version: 3.x\r\n- Framework and version [e.g. Django 2.1]:\r\n- APM Server version: N/A\r\n- Agent version: 6.7.2\r\n\r\n\r\n**Additional context**\r\nCreated a case here https://discuss.elastic.co/t/python-client-apm-very-slow-when-local-list-variable-is-very-large/309935. See this for additional details.\n", "before_files": [{"content": "# BSD 3-Clause License\n#\n# Copyright (c) 2012, the Sentry Team, see AUTHORS for more details\n# Copyright (c) 2019, Elasticsearch BV\n# All rights reserved.\n#\n# Redistribution and use in source and binary forms, with or without\n# modification, are permitted provided that the following conditions are met:\n#\n# * Redistributions of source code must retain the above copyright notice, this\n# list of conditions and the following disclaimer.\n#\n# * Redistributions in binary form must reproduce the above copyright notice,\n# this list of conditions and the following disclaimer in the documentation\n# and/or other materials provided with the distribution.\n#\n# * Neither the name of the copyright holder nor the names of its\n# contributors may be used to endorse or promote products derived from\n# this software without specific prior written permission.\n#\n# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS \"AS IS\"\n# AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE\n# IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE\n# DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE\n# FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL\n# DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR\n# SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER\n# CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,\n# OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE\n\nimport base64\nimport os\nimport re\nimport urllib.parse\nfrom functools import partial\nfrom types import FunctionType\nfrom typing import Pattern\n\nfrom elasticapm.conf import constants\nfrom elasticapm.utils import encoding\n\ntry:\n from functools import partialmethod\n\n partial_types = (partial, partialmethod)\nexcept ImportError:\n # Python 2\n partial_types = (partial,)\n\n\ndefault_ports = {\"https\": 443, \"http\": 80, \"postgresql\": 5432, \"mysql\": 3306, \"mssql\": 1433}\n\n\ndef varmap(func, var, context=None, name=None, **kwargs):\n \"\"\"\n Executes ``func(key_name, value)`` on all values,\n recursively discovering dict and list scoped\n values.\n \"\"\"\n if context is None:\n context = set()\n objid = id(var)\n if objid in context:\n return func(name, \"<...>\", **kwargs)\n context.add(objid)\n if isinstance(var, dict):\n # iterate over a copy of the dictionary to avoid \"dictionary changed size during iteration\" issues\n ret = func(name, dict((k, varmap(func, v, context, k, **kwargs)) for k, v in var.copy().items()), **kwargs)\n elif isinstance(var, (list, tuple)):\n ret = func(name, [varmap(func, f, context, name, **kwargs) for f in var], **kwargs)\n else:\n ret = func(name, var, **kwargs)\n context.remove(objid)\n return ret\n\n\ndef get_name_from_func(func: FunctionType) -> str:\n # partials don't have `__module__` or `__name__`, so we use the values from the \"inner\" function\n if isinstance(func, partial_types):\n return \"partial({})\".format(get_name_from_func(func.func))\n elif hasattr(func, \"_partialmethod\") and hasattr(func._partialmethod, \"func\"):\n return \"partial({})\".format(get_name_from_func(func._partialmethod.func))\n\n module = func.__module__\n\n if hasattr(func, \"view_class\"):\n view_name = func.view_class.__name__\n elif hasattr(func, \"__name__\"):\n view_name = func.__name__\n else: # Fall back if there's no __name__\n view_name = func.__class__.__name__\n\n return \"{0}.{1}\".format(module, view_name)\n\n\ndef build_name_with_http_method_prefix(name, request):\n return \" \".join((request.method, name)) if name else name\n\n\ndef is_master_process() -> bool:\n # currently only recognizes uwsgi master process\n try:\n import uwsgi\n\n return os.getpid() == uwsgi.masterpid()\n except ImportError:\n return False\n\n\ndef get_url_dict(url: str) -> dict:\n parse_result = urllib.parse.urlparse(url)\n\n url_dict = {\n \"full\": encoding.keyword_field(url),\n \"protocol\": parse_result.scheme + \":\",\n \"hostname\": encoding.keyword_field(parse_result.hostname),\n \"pathname\": encoding.keyword_field(parse_result.path),\n }\n\n port = None if parse_result.port is None else str(parse_result.port)\n\n if port:\n url_dict[\"port\"] = port\n if parse_result.query:\n url_dict[\"search\"] = encoding.keyword_field(\"?\" + parse_result.query)\n return url_dict\n\n\ndef sanitize_url(url: str) -> str:\n if \"@\" not in url:\n return url\n parts = urllib.parse.urlparse(url)\n return url.replace(\"%s:%s\" % (parts.username, parts.password), \"%s:%s\" % (parts.username, constants.MASK))\n\n\ndef get_host_from_url(url: str) -> str:\n parsed_url = urllib.parse.urlparse(url)\n host = parsed_url.hostname or \" \"\n\n if parsed_url.port and default_ports.get(parsed_url.scheme) != parsed_url.port:\n host += \":\" + str(parsed_url.port)\n\n return host\n\n\ndef url_to_destination_resource(url: str) -> str:\n parts = urllib.parse.urlsplit(url)\n hostname = parts.hostname if parts.hostname else \"\"\n # preserve brackets for IPv6 URLs\n if \"://[\" in url:\n hostname = \"[%s]\" % hostname\n try:\n port = parts.port\n except ValueError:\n # Malformed port, just use None rather than raising an exception\n port = None\n default_port = default_ports.get(parts.scheme, None)\n name = \"%s://%s\" % (parts.scheme, hostname)\n resource = hostname\n if not port and parts.scheme in default_ports:\n port = default_ports[parts.scheme]\n if port:\n if port != default_port:\n name += \":%d\" % port\n resource += \":%d\" % port\n return resource\n\n\ndef read_pem_file(file_obj) -> bytes:\n cert = b\"\"\n for line in file_obj:\n if line.startswith(b\"-----BEGIN CERTIFICATE-----\"):\n break\n # scan until we find the first END CERTIFICATE marker\n for line in file_obj:\n if line.startswith(b\"-----END CERTIFICATE-----\"):\n break\n cert += line.strip()\n return base64.b64decode(cert)\n\n\ndef starmatch_to_regex(pattern: str) -> Pattern:\n options = re.DOTALL\n # check if we are case sensitive\n if pattern.startswith(\"(?-i)\"):\n pattern = pattern[5:]\n else:\n options |= re.IGNORECASE\n i, n = 0, len(pattern)\n res = []\n while i < n:\n c = pattern[i]\n i = i + 1\n if c == \"*\":\n res.append(\".*\")\n else:\n res.append(re.escape(c))\n return re.compile(r\"(?:%s)\\Z\" % \"\".join(res), options)\n\n\ndef nested_key(d: dict, *args):\n \"\"\"\n Traverses a dictionary for nested keys. Returns `None` if the at any point\n in the traversal a key cannot be found.\n\n Example:\n\n >>> from elasticapm.utils import nested_key\n >>> d = {\"a\": {\"b\": {\"c\": 0}}}\n >>> nested_key(d, \"a\", \"b\", \"c\")\n 0\n >>> nested_key(d, \"a\", \"b\", \"d\")\n None\n \"\"\"\n for arg in args:\n try:\n d = d[arg]\n except (TypeError, KeyError):\n d = None\n break\n return d\n", "path": "elasticapm/utils/__init__.py"}]}
| 3,237 | 332 |
gh_patches_debug_21235
|
rasdani/github-patches
|
git_diff
|
facebookresearch__hydra-1630
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Document hydra.callbacks api.
- [ ] Document API
- [ ] Add news fragment if missing
</issue>
<code>
[start of hydra/experimental/callback.py]
1 # Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved
2 import logging
3 from typing import Any
4
5 from omegaconf import DictConfig
6
7 from hydra.core.utils import JobReturn
8
9 logger = logging.getLogger(__name__)
10
11
12 class Callback:
13 def on_run_start(self, config: DictConfig, **kwargs: Any) -> None:
14 """
15 Called in RUN mode before job starts.
16 """
17 ...
18
19 def on_run_end(self, config: DictConfig, **kwargs: Any) -> None:
20 """
21 Called in RUN mode after job ends.
22 """
23 ...
24
25 def on_multirun_start(self, config: DictConfig, **kwargs: Any) -> None:
26 """
27 Called in MULTIRUN mode before any job starts.
28 """
29 ...
30
31 def on_multirun_end(self, config: DictConfig, **kwargs: Any) -> None:
32 """
33 Called in MULTIRUN mode after all job end.
34 """
35 ...
36
37 def on_job_start(self, config: DictConfig, **kwargs: Any) -> None:
38 """
39 Called in both RUN and MULTIRUN modes inside a Hydra job; before running
40 application code.
41 """
42 ...
43
44 def on_job_end(
45 self, config: DictConfig, job_return: JobReturn, **kwargs: Any
46 ) -> None:
47 """
48 Called in both RUN and MULTIRUN modes inside a Hydra job; after running
49 application code.
50 """
51 ...
52
[end of hydra/experimental/callback.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/hydra/experimental/callback.py b/hydra/experimental/callback.py
--- a/hydra/experimental/callback.py
+++ b/hydra/experimental/callback.py
@@ -30,14 +30,14 @@
def on_multirun_end(self, config: DictConfig, **kwargs: Any) -> None:
"""
- Called in MULTIRUN mode after all job end.
+ Called in MULTIRUN mode after all jobs end.
"""
...
def on_job_start(self, config: DictConfig, **kwargs: Any) -> None:
"""
- Called in both RUN and MULTIRUN modes inside a Hydra job; before running
- application code.
+ Called in both RUN and MULTIRUN modes, once for each Hydra job (before running
+ application code).
"""
...
@@ -45,7 +45,7 @@
self, config: DictConfig, job_return: JobReturn, **kwargs: Any
) -> None:
"""
- Called in both RUN and MULTIRUN modes inside a Hydra job; after running
- application code.
+ Called in both RUN and MULTIRUN modes, once for each Hydra job (after running
+ application code).
"""
...
|
{"golden_diff": "diff --git a/hydra/experimental/callback.py b/hydra/experimental/callback.py\n--- a/hydra/experimental/callback.py\n+++ b/hydra/experimental/callback.py\n@@ -30,14 +30,14 @@\n \n def on_multirun_end(self, config: DictConfig, **kwargs: Any) -> None:\n \"\"\"\n- Called in MULTIRUN mode after all job end.\n+ Called in MULTIRUN mode after all jobs end.\n \"\"\"\n ...\n \n def on_job_start(self, config: DictConfig, **kwargs: Any) -> None:\n \"\"\"\n- Called in both RUN and MULTIRUN modes inside a Hydra job; before running\n- application code.\n+ Called in both RUN and MULTIRUN modes, once for each Hydra job (before running\n+ application code).\n \"\"\"\n ...\n \n@@ -45,7 +45,7 @@\n self, config: DictConfig, job_return: JobReturn, **kwargs: Any\n ) -> None:\n \"\"\"\n- Called in both RUN and MULTIRUN modes inside a Hydra job; after running\n- application code.\n+ Called in both RUN and MULTIRUN modes, once for each Hydra job (after running\n+ application code).\n \"\"\"\n ...\n", "issue": "Document hydra.callbacks api.\n- [ ] Document API\r\n- [ ] Add news fragment if missing\n", "before_files": [{"content": "# Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved\nimport logging\nfrom typing import Any\n\nfrom omegaconf import DictConfig\n\nfrom hydra.core.utils import JobReturn\n\nlogger = logging.getLogger(__name__)\n\n\nclass Callback:\n def on_run_start(self, config: DictConfig, **kwargs: Any) -> None:\n \"\"\"\n Called in RUN mode before job starts.\n \"\"\"\n ...\n\n def on_run_end(self, config: DictConfig, **kwargs: Any) -> None:\n \"\"\"\n Called in RUN mode after job ends.\n \"\"\"\n ...\n\n def on_multirun_start(self, config: DictConfig, **kwargs: Any) -> None:\n \"\"\"\n Called in MULTIRUN mode before any job starts.\n \"\"\"\n ...\n\n def on_multirun_end(self, config: DictConfig, **kwargs: Any) -> None:\n \"\"\"\n Called in MULTIRUN mode after all job end.\n \"\"\"\n ...\n\n def on_job_start(self, config: DictConfig, **kwargs: Any) -> None:\n \"\"\"\n Called in both RUN and MULTIRUN modes inside a Hydra job; before running\n application code.\n \"\"\"\n ...\n\n def on_job_end(\n self, config: DictConfig, job_return: JobReturn, **kwargs: Any\n ) -> None:\n \"\"\"\n Called in both RUN and MULTIRUN modes inside a Hydra job; after running\n application code.\n \"\"\"\n ...\n", "path": "hydra/experimental/callback.py"}]}
| 972 | 279 |
gh_patches_debug_37206
|
rasdani/github-patches
|
git_diff
|
PokemonGoF__PokemonGo-Bot-5067
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Assign to Team after reaching lvl 5
<!-- ===============FEATURE REQUEST SECTION===============
Before you create a Feature Request, please check the following:
1. Have you [searched our feature tracker](https://github.com/PokemonGoF/PokemonGo-Bot/labels/Feature%20Request) first to ensure someone else hasn't already come up with the same great idea. If so then be sure to +1 it
2. While you're there, be sure to vote on other feature requests to let the devs know what is important to you.
-->
<!-- Note: Delete this line and everything ABOVE if creating a Feature Request -->
### Short Description
<!-- Tell us a short description of your request -->
Add the ability to assign to a team (blue, red, yellow) after bot is reaching lvl5
<!-- ==========END OF FEATURE REQUEST SECTION========== -->
</issue>
<code>
[start of pokemongo_bot/cell_workers/complete_tutorial.py]
1 import random
2
3 from pokemongo_bot import logger
4 from pokemongo_bot.base_task import BaseTask
5 from pokemongo_bot.worker_result import WorkerResult
6 from pokemongo_bot.human_behaviour import sleep
7
8
9
10 class CompleteTutorial(BaseTask):
11
12 SUPPORTED_TASK_API_VERSION = 1
13
14 def initialize(self):
15 self.api = self.bot.api
16 self.nickname = self.config.get('nickname','')
17 self.team = self.config.get('team',0)
18 self.may_run = True
19
20 def should_run(self):
21 return self.may_run
22
23 def work(self):
24
25 if not self.should_run():
26 return WorkerResult.SUCCESS
27
28 # Only execute the worker once to avoid error loop
29 self.may_run = False
30
31 if self._check_tutorial_state():
32 return WorkerResult.SUCCESS
33 else:
34 return WorkerResult.ERROR
35
36 def _check_tutorial_state(self):
37 self._player=self.bot.player_data
38
39 tutorial_state = self._player.get('tutorial_state', [])
40 # LEGAL_SCREEN = 0
41 if not 0 in tutorial_state:
42 sleep(2)
43 if self._set_tutorial_state(0):
44 self.logger.info('Completed legal screen')
45 tutorial_state = self._player.get('tutorial_state', [])
46 else:
47 return False
48
49 # AVATAR_SELECTION = 1
50 if not 1 in tutorial_state:
51 sleep(7)
52 if self._set_avatar():
53 if self._set_tutorial_state(1):
54 self.logger.info('Completed avatar selection')
55 tutorial_state = self._player.get('tutorial_state', [])
56 else:
57 return False
58 else:
59 self.logger.error('Error during avatar selection')
60 return False
61
62 # POKEMON_CAPTURE = 3
63 if not 3 in tutorial_state:
64 sleep(10)
65 if self._encounter_tutorial():
66 self.logger.info('Completed first capture')
67 else:
68 self.logger.error('Error during first capture')
69 return False
70
71 # NAME_SELECTION = 4
72 if not 4 in tutorial_state:
73 if not self.nickname:
74 self.logger.info("No nickname defined in config")
75 return False
76
77 self.logger.info(u'Trying to set {} as nickname'.format(self.nickname))
78 sleep(5)
79 if self._set_nickname(self.nickname):
80 self._set_tutorial_state(4)
81 tutorial_state = self._player.get('tutorial_state', [])
82 else:
83 self.logger.error('Error trying to set nickname')
84 return False
85
86 # FIRST_TIME_EXPERIENCE_COMPLETE = 7
87 if not 7 in tutorial_state:
88 if self._set_tutorial_state(7):
89 self.logger.info('Completed first time experience')
90 else:
91 return False
92
93 return True
94
95 def _encounter_tutorial(self):
96 # You just need to call the API with the pokemon you choose
97 # Probably can't get MewTwo as first pokemon though
98 first_pokemon_id = random.choice([1, 4, 7])
99 response_dict = self.api.encounter_tutorial_complete(
100 pokemon_id=first_pokemon_id)
101 try:
102 if response_dict['responses']['ENCOUNTER_TUTORIAL_COMPLETE']['result'] == 1:
103 return True
104 else:
105 self.logger.error("Error during encouter tutorial")
106 return False
107 except KeyError:
108 self.logger.error("KeyError during encouter tutorial")
109 return False
110
111 def _random_avatar(self):
112 avatar= {}
113 # 0 = Male, 1 = Female
114 avatar['gender']=random.randint(0,1)
115 # What is the max value of each parameter ?
116 # Default is 0, anyway human player will stop
117 # at the first choices in general, so fully
118 # random on the whole avatar space is not the way to go either
119 avatar['skin']=random.randint(0,3)
120 avatar['hair']=random.randint(0,3)
121 avatar['shirt']=random.randint(0,3)
122 avatar['pants']=random.randint(0,3)
123 avatar['hat']=random.randint(0,3)
124 avatar['shoes']=random.randint(0,3)
125 avatar['eyes']=random.randint(0,3)
126 avatar['backpack']=random.randint(0,3)
127 return avatar
128
129 def _set_avatar(self):
130 avatar = self._random_avatar()
131 response_dict = self.api.set_avatar(player_avatar=avatar)
132 status = response_dict['responses']['SET_AVATAR']['status']
133 try:
134 if status == 1:
135 return True
136 else:
137 error_codes = {
138 0: 'UNSET',
139 1: 'SUCCESS',
140 2: 'AVATAR_ALREADY_SET',
141 3: 'FAILURE',
142 }
143 self.logger.error("Error during avatar selection : {}".format(error_codes[status]))
144 return False
145 except KeyError:
146 self.logger.error("KeyError during avatar selection")
147 return False
148
149 def _set_nickname(self, nickname):
150 response_dict = self.api.claim_codename(codename=nickname)
151 try:
152 result = response_dict['responses']['CLAIM_CODENAME']['status']
153 if result == 1:
154 self.logger.info(u'Name changed to {}'.format(nickname))
155 return True
156 else:
157 # Would be nice to get the text directly from the proto Enum
158 error_codes = {
159 0: 'UNSET',
160 1: 'SUCCESS',
161 2: 'CODENAME_NOT_AVAILABLE',
162 3: 'CODENAME_NOT_VALID',
163 4: 'CURRENT_OWNER',
164 5: 'CODENAME_CHANGE_NOT_ALLOWED'
165 }
166 self.logger.error(
167 u'Error while changing nickname : {}'.format(error_codes[result]))
168 return False
169 except KeyError:
170 return False
171
172 def _set_tutorial_state(self, completed):
173 response_dict = self.api.mark_tutorial_complete(tutorials_completed=[
174 completed], send_marketing_emails=False, send_push_notifications=False)
175 try:
176 self._player = response_dict['responses'][
177 'MARK_TUTORIAL_COMPLETE']['player_data']
178 return response_dict['responses']['MARK_TUTORIAL_COMPLETE']['success']
179 except KeyError:
180 self.logger.error("KeyError while setting tutorial state")
181 return False
182
[end of pokemongo_bot/cell_workers/complete_tutorial.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/pokemongo_bot/cell_workers/complete_tutorial.py b/pokemongo_bot/cell_workers/complete_tutorial.py
--- a/pokemongo_bot/cell_workers/complete_tutorial.py
+++ b/pokemongo_bot/cell_workers/complete_tutorial.py
@@ -1,6 +1,7 @@
import random
from pokemongo_bot import logger
+from pokemongo_bot.inventory import player
from pokemongo_bot.base_task import BaseTask
from pokemongo_bot.worker_result import WorkerResult
from pokemongo_bot.human_behaviour import sleep
@@ -15,23 +16,22 @@
self.api = self.bot.api
self.nickname = self.config.get('nickname','')
self.team = self.config.get('team',0)
- self.may_run = True
-
- def should_run(self):
- return self.may_run
+ self.tutorial_run = True
+ self.team_run = True
def work(self):
- if not self.should_run():
- return WorkerResult.SUCCESS
+ if self.tutorial_run:
+ self.tutorial_run = False
+ if not self._check_tutorial_state():
+ return WorkerResult.ERROR
- # Only execute the worker once to avoid error loop
- self.may_run = False
+ if self.team_run and player()._level >= 5:
+ self.team_run = False
+ if not self._set_team():
+ return WorkerResult.ERROR
- if self._check_tutorial_state():
- return WorkerResult.SUCCESS
- else:
- return WorkerResult.ERROR
+ return WorkerResult.SUCCESS
def _check_tutorial_state(self):
self._player=self.bot.player_data
@@ -179,3 +179,36 @@
except KeyError:
self.logger.error("KeyError while setting tutorial state")
return False
+
+ def _set_team(self):
+ if self.team == 0:
+ return True
+
+ if self.bot.player_data.get('team', 0) != 0:
+ self.logger.info(u'Team already picked')
+ return True
+
+ sleep(10)
+ response_dict = self.api.set_player_team(team=self.team)
+ try:
+ result = response_dict['responses']['SET_PLAYER_TEAM']['status']
+ if result == 1:
+ team_codes = {
+ 1: 'Mystic (BLUE)',
+ 2: 'Valor (RED)',
+ 3: 'Instinct (YELLOW)'
+ }
+ self.logger.info(u'Picked Team {}.'.format(team_codes[self.team]))
+ return True
+ else:
+ error_codes = {
+ 0: 'UNSET',
+ 1: 'SUCCESS',
+ 2: 'TEAM_ALREADY_SET',
+ 3: 'FAILURE'
+ }
+ self.logger.error(u'Error while picking team : {}'.format(error_codes[result]))
+ return False
+ except KeyError:
+ return False
+
|
{"golden_diff": "diff --git a/pokemongo_bot/cell_workers/complete_tutorial.py b/pokemongo_bot/cell_workers/complete_tutorial.py\n--- a/pokemongo_bot/cell_workers/complete_tutorial.py\n+++ b/pokemongo_bot/cell_workers/complete_tutorial.py\n@@ -1,6 +1,7 @@\n import random\n \n from pokemongo_bot import logger\n+from pokemongo_bot.inventory import player\n from pokemongo_bot.base_task import BaseTask\n from pokemongo_bot.worker_result import WorkerResult\n from pokemongo_bot.human_behaviour import sleep\n@@ -15,23 +16,22 @@\n self.api = self.bot.api\n self.nickname = self.config.get('nickname','')\n self.team = self.config.get('team',0)\n- self.may_run = True\n-\n- def should_run(self):\n- return self.may_run\n+ self.tutorial_run = True\n+ self.team_run = True\n \n def work(self):\n \n- if not self.should_run():\n- return WorkerResult.SUCCESS\n+ if self.tutorial_run:\n+ self.tutorial_run = False\n+ if not self._check_tutorial_state():\n+ return WorkerResult.ERROR\n \n- # Only execute the worker once to avoid error loop\n- self.may_run = False\n+ if self.team_run and player()._level >= 5:\n+ self.team_run = False\n+ if not self._set_team():\n+ return WorkerResult.ERROR\n \n- if self._check_tutorial_state():\n- return WorkerResult.SUCCESS\n- else:\n- return WorkerResult.ERROR\n+ return WorkerResult.SUCCESS\n \n def _check_tutorial_state(self):\n self._player=self.bot.player_data\n@@ -179,3 +179,36 @@\n except KeyError:\n self.logger.error(\"KeyError while setting tutorial state\")\n return False\n+\n+ def _set_team(self):\n+ if self.team == 0:\n+ return True\n+\n+ if self.bot.player_data.get('team', 0) != 0:\n+ self.logger.info(u'Team already picked')\n+ return True\n+\n+ sleep(10)\n+ response_dict = self.api.set_player_team(team=self.team)\n+ try:\n+ result = response_dict['responses']['SET_PLAYER_TEAM']['status']\n+ if result == 1:\n+ team_codes = {\n+ 1: 'Mystic (BLUE)',\n+ 2: 'Valor (RED)',\n+ 3: 'Instinct (YELLOW)'\n+ }\n+ self.logger.info(u'Picked Team {}.'.format(team_codes[self.team]))\n+ return True\n+ else:\n+ error_codes = {\n+ 0: 'UNSET',\n+ 1: 'SUCCESS',\n+ 2: 'TEAM_ALREADY_SET',\n+ 3: 'FAILURE'\n+ }\n+ self.logger.error(u'Error while picking team : {}'.format(error_codes[result]))\n+ return False\n+ except KeyError:\n+ return False\n+\n", "issue": "Assign to Team after reaching lvl 5\n<!-- ===============FEATURE REQUEST SECTION===============\n\nBefore you create a Feature Request, please check the following:\n\n1. Have you [searched our feature tracker](https://github.com/PokemonGoF/PokemonGo-Bot/labels/Feature%20Request) first to ensure someone else hasn't already come up with the same great idea. If so then be sure to +1 it \n2. While you're there, be sure to vote on other feature requests to let the devs know what is important to you.\n-->\n\n<!-- Note: Delete this line and everything ABOVE if creating a Feature Request -->\n### Short Description\n\n<!-- Tell us a short description of your request -->\n\nAdd the ability to assign to a team (blue, red, yellow) after bot is reaching lvl5\n\n<!-- ==========END OF FEATURE REQUEST SECTION========== -->\n\n", "before_files": [{"content": "import random\n\nfrom pokemongo_bot import logger\nfrom pokemongo_bot.base_task import BaseTask\nfrom pokemongo_bot.worker_result import WorkerResult\nfrom pokemongo_bot.human_behaviour import sleep\n\n\n\nclass CompleteTutorial(BaseTask):\n\n SUPPORTED_TASK_API_VERSION = 1\n\n def initialize(self):\n self.api = self.bot.api\n self.nickname = self.config.get('nickname','')\n self.team = self.config.get('team',0)\n self.may_run = True\n\n def should_run(self):\n return self.may_run\n\n def work(self):\n\n if not self.should_run():\n return WorkerResult.SUCCESS\n\n # Only execute the worker once to avoid error loop\n self.may_run = False\n\n if self._check_tutorial_state():\n return WorkerResult.SUCCESS\n else:\n return WorkerResult.ERROR\n\n def _check_tutorial_state(self):\n self._player=self.bot.player_data\n\n tutorial_state = self._player.get('tutorial_state', [])\n # LEGAL_SCREEN = 0\n if not 0 in tutorial_state:\n sleep(2)\n if self._set_tutorial_state(0):\n self.logger.info('Completed legal screen')\n tutorial_state = self._player.get('tutorial_state', [])\n else:\n return False\n\n # AVATAR_SELECTION = 1\n if not 1 in tutorial_state:\n sleep(7)\n if self._set_avatar():\n if self._set_tutorial_state(1):\n self.logger.info('Completed avatar selection')\n tutorial_state = self._player.get('tutorial_state', [])\n else:\n return False\n else:\n self.logger.error('Error during avatar selection')\n return False\n\n # POKEMON_CAPTURE = 3\n if not 3 in tutorial_state:\n sleep(10)\n if self._encounter_tutorial():\n self.logger.info('Completed first capture')\n else:\n self.logger.error('Error during first capture')\n return False\n\n # NAME_SELECTION = 4\n if not 4 in tutorial_state:\n if not self.nickname:\n self.logger.info(\"No nickname defined in config\")\n return False\n\n self.logger.info(u'Trying to set {} as nickname'.format(self.nickname))\n sleep(5)\n if self._set_nickname(self.nickname):\n self._set_tutorial_state(4)\n tutorial_state = self._player.get('tutorial_state', [])\n else:\n self.logger.error('Error trying to set nickname')\n return False\n\n # FIRST_TIME_EXPERIENCE_COMPLETE = 7\n if not 7 in tutorial_state:\n if self._set_tutorial_state(7):\n self.logger.info('Completed first time experience')\n else:\n return False\n\n return True\n\n def _encounter_tutorial(self):\n # You just need to call the API with the pokemon you choose\n # Probably can't get MewTwo as first pokemon though\n first_pokemon_id = random.choice([1, 4, 7])\n response_dict = self.api.encounter_tutorial_complete(\n pokemon_id=first_pokemon_id)\n try:\n if response_dict['responses']['ENCOUNTER_TUTORIAL_COMPLETE']['result'] == 1:\n return True\n else:\n self.logger.error(\"Error during encouter tutorial\")\n return False\n except KeyError:\n self.logger.error(\"KeyError during encouter tutorial\")\n return False\n\n def _random_avatar(self):\n avatar= {}\n # 0 = Male, 1 = Female\n avatar['gender']=random.randint(0,1)\n # What is the max value of each parameter ?\n # Default is 0, anyway human player will stop\n # at the first choices in general, so fully\n # random on the whole avatar space is not the way to go either\n avatar['skin']=random.randint(0,3)\n avatar['hair']=random.randint(0,3)\n avatar['shirt']=random.randint(0,3)\n avatar['pants']=random.randint(0,3)\n avatar['hat']=random.randint(0,3)\n avatar['shoes']=random.randint(0,3)\n avatar['eyes']=random.randint(0,3)\n avatar['backpack']=random.randint(0,3)\n return avatar\n\n def _set_avatar(self):\n avatar = self._random_avatar()\n response_dict = self.api.set_avatar(player_avatar=avatar)\n status = response_dict['responses']['SET_AVATAR']['status']\n try:\n if status == 1:\n return True\n else:\n error_codes = {\n 0: 'UNSET',\n 1: 'SUCCESS',\n 2: 'AVATAR_ALREADY_SET',\n 3: 'FAILURE',\n }\n self.logger.error(\"Error during avatar selection : {}\".format(error_codes[status]))\n return False\n except KeyError:\n self.logger.error(\"KeyError during avatar selection\")\n return False\n\n def _set_nickname(self, nickname):\n response_dict = self.api.claim_codename(codename=nickname)\n try:\n result = response_dict['responses']['CLAIM_CODENAME']['status']\n if result == 1:\n self.logger.info(u'Name changed to {}'.format(nickname))\n return True\n else:\n # Would be nice to get the text directly from the proto Enum\n error_codes = {\n 0: 'UNSET',\n 1: 'SUCCESS',\n 2: 'CODENAME_NOT_AVAILABLE',\n 3: 'CODENAME_NOT_VALID',\n 4: 'CURRENT_OWNER',\n 5: 'CODENAME_CHANGE_NOT_ALLOWED'\n }\n self.logger.error(\n u'Error while changing nickname : {}'.format(error_codes[result]))\n return False\n except KeyError:\n return False\n\n def _set_tutorial_state(self, completed):\n response_dict = self.api.mark_tutorial_complete(tutorials_completed=[\n completed], send_marketing_emails=False, send_push_notifications=False)\n try:\n self._player = response_dict['responses'][\n 'MARK_TUTORIAL_COMPLETE']['player_data']\n return response_dict['responses']['MARK_TUTORIAL_COMPLETE']['success']\n except KeyError:\n self.logger.error(\"KeyError while setting tutorial state\")\n return False\n", "path": "pokemongo_bot/cell_workers/complete_tutorial.py"}]}
| 2,517 | 674 |
gh_patches_debug_10147
|
rasdani/github-patches
|
git_diff
|
mesonbuild__meson-9561
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Require parallel tools when depending on HDF5
I have an application that specifically requires the availability of `h5pcc` instead of `h5cc`: https://github.com/mesonbuild/meson/blob/39ceb677744e1f959a0db6406d51a26d78f0ab68/mesonbuild/dependencies/hdf5.py#L97-L99
In my case HDF5 is only build with parallel compatibility and therefore `h5cc` is not available.
My current workaround is to simple edit this line and specify the other tool directly.
It there a more appropriate workaround for this issue?
Alternatively: Is it possible to add an optional parameter to `hdf5_dep = dependency('hdf5')` e.g. `dependency('hdf5', parallel: true)`
</issue>
<code>
[start of mesonbuild/dependencies/hdf5.py]
1 # Copyright 2013-2019 The Meson development team
2
3 # Licensed under the Apache License, Version 2.0 (the "License");
4 # you may not use this file except in compliance with the License.
5 # You may obtain a copy of the License at
6
7 # http://www.apache.org/licenses/LICENSE-2.0
8
9 # Unless required by applicable law or agreed to in writing, software
10 # distributed under the License is distributed on an "AS IS" BASIS,
11 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12 # See the License for the specific language governing permissions and
13 # limitations under the License.
14
15 # This file contains the detection logic for miscellaneous external dependencies.
16
17 import functools
18 import os
19 import re
20 import shutil
21 import subprocess
22 from pathlib import Path
23
24 from ..mesonlib import OrderedSet, join_args
25 from .base import DependencyException, DependencyMethods
26 from .configtool import ConfigToolDependency
27 from .pkgconfig import PkgConfigDependency
28 from .factory import factory_methods
29 import typing as T
30
31 if T.TYPE_CHECKING:
32 from .factory import DependencyGenerator
33 from ..environment import Environment
34 from ..mesonlib import MachineChoice
35
36
37 class HDF5PkgConfigDependency(PkgConfigDependency):
38
39 """Handle brokenness in the HDF5 pkg-config files."""
40
41 def __init__(self, name: str, environment: 'Environment', kwargs: T.Dict[str, T.Any], language: T.Optional[str] = None) -> None:
42 language = language or 'c'
43 if language not in {'c', 'cpp', 'fortran'}:
44 raise DependencyException(f'Language {language} is not supported with HDF5.')
45
46 super().__init__(name, environment, kwargs, language)
47 if not self.is_found:
48 return
49
50 # some broken pkgconfig don't actually list the full path to the needed includes
51 newinc = [] # type: T.List[str]
52 for arg in self.compile_args:
53 if arg.startswith('-I'):
54 stem = 'static' if kwargs.get('static', False) else 'shared'
55 if (Path(arg[2:]) / stem).is_dir():
56 newinc.append('-I' + str(Path(arg[2:]) / stem))
57 self.compile_args += newinc
58
59 link_args = [] # type: T.List[str]
60 for larg in self.get_link_args():
61 lpath = Path(larg)
62 # some pkg-config hdf5.pc (e.g. Ubuntu) don't include the commonly-used HL HDF5 libraries,
63 # so let's add them if they exist
64 # additionally, some pkgconfig HDF5 HL files are malformed so let's be sure to find HL anyway
65 if lpath.is_file():
66 hl = []
67 if language == 'cpp':
68 hl += ['_hl_cpp', '_cpp']
69 elif language == 'fortran':
70 hl += ['_hl_fortran', 'hl_fortran', '_fortran']
71 hl += ['_hl'] # C HL library, always needed
72
73 suffix = '.' + lpath.name.split('.', 1)[1] # in case of .dll.a
74 for h in hl:
75 hlfn = lpath.parent / (lpath.name.split('.', 1)[0] + h + suffix)
76 if hlfn.is_file():
77 link_args.append(str(hlfn))
78 # HDF5 C libs are required by other HDF5 languages
79 link_args.append(larg)
80 else:
81 link_args.append(larg)
82
83 self.link_args = link_args
84
85
86 class HDF5ConfigToolDependency(ConfigToolDependency):
87
88 """Wrapper around hdf5 binary config tools."""
89
90 version_arg = '-showconfig'
91
92 def __init__(self, name: str, environment: 'Environment', kwargs: T.Dict[str, T.Any], language: T.Optional[str] = None) -> None:
93 language = language or 'c'
94 if language not in {'c', 'cpp', 'fortran'}:
95 raise DependencyException(f'Language {language} is not supported with HDF5.')
96
97 if language == 'c':
98 cenv = 'CC'
99 tools = ['h5cc']
100 elif language == 'cpp':
101 cenv = 'CXX'
102 tools = ['h5c++']
103 elif language == 'fortran':
104 cenv = 'FC'
105 tools = ['h5fc']
106 else:
107 raise DependencyException('How did you get here?')
108
109 # We need this before we call super()
110 for_machine = self.get_for_machine_from_kwargs(kwargs)
111
112 nkwargs = kwargs.copy()
113 nkwargs['tools'] = tools
114
115 # Override the compiler that the config tools are going to use by
116 # setting the environment variables that they use for the compiler and
117 # linkers.
118 compiler = environment.coredata.compilers[for_machine][language]
119 try:
120 os.environ[f'HDF5_{cenv}'] = join_args(compiler.get_exelist())
121 os.environ[f'HDF5_{cenv}LINKER'] = join_args(compiler.get_linker_exelist())
122 super().__init__(name, environment, nkwargs, language)
123 finally:
124 del os.environ[f'HDF5_{cenv}']
125 del os.environ[f'HDF5_{cenv}LINKER']
126 if not self.is_found:
127 return
128
129 # We first need to call the tool with -c to get the compile arguments
130 # and then without -c to get the link arguments.
131 args = self.get_config_value(['-show', '-c'], 'args')[1:]
132 args += self.get_config_value(['-show', '-noshlib' if kwargs.get('static', False) else '-shlib'], 'args')[1:]
133 for arg in args:
134 if arg.startswith(('-I', '-f', '-D')) or arg == '-pthread':
135 self.compile_args.append(arg)
136 elif arg.startswith(('-L', '-l', '-Wl')):
137 self.link_args.append(arg)
138 elif Path(arg).is_file():
139 self.link_args.append(arg)
140
141 # If the language is not C we need to add C as a subdependency
142 if language != 'c':
143 nkwargs = kwargs.copy()
144 nkwargs['language'] = 'c'
145 # I'm being too clever for mypy and pylint
146 self.is_found = self._add_sub_dependency(hdf5_factory(environment, for_machine, nkwargs)) # pylint: disable=no-value-for-parameter
147
148 def _sanitize_version(self, ver: str) -> str:
149 v = re.search(r'\s*HDF5 Version: (\d+\.\d+\.\d+)', ver)
150 return v.group(1)
151
152
153 @factory_methods({DependencyMethods.PKGCONFIG, DependencyMethods.CONFIG_TOOL})
154 def hdf5_factory(env: 'Environment', for_machine: 'MachineChoice',
155 kwargs: T.Dict[str, T.Any], methods: T.List[DependencyMethods]) -> T.List['DependencyGenerator']:
156 language = kwargs.get('language')
157 candidates: T.List['DependencyGenerator'] = []
158
159 if DependencyMethods.PKGCONFIG in methods:
160 # Use an ordered set so that these remain the first tried pkg-config files
161 pkgconfig_files = OrderedSet(['hdf5', 'hdf5-serial'])
162 # FIXME: This won't honor pkg-config paths, and cross-native files
163 PCEXE = shutil.which('pkg-config')
164 if PCEXE:
165 # some distros put hdf5-1.2.3.pc with version number in .pc filename.
166 ret = subprocess.run([PCEXE, '--list-all'], stdout=subprocess.PIPE, stderr=subprocess.DEVNULL,
167 universal_newlines=True)
168 if ret.returncode == 0:
169 for pkg in ret.stdout.split('\n'):
170 if pkg.startswith('hdf5'):
171 pkgconfig_files.add(pkg.split(' ', 1)[0])
172
173 for pkg in pkgconfig_files:
174 candidates.append(functools.partial(HDF5PkgConfigDependency, pkg, env, kwargs, language))
175
176 if DependencyMethods.CONFIG_TOOL in methods:
177 candidates.append(functools.partial(HDF5ConfigToolDependency, 'hdf5', env, kwargs, language))
178
179 return candidates
180
[end of mesonbuild/dependencies/hdf5.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/mesonbuild/dependencies/hdf5.py b/mesonbuild/dependencies/hdf5.py
--- a/mesonbuild/dependencies/hdf5.py
+++ b/mesonbuild/dependencies/hdf5.py
@@ -96,13 +96,13 @@
if language == 'c':
cenv = 'CC'
- tools = ['h5cc']
+ tools = ['h5cc', 'h5pcc']
elif language == 'cpp':
cenv = 'CXX'
- tools = ['h5c++']
+ tools = ['h5c++', 'h5pc++']
elif language == 'fortran':
cenv = 'FC'
- tools = ['h5fc']
+ tools = ['h5fc', 'h5pfc']
else:
raise DependencyException('How did you get here?')
|
{"golden_diff": "diff --git a/mesonbuild/dependencies/hdf5.py b/mesonbuild/dependencies/hdf5.py\n--- a/mesonbuild/dependencies/hdf5.py\n+++ b/mesonbuild/dependencies/hdf5.py\n@@ -96,13 +96,13 @@\n \n if language == 'c':\n cenv = 'CC'\n- tools = ['h5cc']\n+ tools = ['h5cc', 'h5pcc']\n elif language == 'cpp':\n cenv = 'CXX'\n- tools = ['h5c++']\n+ tools = ['h5c++', 'h5pc++']\n elif language == 'fortran':\n cenv = 'FC'\n- tools = ['h5fc']\n+ tools = ['h5fc', 'h5pfc']\n else:\n raise DependencyException('How did you get here?')\n", "issue": "Require parallel tools when depending on HDF5\nI have an application that specifically requires the availability of `h5pcc` instead of `h5cc`: https://github.com/mesonbuild/meson/blob/39ceb677744e1f959a0db6406d51a26d78f0ab68/mesonbuild/dependencies/hdf5.py#L97-L99\r\n\r\nIn my case HDF5 is only build with parallel compatibility and therefore `h5cc` is not available. \r\nMy current workaround is to simple edit this line and specify the other tool directly. \r\n\r\nIt there a more appropriate workaround for this issue?\r\nAlternatively: Is it possible to add an optional parameter to `hdf5_dep = dependency('hdf5')` e.g. `dependency('hdf5', parallel: true)`\n", "before_files": [{"content": "# Copyright 2013-2019 The Meson development team\n\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n\n# http://www.apache.org/licenses/LICENSE-2.0\n\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\n# This file contains the detection logic for miscellaneous external dependencies.\n\nimport functools\nimport os\nimport re\nimport shutil\nimport subprocess\nfrom pathlib import Path\n\nfrom ..mesonlib import OrderedSet, join_args\nfrom .base import DependencyException, DependencyMethods\nfrom .configtool import ConfigToolDependency\nfrom .pkgconfig import PkgConfigDependency\nfrom .factory import factory_methods\nimport typing as T\n\nif T.TYPE_CHECKING:\n from .factory import DependencyGenerator\n from ..environment import Environment\n from ..mesonlib import MachineChoice\n\n\nclass HDF5PkgConfigDependency(PkgConfigDependency):\n\n \"\"\"Handle brokenness in the HDF5 pkg-config files.\"\"\"\n\n def __init__(self, name: str, environment: 'Environment', kwargs: T.Dict[str, T.Any], language: T.Optional[str] = None) -> None:\n language = language or 'c'\n if language not in {'c', 'cpp', 'fortran'}:\n raise DependencyException(f'Language {language} is not supported with HDF5.')\n\n super().__init__(name, environment, kwargs, language)\n if not self.is_found:\n return\n\n # some broken pkgconfig don't actually list the full path to the needed includes\n newinc = [] # type: T.List[str]\n for arg in self.compile_args:\n if arg.startswith('-I'):\n stem = 'static' if kwargs.get('static', False) else 'shared'\n if (Path(arg[2:]) / stem).is_dir():\n newinc.append('-I' + str(Path(arg[2:]) / stem))\n self.compile_args += newinc\n\n link_args = [] # type: T.List[str]\n for larg in self.get_link_args():\n lpath = Path(larg)\n # some pkg-config hdf5.pc (e.g. Ubuntu) don't include the commonly-used HL HDF5 libraries,\n # so let's add them if they exist\n # additionally, some pkgconfig HDF5 HL files are malformed so let's be sure to find HL anyway\n if lpath.is_file():\n hl = []\n if language == 'cpp':\n hl += ['_hl_cpp', '_cpp']\n elif language == 'fortran':\n hl += ['_hl_fortran', 'hl_fortran', '_fortran']\n hl += ['_hl'] # C HL library, always needed\n\n suffix = '.' + lpath.name.split('.', 1)[1] # in case of .dll.a\n for h in hl:\n hlfn = lpath.parent / (lpath.name.split('.', 1)[0] + h + suffix)\n if hlfn.is_file():\n link_args.append(str(hlfn))\n # HDF5 C libs are required by other HDF5 languages\n link_args.append(larg)\n else:\n link_args.append(larg)\n\n self.link_args = link_args\n\n\nclass HDF5ConfigToolDependency(ConfigToolDependency):\n\n \"\"\"Wrapper around hdf5 binary config tools.\"\"\"\n\n version_arg = '-showconfig'\n\n def __init__(self, name: str, environment: 'Environment', kwargs: T.Dict[str, T.Any], language: T.Optional[str] = None) -> None:\n language = language or 'c'\n if language not in {'c', 'cpp', 'fortran'}:\n raise DependencyException(f'Language {language} is not supported with HDF5.')\n\n if language == 'c':\n cenv = 'CC'\n tools = ['h5cc']\n elif language == 'cpp':\n cenv = 'CXX'\n tools = ['h5c++']\n elif language == 'fortran':\n cenv = 'FC'\n tools = ['h5fc']\n else:\n raise DependencyException('How did you get here?')\n\n # We need this before we call super()\n for_machine = self.get_for_machine_from_kwargs(kwargs)\n\n nkwargs = kwargs.copy()\n nkwargs['tools'] = tools\n\n # Override the compiler that the config tools are going to use by\n # setting the environment variables that they use for the compiler and\n # linkers.\n compiler = environment.coredata.compilers[for_machine][language]\n try:\n os.environ[f'HDF5_{cenv}'] = join_args(compiler.get_exelist())\n os.environ[f'HDF5_{cenv}LINKER'] = join_args(compiler.get_linker_exelist())\n super().__init__(name, environment, nkwargs, language)\n finally:\n del os.environ[f'HDF5_{cenv}']\n del os.environ[f'HDF5_{cenv}LINKER']\n if not self.is_found:\n return\n\n # We first need to call the tool with -c to get the compile arguments\n # and then without -c to get the link arguments.\n args = self.get_config_value(['-show', '-c'], 'args')[1:]\n args += self.get_config_value(['-show', '-noshlib' if kwargs.get('static', False) else '-shlib'], 'args')[1:]\n for arg in args:\n if arg.startswith(('-I', '-f', '-D')) or arg == '-pthread':\n self.compile_args.append(arg)\n elif arg.startswith(('-L', '-l', '-Wl')):\n self.link_args.append(arg)\n elif Path(arg).is_file():\n self.link_args.append(arg)\n\n # If the language is not C we need to add C as a subdependency\n if language != 'c':\n nkwargs = kwargs.copy()\n nkwargs['language'] = 'c'\n # I'm being too clever for mypy and pylint\n self.is_found = self._add_sub_dependency(hdf5_factory(environment, for_machine, nkwargs)) # pylint: disable=no-value-for-parameter\n\n def _sanitize_version(self, ver: str) -> str:\n v = re.search(r'\\s*HDF5 Version: (\\d+\\.\\d+\\.\\d+)', ver)\n return v.group(1)\n\n\n@factory_methods({DependencyMethods.PKGCONFIG, DependencyMethods.CONFIG_TOOL})\ndef hdf5_factory(env: 'Environment', for_machine: 'MachineChoice',\n kwargs: T.Dict[str, T.Any], methods: T.List[DependencyMethods]) -> T.List['DependencyGenerator']:\n language = kwargs.get('language')\n candidates: T.List['DependencyGenerator'] = []\n\n if DependencyMethods.PKGCONFIG in methods:\n # Use an ordered set so that these remain the first tried pkg-config files\n pkgconfig_files = OrderedSet(['hdf5', 'hdf5-serial'])\n # FIXME: This won't honor pkg-config paths, and cross-native files\n PCEXE = shutil.which('pkg-config')\n if PCEXE:\n # some distros put hdf5-1.2.3.pc with version number in .pc filename.\n ret = subprocess.run([PCEXE, '--list-all'], stdout=subprocess.PIPE, stderr=subprocess.DEVNULL,\n universal_newlines=True)\n if ret.returncode == 0:\n for pkg in ret.stdout.split('\\n'):\n if pkg.startswith('hdf5'):\n pkgconfig_files.add(pkg.split(' ', 1)[0])\n\n for pkg in pkgconfig_files:\n candidates.append(functools.partial(HDF5PkgConfigDependency, pkg, env, kwargs, language))\n\n if DependencyMethods.CONFIG_TOOL in methods:\n candidates.append(functools.partial(HDF5ConfigToolDependency, 'hdf5', env, kwargs, language))\n\n return candidates\n", "path": "mesonbuild/dependencies/hdf5.py"}]}
| 2,932 | 196 |
gh_patches_debug_17463
|
rasdani/github-patches
|
git_diff
|
chainer__chainer-7185
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
Why `F.dropout` usually does not use CuDNN?
Currently, `F.dropout` uses cudnn only when `using_config('use_cudnn', 'always')`
https://github.com/chainer/chainer/blob/d7adbfb45d46fafa6bd0bbb9f0cebe5894489437/chainer/functions/noise/dropout.py#L46
But IMO, cudnn should be enabled when `>=auto`.
</issue>
<code>
[start of chainer/functions/noise/dropout.py]
1 import numpy
2
3 import chainer
4 from chainer.backends import cuda
5 from chainer.backends import intel64
6 from chainer import configuration
7 from chainer import function_node
8 from chainer.utils import argument
9 from chainer.utils import type_check
10
11 if cuda.cudnn_enabled:
12 cudnn = cuda.cudnn
13
14
15 class Dropout(function_node.FunctionNode):
16
17 """Dropout regularization."""
18
19 def __init__(self, dropout_ratio, mask=None, return_mask=False):
20 if not 0.0 <= dropout_ratio < 1.0:
21 raise ValueError('dropout_ratio must be in the range [0, 1)')
22 self.dropout_ratio = dropout_ratio
23 self.mask = mask
24 self.return_mask = return_mask
25 self._use_cudnn = False
26
27 def check_type_forward(self, in_types):
28 type_check._argname(in_types, ('x',))
29 type_check.expect(in_types[0].dtype.kind == 'f')
30
31 def forward_cpu(self, x):
32 if (intel64.should_use_ideep('>=auto')
33 and intel64.inputs_all_ready(x)
34 and self.mask is None):
35 return self._forward_ideep(x)
36
37 if self.mask is not None:
38 y = x[0] * self.mask
39 else:
40 scale = x[0].dtype.type(1. / (1 - self.dropout_ratio))
41 flag = numpy.random.rand(*x[0].shape) >= self.dropout_ratio
42 self.mask = scale * flag
43 y = x[0] * self.mask
44 return y,
45
46 def forward_gpu(self, x):
47 if (chainer.should_use_cudnn('==always', 5000)
48 and x[0].flags.c_contiguous
49 and self.mask is None
50 and not self.return_mask):
51 self._use_cudnn = True
52
53 if hasattr(self, 'states'):
54 # if we already have a dropout mask,
55 # the forward operation is equal to backward.
56 return cuda.get_cudnn_dropout_states().backward(
57 None, x[0], self.dropout_ratio, self.states),
58
59 self.states, y = cuda.get_cudnn_dropout_states().forward(
60 None, x[0], self.dropout_ratio)
61 return y,
62 else:
63 if self.mask is not None:
64 y = x[0] * self.mask
65 else:
66 rand = cuda.cupy.random.rand(*x[0].shape, dtype=numpy.float32)
67 scale = x[0].dtype.type(1. / (1 - self.dropout_ratio))
68 self.mask, y = cuda.elementwise(
69 'T x, R r, T scale, T ratio', 'T mask, T y',
70 '''
71 mask = (r >= ratio) * scale;
72 y = x * mask;
73 ''',
74 'dropout_fwd',
75 )(x[0], rand, scale, self.dropout_ratio)
76 return y,
77
78 def _forward_ideep(self, x):
79 mask, y = intel64.ideep.dropout.Forward(
80 intel64.ideep.array(x[0]),
81 self.dropout_ratio)
82 self.mask = mask
83 return y,
84
85 def backward(self, x, gy):
86 if chainer.should_use_cudnn('==always', 5000) and self._use_cudnn:
87 return DropoutGradCuDNN(self.states, self.dropout_ratio).apply(gy)
88 else:
89 return DropoutGrad(self.mask).apply(gy)
90
91
92 class DropoutGrad(function_node.FunctionNode):
93 """Computes the gradient of the Dropout function."""
94
95 def __init__(self, mask):
96 self.mask = mask
97
98 def forward(self, inputs):
99 if (intel64.should_use_ideep('>=auto')
100 and intel64.inputs_all_ready(inputs)):
101 return self._forward_ideep(inputs)
102
103 y = inputs[0] * self.mask
104 return y,
105
106 def _forward_ideep(self, inputs):
107 return intel64.ideep.dropout.Backward(
108 intel64.ideep.array(self.mask),
109 intel64.ideep.array(inputs[0])),
110
111 def backward(self, indexes, gy):
112 return DropoutGrad(self.mask).apply(gy)
113
114
115 class DropoutGradCuDNN(function_node.FunctionNode):
116 """Computes the gradient of the Dropout function with cuDNN support."""
117
118 def __init__(self, states, dropout_ratio):
119 self.states = states
120 self.dropout_ratio = dropout_ratio
121
122 def forward(self, inputs):
123 return cuda.get_cudnn_dropout_states().backward(
124 None, inputs[0], self.dropout_ratio, self.states),
125
126 def backward(self, indexes, gy):
127 return DropoutGradCuDNN(self.states, self.dropout_ratio).apply(gy)
128
129
130 def dropout(x, ratio=.5, **kwargs):
131 """dropout(x, ratio=.5, *, mask=None, return_mask=False)
132
133 Drops elements of input variable randomly.
134
135 This function drops input elements randomly with probability ``ratio`` and
136 scales the remaining elements by factor ``1 / (1 - ratio)``. In testing
137 mode (i.e., ``chainer.config.train`` is set to ``False``), it does nothing
138 and just returns ``x``.
139
140 Args:
141 x (:class:`~chainer.Variable` or :ref:`ndarray`):
142 Input variable. A :math:`(s_1, s_2, ..., s_N)` -shaped float array.
143 ratio (float):
144 Dropout ratio. The ``ratio`` must be ``0.0 <= ratio < 1.0``.
145 mask (:ref:`ndarray` or None):
146 The mask to be used for dropout.
147 You do not have to specify this value, unless you need to make
148 results deterministic.
149 If ``mask`` is not specified or set to ``None``, a mask will be
150 generated randomly according to the given ``ratio``.
151 If ``mask`` is specified, ``ratio`` will be ignored.
152 The shape and dtype must be the same as ``x`` and should be on the
153 same device.
154 Note that iDeep and cuDNN will not be used for this function if
155 mask is specified, as iDeep and cuDNN do not support it.
156 return_mask (bool):
157 If ``True``, the mask used for dropout is returned together with
158 the output variable.
159 The returned mask can later be reused by passing it to ``mask``
160 argument.
161
162 Returns:
163 ~chainer.Variable or tuple:
164 When ``return_mask`` is ``False`` (default), returns the output
165 variable.
166 When ``True``, returns the tuple of the output variable and
167 mask (:ref:`ndarray`). The mask will be on the same device as the
168 input. The mask will become ``None`` when ``chainer.config.train``
169 is set to ``False``.
170
171 See the paper by G. Hinton: `Improving neural networks by preventing
172 co-adaptation of feature detectors <https://arxiv.org/abs/1207.0580>`_.
173
174 .. admonition:: Example
175
176 >>> x = np.array([[-1, 0], [2, -3], [-2, 1]], np.float32)
177 >>> with chainer.using_config('train', True):
178 ... y = F.dropout(x)
179 >>> y.array
180 array([[-2., 0.],
181 [ 4., -6.],
182 [-0., 2.]], dtype=float32)
183 >>> with chainer.using_config('train', True):
184 ... y = F.dropout(x, ratio=0.0) \
185 # dropout returns original input if ratio=0.0
186 >>> (x == y.array).all()
187 True
188 >>> with chainer.using_config('train', False):
189 ... y = F.dropout(x) \
190 # dropout in test mode returns original input
191 >>> (x == y.array).all()
192 True
193
194 """
195 mask = None
196 return_mask = False
197 if kwargs:
198 mask, return_mask = argument.parse_kwargs(
199 kwargs, ('mask', mask), ('return_mask', return_mask),
200 train='train argument is not supported anymore. '
201 'Use chainer.using_config')
202
203 if configuration.config.train:
204 func = Dropout(ratio, mask, return_mask)
205 out, = func.apply((x,))
206 mask = func.mask
207 else:
208 out = chainer.as_variable(x)
209 mask = None
210
211 if return_mask:
212 return out, mask
213 return out
214
[end of chainer/functions/noise/dropout.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/chainer/functions/noise/dropout.py b/chainer/functions/noise/dropout.py
--- a/chainer/functions/noise/dropout.py
+++ b/chainer/functions/noise/dropout.py
@@ -44,7 +44,7 @@
return y,
def forward_gpu(self, x):
- if (chainer.should_use_cudnn('==always', 5000)
+ if (chainer.should_use_cudnn('>=auto', 5000)
and x[0].flags.c_contiguous
and self.mask is None
and not self.return_mask):
@@ -83,7 +83,7 @@
return y,
def backward(self, x, gy):
- if chainer.should_use_cudnn('==always', 5000) and self._use_cudnn:
+ if chainer.should_use_cudnn('>=auto', 5000) and self._use_cudnn:
return DropoutGradCuDNN(self.states, self.dropout_ratio).apply(gy)
else:
return DropoutGrad(self.mask).apply(gy)
|
{"golden_diff": "diff --git a/chainer/functions/noise/dropout.py b/chainer/functions/noise/dropout.py\n--- a/chainer/functions/noise/dropout.py\n+++ b/chainer/functions/noise/dropout.py\n@@ -44,7 +44,7 @@\n return y,\n \n def forward_gpu(self, x):\n- if (chainer.should_use_cudnn('==always', 5000)\n+ if (chainer.should_use_cudnn('>=auto', 5000)\n and x[0].flags.c_contiguous\n and self.mask is None\n and not self.return_mask):\n@@ -83,7 +83,7 @@\n return y,\n \n def backward(self, x, gy):\n- if chainer.should_use_cudnn('==always', 5000) and self._use_cudnn:\n+ if chainer.should_use_cudnn('>=auto', 5000) and self._use_cudnn:\n return DropoutGradCuDNN(self.states, self.dropout_ratio).apply(gy)\n else:\n return DropoutGrad(self.mask).apply(gy)\n", "issue": "Why `F.dropout` usually does not use CuDNN?\nCurrently, `F.dropout` uses cudnn only when `using_config('use_cudnn', 'always')`\r\nhttps://github.com/chainer/chainer/blob/d7adbfb45d46fafa6bd0bbb9f0cebe5894489437/chainer/functions/noise/dropout.py#L46\r\n\r\nBut IMO, cudnn should be enabled when `>=auto`.\n", "before_files": [{"content": "import numpy\n\nimport chainer\nfrom chainer.backends import cuda\nfrom chainer.backends import intel64\nfrom chainer import configuration\nfrom chainer import function_node\nfrom chainer.utils import argument\nfrom chainer.utils import type_check\n\nif cuda.cudnn_enabled:\n cudnn = cuda.cudnn\n\n\nclass Dropout(function_node.FunctionNode):\n\n \"\"\"Dropout regularization.\"\"\"\n\n def __init__(self, dropout_ratio, mask=None, return_mask=False):\n if not 0.0 <= dropout_ratio < 1.0:\n raise ValueError('dropout_ratio must be in the range [0, 1)')\n self.dropout_ratio = dropout_ratio\n self.mask = mask\n self.return_mask = return_mask\n self._use_cudnn = False\n\n def check_type_forward(self, in_types):\n type_check._argname(in_types, ('x',))\n type_check.expect(in_types[0].dtype.kind == 'f')\n\n def forward_cpu(self, x):\n if (intel64.should_use_ideep('>=auto')\n and intel64.inputs_all_ready(x)\n and self.mask is None):\n return self._forward_ideep(x)\n\n if self.mask is not None:\n y = x[0] * self.mask\n else:\n scale = x[0].dtype.type(1. / (1 - self.dropout_ratio))\n flag = numpy.random.rand(*x[0].shape) >= self.dropout_ratio\n self.mask = scale * flag\n y = x[0] * self.mask\n return y,\n\n def forward_gpu(self, x):\n if (chainer.should_use_cudnn('==always', 5000)\n and x[0].flags.c_contiguous\n and self.mask is None\n and not self.return_mask):\n self._use_cudnn = True\n\n if hasattr(self, 'states'):\n # if we already have a dropout mask,\n # the forward operation is equal to backward.\n return cuda.get_cudnn_dropout_states().backward(\n None, x[0], self.dropout_ratio, self.states),\n\n self.states, y = cuda.get_cudnn_dropout_states().forward(\n None, x[0], self.dropout_ratio)\n return y,\n else:\n if self.mask is not None:\n y = x[0] * self.mask\n else:\n rand = cuda.cupy.random.rand(*x[0].shape, dtype=numpy.float32)\n scale = x[0].dtype.type(1. / (1 - self.dropout_ratio))\n self.mask, y = cuda.elementwise(\n 'T x, R r, T scale, T ratio', 'T mask, T y',\n '''\n mask = (r >= ratio) * scale;\n y = x * mask;\n ''',\n 'dropout_fwd',\n )(x[0], rand, scale, self.dropout_ratio)\n return y,\n\n def _forward_ideep(self, x):\n mask, y = intel64.ideep.dropout.Forward(\n intel64.ideep.array(x[0]),\n self.dropout_ratio)\n self.mask = mask\n return y,\n\n def backward(self, x, gy):\n if chainer.should_use_cudnn('==always', 5000) and self._use_cudnn:\n return DropoutGradCuDNN(self.states, self.dropout_ratio).apply(gy)\n else:\n return DropoutGrad(self.mask).apply(gy)\n\n\nclass DropoutGrad(function_node.FunctionNode):\n \"\"\"Computes the gradient of the Dropout function.\"\"\"\n\n def __init__(self, mask):\n self.mask = mask\n\n def forward(self, inputs):\n if (intel64.should_use_ideep('>=auto')\n and intel64.inputs_all_ready(inputs)):\n return self._forward_ideep(inputs)\n\n y = inputs[0] * self.mask\n return y,\n\n def _forward_ideep(self, inputs):\n return intel64.ideep.dropout.Backward(\n intel64.ideep.array(self.mask),\n intel64.ideep.array(inputs[0])),\n\n def backward(self, indexes, gy):\n return DropoutGrad(self.mask).apply(gy)\n\n\nclass DropoutGradCuDNN(function_node.FunctionNode):\n \"\"\"Computes the gradient of the Dropout function with cuDNN support.\"\"\"\n\n def __init__(self, states, dropout_ratio):\n self.states = states\n self.dropout_ratio = dropout_ratio\n\n def forward(self, inputs):\n return cuda.get_cudnn_dropout_states().backward(\n None, inputs[0], self.dropout_ratio, self.states),\n\n def backward(self, indexes, gy):\n return DropoutGradCuDNN(self.states, self.dropout_ratio).apply(gy)\n\n\ndef dropout(x, ratio=.5, **kwargs):\n \"\"\"dropout(x, ratio=.5, *, mask=None, return_mask=False)\n\n Drops elements of input variable randomly.\n\n This function drops input elements randomly with probability ``ratio`` and\n scales the remaining elements by factor ``1 / (1 - ratio)``. In testing\n mode (i.e., ``chainer.config.train`` is set to ``False``), it does nothing\n and just returns ``x``.\n\n Args:\n x (:class:`~chainer.Variable` or :ref:`ndarray`):\n Input variable. A :math:`(s_1, s_2, ..., s_N)` -shaped float array.\n ratio (float):\n Dropout ratio. The ``ratio`` must be ``0.0 <= ratio < 1.0``.\n mask (:ref:`ndarray` or None):\n The mask to be used for dropout.\n You do not have to specify this value, unless you need to make\n results deterministic.\n If ``mask`` is not specified or set to ``None``, a mask will be\n generated randomly according to the given ``ratio``.\n If ``mask`` is specified, ``ratio`` will be ignored.\n The shape and dtype must be the same as ``x`` and should be on the\n same device.\n Note that iDeep and cuDNN will not be used for this function if\n mask is specified, as iDeep and cuDNN do not support it.\n return_mask (bool):\n If ``True``, the mask used for dropout is returned together with\n the output variable.\n The returned mask can later be reused by passing it to ``mask``\n argument.\n\n Returns:\n ~chainer.Variable or tuple:\n When ``return_mask`` is ``False`` (default), returns the output\n variable.\n When ``True``, returns the tuple of the output variable and\n mask (:ref:`ndarray`). The mask will be on the same device as the\n input. The mask will become ``None`` when ``chainer.config.train``\n is set to ``False``.\n\n See the paper by G. Hinton: `Improving neural networks by preventing\n co-adaptation of feature detectors <https://arxiv.org/abs/1207.0580>`_.\n\n .. admonition:: Example\n\n >>> x = np.array([[-1, 0], [2, -3], [-2, 1]], np.float32)\n >>> with chainer.using_config('train', True):\n ... y = F.dropout(x)\n >>> y.array\n array([[-2., 0.],\n [ 4., -6.],\n [-0., 2.]], dtype=float32)\n >>> with chainer.using_config('train', True):\n ... y = F.dropout(x, ratio=0.0) \\\n# dropout returns original input if ratio=0.0\n >>> (x == y.array).all()\n True\n >>> with chainer.using_config('train', False):\n ... y = F.dropout(x) \\\n# dropout in test mode returns original input\n >>> (x == y.array).all()\n True\n\n \"\"\"\n mask = None\n return_mask = False\n if kwargs:\n mask, return_mask = argument.parse_kwargs(\n kwargs, ('mask', mask), ('return_mask', return_mask),\n train='train argument is not supported anymore. '\n 'Use chainer.using_config')\n\n if configuration.config.train:\n func = Dropout(ratio, mask, return_mask)\n out, = func.apply((x,))\n mask = func.mask\n else:\n out = chainer.as_variable(x)\n mask = None\n\n if return_mask:\n return out, mask\n return out\n", "path": "chainer/functions/noise/dropout.py"}]}
| 3,052 | 254 |
gh_patches_debug_16806
|
rasdani/github-patches
|
git_diff
|
AUTOMATIC1111__stable-diffusion-webui-8931
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
[Bug]: loopback.py match-case traceback on Google Colab
### Is there an existing issue for this?
- [X] I have searched the existing issues and checked the recent builds/commits
### What happened?
what happened? I don't really know.
all of a sudden I'm getting tracebacks and the WebUI won't load properly, making it unuseable.


### Steps to reproduce the problem
1. open [this colab notebook](https://colab.research.google.com/github/shirooo39/MiXLab/blob/master/MiXLab.ipynb#scrollTo=6_Sxep-D4HEf)
2. change the runtime type to GPU
3. run the cell: Initialize
4. urn the cell: Download Diffusion Models
5. run the cell: Start WebUI
6. wait for error
### What should have happened?
it shouldn't've thrown those tracebacks and the WebUI loads properly.
### Commit where the problem happens
91ae48fd
### What platforms do you use to access the UI ?
Other/Cloud
### What browsers do you use to access the UI ?
Mozilla Firefox
### Command Line Arguments
```Shell
!COMMANDLINE_ARGS="--share --api --gradio-debug --gradio-queue --medvram --xformers --disable-safe-unpickle --enable-insecure-extension-access --no-half-vae --no-download-sd-model --disable-console-progressbars" REQS_FILE="requirements.txt" python launch.py
```
### List of extensions
No
### Console logs
```Shell
https://rentry.org/a1111webui91ae48fd
```
### Additional information
I'm the author of MiXLab notebook and I wrote my own simple Colab implementation of your WebUI.
it was working fine last night as of 2023-03-24, around 11 PM UTC+07:00, I was able to launch the WebUI without getting any tracebacks.
I was able to generate images at that time (the screenshot below is an image I generated on Colab and downloaded locally)

I'm constantly making changes to my notebook BUT I did not touch to the "Initialize" and "Start WebUI" cells.
I only touch to the model downloader cells.
so I don't think the issue is within my notebook, but I haven't tried TheLastBen's notebook either...
Colab still uses Python 3.7 and the match-case used in the loopback.py was only introduced in Python 3.10 and newer.
that could be the reason why Colab is throwing out tracebacks.
</issue>
<code>
[start of scripts/loopback.py]
1 import math
2
3 import gradio as gr
4 import modules.scripts as scripts
5 from modules import deepbooru, images, processing, shared
6 from modules.processing import Processed
7 from modules.shared import opts, state
8
9
10 class Script(scripts.Script):
11 def title(self):
12 return "Loopback"
13
14 def show(self, is_img2img):
15 return is_img2img
16
17 def ui(self, is_img2img):
18 loops = gr.Slider(minimum=1, maximum=32, step=1, label='Loops', value=4, elem_id=self.elem_id("loops"))
19 final_denoising_strength = gr.Slider(minimum=0, maximum=1, step=0.01, label='Final denoising strength', value=0.5, elem_id=self.elem_id("final_denoising_strength"))
20 denoising_curve = gr.Dropdown(label="Denoising strength curve", choices=["Aggressive", "Linear", "Lazy"], value="Linear")
21 append_interrogation = gr.Dropdown(label="Append interrogated prompt at each iteration", choices=["None", "CLIP", "DeepBooru"], value="None")
22
23 return [loops, final_denoising_strength, denoising_curve, append_interrogation]
24
25 def run(self, p, loops, final_denoising_strength, denoising_curve, append_interrogation):
26 processing.fix_seed(p)
27 batch_count = p.n_iter
28 p.extra_generation_params = {
29 "Final denoising strength": final_denoising_strength,
30 "Denoising curve": denoising_curve
31 }
32
33 p.batch_size = 1
34 p.n_iter = 1
35
36 info = None
37 initial_seed = None
38 initial_info = None
39 initial_denoising_strength = p.denoising_strength
40
41 grids = []
42 all_images = []
43 original_init_image = p.init_images
44 original_prompt = p.prompt
45 original_inpainting_fill = p.inpainting_fill
46 state.job_count = loops * batch_count
47
48 initial_color_corrections = [processing.setup_color_correction(p.init_images[0])]
49
50 def calculate_denoising_strength(loop):
51 strength = initial_denoising_strength
52
53 if loops == 1:
54 return strength
55
56 progress = loop / (loops - 1)
57 match denoising_curve:
58 case "Aggressive":
59 strength = math.sin((progress) * math.pi * 0.5)
60
61 case "Lazy":
62 strength = 1 - math.cos((progress) * math.pi * 0.5)
63
64 case _:
65 strength = progress
66
67 change = (final_denoising_strength - initial_denoising_strength) * strength
68 return initial_denoising_strength + change
69
70 history = []
71
72 for n in range(batch_count):
73 # Reset to original init image at the start of each batch
74 p.init_images = original_init_image
75
76 # Reset to original denoising strength
77 p.denoising_strength = initial_denoising_strength
78
79 last_image = None
80
81 for i in range(loops):
82 p.n_iter = 1
83 p.batch_size = 1
84 p.do_not_save_grid = True
85
86 if opts.img2img_color_correction:
87 p.color_corrections = initial_color_corrections
88
89 if append_interrogation != "None":
90 p.prompt = original_prompt + ", " if original_prompt != "" else ""
91 if append_interrogation == "CLIP":
92 p.prompt += shared.interrogator.interrogate(p.init_images[0])
93 elif append_interrogation == "DeepBooru":
94 p.prompt += deepbooru.model.tag(p.init_images[0])
95
96 state.job = f"Iteration {i + 1}/{loops}, batch {n + 1}/{batch_count}"
97
98 processed = processing.process_images(p)
99
100 # Generation cancelled.
101 if state.interrupted:
102 break
103
104 if initial_seed is None:
105 initial_seed = processed.seed
106 initial_info = processed.info
107
108 p.seed = processed.seed + 1
109 p.denoising_strength = calculate_denoising_strength(i + 1)
110
111 if state.skipped:
112 break
113
114 last_image = processed.images[0]
115 p.init_images = [last_image]
116 p.inpainting_fill = 1 # Set "masked content" to "original" for next loop.
117
118 if batch_count == 1:
119 history.append(last_image)
120 all_images.append(last_image)
121
122 if batch_count > 1 and not state.skipped and not state.interrupted:
123 history.append(last_image)
124 all_images.append(last_image)
125
126 p.inpainting_fill = original_inpainting_fill
127
128 if state.interrupted:
129 break
130
131 if len(history) > 1:
132 grid = images.image_grid(history, rows=1)
133 if opts.grid_save:
134 images.save_image(grid, p.outpath_grids, "grid", initial_seed, p.prompt, opts.grid_format, info=info, short_filename=not opts.grid_extended_filename, grid=True, p=p)
135
136 if opts.return_grid:
137 grids.append(grid)
138
139 all_images = grids + all_images
140
141 processed = Processed(p, all_images, initial_seed, initial_info)
142
143 return processed
144
[end of scripts/loopback.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/scripts/loopback.py b/scripts/loopback.py
--- a/scripts/loopback.py
+++ b/scripts/loopback.py
@@ -54,15 +54,12 @@
return strength
progress = loop / (loops - 1)
- match denoising_curve:
- case "Aggressive":
- strength = math.sin((progress) * math.pi * 0.5)
-
- case "Lazy":
- strength = 1 - math.cos((progress) * math.pi * 0.5)
-
- case _:
- strength = progress
+ if denoising_curve == "Aggressive":
+ strength = math.sin((progress) * math.pi * 0.5)
+ elif denoising_curve == "Lazy":
+ strength = 1 - math.cos((progress) * math.pi * 0.5)
+ else:
+ strength = progress
change = (final_denoising_strength - initial_denoising_strength) * strength
return initial_denoising_strength + change
|
{"golden_diff": "diff --git a/scripts/loopback.py b/scripts/loopback.py\n--- a/scripts/loopback.py\n+++ b/scripts/loopback.py\n@@ -54,15 +54,12 @@\n return strength\r\n \r\n progress = loop / (loops - 1)\r\n- match denoising_curve:\r\n- case \"Aggressive\":\r\n- strength = math.sin((progress) * math.pi * 0.5)\r\n-\r\n- case \"Lazy\":\r\n- strength = 1 - math.cos((progress) * math.pi * 0.5)\r\n-\r\n- case _:\r\n- strength = progress\r\n+ if denoising_curve == \"Aggressive\":\r\n+ strength = math.sin((progress) * math.pi * 0.5)\r\n+ elif denoising_curve == \"Lazy\":\r\n+ strength = 1 - math.cos((progress) * math.pi * 0.5)\r\n+ else:\r\n+ strength = progress\r\n \r\n change = (final_denoising_strength - initial_denoising_strength) * strength\r\n return initial_denoising_strength + change\n", "issue": "[Bug]: loopback.py match-case traceback on Google Colab\n### Is there an existing issue for this?\n\n- [X] I have searched the existing issues and checked the recent builds/commits\n\n### What happened?\n\nwhat happened? I don't really know.\r\nall of a sudden I'm getting tracebacks and the WebUI won't load properly, making it unuseable.\r\n\r\n\r\n\n\n### Steps to reproduce the problem\n\n1. open [this colab notebook](https://colab.research.google.com/github/shirooo39/MiXLab/blob/master/MiXLab.ipynb#scrollTo=6_Sxep-D4HEf)\r\n2. change the runtime type to GPU\r\n3. run the cell: Initialize\r\n4. urn the cell: Download Diffusion Models\r\n5. run the cell: Start WebUI\r\n6. wait for error\n\n### What should have happened?\n\nit shouldn't've thrown those tracebacks and the WebUI loads properly.\n\n### Commit where the problem happens\n\n91ae48fd\n\n### What platforms do you use to access the UI ?\n\nOther/Cloud\n\n### What browsers do you use to access the UI ?\n\nMozilla Firefox\n\n### Command Line Arguments\n\n```Shell\n!COMMANDLINE_ARGS=\"--share --api --gradio-debug --gradio-queue --medvram --xformers --disable-safe-unpickle --enable-insecure-extension-access --no-half-vae --no-download-sd-model --disable-console-progressbars\" REQS_FILE=\"requirements.txt\" python launch.py\n```\n\n\n### List of extensions\n\nNo\n\n### Console logs\n\n```Shell\nhttps://rentry.org/a1111webui91ae48fd\n```\n\n\n### Additional information\n\nI'm the author of MiXLab notebook and I wrote my own simple Colab implementation of your WebUI.\r\nit was working fine last night as of 2023-03-24, around 11 PM UTC+07:00, I was able to launch the WebUI without getting any tracebacks.\r\n\r\nI was able to generate images at that time (the screenshot below is an image I generated on Colab and downloaded locally)\r\n\r\n\r\nI'm constantly making changes to my notebook BUT I did not touch to the \"Initialize\" and \"Start WebUI\" cells. \r\nI only touch to the model downloader cells. \r\nso I don't think the issue is within my notebook, but I haven't tried TheLastBen's notebook either...\r\n\r\nColab still uses Python 3.7 and the match-case used in the loopback.py was only introduced in Python 3.10 and newer. \r\nthat could be the reason why Colab is throwing out tracebacks.\n", "before_files": [{"content": "import math\r\n\r\nimport gradio as gr\r\nimport modules.scripts as scripts\r\nfrom modules import deepbooru, images, processing, shared\r\nfrom modules.processing import Processed\r\nfrom modules.shared import opts, state\r\n\r\n\r\nclass Script(scripts.Script):\r\n def title(self):\r\n return \"Loopback\"\r\n\r\n def show(self, is_img2img):\r\n return is_img2img\r\n\r\n def ui(self, is_img2img): \r\n loops = gr.Slider(minimum=1, maximum=32, step=1, label='Loops', value=4, elem_id=self.elem_id(\"loops\"))\r\n final_denoising_strength = gr.Slider(minimum=0, maximum=1, step=0.01, label='Final denoising strength', value=0.5, elem_id=self.elem_id(\"final_denoising_strength\"))\r\n denoising_curve = gr.Dropdown(label=\"Denoising strength curve\", choices=[\"Aggressive\", \"Linear\", \"Lazy\"], value=\"Linear\")\r\n append_interrogation = gr.Dropdown(label=\"Append interrogated prompt at each iteration\", choices=[\"None\", \"CLIP\", \"DeepBooru\"], value=\"None\")\r\n\r\n return [loops, final_denoising_strength, denoising_curve, append_interrogation]\r\n\r\n def run(self, p, loops, final_denoising_strength, denoising_curve, append_interrogation):\r\n processing.fix_seed(p)\r\n batch_count = p.n_iter\r\n p.extra_generation_params = {\r\n \"Final denoising strength\": final_denoising_strength,\r\n \"Denoising curve\": denoising_curve\r\n }\r\n\r\n p.batch_size = 1\r\n p.n_iter = 1\r\n\r\n info = None\r\n initial_seed = None\r\n initial_info = None\r\n initial_denoising_strength = p.denoising_strength\r\n\r\n grids = []\r\n all_images = []\r\n original_init_image = p.init_images\r\n original_prompt = p.prompt\r\n original_inpainting_fill = p.inpainting_fill\r\n state.job_count = loops * batch_count\r\n\r\n initial_color_corrections = [processing.setup_color_correction(p.init_images[0])]\r\n\r\n def calculate_denoising_strength(loop):\r\n strength = initial_denoising_strength\r\n\r\n if loops == 1:\r\n return strength\r\n\r\n progress = loop / (loops - 1)\r\n match denoising_curve:\r\n case \"Aggressive\":\r\n strength = math.sin((progress) * math.pi * 0.5)\r\n\r\n case \"Lazy\":\r\n strength = 1 - math.cos((progress) * math.pi * 0.5)\r\n\r\n case _:\r\n strength = progress\r\n\r\n change = (final_denoising_strength - initial_denoising_strength) * strength\r\n return initial_denoising_strength + change\r\n\r\n history = []\r\n\r\n for n in range(batch_count):\r\n # Reset to original init image at the start of each batch\r\n p.init_images = original_init_image\r\n\r\n # Reset to original denoising strength\r\n p.denoising_strength = initial_denoising_strength\r\n\r\n last_image = None\r\n\r\n for i in range(loops):\r\n p.n_iter = 1\r\n p.batch_size = 1\r\n p.do_not_save_grid = True\r\n\r\n if opts.img2img_color_correction:\r\n p.color_corrections = initial_color_corrections\r\n\r\n if append_interrogation != \"None\":\r\n p.prompt = original_prompt + \", \" if original_prompt != \"\" else \"\"\r\n if append_interrogation == \"CLIP\":\r\n p.prompt += shared.interrogator.interrogate(p.init_images[0])\r\n elif append_interrogation == \"DeepBooru\":\r\n p.prompt += deepbooru.model.tag(p.init_images[0])\r\n\r\n state.job = f\"Iteration {i + 1}/{loops}, batch {n + 1}/{batch_count}\"\r\n\r\n processed = processing.process_images(p)\r\n\r\n # Generation cancelled.\r\n if state.interrupted:\r\n break\r\n\r\n if initial_seed is None:\r\n initial_seed = processed.seed\r\n initial_info = processed.info\r\n\r\n p.seed = processed.seed + 1\r\n p.denoising_strength = calculate_denoising_strength(i + 1)\r\n \r\n if state.skipped:\r\n break\r\n\r\n last_image = processed.images[0]\r\n p.init_images = [last_image]\r\n p.inpainting_fill = 1 # Set \"masked content\" to \"original\" for next loop.\r\n\r\n if batch_count == 1:\r\n history.append(last_image)\r\n all_images.append(last_image)\r\n\r\n if batch_count > 1 and not state.skipped and not state.interrupted:\r\n history.append(last_image)\r\n all_images.append(last_image)\r\n\r\n p.inpainting_fill = original_inpainting_fill\r\n \r\n if state.interrupted:\r\n break\r\n\r\n if len(history) > 1:\r\n grid = images.image_grid(history, rows=1)\r\n if opts.grid_save:\r\n images.save_image(grid, p.outpath_grids, \"grid\", initial_seed, p.prompt, opts.grid_format, info=info, short_filename=not opts.grid_extended_filename, grid=True, p=p)\r\n\r\n if opts.return_grid:\r\n grids.append(grid)\r\n \r\n all_images = grids + all_images\r\n\r\n processed = Processed(p, all_images, initial_seed, initial_info)\r\n\r\n return processed\r\n", "path": "scripts/loopback.py"}]}
| 2,769 | 237 |
gh_patches_debug_21670
|
rasdani/github-patches
|
git_diff
|
open-mmlab__mmpretrain-474
|
You will be provided with a partial code base and an issue statement explaining a problem to resolve.
<issue>
RuntimeError: Pin memory thread exited unexpectedly
2021-10-04 15:14:24,792 - mmcls - INFO - workflow: [('train', 1)], max: 200 epochs
2021-10-04 15:14:33,806 - mmcls - INFO - Epoch [1][100/391] lr: 1.000e-01, eta: 1:57:16, time: 0.090, data_time: 0.022, memory: 76, loss: 2.0658
2021-10-04 15:14:39,876 - mmcls - INFO - Epoch [1][200/391] lr: 1.000e-01, eta: 1:38:01, time: 0.061, data_time: 0.000, memory: 76, loss: 1.8056
2021-10-04 15:14:45,923 - mmcls - INFO - Epoch [1][300/391] lr: 1.000e-01, eta: 1:31:25, time: 0.060, data_time: 0.000, memory: 76, loss: 1.6816
2021-10-04 15:14:51,428 - mmcls - INFO - Saving checkpoint at 1 epochs
[>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>] 10000/10000, 7696.3 task/s, elapsed: 1s, ETA: 0s2021-10-04 15:14:52,802 - mmcls - INFO - Epoch(val) [1][79] accuracy_top-1: 32.6600, accuracy_top-5: 84.0200
Exception in thread Thread-2:
Traceback (most recent call last):
File "/home/songyuan/anaconda3/envs/pytorch1.7.1/lib/python3.7/threading.py", line 926, in _bootstrap_inner
self.run()
File "/home/songyuan/anaconda3/envs/pytorch1.7.1/lib/python3.7/threading.py", line 870, in run
self._target(*self._args, **self._kwargs)
File "/home/songyuan/anaconda3/envs/pytorch1.7.1/lib/python3.7/site-packages/torch/utils/data/_utils/pin_memory.py", line 28, in _pin_memory_loop
idx, data = r
ValueError: not enough values to unpack (expected 2, got 0)
Traceback (most recent call last):
File "tools/train.py", line 157, in <module>
main()
File "tools/train.py", line 153, in main
meta=meta)
File "/home/songyuan/yty/mmclassification-0.15.0/mmcls/apis/train.py", line 160, in train_model
runner.run(data_loaders, cfg.workflow)
File "/home/songyuan/anaconda3/envs/pytorch1.7.1/lib/python3.7/site-packages/mmcv/runner/epoch_based_runner.py", line 127, in run
epoch_runner(data_loaders[i], **kwargs)
File "/home/songyuan/anaconda3/envs/pytorch1.7.1/lib/python3.7/site-packages/mmcv/runner/epoch_based_runner.py", line 47, in train
for i, data_batch in enumerate(self.data_loader):
File "/home/songyuan/anaconda3/envs/pytorch1.7.1/lib/python3.7/site-packages/torch/utils/data/dataloader.py", line 349, in __iter__
self._iterator._reset(self)
File "/home/songyuan/anaconda3/envs/pytorch1.7.1/lib/python3.7/site-packages/torch/utils/data/dataloader.py", line 852, in _reset
data = self._get_data()
File "/home/songyuan/anaconda3/envs/pytorch1.7.1/lib/python3.7/site-packages/torch/utils/data/dataloader.py", line 1029, in _get_data
raise RuntimeError('Pin memory thread exited unexpectedly')
RuntimeError: Pin memory thread exited unexpectedly
我用mmclassification-0.15.0在3090上跑的,pytorch版本为1.7,mmcv为1.3.14,然后训了一个epoch并完成验证后,就出现了这个问题。请问这是什么原因?
</issue>
<code>
[start of mmcls/datasets/builder.py]
1 # Copyright (c) OpenMMLab. All rights reserved.
2 import platform
3 import random
4 from distutils.version import LooseVersion
5 from functools import partial
6
7 import numpy as np
8 import torch
9 from mmcv.parallel import collate
10 from mmcv.runner import get_dist_info
11 from mmcv.utils import Registry, build_from_cfg
12 from torch.utils.data import DataLoader
13
14 from .samplers import DistributedSampler
15
16 if platform.system() != 'Windows':
17 # https://github.com/pytorch/pytorch/issues/973
18 import resource
19 rlimit = resource.getrlimit(resource.RLIMIT_NOFILE)
20 hard_limit = rlimit[1]
21 soft_limit = min(4096, hard_limit)
22 resource.setrlimit(resource.RLIMIT_NOFILE, (soft_limit, hard_limit))
23
24 DATASETS = Registry('dataset')
25 PIPELINES = Registry('pipeline')
26
27
28 def build_dataset(cfg, default_args=None):
29 from .dataset_wrappers import (ConcatDataset, RepeatDataset,
30 ClassBalancedDataset)
31 if isinstance(cfg, (list, tuple)):
32 dataset = ConcatDataset([build_dataset(c, default_args) for c in cfg])
33 elif cfg['type'] == 'RepeatDataset':
34 dataset = RepeatDataset(
35 build_dataset(cfg['dataset'], default_args), cfg['times'])
36 elif cfg['type'] == 'ClassBalancedDataset':
37 dataset = ClassBalancedDataset(
38 build_dataset(cfg['dataset'], default_args), cfg['oversample_thr'])
39 else:
40 dataset = build_from_cfg(cfg, DATASETS, default_args)
41
42 return dataset
43
44
45 def build_dataloader(dataset,
46 samples_per_gpu,
47 workers_per_gpu,
48 num_gpus=1,
49 dist=True,
50 shuffle=True,
51 round_up=True,
52 seed=None,
53 pin_memory=True,
54 persistent_workers=True,
55 **kwargs):
56 """Build PyTorch DataLoader.
57
58 In distributed training, each GPU/process has a dataloader.
59 In non-distributed training, there is only one dataloader for all GPUs.
60
61 Args:
62 dataset (Dataset): A PyTorch dataset.
63 samples_per_gpu (int): Number of training samples on each GPU, i.e.,
64 batch size of each GPU.
65 workers_per_gpu (int): How many subprocesses to use for data loading
66 for each GPU.
67 num_gpus (int): Number of GPUs. Only used in non-distributed training.
68 dist (bool): Distributed training/test or not. Default: True.
69 shuffle (bool): Whether to shuffle the data at every epoch.
70 Default: True.
71 round_up (bool): Whether to round up the length of dataset by adding
72 extra samples to make it evenly divisible. Default: True.
73 pin_memory (bool): Whether to use pin_memory in DataLoader.
74 Default: True
75 persistent_workers (bool): If True, the data loader will not shutdown
76 the worker processes after a dataset has been consumed once.
77 This allows to maintain the workers Dataset instances alive.
78 The argument also has effect in PyTorch>=1.7.0.
79 Default: True
80 kwargs: any keyword argument to be used to initialize DataLoader
81
82 Returns:
83 DataLoader: A PyTorch dataloader.
84 """
85 rank, world_size = get_dist_info()
86 if dist:
87 sampler = DistributedSampler(
88 dataset, world_size, rank, shuffle=shuffle, round_up=round_up)
89 shuffle = False
90 batch_size = samples_per_gpu
91 num_workers = workers_per_gpu
92 else:
93 sampler = None
94 batch_size = num_gpus * samples_per_gpu
95 num_workers = num_gpus * workers_per_gpu
96
97 init_fn = partial(
98 worker_init_fn, num_workers=num_workers, rank=rank,
99 seed=seed) if seed is not None else None
100
101 if LooseVersion(torch.__version__) >= LooseVersion('1.7.0'):
102 kwargs['persistent_workers'] = persistent_workers
103
104 data_loader = DataLoader(
105 dataset,
106 batch_size=batch_size,
107 sampler=sampler,
108 num_workers=num_workers,
109 collate_fn=partial(collate, samples_per_gpu=samples_per_gpu),
110 pin_memory=pin_memory,
111 shuffle=shuffle,
112 worker_init_fn=init_fn,
113 **kwargs)
114
115 return data_loader
116
117
118 def worker_init_fn(worker_id, num_workers, rank, seed):
119 # The seed of each worker equals to
120 # num_worker * rank + worker_id + user_seed
121 worker_seed = num_workers * rank + worker_id + seed
122 np.random.seed(worker_seed)
123 random.seed(worker_seed)
124
[end of mmcls/datasets/builder.py]
</code>
I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format.
<patch>
diff --git a/file.py b/file.py
--- a/file.py
+++ b/file.py
@@ -1,27 +1,35 @@
def euclidean(a, b):
- while b:
- a, b = b, a % b
- return a
+ if b == 0:
+ return a
+ return euclidean(b, a % b)
def bresenham(x0, y0, x1, y1):
points = []
dx = abs(x1 - x0)
dy = abs(y1 - y0)
- sx = 1 if x0 < x1 else -1
- sy = 1 if y0 < y1 else -1
- err = dx - dy
+ x, y = x0, y0
+ sx = -1 if x0 > x1 else 1
+ sy = -1 if y0 > y1 else 1
- while True:
- points.append((x0, y0))
- if x0 == x1 and y0 == y1:
- break
- e2 = 2 * err
- if e2 > -dy:
- err -= dy
- x0 += sx
- if e2 < dx:
- err += dx
- y0 += sy
+ if dx > dy:
+ err = dx / 2.0
+ while x != x1:
+ points.append((x, y))
+ err -= dy
+ if err < 0:
+ y += sy
+ err += dx
+ x += sx
+ else:
+ err = dy / 2.0
+ while y != y1:
+ points.append((x, y))
+ err -= dx
+ if err < 0:
+ x += sx
+ err += dy
+ y += sy
+
+ points.append((x, y))
return points
</patch>
|
diff --git a/mmcls/datasets/builder.py b/mmcls/datasets/builder.py
--- a/mmcls/datasets/builder.py
+++ b/mmcls/datasets/builder.py
@@ -1,14 +1,13 @@
# Copyright (c) OpenMMLab. All rights reserved.
import platform
import random
-from distutils.version import LooseVersion
from functools import partial
import numpy as np
import torch
from mmcv.parallel import collate
from mmcv.runner import get_dist_info
-from mmcv.utils import Registry, build_from_cfg
+from mmcv.utils import Registry, build_from_cfg, digit_version
from torch.utils.data import DataLoader
from .samplers import DistributedSampler
@@ -98,7 +97,7 @@
worker_init_fn, num_workers=num_workers, rank=rank,
seed=seed) if seed is not None else None
- if LooseVersion(torch.__version__) >= LooseVersion('1.7.0'):
+ if digit_version(torch.__version__) >= digit_version('1.8.0'):
kwargs['persistent_workers'] = persistent_workers
data_loader = DataLoader(
|
{"golden_diff": "diff --git a/mmcls/datasets/builder.py b/mmcls/datasets/builder.py\n--- a/mmcls/datasets/builder.py\n+++ b/mmcls/datasets/builder.py\n@@ -1,14 +1,13 @@\n # Copyright (c) OpenMMLab. All rights reserved.\n import platform\n import random\n-from distutils.version import LooseVersion\n from functools import partial\n \n import numpy as np\n import torch\n from mmcv.parallel import collate\n from mmcv.runner import get_dist_info\n-from mmcv.utils import Registry, build_from_cfg\n+from mmcv.utils import Registry, build_from_cfg, digit_version\n from torch.utils.data import DataLoader\n \n from .samplers import DistributedSampler\n@@ -98,7 +97,7 @@\n worker_init_fn, num_workers=num_workers, rank=rank,\n seed=seed) if seed is not None else None\n \n- if LooseVersion(torch.__version__) >= LooseVersion('1.7.0'):\n+ if digit_version(torch.__version__) >= digit_version('1.8.0'):\n kwargs['persistent_workers'] = persistent_workers\n \n data_loader = DataLoader(\n", "issue": "RuntimeError: Pin memory thread exited unexpectedly\n2021-10-04 15:14:24,792 - mmcls - INFO - workflow: [('train', 1)], max: 200 epochs\r\n2021-10-04 15:14:33,806 - mmcls - INFO - Epoch [1][100/391]\tlr: 1.000e-01, eta: 1:57:16, time: 0.090, data_time: 0.022, memory: 76, loss: 2.0658\r\n2021-10-04 15:14:39,876 - mmcls - INFO - Epoch [1][200/391]\tlr: 1.000e-01, eta: 1:38:01, time: 0.061, data_time: 0.000, memory: 76, loss: 1.8056\r\n2021-10-04 15:14:45,923 - mmcls - INFO - Epoch [1][300/391]\tlr: 1.000e-01, eta: 1:31:25, time: 0.060, data_time: 0.000, memory: 76, loss: 1.6816\r\n2021-10-04 15:14:51,428 - mmcls - INFO - Saving checkpoint at 1 epochs\r\n[>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>] 10000/10000, 7696.3 task/s, elapsed: 1s, ETA: 0s2021-10-04 15:14:52,802 - mmcls - INFO - Epoch(val) [1][79]\taccuracy_top-1: 32.6600, accuracy_top-5: 84.0200\r\nException in thread Thread-2:\r\nTraceback (most recent call last):\r\n File \"/home/songyuan/anaconda3/envs/pytorch1.7.1/lib/python3.7/threading.py\", line 926, in _bootstrap_inner\r\n self.run()\r\n File \"/home/songyuan/anaconda3/envs/pytorch1.7.1/lib/python3.7/threading.py\", line 870, in run\r\n self._target(*self._args, **self._kwargs)\r\n File \"/home/songyuan/anaconda3/envs/pytorch1.7.1/lib/python3.7/site-packages/torch/utils/data/_utils/pin_memory.py\", line 28, in _pin_memory_loop\r\n idx, data = r\r\nValueError: not enough values to unpack (expected 2, got 0)\r\n\r\nTraceback (most recent call last):\r\n File \"tools/train.py\", line 157, in <module>\r\n main()\r\n File \"tools/train.py\", line 153, in main\r\n meta=meta)\r\n File \"/home/songyuan/yty/mmclassification-0.15.0/mmcls/apis/train.py\", line 160, in train_model\r\n runner.run(data_loaders, cfg.workflow)\r\n File \"/home/songyuan/anaconda3/envs/pytorch1.7.1/lib/python3.7/site-packages/mmcv/runner/epoch_based_runner.py\", line 127, in run\r\n epoch_runner(data_loaders[i], **kwargs)\r\n File \"/home/songyuan/anaconda3/envs/pytorch1.7.1/lib/python3.7/site-packages/mmcv/runner/epoch_based_runner.py\", line 47, in train\r\n for i, data_batch in enumerate(self.data_loader):\r\n File \"/home/songyuan/anaconda3/envs/pytorch1.7.1/lib/python3.7/site-packages/torch/utils/data/dataloader.py\", line 349, in __iter__\r\n self._iterator._reset(self)\r\n File \"/home/songyuan/anaconda3/envs/pytorch1.7.1/lib/python3.7/site-packages/torch/utils/data/dataloader.py\", line 852, in _reset\r\n data = self._get_data()\r\n File \"/home/songyuan/anaconda3/envs/pytorch1.7.1/lib/python3.7/site-packages/torch/utils/data/dataloader.py\", line 1029, in _get_data\r\n raise RuntimeError('Pin memory thread exited unexpectedly')\r\nRuntimeError: Pin memory thread exited unexpectedly\r\n\r\n\r\n\u6211\u7528mmclassification-0.15.0\u57283090\u4e0a\u8dd1\u7684\uff0cpytorch\u7248\u672c\u4e3a1.7\uff0cmmcv\u4e3a1.3.14\uff0c\u7136\u540e\u8bad\u4e86\u4e00\u4e2aepoch\u5e76\u5b8c\u6210\u9a8c\u8bc1\u540e\uff0c\u5c31\u51fa\u73b0\u4e86\u8fd9\u4e2a\u95ee\u9898\u3002\u8bf7\u95ee\u8fd9\u662f\u4ec0\u4e48\u539f\u56e0?\n", "before_files": [{"content": "# Copyright (c) OpenMMLab. All rights reserved.\nimport platform\nimport random\nfrom distutils.version import LooseVersion\nfrom functools import partial\n\nimport numpy as np\nimport torch\nfrom mmcv.parallel import collate\nfrom mmcv.runner import get_dist_info\nfrom mmcv.utils import Registry, build_from_cfg\nfrom torch.utils.data import DataLoader\n\nfrom .samplers import DistributedSampler\n\nif platform.system() != 'Windows':\n # https://github.com/pytorch/pytorch/issues/973\n import resource\n rlimit = resource.getrlimit(resource.RLIMIT_NOFILE)\n hard_limit = rlimit[1]\n soft_limit = min(4096, hard_limit)\n resource.setrlimit(resource.RLIMIT_NOFILE, (soft_limit, hard_limit))\n\nDATASETS = Registry('dataset')\nPIPELINES = Registry('pipeline')\n\n\ndef build_dataset(cfg, default_args=None):\n from .dataset_wrappers import (ConcatDataset, RepeatDataset,\n ClassBalancedDataset)\n if isinstance(cfg, (list, tuple)):\n dataset = ConcatDataset([build_dataset(c, default_args) for c in cfg])\n elif cfg['type'] == 'RepeatDataset':\n dataset = RepeatDataset(\n build_dataset(cfg['dataset'], default_args), cfg['times'])\n elif cfg['type'] == 'ClassBalancedDataset':\n dataset = ClassBalancedDataset(\n build_dataset(cfg['dataset'], default_args), cfg['oversample_thr'])\n else:\n dataset = build_from_cfg(cfg, DATASETS, default_args)\n\n return dataset\n\n\ndef build_dataloader(dataset,\n samples_per_gpu,\n workers_per_gpu,\n num_gpus=1,\n dist=True,\n shuffle=True,\n round_up=True,\n seed=None,\n pin_memory=True,\n persistent_workers=True,\n **kwargs):\n \"\"\"Build PyTorch DataLoader.\n\n In distributed training, each GPU/process has a dataloader.\n In non-distributed training, there is only one dataloader for all GPUs.\n\n Args:\n dataset (Dataset): A PyTorch dataset.\n samples_per_gpu (int): Number of training samples on each GPU, i.e.,\n batch size of each GPU.\n workers_per_gpu (int): How many subprocesses to use for data loading\n for each GPU.\n num_gpus (int): Number of GPUs. Only used in non-distributed training.\n dist (bool): Distributed training/test or not. Default: True.\n shuffle (bool): Whether to shuffle the data at every epoch.\n Default: True.\n round_up (bool): Whether to round up the length of dataset by adding\n extra samples to make it evenly divisible. Default: True.\n pin_memory (bool): Whether to use pin_memory in DataLoader.\n Default: True\n persistent_workers (bool): If True, the data loader will not shutdown\n the worker processes after a dataset has been consumed once.\n This allows to maintain the workers Dataset instances alive.\n The argument also has effect in PyTorch>=1.7.0.\n Default: True\n kwargs: any keyword argument to be used to initialize DataLoader\n\n Returns:\n DataLoader: A PyTorch dataloader.\n \"\"\"\n rank, world_size = get_dist_info()\n if dist:\n sampler = DistributedSampler(\n dataset, world_size, rank, shuffle=shuffle, round_up=round_up)\n shuffle = False\n batch_size = samples_per_gpu\n num_workers = workers_per_gpu\n else:\n sampler = None\n batch_size = num_gpus * samples_per_gpu\n num_workers = num_gpus * workers_per_gpu\n\n init_fn = partial(\n worker_init_fn, num_workers=num_workers, rank=rank,\n seed=seed) if seed is not None else None\n\n if LooseVersion(torch.__version__) >= LooseVersion('1.7.0'):\n kwargs['persistent_workers'] = persistent_workers\n\n data_loader = DataLoader(\n dataset,\n batch_size=batch_size,\n sampler=sampler,\n num_workers=num_workers,\n collate_fn=partial(collate, samples_per_gpu=samples_per_gpu),\n pin_memory=pin_memory,\n shuffle=shuffle,\n worker_init_fn=init_fn,\n **kwargs)\n\n return data_loader\n\n\ndef worker_init_fn(worker_id, num_workers, rank, seed):\n # The seed of each worker equals to\n # num_worker * rank + worker_id + user_seed\n worker_seed = num_workers * rank + worker_id + seed\n np.random.seed(worker_seed)\n random.seed(worker_seed)\n", "path": "mmcls/datasets/builder.py"}]}
| 2,919 | 244 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.