problem_id
stringlengths
18
22
source
stringclasses
1 value
task_type
stringclasses
1 value
in_source_id
stringlengths
13
58
prompt
stringlengths
1.71k
18.9k
golden_diff
stringlengths
145
5.13k
verification_info
stringlengths
465
23.6k
num_tokens_prompt
int64
556
4.1k
num_tokens_diff
int64
47
1.02k
gh_patches_debug_23360
rasdani/github-patches
git_diff
allegro__ralph-3159
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> Doc fixes Some minor doc fixes with a bit of style change </issue> <code> [start of src/ralph/dashboards/management/commands/push_graphs_to_statsd.py] 1 # -*- coding: utf-8 -*- 2 import logging 3 import textwrap 4 5 from django.conf import settings 6 from django.core.management.base import BaseCommand 7 from django.utils.text import slugify 8 9 from ralph.dashboards.models import Graph 10 from ralph.lib.metrics import build_statsd_client 11 12 logger = logging.getLogger(__name__) 13 PREFIX = settings.STATSD_GRAPHS_PREFIX 14 STATSD_PATH = '{}.{{}}.{{}}'.format(PREFIX) 15 16 17 def normalize(s): 18 s = slugify(s) 19 return s.replace('-', '_') 20 21 22 class Command(BaseCommand): 23 """Push to statsd data generated by graphs.""" 24 help = textwrap.dedent(__doc__).strip() 25 26 def handle(self, *args, **kwargs): 27 statsd = build_statsd_client(prefix=STATSD_PATH) 28 graphs = Graph.objects.filter(push_to_statsd=True) 29 for graph in graphs: 30 graph_data = graph.get_data() 31 graph_name = normalize(graph.name) 32 for label, value in zip(graph_data['labels'], graph_data['series']): 33 path = STATSD_PATH.format(graph_name, normalize(label)) 34 statsd.gauge(path, value) 35 [end of src/ralph/dashboards/management/commands/push_graphs_to_statsd.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/src/ralph/dashboards/management/commands/push_graphs_to_statsd.py b/src/ralph/dashboards/management/commands/push_graphs_to_statsd.py --- a/src/ralph/dashboards/management/commands/push_graphs_to_statsd.py +++ b/src/ralph/dashboards/management/commands/push_graphs_to_statsd.py @@ -10,8 +10,6 @@ from ralph.lib.metrics import build_statsd_client logger = logging.getLogger(__name__) -PREFIX = settings.STATSD_GRAPHS_PREFIX -STATSD_PATH = '{}.{{}}.{{}}'.format(PREFIX) def normalize(s): @@ -24,11 +22,11 @@ help = textwrap.dedent(__doc__).strip() def handle(self, *args, **kwargs): - statsd = build_statsd_client(prefix=STATSD_PATH) + statsd = build_statsd_client(prefix=settings.STATSD_GRAPHS_PREFIX) graphs = Graph.objects.filter(push_to_statsd=True) for graph in graphs: graph_data = graph.get_data() graph_name = normalize(graph.name) for label, value in zip(graph_data['labels'], graph_data['series']): - path = STATSD_PATH.format(graph_name, normalize(label)) + path = '.'.join((graph_name, normalize(label))) statsd.gauge(path, value)
{"golden_diff": "diff --git a/src/ralph/dashboards/management/commands/push_graphs_to_statsd.py b/src/ralph/dashboards/management/commands/push_graphs_to_statsd.py\n--- a/src/ralph/dashboards/management/commands/push_graphs_to_statsd.py\n+++ b/src/ralph/dashboards/management/commands/push_graphs_to_statsd.py\n@@ -10,8 +10,6 @@\n from ralph.lib.metrics import build_statsd_client\n \n logger = logging.getLogger(__name__)\n-PREFIX = settings.STATSD_GRAPHS_PREFIX\n-STATSD_PATH = '{}.{{}}.{{}}'.format(PREFIX)\n \n \n def normalize(s):\n@@ -24,11 +22,11 @@\n help = textwrap.dedent(__doc__).strip()\n \n def handle(self, *args, **kwargs):\n- statsd = build_statsd_client(prefix=STATSD_PATH)\n+ statsd = build_statsd_client(prefix=settings.STATSD_GRAPHS_PREFIX)\n graphs = Graph.objects.filter(push_to_statsd=True)\n for graph in graphs:\n graph_data = graph.get_data()\n graph_name = normalize(graph.name)\n for label, value in zip(graph_data['labels'], graph_data['series']):\n- path = STATSD_PATH.format(graph_name, normalize(label))\n+ path = '.'.join((graph_name, normalize(label)))\n statsd.gauge(path, value)\n", "issue": "Doc fixes\nSome minor doc fixes with a bit of style change\n\n", "before_files": [{"content": "# -*- coding: utf-8 -*-\nimport logging\nimport textwrap\n\nfrom django.conf import settings\nfrom django.core.management.base import BaseCommand\nfrom django.utils.text import slugify\n\nfrom ralph.dashboards.models import Graph\nfrom ralph.lib.metrics import build_statsd_client\n\nlogger = logging.getLogger(__name__)\nPREFIX = settings.STATSD_GRAPHS_PREFIX\nSTATSD_PATH = '{}.{{}}.{{}}'.format(PREFIX)\n\n\ndef normalize(s):\n s = slugify(s)\n return s.replace('-', '_')\n\n\nclass Command(BaseCommand):\n \"\"\"Push to statsd data generated by graphs.\"\"\"\n help = textwrap.dedent(__doc__).strip()\n\n def handle(self, *args, **kwargs):\n statsd = build_statsd_client(prefix=STATSD_PATH)\n graphs = Graph.objects.filter(push_to_statsd=True)\n for graph in graphs:\n graph_data = graph.get_data()\n graph_name = normalize(graph.name)\n for label, value in zip(graph_data['labels'], graph_data['series']):\n path = STATSD_PATH.format(graph_name, normalize(label))\n statsd.gauge(path, value)\n", "path": "src/ralph/dashboards/management/commands/push_graphs_to_statsd.py"}]}
884
313
gh_patches_debug_14848
rasdani/github-patches
git_diff
MongoEngine__mongoengine-1858
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> select_related() ConnectionError Hi, I am having an interesting situation. We have been using MongoEngine in our project for a good few years now and this issue appeared out of nowhere (no update, no change to the code). Calling `select_related()` on the results of a query raises a `mongoengine.connection.ConnectionError: You have not defined a default connection` Exception The call that causes the issue: ```credentials = Credential.objects(id__in=ids).select_related()``` I noticed that I am getting the exception only the first time time I call `select_related()` and only on this object. Any subsequent calls (on the same collection) work as expected. While trying to debug, I got to this: ``` > /usr/local/lib/python2.7/dist-packages/mongoengine/dereference.py(129)_fetch_objects() 127 object_map = {} 128 for collection, dbrefs in self.reference_map.iteritems(): --> 129 if hasattr(collection, 'objects'): # We have a document class for the refs 130 col_name = collection._get_collection_name() 131 refs = [dbref for dbref in dbrefs ipdb> collection <class 'insights.models.company.Company'> ipdb> hasattr(collection, 'objects') False ipdb> hasattr(collection, 'objects') True ``` Can this be caused by some lazy loading or some configuration issue? Thank you! </issue> <code> [start of mongoengine/dereference.py] 1 from bson import DBRef, SON 2 import six 3 4 from mongoengine.base import (BaseDict, BaseList, EmbeddedDocumentList, 5 TopLevelDocumentMetaclass, get_document) 6 from mongoengine.base.datastructures import LazyReference 7 from mongoengine.connection import get_db 8 from mongoengine.document import Document, EmbeddedDocument 9 from mongoengine.fields import DictField, ListField, MapField, ReferenceField 10 from mongoengine.queryset import QuerySet 11 12 13 class DeReference(object): 14 def __call__(self, items, max_depth=1, instance=None, name=None): 15 """ 16 Cheaply dereferences the items to a set depth. 17 Also handles the conversion of complex data types. 18 19 :param items: The iterable (dict, list, queryset) to be dereferenced. 20 :param max_depth: The maximum depth to recurse to 21 :param instance: The owning instance used for tracking changes by 22 :class:`~mongoengine.base.ComplexBaseField` 23 :param name: The name of the field, used for tracking changes by 24 :class:`~mongoengine.base.ComplexBaseField` 25 :param get: A boolean determining if being called by __get__ 26 """ 27 if items is None or isinstance(items, six.string_types): 28 return items 29 30 # cheapest way to convert a queryset to a list 31 # list(queryset) uses a count() query to determine length 32 if isinstance(items, QuerySet): 33 items = [i for i in items] 34 35 self.max_depth = max_depth 36 doc_type = None 37 38 if instance and isinstance(instance, (Document, EmbeddedDocument, 39 TopLevelDocumentMetaclass)): 40 doc_type = instance._fields.get(name) 41 while hasattr(doc_type, 'field'): 42 doc_type = doc_type.field 43 44 if isinstance(doc_type, ReferenceField): 45 field = doc_type 46 doc_type = doc_type.document_type 47 is_list = not hasattr(items, 'items') 48 49 if is_list and all([i.__class__ == doc_type for i in items]): 50 return items 51 elif not is_list and all( 52 [i.__class__ == doc_type for i in items.values()]): 53 return items 54 elif not field.dbref: 55 if not hasattr(items, 'items'): 56 57 def _get_items(items): 58 new_items = [] 59 for v in items: 60 if isinstance(v, list): 61 new_items.append(_get_items(v)) 62 elif not isinstance(v, (DBRef, Document)): 63 new_items.append(field.to_python(v)) 64 else: 65 new_items.append(v) 66 return new_items 67 68 items = _get_items(items) 69 else: 70 items = { 71 k: (v if isinstance(v, (DBRef, Document)) 72 else field.to_python(v)) 73 for k, v in items.iteritems() 74 } 75 76 self.reference_map = self._find_references(items) 77 self.object_map = self._fetch_objects(doc_type=doc_type) 78 return self._attach_objects(items, 0, instance, name) 79 80 def _find_references(self, items, depth=0): 81 """ 82 Recursively finds all db references to be dereferenced 83 84 :param items: The iterable (dict, list, queryset) 85 :param depth: The current depth of recursion 86 """ 87 reference_map = {} 88 if not items or depth >= self.max_depth: 89 return reference_map 90 91 # Determine the iterator to use 92 if isinstance(items, dict): 93 iterator = items.values() 94 else: 95 iterator = items 96 97 # Recursively find dbreferences 98 depth += 1 99 for item in iterator: 100 if isinstance(item, (Document, EmbeddedDocument)): 101 for field_name, field in item._fields.iteritems(): 102 v = item._data.get(field_name, None) 103 if isinstance(v, LazyReference): 104 # LazyReference inherits DBRef but should not be dereferenced here ! 105 continue 106 elif isinstance(v, DBRef): 107 reference_map.setdefault(field.document_type, set()).add(v.id) 108 elif isinstance(v, (dict, SON)) and '_ref' in v: 109 reference_map.setdefault(get_document(v['_cls']), set()).add(v['_ref'].id) 110 elif isinstance(v, (dict, list, tuple)) and depth <= self.max_depth: 111 field_cls = getattr(getattr(field, 'field', None), 'document_type', None) 112 references = self._find_references(v, depth) 113 for key, refs in references.iteritems(): 114 if isinstance(field_cls, (Document, TopLevelDocumentMetaclass)): 115 key = field_cls 116 reference_map.setdefault(key, set()).update(refs) 117 elif isinstance(item, LazyReference): 118 # LazyReference inherits DBRef but should not be dereferenced here ! 119 continue 120 elif isinstance(item, DBRef): 121 reference_map.setdefault(item.collection, set()).add(item.id) 122 elif isinstance(item, (dict, SON)) and '_ref' in item: 123 reference_map.setdefault(get_document(item['_cls']), set()).add(item['_ref'].id) 124 elif isinstance(item, (dict, list, tuple)) and depth - 1 <= self.max_depth: 125 references = self._find_references(item, depth - 1) 126 for key, refs in references.iteritems(): 127 reference_map.setdefault(key, set()).update(refs) 128 129 return reference_map 130 131 def _fetch_objects(self, doc_type=None): 132 """Fetch all references and convert to their document objects 133 """ 134 object_map = {} 135 for collection, dbrefs in self.reference_map.iteritems(): 136 if hasattr(collection, 'objects'): # We have a document class for the refs 137 col_name = collection._get_collection_name() 138 refs = [dbref for dbref in dbrefs 139 if (col_name, dbref) not in object_map] 140 references = collection.objects.in_bulk(refs) 141 for key, doc in references.iteritems(): 142 object_map[(col_name, key)] = doc 143 else: # Generic reference: use the refs data to convert to document 144 if isinstance(doc_type, (ListField, DictField, MapField,)): 145 continue 146 147 refs = [dbref for dbref in dbrefs 148 if (collection, dbref) not in object_map] 149 150 if doc_type: 151 references = doc_type._get_db()[collection].find({'_id': {'$in': refs}}) 152 for ref in references: 153 doc = doc_type._from_son(ref) 154 object_map[(collection, doc.id)] = doc 155 else: 156 references = get_db()[collection].find({'_id': {'$in': refs}}) 157 for ref in references: 158 if '_cls' in ref: 159 doc = get_document(ref['_cls'])._from_son(ref) 160 elif doc_type is None: 161 doc = get_document( 162 ''.join(x.capitalize() 163 for x in collection.split('_')))._from_son(ref) 164 else: 165 doc = doc_type._from_son(ref) 166 object_map[(collection, doc.id)] = doc 167 return object_map 168 169 def _attach_objects(self, items, depth=0, instance=None, name=None): 170 """ 171 Recursively finds all db references to be dereferenced 172 173 :param items: The iterable (dict, list, queryset) 174 :param depth: The current depth of recursion 175 :param instance: The owning instance used for tracking changes by 176 :class:`~mongoengine.base.ComplexBaseField` 177 :param name: The name of the field, used for tracking changes by 178 :class:`~mongoengine.base.ComplexBaseField` 179 """ 180 if not items: 181 if isinstance(items, (BaseDict, BaseList)): 182 return items 183 184 if instance: 185 if isinstance(items, dict): 186 return BaseDict(items, instance, name) 187 else: 188 return BaseList(items, instance, name) 189 190 if isinstance(items, (dict, SON)): 191 if '_ref' in items: 192 return self.object_map.get( 193 (items['_ref'].collection, items['_ref'].id), items) 194 elif '_cls' in items: 195 doc = get_document(items['_cls'])._from_son(items) 196 _cls = doc._data.pop('_cls', None) 197 del items['_cls'] 198 doc._data = self._attach_objects(doc._data, depth, doc, None) 199 if _cls is not None: 200 doc._data['_cls'] = _cls 201 return doc 202 203 if not hasattr(items, 'items'): 204 is_list = True 205 list_type = BaseList 206 if isinstance(items, EmbeddedDocumentList): 207 list_type = EmbeddedDocumentList 208 as_tuple = isinstance(items, tuple) 209 iterator = enumerate(items) 210 data = [] 211 else: 212 is_list = False 213 iterator = items.iteritems() 214 data = {} 215 216 depth += 1 217 for k, v in iterator: 218 if is_list: 219 data.append(v) 220 else: 221 data[k] = v 222 223 if k in self.object_map and not is_list: 224 data[k] = self.object_map[k] 225 elif isinstance(v, (Document, EmbeddedDocument)): 226 for field_name in v._fields: 227 v = data[k]._data.get(field_name, None) 228 if isinstance(v, DBRef): 229 data[k]._data[field_name] = self.object_map.get( 230 (v.collection, v.id), v) 231 elif isinstance(v, (dict, SON)) and '_ref' in v: 232 data[k]._data[field_name] = self.object_map.get( 233 (v['_ref'].collection, v['_ref'].id), v) 234 elif isinstance(v, (dict, list, tuple)) and depth <= self.max_depth: 235 item_name = six.text_type('{0}.{1}.{2}').format(name, k, field_name) 236 data[k]._data[field_name] = self._attach_objects(v, depth, instance=instance, name=item_name) 237 elif isinstance(v, (dict, list, tuple)) and depth <= self.max_depth: 238 item_name = '%s.%s' % (name, k) if name else name 239 data[k] = self._attach_objects(v, depth - 1, instance=instance, name=item_name) 240 elif isinstance(v, DBRef) and hasattr(v, 'id'): 241 data[k] = self.object_map.get((v.collection, v.id), v) 242 243 if instance and name: 244 if is_list: 245 return tuple(data) if as_tuple else list_type(data, instance, name) 246 return BaseDict(data, instance, name) 247 depth += 1 248 return data 249 [end of mongoengine/dereference.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/mongoengine/dereference.py b/mongoengine/dereference.py --- a/mongoengine/dereference.py +++ b/mongoengine/dereference.py @@ -133,7 +133,12 @@ """ object_map = {} for collection, dbrefs in self.reference_map.iteritems(): - if hasattr(collection, 'objects'): # We have a document class for the refs + + # we use getattr instead of hasattr because as hasattr swallows any exception under python2 + # so it could hide nasty things without raising exceptions (cfr bug #1688)) + ref_document_cls_exists = (getattr(collection, 'objects', None) is not None) + + if ref_document_cls_exists: col_name = collection._get_collection_name() refs = [dbref for dbref in dbrefs if (col_name, dbref) not in object_map]
{"golden_diff": "diff --git a/mongoengine/dereference.py b/mongoengine/dereference.py\n--- a/mongoengine/dereference.py\n+++ b/mongoengine/dereference.py\n@@ -133,7 +133,12 @@\n \"\"\"\n object_map = {}\n for collection, dbrefs in self.reference_map.iteritems():\n- if hasattr(collection, 'objects'): # We have a document class for the refs\n+\n+ # we use getattr instead of hasattr because as hasattr swallows any exception under python2\n+ # so it could hide nasty things without raising exceptions (cfr bug #1688))\n+ ref_document_cls_exists = (getattr(collection, 'objects', None) is not None)\n+\n+ if ref_document_cls_exists:\n col_name = collection._get_collection_name()\n refs = [dbref for dbref in dbrefs\n if (col_name, dbref) not in object_map]\n", "issue": "select_related() ConnectionError \nHi,\r\n\r\nI am having an interesting situation. We have been using MongoEngine in our project for a good few years now and this issue appeared out of nowhere (no update, no change to the code).\r\n\r\nCalling `select_related()` on the results of a query raises a `mongoengine.connection.ConnectionError: You have not defined a default connection` Exception\r\n\r\nThe call that causes the issue:\r\n```credentials = Credential.objects(id__in=ids).select_related()```\r\n\r\nI noticed that I am getting the exception only the first time time I call `select_related()` and only on this object. Any subsequent calls (on the same collection) work as expected. \r\n\r\nWhile trying to debug, I got to this:\r\n```\r\n> /usr/local/lib/python2.7/dist-packages/mongoengine/dereference.py(129)_fetch_objects()\r\n 127 object_map = {}\r\n 128 for collection, dbrefs in self.reference_map.iteritems():\r\n--> 129 if hasattr(collection, 'objects'): # We have a document class for the refs\r\n 130 col_name = collection._get_collection_name()\r\n 131 refs = [dbref for dbref in dbrefs\r\n\r\nipdb> collection\r\n<class 'insights.models.company.Company'>\r\nipdb> hasattr(collection, 'objects')\r\nFalse\r\nipdb> hasattr(collection, 'objects')\r\nTrue\r\n```\r\n\r\nCan this be caused by some lazy loading or some configuration issue?\r\n\r\nThank you!\n", "before_files": [{"content": "from bson import DBRef, SON\nimport six\n\nfrom mongoengine.base import (BaseDict, BaseList, EmbeddedDocumentList,\n TopLevelDocumentMetaclass, get_document)\nfrom mongoengine.base.datastructures import LazyReference\nfrom mongoengine.connection import get_db\nfrom mongoengine.document import Document, EmbeddedDocument\nfrom mongoengine.fields import DictField, ListField, MapField, ReferenceField\nfrom mongoengine.queryset import QuerySet\n\n\nclass DeReference(object):\n def __call__(self, items, max_depth=1, instance=None, name=None):\n \"\"\"\n Cheaply dereferences the items to a set depth.\n Also handles the conversion of complex data types.\n\n :param items: The iterable (dict, list, queryset) to be dereferenced.\n :param max_depth: The maximum depth to recurse to\n :param instance: The owning instance used for tracking changes by\n :class:`~mongoengine.base.ComplexBaseField`\n :param name: The name of the field, used for tracking changes by\n :class:`~mongoengine.base.ComplexBaseField`\n :param get: A boolean determining if being called by __get__\n \"\"\"\n if items is None or isinstance(items, six.string_types):\n return items\n\n # cheapest way to convert a queryset to a list\n # list(queryset) uses a count() query to determine length\n if isinstance(items, QuerySet):\n items = [i for i in items]\n\n self.max_depth = max_depth\n doc_type = None\n\n if instance and isinstance(instance, (Document, EmbeddedDocument,\n TopLevelDocumentMetaclass)):\n doc_type = instance._fields.get(name)\n while hasattr(doc_type, 'field'):\n doc_type = doc_type.field\n\n if isinstance(doc_type, ReferenceField):\n field = doc_type\n doc_type = doc_type.document_type\n is_list = not hasattr(items, 'items')\n\n if is_list and all([i.__class__ == doc_type for i in items]):\n return items\n elif not is_list and all(\n [i.__class__ == doc_type for i in items.values()]):\n return items\n elif not field.dbref:\n if not hasattr(items, 'items'):\n\n def _get_items(items):\n new_items = []\n for v in items:\n if isinstance(v, list):\n new_items.append(_get_items(v))\n elif not isinstance(v, (DBRef, Document)):\n new_items.append(field.to_python(v))\n else:\n new_items.append(v)\n return new_items\n\n items = _get_items(items)\n else:\n items = {\n k: (v if isinstance(v, (DBRef, Document))\n else field.to_python(v))\n for k, v in items.iteritems()\n }\n\n self.reference_map = self._find_references(items)\n self.object_map = self._fetch_objects(doc_type=doc_type)\n return self._attach_objects(items, 0, instance, name)\n\n def _find_references(self, items, depth=0):\n \"\"\"\n Recursively finds all db references to be dereferenced\n\n :param items: The iterable (dict, list, queryset)\n :param depth: The current depth of recursion\n \"\"\"\n reference_map = {}\n if not items or depth >= self.max_depth:\n return reference_map\n\n # Determine the iterator to use\n if isinstance(items, dict):\n iterator = items.values()\n else:\n iterator = items\n\n # Recursively find dbreferences\n depth += 1\n for item in iterator:\n if isinstance(item, (Document, EmbeddedDocument)):\n for field_name, field in item._fields.iteritems():\n v = item._data.get(field_name, None)\n if isinstance(v, LazyReference):\n # LazyReference inherits DBRef but should not be dereferenced here !\n continue\n elif isinstance(v, DBRef):\n reference_map.setdefault(field.document_type, set()).add(v.id)\n elif isinstance(v, (dict, SON)) and '_ref' in v:\n reference_map.setdefault(get_document(v['_cls']), set()).add(v['_ref'].id)\n elif isinstance(v, (dict, list, tuple)) and depth <= self.max_depth:\n field_cls = getattr(getattr(field, 'field', None), 'document_type', None)\n references = self._find_references(v, depth)\n for key, refs in references.iteritems():\n if isinstance(field_cls, (Document, TopLevelDocumentMetaclass)):\n key = field_cls\n reference_map.setdefault(key, set()).update(refs)\n elif isinstance(item, LazyReference):\n # LazyReference inherits DBRef but should not be dereferenced here !\n continue\n elif isinstance(item, DBRef):\n reference_map.setdefault(item.collection, set()).add(item.id)\n elif isinstance(item, (dict, SON)) and '_ref' in item:\n reference_map.setdefault(get_document(item['_cls']), set()).add(item['_ref'].id)\n elif isinstance(item, (dict, list, tuple)) and depth - 1 <= self.max_depth:\n references = self._find_references(item, depth - 1)\n for key, refs in references.iteritems():\n reference_map.setdefault(key, set()).update(refs)\n\n return reference_map\n\n def _fetch_objects(self, doc_type=None):\n \"\"\"Fetch all references and convert to their document objects\n \"\"\"\n object_map = {}\n for collection, dbrefs in self.reference_map.iteritems():\n if hasattr(collection, 'objects'): # We have a document class for the refs\n col_name = collection._get_collection_name()\n refs = [dbref for dbref in dbrefs\n if (col_name, dbref) not in object_map]\n references = collection.objects.in_bulk(refs)\n for key, doc in references.iteritems():\n object_map[(col_name, key)] = doc\n else: # Generic reference: use the refs data to convert to document\n if isinstance(doc_type, (ListField, DictField, MapField,)):\n continue\n\n refs = [dbref for dbref in dbrefs\n if (collection, dbref) not in object_map]\n\n if doc_type:\n references = doc_type._get_db()[collection].find({'_id': {'$in': refs}})\n for ref in references:\n doc = doc_type._from_son(ref)\n object_map[(collection, doc.id)] = doc\n else:\n references = get_db()[collection].find({'_id': {'$in': refs}})\n for ref in references:\n if '_cls' in ref:\n doc = get_document(ref['_cls'])._from_son(ref)\n elif doc_type is None:\n doc = get_document(\n ''.join(x.capitalize()\n for x in collection.split('_')))._from_son(ref)\n else:\n doc = doc_type._from_son(ref)\n object_map[(collection, doc.id)] = doc\n return object_map\n\n def _attach_objects(self, items, depth=0, instance=None, name=None):\n \"\"\"\n Recursively finds all db references to be dereferenced\n\n :param items: The iterable (dict, list, queryset)\n :param depth: The current depth of recursion\n :param instance: The owning instance used for tracking changes by\n :class:`~mongoengine.base.ComplexBaseField`\n :param name: The name of the field, used for tracking changes by\n :class:`~mongoengine.base.ComplexBaseField`\n \"\"\"\n if not items:\n if isinstance(items, (BaseDict, BaseList)):\n return items\n\n if instance:\n if isinstance(items, dict):\n return BaseDict(items, instance, name)\n else:\n return BaseList(items, instance, name)\n\n if isinstance(items, (dict, SON)):\n if '_ref' in items:\n return self.object_map.get(\n (items['_ref'].collection, items['_ref'].id), items)\n elif '_cls' in items:\n doc = get_document(items['_cls'])._from_son(items)\n _cls = doc._data.pop('_cls', None)\n del items['_cls']\n doc._data = self._attach_objects(doc._data, depth, doc, None)\n if _cls is not None:\n doc._data['_cls'] = _cls\n return doc\n\n if not hasattr(items, 'items'):\n is_list = True\n list_type = BaseList\n if isinstance(items, EmbeddedDocumentList):\n list_type = EmbeddedDocumentList\n as_tuple = isinstance(items, tuple)\n iterator = enumerate(items)\n data = []\n else:\n is_list = False\n iterator = items.iteritems()\n data = {}\n\n depth += 1\n for k, v in iterator:\n if is_list:\n data.append(v)\n else:\n data[k] = v\n\n if k in self.object_map and not is_list:\n data[k] = self.object_map[k]\n elif isinstance(v, (Document, EmbeddedDocument)):\n for field_name in v._fields:\n v = data[k]._data.get(field_name, None)\n if isinstance(v, DBRef):\n data[k]._data[field_name] = self.object_map.get(\n (v.collection, v.id), v)\n elif isinstance(v, (dict, SON)) and '_ref' in v:\n data[k]._data[field_name] = self.object_map.get(\n (v['_ref'].collection, v['_ref'].id), v)\n elif isinstance(v, (dict, list, tuple)) and depth <= self.max_depth:\n item_name = six.text_type('{0}.{1}.{2}').format(name, k, field_name)\n data[k]._data[field_name] = self._attach_objects(v, depth, instance=instance, name=item_name)\n elif isinstance(v, (dict, list, tuple)) and depth <= self.max_depth:\n item_name = '%s.%s' % (name, k) if name else name\n data[k] = self._attach_objects(v, depth - 1, instance=instance, name=item_name)\n elif isinstance(v, DBRef) and hasattr(v, 'id'):\n data[k] = self.object_map.get((v.collection, v.id), v)\n\n if instance and name:\n if is_list:\n return tuple(data) if as_tuple else list_type(data, instance, name)\n return BaseDict(data, instance, name)\n depth += 1\n return data\n", "path": "mongoengine/dereference.py"}]}
3,746
203
gh_patches_debug_23048
rasdani/github-patches
git_diff
cupy__cupy-5759
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> `cupy.concatenate()` misses arguments `dtype` and `casting` Refs: - NumPy: https://numpy.org/doc/stable/reference/generated/numpy.concatenate.html - CuPy: https://docs.cupy.dev/en/stable/reference/generated/cupy.concatenate.html The `dtype` argument is needed by the Array API standard (#5698, #4789). </issue> <code> [start of cupy/_manipulation/join.py] 1 import cupy 2 from cupy import _core 3 4 5 def column_stack(tup): 6 """Stacks 1-D and 2-D arrays as columns into a 2-D array. 7 8 A 1-D array is first converted to a 2-D column array. Then, the 2-D arrays 9 are concatenated along the second axis. 10 11 Args: 12 tup (sequence of arrays): 1-D or 2-D arrays to be stacked. 13 14 Returns: 15 cupy.ndarray: A new 2-D array of stacked columns. 16 17 .. seealso:: :func:`numpy.column_stack` 18 19 """ 20 if any(not isinstance(a, cupy.ndarray) for a in tup): 21 raise TypeError('Only cupy arrays can be column stacked') 22 23 lst = list(tup) 24 for i, a in enumerate(lst): 25 if a.ndim == 1: 26 a = a[:, cupy.newaxis] 27 lst[i] = a 28 elif a.ndim != 2: 29 raise ValueError( 30 'Only 1 or 2 dimensional arrays can be column stacked') 31 32 return concatenate(lst, axis=1) 33 34 35 def concatenate(tup, axis=0, out=None): 36 """Joins arrays along an axis. 37 38 Args: 39 tup (sequence of arrays): Arrays to be joined. All of these should have 40 same dimensionalities except the specified axis. 41 axis (int or None): The axis to join arrays along. 42 If axis is None, arrays are flattened before use. 43 Default is 0. 44 out (cupy.ndarray): Output array. 45 46 Returns: 47 cupy.ndarray: Joined array. 48 49 .. seealso:: :func:`numpy.concatenate` 50 51 """ 52 if axis is None: 53 tup = [m.ravel() for m in tup] 54 axis = 0 55 return _core.concatenate_method(tup, axis, out) 56 57 58 def dstack(tup): 59 """Stacks arrays along the third axis. 60 61 Args: 62 tup (sequence of arrays): Arrays to be stacked. Each array is converted 63 by :func:`cupy.atleast_3d` before stacking. 64 65 Returns: 66 cupy.ndarray: Stacked array. 67 68 .. seealso:: :func:`numpy.dstack` 69 70 """ 71 return concatenate([cupy.atleast_3d(m) for m in tup], 2) 72 73 74 def hstack(tup): 75 """Stacks arrays horizontally. 76 77 If an input array has one dimension, then the array is treated as a 78 horizontal vector and stacked along the first axis. Otherwise, the array is 79 stacked along the second axis. 80 81 Args: 82 tup (sequence of arrays): Arrays to be stacked. 83 84 Returns: 85 cupy.ndarray: Stacked array. 86 87 .. seealso:: :func:`numpy.hstack` 88 89 """ 90 arrs = [cupy.atleast_1d(a) for a in tup] 91 axis = 1 92 if arrs[0].ndim == 1: 93 axis = 0 94 return concatenate(arrs, axis) 95 96 97 def vstack(tup): 98 """Stacks arrays vertically. 99 100 If an input array has one dimension, then the array is treated as a 101 horizontal vector and stacked along the additional axis at the head. 102 Otherwise, the array is stacked along the first axis. 103 104 Args: 105 tup (sequence of arrays): Arrays to be stacked. Each array is converted 106 by :func:`cupy.atleast_2d` before stacking. 107 108 Returns: 109 cupy.ndarray: Stacked array. 110 111 .. seealso:: :func:`numpy.dstack` 112 113 """ 114 return concatenate([cupy.atleast_2d(m) for m in tup], 0) 115 116 117 def stack(tup, axis=0, out=None): 118 """Stacks arrays along a new axis. 119 120 Args: 121 tup (sequence of arrays): Arrays to be stacked. 122 axis (int): Axis along which the arrays are stacked. 123 out (cupy.ndarray): Output array. 124 125 Returns: 126 cupy.ndarray: Stacked array. 127 128 .. seealso:: :func:`numpy.stack` 129 """ 130 return concatenate([cupy.expand_dims(x, axis) for x in tup], axis, out) 131 [end of cupy/_manipulation/join.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/cupy/_manipulation/join.py b/cupy/_manipulation/join.py --- a/cupy/_manipulation/join.py +++ b/cupy/_manipulation/join.py @@ -32,7 +32,7 @@ return concatenate(lst, axis=1) -def concatenate(tup, axis=0, out=None): +def concatenate(tup, axis=0, out=None, *, dtype=None, casting='same_kind'): """Joins arrays along an axis. Args: @@ -42,6 +42,11 @@ If axis is None, arrays are flattened before use. Default is 0. out (cupy.ndarray): Output array. + dtype (str or dtype): If provided, the destination array will have this + dtype. Cannot be provided together with ``out``. + casting ({‘no’, ‘equiv’, ‘safe’, ‘same_kind’, ‘unsafe’}, optional): + Controls what kind of data casting may occur. Defaults to + ``'same_kind'``. Returns: cupy.ndarray: Joined array. @@ -52,7 +57,7 @@ if axis is None: tup = [m.ravel() for m in tup] axis = 0 - return _core.concatenate_method(tup, axis, out) + return _core.concatenate_method(tup, axis, out, dtype, casting) def dstack(tup):
{"golden_diff": "diff --git a/cupy/_manipulation/join.py b/cupy/_manipulation/join.py\n--- a/cupy/_manipulation/join.py\n+++ b/cupy/_manipulation/join.py\n@@ -32,7 +32,7 @@\n return concatenate(lst, axis=1)\n \n \n-def concatenate(tup, axis=0, out=None):\n+def concatenate(tup, axis=0, out=None, *, dtype=None, casting='same_kind'):\n \"\"\"Joins arrays along an axis.\n \n Args:\n@@ -42,6 +42,11 @@\n If axis is None, arrays are flattened before use.\n Default is 0.\n out (cupy.ndarray): Output array.\n+ dtype (str or dtype): If provided, the destination array will have this\n+ dtype. Cannot be provided together with ``out``.\n+ casting ({\u2018no\u2019, \u2018equiv\u2019, \u2018safe\u2019, \u2018same_kind\u2019, \u2018unsafe\u2019}, optional):\n+ Controls what kind of data casting may occur. Defaults to\n+ ``'same_kind'``.\n \n Returns:\n cupy.ndarray: Joined array.\n@@ -52,7 +57,7 @@\n if axis is None:\n tup = [m.ravel() for m in tup]\n axis = 0\n- return _core.concatenate_method(tup, axis, out)\n+ return _core.concatenate_method(tup, axis, out, dtype, casting)\n \n \n def dstack(tup):\n", "issue": "`cupy.concatenate()` misses arguments `dtype` and `casting`\nRefs:\r\n- NumPy: https://numpy.org/doc/stable/reference/generated/numpy.concatenate.html\r\n- CuPy: https://docs.cupy.dev/en/stable/reference/generated/cupy.concatenate.html\r\n\r\nThe `dtype` argument is needed by the Array API standard (#5698, #4789).\n", "before_files": [{"content": "import cupy\nfrom cupy import _core\n\n\ndef column_stack(tup):\n \"\"\"Stacks 1-D and 2-D arrays as columns into a 2-D array.\n\n A 1-D array is first converted to a 2-D column array. Then, the 2-D arrays\n are concatenated along the second axis.\n\n Args:\n tup (sequence of arrays): 1-D or 2-D arrays to be stacked.\n\n Returns:\n cupy.ndarray: A new 2-D array of stacked columns.\n\n .. seealso:: :func:`numpy.column_stack`\n\n \"\"\"\n if any(not isinstance(a, cupy.ndarray) for a in tup):\n raise TypeError('Only cupy arrays can be column stacked')\n\n lst = list(tup)\n for i, a in enumerate(lst):\n if a.ndim == 1:\n a = a[:, cupy.newaxis]\n lst[i] = a\n elif a.ndim != 2:\n raise ValueError(\n 'Only 1 or 2 dimensional arrays can be column stacked')\n\n return concatenate(lst, axis=1)\n\n\ndef concatenate(tup, axis=0, out=None):\n \"\"\"Joins arrays along an axis.\n\n Args:\n tup (sequence of arrays): Arrays to be joined. All of these should have\n same dimensionalities except the specified axis.\n axis (int or None): The axis to join arrays along.\n If axis is None, arrays are flattened before use.\n Default is 0.\n out (cupy.ndarray): Output array.\n\n Returns:\n cupy.ndarray: Joined array.\n\n .. seealso:: :func:`numpy.concatenate`\n\n \"\"\"\n if axis is None:\n tup = [m.ravel() for m in tup]\n axis = 0\n return _core.concatenate_method(tup, axis, out)\n\n\ndef dstack(tup):\n \"\"\"Stacks arrays along the third axis.\n\n Args:\n tup (sequence of arrays): Arrays to be stacked. Each array is converted\n by :func:`cupy.atleast_3d` before stacking.\n\n Returns:\n cupy.ndarray: Stacked array.\n\n .. seealso:: :func:`numpy.dstack`\n\n \"\"\"\n return concatenate([cupy.atleast_3d(m) for m in tup], 2)\n\n\ndef hstack(tup):\n \"\"\"Stacks arrays horizontally.\n\n If an input array has one dimension, then the array is treated as a\n horizontal vector and stacked along the first axis. Otherwise, the array is\n stacked along the second axis.\n\n Args:\n tup (sequence of arrays): Arrays to be stacked.\n\n Returns:\n cupy.ndarray: Stacked array.\n\n .. seealso:: :func:`numpy.hstack`\n\n \"\"\"\n arrs = [cupy.atleast_1d(a) for a in tup]\n axis = 1\n if arrs[0].ndim == 1:\n axis = 0\n return concatenate(arrs, axis)\n\n\ndef vstack(tup):\n \"\"\"Stacks arrays vertically.\n\n If an input array has one dimension, then the array is treated as a\n horizontal vector and stacked along the additional axis at the head.\n Otherwise, the array is stacked along the first axis.\n\n Args:\n tup (sequence of arrays): Arrays to be stacked. Each array is converted\n by :func:`cupy.atleast_2d` before stacking.\n\n Returns:\n cupy.ndarray: Stacked array.\n\n .. seealso:: :func:`numpy.dstack`\n\n \"\"\"\n return concatenate([cupy.atleast_2d(m) for m in tup], 0)\n\n\ndef stack(tup, axis=0, out=None):\n \"\"\"Stacks arrays along a new axis.\n\n Args:\n tup (sequence of arrays): Arrays to be stacked.\n axis (int): Axis along which the arrays are stacked.\n out (cupy.ndarray): Output array.\n\n Returns:\n cupy.ndarray: Stacked array.\n\n .. seealso:: :func:`numpy.stack`\n \"\"\"\n return concatenate([cupy.expand_dims(x, axis) for x in tup], axis, out)\n", "path": "cupy/_manipulation/join.py"}]}
1,818
322
gh_patches_debug_25497
rasdani/github-patches
git_diff
tensorflow__addons-2048
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> Deprecate GELU Per https://github.com/tensorflow/community/pull/252, oonce https://github.com/tensorflow/tensorflow/pull/41178 merges we'll need to deprecate our GELU for versions of TensorFlow that include it within core. </issue> <code> [start of tensorflow_addons/layers/gelu.py] 1 # Copyright 2019 The TensorFlow Authors. All Rights Reserved. 2 # 3 # Licensed under the Apache License, Version 2.0 (the "License"); 4 # you may not use this file except in compliance with the License. 5 # You may obtain a copy of the License at 6 # 7 # http://www.apache.org/licenses/LICENSE-2.0 8 # 9 # Unless required by applicable law or agreed to in writing, software 10 # distributed under the License is distributed on an "AS IS" BASIS, 11 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 12 # See the License for the specific language governing permissions and 13 # limitations under the License. 14 # ============================================================================== 15 """Implements GELU activation.""" 16 17 import tensorflow as tf 18 from tensorflow_addons.activations import gelu 19 from typeguard import typechecked 20 21 22 @tf.keras.utils.register_keras_serializable(package="Addons") 23 class GELU(tf.keras.layers.Layer): 24 """Gaussian Error Linear Unit. 25 26 A smoother version of ReLU generally used 27 in the BERT or BERT architecture based models. 28 Original paper: https://arxiv.org/abs/1606.08415 29 30 Input shape: 31 Arbitrary. Use the keyword argument `input_shape` 32 (tuple of integers, does not include the samples axis) 33 when using this layer as the first layer in a model. 34 35 Output shape: 36 Same shape as the input. 37 """ 38 39 @typechecked 40 def __init__(self, approximate: bool = True, **kwargs): 41 super().__init__(**kwargs) 42 self.approximate = approximate 43 self.supports_masking = True 44 45 def call(self, inputs): 46 return gelu(inputs, approximate=self.approximate) 47 48 def get_config(self): 49 config = {"approximate": self.approximate} 50 base_config = super().get_config() 51 return {**base_config, **config} 52 53 def compute_output_shape(self, input_shape): 54 return input_shape 55 [end of tensorflow_addons/layers/gelu.py] [start of tensorflow_addons/activations/gelu.py] 1 # Copyright 2019 The TensorFlow Authors. All Rights Reserved. 2 # 3 # Licensed under the Apache License, Version 2.0 (the "License"); 4 # you may not use this file except in compliance with the License. 5 # You may obtain a copy of the License at 6 # 7 # http://www.apache.org/licenses/LICENSE-2.0 8 # 9 # Unless required by applicable law or agreed to in writing, software 10 # distributed under the License is distributed on an "AS IS" BASIS, 11 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 12 # See the License for the specific language governing permissions and 13 # limitations under the License. 14 # ============================================================================== 15 16 import tensorflow as tf 17 import math 18 import warnings 19 20 from tensorflow_addons.utils import types 21 from tensorflow_addons.utils.resource_loader import LazySO 22 from tensorflow_addons import options 23 24 _activation_so = LazySO("custom_ops/activations/_activation_ops.so") 25 26 27 @tf.keras.utils.register_keras_serializable(package="Addons") 28 def gelu(x: types.TensorLike, approximate: bool = True) -> tf.Tensor: 29 """Gaussian Error Linear Unit. 30 31 Computes gaussian error linear: 32 `0.5 * x * (1 + tanh(sqrt(2 / pi) * (x + 0.044715 * x^3)))` or 33 `x * P(X <= x) = 0.5 * x * (1 + erf(x / sqrt(2)))`, where P(X) ~ N(0, 1), 34 depending on whether approximation is enabled. 35 36 See [Gaussian Error Linear Units (GELUs)](https://arxiv.org/abs/1606.08415) 37 and [BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding](https://arxiv.org/abs/1810.04805). 38 39 Args: 40 x: A `Tensor`. Must be one of the following types: 41 `float16`, `float32`, `float64`. 42 approximate: bool, whether to enable approximation. 43 Returns: 44 A `Tensor`. Has the same type as `x`. 45 """ 46 x = tf.convert_to_tensor(x) 47 48 if not options.TF_ADDONS_PY_OPS: 49 try: 50 return _gelu_custom_op(x, approximate) 51 except tf.errors.NotFoundError: 52 options.warn_fallback("gelu") 53 54 return _gelu_py(x, approximate) 55 56 57 def _gelu_custom_op(x, approximate): 58 warnings.warn( 59 "The activations custom ops are deprecated and will be removed in TensorFlow Addons " 60 "v0.12.0. \nPlease use the pure python version of Gelu instead by using the " 61 "`TF_ADDONS_PY_OPS` flag. \nFor more info about this flag, see " 62 "https://github.com/tensorflow/addons#gpucpu-custom-ops ", 63 DeprecationWarning, 64 ) 65 return _activation_so.ops.addons_gelu(x, approximate) 66 67 68 @tf.RegisterGradient("Addons>Gelu") 69 def _gelu_grad(op, grad): 70 return _activation_so.ops.addons_gelu_grad( 71 grad, op.inputs[0], op.get_attr("approximate") 72 ) 73 74 75 def _gelu_py(x: types.TensorLike, approximate: bool = True) -> tf.Tensor: 76 x = tf.convert_to_tensor(x) 77 if approximate: 78 pi = tf.cast(math.pi, x.dtype) 79 coeff = tf.cast(0.044715, x.dtype) 80 return 0.5 * x * (1.0 + tf.tanh(tf.sqrt(2.0 / pi) * (x + coeff * tf.pow(x, 3)))) 81 else: 82 return 0.5 * x * (1.0 + tf.math.erf(x / tf.cast(tf.sqrt(2.0), x.dtype))) 83 [end of tensorflow_addons/activations/gelu.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/tensorflow_addons/activations/gelu.py b/tensorflow_addons/activations/gelu.py --- a/tensorflow_addons/activations/gelu.py +++ b/tensorflow_addons/activations/gelu.py @@ -43,6 +43,12 @@ Returns: A `Tensor`. Has the same type as `x`. """ + warnings.warn( + "gelu activation has been migrated to core TensorFlow, " + "and will be deprecated in Addons 0.12.", + DeprecationWarning, + ) + x = tf.convert_to_tensor(x) if not options.TF_ADDONS_PY_OPS: diff --git a/tensorflow_addons/layers/gelu.py b/tensorflow_addons/layers/gelu.py --- a/tensorflow_addons/layers/gelu.py +++ b/tensorflow_addons/layers/gelu.py @@ -14,6 +14,8 @@ # ============================================================================== """Implements GELU activation.""" +import warnings + import tensorflow as tf from tensorflow_addons.activations import gelu from typeguard import typechecked @@ -38,6 +40,12 @@ @typechecked def __init__(self, approximate: bool = True, **kwargs): + warnings.warn( + "gelu activation has been migrated to core TensorFlow, " + "and will be deprecated in Addons 0.12.", + DeprecationWarning, + ) + super().__init__(**kwargs) self.approximate = approximate self.supports_masking = True
{"golden_diff": "diff --git a/tensorflow_addons/activations/gelu.py b/tensorflow_addons/activations/gelu.py\n--- a/tensorflow_addons/activations/gelu.py\n+++ b/tensorflow_addons/activations/gelu.py\n@@ -43,6 +43,12 @@\n Returns:\n A `Tensor`. Has the same type as `x`.\n \"\"\"\n+ warnings.warn(\n+ \"gelu activation has been migrated to core TensorFlow, \"\n+ \"and will be deprecated in Addons 0.12.\",\n+ DeprecationWarning,\n+ )\n+\n x = tf.convert_to_tensor(x)\n \n if not options.TF_ADDONS_PY_OPS:\ndiff --git a/tensorflow_addons/layers/gelu.py b/tensorflow_addons/layers/gelu.py\n--- a/tensorflow_addons/layers/gelu.py\n+++ b/tensorflow_addons/layers/gelu.py\n@@ -14,6 +14,8 @@\n # ==============================================================================\n \"\"\"Implements GELU activation.\"\"\"\n \n+import warnings\n+\n import tensorflow as tf\n from tensorflow_addons.activations import gelu\n from typeguard import typechecked\n@@ -38,6 +40,12 @@\n \n @typechecked\n def __init__(self, approximate: bool = True, **kwargs):\n+ warnings.warn(\n+ \"gelu activation has been migrated to core TensorFlow, \"\n+ \"and will be deprecated in Addons 0.12.\",\n+ DeprecationWarning,\n+ )\n+\n super().__init__(**kwargs)\n self.approximate = approximate\n self.supports_masking = True\n", "issue": "Deprecate GELU \nPer https://github.com/tensorflow/community/pull/252, oonce https://github.com/tensorflow/tensorflow/pull/41178 merges we'll need to deprecate our GELU for versions of TensorFlow that include it within core.\n", "before_files": [{"content": "# Copyright 2019 The TensorFlow Authors. All Rights Reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n# ==============================================================================\n\"\"\"Implements GELU activation.\"\"\"\n\nimport tensorflow as tf\nfrom tensorflow_addons.activations import gelu\nfrom typeguard import typechecked\n\n\[email protected]_keras_serializable(package=\"Addons\")\nclass GELU(tf.keras.layers.Layer):\n \"\"\"Gaussian Error Linear Unit.\n\n A smoother version of ReLU generally used\n in the BERT or BERT architecture based models.\n Original paper: https://arxiv.org/abs/1606.08415\n\n Input shape:\n Arbitrary. Use the keyword argument `input_shape`\n (tuple of integers, does not include the samples axis)\n when using this layer as the first layer in a model.\n\n Output shape:\n Same shape as the input.\n \"\"\"\n\n @typechecked\n def __init__(self, approximate: bool = True, **kwargs):\n super().__init__(**kwargs)\n self.approximate = approximate\n self.supports_masking = True\n\n def call(self, inputs):\n return gelu(inputs, approximate=self.approximate)\n\n def get_config(self):\n config = {\"approximate\": self.approximate}\n base_config = super().get_config()\n return {**base_config, **config}\n\n def compute_output_shape(self, input_shape):\n return input_shape\n", "path": "tensorflow_addons/layers/gelu.py"}, {"content": "# Copyright 2019 The TensorFlow Authors. All Rights Reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n# ==============================================================================\n\nimport tensorflow as tf\nimport math\nimport warnings\n\nfrom tensorflow_addons.utils import types\nfrom tensorflow_addons.utils.resource_loader import LazySO\nfrom tensorflow_addons import options\n\n_activation_so = LazySO(\"custom_ops/activations/_activation_ops.so\")\n\n\[email protected]_keras_serializable(package=\"Addons\")\ndef gelu(x: types.TensorLike, approximate: bool = True) -> tf.Tensor:\n \"\"\"Gaussian Error Linear Unit.\n\n Computes gaussian error linear:\n `0.5 * x * (1 + tanh(sqrt(2 / pi) * (x + 0.044715 * x^3)))` or\n `x * P(X <= x) = 0.5 * x * (1 + erf(x / sqrt(2)))`, where P(X) ~ N(0, 1),\n depending on whether approximation is enabled.\n\n See [Gaussian Error Linear Units (GELUs)](https://arxiv.org/abs/1606.08415)\n and [BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding](https://arxiv.org/abs/1810.04805).\n\n Args:\n x: A `Tensor`. Must be one of the following types:\n `float16`, `float32`, `float64`.\n approximate: bool, whether to enable approximation.\n Returns:\n A `Tensor`. Has the same type as `x`.\n \"\"\"\n x = tf.convert_to_tensor(x)\n\n if not options.TF_ADDONS_PY_OPS:\n try:\n return _gelu_custom_op(x, approximate)\n except tf.errors.NotFoundError:\n options.warn_fallback(\"gelu\")\n\n return _gelu_py(x, approximate)\n\n\ndef _gelu_custom_op(x, approximate):\n warnings.warn(\n \"The activations custom ops are deprecated and will be removed in TensorFlow Addons \"\n \"v0.12.0. \\nPlease use the pure python version of Gelu instead by using the \"\n \"`TF_ADDONS_PY_OPS` flag. \\nFor more info about this flag, see \"\n \"https://github.com/tensorflow/addons#gpucpu-custom-ops \",\n DeprecationWarning,\n )\n return _activation_so.ops.addons_gelu(x, approximate)\n\n\[email protected](\"Addons>Gelu\")\ndef _gelu_grad(op, grad):\n return _activation_so.ops.addons_gelu_grad(\n grad, op.inputs[0], op.get_attr(\"approximate\")\n )\n\n\ndef _gelu_py(x: types.TensorLike, approximate: bool = True) -> tf.Tensor:\n x = tf.convert_to_tensor(x)\n if approximate:\n pi = tf.cast(math.pi, x.dtype)\n coeff = tf.cast(0.044715, x.dtype)\n return 0.5 * x * (1.0 + tf.tanh(tf.sqrt(2.0 / pi) * (x + coeff * tf.pow(x, 3))))\n else:\n return 0.5 * x * (1.0 + tf.math.erf(x / tf.cast(tf.sqrt(2.0), x.dtype)))\n", "path": "tensorflow_addons/activations/gelu.py"}]}
2,154
364
gh_patches_debug_13722
rasdani/github-patches
git_diff
open-telemetry__opentelemetry-python-contrib-935
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> Support for pymemcache 2.x and 3.x The `pymemcache` integration is currently locked to version 1.x. Current version is ~=3.5. https://github.com/open-telemetry/opentelemetry-python-contrib/blob/0b9e96dae0bfcafa1e0b28dc695e3697fbcf2864/instrumentation/opentelemetry-instrumentation-pymemcache/src/opentelemetry/instrumentation/pymemcache/package.py#L16 </issue> <code> [start of instrumentation/opentelemetry-instrumentation-pymemcache/src/opentelemetry/instrumentation/pymemcache/package.py] 1 # Copyright The OpenTelemetry Authors 2 # 3 # Licensed under the Apache License, Version 2.0 (the "License"); 4 # you may not use this file except in compliance with the License. 5 # You may obtain a copy of the License at 6 # 7 # http://www.apache.org/licenses/LICENSE-2.0 8 # 9 # Unless required by applicable law or agreed to in writing, software 10 # distributed under the License is distributed on an "AS IS" BASIS, 11 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 12 # See the License for the specific language governing permissions and 13 # limitations under the License. 14 15 16 _instruments = ("pymemcache ~= 1.3",) 17 [end of instrumentation/opentelemetry-instrumentation-pymemcache/src/opentelemetry/instrumentation/pymemcache/package.py] [start of opentelemetry-instrumentation/src/opentelemetry/instrumentation/bootstrap_gen.py] 1 # Copyright The OpenTelemetry Authors 2 # 3 # Licensed under the Apache License, Version 2.0 (the "License"); 4 # you may not use this file except in compliance with the License. 5 # You may obtain a copy of the License at 6 # 7 # http://www.apache.org/licenses/LICENSE-2.0 8 # 9 # Unless required by applicable law or agreed to in writing, software 10 # distributed under the License is distributed on an "AS IS" BASIS, 11 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 12 # See the License for the specific language governing permissions and 13 # limitations under the License. 14 15 # DO NOT EDIT. THIS FILE WAS AUTOGENERATED FROM INSTRUMENTATION PACKAGES. 16 # RUN `python scripts/generate_instrumentation_bootstrap.py` TO REGENERATE. 17 18 libraries = { 19 "aiohttp": { 20 "library": "aiohttp ~= 3.0", 21 "instrumentation": "opentelemetry-instrumentation-aiohttp-client==0.29b0", 22 }, 23 "aiopg": { 24 "library": "aiopg >= 0.13.0, < 1.3.0", 25 "instrumentation": "opentelemetry-instrumentation-aiopg==0.29b0", 26 }, 27 "asgiref": { 28 "library": "asgiref ~= 3.0", 29 "instrumentation": "opentelemetry-instrumentation-asgi==0.29b0", 30 }, 31 "asyncpg": { 32 "library": "asyncpg >= 0.12.0", 33 "instrumentation": "opentelemetry-instrumentation-asyncpg==0.29b0", 34 }, 35 "boto": { 36 "library": "boto~=2.0", 37 "instrumentation": "opentelemetry-instrumentation-boto==0.29b0", 38 }, 39 "botocore": { 40 "library": "botocore ~= 1.0", 41 "instrumentation": "opentelemetry-instrumentation-botocore==0.29b0", 42 }, 43 "celery": { 44 "library": "celery >= 4.0, < 6.0", 45 "instrumentation": "opentelemetry-instrumentation-celery==0.29b0", 46 }, 47 "django": { 48 "library": "django >= 1.10", 49 "instrumentation": "opentelemetry-instrumentation-django==0.29b0", 50 }, 51 "elasticsearch": { 52 "library": "elasticsearch >= 2.0", 53 "instrumentation": "opentelemetry-instrumentation-elasticsearch==0.29b0", 54 }, 55 "falcon": { 56 "library": "falcon >= 2.0.0, < 4.0.0", 57 "instrumentation": "opentelemetry-instrumentation-falcon==0.29b0", 58 }, 59 "fastapi": { 60 "library": "fastapi ~= 0.58", 61 "instrumentation": "opentelemetry-instrumentation-fastapi==0.29b0", 62 }, 63 "flask": { 64 "library": "flask >= 1.0, < 3.0", 65 "instrumentation": "opentelemetry-instrumentation-flask==0.29b0", 66 }, 67 "grpcio": { 68 "library": "grpcio ~= 1.27", 69 "instrumentation": "opentelemetry-instrumentation-grpc==0.29b0", 70 }, 71 "httpx": { 72 "library": "httpx >= 0.18.0", 73 "instrumentation": "opentelemetry-instrumentation-httpx==0.29b0", 74 }, 75 "jinja2": { 76 "library": "jinja2 >= 2.7, < 4.0", 77 "instrumentation": "opentelemetry-instrumentation-jinja2==0.29b0", 78 }, 79 "kafka-python": { 80 "library": "kafka-python >= 2.0", 81 "instrumentation": "opentelemetry-instrumentation-kafka-python==0.29b0", 82 }, 83 "mysql-connector-python": { 84 "library": "mysql-connector-python ~= 8.0", 85 "instrumentation": "opentelemetry-instrumentation-mysql==0.29b0", 86 }, 87 "pika": { 88 "library": "pika >= 0.12.0", 89 "instrumentation": "opentelemetry-instrumentation-pika==0.29b0", 90 }, 91 "psycopg2": { 92 "library": "psycopg2 >= 2.7.3.1", 93 "instrumentation": "opentelemetry-instrumentation-psycopg2==0.29b0", 94 }, 95 "pymemcache": { 96 "library": "pymemcache ~= 1.3", 97 "instrumentation": "opentelemetry-instrumentation-pymemcache==0.29b0", 98 }, 99 "pymongo": { 100 "library": "pymongo >= 3.1, < 5.0", 101 "instrumentation": "opentelemetry-instrumentation-pymongo==0.29b0", 102 }, 103 "PyMySQL": { 104 "library": "PyMySQL < 2", 105 "instrumentation": "opentelemetry-instrumentation-pymysql==0.29b0", 106 }, 107 "pyramid": { 108 "library": "pyramid >= 1.7", 109 "instrumentation": "opentelemetry-instrumentation-pyramid==0.29b0", 110 }, 111 "redis": { 112 "library": "redis >= 2.6", 113 "instrumentation": "opentelemetry-instrumentation-redis==0.29b0", 114 }, 115 "requests": { 116 "library": "requests ~= 2.0", 117 "instrumentation": "opentelemetry-instrumentation-requests==0.29b0", 118 }, 119 "scikit-learn": { 120 "library": "scikit-learn ~= 0.24.0", 121 "instrumentation": "opentelemetry-instrumentation-sklearn==0.29b0", 122 }, 123 "sqlalchemy": { 124 "library": "sqlalchemy", 125 "instrumentation": "opentelemetry-instrumentation-sqlalchemy==0.29b0", 126 }, 127 "starlette": { 128 "library": "starlette ~= 0.13.0", 129 "instrumentation": "opentelemetry-instrumentation-starlette==0.29b0", 130 }, 131 "tornado": { 132 "library": "tornado >= 5.1.1", 133 "instrumentation": "opentelemetry-instrumentation-tornado==0.29b0", 134 }, 135 "urllib3": { 136 "library": "urllib3 >= 1.0.0, < 2.0.0", 137 "instrumentation": "opentelemetry-instrumentation-urllib3==0.29b0", 138 }, 139 } 140 default_instrumentations = [ 141 "opentelemetry-instrumentation-aws-lambda==0.29b0", 142 "opentelemetry-instrumentation-dbapi==0.29b0", 143 "opentelemetry-instrumentation-logging==0.29b0", 144 "opentelemetry-instrumentation-sqlite3==0.29b0", 145 "opentelemetry-instrumentation-urllib==0.29b0", 146 "opentelemetry-instrumentation-wsgi==0.29b0", 147 ] 148 [end of opentelemetry-instrumentation/src/opentelemetry/instrumentation/bootstrap_gen.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/instrumentation/opentelemetry-instrumentation-pymemcache/src/opentelemetry/instrumentation/pymemcache/package.py b/instrumentation/opentelemetry-instrumentation-pymemcache/src/opentelemetry/instrumentation/pymemcache/package.py --- a/instrumentation/opentelemetry-instrumentation-pymemcache/src/opentelemetry/instrumentation/pymemcache/package.py +++ b/instrumentation/opentelemetry-instrumentation-pymemcache/src/opentelemetry/instrumentation/pymemcache/package.py @@ -13,4 +13,4 @@ # limitations under the License. -_instruments = ("pymemcache ~= 1.3",) +_instruments = ("pymemcache >= 1.3.5, < 4",) diff --git a/opentelemetry-instrumentation/src/opentelemetry/instrumentation/bootstrap_gen.py b/opentelemetry-instrumentation/src/opentelemetry/instrumentation/bootstrap_gen.py --- a/opentelemetry-instrumentation/src/opentelemetry/instrumentation/bootstrap_gen.py +++ b/opentelemetry-instrumentation/src/opentelemetry/instrumentation/bootstrap_gen.py @@ -93,7 +93,7 @@ "instrumentation": "opentelemetry-instrumentation-psycopg2==0.29b0", }, "pymemcache": { - "library": "pymemcache ~= 1.3", + "library": "pymemcache >= 1.3.5, < 4", "instrumentation": "opentelemetry-instrumentation-pymemcache==0.29b0", }, "pymongo": {
{"golden_diff": "diff --git a/instrumentation/opentelemetry-instrumentation-pymemcache/src/opentelemetry/instrumentation/pymemcache/package.py b/instrumentation/opentelemetry-instrumentation-pymemcache/src/opentelemetry/instrumentation/pymemcache/package.py\n--- a/instrumentation/opentelemetry-instrumentation-pymemcache/src/opentelemetry/instrumentation/pymemcache/package.py\n+++ b/instrumentation/opentelemetry-instrumentation-pymemcache/src/opentelemetry/instrumentation/pymemcache/package.py\n@@ -13,4 +13,4 @@\n # limitations under the License.\n \n \n-_instruments = (\"pymemcache ~= 1.3\",)\n+_instruments = (\"pymemcache >= 1.3.5, < 4\",)\ndiff --git a/opentelemetry-instrumentation/src/opentelemetry/instrumentation/bootstrap_gen.py b/opentelemetry-instrumentation/src/opentelemetry/instrumentation/bootstrap_gen.py\n--- a/opentelemetry-instrumentation/src/opentelemetry/instrumentation/bootstrap_gen.py\n+++ b/opentelemetry-instrumentation/src/opentelemetry/instrumentation/bootstrap_gen.py\n@@ -93,7 +93,7 @@\n \"instrumentation\": \"opentelemetry-instrumentation-psycopg2==0.29b0\",\n },\n \"pymemcache\": {\n- \"library\": \"pymemcache ~= 1.3\",\n+ \"library\": \"pymemcache >= 1.3.5, < 4\",\n \"instrumentation\": \"opentelemetry-instrumentation-pymemcache==0.29b0\",\n },\n \"pymongo\": {\n", "issue": "Support for pymemcache 2.x and 3.x\nThe `pymemcache` integration is currently locked to version 1.x. Current version is ~=3.5.\r\n\r\nhttps://github.com/open-telemetry/opentelemetry-python-contrib/blob/0b9e96dae0bfcafa1e0b28dc695e3697fbcf2864/instrumentation/opentelemetry-instrumentation-pymemcache/src/opentelemetry/instrumentation/pymemcache/package.py#L16\n", "before_files": [{"content": "# Copyright The OpenTelemetry Authors\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\n\n_instruments = (\"pymemcache ~= 1.3\",)\n", "path": "instrumentation/opentelemetry-instrumentation-pymemcache/src/opentelemetry/instrumentation/pymemcache/package.py"}, {"content": "# Copyright The OpenTelemetry Authors\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\n# DO NOT EDIT. THIS FILE WAS AUTOGENERATED FROM INSTRUMENTATION PACKAGES.\n# RUN `python scripts/generate_instrumentation_bootstrap.py` TO REGENERATE.\n\nlibraries = {\n \"aiohttp\": {\n \"library\": \"aiohttp ~= 3.0\",\n \"instrumentation\": \"opentelemetry-instrumentation-aiohttp-client==0.29b0\",\n },\n \"aiopg\": {\n \"library\": \"aiopg >= 0.13.0, < 1.3.0\",\n \"instrumentation\": \"opentelemetry-instrumentation-aiopg==0.29b0\",\n },\n \"asgiref\": {\n \"library\": \"asgiref ~= 3.0\",\n \"instrumentation\": \"opentelemetry-instrumentation-asgi==0.29b0\",\n },\n \"asyncpg\": {\n \"library\": \"asyncpg >= 0.12.0\",\n \"instrumentation\": \"opentelemetry-instrumentation-asyncpg==0.29b0\",\n },\n \"boto\": {\n \"library\": \"boto~=2.0\",\n \"instrumentation\": \"opentelemetry-instrumentation-boto==0.29b0\",\n },\n \"botocore\": {\n \"library\": \"botocore ~= 1.0\",\n \"instrumentation\": \"opentelemetry-instrumentation-botocore==0.29b0\",\n },\n \"celery\": {\n \"library\": \"celery >= 4.0, < 6.0\",\n \"instrumentation\": \"opentelemetry-instrumentation-celery==0.29b0\",\n },\n \"django\": {\n \"library\": \"django >= 1.10\",\n \"instrumentation\": \"opentelemetry-instrumentation-django==0.29b0\",\n },\n \"elasticsearch\": {\n \"library\": \"elasticsearch >= 2.0\",\n \"instrumentation\": \"opentelemetry-instrumentation-elasticsearch==0.29b0\",\n },\n \"falcon\": {\n \"library\": \"falcon >= 2.0.0, < 4.0.0\",\n \"instrumentation\": \"opentelemetry-instrumentation-falcon==0.29b0\",\n },\n \"fastapi\": {\n \"library\": \"fastapi ~= 0.58\",\n \"instrumentation\": \"opentelemetry-instrumentation-fastapi==0.29b0\",\n },\n \"flask\": {\n \"library\": \"flask >= 1.0, < 3.0\",\n \"instrumentation\": \"opentelemetry-instrumentation-flask==0.29b0\",\n },\n \"grpcio\": {\n \"library\": \"grpcio ~= 1.27\",\n \"instrumentation\": \"opentelemetry-instrumentation-grpc==0.29b0\",\n },\n \"httpx\": {\n \"library\": \"httpx >= 0.18.0\",\n \"instrumentation\": \"opentelemetry-instrumentation-httpx==0.29b0\",\n },\n \"jinja2\": {\n \"library\": \"jinja2 >= 2.7, < 4.0\",\n \"instrumentation\": \"opentelemetry-instrumentation-jinja2==0.29b0\",\n },\n \"kafka-python\": {\n \"library\": \"kafka-python >= 2.0\",\n \"instrumentation\": \"opentelemetry-instrumentation-kafka-python==0.29b0\",\n },\n \"mysql-connector-python\": {\n \"library\": \"mysql-connector-python ~= 8.0\",\n \"instrumentation\": \"opentelemetry-instrumentation-mysql==0.29b0\",\n },\n \"pika\": {\n \"library\": \"pika >= 0.12.0\",\n \"instrumentation\": \"opentelemetry-instrumentation-pika==0.29b0\",\n },\n \"psycopg2\": {\n \"library\": \"psycopg2 >= 2.7.3.1\",\n \"instrumentation\": \"opentelemetry-instrumentation-psycopg2==0.29b0\",\n },\n \"pymemcache\": {\n \"library\": \"pymemcache ~= 1.3\",\n \"instrumentation\": \"opentelemetry-instrumentation-pymemcache==0.29b0\",\n },\n \"pymongo\": {\n \"library\": \"pymongo >= 3.1, < 5.0\",\n \"instrumentation\": \"opentelemetry-instrumentation-pymongo==0.29b0\",\n },\n \"PyMySQL\": {\n \"library\": \"PyMySQL < 2\",\n \"instrumentation\": \"opentelemetry-instrumentation-pymysql==0.29b0\",\n },\n \"pyramid\": {\n \"library\": \"pyramid >= 1.7\",\n \"instrumentation\": \"opentelemetry-instrumentation-pyramid==0.29b0\",\n },\n \"redis\": {\n \"library\": \"redis >= 2.6\",\n \"instrumentation\": \"opentelemetry-instrumentation-redis==0.29b0\",\n },\n \"requests\": {\n \"library\": \"requests ~= 2.0\",\n \"instrumentation\": \"opentelemetry-instrumentation-requests==0.29b0\",\n },\n \"scikit-learn\": {\n \"library\": \"scikit-learn ~= 0.24.0\",\n \"instrumentation\": \"opentelemetry-instrumentation-sklearn==0.29b0\",\n },\n \"sqlalchemy\": {\n \"library\": \"sqlalchemy\",\n \"instrumentation\": \"opentelemetry-instrumentation-sqlalchemy==0.29b0\",\n },\n \"starlette\": {\n \"library\": \"starlette ~= 0.13.0\",\n \"instrumentation\": \"opentelemetry-instrumentation-starlette==0.29b0\",\n },\n \"tornado\": {\n \"library\": \"tornado >= 5.1.1\",\n \"instrumentation\": \"opentelemetry-instrumentation-tornado==0.29b0\",\n },\n \"urllib3\": {\n \"library\": \"urllib3 >= 1.0.0, < 2.0.0\",\n \"instrumentation\": \"opentelemetry-instrumentation-urllib3==0.29b0\",\n },\n}\ndefault_instrumentations = [\n \"opentelemetry-instrumentation-aws-lambda==0.29b0\",\n \"opentelemetry-instrumentation-dbapi==0.29b0\",\n \"opentelemetry-instrumentation-logging==0.29b0\",\n \"opentelemetry-instrumentation-sqlite3==0.29b0\",\n \"opentelemetry-instrumentation-urllib==0.29b0\",\n \"opentelemetry-instrumentation-wsgi==0.29b0\",\n]\n", "path": "opentelemetry-instrumentation/src/opentelemetry/instrumentation/bootstrap_gen.py"}]}
2,931
365
gh_patches_debug_7541
rasdani/github-patches
git_diff
twisted__twisted-12106
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> Release 24.1.0 to unbreak users who use the latest PyPy #12084 is breaking CI for Tahoe-LAFS, so probably is breaking real-world usage for someone somewhere too. So it'd be good to have a release sooner rather than later. </issue> <code> [start of src/twisted/_version.py] 1 """ 2 Provides Twisted version information. 3 """ 4 5 # This file is auto-generated! Do not edit! 6 # Use `python -m incremental.update Twisted` to change this file. 7 8 from incremental import Version 9 10 __version__ = Version("Twisted", 23, 10, 0, post=0) 11 __all__ = ["__version__"] 12 [end of src/twisted/_version.py] [start of src/twisted/copyright.py] 1 # Copyright (c) Twisted Matrix Laboratories. 2 # See LICENSE for details. 3 4 """ 5 Copyright information for Twisted. 6 """ 7 8 9 __all__ = ["copyright", "disclaimer", "longversion", "version"] 10 11 from twisted import __version__ as version, version as _longversion 12 13 longversion = str(_longversion) 14 15 copyright = """\ 16 Copyright (c) 2001-2023 Twisted Matrix Laboratories. 17 See LICENSE for details.""" 18 19 disclaimer = """ 20 Twisted, the Framework of Your Internet 21 {} 22 23 Permission is hereby granted, free of charge, to any person obtaining 24 a copy of this software and associated documentation files (the 25 "Software"), to deal in the Software without restriction, including 26 without limitation the rights to use, copy, modify, merge, publish, 27 distribute, sublicense, and/or sell copies of the Software, and to 28 permit persons to whom the Software is furnished to do so, subject to 29 the following conditions: 30 31 The above copyright notice and this permission notice shall be 32 included in all copies or substantial portions of the Software. 33 34 THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, 35 EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF 36 MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND 37 NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE 38 LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION 39 OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION 40 WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. 41 42 """.format( 43 copyright, 44 ) 45 [end of src/twisted/copyright.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/src/twisted/_version.py b/src/twisted/_version.py --- a/src/twisted/_version.py +++ b/src/twisted/_version.py @@ -7,5 +7,5 @@ from incremental import Version -__version__ = Version("Twisted", 23, 10, 0, post=0) +__version__ = Version("Twisted", 24, 3, 0, post=0) __all__ = ["__version__"] diff --git a/src/twisted/copyright.py b/src/twisted/copyright.py --- a/src/twisted/copyright.py +++ b/src/twisted/copyright.py @@ -13,7 +13,7 @@ longversion = str(_longversion) copyright = """\ -Copyright (c) 2001-2023 Twisted Matrix Laboratories. +Copyright (c) 2001-2024 Twisted Matrix Laboratories. See LICENSE for details.""" disclaimer = """
{"golden_diff": "diff --git a/src/twisted/_version.py b/src/twisted/_version.py\n--- a/src/twisted/_version.py\n+++ b/src/twisted/_version.py\n@@ -7,5 +7,5 @@\n \n from incremental import Version\n \n-__version__ = Version(\"Twisted\", 23, 10, 0, post=0)\n+__version__ = Version(\"Twisted\", 24, 3, 0, post=0)\n __all__ = [\"__version__\"]\ndiff --git a/src/twisted/copyright.py b/src/twisted/copyright.py\n--- a/src/twisted/copyright.py\n+++ b/src/twisted/copyright.py\n@@ -13,7 +13,7 @@\n longversion = str(_longversion)\n \n copyright = \"\"\"\\\n-Copyright (c) 2001-2023 Twisted Matrix Laboratories.\n+Copyright (c) 2001-2024 Twisted Matrix Laboratories.\n See LICENSE for details.\"\"\"\n \n disclaimer = \"\"\"\n", "issue": "Release 24.1.0 to unbreak users who use the latest PyPy\n#12084 is breaking CI for Tahoe-LAFS, so probably is breaking real-world usage for someone somewhere too. So it'd be good to have a release sooner rather than later.\n", "before_files": [{"content": "\"\"\"\nProvides Twisted version information.\n\"\"\"\n\n# This file is auto-generated! Do not edit!\n# Use `python -m incremental.update Twisted` to change this file.\n\nfrom incremental import Version\n\n__version__ = Version(\"Twisted\", 23, 10, 0, post=0)\n__all__ = [\"__version__\"]\n", "path": "src/twisted/_version.py"}, {"content": "# Copyright (c) Twisted Matrix Laboratories.\n# See LICENSE for details.\n\n\"\"\"\nCopyright information for Twisted.\n\"\"\"\n\n\n__all__ = [\"copyright\", \"disclaimer\", \"longversion\", \"version\"]\n\nfrom twisted import __version__ as version, version as _longversion\n\nlongversion = str(_longversion)\n\ncopyright = \"\"\"\\\nCopyright (c) 2001-2023 Twisted Matrix Laboratories.\nSee LICENSE for details.\"\"\"\n\ndisclaimer = \"\"\"\nTwisted, the Framework of Your Internet\n{}\n\nPermission is hereby granted, free of charge, to any person obtaining\na copy of this software and associated documentation files (the\n\"Software\"), to deal in the Software without restriction, including\nwithout limitation the rights to use, copy, modify, merge, publish,\ndistribute, sublicense, and/or sell copies of the Software, and to\npermit persons to whom the Software is furnished to do so, subject to\nthe following conditions:\n\nThe above copyright notice and this permission notice shall be\nincluded in all copies or substantial portions of the Software.\n\nTHE SOFTWARE IS PROVIDED \"AS IS\", WITHOUT WARRANTY OF ANY KIND,\nEXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF\nMERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND\nNONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE\nLIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION\nOF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION\nWITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.\n\n\"\"\".format(\n copyright,\n)\n", "path": "src/twisted/copyright.py"}]}
1,119
228
gh_patches_debug_6887
rasdani/github-patches
git_diff
sherlock-project__sherlock-911
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> [site_list.py] change numbering to reduce commit size letting the markdown renderer do the counting lets us reduce commit size and avoide possible merge conflicts. --- ``` 1. 1. 1. ``` renders to: 1. 1. 1. </issue> <code> [start of site_list.py] 1 """Sherlock: Supported Site Listing 2 This module generates the listing of supported sites 3 which can be found in sites.md 4 It also organizes all the sites in alphanumeric order 5 """ 6 import json 7 8 pool = list() 9 10 with open("sherlock/resources/data.json", "r", encoding="utf-8") as data_file: 11 data = json.load(data_file) 12 13 with open("sites.md", "w") as site_file: 14 data_length = len(data) 15 site_file.write(f'## List Of Supported Sites ({data_length} Sites In Total!)\n') 16 17 for social_network in data: 18 url_main = data.get(social_network).get("urlMain") 19 pool.append((social_network, url_main)) 20 21 index = 1 22 for social_network, url_main in pool: 23 site_file.write(f'{index}. [{social_network}]({url_main})\n') 24 index = index + 1 25 26 27 sorted_json_data = json.dumps(data, indent=2, sort_keys=True) 28 29 with open("sherlock/resources/data.json", "w") as data_file: 30 data_file.write(sorted_json_data) 31 32 print("Finished updating supported site listing!") 33 [end of site_list.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/site_list.py b/site_list.py --- a/site_list.py +++ b/site_list.py @@ -18,11 +18,8 @@ url_main = data.get(social_network).get("urlMain") pool.append((social_network, url_main)) - index = 1 for social_network, url_main in pool: - site_file.write(f'{index}. [{social_network}]({url_main})\n') - index = index + 1 - + site_file.write(f'1. [{social_network}]({url_main})\n') sorted_json_data = json.dumps(data, indent=2, sort_keys=True)
{"golden_diff": "diff --git a/site_list.py b/site_list.py\n--- a/site_list.py\n+++ b/site_list.py\n@@ -18,11 +18,8 @@\n url_main = data.get(social_network).get(\"urlMain\")\n pool.append((social_network, url_main))\n \n- index = 1\n for social_network, url_main in pool:\n- site_file.write(f'{index}. [{social_network}]({url_main})\\n')\n- index = index + 1\n-\n+ site_file.write(f'1. [{social_network}]({url_main})\\n')\n \n sorted_json_data = json.dumps(data, indent=2, sort_keys=True)\n", "issue": "[site_list.py] change numbering to reduce commit size\nletting the markdown renderer do the counting lets us reduce commit size and avoide possible merge conflicts.\r\n\r\n---\r\n\r\n```\r\n1.\r\n1.\r\n1.\r\n```\r\nrenders to:\r\n\r\n1.\r\n1.\r\n1.\n", "before_files": [{"content": "\"\"\"Sherlock: Supported Site Listing\nThis module generates the listing of supported sites\nwhich can be found in sites.md\nIt also organizes all the sites in alphanumeric order\n\"\"\"\nimport json\n\npool = list()\n\nwith open(\"sherlock/resources/data.json\", \"r\", encoding=\"utf-8\") as data_file:\n data = json.load(data_file)\n\nwith open(\"sites.md\", \"w\") as site_file:\n data_length = len(data)\n site_file.write(f'## List Of Supported Sites ({data_length} Sites In Total!)\\n')\n\n for social_network in data:\n url_main = data.get(social_network).get(\"urlMain\")\n pool.append((social_network, url_main))\n\n index = 1\n for social_network, url_main in pool:\n site_file.write(f'{index}. [{social_network}]({url_main})\\n')\n index = index + 1\n\n\nsorted_json_data = json.dumps(data, indent=2, sort_keys=True)\n\nwith open(\"sherlock/resources/data.json\", \"w\") as data_file:\n data_file.write(sorted_json_data)\n\nprint(\"Finished updating supported site listing!\")\n", "path": "site_list.py"}]}
890
147
gh_patches_debug_10290
rasdani/github-patches
git_diff
goauthentik__authentik-8139
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> 2023.10.6 - "Please select a username" after Azure AD login **Describe your question/** Is it now a expected behavior in 2023.10.6 version to ask every user for username input after logging in with azure ad? ![image](https://github.com/goauthentik/authentik/assets/7706996/7c35f530-62a2-4afd-92aa-c290ad221c5f) In previous versions it was simply authenticating without any prompt, using email address from Azure AD as username. Now it expects user to input username (and it leads to duplicated accounts, because users with mail as username already exist), and if you enter already existing mail as username it shows error: ![image](https://github.com/goauthentik/authentik/assets/7706996/a8aaaf90-b309-425c-92a1-174739d01fc6) I think it can be related to this fix: https://github.com/goauthentik/authentik/pull/7970 Is it possible somehow to set this username automatically, or revert back to using email address so old user accounts will work again? **Version and Deployment (please complete the following information):** - authentik version: 2023.10.6 - Deployment: helm </issue> <code> [start of authentik/sources/oauth/types/azure_ad.py] 1 """AzureAD OAuth2 Views""" 2 from typing import Any 3 4 from structlog.stdlib import get_logger 5 6 from authentik.sources.oauth.clients.oauth2 import UserprofileHeaderAuthClient 7 from authentik.sources.oauth.types.oidc import OpenIDConnectOAuth2Callback 8 from authentik.sources.oauth.types.registry import SourceType, registry 9 from authentik.sources.oauth.views.redirect import OAuthRedirect 10 11 LOGGER = get_logger() 12 13 14 class AzureADOAuthRedirect(OAuthRedirect): 15 """Azure AD OAuth2 Redirect""" 16 17 def get_additional_parameters(self, source): # pragma: no cover 18 return { 19 "scope": ["openid", "https://graph.microsoft.com/User.Read"], 20 } 21 22 23 class AzureADOAuthCallback(OpenIDConnectOAuth2Callback): 24 """AzureAD OAuth2 Callback""" 25 26 client_class = UserprofileHeaderAuthClient 27 28 def get_user_enroll_context( 29 self, 30 info: dict[str, Any], 31 ) -> dict[str, Any]: 32 mail = info.get("mail", None) or info.get("otherMails", [None])[0] 33 return { 34 "username": info.get("userPrincipalName"), 35 "email": mail, 36 "name": info.get("displayName"), 37 } 38 39 40 @registry.register() 41 class AzureADType(SourceType): 42 """Azure AD Type definition""" 43 44 callback_view = AzureADOAuthCallback 45 redirect_view = AzureADOAuthRedirect 46 verbose_name = "Azure AD" 47 name = "azuread" 48 49 urls_customizable = True 50 51 authorization_url = "https://login.microsoftonline.com/common/oauth2/v2.0/authorize" 52 access_token_url = "https://login.microsoftonline.com/common/oauth2/v2.0/token" # nosec 53 profile_url = "https://login.microsoftonline.com/common/openid/userinfo" 54 oidc_well_known_url = ( 55 "https://login.microsoftonline.com/common/.well-known/openid-configuration" 56 ) 57 oidc_jwks_url = "https://login.microsoftonline.com/common/discovery/keys" 58 [end of authentik/sources/oauth/types/azure_ad.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/authentik/sources/oauth/types/azure_ad.py b/authentik/sources/oauth/types/azure_ad.py --- a/authentik/sources/oauth/types/azure_ad.py +++ b/authentik/sources/oauth/types/azure_ad.py @@ -50,7 +50,7 @@ authorization_url = "https://login.microsoftonline.com/common/oauth2/v2.0/authorize" access_token_url = "https://login.microsoftonline.com/common/oauth2/v2.0/token" # nosec - profile_url = "https://login.microsoftonline.com/common/openid/userinfo" + profile_url = "https://graph.microsoft.com/v1.0/me" oidc_well_known_url = ( "https://login.microsoftonline.com/common/.well-known/openid-configuration" )
{"golden_diff": "diff --git a/authentik/sources/oauth/types/azure_ad.py b/authentik/sources/oauth/types/azure_ad.py\n--- a/authentik/sources/oauth/types/azure_ad.py\n+++ b/authentik/sources/oauth/types/azure_ad.py\n@@ -50,7 +50,7 @@\n \n authorization_url = \"https://login.microsoftonline.com/common/oauth2/v2.0/authorize\"\n access_token_url = \"https://login.microsoftonline.com/common/oauth2/v2.0/token\" # nosec\n- profile_url = \"https://login.microsoftonline.com/common/openid/userinfo\"\n+ profile_url = \"https://graph.microsoft.com/v1.0/me\"\n oidc_well_known_url = (\n \"https://login.microsoftonline.com/common/.well-known/openid-configuration\"\n )\n", "issue": "2023.10.6 - \"Please select a username\" after Azure AD login\n**Describe your question/**\r\n\r\nIs it now a expected behavior in 2023.10.6 version to ask every user for username input after logging in with azure ad?\r\n![image](https://github.com/goauthentik/authentik/assets/7706996/7c35f530-62a2-4afd-92aa-c290ad221c5f)\r\n\r\nIn previous versions it was simply authenticating without any prompt, using email address from Azure AD as username.\r\n\r\nNow it expects user to input username (and it leads to duplicated accounts, because users with mail as username already exist), and if you enter already existing mail as username it shows error:\r\n![image](https://github.com/goauthentik/authentik/assets/7706996/a8aaaf90-b309-425c-92a1-174739d01fc6)\r\n\r\nI think it can be related to this fix:\r\nhttps://github.com/goauthentik/authentik/pull/7970\r\n\r\nIs it possible somehow to set this username automatically, or revert back to using email address so old user accounts will work again?\r\n\r\n**Version and Deployment (please complete the following information):**\r\n\r\n- authentik version: 2023.10.6\r\n- Deployment: helm\r\n\r\n\n", "before_files": [{"content": "\"\"\"AzureAD OAuth2 Views\"\"\"\nfrom typing import Any\n\nfrom structlog.stdlib import get_logger\n\nfrom authentik.sources.oauth.clients.oauth2 import UserprofileHeaderAuthClient\nfrom authentik.sources.oauth.types.oidc import OpenIDConnectOAuth2Callback\nfrom authentik.sources.oauth.types.registry import SourceType, registry\nfrom authentik.sources.oauth.views.redirect import OAuthRedirect\n\nLOGGER = get_logger()\n\n\nclass AzureADOAuthRedirect(OAuthRedirect):\n \"\"\"Azure AD OAuth2 Redirect\"\"\"\n\n def get_additional_parameters(self, source): # pragma: no cover\n return {\n \"scope\": [\"openid\", \"https://graph.microsoft.com/User.Read\"],\n }\n\n\nclass AzureADOAuthCallback(OpenIDConnectOAuth2Callback):\n \"\"\"AzureAD OAuth2 Callback\"\"\"\n\n client_class = UserprofileHeaderAuthClient\n\n def get_user_enroll_context(\n self,\n info: dict[str, Any],\n ) -> dict[str, Any]:\n mail = info.get(\"mail\", None) or info.get(\"otherMails\", [None])[0]\n return {\n \"username\": info.get(\"userPrincipalName\"),\n \"email\": mail,\n \"name\": info.get(\"displayName\"),\n }\n\n\[email protected]()\nclass AzureADType(SourceType):\n \"\"\"Azure AD Type definition\"\"\"\n\n callback_view = AzureADOAuthCallback\n redirect_view = AzureADOAuthRedirect\n verbose_name = \"Azure AD\"\n name = \"azuread\"\n\n urls_customizable = True\n\n authorization_url = \"https://login.microsoftonline.com/common/oauth2/v2.0/authorize\"\n access_token_url = \"https://login.microsoftonline.com/common/oauth2/v2.0/token\" # nosec\n profile_url = \"https://login.microsoftonline.com/common/openid/userinfo\"\n oidc_well_known_url = (\n \"https://login.microsoftonline.com/common/.well-known/openid-configuration\"\n )\n oidc_jwks_url = \"https://login.microsoftonline.com/common/discovery/keys\"\n", "path": "authentik/sources/oauth/types/azure_ad.py"}]}
1,413
175
gh_patches_debug_9688
rasdani/github-patches
git_diff
googleapis__google-cloud-python-4170
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> Spanner: IndexError: pop from empty list when merge_array ``` return merger(lhs, rhs, type_) File ".../google/cloud/spanner_v1/streamed.py", line 259, in _merge_array first = rhs.pop(0) IndexError: pop from empty list ``` </issue> <code> [start of spanner/google/cloud/spanner_v1/streamed.py] 1 # Copyright 2016 Google Inc. All rights reserved. 2 # 3 # Licensed under the Apache License, Version 2.0 (the "License"); 4 # you may not use this file except in compliance with the License. 5 # You may obtain a copy of the License at 6 # 7 # http://www.apache.org/licenses/LICENSE-2.0 8 # 9 # Unless required by applicable law or agreed to in writing, software 10 # distributed under the License is distributed on an "AS IS" BASIS, 11 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 12 # See the License for the specific language governing permissions and 13 # limitations under the License. 14 15 """Wrapper for streaming results.""" 16 17 from google.protobuf.struct_pb2 import ListValue 18 from google.protobuf.struct_pb2 import Value 19 from google.cloud import exceptions 20 from google.cloud.spanner_v1.proto import type_pb2 21 import six 22 23 # pylint: disable=ungrouped-imports 24 from google.cloud.spanner_v1._helpers import _parse_value_pb 25 # pylint: enable=ungrouped-imports 26 27 28 class StreamedResultSet(object): 29 """Process a sequence of partial result sets into a single set of row data. 30 31 :type response_iterator: 32 :param response_iterator: 33 Iterator yielding 34 :class:`google.cloud.spanner_v1.proto.result_set_pb2.PartialResultSet` 35 instances. 36 37 :type source: :class:`~google.cloud.spanner_v1.snapshot.Snapshot` 38 :param source: Snapshot from which the result set was fetched. 39 """ 40 def __init__(self, response_iterator, source=None): 41 self._response_iterator = response_iterator 42 self._rows = [] # Fully-processed rows 43 self._counter = 0 # Counter for processed responses 44 self._metadata = None # Until set from first PRS 45 self._stats = None # Until set from last PRS 46 self._current_row = [] # Accumulated values for incomplete row 47 self._pending_chunk = None # Incomplete value 48 self._source = source # Source snapshot 49 50 @property 51 def rows(self): 52 """Fully-processed rows. 53 54 :rtype: list of row-data lists. 55 :returns: list of completed row data, from proceesd PRS responses. 56 """ 57 return self._rows 58 59 @property 60 def fields(self): 61 """Field descriptors for result set columns. 62 63 :rtype: list of :class:`~google.cloud.spanner_v1.proto.type_pb2.Field` 64 :returns: list of fields describing column names / types. 65 """ 66 return self._metadata.row_type.fields 67 68 @property 69 def metadata(self): 70 """Result set metadata 71 72 :rtype: :class:`~.result_set_pb2.ResultSetMetadata` 73 :returns: structure describing the results 74 """ 75 return self._metadata 76 77 @property 78 def stats(self): 79 """Result set statistics 80 81 :rtype: 82 :class:`~google.cloud.spanner_v1.proto.result_set_pb2.ResultSetStats` 83 :returns: structure describing status about the response 84 """ 85 return self._stats 86 87 def _merge_chunk(self, value): 88 """Merge pending chunk with next value. 89 90 :type value: :class:`~google.protobuf.struct_pb2.Value` 91 :param value: continuation of chunked value from previous 92 partial result set. 93 94 :rtype: :class:`~google.protobuf.struct_pb2.Value` 95 :returns: the merged value 96 """ 97 current_column = len(self._current_row) 98 field = self.fields[current_column] 99 merged = _merge_by_type(self._pending_chunk, value, field.type) 100 self._pending_chunk = None 101 return merged 102 103 def _merge_values(self, values): 104 """Merge values into rows. 105 106 :type values: list of :class:`~google.protobuf.struct_pb2.Value` 107 :param values: non-chunked values from partial result set. 108 """ 109 width = len(self.fields) 110 for value in values: 111 index = len(self._current_row) 112 field = self.fields[index] 113 self._current_row.append(_parse_value_pb(value, field.type)) 114 if len(self._current_row) == width: 115 self._rows.append(self._current_row) 116 self._current_row = [] 117 118 def consume_next(self): 119 """Consume the next partial result set from the stream. 120 121 Parse the result set into new/existing rows in :attr:`_rows` 122 """ 123 response = six.next(self._response_iterator) 124 self._counter += 1 125 126 if self._metadata is None: # first response 127 metadata = self._metadata = response.metadata 128 129 source = self._source 130 if source is not None and source._transaction_id is None: 131 source._transaction_id = metadata.transaction.id 132 133 if response.HasField('stats'): # last response 134 self._stats = response.stats 135 136 values = list(response.values) 137 if self._pending_chunk is not None: 138 values[0] = self._merge_chunk(values[0]) 139 140 if response.chunked_value: 141 self._pending_chunk = values.pop() 142 143 self._merge_values(values) 144 145 def consume_all(self): 146 """Consume the streamed responses until there are no more.""" 147 while True: 148 try: 149 self.consume_next() 150 except StopIteration: 151 break 152 153 def __iter__(self): 154 iter_rows, self._rows[:] = self._rows[:], () 155 while True: 156 if not iter_rows: 157 self.consume_next() # raises StopIteration 158 iter_rows, self._rows[:] = self._rows[:], () 159 while iter_rows: 160 yield iter_rows.pop(0) 161 162 def one(self): 163 """Return exactly one result, or raise an exception. 164 165 :raises: :exc:`NotFound`: If there are no results. 166 :raises: :exc:`ValueError`: If there are multiple results. 167 :raises: :exc:`RuntimeError`: If consumption has already occurred, 168 in whole or in part. 169 """ 170 answer = self.one_or_none() 171 if answer is None: 172 raise exceptions.NotFound('No rows matched the given query.') 173 return answer 174 175 def one_or_none(self): 176 """Return exactly one result, or None if there are no results. 177 178 :raises: :exc:`ValueError`: If there are multiple results. 179 :raises: :exc:`RuntimeError`: If consumption has already occurred, 180 in whole or in part. 181 """ 182 # Sanity check: Has consumption of this query already started? 183 # If it has, then this is an exception. 184 if self._metadata is not None: 185 raise RuntimeError('Can not call `.one` or `.one_or_none` after ' 186 'stream consumption has already started.') 187 188 # Consume the first result of the stream. 189 # If there is no first result, then return None. 190 iterator = iter(self) 191 try: 192 answer = next(iterator) 193 except StopIteration: 194 return None 195 196 # Attempt to consume more. This should no-op; if we get additional 197 # rows, then this is an error case. 198 try: 199 next(iterator) 200 raise ValueError('Expected one result; got more.') 201 except StopIteration: 202 return answer 203 204 205 class Unmergeable(ValueError): 206 """Unable to merge two values. 207 208 :type lhs: :class:`google.protobuf.struct_pb2.Value` 209 :param lhs: pending value to be merged 210 211 :type rhs: :class:`google.protobuf.struct_pb2.Value` 212 :param rhs: remaining value to be merged 213 214 :type type_: :class:`google.cloud.spanner_v1.proto.type_pb2.Type` 215 :param type_: field type of values being merged 216 """ 217 def __init__(self, lhs, rhs, type_): 218 message = "Cannot merge %s values: %s %s" % ( 219 type_pb2.TypeCode.Name(type_.code), lhs, rhs) 220 super(Unmergeable, self).__init__(message) 221 222 223 def _unmergeable(lhs, rhs, type_): 224 """Helper for '_merge_by_type'.""" 225 raise Unmergeable(lhs, rhs, type_) 226 227 228 def _merge_float64(lhs, rhs, type_): # pylint: disable=unused-argument 229 """Helper for '_merge_by_type'.""" 230 lhs_kind = lhs.WhichOneof('kind') 231 if lhs_kind == 'string_value': 232 return Value(string_value=lhs.string_value + rhs.string_value) 233 rhs_kind = rhs.WhichOneof('kind') 234 array_continuation = ( 235 lhs_kind == 'number_value' and 236 rhs_kind == 'string_value' and 237 rhs.string_value == '') 238 if array_continuation: 239 return lhs 240 raise Unmergeable(lhs, rhs, type_) 241 242 243 def _merge_string(lhs, rhs, type_): # pylint: disable=unused-argument 244 """Helper for '_merge_by_type'.""" 245 return Value(string_value=lhs.string_value + rhs.string_value) 246 247 248 _UNMERGEABLE_TYPES = (type_pb2.BOOL,) 249 250 251 def _merge_array(lhs, rhs, type_): 252 """Helper for '_merge_by_type'.""" 253 element_type = type_.array_element_type 254 if element_type.code in _UNMERGEABLE_TYPES: 255 # Individual values cannot be merged, just concatenate 256 lhs.list_value.values.extend(rhs.list_value.values) 257 return lhs 258 lhs, rhs = list(lhs.list_value.values), list(rhs.list_value.values) 259 first = rhs.pop(0) 260 if first.HasField('null_value'): # can't merge 261 lhs.append(first) 262 else: 263 last = lhs.pop() 264 try: 265 merged = _merge_by_type(last, first, element_type) 266 except Unmergeable: 267 lhs.append(last) 268 lhs.append(first) 269 else: 270 lhs.append(merged) 271 return Value(list_value=ListValue(values=(lhs + rhs))) 272 273 274 def _merge_struct(lhs, rhs, type_): 275 """Helper for '_merge_by_type'.""" 276 fields = type_.struct_type.fields 277 lhs, rhs = list(lhs.list_value.values), list(rhs.list_value.values) 278 candidate_type = fields[len(lhs) - 1].type 279 first = rhs.pop(0) 280 if (first.HasField('null_value') or 281 candidate_type.code in _UNMERGEABLE_TYPES): 282 lhs.append(first) 283 else: 284 last = lhs.pop() 285 lhs.append(_merge_by_type(last, first, candidate_type)) 286 return Value(list_value=ListValue(values=lhs + rhs)) 287 288 289 _MERGE_BY_TYPE = { 290 type_pb2.ARRAY: _merge_array, 291 type_pb2.BOOL: _unmergeable, 292 type_pb2.BYTES: _merge_string, 293 type_pb2.DATE: _merge_string, 294 type_pb2.FLOAT64: _merge_float64, 295 type_pb2.INT64: _merge_string, 296 type_pb2.STRING: _merge_string, 297 type_pb2.STRUCT: _merge_struct, 298 type_pb2.TIMESTAMP: _merge_string, 299 } 300 301 302 def _merge_by_type(lhs, rhs, type_): 303 """Helper for '_merge_chunk'.""" 304 merger = _MERGE_BY_TYPE[type_.code] 305 return merger(lhs, rhs, type_) 306 [end of spanner/google/cloud/spanner_v1/streamed.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/spanner/google/cloud/spanner_v1/streamed.py b/spanner/google/cloud/spanner_v1/streamed.py --- a/spanner/google/cloud/spanner_v1/streamed.py +++ b/spanner/google/cloud/spanner_v1/streamed.py @@ -256,6 +256,12 @@ lhs.list_value.values.extend(rhs.list_value.values) return lhs lhs, rhs = list(lhs.list_value.values), list(rhs.list_value.values) + + # Sanity check: If either list is empty, short-circuit. + # This is effectively a no-op. + if not len(lhs) or not len(rhs): + return Value(list_value=ListValue(values=(lhs + rhs))) + first = rhs.pop(0) if first.HasField('null_value'): # can't merge lhs.append(first)
{"golden_diff": "diff --git a/spanner/google/cloud/spanner_v1/streamed.py b/spanner/google/cloud/spanner_v1/streamed.py\n--- a/spanner/google/cloud/spanner_v1/streamed.py\n+++ b/spanner/google/cloud/spanner_v1/streamed.py\n@@ -256,6 +256,12 @@\n lhs.list_value.values.extend(rhs.list_value.values)\n return lhs\n lhs, rhs = list(lhs.list_value.values), list(rhs.list_value.values)\n+\n+ # Sanity check: If either list is empty, short-circuit.\n+ # This is effectively a no-op.\n+ if not len(lhs) or not len(rhs):\n+ return Value(list_value=ListValue(values=(lhs + rhs)))\n+\n first = rhs.pop(0)\n if first.HasField('null_value'): # can't merge\n lhs.append(first)\n", "issue": "Spanner: IndexError: pop from empty list when merge_array\n```\r\nreturn merger(lhs, rhs, type_)\r\n File \".../google/cloud/spanner_v1/streamed.py\", line 259, in _merge_array\r\n first = rhs.pop(0)\r\nIndexError: pop from empty list\r\n```\n", "before_files": [{"content": "# Copyright 2016 Google Inc. All rights reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\n\"\"\"Wrapper for streaming results.\"\"\"\n\nfrom google.protobuf.struct_pb2 import ListValue\nfrom google.protobuf.struct_pb2 import Value\nfrom google.cloud import exceptions\nfrom google.cloud.spanner_v1.proto import type_pb2\nimport six\n\n# pylint: disable=ungrouped-imports\nfrom google.cloud.spanner_v1._helpers import _parse_value_pb\n# pylint: enable=ungrouped-imports\n\n\nclass StreamedResultSet(object):\n \"\"\"Process a sequence of partial result sets into a single set of row data.\n\n :type response_iterator:\n :param response_iterator:\n Iterator yielding\n :class:`google.cloud.spanner_v1.proto.result_set_pb2.PartialResultSet`\n instances.\n\n :type source: :class:`~google.cloud.spanner_v1.snapshot.Snapshot`\n :param source: Snapshot from which the result set was fetched.\n \"\"\"\n def __init__(self, response_iterator, source=None):\n self._response_iterator = response_iterator\n self._rows = [] # Fully-processed rows\n self._counter = 0 # Counter for processed responses\n self._metadata = None # Until set from first PRS\n self._stats = None # Until set from last PRS\n self._current_row = [] # Accumulated values for incomplete row\n self._pending_chunk = None # Incomplete value\n self._source = source # Source snapshot\n\n @property\n def rows(self):\n \"\"\"Fully-processed rows.\n\n :rtype: list of row-data lists.\n :returns: list of completed row data, from proceesd PRS responses.\n \"\"\"\n return self._rows\n\n @property\n def fields(self):\n \"\"\"Field descriptors for result set columns.\n\n :rtype: list of :class:`~google.cloud.spanner_v1.proto.type_pb2.Field`\n :returns: list of fields describing column names / types.\n \"\"\"\n return self._metadata.row_type.fields\n\n @property\n def metadata(self):\n \"\"\"Result set metadata\n\n :rtype: :class:`~.result_set_pb2.ResultSetMetadata`\n :returns: structure describing the results\n \"\"\"\n return self._metadata\n\n @property\n def stats(self):\n \"\"\"Result set statistics\n\n :rtype:\n :class:`~google.cloud.spanner_v1.proto.result_set_pb2.ResultSetStats`\n :returns: structure describing status about the response\n \"\"\"\n return self._stats\n\n def _merge_chunk(self, value):\n \"\"\"Merge pending chunk with next value.\n\n :type value: :class:`~google.protobuf.struct_pb2.Value`\n :param value: continuation of chunked value from previous\n partial result set.\n\n :rtype: :class:`~google.protobuf.struct_pb2.Value`\n :returns: the merged value\n \"\"\"\n current_column = len(self._current_row)\n field = self.fields[current_column]\n merged = _merge_by_type(self._pending_chunk, value, field.type)\n self._pending_chunk = None\n return merged\n\n def _merge_values(self, values):\n \"\"\"Merge values into rows.\n\n :type values: list of :class:`~google.protobuf.struct_pb2.Value`\n :param values: non-chunked values from partial result set.\n \"\"\"\n width = len(self.fields)\n for value in values:\n index = len(self._current_row)\n field = self.fields[index]\n self._current_row.append(_parse_value_pb(value, field.type))\n if len(self._current_row) == width:\n self._rows.append(self._current_row)\n self._current_row = []\n\n def consume_next(self):\n \"\"\"Consume the next partial result set from the stream.\n\n Parse the result set into new/existing rows in :attr:`_rows`\n \"\"\"\n response = six.next(self._response_iterator)\n self._counter += 1\n\n if self._metadata is None: # first response\n metadata = self._metadata = response.metadata\n\n source = self._source\n if source is not None and source._transaction_id is None:\n source._transaction_id = metadata.transaction.id\n\n if response.HasField('stats'): # last response\n self._stats = response.stats\n\n values = list(response.values)\n if self._pending_chunk is not None:\n values[0] = self._merge_chunk(values[0])\n\n if response.chunked_value:\n self._pending_chunk = values.pop()\n\n self._merge_values(values)\n\n def consume_all(self):\n \"\"\"Consume the streamed responses until there are no more.\"\"\"\n while True:\n try:\n self.consume_next()\n except StopIteration:\n break\n\n def __iter__(self):\n iter_rows, self._rows[:] = self._rows[:], ()\n while True:\n if not iter_rows:\n self.consume_next() # raises StopIteration\n iter_rows, self._rows[:] = self._rows[:], ()\n while iter_rows:\n yield iter_rows.pop(0)\n\n def one(self):\n \"\"\"Return exactly one result, or raise an exception.\n\n :raises: :exc:`NotFound`: If there are no results.\n :raises: :exc:`ValueError`: If there are multiple results.\n :raises: :exc:`RuntimeError`: If consumption has already occurred,\n in whole or in part.\n \"\"\"\n answer = self.one_or_none()\n if answer is None:\n raise exceptions.NotFound('No rows matched the given query.')\n return answer\n\n def one_or_none(self):\n \"\"\"Return exactly one result, or None if there are no results.\n\n :raises: :exc:`ValueError`: If there are multiple results.\n :raises: :exc:`RuntimeError`: If consumption has already occurred,\n in whole or in part.\n \"\"\"\n # Sanity check: Has consumption of this query already started?\n # If it has, then this is an exception.\n if self._metadata is not None:\n raise RuntimeError('Can not call `.one` or `.one_or_none` after '\n 'stream consumption has already started.')\n\n # Consume the first result of the stream.\n # If there is no first result, then return None.\n iterator = iter(self)\n try:\n answer = next(iterator)\n except StopIteration:\n return None\n\n # Attempt to consume more. This should no-op; if we get additional\n # rows, then this is an error case.\n try:\n next(iterator)\n raise ValueError('Expected one result; got more.')\n except StopIteration:\n return answer\n\n\nclass Unmergeable(ValueError):\n \"\"\"Unable to merge two values.\n\n :type lhs: :class:`google.protobuf.struct_pb2.Value`\n :param lhs: pending value to be merged\n\n :type rhs: :class:`google.protobuf.struct_pb2.Value`\n :param rhs: remaining value to be merged\n\n :type type_: :class:`google.cloud.spanner_v1.proto.type_pb2.Type`\n :param type_: field type of values being merged\n \"\"\"\n def __init__(self, lhs, rhs, type_):\n message = \"Cannot merge %s values: %s %s\" % (\n type_pb2.TypeCode.Name(type_.code), lhs, rhs)\n super(Unmergeable, self).__init__(message)\n\n\ndef _unmergeable(lhs, rhs, type_):\n \"\"\"Helper for '_merge_by_type'.\"\"\"\n raise Unmergeable(lhs, rhs, type_)\n\n\ndef _merge_float64(lhs, rhs, type_): # pylint: disable=unused-argument\n \"\"\"Helper for '_merge_by_type'.\"\"\"\n lhs_kind = lhs.WhichOneof('kind')\n if lhs_kind == 'string_value':\n return Value(string_value=lhs.string_value + rhs.string_value)\n rhs_kind = rhs.WhichOneof('kind')\n array_continuation = (\n lhs_kind == 'number_value' and\n rhs_kind == 'string_value' and\n rhs.string_value == '')\n if array_continuation:\n return lhs\n raise Unmergeable(lhs, rhs, type_)\n\n\ndef _merge_string(lhs, rhs, type_): # pylint: disable=unused-argument\n \"\"\"Helper for '_merge_by_type'.\"\"\"\n return Value(string_value=lhs.string_value + rhs.string_value)\n\n\n_UNMERGEABLE_TYPES = (type_pb2.BOOL,)\n\n\ndef _merge_array(lhs, rhs, type_):\n \"\"\"Helper for '_merge_by_type'.\"\"\"\n element_type = type_.array_element_type\n if element_type.code in _UNMERGEABLE_TYPES:\n # Individual values cannot be merged, just concatenate\n lhs.list_value.values.extend(rhs.list_value.values)\n return lhs\n lhs, rhs = list(lhs.list_value.values), list(rhs.list_value.values)\n first = rhs.pop(0)\n if first.HasField('null_value'): # can't merge\n lhs.append(first)\n else:\n last = lhs.pop()\n try:\n merged = _merge_by_type(last, first, element_type)\n except Unmergeable:\n lhs.append(last)\n lhs.append(first)\n else:\n lhs.append(merged)\n return Value(list_value=ListValue(values=(lhs + rhs)))\n\n\ndef _merge_struct(lhs, rhs, type_):\n \"\"\"Helper for '_merge_by_type'.\"\"\"\n fields = type_.struct_type.fields\n lhs, rhs = list(lhs.list_value.values), list(rhs.list_value.values)\n candidate_type = fields[len(lhs) - 1].type\n first = rhs.pop(0)\n if (first.HasField('null_value') or\n candidate_type.code in _UNMERGEABLE_TYPES):\n lhs.append(first)\n else:\n last = lhs.pop()\n lhs.append(_merge_by_type(last, first, candidate_type))\n return Value(list_value=ListValue(values=lhs + rhs))\n\n\n_MERGE_BY_TYPE = {\n type_pb2.ARRAY: _merge_array,\n type_pb2.BOOL: _unmergeable,\n type_pb2.BYTES: _merge_string,\n type_pb2.DATE: _merge_string,\n type_pb2.FLOAT64: _merge_float64,\n type_pb2.INT64: _merge_string,\n type_pb2.STRING: _merge_string,\n type_pb2.STRUCT: _merge_struct,\n type_pb2.TIMESTAMP: _merge_string,\n}\n\n\ndef _merge_by_type(lhs, rhs, type_):\n \"\"\"Helper for '_merge_chunk'.\"\"\"\n merger = _MERGE_BY_TYPE[type_.code]\n return merger(lhs, rhs, type_)\n", "path": "spanner/google/cloud/spanner_v1/streamed.py"}]}
3,850
186
gh_patches_debug_15565
rasdani/github-patches
git_diff
deepset-ai__haystack-7796
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> [V2.2.0] ChatPromptBuilder is not export **Describe the bug** v2.2.0 => ChatPromptBuilder is not export **Error message** <img width="1102" alt="image" src="https://github.com/deepset-ai/haystack/assets/15232298/b9372767-42f5-464c-832f-cca38a00cf60"> </issue> <code> [start of haystack/components/builders/__init__.py] 1 # SPDX-FileCopyrightText: 2022-present deepset GmbH <[email protected]> 2 # 3 # SPDX-License-Identifier: Apache-2.0 4 5 from haystack.components.builders.answer_builder import AnswerBuilder 6 from haystack.components.builders.dynamic_chat_prompt_builder import DynamicChatPromptBuilder 7 from haystack.components.builders.dynamic_prompt_builder import DynamicPromptBuilder 8 from haystack.components.builders.prompt_builder import PromptBuilder 9 10 __all__ = ["AnswerBuilder", "PromptBuilder", "DynamicPromptBuilder", "DynamicChatPromptBuilder"] 11 [end of haystack/components/builders/__init__.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/haystack/components/builders/__init__.py b/haystack/components/builders/__init__.py --- a/haystack/components/builders/__init__.py +++ b/haystack/components/builders/__init__.py @@ -3,8 +3,9 @@ # SPDX-License-Identifier: Apache-2.0 from haystack.components.builders.answer_builder import AnswerBuilder +from haystack.components.builders.chat_prompt_builder import ChatPromptBuilder from haystack.components.builders.dynamic_chat_prompt_builder import DynamicChatPromptBuilder from haystack.components.builders.dynamic_prompt_builder import DynamicPromptBuilder from haystack.components.builders.prompt_builder import PromptBuilder -__all__ = ["AnswerBuilder", "PromptBuilder", "DynamicPromptBuilder", "DynamicChatPromptBuilder"] +__all__ = ["AnswerBuilder", "PromptBuilder", "DynamicPromptBuilder", "DynamicChatPromptBuilder", "ChatPromptBuilder"]
{"golden_diff": "diff --git a/haystack/components/builders/__init__.py b/haystack/components/builders/__init__.py\n--- a/haystack/components/builders/__init__.py\n+++ b/haystack/components/builders/__init__.py\n@@ -3,8 +3,9 @@\n # SPDX-License-Identifier: Apache-2.0\n \n from haystack.components.builders.answer_builder import AnswerBuilder\n+from haystack.components.builders.chat_prompt_builder import ChatPromptBuilder\n from haystack.components.builders.dynamic_chat_prompt_builder import DynamicChatPromptBuilder\n from haystack.components.builders.dynamic_prompt_builder import DynamicPromptBuilder\n from haystack.components.builders.prompt_builder import PromptBuilder\n \n-__all__ = [\"AnswerBuilder\", \"PromptBuilder\", \"DynamicPromptBuilder\", \"DynamicChatPromptBuilder\"]\n+__all__ = [\"AnswerBuilder\", \"PromptBuilder\", \"DynamicPromptBuilder\", \"DynamicChatPromptBuilder\", \"ChatPromptBuilder\"]\n", "issue": "[V2.2.0] ChatPromptBuilder is not export\n**Describe the bug**\r\nv2.2.0 => ChatPromptBuilder is not export\r\n\r\n**Error message**\r\n<img width=\"1102\" alt=\"image\" src=\"https://github.com/deepset-ai/haystack/assets/15232298/b9372767-42f5-464c-832f-cca38a00cf60\">\r\n\r\n\n", "before_files": [{"content": "# SPDX-FileCopyrightText: 2022-present deepset GmbH <[email protected]>\n#\n# SPDX-License-Identifier: Apache-2.0\n\nfrom haystack.components.builders.answer_builder import AnswerBuilder\nfrom haystack.components.builders.dynamic_chat_prompt_builder import DynamicChatPromptBuilder\nfrom haystack.components.builders.dynamic_prompt_builder import DynamicPromptBuilder\nfrom haystack.components.builders.prompt_builder import PromptBuilder\n\n__all__ = [\"AnswerBuilder\", \"PromptBuilder\", \"DynamicPromptBuilder\", \"DynamicChatPromptBuilder\"]\n", "path": "haystack/components/builders/__init__.py"}]}
769
185
gh_patches_debug_36530
rasdani/github-patches
git_diff
getsentry__sentry-python-541
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> 0.12.0 breaks Django function-based middleware Similar to #504, but a different stack trace: AttributeError: 'method-wrapper' object has no attribute '__module__' File "django/core/handlers/exception.py", line 41, in inner response = get_response(request) File "functools.py", line 33, in update_wrapper setattr(wrapper, attr, getattr(wrapped, attr)) According to sentry (kind-of neat how I get this in this case...), the `get_response` object at that point in time is `<sentry_sdk.integrations.django.middleware.AuditMiddleware object at 0x7f37d64d4450>`. This problem only occurs in 0.12.0 and newer, and with Django 1.11.x </issue> <code> [start of sentry_sdk/integrations/django/middleware.py] 1 """ 2 Create spans from Django middleware invocations 3 """ 4 5 from functools import wraps 6 7 from django import VERSION as DJANGO_VERSION 8 9 from sentry_sdk import Hub 10 from sentry_sdk.utils import ContextVar, transaction_from_function 11 12 from sentry_sdk._types import MYPY 13 14 if MYPY: 15 from typing import Any 16 from typing import Callable 17 from typing import TypeVar 18 19 F = TypeVar("F", bound=Callable[..., Any]) 20 21 _import_string_should_wrap_middleware = ContextVar( 22 "import_string_should_wrap_middleware" 23 ) 24 25 if DJANGO_VERSION < (1, 7): 26 import_string_name = "import_by_path" 27 else: 28 import_string_name = "import_string" 29 30 31 def patch_django_middlewares(): 32 # type: () -> None 33 from django.core.handlers import base 34 35 old_import_string = getattr(base, import_string_name) 36 37 def sentry_patched_import_string(dotted_path): 38 # type: (str) -> Any 39 rv = old_import_string(dotted_path) 40 41 if _import_string_should_wrap_middleware.get(None): 42 rv = _wrap_middleware(rv, dotted_path) 43 44 return rv 45 46 setattr(base, import_string_name, sentry_patched_import_string) 47 48 old_load_middleware = base.BaseHandler.load_middleware 49 50 def sentry_patched_load_middleware(self): 51 # type: (base.BaseHandler) -> Any 52 _import_string_should_wrap_middleware.set(True) 53 try: 54 return old_load_middleware(self) 55 finally: 56 _import_string_should_wrap_middleware.set(False) 57 58 base.BaseHandler.load_middleware = sentry_patched_load_middleware 59 60 61 def _wrap_middleware(middleware, middleware_name): 62 # type: (Any, str) -> Any 63 from sentry_sdk.integrations.django import DjangoIntegration 64 65 def _get_wrapped_method(old_method): 66 # type: (F) -> F 67 @wraps(old_method) 68 def sentry_wrapped_method(*args, **kwargs): 69 # type: (*Any, **Any) -> Any 70 hub = Hub.current 71 integration = hub.get_integration(DjangoIntegration) 72 if integration is None or not integration.middleware_spans: 73 return old_method(*args, **kwargs) 74 75 function_name = transaction_from_function(old_method) 76 77 description = middleware_name 78 function_basename = getattr(old_method, "__name__", None) 79 if function_basename: 80 description = "{}.{}".format(description, function_basename) 81 82 with hub.start_span( 83 op="django.middleware", description=description 84 ) as span: 85 span.set_tag("django.function_name", function_name) 86 span.set_tag("django.middleware_name", middleware_name) 87 return old_method(*args, **kwargs) 88 89 return sentry_wrapped_method # type: ignore 90 91 class SentryWrappingMiddleware(object): 92 def __init__(self, *args, **kwargs): 93 # type: (*Any, **Any) -> None 94 self._inner = middleware(*args, **kwargs) 95 self._call_method = None 96 97 # We need correct behavior for `hasattr()`, which we can only determine 98 # when we have an instance of the middleware we're wrapping. 99 def __getattr__(self, method_name): 100 # type: (str) -> Any 101 if method_name not in ( 102 "process_request", 103 "process_view", 104 "process_template_response", 105 "process_response", 106 "process_exception", 107 ): 108 raise AttributeError() 109 110 old_method = getattr(self._inner, method_name) 111 rv = _get_wrapped_method(old_method) 112 self.__dict__[method_name] = rv 113 return rv 114 115 def __call__(self, *args, **kwargs): 116 # type: (*Any, **Any) -> Any 117 f = self._call_method 118 if f is None: 119 self._call_method = f = _get_wrapped_method(self._inner.__call__) 120 return f(*args, **kwargs) 121 122 if hasattr(middleware, "__name__"): 123 SentryWrappingMiddleware.__name__ = middleware.__name__ 124 125 return SentryWrappingMiddleware 126 [end of sentry_sdk/integrations/django/middleware.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/sentry_sdk/integrations/django/middleware.py b/sentry_sdk/integrations/django/middleware.py --- a/sentry_sdk/integrations/django/middleware.py +++ b/sentry_sdk/integrations/django/middleware.py @@ -7,7 +7,11 @@ from django import VERSION as DJANGO_VERSION from sentry_sdk import Hub -from sentry_sdk.utils import ContextVar, transaction_from_function +from sentry_sdk.utils import ( + ContextVar, + transaction_from_function, + capture_internal_exceptions, +) from sentry_sdk._types import MYPY @@ -64,29 +68,36 @@ def _get_wrapped_method(old_method): # type: (F) -> F - @wraps(old_method) - def sentry_wrapped_method(*args, **kwargs): - # type: (*Any, **Any) -> Any - hub = Hub.current - integration = hub.get_integration(DjangoIntegration) - if integration is None or not integration.middleware_spans: - return old_method(*args, **kwargs) - - function_name = transaction_from_function(old_method) - - description = middleware_name - function_basename = getattr(old_method, "__name__", None) - if function_basename: - description = "{}.{}".format(description, function_basename) - - with hub.start_span( - op="django.middleware", description=description - ) as span: - span.set_tag("django.function_name", function_name) - span.set_tag("django.middleware_name", middleware_name) - return old_method(*args, **kwargs) - - return sentry_wrapped_method # type: ignore + with capture_internal_exceptions(): + + def sentry_wrapped_method(*args, **kwargs): + # type: (*Any, **Any) -> Any + hub = Hub.current + integration = hub.get_integration(DjangoIntegration) + if integration is None or not integration.middleware_spans: + return old_method(*args, **kwargs) + + function_name = transaction_from_function(old_method) + + description = middleware_name + function_basename = getattr(old_method, "__name__", None) + if function_basename: + description = "{}.{}".format(description, function_basename) + + with hub.start_span( + op="django.middleware", description=description + ) as span: + span.set_tag("django.function_name", function_name) + span.set_tag("django.middleware_name", middleware_name) + return old_method(*args, **kwargs) + + try: + # fails for __call__ of function on Python 2 (see py2.7-django-1.11) + return wraps(old_method)(sentry_wrapped_method) # type: ignore + except Exception: + return sentry_wrapped_method # type: ignore + + return old_method class SentryWrappingMiddleware(object): def __init__(self, *args, **kwargs):
{"golden_diff": "diff --git a/sentry_sdk/integrations/django/middleware.py b/sentry_sdk/integrations/django/middleware.py\n--- a/sentry_sdk/integrations/django/middleware.py\n+++ b/sentry_sdk/integrations/django/middleware.py\n@@ -7,7 +7,11 @@\n from django import VERSION as DJANGO_VERSION\n \n from sentry_sdk import Hub\n-from sentry_sdk.utils import ContextVar, transaction_from_function\n+from sentry_sdk.utils import (\n+ ContextVar,\n+ transaction_from_function,\n+ capture_internal_exceptions,\n+)\n \n from sentry_sdk._types import MYPY\n \n@@ -64,29 +68,36 @@\n \n def _get_wrapped_method(old_method):\n # type: (F) -> F\n- @wraps(old_method)\n- def sentry_wrapped_method(*args, **kwargs):\n- # type: (*Any, **Any) -> Any\n- hub = Hub.current\n- integration = hub.get_integration(DjangoIntegration)\n- if integration is None or not integration.middleware_spans:\n- return old_method(*args, **kwargs)\n-\n- function_name = transaction_from_function(old_method)\n-\n- description = middleware_name\n- function_basename = getattr(old_method, \"__name__\", None)\n- if function_basename:\n- description = \"{}.{}\".format(description, function_basename)\n-\n- with hub.start_span(\n- op=\"django.middleware\", description=description\n- ) as span:\n- span.set_tag(\"django.function_name\", function_name)\n- span.set_tag(\"django.middleware_name\", middleware_name)\n- return old_method(*args, **kwargs)\n-\n- return sentry_wrapped_method # type: ignore\n+ with capture_internal_exceptions():\n+\n+ def sentry_wrapped_method(*args, **kwargs):\n+ # type: (*Any, **Any) -> Any\n+ hub = Hub.current\n+ integration = hub.get_integration(DjangoIntegration)\n+ if integration is None or not integration.middleware_spans:\n+ return old_method(*args, **kwargs)\n+\n+ function_name = transaction_from_function(old_method)\n+\n+ description = middleware_name\n+ function_basename = getattr(old_method, \"__name__\", None)\n+ if function_basename:\n+ description = \"{}.{}\".format(description, function_basename)\n+\n+ with hub.start_span(\n+ op=\"django.middleware\", description=description\n+ ) as span:\n+ span.set_tag(\"django.function_name\", function_name)\n+ span.set_tag(\"django.middleware_name\", middleware_name)\n+ return old_method(*args, **kwargs)\n+\n+ try:\n+ # fails for __call__ of function on Python 2 (see py2.7-django-1.11)\n+ return wraps(old_method)(sentry_wrapped_method) # type: ignore\n+ except Exception:\n+ return sentry_wrapped_method # type: ignore\n+\n+ return old_method\n \n class SentryWrappingMiddleware(object):\n def __init__(self, *args, **kwargs):\n", "issue": "0.12.0 breaks Django function-based middleware\nSimilar to #504, but a different stack trace:\r\n\r\n AttributeError: 'method-wrapper' object has no attribute '__module__'\r\n File \"django/core/handlers/exception.py\", line 41, in inner\r\n response = get_response(request)\r\n File \"functools.py\", line 33, in update_wrapper\r\n setattr(wrapper, attr, getattr(wrapped, attr))\r\n\r\nAccording to sentry (kind-of neat how I get this in this case...), the `get_response` object at that point in time is `<sentry_sdk.integrations.django.middleware.AuditMiddleware object at 0x7f37d64d4450>`.\r\n\r\nThis problem only occurs in 0.12.0 and newer, and with Django 1.11.x\n", "before_files": [{"content": "\"\"\"\nCreate spans from Django middleware invocations\n\"\"\"\n\nfrom functools import wraps\n\nfrom django import VERSION as DJANGO_VERSION\n\nfrom sentry_sdk import Hub\nfrom sentry_sdk.utils import ContextVar, transaction_from_function\n\nfrom sentry_sdk._types import MYPY\n\nif MYPY:\n from typing import Any\n from typing import Callable\n from typing import TypeVar\n\n F = TypeVar(\"F\", bound=Callable[..., Any])\n\n_import_string_should_wrap_middleware = ContextVar(\n \"import_string_should_wrap_middleware\"\n)\n\nif DJANGO_VERSION < (1, 7):\n import_string_name = \"import_by_path\"\nelse:\n import_string_name = \"import_string\"\n\n\ndef patch_django_middlewares():\n # type: () -> None\n from django.core.handlers import base\n\n old_import_string = getattr(base, import_string_name)\n\n def sentry_patched_import_string(dotted_path):\n # type: (str) -> Any\n rv = old_import_string(dotted_path)\n\n if _import_string_should_wrap_middleware.get(None):\n rv = _wrap_middleware(rv, dotted_path)\n\n return rv\n\n setattr(base, import_string_name, sentry_patched_import_string)\n\n old_load_middleware = base.BaseHandler.load_middleware\n\n def sentry_patched_load_middleware(self):\n # type: (base.BaseHandler) -> Any\n _import_string_should_wrap_middleware.set(True)\n try:\n return old_load_middleware(self)\n finally:\n _import_string_should_wrap_middleware.set(False)\n\n base.BaseHandler.load_middleware = sentry_patched_load_middleware\n\n\ndef _wrap_middleware(middleware, middleware_name):\n # type: (Any, str) -> Any\n from sentry_sdk.integrations.django import DjangoIntegration\n\n def _get_wrapped_method(old_method):\n # type: (F) -> F\n @wraps(old_method)\n def sentry_wrapped_method(*args, **kwargs):\n # type: (*Any, **Any) -> Any\n hub = Hub.current\n integration = hub.get_integration(DjangoIntegration)\n if integration is None or not integration.middleware_spans:\n return old_method(*args, **kwargs)\n\n function_name = transaction_from_function(old_method)\n\n description = middleware_name\n function_basename = getattr(old_method, \"__name__\", None)\n if function_basename:\n description = \"{}.{}\".format(description, function_basename)\n\n with hub.start_span(\n op=\"django.middleware\", description=description\n ) as span:\n span.set_tag(\"django.function_name\", function_name)\n span.set_tag(\"django.middleware_name\", middleware_name)\n return old_method(*args, **kwargs)\n\n return sentry_wrapped_method # type: ignore\n\n class SentryWrappingMiddleware(object):\n def __init__(self, *args, **kwargs):\n # type: (*Any, **Any) -> None\n self._inner = middleware(*args, **kwargs)\n self._call_method = None\n\n # We need correct behavior for `hasattr()`, which we can only determine\n # when we have an instance of the middleware we're wrapping.\n def __getattr__(self, method_name):\n # type: (str) -> Any\n if method_name not in (\n \"process_request\",\n \"process_view\",\n \"process_template_response\",\n \"process_response\",\n \"process_exception\",\n ):\n raise AttributeError()\n\n old_method = getattr(self._inner, method_name)\n rv = _get_wrapped_method(old_method)\n self.__dict__[method_name] = rv\n return rv\n\n def __call__(self, *args, **kwargs):\n # type: (*Any, **Any) -> Any\n f = self._call_method\n if f is None:\n self._call_method = f = _get_wrapped_method(self._inner.__call__)\n return f(*args, **kwargs)\n\n if hasattr(middleware, \"__name__\"):\n SentryWrappingMiddleware.__name__ = middleware.__name__\n\n return SentryWrappingMiddleware\n", "path": "sentry_sdk/integrations/django/middleware.py"}]}
1,898
672
gh_patches_debug_28409
rasdani/github-patches
git_diff
mampfes__hacs_waste_collection_schedule-182
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> recycleapp_be not working for some addresses when I enter my address into the configuration.yaml I receive this error on restart: ``` fetch failed for source Recycle!: Traceback (most recent call last): File "/config/custom_components/waste_collection_schedule/waste_collection_schedule/scraper.py", line 116, in fetch entries = self._source.fetch() File "/config/custom_components/waste_collection_schedule/waste_collection_schedule/source/recycleapp_be.py", line 79, in fetch entries.append(Collection(date, item["fraction"]["name"]["en"])) KeyError: 'name' ``` when I use the example address or some other addresses everything works fine. Is it a problem with my city? Because other addresses of this city also don't work, even though those addresses work on [Recycle!](https://recycleapp.be/home). this is what I have in configuration.yaml ``` waste_collection_schedule: sources: - name: recycleapp_be args: postcode: 3001 street: Waversebaan house_number: 276 ``` </issue> <code> [start of custom_components/waste_collection_schedule/waste_collection_schedule/source/recycleapp_be.py] 1 import logging 2 from datetime import datetime, timedelta 3 4 import requests 5 from waste_collection_schedule import Collection # type: ignore[attr-defined] 6 7 TITLE = "Recycle!" 8 DESCRIPTION = "Source for RecycleApp.be" 9 URL = "https://www.recycleapp.be" 10 TEST_CASES = { 11 "1140 Evere, Bazellaan 1": { 12 "postcode": 1140, 13 "street": "Bazellaan", 14 "house_number": 1, 15 } 16 } 17 18 _LOGGER = logging.getLogger(__name__) 19 20 21 class Source: 22 def __init__(self, postcode, street, house_number): 23 self._postcode = postcode 24 self._street = street 25 self._house_number = house_number 26 27 def fetch(self): 28 url = "https://recycleapp.be/api/app/v1" 29 headers = { 30 "x-secret": "Crgja3EGWe8jdapyr4EEoMBgZACYYjRRcRpaMQrLDW9HJBvmgkfGQyYqLgeXPavAGvnJqkV87PBB2b8zx43q46sUgzqio4yRZbABhtKeagkVKypTEDjKfPgGycjLyJTtLHYpzwJgp4YmmCuJZN9ZmJY8CGEoFs8MKfdJpU9RjkEVfngmmk2LYD4QzFegLNKUbcCeAdEW", 31 "x-consumer": "recycleapp.be", 32 "User-Agent": "", 33 "Authorization": "", 34 } 35 r = requests.get(f"{url}/access-token", headers=headers) 36 headers["Authorization"] = r.json()["accessToken"] 37 38 params = {"q": self._postcode} 39 r = requests.get(f"{url}/zipcodes", params=params, headers=headers) 40 if r.status_code != 200: 41 _LOGGER.error("Get zip code failed") 42 return [] 43 zipcodeId = r.json()["items"][0]["id"] 44 45 params = {"q": self._street, "zipcodes": zipcodeId} 46 r = requests.get(f"{url}/streets", params=params, headers=headers) 47 if r.status_code != 200: 48 _LOGGER.error("Get street id failed") 49 return [] 50 51 for item in r.json()["items"]: 52 if item["name"] == self._street: 53 streetId = item["id"] 54 if streetId is None: 55 streetId = r.json()["items"][0]["id"] 56 57 now = datetime.now() 58 fromDate = now.strftime("%Y-%m-%d") 59 untilDate = (now + timedelta(days=365)).strftime("%Y-%m-%d") 60 params = { 61 "zipcodeId": zipcodeId, 62 "streetId": streetId, 63 "houseNumber": self._house_number, 64 "fromDate": fromDate, 65 "untilDate": untilDate, 66 # "size":100, 67 } 68 r = requests.get(f"{url}/collections", params=params, headers=headers) 69 if r.status_code != 200: 70 _LOGGER.error("Get data failed") 71 return [] 72 73 entries = [] 74 for item in r.json()["items"]: 75 if "exception" in item and "replacedBy" in item["exception"]: 76 continue 77 78 date = datetime.strptime(item["timestamp"], "%Y-%m-%dT%H:%M:%S.000Z").date() 79 entries.append(Collection(date, item["fraction"]["name"]["en"])) 80 return entries 81 [end of custom_components/waste_collection_schedule/waste_collection_schedule/source/recycleapp_be.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/custom_components/waste_collection_schedule/waste_collection_schedule/source/recycleapp_be.py b/custom_components/waste_collection_schedule/waste_collection_schedule/source/recycleapp_be.py --- a/custom_components/waste_collection_schedule/waste_collection_schedule/source/recycleapp_be.py +++ b/custom_components/waste_collection_schedule/waste_collection_schedule/source/recycleapp_be.py @@ -12,17 +12,29 @@ "postcode": 1140, "street": "Bazellaan", "house_number": 1, - } + }, + "3001, Waversebaan 276 with events": { + "postcode": 3001, + "street": "Waversebaan", + "house_number": 276, + }, + "3001, Waversebaan 276 without events": { + "postcode": 3001, + "street": "Waversebaan", + "house_number": 276, + "add_events": False, + }, } _LOGGER = logging.getLogger(__name__) class Source: - def __init__(self, postcode, street, house_number): + def __init__(self, postcode, street, house_number, add_events=True): self._postcode = postcode self._street = street self._house_number = house_number + self._add_events = add_events def fetch(self): url = "https://recycleapp.be/api/app/v1" @@ -76,5 +88,9 @@ continue date = datetime.strptime(item["timestamp"], "%Y-%m-%dT%H:%M:%S.000Z").date() - entries.append(Collection(date, item["fraction"]["name"]["en"])) + if item["type"] == "collection": + entries.append(Collection(date, item["fraction"]["name"]["en"])) + elif item["type"] == "event" and self._add_events: + entries.append(Collection(date, item["event"]["title"]["en"])) + return entries
{"golden_diff": "diff --git a/custom_components/waste_collection_schedule/waste_collection_schedule/source/recycleapp_be.py b/custom_components/waste_collection_schedule/waste_collection_schedule/source/recycleapp_be.py\n--- a/custom_components/waste_collection_schedule/waste_collection_schedule/source/recycleapp_be.py\n+++ b/custom_components/waste_collection_schedule/waste_collection_schedule/source/recycleapp_be.py\n@@ -12,17 +12,29 @@\n \"postcode\": 1140,\n \"street\": \"Bazellaan\",\n \"house_number\": 1,\n- }\n+ },\n+ \"3001, Waversebaan 276 with events\": {\n+ \"postcode\": 3001,\n+ \"street\": \"Waversebaan\",\n+ \"house_number\": 276,\n+ },\n+ \"3001, Waversebaan 276 without events\": {\n+ \"postcode\": 3001,\n+ \"street\": \"Waversebaan\",\n+ \"house_number\": 276,\n+ \"add_events\": False,\n+ },\n }\n \n _LOGGER = logging.getLogger(__name__)\n \n \n class Source:\n- def __init__(self, postcode, street, house_number):\n+ def __init__(self, postcode, street, house_number, add_events=True):\n self._postcode = postcode\n self._street = street\n self._house_number = house_number\n+ self._add_events = add_events\n \n def fetch(self):\n url = \"https://recycleapp.be/api/app/v1\"\n@@ -76,5 +88,9 @@\n continue\n \n date = datetime.strptime(item[\"timestamp\"], \"%Y-%m-%dT%H:%M:%S.000Z\").date()\n- entries.append(Collection(date, item[\"fraction\"][\"name\"][\"en\"]))\n+ if item[\"type\"] == \"collection\":\n+ entries.append(Collection(date, item[\"fraction\"][\"name\"][\"en\"]))\n+ elif item[\"type\"] == \"event\" and self._add_events:\n+ entries.append(Collection(date, item[\"event\"][\"title\"][\"en\"]))\n+\n return entries\n", "issue": "recycleapp_be not working for some addresses\nwhen I enter my address into the configuration.yaml I receive this error on restart:\r\n```\r\nfetch failed for source Recycle!: Traceback (most recent call last): File \"/config/custom_components/waste_collection_schedule/waste_collection_schedule/scraper.py\", \r\nline 116, in fetch entries = self._source.fetch() File \"/config/custom_components/waste_collection_schedule/waste_collection_schedule/source/recycleapp_be.py\", \r\nline 79, in fetch entries.append(Collection(date, item[\"fraction\"][\"name\"][\"en\"])) KeyError: 'name'\r\n```\r\nwhen I use the example address or some other addresses everything works fine. Is it a problem with my city? Because other addresses of this city also don't work, even though those addresses work on [Recycle!](https://recycleapp.be/home).\r\nthis is what I have in configuration.yaml\r\n```\r\nwaste_collection_schedule:\r\n sources:\r\n - name: recycleapp_be\r\n args:\r\n postcode: 3001\r\n street: Waversebaan\r\n house_number: 276\r\n```\n", "before_files": [{"content": "import logging\nfrom datetime import datetime, timedelta\n\nimport requests\nfrom waste_collection_schedule import Collection # type: ignore[attr-defined]\n\nTITLE = \"Recycle!\"\nDESCRIPTION = \"Source for RecycleApp.be\"\nURL = \"https://www.recycleapp.be\"\nTEST_CASES = {\n \"1140 Evere, Bazellaan 1\": {\n \"postcode\": 1140,\n \"street\": \"Bazellaan\",\n \"house_number\": 1,\n }\n}\n\n_LOGGER = logging.getLogger(__name__)\n\n\nclass Source:\n def __init__(self, postcode, street, house_number):\n self._postcode = postcode\n self._street = street\n self._house_number = house_number\n\n def fetch(self):\n url = \"https://recycleapp.be/api/app/v1\"\n headers = {\n \"x-secret\": \"Crgja3EGWe8jdapyr4EEoMBgZACYYjRRcRpaMQrLDW9HJBvmgkfGQyYqLgeXPavAGvnJqkV87PBB2b8zx43q46sUgzqio4yRZbABhtKeagkVKypTEDjKfPgGycjLyJTtLHYpzwJgp4YmmCuJZN9ZmJY8CGEoFs8MKfdJpU9RjkEVfngmmk2LYD4QzFegLNKUbcCeAdEW\",\n \"x-consumer\": \"recycleapp.be\",\n \"User-Agent\": \"\",\n \"Authorization\": \"\",\n }\n r = requests.get(f\"{url}/access-token\", headers=headers)\n headers[\"Authorization\"] = r.json()[\"accessToken\"]\n\n params = {\"q\": self._postcode}\n r = requests.get(f\"{url}/zipcodes\", params=params, headers=headers)\n if r.status_code != 200:\n _LOGGER.error(\"Get zip code failed\")\n return []\n zipcodeId = r.json()[\"items\"][0][\"id\"]\n\n params = {\"q\": self._street, \"zipcodes\": zipcodeId}\n r = requests.get(f\"{url}/streets\", params=params, headers=headers)\n if r.status_code != 200:\n _LOGGER.error(\"Get street id failed\")\n return []\n\n for item in r.json()[\"items\"]:\n if item[\"name\"] == self._street:\n streetId = item[\"id\"]\n if streetId is None:\n streetId = r.json()[\"items\"][0][\"id\"]\n\n now = datetime.now()\n fromDate = now.strftime(\"%Y-%m-%d\")\n untilDate = (now + timedelta(days=365)).strftime(\"%Y-%m-%d\")\n params = {\n \"zipcodeId\": zipcodeId,\n \"streetId\": streetId,\n \"houseNumber\": self._house_number,\n \"fromDate\": fromDate,\n \"untilDate\": untilDate,\n # \"size\":100,\n }\n r = requests.get(f\"{url}/collections\", params=params, headers=headers)\n if r.status_code != 200:\n _LOGGER.error(\"Get data failed\")\n return []\n\n entries = []\n for item in r.json()[\"items\"]:\n if \"exception\" in item and \"replacedBy\" in item[\"exception\"]:\n continue\n\n date = datetime.strptime(item[\"timestamp\"], \"%Y-%m-%dT%H:%M:%S.000Z\").date()\n entries.append(Collection(date, item[\"fraction\"][\"name\"][\"en\"]))\n return entries\n", "path": "custom_components/waste_collection_schedule/waste_collection_schedule/source/recycleapp_be.py"}]}
1,721
466
gh_patches_debug_31901
rasdani/github-patches
git_diff
fidals__shopelectro-753
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> Add sending mail action for given orders to admin panel We have got connection issue with smtp and didn't notify 5 client https://sentry.fidals.com/fidals/shopelectro/issues/36/ </issue> <code> [start of shopelectro/admin.py] 1 from django.contrib import admin 2 from django.contrib.admin.widgets import FilteredSelectMultiple 3 from django.contrib.redirects.models import Redirect 4 from django.db import models as django_models 5 from django.urls import reverse 6 from django.utils.html import format_html 7 from django.utils.translation import ugettext_lazy as _ 8 9 from django_select2.forms import ModelSelect2Widget 10 11 from ecommerce.models import Position 12 from pages.models import CustomPage, FlatPage, PageTemplate 13 from generic_admin import inlines, mixins, models, sites 14 15 from shopelectro import models as se_models 16 from shopelectro.views.admin import TableEditor 17 18 19 class SEAdminSite(sites.SiteWithTableEditor): 20 21 site_header = 'Shopelectro administration' 22 table_editor_view = TableEditor 23 24 25 def prepare_has_filter_queryset(value, db_table, queryset): 26 if not value: 27 return 28 29 query = '{}__tags__isnull'.format(db_table) 30 31 # Use brackets, because `Explicit is better than implicit`. 32 return queryset.filter(**{query: value != 'yes'}) 33 34 35 class HasTagsFilter(admin.SimpleListFilter): 36 37 product_model = se_models.Product 38 title = _('has tags') 39 parameter_name = 'has_tags' 40 41 def lookups(self, request, model_admin): 42 return ( 43 ('yes', _('Has tags')), 44 ('no', _('Has no tags')), 45 ) 46 47 def queryset(self, request, queryset): 48 return prepare_has_filter_queryset( 49 self.value(), 50 self.product_model._meta.db_table, 51 queryset 52 ) 53 54 55 class HasCategoryFilter(admin.SimpleListFilter): 56 57 product_model = se_models.Product 58 title = _('has category') 59 parameter_name = 'has_category' 60 61 def lookups(self, request, model_admin): 62 return ( 63 ('yes', _('Has category')), 64 ('no', _('Has no category')), 65 ) 66 67 def queryset(self, request, queryset): 68 return prepare_has_filter_queryset( 69 self.value(), 70 self.product_model._meta.db_table, 71 queryset 72 ) 73 74 75 class TagInline(admin.StackedInline): 76 77 model = se_models.Tag 78 extra = 0 79 80 81 class CategoryInline(inlines.CategoryInline): 82 83 model = se_models.Category 84 85 fieldsets = ((None, { 86 'classes': ('primary-chars', ), 87 'fields': ( 88 ('name', 'id'), 89 ('parent', 'correct_parent_id'), 90 ) 91 }),) 92 93 94 class ProductInline(inlines.ProductInline): 95 96 model = se_models.Product 97 98 formfield_overrides = { 99 django_models.ManyToManyField: { 100 'widget': FilteredSelectMultiple(verbose_name='Tags', is_stacked=False) 101 }, 102 } 103 104 fieldsets = ((None, { 105 'classes': ('primary-chars', ), 106 'fields': ( 107 ('name', 'id', 'vendor_code'), 108 ('category', 'correct_category_id'), 109 ('price', 'in_stock'), 110 'is_popular', 111 ('purchase_price', 'wholesale_small'), 112 ('wholesale_medium', 'wholesale_large'), 113 'tags', 114 ) 115 }),) 116 117 def formfield_for_dbfield(self, db_field, **kwargs): 118 if db_field.name == 'category': 119 kwargs['widget'] = ModelSelect2Widget( 120 model=se_models.Category, 121 search_fields=[ 122 'name__icontains', 123 'pk__startswith', 124 ], 125 ) 126 return super(ProductInline, self).formfield_for_dbfield( 127 db_field, 128 **kwargs, 129 ) 130 131 132 class PositionInline(admin.StackedInline): 133 134 model = Position 135 136 137 class CategoryPageAdmin(models.CategoryPageAdmin): 138 139 add = True 140 delete = False 141 inlines = [CategoryInline, inlines.ImageInline] 142 143 def get_queryset(self, request): 144 return ( 145 super(CategoryPageAdmin, self) 146 .get_queryset(request) 147 .select_related('shopelectro_category') 148 ) 149 150 151 class ProductPageAdmin(models.ProductPageAdmin): 152 153 add = False 154 delete = False 155 category_page_model = se_models.CategoryPage 156 list_filter = [*models.ProductPageAdmin.list_filter, HasTagsFilter, HasCategoryFilter] 157 inlines = [ProductInline, inlines.ImageInline] 158 search_fields = [ 159 'shopelectro_product__vendor_code', 'name', 'slug', 160 ] 161 162 def model_id(self, obj): 163 return obj.model.vendor_code 164 165 model_id.short_description = _('Vendor code') 166 model_id.admin_order_field = 'shopelectro_product__vendor_code' 167 168 def get_queryset(self, request): 169 return ( 170 super(ProductPageAdmin, self) 171 .get_queryset(request) 172 .select_related('shopelectro_product') 173 ) 174 175 176 class ProductFeedbackPageAdmin(admin.ModelAdmin): 177 178 add = False 179 delete = False 180 list_filter = ['rating'] 181 list_display = ['rating', 'name', 'dignities', 'limitations', 'general', 'links'] 182 183 def links(self, obj): 184 return format_html( 185 ''' 186 <a href="{url}" class="field-link" title="Посмотреть на сайте" target="_blank"> 187 <i class="fa fa-link" aria-hidden="true"></i> 188 </a> 189 '''.format(url=obj.product.url)) 190 191 links.short_description = _('Link') 192 193 def get_queryset(self, request): 194 return ( 195 super(ProductFeedbackPageAdmin, self) 196 .get_queryset(request) 197 .select_related('product') 198 ) 199 200 201 class TagGroupAdmin(admin.ModelAdmin): 202 203 list_display = ['id', 'name', 'position', 'count_tags'] 204 list_display_links = ['name'] 205 206 inlines = [TagInline] 207 208 def get_queryset(self, request): 209 return super(TagGroupAdmin, self).get_queryset(request).prefetch_related('tags') 210 211 def count_tags(self, obj): 212 return obj.tags.count() 213 214 215 class TagAdmin(admin.ModelAdmin): 216 217 search_fields = ['id', 'name'] 218 list_display = ['id', 'name', 'position', 'custom_group'] 219 list_display_links = ['name'] 220 221 def get_queryset(self, request): 222 return super(TagAdmin, self).get_queryset(request).prefetch_related('group') 223 224 def custom_group(self, obj): 225 group = obj.group 226 227 return format_html( 228 '<a href="{url}">{group}</a>', 229 group=group, 230 url=reverse('admin:shopelectro_taggroup_change', args=(group.id, )) 231 ) 232 233 custom_group.admin_order_field = 'group' 234 custom_group.short_description = _('Group') 235 236 237 class OrderAdmin(mixins.PermissionsControl): 238 239 add = False 240 inlines = [PositionInline] 241 list_display = ['id_', 'name', 'email', 'phone', 'total_price', 'payment_type', 'paid'] 242 search_fields = ['name', 'email', 'phone'] 243 list_display_links = ['name'] 244 245 def id_(self, obj): 246 return obj.fake_order_number 247 248 249 se_admin = SEAdminSite(name='se_admin') 250 se_admin.register(CustomPage, models.CustomPageAdmin) 251 se_admin.register(FlatPage, models.FlatPageAdmin) 252 se_admin.register(PageTemplate, models.CustomPageTemplateAdmin) 253 254 se_admin.register(se_models.CategoryPage, CategoryPageAdmin) 255 se_admin.register(se_models.Order, OrderAdmin) 256 se_admin.register(se_models.ProductPage, ProductPageAdmin) 257 se_admin.register(se_models.ProductFeedback, ProductFeedbackPageAdmin) 258 se_admin.register(se_models.TagGroup, TagGroupAdmin) 259 se_admin.register(se_models.Tag, TagAdmin) 260 261 se_admin.register(Redirect) 262 [end of shopelectro/admin.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/shopelectro/admin.py b/shopelectro/admin.py --- a/shopelectro/admin.py +++ b/shopelectro/admin.py @@ -1,3 +1,4 @@ +from django.conf import settings from django.contrib import admin from django.contrib.admin.widgets import FilteredSelectMultiple from django.contrib.redirects.models import Redirect @@ -8,6 +9,7 @@ from django_select2.forms import ModelSelect2Widget +from ecommerce import mailer from ecommerce.models import Position from pages.models import CustomPage, FlatPage, PageTemplate from generic_admin import inlines, mixins, models, sites @@ -234,6 +236,27 @@ custom_group.short_description = _('Group') +def send_order_emails(admin_model, request, order_qs): + for order in order_qs: + context = {'shop': settings.SHOP} + if order.email: + mailer.send_order( + subject=settings.EMAIL_SUBJECTS['order'], + order=order, + extra_context=context, + ) + else: + mailer.send_order( + subject=settings.EMAIL_SUBJECTS['one_click'], + order=order, + to_customer=False, + extra_context=context, + ) + + +send_order_emails.short_description = _('Sends email notifications about placed orders') + + class OrderAdmin(mixins.PermissionsControl): add = False @@ -241,6 +264,7 @@ list_display = ['id_', 'name', 'email', 'phone', 'total_price', 'payment_type', 'paid'] search_fields = ['name', 'email', 'phone'] list_display_links = ['name'] + actions = [send_order_emails] def id_(self, obj): return obj.fake_order_number
{"golden_diff": "diff --git a/shopelectro/admin.py b/shopelectro/admin.py\n--- a/shopelectro/admin.py\n+++ b/shopelectro/admin.py\n@@ -1,3 +1,4 @@\n+from django.conf import settings\n from django.contrib import admin\n from django.contrib.admin.widgets import FilteredSelectMultiple\n from django.contrib.redirects.models import Redirect\n@@ -8,6 +9,7 @@\n \n from django_select2.forms import ModelSelect2Widget\n \n+from ecommerce import mailer\n from ecommerce.models import Position\n from pages.models import CustomPage, FlatPage, PageTemplate\n from generic_admin import inlines, mixins, models, sites\n@@ -234,6 +236,27 @@\n custom_group.short_description = _('Group')\n \n \n+def send_order_emails(admin_model, request, order_qs):\n+ for order in order_qs:\n+ context = {'shop': settings.SHOP}\n+ if order.email:\n+ mailer.send_order(\n+ subject=settings.EMAIL_SUBJECTS['order'],\n+ order=order,\n+ extra_context=context,\n+ )\n+ else:\n+ mailer.send_order(\n+ subject=settings.EMAIL_SUBJECTS['one_click'],\n+ order=order,\n+ to_customer=False,\n+ extra_context=context,\n+ )\n+\n+\n+send_order_emails.short_description = _('Sends email notifications about placed orders')\n+\n+\n class OrderAdmin(mixins.PermissionsControl):\n \n add = False\n@@ -241,6 +264,7 @@\n list_display = ['id_', 'name', 'email', 'phone', 'total_price', 'payment_type', 'paid']\n search_fields = ['name', 'email', 'phone']\n list_display_links = ['name']\n+ actions = [send_order_emails]\n \n def id_(self, obj):\n return obj.fake_order_number\n", "issue": "Add sending mail action for given orders to admin panel\nWe have got connection issue with smtp and didn't notify 5 client\r\nhttps://sentry.fidals.com/fidals/shopelectro/issues/36/\n", "before_files": [{"content": "from django.contrib import admin\nfrom django.contrib.admin.widgets import FilteredSelectMultiple\nfrom django.contrib.redirects.models import Redirect\nfrom django.db import models as django_models\nfrom django.urls import reverse\nfrom django.utils.html import format_html\nfrom django.utils.translation import ugettext_lazy as _\n\nfrom django_select2.forms import ModelSelect2Widget\n\nfrom ecommerce.models import Position\nfrom pages.models import CustomPage, FlatPage, PageTemplate\nfrom generic_admin import inlines, mixins, models, sites\n\nfrom shopelectro import models as se_models\nfrom shopelectro.views.admin import TableEditor\n\n\nclass SEAdminSite(sites.SiteWithTableEditor):\n\n site_header = 'Shopelectro administration'\n table_editor_view = TableEditor\n\n\ndef prepare_has_filter_queryset(value, db_table, queryset):\n if not value:\n return\n\n query = '{}__tags__isnull'.format(db_table)\n\n # Use brackets, because `Explicit is better than implicit`.\n return queryset.filter(**{query: value != 'yes'})\n\n\nclass HasTagsFilter(admin.SimpleListFilter):\n\n product_model = se_models.Product\n title = _('has tags')\n parameter_name = 'has_tags'\n\n def lookups(self, request, model_admin):\n return (\n ('yes', _('Has tags')),\n ('no', _('Has no tags')),\n )\n\n def queryset(self, request, queryset):\n return prepare_has_filter_queryset(\n self.value(),\n self.product_model._meta.db_table,\n queryset\n )\n\n\nclass HasCategoryFilter(admin.SimpleListFilter):\n\n product_model = se_models.Product\n title = _('has category')\n parameter_name = 'has_category'\n\n def lookups(self, request, model_admin):\n return (\n ('yes', _('Has category')),\n ('no', _('Has no category')),\n )\n\n def queryset(self, request, queryset):\n return prepare_has_filter_queryset(\n self.value(),\n self.product_model._meta.db_table,\n queryset\n )\n\n\nclass TagInline(admin.StackedInline):\n\n model = se_models.Tag\n extra = 0\n\n\nclass CategoryInline(inlines.CategoryInline):\n\n model = se_models.Category\n\n fieldsets = ((None, {\n 'classes': ('primary-chars', ),\n 'fields': (\n ('name', 'id'),\n ('parent', 'correct_parent_id'),\n )\n }),)\n\n\nclass ProductInline(inlines.ProductInline):\n\n model = se_models.Product\n\n formfield_overrides = {\n django_models.ManyToManyField: {\n 'widget': FilteredSelectMultiple(verbose_name='Tags', is_stacked=False)\n },\n }\n\n fieldsets = ((None, {\n 'classes': ('primary-chars', ),\n 'fields': (\n ('name', 'id', 'vendor_code'),\n ('category', 'correct_category_id'),\n ('price', 'in_stock'),\n 'is_popular',\n ('purchase_price', 'wholesale_small'),\n ('wholesale_medium', 'wholesale_large'),\n 'tags',\n )\n }),)\n\n def formfield_for_dbfield(self, db_field, **kwargs):\n if db_field.name == 'category':\n kwargs['widget'] = ModelSelect2Widget(\n model=se_models.Category,\n search_fields=[\n 'name__icontains',\n 'pk__startswith',\n ],\n )\n return super(ProductInline, self).formfield_for_dbfield(\n db_field,\n **kwargs,\n )\n\n\nclass PositionInline(admin.StackedInline):\n\n model = Position\n\n\nclass CategoryPageAdmin(models.CategoryPageAdmin):\n\n add = True\n delete = False\n inlines = [CategoryInline, inlines.ImageInline]\n\n def get_queryset(self, request):\n return (\n super(CategoryPageAdmin, self)\n .get_queryset(request)\n .select_related('shopelectro_category')\n )\n\n\nclass ProductPageAdmin(models.ProductPageAdmin):\n\n add = False\n delete = False\n category_page_model = se_models.CategoryPage\n list_filter = [*models.ProductPageAdmin.list_filter, HasTagsFilter, HasCategoryFilter]\n inlines = [ProductInline, inlines.ImageInline]\n search_fields = [\n 'shopelectro_product__vendor_code', 'name', 'slug',\n ]\n\n def model_id(self, obj):\n return obj.model.vendor_code\n\n model_id.short_description = _('Vendor code')\n model_id.admin_order_field = 'shopelectro_product__vendor_code'\n\n def get_queryset(self, request):\n return (\n super(ProductPageAdmin, self)\n .get_queryset(request)\n .select_related('shopelectro_product')\n )\n\n\nclass ProductFeedbackPageAdmin(admin.ModelAdmin):\n\n add = False\n delete = False\n list_filter = ['rating']\n list_display = ['rating', 'name', 'dignities', 'limitations', 'general', 'links']\n\n def links(self, obj):\n return format_html(\n '''\n <a href=\"{url}\" class=\"field-link\" title=\"\u041f\u043e\u0441\u043c\u043e\u0442\u0440\u0435\u0442\u044c \u043d\u0430 \u0441\u0430\u0439\u0442\u0435\" target=\"_blank\">\n <i class=\"fa fa-link\" aria-hidden=\"true\"></i>\n </a>\n '''.format(url=obj.product.url))\n\n links.short_description = _('Link')\n\n def get_queryset(self, request):\n return (\n super(ProductFeedbackPageAdmin, self)\n .get_queryset(request)\n .select_related('product')\n )\n\n\nclass TagGroupAdmin(admin.ModelAdmin):\n\n list_display = ['id', 'name', 'position', 'count_tags']\n list_display_links = ['name']\n\n inlines = [TagInline]\n\n def get_queryset(self, request):\n return super(TagGroupAdmin, self).get_queryset(request).prefetch_related('tags')\n\n def count_tags(self, obj):\n return obj.tags.count()\n\n\nclass TagAdmin(admin.ModelAdmin):\n\n search_fields = ['id', 'name']\n list_display = ['id', 'name', 'position', 'custom_group']\n list_display_links = ['name']\n\n def get_queryset(self, request):\n return super(TagAdmin, self).get_queryset(request).prefetch_related('group')\n\n def custom_group(self, obj):\n group = obj.group\n\n return format_html(\n '<a href=\"{url}\">{group}</a>',\n group=group,\n url=reverse('admin:shopelectro_taggroup_change', args=(group.id, ))\n )\n\n custom_group.admin_order_field = 'group'\n custom_group.short_description = _('Group')\n\n\nclass OrderAdmin(mixins.PermissionsControl):\n\n add = False\n inlines = [PositionInline]\n list_display = ['id_', 'name', 'email', 'phone', 'total_price', 'payment_type', 'paid']\n search_fields = ['name', 'email', 'phone']\n list_display_links = ['name']\n\n def id_(self, obj):\n return obj.fake_order_number\n\n\nse_admin = SEAdminSite(name='se_admin')\nse_admin.register(CustomPage, models.CustomPageAdmin)\nse_admin.register(FlatPage, models.FlatPageAdmin)\nse_admin.register(PageTemplate, models.CustomPageTemplateAdmin)\n\nse_admin.register(se_models.CategoryPage, CategoryPageAdmin)\nse_admin.register(se_models.Order, OrderAdmin)\nse_admin.register(se_models.ProductPage, ProductPageAdmin)\nse_admin.register(se_models.ProductFeedback, ProductFeedbackPageAdmin)\nse_admin.register(se_models.TagGroup, TagGroupAdmin)\nse_admin.register(se_models.Tag, TagAdmin)\n\nse_admin.register(Redirect)\n", "path": "shopelectro/admin.py"}]}
2,893
401
gh_patches_debug_27458
rasdani/github-patches
git_diff
pyinstaller__pyinstaller-4889
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> Py3.6: Unable to find .../site-packages/importlib_resources/version.txt" Hello, On latest version of pyinstaller, the hook for importlib_resource seems to look for a non existing version.txt file. It is not provided by the latest version 1.2.0 of the backport: https://gitlab.com/python-devs/importlib_resources </issue> <code> [start of PyInstaller/hooks/hook-importlib_resources.py] 1 #----------------------------------------------------------------------------- 2 # Copyright (c) 2019-2020, PyInstaller Development Team. 3 # 4 # Distributed under the terms of the GNU General Public License (version 2 5 # or later) with exception for distributing the bootloader. 6 # 7 # The full license is in the file COPYING.txt, distributed with this software. 8 # 9 # SPDX-License-Identifier: (GPL-2.0-or-later WITH Bootloader-exception) 10 #----------------------------------------------------------------------------- 11 """ 12 `importlib_resources` is a backport of the 3.7+ module `importlib.resources` 13 """ 14 15 import os 16 from PyInstaller.compat import is_py37 17 from PyInstaller.utils.hooks import get_module_file_attribute 18 19 # Include the version.txt file, used to set __version__ 20 res_loc = os.path.dirname(get_module_file_attribute('importlib_resources')) 21 datas = [ 22 (os.path.join(res_loc, 'version.txt'), 'importlib_resources'), 23 ] 24 25 # Replicate the module's version checks to exclude unused modules. 26 if is_py37: 27 # Stdlib now has the implmentation of this, so the backports 28 # aren't used at all 29 excludedmodules = [ 30 'importlib_resources._py2', 31 'importlib_resources._py3', 32 ] 33 else: 34 excludedmodules = ['importlib_resources._py2'] 35 [end of PyInstaller/hooks/hook-importlib_resources.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/PyInstaller/hooks/hook-importlib_resources.py b/PyInstaller/hooks/hook-importlib_resources.py --- a/PyInstaller/hooks/hook-importlib_resources.py +++ b/PyInstaller/hooks/hook-importlib_resources.py @@ -9,26 +9,25 @@ # SPDX-License-Identifier: (GPL-2.0-or-later WITH Bootloader-exception) #----------------------------------------------------------------------------- """ -`importlib_resources` is a backport of the 3.7+ module `importlib.resources` +`importlib_resources` is a backport of the 3.9+ module `importlib.resources` """ import os -from PyInstaller.compat import is_py37 -from PyInstaller.utils.hooks import get_module_file_attribute +from PyInstaller.utils.hooks import get_module_file_attribute, \ + is_module_satisfies, copy_metadata -# Include the version.txt file, used to set __version__ -res_loc = os.path.dirname(get_module_file_attribute('importlib_resources')) -datas = [ - (os.path.join(res_loc, 'version.txt'), 'importlib_resources'), -] - -# Replicate the module's version checks to exclude unused modules. -if is_py37: - # Stdlib now has the implmentation of this, so the backports - # aren't used at all - excludedmodules = [ - 'importlib_resources._py2', - 'importlib_resources._py3', - ] +if is_module_satisfies("importlib_resources >= 1.2.0"): + # since 1.2.0 importlib.metadata is used + datas = copy_metadata('importlib_resources') else: - excludedmodules = ['importlib_resources._py2'] + # include the version.txt file, used to set __version__ + res_loc = os.path.dirname(get_module_file_attribute('importlib_resources')) + datas = [ + (os.path.join(res_loc, 'version.txt'), 'importlib_resources'), + ] + +if is_module_satisfies("importlib_resources >= 1.3.1"): + hiddenimports = ['importlib_resources.trees'] + +# this is only required for python2 support +excludedimports = ['importlib_resources._py2']
{"golden_diff": "diff --git a/PyInstaller/hooks/hook-importlib_resources.py b/PyInstaller/hooks/hook-importlib_resources.py\n--- a/PyInstaller/hooks/hook-importlib_resources.py\n+++ b/PyInstaller/hooks/hook-importlib_resources.py\n@@ -9,26 +9,25 @@\n # SPDX-License-Identifier: (GPL-2.0-or-later WITH Bootloader-exception)\n #-----------------------------------------------------------------------------\n \"\"\"\n-`importlib_resources` is a backport of the 3.7+ module `importlib.resources`\n+`importlib_resources` is a backport of the 3.9+ module `importlib.resources`\n \"\"\"\n \n import os\n-from PyInstaller.compat import is_py37\n-from PyInstaller.utils.hooks import get_module_file_attribute\n+from PyInstaller.utils.hooks import get_module_file_attribute, \\\n+ is_module_satisfies, copy_metadata\n \n-# Include the version.txt file, used to set __version__\n-res_loc = os.path.dirname(get_module_file_attribute('importlib_resources'))\n-datas = [\n- (os.path.join(res_loc, 'version.txt'), 'importlib_resources'),\n-]\n-\n-# Replicate the module's version checks to exclude unused modules.\n-if is_py37:\n- # Stdlib now has the implmentation of this, so the backports\n- # aren't used at all\n- excludedmodules = [\n- 'importlib_resources._py2',\n- 'importlib_resources._py3',\n- ]\n+if is_module_satisfies(\"importlib_resources >= 1.2.0\"):\n+ # since 1.2.0 importlib.metadata is used\n+ datas = copy_metadata('importlib_resources')\n else:\n- excludedmodules = ['importlib_resources._py2']\n+ # include the version.txt file, used to set __version__\n+ res_loc = os.path.dirname(get_module_file_attribute('importlib_resources'))\n+ datas = [\n+ (os.path.join(res_loc, 'version.txt'), 'importlib_resources'),\n+ ]\n+\n+if is_module_satisfies(\"importlib_resources >= 1.3.1\"):\n+ hiddenimports = ['importlib_resources.trees']\n+\n+# this is only required for python2 support\n+excludedimports = ['importlib_resources._py2']\n", "issue": "Py3.6: Unable to find .../site-packages/importlib_resources/version.txt\"\nHello,\r\n\r\nOn latest version of pyinstaller, the hook for importlib_resource seems to look for a non existing version.txt file. It is not provided by the latest version 1.2.0 of the backport: https://gitlab.com/python-devs/importlib_resources\n", "before_files": [{"content": "#-----------------------------------------------------------------------------\n# Copyright (c) 2019-2020, PyInstaller Development Team.\n#\n# Distributed under the terms of the GNU General Public License (version 2\n# or later) with exception for distributing the bootloader.\n#\n# The full license is in the file COPYING.txt, distributed with this software.\n#\n# SPDX-License-Identifier: (GPL-2.0-or-later WITH Bootloader-exception)\n#-----------------------------------------------------------------------------\n\"\"\"\n`importlib_resources` is a backport of the 3.7+ module `importlib.resources`\n\"\"\"\n\nimport os\nfrom PyInstaller.compat import is_py37\nfrom PyInstaller.utils.hooks import get_module_file_attribute\n\n# Include the version.txt file, used to set __version__\nres_loc = os.path.dirname(get_module_file_attribute('importlib_resources'))\ndatas = [\n (os.path.join(res_loc, 'version.txt'), 'importlib_resources'),\n]\n\n# Replicate the module's version checks to exclude unused modules.\nif is_py37:\n # Stdlib now has the implmentation of this, so the backports\n # aren't used at all\n excludedmodules = [\n 'importlib_resources._py2',\n 'importlib_resources._py3',\n ]\nelse:\n excludedmodules = ['importlib_resources._py2']\n", "path": "PyInstaller/hooks/hook-importlib_resources.py"}]}
956
491
gh_patches_debug_27546
rasdani/github-patches
git_diff
TheAlgorithms__Python-10822
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> Improve our test coverage ### Feature description Many of our existing algorithm files have little to no unit testing. This is problematic because this can easily let bugs slip through. We want some assurance that the code we currently have is correct and functional. We welcome all contributors to open PRs to help us add tests to our codebase. ### How to find low-coverage files Go to the Actions tab in this repository and find the most recent **build** workflow run. Open the logs under "Run Tests" and scroll down until you find the section on code coverage: ``` ---------- coverage: platform linux, python 3.12.0-final-0 ----------- Name Stmts Miss Cover Missing ----------------------------------------------------------------------------------------------------------- quantum/q_fourier_transform.py 30 30 0% 14-93 scripts/validate_solutions.py 54 54 0% 2-94 strings/min_cost_string_conversion.py 78 75 4% 20-57, 61-75, 79-129 ... ``` The "Cover" column tells you what percentage of the lines in that file are covered by tests. We want to increase this percentage for existing files. Find a file with low coverage percentage that you wish to write tests for, add doctests for each function, and open a PR with your changes. You do not need to have a perfect coverage percentage, but all functions should have doctests. Some files will naturally be hard to write tests for. For example, the file may be poorly written because they lack any functions. Other files might be how-tos, meaning they simply demonstrate how to use an existing library's functions rather than implementing the algorithm themselves. Ignore these kinds of files, as they will need to be rewritten eventually. Furthermore, ignore files in the `web_programming` and `project_euler` directories. Web programming files are inherently hard to test and Project Euler files have their own validation workflow, so don't worry about their test coverage. _**When you open your PR, put "Contributes to #9943" in the PR description.**_ Do not use the word "fixes", "resolves", or "closes". This issue is an ongoing one, and your PR will not single-handedly resolve this issue. ### How to add doctests A doctest is a unit test that is contained within the documentation comment (docstring) for a function. Here is an example of what doctests look like within a docstring: ```py def add(a: int, b: int) -> int: """ Adds two non-negative numbers. >>> add(1, 1) 2 >>> add(2, 5) 7 >>> add(1, 0) 1 >>> add(-1, -1) Traceback (most recent last): ... ValueError: Numbers must be non-negative """ ``` For every function in the file you choose, you should write doctests like the ones shown above in its docstring. If a function doesn't have a docstring, add one. Your doctests should be comprehensive but not excessive: you should write just enough tests to cover all basic cases as well as all edge cases (e.g., negative numbers, empty lists, etc). Do not simply run a function on some example inputs and put its output as the expected output for a doctest. This assumes that the function is implemented correctly when it might not be. Verify independently that your doctests and their expected outputs are correct. **Your PR will not be merged if it has failing tests.** If you happen to discover a bug while writing doctests, please fix it. _**Please read our [contributing guidelines](https://github.com/TheAlgorithms/Python/blob/master/CONTRIBUTING.md) before you contribute.**_ </issue> <code> [start of maths/binomial_coefficient.py] 1 def binomial_coefficient(n: int, r: int) -> int: 2 """ 3 Find binomial coefficient using pascals triangle. 4 5 >>> binomial_coefficient(10, 5) 6 252 7 """ 8 c = [0 for i in range(r + 1)] 9 # nc0 = 1 10 c[0] = 1 11 for i in range(1, n + 1): 12 # to compute current row from previous row. 13 j = min(i, r) 14 while j > 0: 15 c[j] += c[j - 1] 16 j -= 1 17 return c[r] 18 19 20 print(binomial_coefficient(n=10, r=5)) 21 [end of maths/binomial_coefficient.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/maths/binomial_coefficient.py b/maths/binomial_coefficient.py --- a/maths/binomial_coefficient.py +++ b/maths/binomial_coefficient.py @@ -1,10 +1,48 @@ def binomial_coefficient(n: int, r: int) -> int: """ - Find binomial coefficient using pascals triangle. + Find binomial coefficient using Pascal's triangle. + + Calculate C(n, r) using Pascal's triangle. + + :param n: The total number of items. + :param r: The number of items to choose. + :return: The binomial coefficient C(n, r). >>> binomial_coefficient(10, 5) 252 + >>> binomial_coefficient(10, 0) + 1 + >>> binomial_coefficient(0, 10) + 1 + >>> binomial_coefficient(10, 10) + 1 + >>> binomial_coefficient(5, 2) + 10 + >>> binomial_coefficient(5, 6) + 0 + >>> binomial_coefficient(3, 5) + 0 + >>> binomial_coefficient(-2, 3) + Traceback (most recent call last): + ... + ValueError: n and r must be non-negative integers + >>> binomial_coefficient(5, -1) + Traceback (most recent call last): + ... + ValueError: n and r must be non-negative integers + >>> binomial_coefficient(10.1, 5) + Traceback (most recent call last): + ... + TypeError: 'float' object cannot be interpreted as an integer + >>> binomial_coefficient(10, 5.1) + Traceback (most recent call last): + ... + TypeError: 'float' object cannot be interpreted as an integer """ + if n < 0 or r < 0: + raise ValueError("n and r must be non-negative integers") + if 0 in (n, r): + return 1 c = [0 for i in range(r + 1)] # nc0 = 1 c[0] = 1 @@ -17,4 +55,8 @@ return c[r] -print(binomial_coefficient(n=10, r=5)) +if __name__ == "__main__": + from doctest import testmod + + testmod() + print(binomial_coefficient(n=10, r=5))
{"golden_diff": "diff --git a/maths/binomial_coefficient.py b/maths/binomial_coefficient.py\n--- a/maths/binomial_coefficient.py\n+++ b/maths/binomial_coefficient.py\n@@ -1,10 +1,48 @@\n def binomial_coefficient(n: int, r: int) -> int:\n \"\"\"\n- Find binomial coefficient using pascals triangle.\n+ Find binomial coefficient using Pascal's triangle.\n+\n+ Calculate C(n, r) using Pascal's triangle.\n+\n+ :param n: The total number of items.\n+ :param r: The number of items to choose.\n+ :return: The binomial coefficient C(n, r).\n \n >>> binomial_coefficient(10, 5)\n 252\n+ >>> binomial_coefficient(10, 0)\n+ 1\n+ >>> binomial_coefficient(0, 10)\n+ 1\n+ >>> binomial_coefficient(10, 10)\n+ 1\n+ >>> binomial_coefficient(5, 2)\n+ 10\n+ >>> binomial_coefficient(5, 6)\n+ 0\n+ >>> binomial_coefficient(3, 5)\n+ 0\n+ >>> binomial_coefficient(-2, 3)\n+ Traceback (most recent call last):\n+ ...\n+ ValueError: n and r must be non-negative integers\n+ >>> binomial_coefficient(5, -1)\n+ Traceback (most recent call last):\n+ ...\n+ ValueError: n and r must be non-negative integers\n+ >>> binomial_coefficient(10.1, 5)\n+ Traceback (most recent call last):\n+ ...\n+ TypeError: 'float' object cannot be interpreted as an integer\n+ >>> binomial_coefficient(10, 5.1)\n+ Traceback (most recent call last):\n+ ...\n+ TypeError: 'float' object cannot be interpreted as an integer\n \"\"\"\n+ if n < 0 or r < 0:\n+ raise ValueError(\"n and r must be non-negative integers\")\n+ if 0 in (n, r):\n+ return 1\n c = [0 for i in range(r + 1)]\n # nc0 = 1\n c[0] = 1\n@@ -17,4 +55,8 @@\n return c[r]\n \n \n-print(binomial_coefficient(n=10, r=5))\n+if __name__ == \"__main__\":\n+ from doctest import testmod\n+\n+ testmod()\n+ print(binomial_coefficient(n=10, r=5))\n", "issue": "Improve our test coverage\n### Feature description\r\n\r\nMany of our existing algorithm files have little to no unit testing. This is problematic because this can easily let bugs slip through. We want some assurance that the code we currently have is correct and functional. We welcome all contributors to open PRs to help us add tests to our codebase.\r\n\r\n### How to find low-coverage files\r\n\r\nGo to the Actions tab in this repository and find the most recent **build** workflow run. Open the logs under \"Run Tests\" and scroll down until you find the section on code coverage:\r\n```\r\n---------- coverage: platform linux, python 3.12.0-final-0 -----------\r\nName Stmts Miss Cover Missing\r\n-----------------------------------------------------------------------------------------------------------\r\nquantum/q_fourier_transform.py 30 30 0% 14-93\r\nscripts/validate_solutions.py 54 54 0% 2-94\r\nstrings/min_cost_string_conversion.py 78 75 4% 20-57, 61-75, 79-129\r\n...\r\n```\r\nThe \"Cover\" column tells you what percentage of the lines in that file are covered by tests. We want to increase this percentage for existing files. Find a file with low coverage percentage that you wish to write tests for, add doctests for each function, and open a PR with your changes. You do not need to have a perfect coverage percentage, but all functions should have doctests.\r\n\r\nSome files will naturally be hard to write tests for. For example, the file may be poorly written because they lack any functions. Other files might be how-tos, meaning they simply demonstrate how to use an existing library's functions rather than implementing the algorithm themselves. Ignore these kinds of files, as they will need to be rewritten eventually. Furthermore, ignore files in the `web_programming` and `project_euler` directories. Web programming files are inherently hard to test and Project Euler files have their own validation workflow, so don't worry about their test coverage.\r\n\r\n_**When you open your PR, put \"Contributes to #9943\" in the PR description.**_ Do not use the word \"fixes\", \"resolves\", or \"closes\". This issue is an ongoing one, and your PR will not single-handedly resolve this issue.\r\n\r\n### How to add doctests\r\n\r\nA doctest is a unit test that is contained within the documentation comment (docstring) for a function. Here is an example of what doctests look like within a docstring:\r\n```py\r\ndef add(a: int, b: int) -> int:\r\n \"\"\"\r\n Adds two non-negative numbers.\r\n >>> add(1, 1)\r\n 2\r\n >>> add(2, 5)\r\n 7\r\n >>> add(1, 0)\r\n 1\r\n >>> add(-1, -1)\r\n Traceback (most recent last):\r\n ...\r\n ValueError: Numbers must be non-negative\r\n \"\"\"\r\n```\r\nFor every function in the file you choose, you should write doctests like the ones shown above in its docstring. If a function doesn't have a docstring, add one. Your doctests should be comprehensive but not excessive: you should write just enough tests to cover all basic cases as well as all edge cases (e.g., negative numbers, empty lists, etc).\r\n\r\nDo not simply run a function on some example inputs and put its output as the expected output for a doctest. This assumes that the function is implemented correctly when it might not be. Verify independently that your doctests and their expected outputs are correct. **Your PR will not be merged if it has failing tests.** If you happen to discover a bug while writing doctests, please fix it.\r\n\r\n_**Please read our [contributing guidelines](https://github.com/TheAlgorithms/Python/blob/master/CONTRIBUTING.md) before you contribute.**_\n", "before_files": [{"content": "def binomial_coefficient(n: int, r: int) -> int:\n \"\"\"\n Find binomial coefficient using pascals triangle.\n\n >>> binomial_coefficient(10, 5)\n 252\n \"\"\"\n c = [0 for i in range(r + 1)]\n # nc0 = 1\n c[0] = 1\n for i in range(1, n + 1):\n # to compute current row from previous row.\n j = min(i, r)\n while j > 0:\n c[j] += c[j - 1]\n j -= 1\n return c[r]\n\n\nprint(binomial_coefficient(n=10, r=5))\n", "path": "maths/binomial_coefficient.py"}]}
1,569
595
gh_patches_debug_16578
rasdani/github-patches
git_diff
doccano__doccano-1668
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> Pagination of the project list When fetching projects in the project list page, is it intentional that all projects are fetched at once even though there is pagination? Endpoint of project list fetching: `/v1/projects` When there are a lot of projects, it takes a long time to display them. Your Environment --------- doccano v1.5.5 </issue> <code> [start of backend/api/views/project.py] 1 from django.conf import settings 2 from rest_framework import generics, status 3 from rest_framework.permissions import IsAdminUser, IsAuthenticated 4 from rest_framework.response import Response 5 6 from members.permissions import IsProjectAdmin, IsProjectStaffAndReadOnly 7 8 from ..models import Project 9 from ..serializers import ProjectPolymorphicSerializer 10 11 12 class ProjectList(generics.ListCreateAPIView): 13 serializer_class = ProjectPolymorphicSerializer 14 pagination_class = None 15 16 def get_permissions(self): 17 if self.request.method == 'GET': 18 self.permission_classes = [IsAuthenticated, ] 19 else: 20 self.permission_classes = [IsAuthenticated & IsAdminUser] 21 return super().get_permissions() 22 23 def get_queryset(self): 24 return Project.objects.filter(role_mappings__user=self.request.user) 25 26 def perform_create(self, serializer): 27 serializer.save(created_by=self.request.user) 28 29 def delete(self, request, *args, **kwargs): 30 delete_ids = request.data['ids'] 31 projects = Project.objects.filter( 32 role_mappings__user=self.request.user, 33 role_mappings__role__name=settings.ROLE_PROJECT_ADMIN, 34 pk__in=delete_ids 35 ) 36 # Todo: I want to use bulk delete. 37 # But it causes the constraint error. 38 # See https://github.com/django-polymorphic/django-polymorphic/issues/229 39 for project in projects: 40 project.delete() 41 return Response(status=status.HTTP_204_NO_CONTENT) 42 43 44 class ProjectDetail(generics.RetrieveUpdateDestroyAPIView): 45 queryset = Project.objects.all() 46 serializer_class = ProjectPolymorphicSerializer 47 lookup_url_kwarg = 'project_id' 48 permission_classes = [IsAuthenticated & (IsProjectAdmin | IsProjectStaffAndReadOnly)] 49 [end of backend/api/views/project.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/backend/api/views/project.py b/backend/api/views/project.py --- a/backend/api/views/project.py +++ b/backend/api/views/project.py @@ -1,5 +1,6 @@ from django.conf import settings -from rest_framework import generics, status +from django_filters.rest_framework import DjangoFilterBackend +from rest_framework import filters, generics, status from rest_framework.permissions import IsAdminUser, IsAuthenticated from rest_framework.response import Response @@ -11,7 +12,8 @@ class ProjectList(generics.ListCreateAPIView): serializer_class = ProjectPolymorphicSerializer - pagination_class = None + filter_backends = (DjangoFilterBackend, filters.SearchFilter, filters.OrderingFilter) + search_fields = ('name', 'description') def get_permissions(self): if self.request.method == 'GET':
{"golden_diff": "diff --git a/backend/api/views/project.py b/backend/api/views/project.py\n--- a/backend/api/views/project.py\n+++ b/backend/api/views/project.py\n@@ -1,5 +1,6 @@\n from django.conf import settings\n-from rest_framework import generics, status\n+from django_filters.rest_framework import DjangoFilterBackend\n+from rest_framework import filters, generics, status\n from rest_framework.permissions import IsAdminUser, IsAuthenticated\n from rest_framework.response import Response\n \n@@ -11,7 +12,8 @@\n \n class ProjectList(generics.ListCreateAPIView):\n serializer_class = ProjectPolymorphicSerializer\n- pagination_class = None\n+ filter_backends = (DjangoFilterBackend, filters.SearchFilter, filters.OrderingFilter)\n+ search_fields = ('name', 'description')\n \n def get_permissions(self):\n if self.request.method == 'GET':\n", "issue": "Pagination of the project list\nWhen fetching projects in the project list page, is it intentional that all projects are fetched at once even though there is pagination?\r\n\r\nEndpoint of project list fetching: `/v1/projects`\r\n\r\nWhen there are a lot of projects, it takes a long time to display them.\r\n\r\nYour Environment\r\n---------\r\ndoccano v1.5.5\n", "before_files": [{"content": "from django.conf import settings\nfrom rest_framework import generics, status\nfrom rest_framework.permissions import IsAdminUser, IsAuthenticated\nfrom rest_framework.response import Response\n\nfrom members.permissions import IsProjectAdmin, IsProjectStaffAndReadOnly\n\nfrom ..models import Project\nfrom ..serializers import ProjectPolymorphicSerializer\n\n\nclass ProjectList(generics.ListCreateAPIView):\n serializer_class = ProjectPolymorphicSerializer\n pagination_class = None\n\n def get_permissions(self):\n if self.request.method == 'GET':\n self.permission_classes = [IsAuthenticated, ]\n else:\n self.permission_classes = [IsAuthenticated & IsAdminUser]\n return super().get_permissions()\n\n def get_queryset(self):\n return Project.objects.filter(role_mappings__user=self.request.user)\n\n def perform_create(self, serializer):\n serializer.save(created_by=self.request.user)\n\n def delete(self, request, *args, **kwargs):\n delete_ids = request.data['ids']\n projects = Project.objects.filter(\n role_mappings__user=self.request.user,\n role_mappings__role__name=settings.ROLE_PROJECT_ADMIN,\n pk__in=delete_ids\n )\n # Todo: I want to use bulk delete.\n # But it causes the constraint error.\n # See https://github.com/django-polymorphic/django-polymorphic/issues/229\n for project in projects:\n project.delete()\n return Response(status=status.HTTP_204_NO_CONTENT)\n\n\nclass ProjectDetail(generics.RetrieveUpdateDestroyAPIView):\n queryset = Project.objects.all()\n serializer_class = ProjectPolymorphicSerializer\n lookup_url_kwarg = 'project_id'\n permission_classes = [IsAuthenticated & (IsProjectAdmin | IsProjectStaffAndReadOnly)]\n", "path": "backend/api/views/project.py"}]}
1,068
185
gh_patches_debug_38925
rasdani/github-patches
git_diff
sktime__sktime-5942
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> [BUG] TDistribution pdf does not consider location/scale **Describe the bug** The location/scale (mu/sigma) are not considered at all in methods: pdf, log_pdf, cdf or ppf For example, the case pdf: ``` def pdf(self, x): """Probability density function.""" d = self.loc[x.index, x.columns] pdf_arr = gamma((d._df + 1) / 2) pdf_arr = pdf_arr / (np.sqrt(np.pi * d._df) * gamma(d._df / 2)) pdf_arr = pdf_arr * (1 + x**2 / d._df) ** (-(d._df + 1) / 2) return pd.DataFrame(pdf_arr, index=x.index, columns=x.columns) ``` Here, in line 6, instead of x**2 one should use ((x - d._mu)/d._sigma)**2 sktime version 0.25.1 [BUG] TDistribution computes incorrectly variance **Describe the bug** TDistribution does not use scale (sigma) for variance computation ``` def var(self): ... df_arr = self._df.copy() df_arr = df_arr.astype(np.float32) df_arr[df_arr <= 2] = np.inf mask = (df_arr > 2) & (df_arr != np.inf) df_arr[mask] = df_arr[mask] / (df_arr[mask] - 2) return pd.DataFrame(df_arr, index=self.index, columns=self.columns) ``` one should correct the previous to last line to: ``` df_arr[mask] = self._sigma** 2 * df_arr[mask] / (df_arr[mask] - 2) ``` sktime version 0.25.1 </issue> <code> [start of sktime/proba/t.py] 1 # copyright: sktime developers, BSD-3-Clause License (see LICENSE file) 2 """Student's t-distribution.""" 3 4 __author__ = ["Alex-JG3"] 5 6 import numpy as np 7 import pandas as pd 8 from scipy.special import betaincinv, gamma, hyp2f1, loggamma 9 10 from sktime.proba.base import BaseDistribution 11 12 13 class TDistribution(BaseDistribution): 14 """Student's t-distribution (sktime native). 15 16 Parameters 17 ---------- 18 mean : float or array of float (1D or 2D) 19 mean of the t-distribution distribution 20 sd : float or array of float (1D or 2D), must be positive 21 standard deviation of the t-distribution distribution 22 df : float or array of float (1D or 2D), must be positive 23 Degrees of freedom of the t-distribution 24 index : pd.Index, optional, default = RangeIndex 25 columns : pd.Index, optional, default = RangeIndex 26 27 Example 28 ------- 29 >>> from sktime.proba.t import TDistribution 30 31 >>> n = TDistribution(mu=[[0, 1], [2, 3], [4, 5]], sigma=1, df=10) 32 """ 33 34 _tags = { 35 "authors": ["Alex-JG3"], 36 "maintainers": ["Alex-JG3"], 37 "capabilities:approx": ["pdfnorm", "energy"], 38 "capabilities:exact": ["mean", "var", "pdf", "log_pdf", "cdf", "ppf"], 39 "distr:measuretype": "continuous", 40 } 41 42 def __init__(self, mu, sigma, df=1, index=None, columns=None): 43 self.mu = mu 44 self.sigma = sigma 45 self.df = df 46 self.index = index 47 self.columns = columns 48 49 self._mu, self._sigma, self._df = self._get_bc_params( 50 self.mu, self.sigma, self.df 51 ) 52 shape = self._mu.shape 53 54 if index is None: 55 index = pd.RangeIndex(shape[0]) 56 57 if columns is None: 58 columns = pd.RangeIndex(shape[1]) 59 60 super().__init__(index=index, columns=columns) 61 62 def mean(self): 63 r"""Return expected value of the distribution. 64 65 Let :math:`X` be a random variable with the distribution of `self`. 66 Returns the expectation :math:`\mathbb{E}[X]`. The expectation, 67 :math:`\mathbb{E}[X]`, as infinite if :math:`\nu \le 1`. 68 69 Returns 70 ------- 71 pd.DataFrame with same rows, columns as `self` 72 expected value of distribution (entry-wise) 73 """ 74 mean_arr = self._mu.copy() 75 if (self._df <= 1).any(): 76 mean_arr = mean_arr.astype(np.float32) 77 mean_arr[self._df <= 1] = np.inf 78 return pd.DataFrame(mean_arr, index=self.index, columns=self.columns) 79 80 def var(self): 81 r"""Return element/entry-wise variance of the distribution. 82 83 Let :math:`X` be a random variable with the distribution of `self`. 84 Returns, 85 86 .. math:: 87 \mathbb{V}[X] = \begin{cases} 88 \frac{\nu}{\nu - 2} & \text{if} \nu > 2, \\ 89 \infty & \text{if} \nu \le 2, \\ 90 \begin{cases} 91 92 Where :math:`\nu` is the degrees of freedom of the t-distribution. 93 94 Returns 95 ------- 96 pd.DataFrame with same rows, columns as `self` 97 variance of distribution (entry-wise) 98 """ 99 df_arr = self._df.copy() 100 df_arr = df_arr.astype(np.float32) 101 df_arr[df_arr <= 2] = np.inf 102 mask = (df_arr > 2) & (df_arr != np.inf) 103 df_arr[mask] = df_arr[mask] / (df_arr[mask] - 2) 104 return pd.DataFrame(df_arr, index=self.index, columns=self.columns) 105 106 def pdf(self, x): 107 """Probability density function.""" 108 d = self.loc[x.index, x.columns] 109 pdf_arr = gamma((d._df + 1) / 2) 110 pdf_arr = pdf_arr / (np.sqrt(np.pi * d._df) * gamma(d._df / 2)) 111 pdf_arr = pdf_arr * (1 + x**2 / d._df) ** (-(d._df + 1) / 2) 112 return pd.DataFrame(pdf_arr, index=x.index, columns=x.columns) 113 114 def log_pdf(self, x): 115 """Logarithmic probability density function.""" 116 d = self.loc[x.index, x.columns] 117 lpdf_arr = loggamma((d._df + 1) / 2) 118 lpdf_arr = lpdf_arr - 0.5 * np.log(d._df * np.pi) 119 lpdf_arr = lpdf_arr - loggamma(d._df / 2) 120 lpdf_arr = lpdf_arr - ((d._df + 1) / 2) * np.log(1 + x**2 / d._df) 121 return pd.DataFrame(lpdf_arr, index=x.index, columns=x.columns) 122 123 def cdf(self, x): 124 """Cumulative distribution function.""" 125 d = self.loc[x.index, x.columns] 126 cdf_arr = x * gamma((d._df + 1) / 2) 127 cdf_arr = cdf_arr * hyp2f1(0.5, (d._df + 1) / 2, 3 / 2, -(x**2) / d._df) 128 cdf_arr = 0.5 + cdf_arr / (np.sqrt(np.pi * d._df) * gamma(d._df / 2)) 129 return pd.DataFrame(cdf_arr, index=x.index, columns=x.columns) 130 131 def ppf(self, p): 132 """Quantile function = percent point function = inverse cdf.""" 133 d = self.loc[p.index, p.columns] 134 ppf_arr = p.to_numpy(copy=True) 135 ppf_arr[p.values == 0.5] = 0.0 136 ppf_arr[p.values <= 0] = -np.inf 137 ppf_arr[p.values >= 1] = np.inf 138 139 mask1 = (p.values < 0.5) & (p.values > 0) 140 mask2 = (p.values < 1) & (p.values > 0.5) 141 ppf_arr[mask1] = 1 / betaincinv(0.5 * d._df[mask1], 0.5, 2 * ppf_arr[mask1]) 142 ppf_arr[mask2] = 1 / betaincinv( 143 0.5 * d._df[mask2], 0.5, 2 * (1 - ppf_arr[mask2]) 144 ) 145 ppf_arr[mask1 | mask2] = np.sqrt(ppf_arr[mask1 | mask2] - 1) 146 ppf_arr[mask1 | mask2] = np.sqrt(d._df[mask1 | mask2]) * ppf_arr[mask1 | mask2] 147 ppf_arr[mask1] = -ppf_arr[mask1] 148 return pd.DataFrame(ppf_arr, index=p.index, columns=p.columns) 149 150 @classmethod 151 def get_test_params(cls, parameter_set="default"): 152 """Return testing parameter settings for the estimator.""" 153 params1 = {"mu": [[0, 1], [2, 3], [4, 5]], "sigma": 1} 154 params2 = { 155 "mu": 0, 156 "sigma": 1, 157 "index": pd.Index([1, 2, 5]), 158 "columns": pd.Index(["a", "b"]), 159 } 160 return [params1, params2] 161 [end of sktime/proba/t.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/sktime/proba/t.py b/sktime/proba/t.py --- a/sktime/proba/t.py +++ b/sktime/proba/t.py @@ -100,7 +100,7 @@ df_arr = df_arr.astype(np.float32) df_arr[df_arr <= 2] = np.inf mask = (df_arr > 2) & (df_arr != np.inf) - df_arr[mask] = df_arr[mask] / (df_arr[mask] - 2) + df_arr[mask] = self._sigma[mask] ** 2 * df_arr[mask] / (df_arr[mask] - 2) return pd.DataFrame(df_arr, index=self.index, columns=self.columns) def pdf(self, x): @@ -108,7 +108,10 @@ d = self.loc[x.index, x.columns] pdf_arr = gamma((d._df + 1) / 2) pdf_arr = pdf_arr / (np.sqrt(np.pi * d._df) * gamma(d._df / 2)) - pdf_arr = pdf_arr * (1 + x**2 / d._df) ** (-(d._df + 1) / 2) + pdf_arr = pdf_arr * (1 + ((x - d._mu) / d._sigma) ** 2 / d._df) ** ( + -(d._df + 1) / 2 + ) + pdf_arr = pdf_arr / d._sigma return pd.DataFrame(pdf_arr, index=x.index, columns=x.columns) def log_pdf(self, x): @@ -117,14 +120,18 @@ lpdf_arr = loggamma((d._df + 1) / 2) lpdf_arr = lpdf_arr - 0.5 * np.log(d._df * np.pi) lpdf_arr = lpdf_arr - loggamma(d._df / 2) - lpdf_arr = lpdf_arr - ((d._df + 1) / 2) * np.log(1 + x**2 / d._df) + lpdf_arr = lpdf_arr - ((d._df + 1) / 2) * np.log( + 1 + ((x - d._mu) / d._sigma) ** 2 / d._df + ) + lpdf_arr = lpdf_arr - np.log(d._sigma) return pd.DataFrame(lpdf_arr, index=x.index, columns=x.columns) def cdf(self, x): """Cumulative distribution function.""" d = self.loc[x.index, x.columns] - cdf_arr = x * gamma((d._df + 1) / 2) - cdf_arr = cdf_arr * hyp2f1(0.5, (d._df + 1) / 2, 3 / 2, -(x**2) / d._df) + x_ = (x - d._mu) / d._sigma + cdf_arr = x_ * gamma((d._df + 1) / 2) + cdf_arr = cdf_arr * hyp2f1(0.5, (d._df + 1) / 2, 3 / 2, -(x_**2) / d._df) cdf_arr = 0.5 + cdf_arr / (np.sqrt(np.pi * d._df) * gamma(d._df / 2)) return pd.DataFrame(cdf_arr, index=x.index, columns=x.columns) @@ -145,6 +152,7 @@ ppf_arr[mask1 | mask2] = np.sqrt(ppf_arr[mask1 | mask2] - 1) ppf_arr[mask1 | mask2] = np.sqrt(d._df[mask1 | mask2]) * ppf_arr[mask1 | mask2] ppf_arr[mask1] = -ppf_arr[mask1] + ppf_arr = d._sigma * ppf_arr + d._mu return pd.DataFrame(ppf_arr, index=p.index, columns=p.columns) @classmethod
{"golden_diff": "diff --git a/sktime/proba/t.py b/sktime/proba/t.py\n--- a/sktime/proba/t.py\n+++ b/sktime/proba/t.py\n@@ -100,7 +100,7 @@\n df_arr = df_arr.astype(np.float32)\n df_arr[df_arr <= 2] = np.inf\n mask = (df_arr > 2) & (df_arr != np.inf)\n- df_arr[mask] = df_arr[mask] / (df_arr[mask] - 2)\n+ df_arr[mask] = self._sigma[mask] ** 2 * df_arr[mask] / (df_arr[mask] - 2)\n return pd.DataFrame(df_arr, index=self.index, columns=self.columns)\n \n def pdf(self, x):\n@@ -108,7 +108,10 @@\n d = self.loc[x.index, x.columns]\n pdf_arr = gamma((d._df + 1) / 2)\n pdf_arr = pdf_arr / (np.sqrt(np.pi * d._df) * gamma(d._df / 2))\n- pdf_arr = pdf_arr * (1 + x**2 / d._df) ** (-(d._df + 1) / 2)\n+ pdf_arr = pdf_arr * (1 + ((x - d._mu) / d._sigma) ** 2 / d._df) ** (\n+ -(d._df + 1) / 2\n+ )\n+ pdf_arr = pdf_arr / d._sigma\n return pd.DataFrame(pdf_arr, index=x.index, columns=x.columns)\n \n def log_pdf(self, x):\n@@ -117,14 +120,18 @@\n lpdf_arr = loggamma((d._df + 1) / 2)\n lpdf_arr = lpdf_arr - 0.5 * np.log(d._df * np.pi)\n lpdf_arr = lpdf_arr - loggamma(d._df / 2)\n- lpdf_arr = lpdf_arr - ((d._df + 1) / 2) * np.log(1 + x**2 / d._df)\n+ lpdf_arr = lpdf_arr - ((d._df + 1) / 2) * np.log(\n+ 1 + ((x - d._mu) / d._sigma) ** 2 / d._df\n+ )\n+ lpdf_arr = lpdf_arr - np.log(d._sigma)\n return pd.DataFrame(lpdf_arr, index=x.index, columns=x.columns)\n \n def cdf(self, x):\n \"\"\"Cumulative distribution function.\"\"\"\n d = self.loc[x.index, x.columns]\n- cdf_arr = x * gamma((d._df + 1) / 2)\n- cdf_arr = cdf_arr * hyp2f1(0.5, (d._df + 1) / 2, 3 / 2, -(x**2) / d._df)\n+ x_ = (x - d._mu) / d._sigma\n+ cdf_arr = x_ * gamma((d._df + 1) / 2)\n+ cdf_arr = cdf_arr * hyp2f1(0.5, (d._df + 1) / 2, 3 / 2, -(x_**2) / d._df)\n cdf_arr = 0.5 + cdf_arr / (np.sqrt(np.pi * d._df) * gamma(d._df / 2))\n return pd.DataFrame(cdf_arr, index=x.index, columns=x.columns)\n \n@@ -145,6 +152,7 @@\n ppf_arr[mask1 | mask2] = np.sqrt(ppf_arr[mask1 | mask2] - 1)\n ppf_arr[mask1 | mask2] = np.sqrt(d._df[mask1 | mask2]) * ppf_arr[mask1 | mask2]\n ppf_arr[mask1] = -ppf_arr[mask1]\n+ ppf_arr = d._sigma * ppf_arr + d._mu\n return pd.DataFrame(ppf_arr, index=p.index, columns=p.columns)\n \n @classmethod\n", "issue": "[BUG] TDistribution pdf does not consider location/scale\n**Describe the bug**\r\nThe location/scale (mu/sigma) are not considered at all in methods: pdf, log_pdf, cdf or ppf\r\n\r\nFor example, the case pdf:\r\n\r\n``` \r\ndef pdf(self, x):\r\n \"\"\"Probability density function.\"\"\"\r\n d = self.loc[x.index, x.columns]\r\n pdf_arr = gamma((d._df + 1) / 2)\r\n pdf_arr = pdf_arr / (np.sqrt(np.pi * d._df) * gamma(d._df / 2))\r\n pdf_arr = pdf_arr * (1 + x**2 / d._df) ** (-(d._df + 1) / 2)\r\n return pd.DataFrame(pdf_arr, index=x.index, columns=x.columns)\r\n```\r\n\r\nHere, in line 6, instead of x**2 one should use ((x - d._mu)/d._sigma)**2\r\n\r\n\r\nsktime version 0.25.1\r\n\r\n\n[BUG] TDistribution computes incorrectly variance \n**Describe the bug**\r\n\r\nTDistribution does not use scale (sigma) for variance computation\r\n\r\n``` \r\ndef var(self):\r\n...\r\n df_arr = self._df.copy()\r\n df_arr = df_arr.astype(np.float32)\r\n df_arr[df_arr <= 2] = np.inf\r\n mask = (df_arr > 2) & (df_arr != np.inf)\r\n df_arr[mask] = df_arr[mask] / (df_arr[mask] - 2)\r\n return pd.DataFrame(df_arr, index=self.index, columns=self.columns)\r\n```\r\n\r\none should correct the previous to last line to:\r\n\r\n```\r\ndf_arr[mask] = self._sigma** 2 * df_arr[mask] / (df_arr[mask] - 2)\r\n```\r\n\r\nsktime version 0.25.1\n", "before_files": [{"content": "# copyright: sktime developers, BSD-3-Clause License (see LICENSE file)\n\"\"\"Student's t-distribution.\"\"\"\n\n__author__ = [\"Alex-JG3\"]\n\nimport numpy as np\nimport pandas as pd\nfrom scipy.special import betaincinv, gamma, hyp2f1, loggamma\n\nfrom sktime.proba.base import BaseDistribution\n\n\nclass TDistribution(BaseDistribution):\n \"\"\"Student's t-distribution (sktime native).\n\n Parameters\n ----------\n mean : float or array of float (1D or 2D)\n mean of the t-distribution distribution\n sd : float or array of float (1D or 2D), must be positive\n standard deviation of the t-distribution distribution\n df : float or array of float (1D or 2D), must be positive\n Degrees of freedom of the t-distribution\n index : pd.Index, optional, default = RangeIndex\n columns : pd.Index, optional, default = RangeIndex\n\n Example\n -------\n >>> from sktime.proba.t import TDistribution\n\n >>> n = TDistribution(mu=[[0, 1], [2, 3], [4, 5]], sigma=1, df=10)\n \"\"\"\n\n _tags = {\n \"authors\": [\"Alex-JG3\"],\n \"maintainers\": [\"Alex-JG3\"],\n \"capabilities:approx\": [\"pdfnorm\", \"energy\"],\n \"capabilities:exact\": [\"mean\", \"var\", \"pdf\", \"log_pdf\", \"cdf\", \"ppf\"],\n \"distr:measuretype\": \"continuous\",\n }\n\n def __init__(self, mu, sigma, df=1, index=None, columns=None):\n self.mu = mu\n self.sigma = sigma\n self.df = df\n self.index = index\n self.columns = columns\n\n self._mu, self._sigma, self._df = self._get_bc_params(\n self.mu, self.sigma, self.df\n )\n shape = self._mu.shape\n\n if index is None:\n index = pd.RangeIndex(shape[0])\n\n if columns is None:\n columns = pd.RangeIndex(shape[1])\n\n super().__init__(index=index, columns=columns)\n\n def mean(self):\n r\"\"\"Return expected value of the distribution.\n\n Let :math:`X` be a random variable with the distribution of `self`.\n Returns the expectation :math:`\\mathbb{E}[X]`. The expectation,\n :math:`\\mathbb{E}[X]`, as infinite if :math:`\\nu \\le 1`.\n\n Returns\n -------\n pd.DataFrame with same rows, columns as `self`\n expected value of distribution (entry-wise)\n \"\"\"\n mean_arr = self._mu.copy()\n if (self._df <= 1).any():\n mean_arr = mean_arr.astype(np.float32)\n mean_arr[self._df <= 1] = np.inf\n return pd.DataFrame(mean_arr, index=self.index, columns=self.columns)\n\n def var(self):\n r\"\"\"Return element/entry-wise variance of the distribution.\n\n Let :math:`X` be a random variable with the distribution of `self`.\n Returns,\n\n .. math::\n \\mathbb{V}[X] = \\begin{cases}\n \\frac{\\nu}{\\nu - 2} & \\text{if} \\nu > 2, \\\\\n \\infty & \\text{if} \\nu \\le 2, \\\\\n \\begin{cases}\n\n Where :math:`\\nu` is the degrees of freedom of the t-distribution.\n\n Returns\n -------\n pd.DataFrame with same rows, columns as `self`\n variance of distribution (entry-wise)\n \"\"\"\n df_arr = self._df.copy()\n df_arr = df_arr.astype(np.float32)\n df_arr[df_arr <= 2] = np.inf\n mask = (df_arr > 2) & (df_arr != np.inf)\n df_arr[mask] = df_arr[mask] / (df_arr[mask] - 2)\n return pd.DataFrame(df_arr, index=self.index, columns=self.columns)\n\n def pdf(self, x):\n \"\"\"Probability density function.\"\"\"\n d = self.loc[x.index, x.columns]\n pdf_arr = gamma((d._df + 1) / 2)\n pdf_arr = pdf_arr / (np.sqrt(np.pi * d._df) * gamma(d._df / 2))\n pdf_arr = pdf_arr * (1 + x**2 / d._df) ** (-(d._df + 1) / 2)\n return pd.DataFrame(pdf_arr, index=x.index, columns=x.columns)\n\n def log_pdf(self, x):\n \"\"\"Logarithmic probability density function.\"\"\"\n d = self.loc[x.index, x.columns]\n lpdf_arr = loggamma((d._df + 1) / 2)\n lpdf_arr = lpdf_arr - 0.5 * np.log(d._df * np.pi)\n lpdf_arr = lpdf_arr - loggamma(d._df / 2)\n lpdf_arr = lpdf_arr - ((d._df + 1) / 2) * np.log(1 + x**2 / d._df)\n return pd.DataFrame(lpdf_arr, index=x.index, columns=x.columns)\n\n def cdf(self, x):\n \"\"\"Cumulative distribution function.\"\"\"\n d = self.loc[x.index, x.columns]\n cdf_arr = x * gamma((d._df + 1) / 2)\n cdf_arr = cdf_arr * hyp2f1(0.5, (d._df + 1) / 2, 3 / 2, -(x**2) / d._df)\n cdf_arr = 0.5 + cdf_arr / (np.sqrt(np.pi * d._df) * gamma(d._df / 2))\n return pd.DataFrame(cdf_arr, index=x.index, columns=x.columns)\n\n def ppf(self, p):\n \"\"\"Quantile function = percent point function = inverse cdf.\"\"\"\n d = self.loc[p.index, p.columns]\n ppf_arr = p.to_numpy(copy=True)\n ppf_arr[p.values == 0.5] = 0.0\n ppf_arr[p.values <= 0] = -np.inf\n ppf_arr[p.values >= 1] = np.inf\n\n mask1 = (p.values < 0.5) & (p.values > 0)\n mask2 = (p.values < 1) & (p.values > 0.5)\n ppf_arr[mask1] = 1 / betaincinv(0.5 * d._df[mask1], 0.5, 2 * ppf_arr[mask1])\n ppf_arr[mask2] = 1 / betaincinv(\n 0.5 * d._df[mask2], 0.5, 2 * (1 - ppf_arr[mask2])\n )\n ppf_arr[mask1 | mask2] = np.sqrt(ppf_arr[mask1 | mask2] - 1)\n ppf_arr[mask1 | mask2] = np.sqrt(d._df[mask1 | mask2]) * ppf_arr[mask1 | mask2]\n ppf_arr[mask1] = -ppf_arr[mask1]\n return pd.DataFrame(ppf_arr, index=p.index, columns=p.columns)\n\n @classmethod\n def get_test_params(cls, parameter_set=\"default\"):\n \"\"\"Return testing parameter settings for the estimator.\"\"\"\n params1 = {\"mu\": [[0, 1], [2, 3], [4, 5]], \"sigma\": 1}\n params2 = {\n \"mu\": 0,\n \"sigma\": 1,\n \"index\": pd.Index([1, 2, 5]),\n \"columns\": pd.Index([\"a\", \"b\"]),\n }\n return [params1, params2]\n", "path": "sktime/proba/t.py"}]}
3,024
915
gh_patches_debug_27773
rasdani/github-patches
git_diff
pypa__pip-12569
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> Show 'home-page' project URL when Home-Page metadata value is not set ### What's the problem this feature will solve? When installing a pyproject.toml-only project, the `Home-Page` metadata won't be set (as there's no way to specify it: PEP 621 only maps to `Project-Url`). This means `pip show` won't provide any URL to the user. ### Describe the solution you'd like If `Home-Page` is not set, instead select one of the project URLs (if present), with the following order of preference: * Home-page * Homepage * homepage * Source * fallback: the first project-URL in the metadata ### Alternative Solutions `pip show -v` shows all project URLs, but that still shows an empty `Home-page:` field ### Additional context Project URLs were included in `pip show -v` in #10827 ### Code of Conduct - [X] I agree to follow the [PSF Code of Conduct](https://www.python.org/psf/conduct/). </issue> <code> [start of src/pip/_internal/commands/show.py] 1 import logging 2 from optparse import Values 3 from typing import Generator, Iterable, Iterator, List, NamedTuple, Optional 4 5 from pip._vendor.packaging.utils import canonicalize_name 6 7 from pip._internal.cli.base_command import Command 8 from pip._internal.cli.status_codes import ERROR, SUCCESS 9 from pip._internal.metadata import BaseDistribution, get_default_environment 10 from pip._internal.utils.misc import write_output 11 12 logger = logging.getLogger(__name__) 13 14 15 class ShowCommand(Command): 16 """ 17 Show information about one or more installed packages. 18 19 The output is in RFC-compliant mail header format. 20 """ 21 22 usage = """ 23 %prog [options] <package> ...""" 24 ignore_require_venv = True 25 26 def add_options(self) -> None: 27 self.cmd_opts.add_option( 28 "-f", 29 "--files", 30 dest="files", 31 action="store_true", 32 default=False, 33 help="Show the full list of installed files for each package.", 34 ) 35 36 self.parser.insert_option_group(0, self.cmd_opts) 37 38 def run(self, options: Values, args: List[str]) -> int: 39 if not args: 40 logger.warning("ERROR: Please provide a package name or names.") 41 return ERROR 42 query = args 43 44 results = search_packages_info(query) 45 if not print_results( 46 results, list_files=options.files, verbose=options.verbose 47 ): 48 return ERROR 49 return SUCCESS 50 51 52 class _PackageInfo(NamedTuple): 53 name: str 54 version: str 55 location: str 56 editable_project_location: Optional[str] 57 requires: List[str] 58 required_by: List[str] 59 installer: str 60 metadata_version: str 61 classifiers: List[str] 62 summary: str 63 homepage: str 64 project_urls: List[str] 65 author: str 66 author_email: str 67 license: str 68 entry_points: List[str] 69 files: Optional[List[str]] 70 71 72 def search_packages_info(query: List[str]) -> Generator[_PackageInfo, None, None]: 73 """ 74 Gather details from installed distributions. Print distribution name, 75 version, location, and installed files. Installed files requires a 76 pip generated 'installed-files.txt' in the distributions '.egg-info' 77 directory. 78 """ 79 env = get_default_environment() 80 81 installed = {dist.canonical_name: dist for dist in env.iter_all_distributions()} 82 query_names = [canonicalize_name(name) for name in query] 83 missing = sorted( 84 [name for name, pkg in zip(query, query_names) if pkg not in installed] 85 ) 86 if missing: 87 logger.warning("Package(s) not found: %s", ", ".join(missing)) 88 89 def _get_requiring_packages(current_dist: BaseDistribution) -> Iterator[str]: 90 return ( 91 dist.metadata["Name"] or "UNKNOWN" 92 for dist in installed.values() 93 if current_dist.canonical_name 94 in {canonicalize_name(d.name) for d in dist.iter_dependencies()} 95 ) 96 97 for query_name in query_names: 98 try: 99 dist = installed[query_name] 100 except KeyError: 101 continue 102 103 requires = sorted( 104 # Avoid duplicates in requirements (e.g. due to environment markers). 105 {req.name for req in dist.iter_dependencies()}, 106 key=str.lower, 107 ) 108 required_by = sorted(_get_requiring_packages(dist), key=str.lower) 109 110 try: 111 entry_points_text = dist.read_text("entry_points.txt") 112 entry_points = entry_points_text.splitlines(keepends=False) 113 except FileNotFoundError: 114 entry_points = [] 115 116 files_iter = dist.iter_declared_entries() 117 if files_iter is None: 118 files: Optional[List[str]] = None 119 else: 120 files = sorted(files_iter) 121 122 metadata = dist.metadata 123 124 yield _PackageInfo( 125 name=dist.raw_name, 126 version=str(dist.version), 127 location=dist.location or "", 128 editable_project_location=dist.editable_project_location, 129 requires=requires, 130 required_by=required_by, 131 installer=dist.installer, 132 metadata_version=dist.metadata_version or "", 133 classifiers=metadata.get_all("Classifier", []), 134 summary=metadata.get("Summary", ""), 135 homepage=metadata.get("Home-page", ""), 136 project_urls=metadata.get_all("Project-URL", []), 137 author=metadata.get("Author", ""), 138 author_email=metadata.get("Author-email", ""), 139 license=metadata.get("License", ""), 140 entry_points=entry_points, 141 files=files, 142 ) 143 144 145 def print_results( 146 distributions: Iterable[_PackageInfo], 147 list_files: bool, 148 verbose: bool, 149 ) -> bool: 150 """ 151 Print the information from installed distributions found. 152 """ 153 results_printed = False 154 for i, dist in enumerate(distributions): 155 results_printed = True 156 if i > 0: 157 write_output("---") 158 159 write_output("Name: %s", dist.name) 160 write_output("Version: %s", dist.version) 161 write_output("Summary: %s", dist.summary) 162 write_output("Home-page: %s", dist.homepage) 163 write_output("Author: %s", dist.author) 164 write_output("Author-email: %s", dist.author_email) 165 write_output("License: %s", dist.license) 166 write_output("Location: %s", dist.location) 167 if dist.editable_project_location is not None: 168 write_output( 169 "Editable project location: %s", dist.editable_project_location 170 ) 171 write_output("Requires: %s", ", ".join(dist.requires)) 172 write_output("Required-by: %s", ", ".join(dist.required_by)) 173 174 if verbose: 175 write_output("Metadata-Version: %s", dist.metadata_version) 176 write_output("Installer: %s", dist.installer) 177 write_output("Classifiers:") 178 for classifier in dist.classifiers: 179 write_output(" %s", classifier) 180 write_output("Entry-points:") 181 for entry in dist.entry_points: 182 write_output(" %s", entry.strip()) 183 write_output("Project-URLs:") 184 for project_url in dist.project_urls: 185 write_output(" %s", project_url) 186 if list_files: 187 write_output("Files:") 188 if dist.files is None: 189 write_output("Cannot locate RECORD or installed-files.txt") 190 else: 191 for line in dist.files: 192 write_output(" %s", line.strip()) 193 return results_printed 194 [end of src/pip/_internal/commands/show.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/src/pip/_internal/commands/show.py b/src/pip/_internal/commands/show.py --- a/src/pip/_internal/commands/show.py +++ b/src/pip/_internal/commands/show.py @@ -121,6 +121,22 @@ metadata = dist.metadata + project_urls = metadata.get_all("Project-URL", []) + homepage = metadata.get("Home-page", "") + if not homepage: + # It's common that there is a "homepage" Project-URL, but Home-page + # remains unset (especially as PEP 621 doesn't surface the field). + # + # This logic was taken from PyPI's codebase. + for url in project_urls: + url_label, url = url.split(",", maxsplit=1) + normalized_label = ( + url_label.casefold().replace("-", "").replace("_", "").strip() + ) + if normalized_label == "homepage": + homepage = url.strip() + break + yield _PackageInfo( name=dist.raw_name, version=str(dist.version), @@ -132,8 +148,8 @@ metadata_version=dist.metadata_version or "", classifiers=metadata.get_all("Classifier", []), summary=metadata.get("Summary", ""), - homepage=metadata.get("Home-page", ""), - project_urls=metadata.get_all("Project-URL", []), + homepage=homepage, + project_urls=project_urls, author=metadata.get("Author", ""), author_email=metadata.get("Author-email", ""), license=metadata.get("License", ""),
{"golden_diff": "diff --git a/src/pip/_internal/commands/show.py b/src/pip/_internal/commands/show.py\n--- a/src/pip/_internal/commands/show.py\n+++ b/src/pip/_internal/commands/show.py\n@@ -121,6 +121,22 @@\n \n metadata = dist.metadata\n \n+ project_urls = metadata.get_all(\"Project-URL\", [])\n+ homepage = metadata.get(\"Home-page\", \"\")\n+ if not homepage:\n+ # It's common that there is a \"homepage\" Project-URL, but Home-page\n+ # remains unset (especially as PEP 621 doesn't surface the field).\n+ #\n+ # This logic was taken from PyPI's codebase.\n+ for url in project_urls:\n+ url_label, url = url.split(\",\", maxsplit=1)\n+ normalized_label = (\n+ url_label.casefold().replace(\"-\", \"\").replace(\"_\", \"\").strip()\n+ )\n+ if normalized_label == \"homepage\":\n+ homepage = url.strip()\n+ break\n+\n yield _PackageInfo(\n name=dist.raw_name,\n version=str(dist.version),\n@@ -132,8 +148,8 @@\n metadata_version=dist.metadata_version or \"\",\n classifiers=metadata.get_all(\"Classifier\", []),\n summary=metadata.get(\"Summary\", \"\"),\n- homepage=metadata.get(\"Home-page\", \"\"),\n- project_urls=metadata.get_all(\"Project-URL\", []),\n+ homepage=homepage,\n+ project_urls=project_urls,\n author=metadata.get(\"Author\", \"\"),\n author_email=metadata.get(\"Author-email\", \"\"),\n license=metadata.get(\"License\", \"\"),\n", "issue": "Show 'home-page' project URL when Home-Page metadata value is not set\n### What's the problem this feature will solve?\n\nWhen installing a pyproject.toml-only project, the `Home-Page` metadata won't be set (as there's no way to specify it: PEP 621 only maps to `Project-Url`). This means `pip show` won't provide any URL to the user.\n\n### Describe the solution you'd like\n\nIf `Home-Page` is not set, instead select one of the project URLs (if present), with the following order of preference:\r\n* Home-page\r\n* Homepage\r\n* homepage\r\n* Source\r\n* fallback: the first project-URL in the metadata\n\n### Alternative Solutions\n\n`pip show -v` shows all project URLs, but that still shows an empty `Home-page:` field\n\n### Additional context\n\nProject URLs were included in `pip show -v` in #10827\n\n### Code of Conduct\n\n- [X] I agree to follow the [PSF Code of Conduct](https://www.python.org/psf/conduct/).\n", "before_files": [{"content": "import logging\nfrom optparse import Values\nfrom typing import Generator, Iterable, Iterator, List, NamedTuple, Optional\n\nfrom pip._vendor.packaging.utils import canonicalize_name\n\nfrom pip._internal.cli.base_command import Command\nfrom pip._internal.cli.status_codes import ERROR, SUCCESS\nfrom pip._internal.metadata import BaseDistribution, get_default_environment\nfrom pip._internal.utils.misc import write_output\n\nlogger = logging.getLogger(__name__)\n\n\nclass ShowCommand(Command):\n \"\"\"\n Show information about one or more installed packages.\n\n The output is in RFC-compliant mail header format.\n \"\"\"\n\n usage = \"\"\"\n %prog [options] <package> ...\"\"\"\n ignore_require_venv = True\n\n def add_options(self) -> None:\n self.cmd_opts.add_option(\n \"-f\",\n \"--files\",\n dest=\"files\",\n action=\"store_true\",\n default=False,\n help=\"Show the full list of installed files for each package.\",\n )\n\n self.parser.insert_option_group(0, self.cmd_opts)\n\n def run(self, options: Values, args: List[str]) -> int:\n if not args:\n logger.warning(\"ERROR: Please provide a package name or names.\")\n return ERROR\n query = args\n\n results = search_packages_info(query)\n if not print_results(\n results, list_files=options.files, verbose=options.verbose\n ):\n return ERROR\n return SUCCESS\n\n\nclass _PackageInfo(NamedTuple):\n name: str\n version: str\n location: str\n editable_project_location: Optional[str]\n requires: List[str]\n required_by: List[str]\n installer: str\n metadata_version: str\n classifiers: List[str]\n summary: str\n homepage: str\n project_urls: List[str]\n author: str\n author_email: str\n license: str\n entry_points: List[str]\n files: Optional[List[str]]\n\n\ndef search_packages_info(query: List[str]) -> Generator[_PackageInfo, None, None]:\n \"\"\"\n Gather details from installed distributions. Print distribution name,\n version, location, and installed files. Installed files requires a\n pip generated 'installed-files.txt' in the distributions '.egg-info'\n directory.\n \"\"\"\n env = get_default_environment()\n\n installed = {dist.canonical_name: dist for dist in env.iter_all_distributions()}\n query_names = [canonicalize_name(name) for name in query]\n missing = sorted(\n [name for name, pkg in zip(query, query_names) if pkg not in installed]\n )\n if missing:\n logger.warning(\"Package(s) not found: %s\", \", \".join(missing))\n\n def _get_requiring_packages(current_dist: BaseDistribution) -> Iterator[str]:\n return (\n dist.metadata[\"Name\"] or \"UNKNOWN\"\n for dist in installed.values()\n if current_dist.canonical_name\n in {canonicalize_name(d.name) for d in dist.iter_dependencies()}\n )\n\n for query_name in query_names:\n try:\n dist = installed[query_name]\n except KeyError:\n continue\n\n requires = sorted(\n # Avoid duplicates in requirements (e.g. due to environment markers).\n {req.name for req in dist.iter_dependencies()},\n key=str.lower,\n )\n required_by = sorted(_get_requiring_packages(dist), key=str.lower)\n\n try:\n entry_points_text = dist.read_text(\"entry_points.txt\")\n entry_points = entry_points_text.splitlines(keepends=False)\n except FileNotFoundError:\n entry_points = []\n\n files_iter = dist.iter_declared_entries()\n if files_iter is None:\n files: Optional[List[str]] = None\n else:\n files = sorted(files_iter)\n\n metadata = dist.metadata\n\n yield _PackageInfo(\n name=dist.raw_name,\n version=str(dist.version),\n location=dist.location or \"\",\n editable_project_location=dist.editable_project_location,\n requires=requires,\n required_by=required_by,\n installer=dist.installer,\n metadata_version=dist.metadata_version or \"\",\n classifiers=metadata.get_all(\"Classifier\", []),\n summary=metadata.get(\"Summary\", \"\"),\n homepage=metadata.get(\"Home-page\", \"\"),\n project_urls=metadata.get_all(\"Project-URL\", []),\n author=metadata.get(\"Author\", \"\"),\n author_email=metadata.get(\"Author-email\", \"\"),\n license=metadata.get(\"License\", \"\"),\n entry_points=entry_points,\n files=files,\n )\n\n\ndef print_results(\n distributions: Iterable[_PackageInfo],\n list_files: bool,\n verbose: bool,\n) -> bool:\n \"\"\"\n Print the information from installed distributions found.\n \"\"\"\n results_printed = False\n for i, dist in enumerate(distributions):\n results_printed = True\n if i > 0:\n write_output(\"---\")\n\n write_output(\"Name: %s\", dist.name)\n write_output(\"Version: %s\", dist.version)\n write_output(\"Summary: %s\", dist.summary)\n write_output(\"Home-page: %s\", dist.homepage)\n write_output(\"Author: %s\", dist.author)\n write_output(\"Author-email: %s\", dist.author_email)\n write_output(\"License: %s\", dist.license)\n write_output(\"Location: %s\", dist.location)\n if dist.editable_project_location is not None:\n write_output(\n \"Editable project location: %s\", dist.editable_project_location\n )\n write_output(\"Requires: %s\", \", \".join(dist.requires))\n write_output(\"Required-by: %s\", \", \".join(dist.required_by))\n\n if verbose:\n write_output(\"Metadata-Version: %s\", dist.metadata_version)\n write_output(\"Installer: %s\", dist.installer)\n write_output(\"Classifiers:\")\n for classifier in dist.classifiers:\n write_output(\" %s\", classifier)\n write_output(\"Entry-points:\")\n for entry in dist.entry_points:\n write_output(\" %s\", entry.strip())\n write_output(\"Project-URLs:\")\n for project_url in dist.project_urls:\n write_output(\" %s\", project_url)\n if list_files:\n write_output(\"Files:\")\n if dist.files is None:\n write_output(\"Cannot locate RECORD or installed-files.txt\")\n else:\n for line in dist.files:\n write_output(\" %s\", line.strip())\n return results_printed\n", "path": "src/pip/_internal/commands/show.py"}]}
2,617
357
gh_patches_debug_42271
rasdani/github-patches
git_diff
google__flax-2204
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> codediff section not respecting html container In the new `sphinx_book_theme` our custom `codediff` directive is rendering its content outside the container, overflowing until the end of the page. As shown in this screenshot, it appear to be rendered underneath the `contents` section because its not respecting its section boundaries: ![image](https://user-images.githubusercontent.com/5862228/173973531-5e548ee8-53f6-4978-9744-f9a3a54b5470.png) </issue> <code> [start of docs/conf.py] 1 # Copyright 2022 The Flax Authors. 2 # 3 # Licensed under the Apache License, Version 2.0 (the "License"); 4 # you may not use this file except in compliance with the License. 5 # You may obtain a copy of the License at 6 # 7 # http://www.apache.org/licenses/LICENSE-2.0 8 # 9 # Unless required by applicable law or agreed to in writing, software 10 # distributed under the License is distributed on an "AS IS" BASIS, 11 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 12 # See the License for the specific language governing permissions and 13 # limitations under the License. 14 15 """Configuration file for the Sphinx documentation builder.""" 16 17 18 # This file only contains a selection of the most common options. For a full 19 # list see the documentation: 20 # https://www.sphinx-doc.org/en/master/usage/configuration.html 21 22 # -- Path setup -------------------------------------------------------------- 23 24 # If extensions (or modules to document with autodoc) are in another directory, 25 # add these directories to sys.path here. If the directory is relative to the 26 # documentation root, use os.path.abspath to make it absolute, like shown here. 27 # 28 # import os 29 # import sys 30 # sys.path.insert(0, os.path.abspath('.')) 31 32 import os 33 import sys 34 sys.path.insert(0, os.path.abspath('..')) 35 # Include local extension. 36 sys.path.append(os.path.abspath('./_ext')) 37 38 # patch sphinx 39 import docs.conf_sphinx_patch 40 # -- Project information ----------------------------------------------------- 41 42 project = 'Flax' 43 copyright = '2020, The Flax authors' # pylint: disable=redefined-builtin 44 author = 'The Flax authors' 45 46 47 # -- General configuration --------------------------------------------------- 48 49 # Add any Sphinx extension module names here, as strings. They can be 50 # extensions coming with Sphinx (named 'sphinx.ext.*') or your custom 51 # ones. 52 extensions = [ 53 'sphinx.ext.autodoc', 54 'sphinx.ext.autosummary', 55 'sphinx.ext.autosectionlabel', 56 'sphinx.ext.doctest', 57 'sphinx.ext.intersphinx', 58 'sphinx.ext.mathjax', 59 'sphinx.ext.napoleon', 60 'sphinx.ext.viewcode', 61 'myst_nb', 62 'codediff', 63 'sphinx_markdown_tables' 64 ] 65 66 # Add any paths that contain templates here, relative to this directory. 67 templates_path = ['_templates'] 68 69 # List of patterns, relative to source directory, that match files and 70 # directories to ignore when looking for source files. 71 # This pattern also affects html_static_path and html_extra_path. 72 exclude_patterns = ['_build', 'Thumbs.db', '.DS_Store'] 73 74 # The suffix(es) of source filenames. 75 # Note: important to list ipynb before md here: we have both md and ipynb 76 # copies of each notebook, and myst will choose which to convert based on 77 # the order in the source_suffix list. Notebooks which are not executed have 78 # outputs stored in ipynb but not in md, so we must convert the ipynb. 79 source_suffix = ['.rst', '.ipynb', '.md'] 80 81 autosummary_generate = True 82 83 master_doc = 'index' 84 85 autodoc_typehints = 'none' 86 87 88 # -- Options for HTML output ------------------------------------------------- 89 90 # The theme to use for HTML and HTML Help pages. See the documentation for 91 # a list of builtin themes. 92 # 93 html_theme = 'sphinx_book_theme' 94 # html_style = 'css/flax_theme.css' 95 96 # The name of an image file (relative to this directory) to place at the top 97 # of the sidebar. 98 html_logo = './flax.png' 99 100 # Add any paths that contain custom static files (such as style sheets) here, 101 # relative to this directory. They are copied after the builtin static files, 102 # so a file named "default.css" will overwrite the builtin "default.css". 103 html_static_path = ['_static'] 104 105 html_theme_options = { 106 "repository_url": "https://github.com/google/flax", 107 "use_repository_button": True, # add a "link to repository" button 108 "use_issues_button": False, # add an "Open an Issue" button 109 "path_to_docs": "docs", # used to compute the path to launch notebooks in colab 110 "launch_buttons": { 111 "colab_url": "https://colab.research.google.com/", 112 }, 113 } 114 115 # -- Options for myst ---------------------------------------------- 116 # Notebook cell execution timeout; defaults to 30. 117 nb_execution_timeout = 100 118 # List of patterns, relative to source directory, that match notebook 119 # files that will not be executed. 120 nb_execution_excludepatterns = [ 121 'notebooks/annotated_mnist.ipynb', # <-- times out 122 ] 123 # raise exceptions on execution so CI can catch errors 124 nb_execution_allow_errors = False 125 nb_execution_raise_on_error = True 126 127 # -- Extension configuration ------------------------------------------------- 128 129 # Tell sphinx-autodoc-typehints to generate stub parameter annotations including 130 # types, even if the parameters aren't explicitly documented. 131 always_document_param_types = True 132 [end of docs/conf.py] [start of docs/_ext/codediff.py] 1 # Copyright 2022 The Flax Authors. 2 # 3 # Licensed under the Apache License, Version 2.0 (the "License"); 4 # you may not use this file except in compliance with the License. 5 # You may obtain a copy of the License at 6 # 7 # http://www.apache.org/licenses/LICENSE-2.0 8 # 9 # Unless required by applicable law or agreed to in writing, software 10 # distributed under the License is distributed on an "AS IS" BASIS, 11 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 12 # See the License for the specific language governing permissions and 13 # limitations under the License. 14 15 """Sphinx directive for creating code diff tables. 16 17 Use directive as follows: 18 19 .. codediff:: 20 :title_left: <LEFT_CODE_BLOCK_TITLE> 21 :title_right: <RIGHT_CODE_BLOCK_TITLE> 22 23 <CODE_BLOCK_LEFT> 24 --- 25 <CODE_BLOCK_RIGHT> 26 27 In order to highlight a line of code, prepend it with "#!". 28 """ 29 import itertools 30 31 from docutils import nodes 32 from docutils.parsers.rst import directives 33 from docutils.statemachine import ViewList 34 35 import sphinx 36 from sphinx.util.docutils import SphinxDirective 37 38 39 class CodeDiffParser: 40 41 def parse(self, lines, title_left='Base', title_right='Diff', code_sep='---'): 42 if code_sep not in lines: 43 raise ValueError('Code separator not found! Code snippets should be ' 44 f'separated by {code_sep}.') 45 idx = lines.index(code_sep) 46 code_left = self._code_block(lines[0: idx]) 47 test_code = lines[idx+1:] 48 code_right = self._code_block(test_code) 49 50 self.max_left = max(len(x) for x in code_left + [title_left]) 51 self.max_right = max(len(x) for x in code_right + [title_right]) 52 53 output = [ 54 self._hline(), 55 self._table_row(title_left, title_right), 56 self._hline(), 57 ] 58 59 for l, r in itertools.zip_longest(code_left, code_right, fillvalue=''): 60 output += [self._table_row(l, r)] 61 62 return output + [self._hline()], test_code 63 64 def _code_block(self, lines): 65 """Creates a codeblock.""" 66 # Remove right trailing whitespace so we can detect the comments. 67 lines = [x.rstrip() for x in lines] 68 highlight = lambda x: x.endswith('#!') 69 code = map(lambda x: x[:-2].rstrip() if highlight(x) else x, lines) 70 highlights = [i + 1 for i in range(len(lines)) if highlight(lines[i])] 71 highlights = ','.join(str(i) for i in highlights) 72 73 directive = ['.. code-block:: python'] 74 if highlights: 75 directive += [f' :emphasize-lines: {highlights}'] 76 77 # Indent code and add empty line so the code is picked up by the directive. 78 return directive + [''] + list(map(lambda x: ' ' + x, code)) 79 80 def _hline(self): 81 return '+' + '-'*(self.max_left+2) + '+' + '-'*(self.max_right+2) + '+' 82 83 def _rfill(self, text, max_len): 84 return text + ' ' * (max_len-len(text)) 85 86 def _table_row(self, left, right): 87 text_left = self._rfill(left, self.max_left) 88 text_right = self._rfill(right, self.max_right) 89 return '| ' + text_left + ' | ' + text_right + ' |' 90 91 92 class CodeDiffDirective(SphinxDirective): 93 has_content = True 94 option_spec = { 95 'title_left': directives.unchanged, 96 'title_right': directives.unchanged, 97 'code_sep': directives.unchanged, 98 } 99 100 def run(self): 101 table_code, test_code = CodeDiffParser().parse( 102 list(self.content), **self.options) 103 104 # Create a test node as a comment node so it won't show up in the docs. 105 # We add attribute "testnodetype" so it is be picked up by the doctest 106 # builder. This functionality is not officially documented but can be found 107 # in the source code: 108 # https://github.com/sphinx-doc/sphinx/blob/3.x/sphinx/ext/doctest.py 109 # (search for 'testnodetype'). 110 test_code = '\n'.join(test_code) 111 test_node = nodes.comment(test_code, test_code, testnodetype='testcode') 112 # Set the source info so the error message is correct when testing. 113 self.set_source_info(test_node) 114 test_node['options'] = {} 115 test_node['language'] = 'python3' 116 117 # The table node is the side-by-side diff view that will be shown on RTD. 118 table_node = nodes.paragraph() 119 self.content = ViewList(table_code, self.content.parent) 120 self.state.nested_parse(self.content, self.content_offset, table_node) 121 122 return [table_node, test_node] 123 124 125 def setup(app): 126 app.add_directive('codediff', CodeDiffDirective) 127 128 return { 129 'version': sphinx.__display_version__, 130 'parallel_read_safe': True, 131 'parallel_write_safe': True, 132 } 133 [end of docs/_ext/codediff.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/docs/_ext/codediff.py b/docs/_ext/codediff.py --- a/docs/_ext/codediff.py +++ b/docs/_ext/codediff.py @@ -24,9 +24,10 @@ --- <CODE_BLOCK_RIGHT> -In order to highlight a line of code, prepend it with "#!". +In order to highlight a line of code, append "#!" to it. """ import itertools +from typing import List, Tuple from docutils import nodes from docutils.parsers.rst import directives @@ -35,10 +36,14 @@ import sphinx from sphinx.util.docutils import SphinxDirective +MISSING = object() class CodeDiffParser: - def parse(self, lines, title_left='Base', title_right='Diff', code_sep='---'): + def parse( + self, lines, title_left='Base', title_right='Diff', code_sep='---', sync=MISSING): + sync = sync is not MISSING + if code_sep not in lines: raise ValueError('Code separator not found! Code snippets should be ' f'separated by {code_sep}.') @@ -47,19 +52,10 @@ test_code = lines[idx+1:] code_right = self._code_block(test_code) - self.max_left = max(len(x) for x in code_left + [title_left]) - self.max_right = max(len(x) for x in code_right + [title_right]) - - output = [ - self._hline(), - self._table_row(title_left, title_right), - self._hline(), - ] + output = self._tabs( + (title_left, code_left), (title_right, code_right), sync=sync) - for l, r in itertools.zip_longest(code_left, code_right, fillvalue=''): - output += [self._table_row(l, r)] - - return output + [self._hline()], test_code + return output, test_code def _code_block(self, lines): """Creates a codeblock.""" @@ -77,17 +73,20 @@ # Indent code and add empty line so the code is picked up by the directive. return directive + [''] + list(map(lambda x: ' ' + x, code)) - def _hline(self): - return '+' + '-'*(self.max_left+2) + '+' + '-'*(self.max_right+2) + '+' - - def _rfill(self, text, max_len): - return text + ' ' * (max_len-len(text)) + def _tabs(self, *contents: Tuple[str, List[str]], sync): + output = ['.. tab-set::'] + [' '] + + for title, content in contents: + output += [f' .. tab-item:: {title}'] + + if sync: + key = title.strip() + output += [f' :sync: {key}'] - def _table_row(self, left, right): - text_left = self._rfill(left, self.max_left) - text_right = self._rfill(right, self.max_right) - return '| ' + text_left + ' | ' + text_right + ' |' + output += [' '] + output += [' ' + line for line in content] + return output class CodeDiffDirective(SphinxDirective): has_content = True @@ -95,6 +94,7 @@ 'title_left': directives.unchanged, 'title_right': directives.unchanged, 'code_sep': directives.unchanged, + 'sync': directives.flag, } def run(self): diff --git a/docs/conf.py b/docs/conf.py --- a/docs/conf.py +++ b/docs/conf.py @@ -60,7 +60,7 @@ 'sphinx.ext.viewcode', 'myst_nb', 'codediff', - 'sphinx_markdown_tables' + 'sphinx_design', ] # Add any paths that contain templates here, relative to this directory. @@ -91,7 +91,7 @@ # a list of builtin themes. # html_theme = 'sphinx_book_theme' -# html_style = 'css/flax_theme.css' +html_css_files = ["css/flax_theme.css"] # The name of an image file (relative to this directory) to place at the top # of the sidebar.
{"golden_diff": "diff --git a/docs/_ext/codediff.py b/docs/_ext/codediff.py\n--- a/docs/_ext/codediff.py\n+++ b/docs/_ext/codediff.py\n@@ -24,9 +24,10 @@\n ---\n <CODE_BLOCK_RIGHT>\n \n-In order to highlight a line of code, prepend it with \"#!\".\n+In order to highlight a line of code, append \"#!\" to it.\n \"\"\"\n import itertools\n+from typing import List, Tuple\n \n from docutils import nodes\n from docutils.parsers.rst import directives\n@@ -35,10 +36,14 @@\n import sphinx\n from sphinx.util.docutils import SphinxDirective\n \n+MISSING = object()\n \n class CodeDiffParser:\n \n- def parse(self, lines, title_left='Base', title_right='Diff', code_sep='---'):\n+ def parse(\n+ self, lines, title_left='Base', title_right='Diff', code_sep='---', sync=MISSING):\n+ sync = sync is not MISSING\n+\n if code_sep not in lines:\n raise ValueError('Code separator not found! Code snippets should be '\n f'separated by {code_sep}.')\n@@ -47,19 +52,10 @@\n test_code = lines[idx+1:]\n code_right = self._code_block(test_code)\n \n- self.max_left = max(len(x) for x in code_left + [title_left])\n- self.max_right = max(len(x) for x in code_right + [title_right])\n-\n- output = [\n- self._hline(),\n- self._table_row(title_left, title_right),\n- self._hline(),\n- ]\n+ output = self._tabs(\n+ (title_left, code_left), (title_right, code_right), sync=sync)\n \n- for l, r in itertools.zip_longest(code_left, code_right, fillvalue=''):\n- output += [self._table_row(l, r)]\n-\n- return output + [self._hline()], test_code\n+ return output, test_code\n \n def _code_block(self, lines):\n \"\"\"Creates a codeblock.\"\"\"\n@@ -77,17 +73,20 @@\n # Indent code and add empty line so the code is picked up by the directive.\n return directive + [''] + list(map(lambda x: ' ' + x, code))\n \n- def _hline(self):\n- return '+' + '-'*(self.max_left+2) + '+' + '-'*(self.max_right+2) + '+'\n-\n- def _rfill(self, text, max_len):\n- return text + ' ' * (max_len-len(text))\n+ def _tabs(self, *contents: Tuple[str, List[str]], sync):\n+ output = ['.. tab-set::'] + [' ']\n+ \n+ for title, content in contents:\n+ output += [f' .. tab-item:: {title}']\n+ \n+ if sync:\n+ key = title.strip()\n+ output += [f' :sync: {key}']\n \n- def _table_row(self, left, right):\n- text_left = self._rfill(left, self.max_left)\n- text_right = self._rfill(right, self.max_right)\n- return '| ' + text_left + ' | ' + text_right + ' |'\n+ output += [' ']\n+ output += [' ' + line for line in content]\n \n+ return output\n \n class CodeDiffDirective(SphinxDirective):\n has_content = True\n@@ -95,6 +94,7 @@\n 'title_left': directives.unchanged,\n 'title_right': directives.unchanged,\n 'code_sep': directives.unchanged,\n+ 'sync': directives.flag,\n }\n \n def run(self):\ndiff --git a/docs/conf.py b/docs/conf.py\n--- a/docs/conf.py\n+++ b/docs/conf.py\n@@ -60,7 +60,7 @@\n 'sphinx.ext.viewcode',\n 'myst_nb',\n 'codediff',\n- 'sphinx_markdown_tables'\n+ 'sphinx_design',\n ]\n \n # Add any paths that contain templates here, relative to this directory.\n@@ -91,7 +91,7 @@\n # a list of builtin themes.\n #\n html_theme = 'sphinx_book_theme'\n-# html_style = 'css/flax_theme.css'\n+html_css_files = [\"css/flax_theme.css\"]\n \n # The name of an image file (relative to this directory) to place at the top\n # of the sidebar.\n", "issue": "codediff section not respecting html container \nIn the new `sphinx_book_theme` our custom `codediff` directive is rendering its content outside the container, overflowing until the end of the page. As shown in this screenshot, it appear to be rendered underneath the `contents` section because its not respecting its section boundaries:\r\n\r\n![image](https://user-images.githubusercontent.com/5862228/173973531-5e548ee8-53f6-4978-9744-f9a3a54b5470.png)\r\n\r\n\n", "before_files": [{"content": "# Copyright 2022 The Flax Authors.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\n\"\"\"Configuration file for the Sphinx documentation builder.\"\"\"\n\n\n# This file only contains a selection of the most common options. For a full\n# list see the documentation:\n# https://www.sphinx-doc.org/en/master/usage/configuration.html\n\n# -- Path setup --------------------------------------------------------------\n\n# If extensions (or modules to document with autodoc) are in another directory,\n# add these directories to sys.path here. If the directory is relative to the\n# documentation root, use os.path.abspath to make it absolute, like shown here.\n#\n# import os\n# import sys\n# sys.path.insert(0, os.path.abspath('.'))\n\nimport os\nimport sys\nsys.path.insert(0, os.path.abspath('..'))\n# Include local extension.\nsys.path.append(os.path.abspath('./_ext'))\n\n# patch sphinx\nimport docs.conf_sphinx_patch\n# -- Project information -----------------------------------------------------\n\nproject = 'Flax'\ncopyright = '2020, The Flax authors' # pylint: disable=redefined-builtin\nauthor = 'The Flax authors'\n\n\n# -- General configuration ---------------------------------------------------\n\n# Add any Sphinx extension module names here, as strings. They can be\n# extensions coming with Sphinx (named 'sphinx.ext.*') or your custom\n# ones.\nextensions = [\n 'sphinx.ext.autodoc',\n 'sphinx.ext.autosummary',\n 'sphinx.ext.autosectionlabel',\n 'sphinx.ext.doctest',\n 'sphinx.ext.intersphinx',\n 'sphinx.ext.mathjax',\n 'sphinx.ext.napoleon',\n 'sphinx.ext.viewcode',\n 'myst_nb',\n 'codediff',\n 'sphinx_markdown_tables'\n]\n\n# Add any paths that contain templates here, relative to this directory.\ntemplates_path = ['_templates']\n\n# List of patterns, relative to source directory, that match files and\n# directories to ignore when looking for source files.\n# This pattern also affects html_static_path and html_extra_path.\nexclude_patterns = ['_build', 'Thumbs.db', '.DS_Store']\n\n# The suffix(es) of source filenames.\n# Note: important to list ipynb before md here: we have both md and ipynb\n# copies of each notebook, and myst will choose which to convert based on\n# the order in the source_suffix list. Notebooks which are not executed have\n# outputs stored in ipynb but not in md, so we must convert the ipynb.\nsource_suffix = ['.rst', '.ipynb', '.md']\n\nautosummary_generate = True\n\nmaster_doc = 'index'\n\nautodoc_typehints = 'none'\n\n\n# -- Options for HTML output -------------------------------------------------\n\n# The theme to use for HTML and HTML Help pages. See the documentation for\n# a list of builtin themes.\n#\nhtml_theme = 'sphinx_book_theme'\n# html_style = 'css/flax_theme.css'\n\n# The name of an image file (relative to this directory) to place at the top\n# of the sidebar.\nhtml_logo = './flax.png'\n\n# Add any paths that contain custom static files (such as style sheets) here,\n# relative to this directory. They are copied after the builtin static files,\n# so a file named \"default.css\" will overwrite the builtin \"default.css\".\nhtml_static_path = ['_static']\n\nhtml_theme_options = {\n \"repository_url\": \"https://github.com/google/flax\",\n \"use_repository_button\": True, # add a \"link to repository\" button\n \"use_issues_button\": False, # add an \"Open an Issue\" button\n \"path_to_docs\": \"docs\", # used to compute the path to launch notebooks in colab\n \"launch_buttons\": {\n \"colab_url\": \"https://colab.research.google.com/\",\n },\n}\n\n# -- Options for myst ----------------------------------------------\n# Notebook cell execution timeout; defaults to 30.\nnb_execution_timeout = 100\n# List of patterns, relative to source directory, that match notebook\n# files that will not be executed.\nnb_execution_excludepatterns = [\n 'notebooks/annotated_mnist.ipynb', # <-- times out \n]\n# raise exceptions on execution so CI can catch errors\nnb_execution_allow_errors = False\nnb_execution_raise_on_error = True\n\n# -- Extension configuration -------------------------------------------------\n\n# Tell sphinx-autodoc-typehints to generate stub parameter annotations including\n# types, even if the parameters aren't explicitly documented.\nalways_document_param_types = True\n", "path": "docs/conf.py"}, {"content": "# Copyright 2022 The Flax Authors.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\n\"\"\"Sphinx directive for creating code diff tables.\n\nUse directive as follows:\n\n.. codediff::\n :title_left: <LEFT_CODE_BLOCK_TITLE>\n :title_right: <RIGHT_CODE_BLOCK_TITLE>\n\n <CODE_BLOCK_LEFT>\n ---\n <CODE_BLOCK_RIGHT>\n\nIn order to highlight a line of code, prepend it with \"#!\".\n\"\"\"\nimport itertools\n\nfrom docutils import nodes\nfrom docutils.parsers.rst import directives\nfrom docutils.statemachine import ViewList\n\nimport sphinx\nfrom sphinx.util.docutils import SphinxDirective\n\n\nclass CodeDiffParser:\n\n def parse(self, lines, title_left='Base', title_right='Diff', code_sep='---'):\n if code_sep not in lines:\n raise ValueError('Code separator not found! Code snippets should be '\n f'separated by {code_sep}.')\n idx = lines.index(code_sep)\n code_left = self._code_block(lines[0: idx])\n test_code = lines[idx+1:]\n code_right = self._code_block(test_code)\n\n self.max_left = max(len(x) for x in code_left + [title_left])\n self.max_right = max(len(x) for x in code_right + [title_right])\n\n output = [\n self._hline(),\n self._table_row(title_left, title_right),\n self._hline(),\n ]\n\n for l, r in itertools.zip_longest(code_left, code_right, fillvalue=''):\n output += [self._table_row(l, r)]\n\n return output + [self._hline()], test_code\n\n def _code_block(self, lines):\n \"\"\"Creates a codeblock.\"\"\"\n # Remove right trailing whitespace so we can detect the comments.\n lines = [x.rstrip() for x in lines]\n highlight = lambda x: x.endswith('#!')\n code = map(lambda x: x[:-2].rstrip() if highlight(x) else x, lines)\n highlights = [i + 1 for i in range(len(lines)) if highlight(lines[i])]\n highlights = ','.join(str(i) for i in highlights)\n\n directive = ['.. code-block:: python']\n if highlights:\n directive += [f' :emphasize-lines: {highlights}']\n\n # Indent code and add empty line so the code is picked up by the directive.\n return directive + [''] + list(map(lambda x: ' ' + x, code))\n\n def _hline(self):\n return '+' + '-'*(self.max_left+2) + '+' + '-'*(self.max_right+2) + '+'\n\n def _rfill(self, text, max_len):\n return text + ' ' * (max_len-len(text))\n\n def _table_row(self, left, right):\n text_left = self._rfill(left, self.max_left)\n text_right = self._rfill(right, self.max_right)\n return '| ' + text_left + ' | ' + text_right + ' |'\n\n\nclass CodeDiffDirective(SphinxDirective):\n has_content = True\n option_spec = {\n 'title_left': directives.unchanged,\n 'title_right': directives.unchanged,\n 'code_sep': directives.unchanged,\n }\n\n def run(self):\n table_code, test_code = CodeDiffParser().parse(\n list(self.content), **self.options)\n\n # Create a test node as a comment node so it won't show up in the docs.\n # We add attribute \"testnodetype\" so it is be picked up by the doctest\n # builder. This functionality is not officially documented but can be found\n # in the source code:\n # https://github.com/sphinx-doc/sphinx/blob/3.x/sphinx/ext/doctest.py\n # (search for 'testnodetype').\n test_code = '\\n'.join(test_code)\n test_node = nodes.comment(test_code, test_code, testnodetype='testcode')\n # Set the source info so the error message is correct when testing.\n self.set_source_info(test_node)\n test_node['options'] = {}\n test_node['language'] = 'python3'\n\n # The table node is the side-by-side diff view that will be shown on RTD.\n table_node = nodes.paragraph()\n self.content = ViewList(table_code, self.content.parent)\n self.state.nested_parse(self.content, self.content_offset, table_node)\n\n return [table_node, test_node]\n\n\ndef setup(app):\n app.add_directive('codediff', CodeDiffDirective)\n\n return {\n 'version': sphinx.__display_version__,\n 'parallel_read_safe': True,\n 'parallel_write_safe': True,\n }\n", "path": "docs/_ext/codediff.py"}]}
3,498
1,007
gh_patches_debug_20614
rasdani/github-patches
git_diff
pytorch__examples-1189
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> Add `save_model` arg to `mnist_hogwild` example Currently the example doesn't support the `--save_model` argument like the other examples </issue> <code> [start of mnist_hogwild/main.py] 1 from __future__ import print_function 2 import argparse 3 import torch 4 import torch.nn as nn 5 import torch.nn.functional as F 6 import torch.multiprocessing as mp 7 from torch.utils.data.sampler import Sampler 8 from torchvision import datasets, transforms 9 10 from train import train, test 11 12 # Training settings 13 parser = argparse.ArgumentParser(description='PyTorch MNIST Example') 14 parser.add_argument('--batch-size', type=int, default=64, metavar='N', 15 help='input batch size for training (default: 64)') 16 parser.add_argument('--test-batch-size', type=int, default=1000, metavar='N', 17 help='input batch size for testing (default: 1000)') 18 parser.add_argument('--epochs', type=int, default=10, metavar='N', 19 help='number of epochs to train (default: 10)') 20 parser.add_argument('--lr', type=float, default=0.01, metavar='LR', 21 help='learning rate (default: 0.01)') 22 parser.add_argument('--momentum', type=float, default=0.5, metavar='M', 23 help='SGD momentum (default: 0.5)') 24 parser.add_argument('--seed', type=int, default=1, metavar='S', 25 help='random seed (default: 1)') 26 parser.add_argument('--log-interval', type=int, default=10, metavar='N', 27 help='how many batches to wait before logging training status') 28 parser.add_argument('--num-processes', type=int, default=2, metavar='N', 29 help='how many training processes to use (default: 2)') 30 parser.add_argument('--cuda', action='store_true', default=False, 31 help='enables CUDA training') 32 parser.add_argument('--mps', action='store_true', default=False, 33 help='enables macOS GPU training') 34 parser.add_argument('--dry-run', action='store_true', default=False, 35 help='quickly check a single pass') 36 37 class Net(nn.Module): 38 def __init__(self): 39 super(Net, self).__init__() 40 self.conv1 = nn.Conv2d(1, 10, kernel_size=5) 41 self.conv2 = nn.Conv2d(10, 20, kernel_size=5) 42 self.conv2_drop = nn.Dropout2d() 43 self.fc1 = nn.Linear(320, 50) 44 self.fc2 = nn.Linear(50, 10) 45 46 def forward(self, x): 47 x = F.relu(F.max_pool2d(self.conv1(x), 2)) 48 x = F.relu(F.max_pool2d(self.conv2_drop(self.conv2(x)), 2)) 49 x = x.view(-1, 320) 50 x = F.relu(self.fc1(x)) 51 x = F.dropout(x, training=self.training) 52 x = self.fc2(x) 53 return F.log_softmax(x, dim=1) 54 55 56 if __name__ == '__main__': 57 args = parser.parse_args() 58 59 use_cuda = args.cuda and torch.cuda.is_available() 60 use_mps = args.mps and torch.backends.mps.is_available() 61 if use_cuda: 62 device = torch.device("cuda") 63 elif use_mps: 64 device = torch.device("mps") 65 else: 66 device = torch.device("cpu") 67 68 transform=transforms.Compose([ 69 transforms.ToTensor(), 70 transforms.Normalize((0.1307,), (0.3081,)) 71 ]) 72 dataset1 = datasets.MNIST('../data', train=True, download=True, 73 transform=transform) 74 dataset2 = datasets.MNIST('../data', train=False, 75 transform=transform) 76 kwargs = {'batch_size': args.batch_size, 77 'shuffle': True} 78 if use_cuda: 79 kwargs.update({'num_workers': 1, 80 'pin_memory': True, 81 }) 82 83 torch.manual_seed(args.seed) 84 mp.set_start_method('spawn', force=True) 85 86 model = Net().to(device) 87 model.share_memory() # gradients are allocated lazily, so they are not shared here 88 89 processes = [] 90 for rank in range(args.num_processes): 91 p = mp.Process(target=train, args=(rank, args, model, device, 92 dataset1, kwargs)) 93 # We first train the model across `num_processes` processes 94 p.start() 95 processes.append(p) 96 for p in processes: 97 p.join() 98 99 # Once training is complete, we can test the model 100 test(args, model, device, dataset2, kwargs) 101 [end of mnist_hogwild/main.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/mnist_hogwild/main.py b/mnist_hogwild/main.py --- a/mnist_hogwild/main.py +++ b/mnist_hogwild/main.py @@ -30,7 +30,9 @@ parser.add_argument('--cuda', action='store_true', default=False, help='enables CUDA training') parser.add_argument('--mps', action='store_true', default=False, - help='enables macOS GPU training') + help='enables macOS GPU training') +parser.add_argument('--save_model', action='store_true', default=False, + help='save the trained model to state_dict') parser.add_argument('--dry-run', action='store_true', default=False, help='quickly check a single pass') @@ -96,5 +98,8 @@ for p in processes: p.join() + if args.save_model: + torch.save(model.state_dict(), "MNIST_hogwild.pt") + # Once training is complete, we can test the model test(args, model, device, dataset2, kwargs)
{"golden_diff": "diff --git a/mnist_hogwild/main.py b/mnist_hogwild/main.py\n--- a/mnist_hogwild/main.py\n+++ b/mnist_hogwild/main.py\n@@ -30,7 +30,9 @@\n parser.add_argument('--cuda', action='store_true', default=False,\n help='enables CUDA training')\n parser.add_argument('--mps', action='store_true', default=False,\n- help='enables macOS GPU training')\n+ help='enables macOS GPU training')\n+parser.add_argument('--save_model', action='store_true', default=False,\n+ help='save the trained model to state_dict')\n parser.add_argument('--dry-run', action='store_true', default=False,\n help='quickly check a single pass')\n \n@@ -96,5 +98,8 @@\n for p in processes:\n p.join()\n \n+ if args.save_model:\n+ torch.save(model.state_dict(), \"MNIST_hogwild.pt\")\n+\n # Once training is complete, we can test the model\n test(args, model, device, dataset2, kwargs)\n", "issue": "Add `save_model` arg to `mnist_hogwild` example\nCurrently the example doesn't support the `--save_model` argument like the other examples\r\n\n", "before_files": [{"content": "from __future__ import print_function\nimport argparse\nimport torch\nimport torch.nn as nn\nimport torch.nn.functional as F\nimport torch.multiprocessing as mp\nfrom torch.utils.data.sampler import Sampler\nfrom torchvision import datasets, transforms\n\nfrom train import train, test\n\n# Training settings\nparser = argparse.ArgumentParser(description='PyTorch MNIST Example')\nparser.add_argument('--batch-size', type=int, default=64, metavar='N',\n help='input batch size for training (default: 64)')\nparser.add_argument('--test-batch-size', type=int, default=1000, metavar='N',\n help='input batch size for testing (default: 1000)')\nparser.add_argument('--epochs', type=int, default=10, metavar='N',\n help='number of epochs to train (default: 10)')\nparser.add_argument('--lr', type=float, default=0.01, metavar='LR',\n help='learning rate (default: 0.01)')\nparser.add_argument('--momentum', type=float, default=0.5, metavar='M',\n help='SGD momentum (default: 0.5)')\nparser.add_argument('--seed', type=int, default=1, metavar='S',\n help='random seed (default: 1)')\nparser.add_argument('--log-interval', type=int, default=10, metavar='N',\n help='how many batches to wait before logging training status')\nparser.add_argument('--num-processes', type=int, default=2, metavar='N',\n help='how many training processes to use (default: 2)')\nparser.add_argument('--cuda', action='store_true', default=False,\n help='enables CUDA training')\nparser.add_argument('--mps', action='store_true', default=False,\n help='enables macOS GPU training')\nparser.add_argument('--dry-run', action='store_true', default=False,\n help='quickly check a single pass')\n\nclass Net(nn.Module):\n def __init__(self):\n super(Net, self).__init__()\n self.conv1 = nn.Conv2d(1, 10, kernel_size=5)\n self.conv2 = nn.Conv2d(10, 20, kernel_size=5)\n self.conv2_drop = nn.Dropout2d()\n self.fc1 = nn.Linear(320, 50)\n self.fc2 = nn.Linear(50, 10)\n\n def forward(self, x):\n x = F.relu(F.max_pool2d(self.conv1(x), 2))\n x = F.relu(F.max_pool2d(self.conv2_drop(self.conv2(x)), 2))\n x = x.view(-1, 320)\n x = F.relu(self.fc1(x))\n x = F.dropout(x, training=self.training)\n x = self.fc2(x)\n return F.log_softmax(x, dim=1)\n\n\nif __name__ == '__main__':\n args = parser.parse_args()\n\n use_cuda = args.cuda and torch.cuda.is_available()\n use_mps = args.mps and torch.backends.mps.is_available()\n if use_cuda:\n device = torch.device(\"cuda\")\n elif use_mps:\n device = torch.device(\"mps\")\n else:\n device = torch.device(\"cpu\")\n\n transform=transforms.Compose([\n transforms.ToTensor(),\n transforms.Normalize((0.1307,), (0.3081,))\n ])\n dataset1 = datasets.MNIST('../data', train=True, download=True,\n transform=transform)\n dataset2 = datasets.MNIST('../data', train=False,\n transform=transform)\n kwargs = {'batch_size': args.batch_size,\n 'shuffle': True}\n if use_cuda:\n kwargs.update({'num_workers': 1,\n 'pin_memory': True,\n })\n\n torch.manual_seed(args.seed)\n mp.set_start_method('spawn', force=True)\n\n model = Net().to(device)\n model.share_memory() # gradients are allocated lazily, so they are not shared here\n\n processes = []\n for rank in range(args.num_processes):\n p = mp.Process(target=train, args=(rank, args, model, device,\n dataset1, kwargs))\n # We first train the model across `num_processes` processes\n p.start()\n processes.append(p)\n for p in processes:\n p.join()\n\n # Once training is complete, we can test the model\n test(args, model, device, dataset2, kwargs)\n", "path": "mnist_hogwild/main.py"}]}
1,736
235
gh_patches_debug_33732
rasdani/github-patches
git_diff
qtile__qtile-1883
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> widget.Pacman needs a non existing file:'checkupdates' ``` 2020-03-30 07:09:12,582 ERROR libqtile base.py:worker():L478 problem polling to update widget pacman Traceback (most recent call last): File "/home/parsrch/.local/lib/python3.8/site-packages/libqtile/widget/base.py", line 474, in worker text = self.poll() File "/home/parsrch/.local/lib/python3.8/site-packages/libqtile/widget/pacman.py", line 49, in poll pacman = self.call_process(['checkupdates']) File "/home/parsrch/.local/lib/python3.8/site-packages/libqtile/widget/base.py", line 266, in call_process output = subprocess.check_output(command, **kwargs) File "/usr/lib/python3.8/subprocess.py", line 411, in check_output return run(*popenargs, stdout=PIPE, timeout=timeout, check=True, File "/usr/lib/python3.8/subprocess.py", line 489, in run with Popen(*popenargs, **kwargs) as process: File "/usr/lib/python3.8/subprocess.py", line 854, in __init__ self._execute_child(args, executable, preexec_fn, close_fds, File "/usr/lib/python3.8/subprocess.py", line 1702, in _execute_child raise child_exception_type(errno_num, err_msg, err_filename) FileNotFoundError: [Errno 2] No such file or directory: 'checkupdates' ``` </issue> <code> [start of libqtile/widget/pacman.py] 1 # -*- coding:utf-8 -*- 2 # 3 # Copyright (C) 2012, Maximilian Köhl <[email protected]> 4 # 5 # This program is free software: you can redistribute it and/or modify 6 # it under the terms of the GNU General Public License as published by 7 # the Free Software Foundation, either version 3 of the License, or 8 # (at your option) any later version. 9 # 10 # This program is distributed in the hope that it will be useful, 11 # but WITHOUT ANY WARRANTY; without even the implied warranty of 12 # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the 13 # GNU General Public License for more details. 14 # 15 # You should have received a copy of the GNU General Public License 16 # along with this program. If not, see <http://www.gnu.org/licenses/>. 17 18 import subprocess 19 20 from libqtile.log_utils import logger 21 from libqtile.widget import base 22 23 24 class Pacman(base.ThreadedPollText): 25 """Shows number of available updates 26 27 Needs the pacman package manager installed. So will only work in Arch Linux 28 installation. 29 """ 30 orientations = base.ORIENTATION_HORIZONTAL 31 defaults = [ 32 ('unavailable', 'ffffff', 'Unavailable Color - no updates.'), 33 ('execute', None, 'Command to execute on click'), 34 ('update_interval', 60, "The update interval."), 35 ] 36 37 def __init__(self, **config): 38 logger.warning("Pacman is deprecated, please use CheckUpdates") 39 base.ThreadedPollText.__init__(self, **config) 40 self.add_defaults(Pacman.defaults) 41 42 def draw(self): 43 if self.text == '0': 44 self.layout.colour = self.unavailable 45 else: 46 self.layout.colour = self.foreground 47 base.ThreadedPollText.draw(self) 48 49 def poll(self): 50 pacman = self.call_process(['checkupdates']) 51 return str(len(pacman.splitlines())) 52 53 def button_press(self, x, y, button): 54 base.ThreadedPollText.button_press(self, x, y, button) 55 if button == 1 and self.execute is not None: 56 subprocess.Popen([self.execute], shell=True) 57 [end of libqtile/widget/pacman.py] [start of libqtile/widget/__init__.py] 1 # Copyright (c) 2014 Rock Neurotiko 2 # Copyright (c) 2014 roger 3 # Copyright (c) 2015 David R. Andersen 4 # 5 # Permission is hereby granted, free of charge, to any person obtaining a copy 6 # of this software and associated documentation files (the "Software"), to deal 7 # in the Software without restriction, including without limitation the rights 8 # to use, copy, modify, merge, publish, distribute, sublicense, and/or sell 9 # copies of the Software, and to permit persons to whom the Software is 10 # furnished to do so, subject to the following conditions: 11 # 12 # The above copyright notice and this permission notice shall be included in 13 # all copies or substantial portions of the Software. 14 # 15 # THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR 16 # IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, 17 # FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE 18 # AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER 19 # LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, 20 # OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE 21 # SOFTWARE. 22 23 from libqtile.utils import safe_import as safe_import_ 24 # only directly import widgets that do not have any third party dependencies 25 # other than those required by qtile, otherwise use the same import function 26 from libqtile.widget.base import Mirror # noqa: F401 27 from libqtile.widget.chord import Chord # noqa: F401 28 from libqtile.widget.clock import Clock # noqa: F401 29 from libqtile.widget.currentlayout import ( # noqa: F401 30 CurrentLayout, 31 CurrentLayoutIcon, 32 ) 33 from libqtile.widget.groupbox import AGroupBox, GroupBox # noqa: F401 34 from libqtile.widget.import_error import make_error 35 from libqtile.widget.prompt import Prompt # noqa: F401 36 from libqtile.widget.quick_exit import QuickExit # noqa: F401 37 from libqtile.widget.systray import Systray # noqa: F401 38 from libqtile.widget.textbox import TextBox # noqa: F401 39 from libqtile.widget.windowname import WindowName # noqa: F401 40 41 42 def safe_import(module_name, class_name): 43 safe_import_( 44 (".widget", module_name), class_name, globals(), fallback=make_error 45 ) 46 47 48 safe_import("backlight", "Backlight") 49 safe_import("battery", ["Battery", "BatteryIcon"]) 50 safe_import("currentscreen", "CurrentScreen") 51 safe_import("debuginfo", "DebugInfo") 52 safe_import( 53 "graph", 54 [ 55 "CPUGraph", 56 "MemoryGraph", 57 "SwapGraph", 58 "NetGraph", 59 "HDDGraph", 60 "HDDBusyGraph", 61 ], 62 ) 63 safe_import("maildir", "Maildir") 64 safe_import("notify", "Notify") 65 safe_import("sensors", "ThermalSensor") 66 safe_import("sep", "Sep") 67 safe_import("she", "She") 68 safe_import("spacer", "Spacer") 69 safe_import("generic_poll_text", ["GenPollText", "GenPollUrl"]) 70 safe_import("volume", "Volume") 71 safe_import("windowtabs", "WindowTabs") 72 safe_import("keyboardlayout", "KeyboardLayout") 73 safe_import("df", "DF") 74 safe_import("image", "Image") 75 safe_import("gmail_checker", "GmailChecker") 76 safe_import("clipboard", "Clipboard") 77 safe_import("countdown", "Countdown") 78 safe_import("tasklist", "TaskList") 79 safe_import("pacman", "Pacman") 80 safe_import("launchbar", "LaunchBar") 81 safe_import("canto", "Canto") 82 safe_import("mpriswidget", "Mpris") 83 safe_import("mpris2widget", "Mpris2") 84 safe_import("mpd2widget", "Mpd2") 85 safe_import("yahoo_weather", "YahooWeather") 86 safe_import("bitcoin_ticker", "BitcoinTicker") 87 safe_import("wlan", "Wlan") 88 safe_import("khal_calendar", "KhalCalendar") 89 safe_import("imapwidget", "ImapWidget") 90 safe_import("net", "Net") 91 safe_import("keyboardkbdd", "KeyboardKbdd") 92 safe_import("cmus", "Cmus") 93 safe_import("wallpaper", "Wallpaper") 94 safe_import("check_updates", "CheckUpdates") 95 safe_import("moc", "Moc") 96 safe_import("memory", "Memory") 97 safe_import("cpu", "CPU") 98 safe_import("idlerpg", "IdleRPG") 99 safe_import("pomodoro", "Pomodoro") 100 safe_import("stock_ticker", "StockTicker") 101 safe_import("caps_num_lock_indicator", "CapsNumLockIndicator") 102 safe_import("quick_exit", "QuickExit") 103 safe_import("pulse_volume", "PulseVolume") 104 safe_import("chord", "Chord") 105 [end of libqtile/widget/__init__.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/libqtile/widget/__init__.py b/libqtile/widget/__init__.py --- a/libqtile/widget/__init__.py +++ b/libqtile/widget/__init__.py @@ -76,7 +76,6 @@ safe_import("clipboard", "Clipboard") safe_import("countdown", "Countdown") safe_import("tasklist", "TaskList") -safe_import("pacman", "Pacman") safe_import("launchbar", "LaunchBar") safe_import("canto", "Canto") safe_import("mpriswidget", "Mpris") diff --git a/libqtile/widget/pacman.py b/libqtile/widget/pacman.py deleted file mode 100644 --- a/libqtile/widget/pacman.py +++ /dev/null @@ -1,56 +0,0 @@ -# -*- coding:utf-8 -*- -# -# Copyright (C) 2012, Maximilian Köhl <[email protected]> -# -# This program is free software: you can redistribute it and/or modify -# it under the terms of the GNU General Public License as published by -# the Free Software Foundation, either version 3 of the License, or -# (at your option) any later version. -# -# This program is distributed in the hope that it will be useful, -# but WITHOUT ANY WARRANTY; without even the implied warranty of -# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the -# GNU General Public License for more details. -# -# You should have received a copy of the GNU General Public License -# along with this program. If not, see <http://www.gnu.org/licenses/>. - -import subprocess - -from libqtile.log_utils import logger -from libqtile.widget import base - - -class Pacman(base.ThreadedPollText): - """Shows number of available updates - - Needs the pacman package manager installed. So will only work in Arch Linux - installation. - """ - orientations = base.ORIENTATION_HORIZONTAL - defaults = [ - ('unavailable', 'ffffff', 'Unavailable Color - no updates.'), - ('execute', None, 'Command to execute on click'), - ('update_interval', 60, "The update interval."), - ] - - def __init__(self, **config): - logger.warning("Pacman is deprecated, please use CheckUpdates") - base.ThreadedPollText.__init__(self, **config) - self.add_defaults(Pacman.defaults) - - def draw(self): - if self.text == '0': - self.layout.colour = self.unavailable - else: - self.layout.colour = self.foreground - base.ThreadedPollText.draw(self) - - def poll(self): - pacman = self.call_process(['checkupdates']) - return str(len(pacman.splitlines())) - - def button_press(self, x, y, button): - base.ThreadedPollText.button_press(self, x, y, button) - if button == 1 and self.execute is not None: - subprocess.Popen([self.execute], shell=True)
{"golden_diff": "diff --git a/libqtile/widget/__init__.py b/libqtile/widget/__init__.py\n--- a/libqtile/widget/__init__.py\n+++ b/libqtile/widget/__init__.py\n@@ -76,7 +76,6 @@\n safe_import(\"clipboard\", \"Clipboard\")\n safe_import(\"countdown\", \"Countdown\")\n safe_import(\"tasklist\", \"TaskList\")\n-safe_import(\"pacman\", \"Pacman\")\n safe_import(\"launchbar\", \"LaunchBar\")\n safe_import(\"canto\", \"Canto\")\n safe_import(\"mpriswidget\", \"Mpris\")\ndiff --git a/libqtile/widget/pacman.py b/libqtile/widget/pacman.py\ndeleted file mode 100644\n--- a/libqtile/widget/pacman.py\n+++ /dev/null\n@@ -1,56 +0,0 @@\n-# -*- coding:utf-8 -*-\n-#\n-# Copyright (C) 2012, Maximilian K\u00f6hl <[email protected]>\n-#\n-# This program is free software: you can redistribute it and/or modify\n-# it under the terms of the GNU General Public License as published by\n-# the Free Software Foundation, either version 3 of the License, or\n-# (at your option) any later version.\n-#\n-# This program is distributed in the hope that it will be useful,\n-# but WITHOUT ANY WARRANTY; without even the implied warranty of\n-# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the\n-# GNU General Public License for more details.\n-#\n-# You should have received a copy of the GNU General Public License\n-# along with this program. If not, see <http://www.gnu.org/licenses/>.\n-\n-import subprocess\n-\n-from libqtile.log_utils import logger\n-from libqtile.widget import base\n-\n-\n-class Pacman(base.ThreadedPollText):\n- \"\"\"Shows number of available updates\n-\n- Needs the pacman package manager installed. So will only work in Arch Linux\n- installation.\n- \"\"\"\n- orientations = base.ORIENTATION_HORIZONTAL\n- defaults = [\n- ('unavailable', 'ffffff', 'Unavailable Color - no updates.'),\n- ('execute', None, 'Command to execute on click'),\n- ('update_interval', 60, \"The update interval.\"),\n- ]\n-\n- def __init__(self, **config):\n- logger.warning(\"Pacman is deprecated, please use CheckUpdates\")\n- base.ThreadedPollText.__init__(self, **config)\n- self.add_defaults(Pacman.defaults)\n-\n- def draw(self):\n- if self.text == '0':\n- self.layout.colour = self.unavailable\n- else:\n- self.layout.colour = self.foreground\n- base.ThreadedPollText.draw(self)\n-\n- def poll(self):\n- pacman = self.call_process(['checkupdates'])\n- return str(len(pacman.splitlines()))\n-\n- def button_press(self, x, y, button):\n- base.ThreadedPollText.button_press(self, x, y, button)\n- if button == 1 and self.execute is not None:\n- subprocess.Popen([self.execute], shell=True)\n", "issue": "widget.Pacman needs a non existing file:'checkupdates'\n```\r\n2020-03-30 07:09:12,582 ERROR libqtile base.py:worker():L478 problem polling to update widget pacman\r\nTraceback (most recent call last):\r\n File \"/home/parsrch/.local/lib/python3.8/site-packages/libqtile/widget/base.py\", line 474, in worker\r\n text = self.poll()\r\n File \"/home/parsrch/.local/lib/python3.8/site-packages/libqtile/widget/pacman.py\", line 49, in poll\r\n pacman = self.call_process(['checkupdates'])\r\n File \"/home/parsrch/.local/lib/python3.8/site-packages/libqtile/widget/base.py\", line 266, in call_process\r\n output = subprocess.check_output(command, **kwargs)\r\n File \"/usr/lib/python3.8/subprocess.py\", line 411, in check_output\r\n return run(*popenargs, stdout=PIPE, timeout=timeout, check=True,\r\n File \"/usr/lib/python3.8/subprocess.py\", line 489, in run\r\n with Popen(*popenargs, **kwargs) as process:\r\n File \"/usr/lib/python3.8/subprocess.py\", line 854, in __init__\r\n self._execute_child(args, executable, preexec_fn, close_fds,\r\n File \"/usr/lib/python3.8/subprocess.py\", line 1702, in _execute_child\r\n raise child_exception_type(errno_num, err_msg, err_filename)\r\nFileNotFoundError: [Errno 2] No such file or directory: 'checkupdates'\r\n```\n", "before_files": [{"content": "# -*- coding:utf-8 -*-\n#\n# Copyright (C) 2012, Maximilian K\u00f6hl <[email protected]>\n#\n# This program is free software: you can redistribute it and/or modify\n# it under the terms of the GNU General Public License as published by\n# the Free Software Foundation, either version 3 of the License, or\n# (at your option) any later version.\n#\n# This program is distributed in the hope that it will be useful,\n# but WITHOUT ANY WARRANTY; without even the implied warranty of\n# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the\n# GNU General Public License for more details.\n#\n# You should have received a copy of the GNU General Public License\n# along with this program. If not, see <http://www.gnu.org/licenses/>.\n\nimport subprocess\n\nfrom libqtile.log_utils import logger\nfrom libqtile.widget import base\n\n\nclass Pacman(base.ThreadedPollText):\n \"\"\"Shows number of available updates\n\n Needs the pacman package manager installed. So will only work in Arch Linux\n installation.\n \"\"\"\n orientations = base.ORIENTATION_HORIZONTAL\n defaults = [\n ('unavailable', 'ffffff', 'Unavailable Color - no updates.'),\n ('execute', None, 'Command to execute on click'),\n ('update_interval', 60, \"The update interval.\"),\n ]\n\n def __init__(self, **config):\n logger.warning(\"Pacman is deprecated, please use CheckUpdates\")\n base.ThreadedPollText.__init__(self, **config)\n self.add_defaults(Pacman.defaults)\n\n def draw(self):\n if self.text == '0':\n self.layout.colour = self.unavailable\n else:\n self.layout.colour = self.foreground\n base.ThreadedPollText.draw(self)\n\n def poll(self):\n pacman = self.call_process(['checkupdates'])\n return str(len(pacman.splitlines()))\n\n def button_press(self, x, y, button):\n base.ThreadedPollText.button_press(self, x, y, button)\n if button == 1 and self.execute is not None:\n subprocess.Popen([self.execute], shell=True)\n", "path": "libqtile/widget/pacman.py"}, {"content": "# Copyright (c) 2014 Rock Neurotiko\n# Copyright (c) 2014 roger\n# Copyright (c) 2015 David R. Andersen\n#\n# Permission is hereby granted, free of charge, to any person obtaining a copy\n# of this software and associated documentation files (the \"Software\"), to deal\n# in the Software without restriction, including without limitation the rights\n# to use, copy, modify, merge, publish, distribute, sublicense, and/or sell\n# copies of the Software, and to permit persons to whom the Software is\n# furnished to do so, subject to the following conditions:\n#\n# The above copyright notice and this permission notice shall be included in\n# all copies or substantial portions of the Software.\n#\n# THE SOFTWARE IS PROVIDED \"AS IS\", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR\n# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,\n# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE\n# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER\n# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,\n# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE\n# SOFTWARE.\n\nfrom libqtile.utils import safe_import as safe_import_\n# only directly import widgets that do not have any third party dependencies\n# other than those required by qtile, otherwise use the same import function\nfrom libqtile.widget.base import Mirror # noqa: F401\nfrom libqtile.widget.chord import Chord # noqa: F401\nfrom libqtile.widget.clock import Clock # noqa: F401\nfrom libqtile.widget.currentlayout import ( # noqa: F401\n CurrentLayout,\n CurrentLayoutIcon,\n)\nfrom libqtile.widget.groupbox import AGroupBox, GroupBox # noqa: F401\nfrom libqtile.widget.import_error import make_error\nfrom libqtile.widget.prompt import Prompt # noqa: F401\nfrom libqtile.widget.quick_exit import QuickExit # noqa: F401\nfrom libqtile.widget.systray import Systray # noqa: F401\nfrom libqtile.widget.textbox import TextBox # noqa: F401\nfrom libqtile.widget.windowname import WindowName # noqa: F401\n\n\ndef safe_import(module_name, class_name):\n safe_import_(\n (\".widget\", module_name), class_name, globals(), fallback=make_error\n )\n\n\nsafe_import(\"backlight\", \"Backlight\")\nsafe_import(\"battery\", [\"Battery\", \"BatteryIcon\"])\nsafe_import(\"currentscreen\", \"CurrentScreen\")\nsafe_import(\"debuginfo\", \"DebugInfo\")\nsafe_import(\n \"graph\",\n [\n \"CPUGraph\",\n \"MemoryGraph\",\n \"SwapGraph\",\n \"NetGraph\",\n \"HDDGraph\",\n \"HDDBusyGraph\",\n ],\n)\nsafe_import(\"maildir\", \"Maildir\")\nsafe_import(\"notify\", \"Notify\")\nsafe_import(\"sensors\", \"ThermalSensor\")\nsafe_import(\"sep\", \"Sep\")\nsafe_import(\"she\", \"She\")\nsafe_import(\"spacer\", \"Spacer\")\nsafe_import(\"generic_poll_text\", [\"GenPollText\", \"GenPollUrl\"])\nsafe_import(\"volume\", \"Volume\")\nsafe_import(\"windowtabs\", \"WindowTabs\")\nsafe_import(\"keyboardlayout\", \"KeyboardLayout\")\nsafe_import(\"df\", \"DF\")\nsafe_import(\"image\", \"Image\")\nsafe_import(\"gmail_checker\", \"GmailChecker\")\nsafe_import(\"clipboard\", \"Clipboard\")\nsafe_import(\"countdown\", \"Countdown\")\nsafe_import(\"tasklist\", \"TaskList\")\nsafe_import(\"pacman\", \"Pacman\")\nsafe_import(\"launchbar\", \"LaunchBar\")\nsafe_import(\"canto\", \"Canto\")\nsafe_import(\"mpriswidget\", \"Mpris\")\nsafe_import(\"mpris2widget\", \"Mpris2\")\nsafe_import(\"mpd2widget\", \"Mpd2\")\nsafe_import(\"yahoo_weather\", \"YahooWeather\")\nsafe_import(\"bitcoin_ticker\", \"BitcoinTicker\")\nsafe_import(\"wlan\", \"Wlan\")\nsafe_import(\"khal_calendar\", \"KhalCalendar\")\nsafe_import(\"imapwidget\", \"ImapWidget\")\nsafe_import(\"net\", \"Net\")\nsafe_import(\"keyboardkbdd\", \"KeyboardKbdd\")\nsafe_import(\"cmus\", \"Cmus\")\nsafe_import(\"wallpaper\", \"Wallpaper\")\nsafe_import(\"check_updates\", \"CheckUpdates\")\nsafe_import(\"moc\", \"Moc\")\nsafe_import(\"memory\", \"Memory\")\nsafe_import(\"cpu\", \"CPU\")\nsafe_import(\"idlerpg\", \"IdleRPG\")\nsafe_import(\"pomodoro\", \"Pomodoro\")\nsafe_import(\"stock_ticker\", \"StockTicker\")\nsafe_import(\"caps_num_lock_indicator\", \"CapsNumLockIndicator\")\nsafe_import(\"quick_exit\", \"QuickExit\")\nsafe_import(\"pulse_volume\", \"PulseVolume\")\nsafe_import(\"chord\", \"Chord\")\n", "path": "libqtile/widget/__init__.py"}]}
2,775
686
gh_patches_debug_5067
rasdani/github-patches
git_diff
jupyterhub__jupyterhub-1111
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> Passwords beginning or ending with a whitespace are not supported Due to POST argument stripping, passwords with a beginning or ending whitespace are not allowed. **How to reproduce the issue** Set up a user password with an ending or beginning whitespace. **What you expected to happen** The user should be allowed to login with the password, given that the password should be any complicated sequence of characters the user can reproduce. **What actually happens** The user is denied access, because the LoginHandler will strip all posted values before considering the password for authentication (line 81, get_argument has a default "strip=True") **Share what version of JupyterHub you are using** HEAD (006488fc749923851df97d47d8850bdf5fd157cf) </issue> <code> [start of jupyterhub/handlers/login.py] 1 """HTTP Handlers for the hub server""" 2 3 # Copyright (c) Jupyter Development Team. 4 # Distributed under the terms of the Modified BSD License. 5 6 from urllib.parse import urlparse 7 8 from tornado.escape import url_escape 9 from tornado import gen 10 from tornado.httputil import url_concat 11 12 from .base import BaseHandler 13 14 15 class LogoutHandler(BaseHandler): 16 """Log a user out by clearing their login cookie.""" 17 def get(self): 18 user = self.get_current_user() 19 if user: 20 self.log.info("User logged out: %s", user.name) 21 self.clear_login_cookie() 22 self.statsd.incr('logout') 23 if self.authenticator.auto_login: 24 self.render('logout.html') 25 else: 26 self.redirect(self.settings['login_url'], permanent=False) 27 28 29 class LoginHandler(BaseHandler): 30 """Render the login page.""" 31 32 def _render(self, login_error=None, username=None): 33 return self.render_template('login.html', 34 next=url_escape(self.get_argument('next', default='')), 35 username=username, 36 login_error=login_error, 37 custom_html=self.authenticator.custom_html, 38 login_url=self.settings['login_url'], 39 authenticator_login_url=self.authenticator.login_url(self.hub.server.base_url), 40 ) 41 42 def get(self): 43 self.statsd.incr('login.request') 44 next_url = self.get_argument('next', '') 45 if (next_url + '/').startswith('%s://%s/' % (self.request.protocol, self.request.host)): 46 # treat absolute URLs for our host as absolute paths: 47 next_url = urlparse(next_url).path 48 elif not next_url.startswith('/'): 49 # disallow non-absolute next URLs (e.g. full URLs to other hosts) 50 next_url = '' 51 user = self.get_current_user() 52 if user: 53 if not next_url: 54 if user.running: 55 next_url = user.url 56 else: 57 next_url = self.hub.server.base_url 58 # set new login cookie 59 # because single-user cookie may have been cleared or incorrect 60 self.set_login_cookie(self.get_current_user()) 61 self.redirect(next_url, permanent=False) 62 else: 63 if self.authenticator.auto_login: 64 auto_login_url = self.authenticator.login_url(self.hub.server.base_url) 65 if auto_login_url == self.settings['login_url']: 66 self.authenticator.auto_login = False 67 self.log.warning("Authenticator.auto_login cannot be used without a custom login_url") 68 else: 69 if next_url: 70 auto_login_url = url_concat(auto_login_url, {'next': next_url}) 71 self.redirect(auto_login_url) 72 return 73 username = self.get_argument('username', default='') 74 self.finish(self._render(username=username)) 75 76 @gen.coroutine 77 def post(self): 78 # parse the arguments dict 79 data = {} 80 for arg in self.request.arguments: 81 data[arg] = self.get_argument(arg) 82 83 auth_timer = self.statsd.timer('login.authenticate').start() 84 username = yield self.authenticate(data) 85 auth_timer.stop(send=False) 86 87 if username: 88 self.statsd.incr('login.success') 89 self.statsd.timing('login.authenticate.success', auth_timer.ms) 90 user = self.user_from_username(username) 91 already_running = False 92 if user.spawner: 93 status = yield user.spawner.poll() 94 already_running = (status == None) 95 if not already_running and not user.spawner.options_form: 96 yield self.spawn_single_user(user) 97 self.set_login_cookie(user) 98 next_url = self.get_argument('next', default='') 99 if not next_url.startswith('/'): 100 next_url = '' 101 next_url = next_url or self.hub.server.base_url 102 self.redirect(next_url) 103 self.log.info("User logged in: %s", username) 104 else: 105 self.statsd.incr('login.failure') 106 self.statsd.timing('login.authenticate.failure', auth_timer.ms) 107 self.log.debug("Failed login for %s", data.get('username', 'unknown user')) 108 html = self._render( 109 login_error='Invalid username or password', 110 username=username, 111 ) 112 self.finish(html) 113 114 115 # /login renders the login page or the "Login with..." link, 116 # so it should always be registered. 117 # /logout clears cookies. 118 default_handlers = [ 119 (r"/login", LoginHandler), 120 (r"/logout", LogoutHandler), 121 ] 122 [end of jupyterhub/handlers/login.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/jupyterhub/handlers/login.py b/jupyterhub/handlers/login.py --- a/jupyterhub/handlers/login.py +++ b/jupyterhub/handlers/login.py @@ -78,7 +78,7 @@ # parse the arguments dict data = {} for arg in self.request.arguments: - data[arg] = self.get_argument(arg) + data[arg] = self.get_argument(arg, strip=False) auth_timer = self.statsd.timer('login.authenticate').start() username = yield self.authenticate(data)
{"golden_diff": "diff --git a/jupyterhub/handlers/login.py b/jupyterhub/handlers/login.py\n--- a/jupyterhub/handlers/login.py\n+++ b/jupyterhub/handlers/login.py\n@@ -78,7 +78,7 @@\n # parse the arguments dict\n data = {}\n for arg in self.request.arguments:\n- data[arg] = self.get_argument(arg)\n+ data[arg] = self.get_argument(arg, strip=False)\n \n auth_timer = self.statsd.timer('login.authenticate').start()\n username = yield self.authenticate(data)\n", "issue": "Passwords beginning or ending with a whitespace are not supported\nDue to POST argument stripping, passwords with a beginning or ending whitespace are not allowed.\r\n\r\n**How to reproduce the issue**\r\nSet up a user password with an ending or beginning whitespace.\r\n\r\n**What you expected to happen**\r\nThe user should be allowed to login with the password, given that the password should be any complicated sequence of characters the user can reproduce.\r\n\r\n**What actually happens**\r\nThe user is denied access, because the LoginHandler will strip all posted values before considering the password for authentication (line 81, get_argument has a default \"strip=True\")\r\n\r\n**Share what version of JupyterHub you are using**\r\nHEAD (006488fc749923851df97d47d8850bdf5fd157cf)\r\n\r\n\r\n\n", "before_files": [{"content": "\"\"\"HTTP Handlers for the hub server\"\"\"\n\n# Copyright (c) Jupyter Development Team.\n# Distributed under the terms of the Modified BSD License.\n\nfrom urllib.parse import urlparse\n\nfrom tornado.escape import url_escape\nfrom tornado import gen\nfrom tornado.httputil import url_concat\n\nfrom .base import BaseHandler\n\n\nclass LogoutHandler(BaseHandler):\n \"\"\"Log a user out by clearing their login cookie.\"\"\"\n def get(self):\n user = self.get_current_user()\n if user:\n self.log.info(\"User logged out: %s\", user.name)\n self.clear_login_cookie()\n self.statsd.incr('logout')\n if self.authenticator.auto_login:\n self.render('logout.html')\n else:\n self.redirect(self.settings['login_url'], permanent=False)\n\n\nclass LoginHandler(BaseHandler):\n \"\"\"Render the login page.\"\"\"\n\n def _render(self, login_error=None, username=None):\n return self.render_template('login.html',\n next=url_escape(self.get_argument('next', default='')),\n username=username,\n login_error=login_error,\n custom_html=self.authenticator.custom_html,\n login_url=self.settings['login_url'],\n authenticator_login_url=self.authenticator.login_url(self.hub.server.base_url),\n )\n\n def get(self):\n self.statsd.incr('login.request')\n next_url = self.get_argument('next', '')\n if (next_url + '/').startswith('%s://%s/' % (self.request.protocol, self.request.host)):\n # treat absolute URLs for our host as absolute paths:\n next_url = urlparse(next_url).path\n elif not next_url.startswith('/'):\n # disallow non-absolute next URLs (e.g. full URLs to other hosts)\n next_url = ''\n user = self.get_current_user()\n if user:\n if not next_url:\n if user.running:\n next_url = user.url\n else:\n next_url = self.hub.server.base_url\n # set new login cookie\n # because single-user cookie may have been cleared or incorrect\n self.set_login_cookie(self.get_current_user())\n self.redirect(next_url, permanent=False)\n else:\n if self.authenticator.auto_login:\n auto_login_url = self.authenticator.login_url(self.hub.server.base_url)\n if auto_login_url == self.settings['login_url']:\n self.authenticator.auto_login = False\n self.log.warning(\"Authenticator.auto_login cannot be used without a custom login_url\")\n else:\n if next_url:\n auto_login_url = url_concat(auto_login_url, {'next': next_url})\n self.redirect(auto_login_url)\n return\n username = self.get_argument('username', default='')\n self.finish(self._render(username=username))\n\n @gen.coroutine\n def post(self):\n # parse the arguments dict\n data = {}\n for arg in self.request.arguments:\n data[arg] = self.get_argument(arg)\n\n auth_timer = self.statsd.timer('login.authenticate').start()\n username = yield self.authenticate(data)\n auth_timer.stop(send=False)\n\n if username:\n self.statsd.incr('login.success')\n self.statsd.timing('login.authenticate.success', auth_timer.ms)\n user = self.user_from_username(username)\n already_running = False\n if user.spawner:\n status = yield user.spawner.poll()\n already_running = (status == None)\n if not already_running and not user.spawner.options_form:\n yield self.spawn_single_user(user)\n self.set_login_cookie(user)\n next_url = self.get_argument('next', default='')\n if not next_url.startswith('/'):\n next_url = ''\n next_url = next_url or self.hub.server.base_url\n self.redirect(next_url)\n self.log.info(\"User logged in: %s\", username)\n else:\n self.statsd.incr('login.failure')\n self.statsd.timing('login.authenticate.failure', auth_timer.ms)\n self.log.debug(\"Failed login for %s\", data.get('username', 'unknown user'))\n html = self._render(\n login_error='Invalid username or password',\n username=username,\n )\n self.finish(html)\n\n\n# /login renders the login page or the \"Login with...\" link,\n# so it should always be registered.\n# /logout clears cookies.\ndefault_handlers = [\n (r\"/login\", LoginHandler),\n (r\"/logout\", LogoutHandler),\n]\n", "path": "jupyterhub/handlers/login.py"}]}
1,901
123
gh_patches_debug_40750
rasdani/github-patches
git_diff
streamlit__streamlit-188
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> Blacklist common virtualenv folders by default We should blacklist the folders below from being watched by Streamlit. This would fix the issue where some people hit the inotify watch limit when running Streamlit from a weird working directory. * */.virtualenv * */.venv * */anaconda3 * */anaconda2 * */miniconda3 * */miniconda2 * Actually let's blacklist every hidden folder: `.*` See also the config option `server.folderWatchBlacklist`. For this fix, you can probably use the same mechanism this config option uses. </issue> <code> [start of lib/streamlit/watcher/LocalSourcesWatcher.py] 1 # -*- coding: utf-8 -*- 2 # Copyright 2018-2019 Streamlit Inc. 3 # 4 # Licensed under the Apache License, Version 2.0 (the "License"); 5 # you may not use this file except in compliance with the License. 6 # You may obtain a copy of the License at 7 # 8 # http://www.apache.org/licenses/LICENSE-2.0 9 # 10 # Unless required by applicable law or agreed to in writing, software 11 # distributed under the License is distributed on an "AS IS" BASIS, 12 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 13 # See the License for the specific language governing permissions and 14 # limitations under the License. 15 16 import os 17 import sys 18 import collections 19 20 try: 21 # Python 2 22 import imp as importlib 23 except ImportError: 24 # Python 3 25 import importlib 26 27 from streamlit import config 28 from streamlit import util 29 30 from streamlit.logger import get_logger 31 LOGGER = get_logger(__name__) 32 33 try: 34 # If the watchdog module is installed. 35 from streamlit.watcher.EventBasedFileWatcher import ( 36 EventBasedFileWatcher as FileWatcher, 37 ) 38 except ImportError: 39 # Fallback that doesn't use watchdog. 40 from streamlit.watcher.PollingFileWatcher import PollingFileWatcher as FileWatcher 41 42 if not config.get_option('global.disableWatchdogWarning'): 43 msg = '\n $ xcode-select --install' if util.is_darwin() else '' 44 45 LOGGER.warning(""" 46 For better performance, install the Watchdog module: 47 %s 48 $ pip install watchdog 49 50 """ % msg) 51 52 53 WatchedModule = collections.namedtuple("WatchedModule", ["watcher", "module_name"]) 54 55 56 class LocalSourcesWatcher(object): 57 def __init__(self, report, on_file_changed): 58 self._report = report 59 self._on_file_changed = on_file_changed 60 self._is_closed = False 61 62 self._folder_blacklist = config.get_option("server.folderWatchBlacklist") 63 64 # A dict of filepath -> WatchedModule. 65 self._watched_modules = {} 66 67 self._register_watcher( 68 self._report.script_path, 69 module_name=None, # Only the root script has None here. 70 ) 71 72 def on_file_changed(self, filepath): 73 if filepath not in self._watched_modules: 74 LOGGER.error("Received event for non-watched file", filepath) 75 return 76 77 wm = self._watched_modules[filepath] 78 79 if wm.module_name is not None and wm.module_name in sys.modules: 80 del sys.modules[wm.module_name] 81 82 self._on_file_changed() 83 84 def close(self): 85 for wm in self._watched_modules.values(): 86 wm.watcher.close() 87 self._watched_modules = {} 88 self._is_closed = True 89 90 def _register_watcher(self, filepath, module_name): 91 wm = WatchedModule( 92 watcher=FileWatcher(filepath, self.on_file_changed), module_name=module_name 93 ) 94 self._watched_modules[filepath] = wm 95 96 def _deregister_watcher(self, filepath): 97 if filepath not in self._watched_modules: 98 return 99 100 if filepath == self._report.script_path: 101 return 102 103 wm = self._watched_modules[filepath] 104 wm.watcher.close() 105 del self._watched_modules[filepath] 106 107 def update_watched_modules(self): 108 if self._is_closed: 109 return 110 111 local_filepaths = [] 112 113 # Clone modules dict here because we may alter the original dict inside 114 # the loop. 115 modules = dict(sys.modules) 116 117 for name, module in modules.items(): 118 try: 119 spec = getattr(module, "__spec__", None) 120 121 if spec is None: 122 filepath = getattr(module, "__file__", None) 123 if filepath is None: 124 # Some modules have neither a spec nor a file. But we 125 # can ignore those since they're not the user-created 126 # modules we want to watch anyway. 127 continue 128 else: 129 filepath = spec.origin 130 131 if filepath is None: 132 # Built-in modules (and other stuff) don't have origins. 133 continue 134 135 filepath = os.path.abspath(filepath) 136 137 if not os.path.isfile(filepath): 138 # There are some modules that have a .origin, but don't 139 # point to real files. For example, there's a module where 140 # .origin is 'built-in'. 141 continue 142 143 folder_is_blacklisted = any( 144 _file_is_in_folder(filepath, blacklisted_folder) 145 for blacklisted_folder in self._folder_blacklist 146 ) 147 148 if folder_is_blacklisted: 149 continue 150 151 file_is_new = filepath not in self._watched_modules 152 file_is_local = _file_is_in_folder(filepath, self._report.script_folder) 153 154 local_filepaths.append(filepath) 155 156 if file_is_local and file_is_new: 157 self._register_watcher(filepath, name) 158 159 except Exception: 160 # In case there's a problem introspecting some specific module, 161 # let's not stop the entire loop from running. For example, 162 # the __spec__ field in some modules (like IPython) is actually 163 # a dynamic property, which can crash if the underlying 164 # module's code has a bug (as discovered by one of our users). 165 continue 166 167 # Clone dict here because we may alter the original dict inside the 168 # loop. 169 watched_modules = dict(self._watched_modules) 170 171 # Remove no-longer-depended-on files from self._watched_modules 172 # Will this ever happen? 173 for filepath in watched_modules: 174 if filepath not in local_filepaths: 175 self._deregister_watcher(filepath) 176 177 178 def _file_is_in_folder(filepath, folderpath): 179 # Assumes filepath is an absolute path, as a teeny tiny optimization. 180 folderpath = os.path.abspath(folderpath) + "/" 181 return filepath.startswith(folderpath) 182 [end of lib/streamlit/watcher/LocalSourcesWatcher.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/lib/streamlit/watcher/LocalSourcesWatcher.py b/lib/streamlit/watcher/LocalSourcesWatcher.py --- a/lib/streamlit/watcher/LocalSourcesWatcher.py +++ b/lib/streamlit/watcher/LocalSourcesWatcher.py @@ -13,6 +13,7 @@ # See the License for the specific language governing permissions and # limitations under the License. +import fnmatch import os import sys import collections @@ -28,6 +29,7 @@ from streamlit import util from streamlit.logger import get_logger + LOGGER = get_logger(__name__) try: @@ -39,15 +41,24 @@ # Fallback that doesn't use watchdog. from streamlit.watcher.PollingFileWatcher import PollingFileWatcher as FileWatcher - if not config.get_option('global.disableWatchdogWarning'): - msg = '\n $ xcode-select --install' if util.is_darwin() else '' + if not config.get_option("global.disableWatchdogWarning"): + msg = "\n $ xcode-select --install" if util.is_darwin() else "" - LOGGER.warning(""" + LOGGER.warning( + """ For better performance, install the Watchdog module: %s $ pip install watchdog - """ % msg) + """ + % msg + ) + + +# Streamlit never watches files in the folders below. +DEFAULT_FOLDER_BLACKLIST = [ + "**/.*", "**/anaconda2", "**/anaconda3", "**/miniconda2", "**/miniconda3" +] WatchedModule = collections.namedtuple("WatchedModule", ["watcher", "module_name"]) @@ -61,6 +72,9 @@ self._folder_blacklist = config.get_option("server.folderWatchBlacklist") + # Blacklist some additional folders, using glob syntax. + self._folder_blacklist.extend(DEFAULT_FOLDER_BLACKLIST) + # A dict of filepath -> WatchedModule. self._watched_modules = {} @@ -140,12 +154,12 @@ # .origin is 'built-in'. continue - folder_is_blacklisted = any( + is_in_blacklisted_folder = any( _file_is_in_folder(filepath, blacklisted_folder) for blacklisted_folder in self._folder_blacklist ) - if folder_is_blacklisted: + if is_in_blacklisted_folder: continue file_is_new = filepath not in self._watched_modules @@ -175,7 +189,10 @@ self._deregister_watcher(filepath) -def _file_is_in_folder(filepath, folderpath): - # Assumes filepath is an absolute path, as a teeny tiny optimization. - folderpath = os.path.abspath(folderpath) + "/" - return filepath.startswith(folderpath) +def _file_is_in_folder(filepath, folderpath_glob): + # Strip trailing slash if it exists + if folderpath_glob.endswith("/"): + folderpath_glob = folderpath_glob[:-1] + + file_dir = os.path.dirname(filepath) + return fnmatch.fnmatch(file_dir, folderpath_glob)
{"golden_diff": "diff --git a/lib/streamlit/watcher/LocalSourcesWatcher.py b/lib/streamlit/watcher/LocalSourcesWatcher.py\n--- a/lib/streamlit/watcher/LocalSourcesWatcher.py\n+++ b/lib/streamlit/watcher/LocalSourcesWatcher.py\n@@ -13,6 +13,7 @@\n # See the License for the specific language governing permissions and\n # limitations under the License.\n \n+import fnmatch\n import os\n import sys\n import collections\n@@ -28,6 +29,7 @@\n from streamlit import util\n \n from streamlit.logger import get_logger\n+\n LOGGER = get_logger(__name__)\n \n try:\n@@ -39,15 +41,24 @@\n # Fallback that doesn't use watchdog.\n from streamlit.watcher.PollingFileWatcher import PollingFileWatcher as FileWatcher\n \n- if not config.get_option('global.disableWatchdogWarning'):\n- msg = '\\n $ xcode-select --install' if util.is_darwin() else ''\n+ if not config.get_option(\"global.disableWatchdogWarning\"):\n+ msg = \"\\n $ xcode-select --install\" if util.is_darwin() else \"\"\n \n- LOGGER.warning(\"\"\"\n+ LOGGER.warning(\n+ \"\"\"\n For better performance, install the Watchdog module:\n %s\n $ pip install watchdog\n \n- \"\"\" % msg)\n+ \"\"\"\n+ % msg\n+ )\n+\n+\n+# Streamlit never watches files in the folders below.\n+DEFAULT_FOLDER_BLACKLIST = [\n+ \"**/.*\", \"**/anaconda2\", \"**/anaconda3\", \"**/miniconda2\", \"**/miniconda3\"\n+]\n \n \n WatchedModule = collections.namedtuple(\"WatchedModule\", [\"watcher\", \"module_name\"])\n@@ -61,6 +72,9 @@\n \n self._folder_blacklist = config.get_option(\"server.folderWatchBlacklist\")\n \n+ # Blacklist some additional folders, using glob syntax.\n+ self._folder_blacklist.extend(DEFAULT_FOLDER_BLACKLIST)\n+\n # A dict of filepath -> WatchedModule.\n self._watched_modules = {}\n \n@@ -140,12 +154,12 @@\n # .origin is 'built-in'.\n continue\n \n- folder_is_blacklisted = any(\n+ is_in_blacklisted_folder = any(\n _file_is_in_folder(filepath, blacklisted_folder)\n for blacklisted_folder in self._folder_blacklist\n )\n \n- if folder_is_blacklisted:\n+ if is_in_blacklisted_folder:\n continue\n \n file_is_new = filepath not in self._watched_modules\n@@ -175,7 +189,10 @@\n self._deregister_watcher(filepath)\n \n \n-def _file_is_in_folder(filepath, folderpath):\n- # Assumes filepath is an absolute path, as a teeny tiny optimization.\n- folderpath = os.path.abspath(folderpath) + \"/\"\n- return filepath.startswith(folderpath)\n+def _file_is_in_folder(filepath, folderpath_glob):\n+ # Strip trailing slash if it exists\n+ if folderpath_glob.endswith(\"/\"):\n+ folderpath_glob = folderpath_glob[:-1]\n+\n+ file_dir = os.path.dirname(filepath)\n+ return fnmatch.fnmatch(file_dir, folderpath_glob)\n", "issue": "Blacklist common virtualenv folders by default\nWe should blacklist the folders below from being watched by Streamlit. This would fix the issue where some people hit the inotify watch limit when running Streamlit from a weird working directory.\r\n\r\n* */.virtualenv\r\n* */.venv\r\n* */anaconda3\r\n* */anaconda2\r\n* */miniconda3\r\n* */miniconda2\r\n* Actually let's blacklist every hidden folder: `.*`\r\n\r\n\r\nSee also the config option `server.folderWatchBlacklist`. For this fix, you can probably use the same mechanism this config option uses.\n", "before_files": [{"content": "# -*- coding: utf-8 -*-\n# Copyright 2018-2019 Streamlit Inc.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nimport os\nimport sys\nimport collections\n\ntry:\n # Python 2\n import imp as importlib\nexcept ImportError:\n # Python 3\n import importlib\n\nfrom streamlit import config\nfrom streamlit import util\n\nfrom streamlit.logger import get_logger\nLOGGER = get_logger(__name__)\n\ntry:\n # If the watchdog module is installed.\n from streamlit.watcher.EventBasedFileWatcher import (\n EventBasedFileWatcher as FileWatcher,\n )\nexcept ImportError:\n # Fallback that doesn't use watchdog.\n from streamlit.watcher.PollingFileWatcher import PollingFileWatcher as FileWatcher\n\n if not config.get_option('global.disableWatchdogWarning'):\n msg = '\\n $ xcode-select --install' if util.is_darwin() else ''\n\n LOGGER.warning(\"\"\"\n For better performance, install the Watchdog module:\n %s\n $ pip install watchdog\n\n \"\"\" % msg)\n\n\nWatchedModule = collections.namedtuple(\"WatchedModule\", [\"watcher\", \"module_name\"])\n\n\nclass LocalSourcesWatcher(object):\n def __init__(self, report, on_file_changed):\n self._report = report\n self._on_file_changed = on_file_changed\n self._is_closed = False\n\n self._folder_blacklist = config.get_option(\"server.folderWatchBlacklist\")\n\n # A dict of filepath -> WatchedModule.\n self._watched_modules = {}\n\n self._register_watcher(\n self._report.script_path,\n module_name=None, # Only the root script has None here.\n )\n\n def on_file_changed(self, filepath):\n if filepath not in self._watched_modules:\n LOGGER.error(\"Received event for non-watched file\", filepath)\n return\n\n wm = self._watched_modules[filepath]\n\n if wm.module_name is not None and wm.module_name in sys.modules:\n del sys.modules[wm.module_name]\n\n self._on_file_changed()\n\n def close(self):\n for wm in self._watched_modules.values():\n wm.watcher.close()\n self._watched_modules = {}\n self._is_closed = True\n\n def _register_watcher(self, filepath, module_name):\n wm = WatchedModule(\n watcher=FileWatcher(filepath, self.on_file_changed), module_name=module_name\n )\n self._watched_modules[filepath] = wm\n\n def _deregister_watcher(self, filepath):\n if filepath not in self._watched_modules:\n return\n\n if filepath == self._report.script_path:\n return\n\n wm = self._watched_modules[filepath]\n wm.watcher.close()\n del self._watched_modules[filepath]\n\n def update_watched_modules(self):\n if self._is_closed:\n return\n\n local_filepaths = []\n\n # Clone modules dict here because we may alter the original dict inside\n # the loop.\n modules = dict(sys.modules)\n\n for name, module in modules.items():\n try:\n spec = getattr(module, \"__spec__\", None)\n\n if spec is None:\n filepath = getattr(module, \"__file__\", None)\n if filepath is None:\n # Some modules have neither a spec nor a file. But we\n # can ignore those since they're not the user-created\n # modules we want to watch anyway.\n continue\n else:\n filepath = spec.origin\n\n if filepath is None:\n # Built-in modules (and other stuff) don't have origins.\n continue\n\n filepath = os.path.abspath(filepath)\n\n if not os.path.isfile(filepath):\n # There are some modules that have a .origin, but don't\n # point to real files. For example, there's a module where\n # .origin is 'built-in'.\n continue\n\n folder_is_blacklisted = any(\n _file_is_in_folder(filepath, blacklisted_folder)\n for blacklisted_folder in self._folder_blacklist\n )\n\n if folder_is_blacklisted:\n continue\n\n file_is_new = filepath not in self._watched_modules\n file_is_local = _file_is_in_folder(filepath, self._report.script_folder)\n\n local_filepaths.append(filepath)\n\n if file_is_local and file_is_new:\n self._register_watcher(filepath, name)\n\n except Exception:\n # In case there's a problem introspecting some specific module,\n # let's not stop the entire loop from running. For example,\n # the __spec__ field in some modules (like IPython) is actually\n # a dynamic property, which can crash if the underlying\n # module's code has a bug (as discovered by one of our users).\n continue\n\n # Clone dict here because we may alter the original dict inside the\n # loop.\n watched_modules = dict(self._watched_modules)\n\n # Remove no-longer-depended-on files from self._watched_modules\n # Will this ever happen?\n for filepath in watched_modules:\n if filepath not in local_filepaths:\n self._deregister_watcher(filepath)\n\n\ndef _file_is_in_folder(filepath, folderpath):\n # Assumes filepath is an absolute path, as a teeny tiny optimization.\n folderpath = os.path.abspath(folderpath) + \"/\"\n return filepath.startswith(folderpath)\n", "path": "lib/streamlit/watcher/LocalSourcesWatcher.py"}]}
2,406
709
gh_patches_debug_9961
rasdani/github-patches
git_diff
pyjanitor-devs__pyjanitor-436
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> Copyright notice is hardcoded to 2018 # Brief Description of Fix <!-- Please describe the fix in terms of a "before" and "after". In other words, what's not so good about the current docs page, and what you would like to see it become. copyright notice is hard-coded to 2018. eg: **©2018, Eric J. Ma. | Powered by Sphinx 2.1.2 & Alabaster 0.7.12 | Page source** I propose modifying /pyjanitor/docs/conf.py to calculate and display current year. Example starter wording is provided. --> Currently, the docs... I would like to propose a change, such that now the docs... # Relevant Context <!-- Please put here, in bullet points, links to the relevant docs page. A few starting template points are available to get you started. --> - [Link to documentation page](http://pyjanitor.readthedocs.io) - [Link to exact file to be edited](https://github.com/ericmjl/pyjanitor/AUTHORS.rst) </issue> <code> [start of docs/conf.py] 1 """Sphinx configuration.""" 2 3 # -*- coding: utf-8 -*- 4 # 5 # Configuration file for the Sphinx documentation builder. 6 # 7 # This file does only contain a selection of the most common options. For a 8 # full list see the documentation: 9 # http://www.sphinx-doc.org/en/stable/config 10 11 # -- Path setup -------------------------------------------------------------- 12 13 # If extensions (or modules to document with autodoc) are in another directory, 14 # add these directories to sys.path here. If the directory is relative to the 15 # documentation root, use os.path.abspath to make it absolute, like shown here. 16 # 17 import os 18 import sys 19 from pathlib import Path 20 21 sys.path.insert(0, os.path.abspath(".")) 22 sys.path.insert(0, os.path.abspath("../examples")) 23 24 # Make a symlink in our sphinx source directory to the top-level 25 # examples/notebooks directory so we can include notebooks in the doc 26 notebooks = Path("./notebooks") 27 if not notebooks.exists(): 28 print("Making symlink to ../examples/notebooks") 29 notebooks.symlink_to("../examples/notebooks") 30 31 32 # -- Project information ----------------------------------------------------- 33 34 project = "pyjanitor" 35 copyright = "2018, Eric J. Ma" 36 author = "Eric J. Ma" 37 38 # The short X.Y version 39 version = "0.1.0" 40 # The full version, including alpha/beta/rc tags 41 release = "" 42 43 44 # -- General configuration --------------------------------------------------- 45 46 # If your documentation needs a minimal Sphinx version, state it here. 47 # 48 # needs_sphinx = '1.0' 49 50 # Add any Sphinx extension module names here, as strings. They can be 51 # extensions coming with Sphinx (named 'sphinx.ext.*') or your custom 52 # ones. 53 extensions = [ 54 "sphinx.ext.autodoc", 55 "sphinx.ext.doctest", 56 "sphinx.ext.intersphinx", 57 "sphinx.ext.todo", 58 "sphinx.ext.coverage", 59 "sphinx.ext.viewcode", 60 "sphinx.ext.githubpages", 61 "sphinxcontrib.fulltoc", 62 "nbsphinx", 63 "sphinx.ext.autosummary", 64 ] 65 66 # Add any paths that contain templates here, relative to this directory. 67 templates_path = ["_templates"] 68 69 # The suffix(es) of source filenames. 70 # You can specify multiple suffix as a list of string: 71 # 72 # source_suffix = ['.rst', '.md'] 73 source_suffix = [".md", ".rst", ".ipynb"] 74 75 # The master toctree document. 76 master_doc = "index" 77 78 # The language for content autogenerated by Sphinx. Refer to documentation 79 # for a list of supported languages. 80 # 81 # This is also used if you do content translation via gettext catalogs. 82 # Usually you set "language" from the command line for these cases. 83 language = None 84 85 # List of patterns, relative to source directory, that match files and 86 # directories to ignore when looking for source files. 87 # This pattern also affects html_static_path and html_extra_path . 88 exclude_patterns = ["_build", "Thumbs.db", ".DS_Store", "**.ipynb_checkpoints"] 89 90 # The name of the Pygments (syntax highlighting) style to use. 91 pygments_style = "sphinx" 92 93 94 # -- Options for HTML output ------------------------------------------------- 95 96 # The theme to use for HTML and HTML Help pages. See the documentation for 97 # a list of builtin themes. 98 # 99 html_theme = "alabaster" 100 101 # Theme options are theme-specific and customize the look and feel of a theme 102 # further. For a list of options available for each theme, see the 103 # documentation. 104 # 105 html_theme_options = {"logo": "logo_title.svg"} 106 107 # Add any paths that contain custom static files (such as style sheets) here, 108 # relative to this directory. They are copied after the builtin static files, 109 # so a file named "default.css" will overwrite the builtin "default.css". 110 html_static_path = ["_static"] 111 112 # Custom sidebar templates, must be a dictionary that maps document names 113 # to template names. 114 # 115 # The default sidebars (for documents that don't match any pattern) are 116 # defined by theme itself. Builtin themes are using these templates by 117 # default: ``['localtoc.html', 'relations.html', 'sourcelink.html', 118 # 'searchbox.html']``. 119 # 120 html_sidebars = { 121 "**": ["about.html", "navigation.html", "relations.html", "searchbox.html"] 122 } 123 124 125 # -- Options for HTMLHelp output --------------------------------------------- 126 127 # Output file base name for HTML help builder. 128 htmlhelp_basename = "pyjanitordoc" 129 130 131 # -- Options for LaTeX output ------------------------------------------------ 132 133 latex_elements = { 134 # The paper size ('letterpaper' or 'a4paper'). 135 # 136 # 'papersize': 'letterpaper', 137 # The font size ('10pt', '11pt' or '12pt'). 138 # 139 # 'pointsize': '10pt', 140 # Additional stuff for the LaTeX preamble. 141 # 142 # 'preamble': '', 143 # Latex figure (float) alignment 144 # 145 # 'figure_align': 'htbp', 146 } 147 148 # Grouping the document tree into LaTeX files. List of tuples 149 # (source start file, target name, title, 150 # author, documentclass [howto, manual, or own class]). 151 latex_documents = [ 152 ( 153 master_doc, 154 "pyjanitor.tex", 155 "pyjanitor Documentation", 156 "Eric J. Ma", 157 "manual", 158 ) 159 ] 160 161 162 # -- Options for manual page output ------------------------------------------ 163 164 # One entry per manual page. List of tuples 165 # (source start file, name, description, authors, manual section). 166 man_pages = [(master_doc, "pyjanitor", "pyjanitor Documentation", [author], 1)] 167 168 169 # -- Options for Texinfo output ---------------------------------------------- 170 171 # Grouping the document tree into Texinfo files. List of tuples 172 # (source start file, target name, title, author, 173 # dir menu entry, description, category) 174 texinfo_documents = [ 175 ( 176 master_doc, 177 "pyjanitor", 178 "pyjanitor Documentation", 179 author, 180 "pyjanitor", 181 "One line description of project.", 182 "Miscellaneous", 183 ) 184 ] 185 186 187 # -- Extension configuration ------------------------------------------------- 188 189 # -- Options for intersphinx extension --------------------------------------- 190 191 # Example configuration for intersphinx: refer to the Python standard library. 192 intersphinx_mapping = { 193 "https://docs.python.org/": None, 194 "https://pandas.pydata.org/pandas-docs/stable": None, 195 } 196 197 # -- Options for todo extension ---------------------------------------------- 198 199 # If true, `todo` and `todoList` produce output, else they produce nothing. 200 todo_include_todos = True 201 202 203 # -- Other options ----------------------------------------------------------- 204 205 autosummary_generate = True # Make _autosummary files and include them 206 [end of docs/conf.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/docs/conf.py b/docs/conf.py --- a/docs/conf.py +++ b/docs/conf.py @@ -16,6 +16,7 @@ # import os import sys +import datetime from pathlib import Path sys.path.insert(0, os.path.abspath(".")) @@ -32,7 +33,11 @@ # -- Project information ----------------------------------------------------- project = "pyjanitor" -copyright = "2018, Eric J. Ma" + + +now = datetime.datetime.now() +CurrentYear = str(now.year) +copyright = CurrentYear+", PyJanitor devs" author = "Eric J. Ma" # The short X.Y version
{"golden_diff": "diff --git a/docs/conf.py b/docs/conf.py\n--- a/docs/conf.py\n+++ b/docs/conf.py\n@@ -16,6 +16,7 @@\n #\n import os\n import sys\n+import datetime\n from pathlib import Path\n \n sys.path.insert(0, os.path.abspath(\".\"))\n@@ -32,7 +33,11 @@\n # -- Project information -----------------------------------------------------\n \n project = \"pyjanitor\"\n-copyright = \"2018, Eric J. Ma\"\n+\n+\n+now = datetime.datetime.now()\n+CurrentYear = str(now.year)\n+copyright = CurrentYear+\", PyJanitor devs\"\n author = \"Eric J. Ma\"\n \n # The short X.Y version\n", "issue": "Copyright notice is hardcoded to 2018\n# Brief Description of Fix\r\n\r\n<!-- Please describe the fix in terms of a \"before\" and \"after\". In other words, what's not so good about the current docs\r\npage, and what you would like to see it become. \r\n\r\ncopyright notice is hard-coded to 2018. eg: \r\n **\u00a92018, Eric J. Ma. | Powered by Sphinx 2.1.2 & Alabaster 0.7.12 | Page source**\r\nI propose modifying /pyjanitor/docs/conf.py to calculate and display current year.\r\n\r\nExample starter wording is provided. -->\r\n\r\nCurrently, the docs...\r\n\r\nI would like to propose a change, such that now the docs...\r\n\r\n# Relevant Context\r\n\r\n<!-- Please put here, in bullet points, links to the relevant docs page. A few starting template points are available\r\nto get you started. -->\r\n\r\n- [Link to documentation page](http://pyjanitor.readthedocs.io)\r\n- [Link to exact file to be edited](https://github.com/ericmjl/pyjanitor/AUTHORS.rst)\r\n\n", "before_files": [{"content": "\"\"\"Sphinx configuration.\"\"\"\n\n# -*- coding: utf-8 -*-\n#\n# Configuration file for the Sphinx documentation builder.\n#\n# This file does only contain a selection of the most common options. For a\n# full list see the documentation:\n# http://www.sphinx-doc.org/en/stable/config\n\n# -- Path setup --------------------------------------------------------------\n\n# If extensions (or modules to document with autodoc) are in another directory,\n# add these directories to sys.path here. If the directory is relative to the\n# documentation root, use os.path.abspath to make it absolute, like shown here.\n#\nimport os\nimport sys\nfrom pathlib import Path\n\nsys.path.insert(0, os.path.abspath(\".\"))\nsys.path.insert(0, os.path.abspath(\"../examples\"))\n\n# Make a symlink in our sphinx source directory to the top-level\n# examples/notebooks directory so we can include notebooks in the doc\nnotebooks = Path(\"./notebooks\")\nif not notebooks.exists():\n print(\"Making symlink to ../examples/notebooks\")\n notebooks.symlink_to(\"../examples/notebooks\")\n\n\n# -- Project information -----------------------------------------------------\n\nproject = \"pyjanitor\"\ncopyright = \"2018, Eric J. Ma\"\nauthor = \"Eric J. Ma\"\n\n# The short X.Y version\nversion = \"0.1.0\"\n# The full version, including alpha/beta/rc tags\nrelease = \"\"\n\n\n# -- General configuration ---------------------------------------------------\n\n# If your documentation needs a minimal Sphinx version, state it here.\n#\n# needs_sphinx = '1.0'\n\n# Add any Sphinx extension module names here, as strings. They can be\n# extensions coming with Sphinx (named 'sphinx.ext.*') or your custom\n# ones.\nextensions = [\n \"sphinx.ext.autodoc\",\n \"sphinx.ext.doctest\",\n \"sphinx.ext.intersphinx\",\n \"sphinx.ext.todo\",\n \"sphinx.ext.coverage\",\n \"sphinx.ext.viewcode\",\n \"sphinx.ext.githubpages\",\n \"sphinxcontrib.fulltoc\",\n \"nbsphinx\",\n \"sphinx.ext.autosummary\",\n]\n\n# Add any paths that contain templates here, relative to this directory.\ntemplates_path = [\"_templates\"]\n\n# The suffix(es) of source filenames.\n# You can specify multiple suffix as a list of string:\n#\n# source_suffix = ['.rst', '.md']\nsource_suffix = [\".md\", \".rst\", \".ipynb\"]\n\n# The master toctree document.\nmaster_doc = \"index\"\n\n# The language for content autogenerated by Sphinx. Refer to documentation\n# for a list of supported languages.\n#\n# This is also used if you do content translation via gettext catalogs.\n# Usually you set \"language\" from the command line for these cases.\nlanguage = None\n\n# List of patterns, relative to source directory, that match files and\n# directories to ignore when looking for source files.\n# This pattern also affects html_static_path and html_extra_path .\nexclude_patterns = [\"_build\", \"Thumbs.db\", \".DS_Store\", \"**.ipynb_checkpoints\"]\n\n# The name of the Pygments (syntax highlighting) style to use.\npygments_style = \"sphinx\"\n\n\n# -- Options for HTML output -------------------------------------------------\n\n# The theme to use for HTML and HTML Help pages. See the documentation for\n# a list of builtin themes.\n#\nhtml_theme = \"alabaster\"\n\n# Theme options are theme-specific and customize the look and feel of a theme\n# further. For a list of options available for each theme, see the\n# documentation.\n#\nhtml_theme_options = {\"logo\": \"logo_title.svg\"}\n\n# Add any paths that contain custom static files (such as style sheets) here,\n# relative to this directory. They are copied after the builtin static files,\n# so a file named \"default.css\" will overwrite the builtin \"default.css\".\nhtml_static_path = [\"_static\"]\n\n# Custom sidebar templates, must be a dictionary that maps document names\n# to template names.\n#\n# The default sidebars (for documents that don't match any pattern) are\n# defined by theme itself. Builtin themes are using these templates by\n# default: ``['localtoc.html', 'relations.html', 'sourcelink.html',\n# 'searchbox.html']``.\n#\nhtml_sidebars = {\n \"**\": [\"about.html\", \"navigation.html\", \"relations.html\", \"searchbox.html\"]\n}\n\n\n# -- Options for HTMLHelp output ---------------------------------------------\n\n# Output file base name for HTML help builder.\nhtmlhelp_basename = \"pyjanitordoc\"\n\n\n# -- Options for LaTeX output ------------------------------------------------\n\nlatex_elements = {\n # The paper size ('letterpaper' or 'a4paper').\n #\n # 'papersize': 'letterpaper',\n # The font size ('10pt', '11pt' or '12pt').\n #\n # 'pointsize': '10pt',\n # Additional stuff for the LaTeX preamble.\n #\n # 'preamble': '',\n # Latex figure (float) alignment\n #\n # 'figure_align': 'htbp',\n}\n\n# Grouping the document tree into LaTeX files. List of tuples\n# (source start file, target name, title,\n# author, documentclass [howto, manual, or own class]).\nlatex_documents = [\n (\n master_doc,\n \"pyjanitor.tex\",\n \"pyjanitor Documentation\",\n \"Eric J. Ma\",\n \"manual\",\n )\n]\n\n\n# -- Options for manual page output ------------------------------------------\n\n# One entry per manual page. List of tuples\n# (source start file, name, description, authors, manual section).\nman_pages = [(master_doc, \"pyjanitor\", \"pyjanitor Documentation\", [author], 1)]\n\n\n# -- Options for Texinfo output ----------------------------------------------\n\n# Grouping the document tree into Texinfo files. List of tuples\n# (source start file, target name, title, author,\n# dir menu entry, description, category)\ntexinfo_documents = [\n (\n master_doc,\n \"pyjanitor\",\n \"pyjanitor Documentation\",\n author,\n \"pyjanitor\",\n \"One line description of project.\",\n \"Miscellaneous\",\n )\n]\n\n\n# -- Extension configuration -------------------------------------------------\n\n# -- Options for intersphinx extension ---------------------------------------\n\n# Example configuration for intersphinx: refer to the Python standard library.\nintersphinx_mapping = {\n \"https://docs.python.org/\": None,\n \"https://pandas.pydata.org/pandas-docs/stable\": None,\n}\n\n# -- Options for todo extension ----------------------------------------------\n\n# If true, `todo` and `todoList` produce output, else they produce nothing.\ntodo_include_todos = True\n\n\n# -- Other options -----------------------------------------------------------\n\nautosummary_generate = True # Make _autosummary files and include them\n", "path": "docs/conf.py"}]}
2,740
148
gh_patches_debug_66426
rasdani/github-patches
git_diff
Textualize__textual-3531
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> Missing link styles in `Toast` notifcations This follows on from discussion #3526 about adding a clickable link in a notification. The `Toast` widget doesn't inherit the styling from the base class, so the "default" links styling defined in Widget is lost. You can obviously just add it back in like I've done here, but it might be nicer if you didn't have to do that. ```python # toast_with_link.py import webbrowser from textual.app import App class ToastLinkApp(App): # Uncomment the CSS_PATH to see the link! # CSS_PATH = "toast_with_link.css" def on_mount(self) -> None: self.notify( "Don't forget to " "[@click=link('https://github.com/Textualize/textual')]star[/] " "the repository!", timeout=10, ) def action_link(self, href: str) -> None: webbrowser.open(href) if __name__ == "__main__": app = ToastLinkApp() app.run() ``` ```css /* toast_with_link.css */ Toast { link-background: ; link-color: $text; link-style: underline; link-hover-background: $accent; link-hover-color: $text; link-hover-style: bold not underline; } ``` </issue> <code> [start of src/textual/widgets/_toast.py] 1 """Widgets for showing notification messages in toasts.""" 2 3 from __future__ import annotations 4 5 from typing import ClassVar 6 7 from rich.console import RenderableType 8 from rich.text import Text 9 10 from .. import on 11 from ..containers import Container 12 from ..css.query import NoMatches 13 from ..events import Click, Mount 14 from ..notifications import Notification, Notifications 15 from ._static import Static 16 17 18 class ToastHolder(Container, inherit_css=False): 19 """Container that holds a single toast. 20 21 Used to control the alignment of each of the toasts in the main toast 22 container. 23 """ 24 25 DEFAULT_CSS = """ 26 ToastHolder { 27 align-horizontal: right; 28 width: 1fr; 29 height: auto; 30 visibility: hidden; 31 } 32 """ 33 34 35 class Toast(Static, inherit_css=False): 36 """A widget for displaying short-lived notifications.""" 37 38 DEFAULT_CSS = """ 39 Toast { 40 width: 60; 41 max-width: 50%; 42 height: auto; 43 visibility: visible; 44 margin-top: 1; 45 padding: 1 1; 46 background: $panel; 47 tint: white 5%; 48 } 49 50 .toast--title { 51 text-style: bold; 52 } 53 54 Toast { 55 border-right: wide $background; 56 } 57 58 Toast.-information { 59 border-left: wide $success; 60 } 61 62 Toast.-information .toast--title { 63 color: $success-darken-1; 64 } 65 66 Toast.-warning { 67 border-left: wide $warning; 68 } 69 70 Toast.-warning .toast--title { 71 color: $warning-darken-1; 72 } 73 74 Toast.-error { 75 border-left: wide $error; 76 } 77 78 Toast.-error .toast--title { 79 color: $error-darken-1; 80 } 81 """ 82 83 COMPONENT_CLASSES: ClassVar[set[str]] = {"toast--title"} 84 """ 85 | Class | Description | 86 | :- | :- | 87 | `toast--title` | Targets the title of the toast. | 88 """ 89 90 def __init__(self, notification: Notification) -> None: 91 """Initialise the toast. 92 93 Args: 94 notification: The notification to show in the toast. 95 """ 96 super().__init__(classes=f"-{notification.severity}") 97 self._notification = notification 98 self._timeout = notification.time_left 99 100 def render(self) -> RenderableType: 101 """Render the toast's content. 102 103 Returns: 104 A Rich renderable for the title and content of the Toast. 105 """ 106 notification = self._notification 107 if notification.title: 108 header_style = self.get_component_rich_style("toast--title") 109 notification_text = Text.assemble( 110 (notification.title, header_style), 111 "\n", 112 Text.from_markup(notification.message), 113 ) 114 else: 115 notification_text = Text.assemble( 116 Text.from_markup(notification.message), 117 ) 118 return notification_text 119 120 def _on_mount(self, _: Mount) -> None: 121 """Set the time running once the toast is mounted.""" 122 self.set_timer(self._timeout, self._expire) 123 124 @on(Click) 125 def _expire(self) -> None: 126 """Remove the toast once the timer has expired.""" 127 # Before we removed ourself, we also call on the app to forget about 128 # the notification that caused us to exist. Note that we tell the 129 # app to not bother refreshing the display on our account, we're 130 # about to handle that anyway. 131 self.app._unnotify(self._notification, refresh=False) 132 # Note that we attempt to remove our parent, because we're wrapped 133 # inside an alignment container. The testing that we are is as much 134 # to keep type checkers happy as anything else. 135 (self.parent if isinstance(self.parent, ToastHolder) else self).remove() 136 137 138 class ToastRack(Container, inherit_css=False): 139 """A container for holding toasts.""" 140 141 DEFAULT_CSS = """ 142 ToastRack { 143 layer: _toastrack; 144 width: 1fr; 145 height: auto; 146 dock: top; 147 align: right bottom; 148 visibility: hidden; 149 layout: vertical; 150 overflow-y: scroll; 151 margin-bottom: 1; 152 margin-right: 1; 153 } 154 """ 155 156 @staticmethod 157 def _toast_id(notification: Notification) -> str: 158 """Create a Textual-DOM-internal ID for the given notification. 159 160 Args: 161 notification: The notification to create the ID for. 162 163 Returns: 164 An ID for the notification that can be used within the DOM. 165 """ 166 return f"--textual-toast-{notification.identity}" 167 168 def show(self, notifications: Notifications) -> None: 169 """Show the notifications as toasts. 170 171 Args: 172 notifications: The notifications to show. 173 """ 174 175 # Look for any stale toasts and remove them. 176 for toast in self.query(Toast): 177 if toast._notification not in notifications: 178 toast.remove() 179 180 # Gather up all the notifications that we don't have toasts for yet. 181 new_toasts: list[Notification] = [] 182 for notification in notifications: 183 try: 184 # See if there's already a toast for that notification. 185 _ = self.get_child_by_id(self._toast_id(notification)) 186 except NoMatches: 187 if not notification.has_expired: 188 new_toasts.append(notification) 189 190 # If we got any... 191 if new_toasts: 192 # ...mount them. 193 self.mount_all( 194 ToastHolder(Toast(toast), id=self._toast_id(toast)) 195 for toast in new_toasts 196 ) 197 self.call_later(self.scroll_end, animate=False, force=True) 198 [end of src/textual/widgets/_toast.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/src/textual/widgets/_toast.py b/src/textual/widgets/_toast.py --- a/src/textual/widgets/_toast.py +++ b/src/textual/widgets/_toast.py @@ -45,6 +45,12 @@ padding: 1 1; background: $panel; tint: white 5%; + link-background:; + link-color: $text; + link-style: underline; + link-hover-background: $accent; + link-hover-color: $text; + link-hover-style: bold not underline; } .toast--title {
{"golden_diff": "diff --git a/src/textual/widgets/_toast.py b/src/textual/widgets/_toast.py\n--- a/src/textual/widgets/_toast.py\n+++ b/src/textual/widgets/_toast.py\n@@ -45,6 +45,12 @@\n padding: 1 1;\n background: $panel;\n tint: white 5%;\n+ link-background:;\n+ link-color: $text;\n+ link-style: underline;\n+ link-hover-background: $accent;\n+ link-hover-color: $text;\n+ link-hover-style: bold not underline;\n }\n \n .toast--title {\n", "issue": "Missing link styles in `Toast` notifcations\nThis follows on from discussion #3526 about adding a clickable link in a notification.\r\n\r\n The `Toast` widget doesn't inherit the styling from the base class, so the \"default\" links styling defined in Widget is lost. You can obviously just add it back in like I've done here, but it might be nicer if you didn't have to do that.\r\n\r\n```python\r\n# toast_with_link.py\r\n\r\nimport webbrowser\r\n\r\nfrom textual.app import App\r\n\r\n\r\nclass ToastLinkApp(App):\r\n # Uncomment the CSS_PATH to see the link!\r\n # CSS_PATH = \"toast_with_link.css\"\r\n\r\n def on_mount(self) -> None:\r\n self.notify(\r\n \"Don't forget to \"\r\n \"[@click=link('https://github.com/Textualize/textual')]star[/] \"\r\n \"the repository!\",\r\n timeout=10,\r\n )\r\n\r\n def action_link(self, href: str) -> None:\r\n webbrowser.open(href)\r\n\r\n\r\nif __name__ == \"__main__\":\r\n app = ToastLinkApp()\r\n app.run()\r\n```\r\n\r\n```css\r\n/* toast_with_link.css */\r\n\r\nToast {\r\n link-background: ;\r\n link-color: $text;\r\n link-style: underline;\r\n link-hover-background: $accent;\r\n link-hover-color: $text;\r\n link-hover-style: bold not underline;\r\n}\r\n```\r\n\r\n\n", "before_files": [{"content": "\"\"\"Widgets for showing notification messages in toasts.\"\"\"\n\nfrom __future__ import annotations\n\nfrom typing import ClassVar\n\nfrom rich.console import RenderableType\nfrom rich.text import Text\n\nfrom .. import on\nfrom ..containers import Container\nfrom ..css.query import NoMatches\nfrom ..events import Click, Mount\nfrom ..notifications import Notification, Notifications\nfrom ._static import Static\n\n\nclass ToastHolder(Container, inherit_css=False):\n \"\"\"Container that holds a single toast.\n\n Used to control the alignment of each of the toasts in the main toast\n container.\n \"\"\"\n\n DEFAULT_CSS = \"\"\"\n ToastHolder {\n align-horizontal: right;\n width: 1fr;\n height: auto;\n visibility: hidden;\n }\n \"\"\"\n\n\nclass Toast(Static, inherit_css=False):\n \"\"\"A widget for displaying short-lived notifications.\"\"\"\n\n DEFAULT_CSS = \"\"\"\n Toast {\n width: 60;\n max-width: 50%;\n height: auto;\n visibility: visible;\n margin-top: 1;\n padding: 1 1;\n background: $panel;\n tint: white 5%;\n }\n\n .toast--title {\n text-style: bold;\n }\n\n Toast {\n border-right: wide $background;\n }\n\n Toast.-information {\n border-left: wide $success;\n }\n\n Toast.-information .toast--title {\n color: $success-darken-1;\n }\n\n Toast.-warning {\n border-left: wide $warning;\n }\n\n Toast.-warning .toast--title {\n color: $warning-darken-1;\n }\n\n Toast.-error {\n border-left: wide $error;\n }\n\n Toast.-error .toast--title {\n color: $error-darken-1;\n }\n \"\"\"\n\n COMPONENT_CLASSES: ClassVar[set[str]] = {\"toast--title\"}\n \"\"\"\n | Class | Description |\n | :- | :- |\n | `toast--title` | Targets the title of the toast. |\n \"\"\"\n\n def __init__(self, notification: Notification) -> None:\n \"\"\"Initialise the toast.\n\n Args:\n notification: The notification to show in the toast.\n \"\"\"\n super().__init__(classes=f\"-{notification.severity}\")\n self._notification = notification\n self._timeout = notification.time_left\n\n def render(self) -> RenderableType:\n \"\"\"Render the toast's content.\n\n Returns:\n A Rich renderable for the title and content of the Toast.\n \"\"\"\n notification = self._notification\n if notification.title:\n header_style = self.get_component_rich_style(\"toast--title\")\n notification_text = Text.assemble(\n (notification.title, header_style),\n \"\\n\",\n Text.from_markup(notification.message),\n )\n else:\n notification_text = Text.assemble(\n Text.from_markup(notification.message),\n )\n return notification_text\n\n def _on_mount(self, _: Mount) -> None:\n \"\"\"Set the time running once the toast is mounted.\"\"\"\n self.set_timer(self._timeout, self._expire)\n\n @on(Click)\n def _expire(self) -> None:\n \"\"\"Remove the toast once the timer has expired.\"\"\"\n # Before we removed ourself, we also call on the app to forget about\n # the notification that caused us to exist. Note that we tell the\n # app to not bother refreshing the display on our account, we're\n # about to handle that anyway.\n self.app._unnotify(self._notification, refresh=False)\n # Note that we attempt to remove our parent, because we're wrapped\n # inside an alignment container. The testing that we are is as much\n # to keep type checkers happy as anything else.\n (self.parent if isinstance(self.parent, ToastHolder) else self).remove()\n\n\nclass ToastRack(Container, inherit_css=False):\n \"\"\"A container for holding toasts.\"\"\"\n\n DEFAULT_CSS = \"\"\"\n ToastRack {\n layer: _toastrack;\n width: 1fr;\n height: auto;\n dock: top;\n align: right bottom;\n visibility: hidden;\n layout: vertical;\n overflow-y: scroll;\n margin-bottom: 1;\n margin-right: 1;\n }\n \"\"\"\n\n @staticmethod\n def _toast_id(notification: Notification) -> str:\n \"\"\"Create a Textual-DOM-internal ID for the given notification.\n\n Args:\n notification: The notification to create the ID for.\n\n Returns:\n An ID for the notification that can be used within the DOM.\n \"\"\"\n return f\"--textual-toast-{notification.identity}\"\n\n def show(self, notifications: Notifications) -> None:\n \"\"\"Show the notifications as toasts.\n\n Args:\n notifications: The notifications to show.\n \"\"\"\n\n # Look for any stale toasts and remove them.\n for toast in self.query(Toast):\n if toast._notification not in notifications:\n toast.remove()\n\n # Gather up all the notifications that we don't have toasts for yet.\n new_toasts: list[Notification] = []\n for notification in notifications:\n try:\n # See if there's already a toast for that notification.\n _ = self.get_child_by_id(self._toast_id(notification))\n except NoMatches:\n if not notification.has_expired:\n new_toasts.append(notification)\n\n # If we got any...\n if new_toasts:\n # ...mount them.\n self.mount_all(\n ToastHolder(Toast(toast), id=self._toast_id(toast))\n for toast in new_toasts\n )\n self.call_later(self.scroll_end, animate=False, force=True)\n", "path": "src/textual/widgets/_toast.py"}]}
2,538
131
gh_patches_debug_29108
rasdani/github-patches
git_diff
rlworkgroup__garage-774
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> Remove conda environment from CI/Dockerfiles Now that we use setup.py for tracking dependencies, there's no need be creating conda environments inside the Dockerfile and CI. We can replace them with a vanilla virtualenv for greatest compatibility. I suggest we add tests which ensure that the package can still be successfully solved and installed by conda and pipenv. If they don't take too long (<2min) then we can run them in the main CI, otherwise they will be cron_job tests (which are run only nightly). </issue> <code> [start of setup.py] 1 from setuptools import find_packages 2 from setuptools import setup 3 4 # Required dependencies 5 required = [ 6 # Please keep alphabetized 7 'akro==0.0.6', 8 'cached_property', 9 'click', 10 'cloudpickle', 11 'cma==1.1.06', 12 # dm_control throws an error during install about not being able to 13 # find a build dependency (absl-py). Later pip executes the `install` 14 # command again and the install succeeds because absl-py has been 15 # installed. This is stupid, but harmless. 16 'dm_control @ https://api.github.com/repos/deepmind/dm_control/tarball/7a36377879c57777e5d5b4da5aae2cd2a29b607a', # noqa: E501 17 'dowel==0.0.2', 18 'gym[all]==0.12.4', 19 'joblib<0.13,>=0.12', 20 'matplotlib', 21 'mujoco-py<2.1,>=2.0', 22 'numpy==1.14.5', 23 'psutil', 24 # Pyglet 1.4.0 introduces some api change which breaks some 25 # gym environments 26 # See: https://github.com/openai/gym/issues/1588 27 'pyglet<1.4.0,>=1.3.0', 28 'pyprind', 29 'python-dateutil', 30 'torch==1.1.0', 31 'scikit-image', 32 'scipy', 33 'tensorflow<1.13,>=1.12.0', 34 'tensorflow-probability<0.6.0,>=0.5.0', # for tensorflow 1.12 35 'torchvision==0.3.0' 36 ] 37 38 # Dependencies for optional features 39 extras = {} 40 extras['all'] = list(set(sum(extras.values(), []))) 41 42 # Intel dependencies not included in all 43 extras['intel'] = ['intel-tensorflow<1.13,>=1.12.0'] 44 45 # Development dependencies (*not* included in "all") 46 extras['dev'] = [ 47 # Please keep alphabetized 48 'baselines @ https://api.github.com/repos/openai/baselines/tarball/f2729693253c0ef4d4086231d36e0a4307ec1cb3', # noqa: E501 49 'flake8', 50 'flake8-docstrings==1.3.0', 51 'flake8-import-order', 52 'pandas', 53 'pep8-naming==0.7.0', 54 'pre-commit', 55 # pydocstyle 4.0.0 breaks flake8-docstrings 1.3.0 56 # See https://gitlab.com/pycqa/flake8-docstrings/issues/36 57 'pydocstyle<4.0.0', 58 'pylint==1.9.2', 59 'pytest>=3.6', # Required for pytest-cov on Python 3.6 60 'pytest-cov', 61 'sphinx', 62 'sphinx_rtd_theme', 63 'yapf', 64 ] 65 66 with open('README.md') as f: 67 readme = f.read() 68 69 # Get the package version dynamically 70 with open('VERSION') as v: 71 version = v.read().strip() 72 73 setup( 74 name='garage', 75 version=version, 76 author='Reinforcement Learning Working Group', 77 description='A framework for reproducible reinforcement learning research', 78 url='https://github.com/rlworkgroup/garage', 79 packages=find_packages(where='src'), 80 package_dir={'': 'src'}, 81 scripts=['scripts/garage'], 82 install_requires=required, 83 extras_require=extras, 84 license='MIT', 85 long_description=readme, 86 long_description_content_type='text/markdown', 87 classifiers=[ 88 'Development Status :: 4 - Beta', 89 'Intended Audience :: Developers', 90 'Intended Audience :: Education', 91 'Intended Audience :: Science/Research', 92 'License :: OSI Approved :: MIT License', 93 'Programming Language :: Python :: 3.5', 94 'Programming Language :: Python :: 3.6', 95 'Programming Language :: Python :: 3.7', 96 'Programming Language :: Python :: 3 :: Only', 97 'Topic :: Scientific/Engineering :: Artificial Intelligence', 98 'Topic :: Scientific/Engineering :: Mathematics', 99 'Topic :: Software Development :: Libraries', 100 ], 101 ) 102 [end of setup.py] [start of src/garage/np/algos/cem.py] 1 from dowel import logger, tabular 2 import numpy as np 3 4 from garage.np.algos import BatchPolopt 5 6 7 class CEM(BatchPolopt): 8 """Cross Entropy Method. 9 10 CEM works by iteratively optimizing a gaussian distribution of policy. 11 12 In each epoch, CEM does the following: 13 1. Sample n_samples policies from a gaussian distribution of 14 mean cur_mean and std cur_std. 15 2. Do rollouts for each policy. 16 3. Update cur_mean and cur_std by doing Maximum Likelihood Estimation 17 over the n_best top policies in terms of return. 18 19 Note: 20 When training CEM with LocalRunner, make sure that n_epoch_cycles for 21 runner equals to n_samples for CEM. 22 23 This implementation leverage n_epoch_cycles to do rollouts for a single 24 policy in an epoch cycle. 25 26 Args: 27 env_spec (garage.envs.EnvSpec): Environment specification. 28 policy (garage.np.policies.Policy): Action policy. 29 baseline(garage.np.baselines.Baseline): Baseline for GAE 30 (Generalized Advantage Estimation). 31 n_samples (int): Number of policies sampled in one epoch. 32 discount (float): Environment reward discount. 33 max_path_length (int): Maximum length of a single rollout. 34 best_frac (float): The best fraction. 35 init_std (float): Initial std for policy param distribution. 36 extra_std (float): Decaying std added to param distribution. 37 extra_decay_time (float): Epochs that it takes to decay extra std. 38 39 """ 40 41 def __init__(self, 42 env_spec, 43 policy, 44 baseline, 45 n_samples, 46 discount=0.99, 47 max_path_length=500, 48 init_std=1, 49 best_frac=0.05, 50 extra_std=1., 51 extra_decay_time=100): 52 super().__init__(policy, baseline, discount, max_path_length, 53 n_samples) 54 self.env_spec = env_spec 55 56 self.init_std = init_std 57 self.best_frac = best_frac 58 self.extra_std = extra_std 59 self.extra_decay_time = extra_decay_time 60 61 # epoch-wise 62 self.cur_std = self.init_std 63 self.cur_mean = self.policy.get_param_values() 64 # epoch-cycle-wise 65 self.cur_params = self.cur_mean 66 self.all_returns = [] 67 self.all_params = [self.cur_mean.copy()] 68 # fixed 69 self.n_best = int(n_samples * best_frac) 70 assert self.n_best >= 1, ( 71 f'n_samples is too low. Make sure that n_samples * best_frac >= 1') 72 self.n_params = len(self.cur_mean) 73 74 def sample_params(self, epoch): 75 extra_var_mult = max(1.0 - epoch / self.extra_decay_time, 0) 76 sample_std = np.sqrt( 77 np.square(self.cur_std) + 78 np.square(self.extra_std) * extra_var_mult) 79 return np.random.standard_normal( 80 self.n_params) * sample_std + self.cur_mean 81 82 def train_once(self, itr, paths): 83 paths = self.process_samples(itr, paths) 84 85 epoch = itr // self.n_samples 86 i_sample = itr - epoch * self.n_samples 87 tabular.record('Epoch', epoch) 88 tabular.record('# Sample', i_sample) 89 # -- Stage: Process path 90 rtn = paths['average_return'] 91 self.all_returns.append(paths['average_return']) 92 93 # -- Stage: Update policy distribution. 94 if (itr + 1) % self.n_samples == 0: 95 avg_rtns = np.array(self.all_returns) 96 best_inds = np.argsort(-avg_rtns)[:self.n_best] 97 best_params = np.array(self.all_params)[best_inds] 98 99 # MLE of normal distribution 100 self.cur_mean = best_params.mean(axis=0) 101 self.cur_std = best_params.std(axis=0) 102 self.policy.set_param_values(self.cur_mean) 103 104 # Clear for next epoch 105 rtn = max(self.all_returns) 106 self.all_returns.clear() 107 self.all_params.clear() 108 109 # -- Stage: Generate a new policy for next path sampling 110 self.cur_params = self.sample_params(itr) 111 self.all_params.append(self.cur_params.copy()) 112 self.policy.set_param_values(self.cur_params) 113 114 logger.log(tabular) 115 return rtn 116 117 def get_itr_snapshot(self, itr): 118 return dict(itr=itr, policy=self.policy, baseline=self.baseline) 119 [end of src/garage/np/algos/cem.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/setup.py b/setup.py --- a/setup.py +++ b/setup.py @@ -1,6 +1,15 @@ from setuptools import find_packages from setuptools import setup +try: + # pylint: disable=unused-import + import numpy # noqa: F401 +except ImportError: + raise RuntimeError( + 'garage requires numpy in the environment to install. ' + 'Please install numpy==1.14.5 and try again. See ' + 'https://github.com/rlworkgroup/garage/issues/800 for more info.') + # Required dependencies required = [ # Please keep alphabetized @@ -58,6 +67,7 @@ 'pylint==1.9.2', 'pytest>=3.6', # Required for pytest-cov on Python 3.6 'pytest-cov', + 'pytest-xdist', 'sphinx', 'sphinx_rtd_theme', 'yapf', @@ -79,6 +89,7 @@ packages=find_packages(where='src'), package_dir={'': 'src'}, scripts=['scripts/garage'], + python_requires='>=3.5', install_requires=required, extras_require=extras, license='MIT', diff --git a/src/garage/np/algos/cem.py b/src/garage/np/algos/cem.py --- a/src/garage/np/algos/cem.py +++ b/src/garage/np/algos/cem.py @@ -68,7 +68,7 @@ # fixed self.n_best = int(n_samples * best_frac) assert self.n_best >= 1, ( - f'n_samples is too low. Make sure that n_samples * best_frac >= 1') + 'n_samples is too low. Make sure that n_samples * best_frac >= 1') self.n_params = len(self.cur_mean) def sample_params(self, epoch):
{"golden_diff": "diff --git a/setup.py b/setup.py\n--- a/setup.py\n+++ b/setup.py\n@@ -1,6 +1,15 @@\n from setuptools import find_packages\n from setuptools import setup\n \n+try:\n+ # pylint: disable=unused-import\n+ import numpy # noqa: F401\n+except ImportError:\n+ raise RuntimeError(\n+ 'garage requires numpy in the environment to install. '\n+ 'Please install numpy==1.14.5 and try again. See '\n+ 'https://github.com/rlworkgroup/garage/issues/800 for more info.')\n+\n # Required dependencies\n required = [\n # Please keep alphabetized\n@@ -58,6 +67,7 @@\n 'pylint==1.9.2',\n 'pytest>=3.6', # Required for pytest-cov on Python 3.6\n 'pytest-cov',\n+ 'pytest-xdist',\n 'sphinx',\n 'sphinx_rtd_theme',\n 'yapf',\n@@ -79,6 +89,7 @@\n packages=find_packages(where='src'),\n package_dir={'': 'src'},\n scripts=['scripts/garage'],\n+ python_requires='>=3.5',\n install_requires=required,\n extras_require=extras,\n license='MIT',\ndiff --git a/src/garage/np/algos/cem.py b/src/garage/np/algos/cem.py\n--- a/src/garage/np/algos/cem.py\n+++ b/src/garage/np/algos/cem.py\n@@ -68,7 +68,7 @@\n # fixed\n self.n_best = int(n_samples * best_frac)\n assert self.n_best >= 1, (\n- f'n_samples is too low. Make sure that n_samples * best_frac >= 1')\n+ 'n_samples is too low. Make sure that n_samples * best_frac >= 1')\n self.n_params = len(self.cur_mean)\n \n def sample_params(self, epoch):\n", "issue": "Remove conda environment from CI/Dockerfiles\nNow that we use setup.py for tracking dependencies, there's no need be creating conda environments inside the Dockerfile and CI. We can replace them with a vanilla virtualenv for greatest compatibility.\r\n\r\nI suggest we add tests which ensure that the package can still be successfully solved and installed by conda and pipenv. If they don't take too long (<2min) then we can run them in the main CI, otherwise they will be cron_job tests (which are run only nightly).\n", "before_files": [{"content": "from setuptools import find_packages\nfrom setuptools import setup\n\n# Required dependencies\nrequired = [\n # Please keep alphabetized\n 'akro==0.0.6',\n 'cached_property',\n 'click',\n 'cloudpickle',\n 'cma==1.1.06',\n # dm_control throws an error during install about not being able to\n # find a build dependency (absl-py). Later pip executes the `install`\n # command again and the install succeeds because absl-py has been\n # installed. This is stupid, but harmless.\n 'dm_control @ https://api.github.com/repos/deepmind/dm_control/tarball/7a36377879c57777e5d5b4da5aae2cd2a29b607a', # noqa: E501\n 'dowel==0.0.2',\n 'gym[all]==0.12.4',\n 'joblib<0.13,>=0.12',\n 'matplotlib',\n 'mujoco-py<2.1,>=2.0',\n 'numpy==1.14.5',\n 'psutil',\n # Pyglet 1.4.0 introduces some api change which breaks some\n # gym environments\n # See: https://github.com/openai/gym/issues/1588\n 'pyglet<1.4.0,>=1.3.0',\n 'pyprind',\n 'python-dateutil',\n 'torch==1.1.0',\n 'scikit-image',\n 'scipy',\n 'tensorflow<1.13,>=1.12.0',\n 'tensorflow-probability<0.6.0,>=0.5.0', # for tensorflow 1.12\n 'torchvision==0.3.0'\n]\n\n# Dependencies for optional features\nextras = {}\nextras['all'] = list(set(sum(extras.values(), [])))\n\n# Intel dependencies not included in all\nextras['intel'] = ['intel-tensorflow<1.13,>=1.12.0']\n\n# Development dependencies (*not* included in \"all\")\nextras['dev'] = [\n # Please keep alphabetized\n 'baselines @ https://api.github.com/repos/openai/baselines/tarball/f2729693253c0ef4d4086231d36e0a4307ec1cb3', # noqa: E501\n 'flake8',\n 'flake8-docstrings==1.3.0',\n 'flake8-import-order',\n 'pandas',\n 'pep8-naming==0.7.0',\n 'pre-commit',\n # pydocstyle 4.0.0 breaks flake8-docstrings 1.3.0\n # See https://gitlab.com/pycqa/flake8-docstrings/issues/36\n 'pydocstyle<4.0.0',\n 'pylint==1.9.2',\n 'pytest>=3.6', # Required for pytest-cov on Python 3.6\n 'pytest-cov',\n 'sphinx',\n 'sphinx_rtd_theme',\n 'yapf',\n]\n\nwith open('README.md') as f:\n readme = f.read()\n\n# Get the package version dynamically\nwith open('VERSION') as v:\n version = v.read().strip()\n\nsetup(\n name='garage',\n version=version,\n author='Reinforcement Learning Working Group',\n description='A framework for reproducible reinforcement learning research',\n url='https://github.com/rlworkgroup/garage',\n packages=find_packages(where='src'),\n package_dir={'': 'src'},\n scripts=['scripts/garage'],\n install_requires=required,\n extras_require=extras,\n license='MIT',\n long_description=readme,\n long_description_content_type='text/markdown',\n classifiers=[\n 'Development Status :: 4 - Beta',\n 'Intended Audience :: Developers',\n 'Intended Audience :: Education',\n 'Intended Audience :: Science/Research',\n 'License :: OSI Approved :: MIT License',\n 'Programming Language :: Python :: 3.5',\n 'Programming Language :: Python :: 3.6',\n 'Programming Language :: Python :: 3.7',\n 'Programming Language :: Python :: 3 :: Only',\n 'Topic :: Scientific/Engineering :: Artificial Intelligence',\n 'Topic :: Scientific/Engineering :: Mathematics',\n 'Topic :: Software Development :: Libraries',\n ],\n)\n", "path": "setup.py"}, {"content": "from dowel import logger, tabular\nimport numpy as np\n\nfrom garage.np.algos import BatchPolopt\n\n\nclass CEM(BatchPolopt):\n \"\"\"Cross Entropy Method.\n\n CEM works by iteratively optimizing a gaussian distribution of policy.\n\n In each epoch, CEM does the following:\n 1. Sample n_samples policies from a gaussian distribution of\n mean cur_mean and std cur_std.\n 2. Do rollouts for each policy.\n 3. Update cur_mean and cur_std by doing Maximum Likelihood Estimation\n over the n_best top policies in terms of return.\n\n Note:\n When training CEM with LocalRunner, make sure that n_epoch_cycles for\n runner equals to n_samples for CEM.\n\n This implementation leverage n_epoch_cycles to do rollouts for a single\n policy in an epoch cycle.\n\n Args:\n env_spec (garage.envs.EnvSpec): Environment specification.\n policy (garage.np.policies.Policy): Action policy.\n baseline(garage.np.baselines.Baseline): Baseline for GAE\n (Generalized Advantage Estimation).\n n_samples (int): Number of policies sampled in one epoch.\n discount (float): Environment reward discount.\n max_path_length (int): Maximum length of a single rollout.\n best_frac (float): The best fraction.\n init_std (float): Initial std for policy param distribution.\n extra_std (float): Decaying std added to param distribution.\n extra_decay_time (float): Epochs that it takes to decay extra std.\n\n \"\"\"\n\n def __init__(self,\n env_spec,\n policy,\n baseline,\n n_samples,\n discount=0.99,\n max_path_length=500,\n init_std=1,\n best_frac=0.05,\n extra_std=1.,\n extra_decay_time=100):\n super().__init__(policy, baseline, discount, max_path_length,\n n_samples)\n self.env_spec = env_spec\n\n self.init_std = init_std\n self.best_frac = best_frac\n self.extra_std = extra_std\n self.extra_decay_time = extra_decay_time\n\n # epoch-wise\n self.cur_std = self.init_std\n self.cur_mean = self.policy.get_param_values()\n # epoch-cycle-wise\n self.cur_params = self.cur_mean\n self.all_returns = []\n self.all_params = [self.cur_mean.copy()]\n # fixed\n self.n_best = int(n_samples * best_frac)\n assert self.n_best >= 1, (\n f'n_samples is too low. Make sure that n_samples * best_frac >= 1')\n self.n_params = len(self.cur_mean)\n\n def sample_params(self, epoch):\n extra_var_mult = max(1.0 - epoch / self.extra_decay_time, 0)\n sample_std = np.sqrt(\n np.square(self.cur_std) +\n np.square(self.extra_std) * extra_var_mult)\n return np.random.standard_normal(\n self.n_params) * sample_std + self.cur_mean\n\n def train_once(self, itr, paths):\n paths = self.process_samples(itr, paths)\n\n epoch = itr // self.n_samples\n i_sample = itr - epoch * self.n_samples\n tabular.record('Epoch', epoch)\n tabular.record('# Sample', i_sample)\n # -- Stage: Process path\n rtn = paths['average_return']\n self.all_returns.append(paths['average_return'])\n\n # -- Stage: Update policy distribution.\n if (itr + 1) % self.n_samples == 0:\n avg_rtns = np.array(self.all_returns)\n best_inds = np.argsort(-avg_rtns)[:self.n_best]\n best_params = np.array(self.all_params)[best_inds]\n\n # MLE of normal distribution\n self.cur_mean = best_params.mean(axis=0)\n self.cur_std = best_params.std(axis=0)\n self.policy.set_param_values(self.cur_mean)\n\n # Clear for next epoch\n rtn = max(self.all_returns)\n self.all_returns.clear()\n self.all_params.clear()\n\n # -- Stage: Generate a new policy for next path sampling\n self.cur_params = self.sample_params(itr)\n self.all_params.append(self.cur_params.copy())\n self.policy.set_param_values(self.cur_params)\n\n logger.log(tabular)\n return rtn\n\n def get_itr_snapshot(self, itr):\n return dict(itr=itr, policy=self.policy, baseline=self.baseline)\n", "path": "src/garage/np/algos/cem.py"}]}
3,101
447
gh_patches_debug_34606
rasdani/github-patches
git_diff
ansible__awx-8016
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> Add insignts_credential paramter to tower_inventory <!-- Issues are for **concrete, actionable bugs and feature requests** only - if you're just asking for debugging help or technical support, please use: - http://webchat.freenode.net/?channels=ansible-awx - https://groups.google.com/forum/#!forum/awx-project We have to limit this because of limited volunteer time to respond to issues! --> ##### ISSUE TYPE - Feature Idea ##### SUMMARY <!-- Briefly describe the problem or desired enhancement. --> Per PR #7963 tower_inventory is missing support for the insights_credential API parameter. </issue> <code> [start of awx_collection/plugins/modules/tower_inventory.py] 1 #!/usr/bin/python 2 # coding: utf-8 -*- 3 4 # (c) 2017, Wayne Witzel III <[email protected]> 5 # GNU General Public License v3.0+ (see COPYING or https://www.gnu.org/licenses/gpl-3.0.txt) 6 7 from __future__ import absolute_import, division, print_function 8 __metaclass__ = type 9 10 11 ANSIBLE_METADATA = {'metadata_version': '1.1', 12 'status': ['preview'], 13 'supported_by': 'community'} 14 15 16 DOCUMENTATION = ''' 17 --- 18 module: tower_inventory 19 author: "Wayne Witzel III (@wwitzel3)" 20 short_description: create, update, or destroy Ansible Tower inventory. 21 description: 22 - Create, update, or destroy Ansible Tower inventories. See 23 U(https://www.ansible.com/tower) for an overview. 24 options: 25 name: 26 description: 27 - The name to use for the inventory. 28 required: True 29 type: str 30 description: 31 description: 32 - The description to use for the inventory. 33 type: str 34 organization: 35 description: 36 - Organization the inventory belongs to. 37 required: True 38 type: str 39 variables: 40 description: 41 - Inventory variables. 42 type: dict 43 kind: 44 description: 45 - The kind field. Cannot be modified after created. 46 default: "" 47 choices: ["", "smart"] 48 type: str 49 host_filter: 50 description: 51 - The host_filter field. Only useful when C(kind=smart). 52 type: str 53 state: 54 description: 55 - Desired state of the resource. 56 default: "present" 57 choices: ["present", "absent"] 58 type: str 59 extends_documentation_fragment: awx.awx.auth 60 ''' 61 62 63 EXAMPLES = ''' 64 - name: Add tower inventory 65 tower_inventory: 66 name: "Foo Inventory" 67 description: "Our Foo Cloud Servers" 68 organization: "Bar Org" 69 state: present 70 tower_config_file: "~/tower_cli.cfg" 71 ''' 72 73 74 from ..module_utils.tower_api import TowerAPIModule 75 import json 76 77 78 def main(): 79 # Any additional arguments that are not fields of the item can be added here 80 argument_spec = dict( 81 name=dict(required=True), 82 description=dict(), 83 organization=dict(required=True), 84 variables=dict(type='dict'), 85 kind=dict(choices=['', 'smart'], default=''), 86 host_filter=dict(), 87 state=dict(choices=['present', 'absent'], default='present'), 88 ) 89 90 # Create a module for ourselves 91 module = TowerAPIModule(argument_spec=argument_spec) 92 93 # Extract our parameters 94 name = module.params.get('name') 95 description = module.params.get('description') 96 organization = module.params.get('organization') 97 variables = module.params.get('variables') 98 state = module.params.get('state') 99 kind = module.params.get('kind') 100 host_filter = module.params.get('host_filter') 101 102 # Attempt to look up the related items the user specified (these will fail the module if not found) 103 org_id = module.resolve_name_to_id('organizations', organization) 104 105 # Attempt to look up inventory based on the provided name and org ID 106 inventory = module.get_one('inventories', **{ 107 'data': { 108 'name': name, 109 'organization': org_id 110 } 111 }) 112 113 if state == 'absent': 114 # If the state was absent we can let the module delete it if needed, the module will handle exiting from this 115 module.delete_if_needed(inventory) 116 117 # Create the data that gets sent for create and update 118 inventory_fields = { 119 'name': name, 120 'organization': org_id, 121 'kind': kind, 122 'host_filter': host_filter, 123 } 124 if description is not None: 125 inventory_fields['description'] = description 126 if variables is not None: 127 inventory_fields['variables'] = json.dumps(variables) 128 129 # We need to perform a check to make sure you are not trying to convert a regular inventory into a smart one. 130 if inventory and inventory['kind'] == '' and inventory_fields['kind'] == 'smart': 131 module.fail_json(msg='You cannot turn a regular inventory into a "smart" inventory.') 132 133 # If the state was present and we can let the module build or update the existing inventory, this will return on its own 134 module.create_or_update_if_needed(inventory, inventory_fields, endpoint='inventories', item_type='inventory') 135 136 137 if __name__ == '__main__': 138 main() 139 [end of awx_collection/plugins/modules/tower_inventory.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/awx_collection/plugins/modules/tower_inventory.py b/awx_collection/plugins/modules/tower_inventory.py --- a/awx_collection/plugins/modules/tower_inventory.py +++ b/awx_collection/plugins/modules/tower_inventory.py @@ -48,7 +48,11 @@ type: str host_filter: description: - - The host_filter field. Only useful when C(kind=smart). + - The host_filter field. Only useful when C(kind=smart). + type: str + insights_credential: + description: + - Credentials to be used by hosts belonging to this inventory when accessing Red Hat Insights API. type: str state: description: @@ -84,6 +88,7 @@ variables=dict(type='dict'), kind=dict(choices=['', 'smart'], default=''), host_filter=dict(), + insights_credential=dict(), state=dict(choices=['present', 'absent'], default='present'), ) @@ -98,6 +103,7 @@ state = module.params.get('state') kind = module.params.get('kind') host_filter = module.params.get('host_filter') + insights_credential = module.params.get('insights_credential') # Attempt to look up the related items the user specified (these will fail the module if not found) org_id = module.resolve_name_to_id('organizations', organization) @@ -125,6 +131,8 @@ inventory_fields['description'] = description if variables is not None: inventory_fields['variables'] = json.dumps(variables) + if insights_credential is not None: + inventory_fields['insights_credential'] = module.resolve_name_to_id('credentials', insights_credential) # We need to perform a check to make sure you are not trying to convert a regular inventory into a smart one. if inventory and inventory['kind'] == '' and inventory_fields['kind'] == 'smart':
{"golden_diff": "diff --git a/awx_collection/plugins/modules/tower_inventory.py b/awx_collection/plugins/modules/tower_inventory.py\n--- a/awx_collection/plugins/modules/tower_inventory.py\n+++ b/awx_collection/plugins/modules/tower_inventory.py\n@@ -48,7 +48,11 @@\n type: str\n host_filter:\n description:\n- - The host_filter field. Only useful when C(kind=smart).\n+ - The host_filter field. Only useful when C(kind=smart).\n+ type: str\n+ insights_credential:\n+ description:\n+ - Credentials to be used by hosts belonging to this inventory when accessing Red Hat Insights API.\n type: str\n state:\n description:\n@@ -84,6 +88,7 @@\n variables=dict(type='dict'),\n kind=dict(choices=['', 'smart'], default=''),\n host_filter=dict(),\n+ insights_credential=dict(),\n state=dict(choices=['present', 'absent'], default='present'),\n )\n \n@@ -98,6 +103,7 @@\n state = module.params.get('state')\n kind = module.params.get('kind')\n host_filter = module.params.get('host_filter')\n+ insights_credential = module.params.get('insights_credential')\n \n # Attempt to look up the related items the user specified (these will fail the module if not found)\n org_id = module.resolve_name_to_id('organizations', organization)\n@@ -125,6 +131,8 @@\n inventory_fields['description'] = description\n if variables is not None:\n inventory_fields['variables'] = json.dumps(variables)\n+ if insights_credential is not None:\n+ inventory_fields['insights_credential'] = module.resolve_name_to_id('credentials', insights_credential)\n \n # We need to perform a check to make sure you are not trying to convert a regular inventory into a smart one.\n if inventory and inventory['kind'] == '' and inventory_fields['kind'] == 'smart':\n", "issue": "Add insignts_credential paramter to tower_inventory\n<!-- Issues are for **concrete, actionable bugs and feature requests** only - if you're just asking for debugging help or technical support, please use:\r\n\r\n- http://webchat.freenode.net/?channels=ansible-awx\r\n- https://groups.google.com/forum/#!forum/awx-project\r\n\r\nWe have to limit this because of limited volunteer time to respond to issues! -->\r\n\r\n##### ISSUE TYPE\r\n - Feature Idea\r\n\r\n##### SUMMARY\r\n<!-- Briefly describe the problem or desired enhancement. -->\r\nPer PR #7963 tower_inventory is missing support for the insights_credential API parameter.\n", "before_files": [{"content": "#!/usr/bin/python\n# coding: utf-8 -*-\n\n# (c) 2017, Wayne Witzel III <[email protected]>\n# GNU General Public License v3.0+ (see COPYING or https://www.gnu.org/licenses/gpl-3.0.txt)\n\nfrom __future__ import absolute_import, division, print_function\n__metaclass__ = type\n\n\nANSIBLE_METADATA = {'metadata_version': '1.1',\n 'status': ['preview'],\n 'supported_by': 'community'}\n\n\nDOCUMENTATION = '''\n---\nmodule: tower_inventory\nauthor: \"Wayne Witzel III (@wwitzel3)\"\nshort_description: create, update, or destroy Ansible Tower inventory.\ndescription:\n - Create, update, or destroy Ansible Tower inventories. See\n U(https://www.ansible.com/tower) for an overview.\noptions:\n name:\n description:\n - The name to use for the inventory.\n required: True\n type: str\n description:\n description:\n - The description to use for the inventory.\n type: str\n organization:\n description:\n - Organization the inventory belongs to.\n required: True\n type: str\n variables:\n description:\n - Inventory variables.\n type: dict\n kind:\n description:\n - The kind field. Cannot be modified after created.\n default: \"\"\n choices: [\"\", \"smart\"]\n type: str\n host_filter:\n description:\n - The host_filter field. Only useful when C(kind=smart).\n type: str\n state:\n description:\n - Desired state of the resource.\n default: \"present\"\n choices: [\"present\", \"absent\"]\n type: str\nextends_documentation_fragment: awx.awx.auth\n'''\n\n\nEXAMPLES = '''\n- name: Add tower inventory\n tower_inventory:\n name: \"Foo Inventory\"\n description: \"Our Foo Cloud Servers\"\n organization: \"Bar Org\"\n state: present\n tower_config_file: \"~/tower_cli.cfg\"\n'''\n\n\nfrom ..module_utils.tower_api import TowerAPIModule\nimport json\n\n\ndef main():\n # Any additional arguments that are not fields of the item can be added here\n argument_spec = dict(\n name=dict(required=True),\n description=dict(),\n organization=dict(required=True),\n variables=dict(type='dict'),\n kind=dict(choices=['', 'smart'], default=''),\n host_filter=dict(),\n state=dict(choices=['present', 'absent'], default='present'),\n )\n\n # Create a module for ourselves\n module = TowerAPIModule(argument_spec=argument_spec)\n\n # Extract our parameters\n name = module.params.get('name')\n description = module.params.get('description')\n organization = module.params.get('organization')\n variables = module.params.get('variables')\n state = module.params.get('state')\n kind = module.params.get('kind')\n host_filter = module.params.get('host_filter')\n\n # Attempt to look up the related items the user specified (these will fail the module if not found)\n org_id = module.resolve_name_to_id('organizations', organization)\n\n # Attempt to look up inventory based on the provided name and org ID\n inventory = module.get_one('inventories', **{\n 'data': {\n 'name': name,\n 'organization': org_id\n }\n })\n\n if state == 'absent':\n # If the state was absent we can let the module delete it if needed, the module will handle exiting from this\n module.delete_if_needed(inventory)\n\n # Create the data that gets sent for create and update\n inventory_fields = {\n 'name': name,\n 'organization': org_id,\n 'kind': kind,\n 'host_filter': host_filter,\n }\n if description is not None:\n inventory_fields['description'] = description\n if variables is not None:\n inventory_fields['variables'] = json.dumps(variables)\n\n # We need to perform a check to make sure you are not trying to convert a regular inventory into a smart one.\n if inventory and inventory['kind'] == '' and inventory_fields['kind'] == 'smart':\n module.fail_json(msg='You cannot turn a regular inventory into a \"smart\" inventory.')\n\n # If the state was present and we can let the module build or update the existing inventory, this will return on its own\n module.create_or_update_if_needed(inventory, inventory_fields, endpoint='inventories', item_type='inventory')\n\n\nif __name__ == '__main__':\n main()\n", "path": "awx_collection/plugins/modules/tower_inventory.py"}]}
1,972
433
gh_patches_debug_12719
rasdani/github-patches
git_diff
microsoft__playwright-python-593
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> Async API - asyncio error After updating to the 1.9.2 version, for **Async API** I get the following error: Exception ignored in: <function BaseSubprocessTransport.__del__ at 0x0000000002E1E9D0> Traceback (most recent call last): File "C:\Users\Matthew\AppData\Local\Programs\Python\Python39\lib\asyncio\base_subprocess.py", line 126, in __del__ self.close() File "C:\Users\Matthew\AppData\Local\Programs\Python\Python39\lib\asyncio\base_subprocess.py", line 104, in close proto.pipe.close() File "C:\Users\Matthew\AppData\Local\Programs\Python\Python39\lib\asyncio\proactor_events.py", line 108, in close self._loop.call_soon(self._call_connection_lost, None) File "C:\Users\Matthew\AppData\Local\Programs\Python\Python39\lib\asyncio\base_events.py", line 746, in call_soon self._check_closed() File "C:\Users\Matthew\AppData\Local\Programs\Python\Python39\lib\asyncio\base_events.py", line 510, in _check_closed raise RuntimeError('Event loop is closed') RuntimeError: Event loop is closed Exception ignored in: <function _ProactorBasePipeTransport.__del__ at 0x0000000002E4A280> Traceback (most recent call last): File "C:\Users\Matthew\AppData\Local\Programs\Python\Python39\lib\asyncio\proactor_events.py", line 116, in __del__ self.close() File "C:\Users\Matthew\AppData\Local\Programs\Python\Python39\lib\asyncio\proactor_events.py", line 108, in close self._loop.call_soon(self._call_connection_lost, None) File "C:\Users\Matthew\AppData\Local\Programs\Python\Python39\lib\asyncio\base_events.py", line 746, in call_soon self._check_closed() File "C:\Users\Matthew\AppData\Local\Programs\Python\Python39\lib\asyncio\base_events.py", line 510, in _check_closed raise RuntimeError('Event loop is closed') RuntimeError: Event loop is closed Versions: playwright=1.9.2 asyncio=3.4.3 </issue> <code> [start of playwright/_impl/_transport.py] 1 # Copyright (c) Microsoft Corporation. 2 # 3 # Licensed under the Apache License, Version 2.0 (the "License"); 4 # you may not use this file except in compliance with the License. 5 # You may obtain a copy of the License at 6 # 7 # http://www.apache.org/licenses/LICENSE-2.0 8 # 9 # Unless required by applicable law or agreed to in writing, software 10 # distributed under the License is distributed on an "AS IS" BASIS, 11 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 12 # See the License for the specific language governing permissions and 13 # limitations under the License. 14 15 import asyncio 16 import io 17 import json 18 import os 19 import sys 20 from pathlib import Path 21 from typing import Dict, Optional 22 23 24 # Sourced from: https://github.com/pytest-dev/pytest/blob/da01ee0a4bb0af780167ecd228ab3ad249511302/src/_pytest/faulthandler.py#L69-L77 25 def _get_stderr_fileno() -> Optional[int]: 26 try: 27 return sys.stderr.fileno() 28 except (AttributeError, io.UnsupportedOperation): 29 # pytest-xdist monkeypatches sys.stderr with an object that is not an actual file. 30 # https://docs.python.org/3/library/faulthandler.html#issue-with-file-descriptors 31 # This is potentially dangerous, but the best we can do. 32 if not hasattr(sys, "__stderr__") or not sys.__stderr__: 33 return None 34 return sys.__stderr__.fileno() 35 36 37 class Transport: 38 def __init__(self, driver_executable: Path) -> None: 39 super().__init__() 40 self.on_message = lambda _: None 41 self._stopped = False 42 self._driver_executable = driver_executable 43 self._loop: asyncio.AbstractEventLoop 44 45 def stop(self) -> None: 46 self._stopped = True 47 self._output.close() 48 49 async def wait_until_stopped(self) -> None: 50 await self._stopped_future 51 52 async def run(self) -> None: 53 self._loop = asyncio.get_running_loop() 54 self._stopped_future: asyncio.Future = asyncio.Future() 55 56 proc = await asyncio.create_subprocess_exec( 57 str(self._driver_executable), 58 "run-driver", 59 stdin=asyncio.subprocess.PIPE, 60 stdout=asyncio.subprocess.PIPE, 61 stderr=_get_stderr_fileno(), 62 limit=32768, 63 ) 64 assert proc.stdout 65 assert proc.stdin 66 self._output = proc.stdin 67 68 while not self._stopped: 69 try: 70 buffer = await proc.stdout.readexactly(4) 71 length = int.from_bytes(buffer, byteorder="little", signed=False) 72 buffer = bytes(0) 73 while length: 74 to_read = min(length, 32768) 75 data = await proc.stdout.readexactly(to_read) 76 length -= to_read 77 if len(buffer): 78 buffer = buffer + data 79 else: 80 buffer = data 81 obj = json.loads(buffer) 82 83 if "DEBUGP" in os.environ: # pragma: no cover 84 print("\x1b[33mRECV>\x1b[0m", json.dumps(obj, indent=2)) 85 self.on_message(obj) 86 except asyncio.IncompleteReadError: 87 break 88 await asyncio.sleep(0) 89 self._stopped_future.set_result(None) 90 91 def send(self, message: Dict) -> None: 92 msg = json.dumps(message) 93 if "DEBUGP" in os.environ: # pragma: no cover 94 print("\x1b[32mSEND>\x1b[0m", json.dumps(message, indent=2)) 95 data = msg.encode() 96 self._output.write( 97 len(data).to_bytes(4, byteorder="little", signed=False) + data 98 ) 99 [end of playwright/_impl/_transport.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/playwright/_impl/_transport.py b/playwright/_impl/_transport.py --- a/playwright/_impl/_transport.py +++ b/playwright/_impl/_transport.py @@ -48,12 +48,13 @@ async def wait_until_stopped(self) -> None: await self._stopped_future + await self._proc.wait() async def run(self) -> None: self._loop = asyncio.get_running_loop() self._stopped_future: asyncio.Future = asyncio.Future() - proc = await asyncio.create_subprocess_exec( + self._proc = proc = await asyncio.create_subprocess_exec( str(self._driver_executable), "run-driver", stdin=asyncio.subprocess.PIPE,
{"golden_diff": "diff --git a/playwright/_impl/_transport.py b/playwright/_impl/_transport.py\n--- a/playwright/_impl/_transport.py\n+++ b/playwright/_impl/_transport.py\n@@ -48,12 +48,13 @@\n \n async def wait_until_stopped(self) -> None:\n await self._stopped_future\n+ await self._proc.wait()\n \n async def run(self) -> None:\n self._loop = asyncio.get_running_loop()\n self._stopped_future: asyncio.Future = asyncio.Future()\n \n- proc = await asyncio.create_subprocess_exec(\n+ self._proc = proc = await asyncio.create_subprocess_exec(\n str(self._driver_executable),\n \"run-driver\",\n stdin=asyncio.subprocess.PIPE,\n", "issue": "Async API - asyncio error\nAfter updating to the 1.9.2 version, for **Async API** I get the following error:\r\n\r\nException ignored in: <function BaseSubprocessTransport.__del__ at 0x0000000002E1E9D0>\r\nTraceback (most recent call last):\r\n File \"C:\\Users\\Matthew\\AppData\\Local\\Programs\\Python\\Python39\\lib\\asyncio\\base_subprocess.py\", line 126, in __del__\r\n self.close()\r\n File \"C:\\Users\\Matthew\\AppData\\Local\\Programs\\Python\\Python39\\lib\\asyncio\\base_subprocess.py\", line 104, in close \r\n proto.pipe.close()\r\n File \"C:\\Users\\Matthew\\AppData\\Local\\Programs\\Python\\Python39\\lib\\asyncio\\proactor_events.py\", line 108, in close \r\n self._loop.call_soon(self._call_connection_lost, None)\r\n File \"C:\\Users\\Matthew\\AppData\\Local\\Programs\\Python\\Python39\\lib\\asyncio\\base_events.py\", line 746, in call_soon\r\n self._check_closed()\r\n File \"C:\\Users\\Matthew\\AppData\\Local\\Programs\\Python\\Python39\\lib\\asyncio\\base_events.py\", line 510, in _check_closed\r\n raise RuntimeError('Event loop is closed')\r\nRuntimeError: Event loop is closed\r\nException ignored in: <function _ProactorBasePipeTransport.__del__ at 0x0000000002E4A280>\r\nTraceback (most recent call last):\r\n File \"C:\\Users\\Matthew\\AppData\\Local\\Programs\\Python\\Python39\\lib\\asyncio\\proactor_events.py\", line 116, in __del__\r\n self.close()\r\n File \"C:\\Users\\Matthew\\AppData\\Local\\Programs\\Python\\Python39\\lib\\asyncio\\proactor_events.py\", line 108, in close\r\n self._loop.call_soon(self._call_connection_lost, None)\r\n File \"C:\\Users\\Matthew\\AppData\\Local\\Programs\\Python\\Python39\\lib\\asyncio\\base_events.py\", line 746, in call_soon\r\n self._check_closed()\r\n File \"C:\\Users\\Matthew\\AppData\\Local\\Programs\\Python\\Python39\\lib\\asyncio\\base_events.py\", line 510, in _check_closed\r\n raise RuntimeError('Event loop is closed')\r\nRuntimeError: Event loop is closed\r\n\r\nVersions:\r\nplaywright=1.9.2\r\nasyncio=3.4.3\n", "before_files": [{"content": "# Copyright (c) Microsoft Corporation.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nimport asyncio\nimport io\nimport json\nimport os\nimport sys\nfrom pathlib import Path\nfrom typing import Dict, Optional\n\n\n# Sourced from: https://github.com/pytest-dev/pytest/blob/da01ee0a4bb0af780167ecd228ab3ad249511302/src/_pytest/faulthandler.py#L69-L77\ndef _get_stderr_fileno() -> Optional[int]:\n try:\n return sys.stderr.fileno()\n except (AttributeError, io.UnsupportedOperation):\n # pytest-xdist monkeypatches sys.stderr with an object that is not an actual file.\n # https://docs.python.org/3/library/faulthandler.html#issue-with-file-descriptors\n # This is potentially dangerous, but the best we can do.\n if not hasattr(sys, \"__stderr__\") or not sys.__stderr__:\n return None\n return sys.__stderr__.fileno()\n\n\nclass Transport:\n def __init__(self, driver_executable: Path) -> None:\n super().__init__()\n self.on_message = lambda _: None\n self._stopped = False\n self._driver_executable = driver_executable\n self._loop: asyncio.AbstractEventLoop\n\n def stop(self) -> None:\n self._stopped = True\n self._output.close()\n\n async def wait_until_stopped(self) -> None:\n await self._stopped_future\n\n async def run(self) -> None:\n self._loop = asyncio.get_running_loop()\n self._stopped_future: asyncio.Future = asyncio.Future()\n\n proc = await asyncio.create_subprocess_exec(\n str(self._driver_executable),\n \"run-driver\",\n stdin=asyncio.subprocess.PIPE,\n stdout=asyncio.subprocess.PIPE,\n stderr=_get_stderr_fileno(),\n limit=32768,\n )\n assert proc.stdout\n assert proc.stdin\n self._output = proc.stdin\n\n while not self._stopped:\n try:\n buffer = await proc.stdout.readexactly(4)\n length = int.from_bytes(buffer, byteorder=\"little\", signed=False)\n buffer = bytes(0)\n while length:\n to_read = min(length, 32768)\n data = await proc.stdout.readexactly(to_read)\n length -= to_read\n if len(buffer):\n buffer = buffer + data\n else:\n buffer = data\n obj = json.loads(buffer)\n\n if \"DEBUGP\" in os.environ: # pragma: no cover\n print(\"\\x1b[33mRECV>\\x1b[0m\", json.dumps(obj, indent=2))\n self.on_message(obj)\n except asyncio.IncompleteReadError:\n break\n await asyncio.sleep(0)\n self._stopped_future.set_result(None)\n\n def send(self, message: Dict) -> None:\n msg = json.dumps(message)\n if \"DEBUGP\" in os.environ: # pragma: no cover\n print(\"\\x1b[32mSEND>\\x1b[0m\", json.dumps(message, indent=2))\n data = msg.encode()\n self._output.write(\n len(data).to_bytes(4, byteorder=\"little\", signed=False) + data\n )\n", "path": "playwright/_impl/_transport.py"}]}
2,149
163
gh_patches_debug_47861
rasdani/github-patches
git_diff
saleor__saleor-4008
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> Duplicate choices in shipping address ### What I'm trying to achieve I'm trying to set shipping user for Russian user and there are duplicate values in "Oblast" selector. ### Steps to reproduce the problem 1. Create new shipping address 2. Country --> Russia 3. There are duplicated values in "Oblast" selector ### What I expected to happen There are no duplicated values in "Oblast" selector ### Screenshots What happens now <img src="https://user-images.githubusercontent.com/13136992/53255369-8a239600-36d6-11e9-84a6-24a10b96a321.png" width="300"> What I expect to see <img src="https://user-images.githubusercontent.com/13136992/53255400-99a2df00-36d6-11e9-8913-ecaec174487a.png" width="300"> **System information** Operating system: Manjaro Linux Browser: Google Chrome 72.0.3626.81 </issue> <code> [start of saleor/account/forms.py] 1 from captcha.fields import ReCaptchaField 2 from django import forms 3 from django.conf import settings 4 from django.contrib.auth import forms as django_forms, update_session_auth_hash 5 from django.utils.translation import pgettext, pgettext_lazy 6 from phonenumbers.phonenumberutil import country_code_for_region 7 8 from ..account.models import User 9 from . import emails 10 from .i18n import AddressMetaForm, get_address_form_class 11 12 13 class FormWithReCaptcha(forms.BaseForm): 14 def __new__(cls, *args, **kwargs): 15 if settings.RECAPTCHA_PUBLIC_KEY and settings.RECAPTCHA_PRIVATE_KEY: 16 # insert a Google reCaptcha field inside the form 17 # note: label is empty, the reCaptcha is self-explanatory making 18 # the form simpler for the user. 19 cls.base_fields['_captcha'] = ReCaptchaField(label='') 20 return super(FormWithReCaptcha, cls).__new__(cls) 21 22 23 def get_address_form( 24 data, country_code, initial=None, instance=None, **kwargs): 25 country_form = AddressMetaForm(data, initial=initial) 26 preview = False 27 if country_form.is_valid(): 28 country_code = country_form.cleaned_data['country'] 29 preview = country_form.cleaned_data['preview'] 30 31 if initial is None and country_code: 32 initial = {} 33 if country_code: 34 initial['phone'] = '+{}'.format(country_code_for_region(country_code)) 35 36 address_form_class = get_address_form_class(country_code) 37 38 if not preview and instance is not None: 39 address_form_class = get_address_form_class(instance.country.code) 40 address_form = address_form_class(data, instance=instance, **kwargs) 41 else: 42 initial_address = ( 43 initial if not preview 44 else data.dict() if data is not None else data) 45 address_form = address_form_class( 46 not preview and data or None, 47 initial=initial_address, 48 **kwargs) 49 return address_form, preview 50 51 52 class ChangePasswordForm(django_forms.PasswordChangeForm): 53 def __init__(self, *args, **kwargs): 54 super().__init__(*args, **kwargs) 55 self.fields['new_password1'].user = self.user 56 self.fields['old_password'].widget.attrs['placeholder'] = '' 57 self.fields['new_password1'].widget.attrs['placeholder'] = '' 58 del self.fields['new_password2'] 59 60 61 def logout_on_password_change(request, user): 62 if (update_session_auth_hash is not None and 63 not settings.LOGOUT_ON_PASSWORD_CHANGE): 64 update_session_auth_hash(request, user) 65 66 67 class LoginForm(django_forms.AuthenticationForm, FormWithReCaptcha): 68 username = forms.EmailField( 69 label=pgettext('Form field', 'Email'), max_length=75) 70 71 def __init__(self, request=None, *args, **kwargs): 72 super().__init__(request=request, *args, **kwargs) 73 if request: 74 email = request.GET.get('email') 75 if email: 76 self.fields['username'].initial = email 77 78 79 class SignupForm(forms.ModelForm, FormWithReCaptcha): 80 password = forms.CharField( 81 widget=forms.PasswordInput, 82 label=pgettext('Password', 'Password')) 83 email = forms.EmailField( 84 label=pgettext('Email', 'Email'), 85 error_messages={ 86 'unique': pgettext_lazy( 87 'Registration error', 88 'This email has already been registered.')}) 89 90 class Meta: 91 model = User 92 fields = ('email',) 93 94 def __init__(self, *args, **kwargs): 95 super().__init__(*args, **kwargs) 96 if self._meta.model.USERNAME_FIELD in self.fields: 97 self.fields[self._meta.model.USERNAME_FIELD].widget.attrs.update( 98 {'autofocus': ''}) 99 100 def save(self, request=None, commit=True): 101 user = super().save(commit=False) 102 password = self.cleaned_data['password'] 103 user.set_password(password) 104 if commit: 105 user.save() 106 return user 107 108 109 class PasswordResetForm(django_forms.PasswordResetForm, FormWithReCaptcha): 110 """Allow resetting passwords. 111 112 This subclass overrides sending emails to use templated email. 113 """ 114 115 def get_users(self, email): 116 active_users = User.objects.filter(email__iexact=email, is_active=True) 117 return active_users 118 119 def send_mail( 120 self, subject_template_name, email_template_name, context, 121 from_email, to_email, html_email_template_name=None): 122 # Passing the user object to the Celery task throws an 123 # error "'User' is not JSON serializable". Since it's not used in our 124 # template, we remove it from the context. 125 del context['user'] 126 emails.send_password_reset_email.delay(context, to_email) 127 128 129 class NameForm(forms.ModelForm): 130 class Meta: 131 model = User 132 fields = ['first_name', 'last_name'] 133 labels = { 134 'first_name': pgettext_lazy( 135 'Customer form: Given name field', 'Given name'), 136 'last_name': pgettext_lazy( 137 'Customer form: Family name field', 'Family name')} 138 [end of saleor/account/forms.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/saleor/account/forms.py b/saleor/account/forms.py --- a/saleor/account/forms.py +++ b/saleor/account/forms.py @@ -46,6 +46,11 @@ not preview and data or None, initial=initial_address, **kwargs) + + if hasattr(address_form.fields['country_area'], 'choices'): + choices = address_form.fields['country_area'].choices + choices = [(choice[1], choice[1]) for choice in choices] + address_form.fields['country_area'].choices = choices return address_form, preview
{"golden_diff": "diff --git a/saleor/account/forms.py b/saleor/account/forms.py\n--- a/saleor/account/forms.py\n+++ b/saleor/account/forms.py\n@@ -46,6 +46,11 @@\n not preview and data or None,\n initial=initial_address,\n **kwargs)\n+\n+ if hasattr(address_form.fields['country_area'], 'choices'):\n+ choices = address_form.fields['country_area'].choices\n+ choices = [(choice[1], choice[1]) for choice in choices]\n+ address_form.fields['country_area'].choices = choices\n return address_form, preview\n", "issue": "Duplicate choices in shipping address\n### What I'm trying to achieve\r\nI'm trying to set shipping user for Russian user and there are duplicate values in \"Oblast\" selector.\r\n\r\n### Steps to reproduce the problem\r\n1. Create new shipping address\r\n2. Country --> Russia\r\n3. There are duplicated values in \"Oblast\" selector\r\n\r\n### What I expected to happen\r\nThere are no duplicated values in \"Oblast\" selector\r\n\r\n### Screenshots\r\nWhat happens now\r\n<img src=\"https://user-images.githubusercontent.com/13136992/53255369-8a239600-36d6-11e9-84a6-24a10b96a321.png\" width=\"300\">\r\n\r\n\r\nWhat I expect to see\r\n<img src=\"https://user-images.githubusercontent.com/13136992/53255400-99a2df00-36d6-11e9-8913-ecaec174487a.png\" width=\"300\">\r\n\r\n\r\n**System information**\r\nOperating system: Manjaro Linux\r\nBrowser: Google Chrome 72.0.3626.81\r\n\n", "before_files": [{"content": "from captcha.fields import ReCaptchaField\nfrom django import forms\nfrom django.conf import settings\nfrom django.contrib.auth import forms as django_forms, update_session_auth_hash\nfrom django.utils.translation import pgettext, pgettext_lazy\nfrom phonenumbers.phonenumberutil import country_code_for_region\n\nfrom ..account.models import User\nfrom . import emails\nfrom .i18n import AddressMetaForm, get_address_form_class\n\n\nclass FormWithReCaptcha(forms.BaseForm):\n def __new__(cls, *args, **kwargs):\n if settings.RECAPTCHA_PUBLIC_KEY and settings.RECAPTCHA_PRIVATE_KEY:\n # insert a Google reCaptcha field inside the form\n # note: label is empty, the reCaptcha is self-explanatory making\n # the form simpler for the user.\n cls.base_fields['_captcha'] = ReCaptchaField(label='')\n return super(FormWithReCaptcha, cls).__new__(cls)\n\n\ndef get_address_form(\n data, country_code, initial=None, instance=None, **kwargs):\n country_form = AddressMetaForm(data, initial=initial)\n preview = False\n if country_form.is_valid():\n country_code = country_form.cleaned_data['country']\n preview = country_form.cleaned_data['preview']\n\n if initial is None and country_code:\n initial = {}\n if country_code:\n initial['phone'] = '+{}'.format(country_code_for_region(country_code))\n\n address_form_class = get_address_form_class(country_code)\n\n if not preview and instance is not None:\n address_form_class = get_address_form_class(instance.country.code)\n address_form = address_form_class(data, instance=instance, **kwargs)\n else:\n initial_address = (\n initial if not preview\n else data.dict() if data is not None else data)\n address_form = address_form_class(\n not preview and data or None,\n initial=initial_address,\n **kwargs)\n return address_form, preview\n\n\nclass ChangePasswordForm(django_forms.PasswordChangeForm):\n def __init__(self, *args, **kwargs):\n super().__init__(*args, **kwargs)\n self.fields['new_password1'].user = self.user\n self.fields['old_password'].widget.attrs['placeholder'] = ''\n self.fields['new_password1'].widget.attrs['placeholder'] = ''\n del self.fields['new_password2']\n\n\ndef logout_on_password_change(request, user):\n if (update_session_auth_hash is not None and\n not settings.LOGOUT_ON_PASSWORD_CHANGE):\n update_session_auth_hash(request, user)\n\n\nclass LoginForm(django_forms.AuthenticationForm, FormWithReCaptcha):\n username = forms.EmailField(\n label=pgettext('Form field', 'Email'), max_length=75)\n\n def __init__(self, request=None, *args, **kwargs):\n super().__init__(request=request, *args, **kwargs)\n if request:\n email = request.GET.get('email')\n if email:\n self.fields['username'].initial = email\n\n\nclass SignupForm(forms.ModelForm, FormWithReCaptcha):\n password = forms.CharField(\n widget=forms.PasswordInput,\n label=pgettext('Password', 'Password'))\n email = forms.EmailField(\n label=pgettext('Email', 'Email'),\n error_messages={\n 'unique': pgettext_lazy(\n 'Registration error',\n 'This email has already been registered.')})\n\n class Meta:\n model = User\n fields = ('email',)\n\n def __init__(self, *args, **kwargs):\n super().__init__(*args, **kwargs)\n if self._meta.model.USERNAME_FIELD in self.fields:\n self.fields[self._meta.model.USERNAME_FIELD].widget.attrs.update(\n {'autofocus': ''})\n\n def save(self, request=None, commit=True):\n user = super().save(commit=False)\n password = self.cleaned_data['password']\n user.set_password(password)\n if commit:\n user.save()\n return user\n\n\nclass PasswordResetForm(django_forms.PasswordResetForm, FormWithReCaptcha):\n \"\"\"Allow resetting passwords.\n\n This subclass overrides sending emails to use templated email.\n \"\"\"\n\n def get_users(self, email):\n active_users = User.objects.filter(email__iexact=email, is_active=True)\n return active_users\n\n def send_mail(\n self, subject_template_name, email_template_name, context,\n from_email, to_email, html_email_template_name=None):\n # Passing the user object to the Celery task throws an\n # error \"'User' is not JSON serializable\". Since it's not used in our\n # template, we remove it from the context.\n del context['user']\n emails.send_password_reset_email.delay(context, to_email)\n\n\nclass NameForm(forms.ModelForm):\n class Meta:\n model = User\n fields = ['first_name', 'last_name']\n labels = {\n 'first_name': pgettext_lazy(\n 'Customer form: Given name field', 'Given name'),\n 'last_name': pgettext_lazy(\n 'Customer form: Family name field', 'Family name')}\n", "path": "saleor/account/forms.py"}]}
2,205
133
gh_patches_debug_25512
rasdani/github-patches
git_diff
openstates__openstates-scrapers-2059
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> RI failing since at least 2018-01-02 RI has been failing since 2018-01-02 Based on automated runs it appears that RI has not run successfully in 2 days (2018-01-02). ``` no pupa_settings on path, using defaults ri (scrape, import) bills: {} people: {} committees: {} 23:00:36 CRITICAL pupa: Session(s) 2018 were reported by RhodeIsland.get_session_list() but were not found in RhodeIsland.legislative_sessions or RhodeIsland.ignored_scraped_sessions. ``` Visit http://bobsled.openstates.org for more info. </issue> <code> [start of billy_metadata/ri.py] 1 import datetime 2 3 metadata = { 4 'name': 'Rhode Island', 5 'abbreviation': 'ri', 6 'legislature_name': 'Rhode Island General Assembly', 7 'legislature_url': 'http://www.rilin.state.ri.us/', 8 'capitol_timezone': 'America/New_York', 9 'chambers': { 10 'upper': {'name': 'Senate', 'title': 'Senator'}, 11 'lower': {'name': 'House', 'title': 'Representative'}, 12 }, 13 'terms': [ 14 { 15 'name': '2012', 16 'start_year': 2012, 17 'end_year': 2012, 18 'sessions': ['2012'], 19 }, 20 { 21 'name': '2013', 22 'start_year': 2013, 23 'end_year': 2013, 24 'sessions': ['2013'], 25 }, 26 27 { 28 'name': '2014', 29 'start_year': 2014, 30 'end_year': 2014, 31 'sessions': ['2014'], 32 }, 33 { 34 'name': '2015', 35 'start_year': 2015, 36 'end_year': 2015, 37 'sessions': ['2015'], 38 }, 39 { 40 'name': '2016', 41 'start_year': 2016, 42 'end_year': 2017, 43 'sessions': ['2016'], 44 }, 45 { 46 'name': '2017', 47 'start_year': 2017, 48 'end_year': 2018, 49 'sessions': ['2017'], 50 }, 51 ], 52 'session_details': { 53 '2012': { 54 'type': 'primary', 55 'start_date': datetime.date(2012, 1, 3), 56 'end_date': datetime.date(2012, 6, 13), 57 'display_name': '2012 Regular Session', 58 '_scraped_name': '2012', 59 }, 60 '2013': { 61 'type': 'primary', 62 'start_date': datetime.date(2013, 1, 1), 63 'end_date': datetime.date(2013, 7, 3), 64 'display_name': '2013 Regular Session', 65 '_scraped_name': '2013', 66 }, 67 '2014': { 68 'type': 'primary', 69 'start_date': datetime.date(2014, 1, 7), 70 'end_date': datetime.date(2014, 6, 21), 71 'display_name': '2014 Regular Session', 72 '_scraped_name': '2014', 73 }, 74 '2015': { 75 'type': 'primary', 76 'start_date': datetime.date(2015, 1, 6), 77 'end_date': datetime.date(2015, 6, 25), 78 'display_name': '2015 Regular Session', 79 '_scraped_name': '2015', 80 }, 81 '2016': { 82 'type': 'primary', 83 'start_date': datetime.date(2016, 1, 5), 84 'display_name': '2016 Regular Session', 85 '_scraped_name': '2016', 86 }, 87 '2017': { 88 'type': 'primary', 89 'display_name': '2017 Regular Session', 90 '_scraped_name': '2017', 91 }, 92 }, 93 'feature_flags': ['subjects', 'events', 'influenceexplorer'], 94 '_ignored_scraped_sessions': [ 95 '2015', 96 '2014', 97 '2013', 98 '2012', 99 '2011', 100 '2010', 101 '2009', 102 '2008', 103 '2007' 104 ], 105 } 106 [end of billy_metadata/ri.py] [start of openstates/ri/__init__.py] 1 from pupa.scrape import Jurisdiction, Organization 2 3 from openstates.utils import url_xpath 4 5 from .bills import RIBillScraper 6 from .events import RIEventScraper 7 from .people import RIPersonScraper 8 from .committees import RICommitteeScraper 9 10 11 class RhodeIsland(Jurisdiction): 12 division_id = "ocd-division/country:us/state:ri" 13 classification = "government" 14 name = "Rhode Island" 15 url = "http://www.ri.gov/" 16 scrapers = { 17 'bills': RIBillScraper, 18 'events': RIEventScraper, 19 'people': RIPersonScraper, 20 'committees': RICommitteeScraper, 21 } 22 parties = [ 23 {'name': 'Republican'}, 24 {'name': 'Democratic'} 25 ] 26 legislative_sessions = [ 27 { 28 "_scraped_name": "2012", 29 "classification": "primary", 30 "end_date": "2012-06-13", 31 "identifier": "2012", 32 "name": "2012 Regular Session", 33 "start_date": "2012-01-03" 34 }, 35 { 36 "_scraped_name": "2013", 37 "classification": "primary", 38 "end_date": "2013-07-03", 39 "identifier": "2013", 40 "name": "2013 Regular Session", 41 "start_date": "2013-01-01" 42 }, 43 { 44 "_scraped_name": "2014", 45 "classification": "primary", 46 "end_date": "2014-06-21", 47 "identifier": "2014", 48 "name": "2014 Regular Session", 49 "start_date": "2014-01-07" 50 }, 51 { 52 "_scraped_name": "2015", 53 "classification": "primary", 54 "end_date": "2015-06-25", 55 "identifier": "2015", 56 "name": "2015 Regular Session", 57 "start_date": "2015-01-06" 58 }, 59 { 60 "_scraped_name": "2016", 61 "classification": "primary", 62 "identifier": "2016", 63 "name": "2016 Regular Session", 64 "start_date": "2016-01-05" 65 }, 66 { 67 "_scraped_name": "2017", 68 "classification": "primary", 69 "identifier": "2017", 70 "name": "2017 Regular Session", 71 "start_date": "2017-01-03", 72 "end_date": "2017-06-30", 73 } 74 ] 75 ignored_scraped_sessions = [ 76 "2015", 77 "2014", 78 "2013", 79 "2012", 80 "2011", 81 "2010", 82 "2009", 83 "2008", 84 "2007" 85 ] 86 87 def get_organizations(self): 88 legislature_name = "Rhode Island General Assembly" 89 lower_chamber_name = "House of Representatives" 90 lower_seats = 75 91 lower_title = "Representative" 92 upper_chamber_name = "Senate" 93 upper_seats = 38 94 upper_title = "Senator" 95 96 legislature = Organization(name=legislature_name, 97 classification="legislature") 98 executive = Organization(name='Office of the Governor', 99 classification="executive") 100 upper = Organization(upper_chamber_name, classification='upper', 101 parent_id=legislature._id) 102 lower = Organization(lower_chamber_name, classification='lower', 103 parent_id=legislature._id) 104 105 for n in range(1, upper_seats + 1): 106 upper.add_post( 107 label=str(n), role=upper_title, 108 division_id='{}/sldu:{}'.format(self.division_id, n)) 109 for n in range(1, lower_seats + 1): 110 lower.add_post( 111 label=str(n), role=lower_title, 112 division_id='{}/sldl:{}'.format(self.division_id, n)) 113 114 yield legislature 115 yield executive 116 yield upper 117 yield lower 118 119 def get_session_list(self): 120 return url_xpath( 121 'http://status.rilin.state.ri.us/bill_history.aspx?mode=previous', 122 '//select[@name="ctl00$rilinContent$cbYear"]/option/text()') 123 [end of openstates/ri/__init__.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/billy_metadata/ri.py b/billy_metadata/ri.py --- a/billy_metadata/ri.py +++ b/billy_metadata/ri.py @@ -46,7 +46,7 @@ 'name': '2017', 'start_year': 2017, 'end_year': 2018, - 'sessions': ['2017'], + 'sessions': ['2017', '2018'], }, ], 'session_details': { @@ -89,6 +89,11 @@ 'display_name': '2017 Regular Session', '_scraped_name': '2017', }, + '2018': { + 'type': 'primary', + 'display_name': '2018 Regular Session', + '_scraped_name': '2018', + }, }, 'feature_flags': ['subjects', 'events', 'influenceexplorer'], '_ignored_scraped_sessions': [ diff --git a/openstates/ri/__init__.py b/openstates/ri/__init__.py --- a/openstates/ri/__init__.py +++ b/openstates/ri/__init__.py @@ -70,7 +70,14 @@ "name": "2017 Regular Session", "start_date": "2017-01-03", "end_date": "2017-06-30", - } + }, + { + "_scraped_name": "2018", + "classification": "primary", + "identifier": "2018", + "name": "2018 Regular Session", + "start_date": "2018-01-02", + }, ] ignored_scraped_sessions = [ "2015",
{"golden_diff": "diff --git a/billy_metadata/ri.py b/billy_metadata/ri.py\n--- a/billy_metadata/ri.py\n+++ b/billy_metadata/ri.py\n@@ -46,7 +46,7 @@\n 'name': '2017',\n 'start_year': 2017,\n 'end_year': 2018,\n- 'sessions': ['2017'],\n+ 'sessions': ['2017', '2018'],\n },\n ],\n 'session_details': {\n@@ -89,6 +89,11 @@\n 'display_name': '2017 Regular Session',\n '_scraped_name': '2017',\n },\n+ '2018': {\n+ 'type': 'primary',\n+ 'display_name': '2018 Regular Session',\n+ '_scraped_name': '2018',\n+ },\n },\n 'feature_flags': ['subjects', 'events', 'influenceexplorer'],\n '_ignored_scraped_sessions': [\ndiff --git a/openstates/ri/__init__.py b/openstates/ri/__init__.py\n--- a/openstates/ri/__init__.py\n+++ b/openstates/ri/__init__.py\n@@ -70,7 +70,14 @@\n \"name\": \"2017 Regular Session\",\n \"start_date\": \"2017-01-03\",\n \"end_date\": \"2017-06-30\",\n- }\n+ },\n+ {\n+ \"_scraped_name\": \"2018\",\n+ \"classification\": \"primary\",\n+ \"identifier\": \"2018\",\n+ \"name\": \"2018 Regular Session\",\n+ \"start_date\": \"2018-01-02\",\n+ },\n ]\n ignored_scraped_sessions = [\n \"2015\",\n", "issue": "RI failing since at least 2018-01-02\nRI has been failing since 2018-01-02\n\nBased on automated runs it appears that RI has not run successfully in 2 days (2018-01-02).\n\n\n```\n no pupa_settings on path, using defaults\nri (scrape, import)\n bills: {}\n people: {}\n committees: {}\n23:00:36 CRITICAL pupa: Session(s) 2018 were reported by RhodeIsland.get_session_list() but were not found in RhodeIsland.legislative_sessions or RhodeIsland.ignored_scraped_sessions.\n```\n\nVisit http://bobsled.openstates.org for more info.\n\n", "before_files": [{"content": "import datetime\n\nmetadata = {\n 'name': 'Rhode Island',\n 'abbreviation': 'ri',\n 'legislature_name': 'Rhode Island General Assembly',\n 'legislature_url': 'http://www.rilin.state.ri.us/',\n 'capitol_timezone': 'America/New_York',\n 'chambers': {\n 'upper': {'name': 'Senate', 'title': 'Senator'},\n 'lower': {'name': 'House', 'title': 'Representative'},\n },\n 'terms': [\n {\n 'name': '2012',\n 'start_year': 2012,\n 'end_year': 2012,\n 'sessions': ['2012'],\n },\n {\n 'name': '2013',\n 'start_year': 2013,\n 'end_year': 2013,\n 'sessions': ['2013'],\n },\n\n {\n 'name': '2014',\n 'start_year': 2014,\n 'end_year': 2014,\n 'sessions': ['2014'],\n },\n {\n 'name': '2015',\n 'start_year': 2015,\n 'end_year': 2015,\n 'sessions': ['2015'],\n },\n {\n 'name': '2016',\n 'start_year': 2016,\n 'end_year': 2017,\n 'sessions': ['2016'],\n },\n {\n 'name': '2017',\n 'start_year': 2017,\n 'end_year': 2018,\n 'sessions': ['2017'],\n },\n ],\n 'session_details': {\n '2012': {\n 'type': 'primary',\n 'start_date': datetime.date(2012, 1, 3),\n 'end_date': datetime.date(2012, 6, 13),\n 'display_name': '2012 Regular Session',\n '_scraped_name': '2012',\n },\n '2013': {\n 'type': 'primary',\n 'start_date': datetime.date(2013, 1, 1),\n 'end_date': datetime.date(2013, 7, 3),\n 'display_name': '2013 Regular Session',\n '_scraped_name': '2013',\n },\n '2014': {\n 'type': 'primary',\n 'start_date': datetime.date(2014, 1, 7),\n 'end_date': datetime.date(2014, 6, 21),\n 'display_name': '2014 Regular Session',\n '_scraped_name': '2014',\n },\n '2015': {\n 'type': 'primary',\n 'start_date': datetime.date(2015, 1, 6),\n 'end_date': datetime.date(2015, 6, 25),\n 'display_name': '2015 Regular Session',\n '_scraped_name': '2015',\n },\n '2016': {\n 'type': 'primary',\n 'start_date': datetime.date(2016, 1, 5),\n 'display_name': '2016 Regular Session',\n '_scraped_name': '2016',\n },\n '2017': {\n 'type': 'primary',\n 'display_name': '2017 Regular Session',\n '_scraped_name': '2017',\n },\n },\n 'feature_flags': ['subjects', 'events', 'influenceexplorer'],\n '_ignored_scraped_sessions': [\n '2015',\n '2014',\n '2013',\n '2012',\n '2011',\n '2010',\n '2009',\n '2008',\n '2007'\n ],\n}\n", "path": "billy_metadata/ri.py"}, {"content": "from pupa.scrape import Jurisdiction, Organization\n\nfrom openstates.utils import url_xpath\n\nfrom .bills import RIBillScraper\nfrom .events import RIEventScraper\nfrom .people import RIPersonScraper\nfrom .committees import RICommitteeScraper\n\n\nclass RhodeIsland(Jurisdiction):\n division_id = \"ocd-division/country:us/state:ri\"\n classification = \"government\"\n name = \"Rhode Island\"\n url = \"http://www.ri.gov/\"\n scrapers = {\n 'bills': RIBillScraper,\n 'events': RIEventScraper,\n 'people': RIPersonScraper,\n 'committees': RICommitteeScraper,\n }\n parties = [\n {'name': 'Republican'},\n {'name': 'Democratic'}\n ]\n legislative_sessions = [\n {\n \"_scraped_name\": \"2012\",\n \"classification\": \"primary\",\n \"end_date\": \"2012-06-13\",\n \"identifier\": \"2012\",\n \"name\": \"2012 Regular Session\",\n \"start_date\": \"2012-01-03\"\n },\n {\n \"_scraped_name\": \"2013\",\n \"classification\": \"primary\",\n \"end_date\": \"2013-07-03\",\n \"identifier\": \"2013\",\n \"name\": \"2013 Regular Session\",\n \"start_date\": \"2013-01-01\"\n },\n {\n \"_scraped_name\": \"2014\",\n \"classification\": \"primary\",\n \"end_date\": \"2014-06-21\",\n \"identifier\": \"2014\",\n \"name\": \"2014 Regular Session\",\n \"start_date\": \"2014-01-07\"\n },\n {\n \"_scraped_name\": \"2015\",\n \"classification\": \"primary\",\n \"end_date\": \"2015-06-25\",\n \"identifier\": \"2015\",\n \"name\": \"2015 Regular Session\",\n \"start_date\": \"2015-01-06\"\n },\n {\n \"_scraped_name\": \"2016\",\n \"classification\": \"primary\",\n \"identifier\": \"2016\",\n \"name\": \"2016 Regular Session\",\n \"start_date\": \"2016-01-05\"\n },\n {\n \"_scraped_name\": \"2017\",\n \"classification\": \"primary\",\n \"identifier\": \"2017\",\n \"name\": \"2017 Regular Session\",\n \"start_date\": \"2017-01-03\",\n \"end_date\": \"2017-06-30\",\n }\n ]\n ignored_scraped_sessions = [\n \"2015\",\n \"2014\",\n \"2013\",\n \"2012\",\n \"2011\",\n \"2010\",\n \"2009\",\n \"2008\",\n \"2007\"\n ]\n\n def get_organizations(self):\n legislature_name = \"Rhode Island General Assembly\"\n lower_chamber_name = \"House of Representatives\"\n lower_seats = 75\n lower_title = \"Representative\"\n upper_chamber_name = \"Senate\"\n upper_seats = 38\n upper_title = \"Senator\"\n\n legislature = Organization(name=legislature_name,\n classification=\"legislature\")\n executive = Organization(name='Office of the Governor',\n classification=\"executive\")\n upper = Organization(upper_chamber_name, classification='upper',\n parent_id=legislature._id)\n lower = Organization(lower_chamber_name, classification='lower',\n parent_id=legislature._id)\n\n for n in range(1, upper_seats + 1):\n upper.add_post(\n label=str(n), role=upper_title,\n division_id='{}/sldu:{}'.format(self.division_id, n))\n for n in range(1, lower_seats + 1):\n lower.add_post(\n label=str(n), role=lower_title,\n division_id='{}/sldl:{}'.format(self.division_id, n))\n\n yield legislature\n yield executive\n yield upper\n yield lower\n\n def get_session_list(self):\n return url_xpath(\n 'http://status.rilin.state.ri.us/bill_history.aspx?mode=previous',\n '//select[@name=\"ctl00$rilinContent$cbYear\"]/option/text()')\n", "path": "openstates/ri/__init__.py"}]}
3,168
431
gh_patches_debug_15196
rasdani/github-patches
git_diff
RedHatInsights__insights-core-3344
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> Pinned jinja2 version is no longer supported and conflicts with the latest MarkupSafe **Current behavior** insights-core has pinned the jinja2 version to <=2.11.3 in #3057. This version is [no longer supported](https://github.com/pallets/jinja/issues/1585#issuecomment-1043892628). **Expected behavior** insights-core can be used with a maintained jinja2 version in applications that depend both on insights-core and jinja2. **Impact** jinja2-2.11.3 fails to import with the latest version of MarkupSafe, 2.1.0: ``` ... .tox/unit-py38/lib/python3.8/site-packages/jinja2/filters.py:13: in <module> from markupsafe import soft_unicode E ImportError: cannot import name 'soft_unicode' from 'markupsafe' (/builds/ccx/ccx-rules-ocp/.tox/unit-py38/lib64/python3.8/site-packages/markupsafe/__init__.py ``` All projects that depend on insights-core have to pin unsupported versions of MarkupSafe now. I would argue that it is important to keep up-to-date for a package that is used for secure handling of user input. </issue> <code> [start of setup.py] 1 import os 2 from setuptools import setup, find_packages 3 4 __here__ = os.path.dirname(os.path.abspath(__file__)) 5 6 package_info = dict.fromkeys(["RELEASE", "COMMIT", "VERSION", "NAME"]) 7 8 for name in package_info: 9 with open(os.path.join(__here__, "insights", name)) as f: 10 package_info[name] = f.read().strip() 11 12 entry_points = { 13 'console_scripts': [ 14 'insights-collect = insights.collect:main', 15 'insights-run = insights:main', 16 'insights = insights.command_parser:main', 17 'insights-cat = insights.tools.cat:main', 18 'insights-dupkeycheck = insights.tools.dupkeycheck:main', 19 'insights-inspect = insights.tools.insights_inspect:main', 20 'insights-info = insights.tools.query:main', 21 'insights-ocpshell= insights.ocpshell:main', 22 'mangle = insights.util.mangle:main' 23 ] 24 } 25 26 runtime = set([ 27 'six', 28 'requests', 29 'redis', 30 'cachecontrol', 31 'cachecontrol[redis]', 32 'cachecontrol[filecache]', 33 'defusedxml', 34 'lockfile', 35 'jinja2<=2.11.3', 36 'pyyaml>=3.10,<=3.13; python_version < "2.7"', 37 'pyyaml; python_version >= "2.7"', 38 ]) 39 40 41 def maybe_require(pkg): 42 try: 43 __import__(pkg) 44 except ImportError: 45 runtime.add(pkg) 46 47 48 maybe_require("importlib") 49 maybe_require("argparse") 50 51 52 client = set([ 53 'requests', 54 'python-gnupg==0.4.6', 55 'oyaml' 56 ]) 57 58 develop = set([ 59 'wheel', 60 ]) 61 62 docs = set([ 63 'docutils', 64 'Sphinx', 65 'nbsphinx', 66 'sphinx_rtd_theme', 67 'ipython', 68 'colorama', 69 'jinja2<=2.11.3', 70 'Pygments', 71 'jedi<0.18.0', # Open issue with jedi 0.18.0 and iPython <= 7.19 72 # https://github.com/davidhalter/jedi/issues/1714 73 ]) 74 75 # python 2.6 requires setuptools~=36.8.0 to support this syntax 76 testing = set([ 77 'coverage==4.3.4; python_version < "2.7"', 78 'coverage; python_version >= "2.7"', 79 'pytest==3.0.6; python_version < "2.7"', 80 'pytest~=4.6.0; python_version == "2.7"', 81 'pytest; python_version >= "3"', 82 'pytest-cov==2.4.0; python_version < "2.7"', 83 'pytest-cov; python_version >= "2.7"', 84 'mock==2.0.0', 85 ]) 86 87 cluster = set([ 88 'ansible', 89 'pandas', 90 'colorama', 91 ]) 92 93 openshift = set([ 94 'openshift' 95 ]) 96 97 linting = set([ 98 'flake8==2.6.2; python_version < "2.7"', 99 'flake8; python_version >= "2.7"' 100 ]) 101 102 optional = set([ 103 'python-cjson', 104 'python-logstash', 105 'python-statsd', 106 'watchdog', 107 ]) 108 109 if __name__ == "__main__": 110 # allows for runtime modification of rpm name 111 name = os.environ.get("INSIGHTS_CORE_NAME", package_info["NAME"]) 112 113 setup( 114 name=name, 115 version=package_info["VERSION"], 116 description="Insights Core is a data collection and analysis framework", 117 long_description=open("README.rst").read(), 118 url="https://github.com/redhatinsights/insights-core", 119 author="Red Hat, Inc.", 120 author_email="[email protected]", 121 packages=find_packages(), 122 install_requires=list(runtime), 123 package_data={'': ['LICENSE']}, 124 license='Apache 2.0', 125 extras_require={ 126 'develop': list(runtime | develop | client | docs | linting | testing | cluster), 127 'develop26': list(runtime | develop | client | linting | testing | cluster), 128 'client': list(runtime | client), 129 'client-develop': list(runtime | develop | client | linting | testing), 130 'cluster': list(runtime | cluster), 131 'openshift': list(runtime | openshift), 132 'optional': list(optional), 133 'docs': list(docs), 134 'linting': list(linting | client), 135 'testing': list(testing | client) 136 }, 137 classifiers=[ 138 'Development Status :: 5 - Production/Stable', 139 'Intended Audience :: Developers', 140 'Natural Language :: English', 141 'License :: OSI Approved :: Apache Software License', 142 'Programming Language :: Python', 143 'Programming Language :: Python :: 2.6', 144 'Programming Language :: Python :: 2.7', 145 'Programming Language :: Python :: 3.3', 146 'Programming Language :: Python :: 3.4', 147 'Programming Language :: Python :: 3.5', 148 'Programming Language :: Python :: 3.6' 149 ], 150 entry_points=entry_points, 151 include_package_data=True 152 ) 153 [end of setup.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/setup.py b/setup.py --- a/setup.py +++ b/setup.py @@ -32,7 +32,8 @@ 'cachecontrol[filecache]', 'defusedxml', 'lockfile', - 'jinja2<=2.11.3', + 'jinja2<=2.11.3; python_version <= "2.7"', + 'jinja2; python_version > "2.7"', 'pyyaml>=3.10,<=3.13; python_version < "2.7"', 'pyyaml; python_version >= "2.7"', ]) @@ -66,7 +67,6 @@ 'sphinx_rtd_theme', 'ipython', 'colorama', - 'jinja2<=2.11.3', 'Pygments', 'jedi<0.18.0', # Open issue with jedi 0.18.0 and iPython <= 7.19 # https://github.com/davidhalter/jedi/issues/1714
{"golden_diff": "diff --git a/setup.py b/setup.py\n--- a/setup.py\n+++ b/setup.py\n@@ -32,7 +32,8 @@\n 'cachecontrol[filecache]',\n 'defusedxml',\n 'lockfile',\n- 'jinja2<=2.11.3',\n+ 'jinja2<=2.11.3; python_version <= \"2.7\"',\n+ 'jinja2; python_version > \"2.7\"',\n 'pyyaml>=3.10,<=3.13; python_version < \"2.7\"',\n 'pyyaml; python_version >= \"2.7\"',\n ])\n@@ -66,7 +67,6 @@\n 'sphinx_rtd_theme',\n 'ipython',\n 'colorama',\n- 'jinja2<=2.11.3',\n 'Pygments',\n 'jedi<0.18.0', # Open issue with jedi 0.18.0 and iPython <= 7.19\n # https://github.com/davidhalter/jedi/issues/1714\n", "issue": "Pinned jinja2 version is no longer supported and conflicts with the latest MarkupSafe\n**Current behavior**\r\n\r\ninsights-core has pinned the jinja2 version to <=2.11.3 in #3057. This version is [no longer supported](https://github.com/pallets/jinja/issues/1585#issuecomment-1043892628).\r\n\r\n**Expected behavior**\r\n\r\ninsights-core can be used with a maintained jinja2 version in applications that depend both on insights-core and jinja2.\r\n\r\n**Impact**\r\n\r\njinja2-2.11.3 fails to import with the latest version of MarkupSafe, 2.1.0:\r\n\r\n```\r\n...\r\n.tox/unit-py38/lib/python3.8/site-packages/jinja2/filters.py:13: in <module>\r\n from markupsafe import soft_unicode\r\nE ImportError: cannot import name 'soft_unicode' from 'markupsafe' (/builds/ccx/ccx-rules-ocp/.tox/unit-py38/lib64/python3.8/site-packages/markupsafe/__init__.py\r\n```\r\n\r\nAll projects that depend on insights-core have to pin unsupported versions of MarkupSafe now. I would argue that it is important to keep up-to-date for a package that is used for secure handling of user input.\n", "before_files": [{"content": "import os\nfrom setuptools import setup, find_packages\n\n__here__ = os.path.dirname(os.path.abspath(__file__))\n\npackage_info = dict.fromkeys([\"RELEASE\", \"COMMIT\", \"VERSION\", \"NAME\"])\n\nfor name in package_info:\n with open(os.path.join(__here__, \"insights\", name)) as f:\n package_info[name] = f.read().strip()\n\nentry_points = {\n 'console_scripts': [\n 'insights-collect = insights.collect:main',\n 'insights-run = insights:main',\n 'insights = insights.command_parser:main',\n 'insights-cat = insights.tools.cat:main',\n 'insights-dupkeycheck = insights.tools.dupkeycheck:main',\n 'insights-inspect = insights.tools.insights_inspect:main',\n 'insights-info = insights.tools.query:main',\n 'insights-ocpshell= insights.ocpshell:main',\n 'mangle = insights.util.mangle:main'\n ]\n}\n\nruntime = set([\n 'six',\n 'requests',\n 'redis',\n 'cachecontrol',\n 'cachecontrol[redis]',\n 'cachecontrol[filecache]',\n 'defusedxml',\n 'lockfile',\n 'jinja2<=2.11.3',\n 'pyyaml>=3.10,<=3.13; python_version < \"2.7\"',\n 'pyyaml; python_version >= \"2.7\"',\n])\n\n\ndef maybe_require(pkg):\n try:\n __import__(pkg)\n except ImportError:\n runtime.add(pkg)\n\n\nmaybe_require(\"importlib\")\nmaybe_require(\"argparse\")\n\n\nclient = set([\n 'requests',\n 'python-gnupg==0.4.6',\n 'oyaml'\n])\n\ndevelop = set([\n 'wheel',\n])\n\ndocs = set([\n 'docutils',\n 'Sphinx',\n 'nbsphinx',\n 'sphinx_rtd_theme',\n 'ipython',\n 'colorama',\n 'jinja2<=2.11.3',\n 'Pygments',\n 'jedi<0.18.0', # Open issue with jedi 0.18.0 and iPython <= 7.19\n # https://github.com/davidhalter/jedi/issues/1714\n])\n\n# python 2.6 requires setuptools~=36.8.0 to support this syntax\ntesting = set([\n 'coverage==4.3.4; python_version < \"2.7\"',\n 'coverage; python_version >= \"2.7\"',\n 'pytest==3.0.6; python_version < \"2.7\"',\n 'pytest~=4.6.0; python_version == \"2.7\"',\n 'pytest; python_version >= \"3\"',\n 'pytest-cov==2.4.0; python_version < \"2.7\"',\n 'pytest-cov; python_version >= \"2.7\"',\n 'mock==2.0.0',\n])\n\ncluster = set([\n 'ansible',\n 'pandas',\n 'colorama',\n])\n\nopenshift = set([\n 'openshift'\n])\n\nlinting = set([\n 'flake8==2.6.2; python_version < \"2.7\"',\n 'flake8; python_version >= \"2.7\"'\n])\n\noptional = set([\n 'python-cjson',\n 'python-logstash',\n 'python-statsd',\n 'watchdog',\n])\n\nif __name__ == \"__main__\":\n # allows for runtime modification of rpm name\n name = os.environ.get(\"INSIGHTS_CORE_NAME\", package_info[\"NAME\"])\n\n setup(\n name=name,\n version=package_info[\"VERSION\"],\n description=\"Insights Core is a data collection and analysis framework\",\n long_description=open(\"README.rst\").read(),\n url=\"https://github.com/redhatinsights/insights-core\",\n author=\"Red Hat, Inc.\",\n author_email=\"[email protected]\",\n packages=find_packages(),\n install_requires=list(runtime),\n package_data={'': ['LICENSE']},\n license='Apache 2.0',\n extras_require={\n 'develop': list(runtime | develop | client | docs | linting | testing | cluster),\n 'develop26': list(runtime | develop | client | linting | testing | cluster),\n 'client': list(runtime | client),\n 'client-develop': list(runtime | develop | client | linting | testing),\n 'cluster': list(runtime | cluster),\n 'openshift': list(runtime | openshift),\n 'optional': list(optional),\n 'docs': list(docs),\n 'linting': list(linting | client),\n 'testing': list(testing | client)\n },\n classifiers=[\n 'Development Status :: 5 - Production/Stable',\n 'Intended Audience :: Developers',\n 'Natural Language :: English',\n 'License :: OSI Approved :: Apache Software License',\n 'Programming Language :: Python',\n 'Programming Language :: Python :: 2.6',\n 'Programming Language :: Python :: 2.7',\n 'Programming Language :: Python :: 3.3',\n 'Programming Language :: Python :: 3.4',\n 'Programming Language :: Python :: 3.5',\n 'Programming Language :: Python :: 3.6'\n ],\n entry_points=entry_points,\n include_package_data=True\n )\n", "path": "setup.py"}]}
2,323
243
gh_patches_debug_32097
rasdani/github-patches
git_diff
dotkom__onlineweb4-1208
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> Posters issues - [ ] Does not send mail [src](https://github.com/dotKom/onlineweb4/blob/f4c45b9d07bd2c20533dec840a3bdcb3b8d50dd3/apps/posters/dashboard/views.py#L106) - [x] Group is set to "Komiteer" [src](https://github.com/dotKom/onlineweb4/blob/f4c45b9d07bd2c20533dec840a3bdcb3b8d50dd3/apps/posters/dashboard/views.py#L93) </issue> <code> [start of apps/posters/dashboard/views.py] 1 # -*- encoding: utf-8 -*- 2 3 import json 4 5 from django.conf import settings 6 from django.contrib import messages 7 from django.contrib.auth.decorators import login_required 8 from django.contrib.auth.models import Group 9 from django.core.mail import EmailMessage 10 from django.core.exceptions import PermissionDenied, ObjectDoesNotExist 11 from django.core.urlresolvers import reverse 12 from django.forms.models import model_to_dict 13 from django.shortcuts import render, redirect, get_object_or_404, HttpResponseRedirect, HttpResponse 14 from django.utils import timezone 15 from django.utils.translation import ugettext as _ 16 from django.views.decorators.csrf import ensure_csrf_cookie 17 18 from guardian.decorators import permission_required 19 from guardian.models import UserObjectPermission, GroupObjectPermission 20 # from guardian.core import ObjectPermissionChecker 21 import guardian 22 23 from datetime import datetime, timedelta 24 25 from apps.authentication.models import OnlineUser as User 26 from apps.dashboard.tools import has_access, get_base_context 27 from apps.posters.models import Poster 28 from apps.posters.forms import AddForm, AddPosterForm, AddBongForm, AddOtherForm, EditPosterForm 29 # from apps.dashboard.posters.models import PosterForm 30 from apps.companyprofile.models import Company 31 from apps.posters.models import Poster, OrderMixin 32 from apps.posters.permissions import has_view_perms, has_view_all_perms 33 34 35 @ensure_csrf_cookie 36 @login_required 37 @permission_required('posters.overview_poster_order', return_403=True) 38 def index(request): 39 if request.is_ajax(): 40 do_ajax_shit = True 41 42 # The group with members who should populate the dropdownlist 43 group = Group.objects.get(name='proKom') 44 users_to_populate = group.user_set.all() 45 46 context = get_base_context(request) 47 48 # View to show if user not in committee, but wanting to see own orders 49 if not has_view_all_perms(request.user): 50 context['your_orders'] = [x for x in Poster.objects.filter(ordered_by=request.user) if request.user.has_perm('view_poster_order', x)] 51 return render(request, 'posters/dashboard/index.html', context) 52 53 orders = Poster.objects.all() 54 55 context['new_orders'] = orders.filter(assigned_to=None) 56 context['active_orders'] = orders.filter(finished=False).exclude(assigned_to=None) 57 context['hanging_orders'] = orders.filter(finished=True, 58 display_to__lte=datetime.now()+timedelta(days=3)) 59 60 context['workers'] = User.objects.filter(groups=Group.objects.get(name='proKom')) 61 62 return render(request, 'posters/dashboard/index.html', context) 63 64 65 @ensure_csrf_cookie 66 @login_required 67 @permission_required('posters.add_poster_order', return_403=True) 68 def add(request, order_type=0): 69 order_type = int(order_type) 70 context = get_base_context(request) 71 type_names = ("Plakat", "Bong", "Generell ") 72 type_name = type_names[order_type-1] 73 74 poster = Poster() 75 form = None 76 77 if request.method == 'POST': 78 if order_type == 1: 79 form = AddPosterForm(data=request.POST, instance=poster) 80 elif order_type == 2: 81 form = AddBongForm(data=request.POST, instance=poster) 82 elif order_type == 3: 83 # poster = GeneralOrder() 84 form = AddOtherForm(data=request.POST, instance=poster) 85 86 if form.is_valid(): 87 poster = form.save(commit=False) 88 if request.POST.get('company'): 89 poster.company = Company.objects.get(pk=request.POST.get('company')) 90 poster.ordered_by = request.user 91 # Should look for a more kosher solution 92 poster.ordered_committee = request.user.groups.exclude(name='Komiteer').filter(name__contains="Kom")[0] 93 poster.order_type = order_type 94 95 poster.save() 96 97 # for b in poster.__class__.__bases__: 98 # poster_mixin = OrderMixin.objects.get(id=poster.id) 99 # Let this user have permissions to show this order 100 UserObjectPermission.objects.assign_perm('view_poster_order', obj=poster, user=request.user) 101 GroupObjectPermission.objects.assign_perm('view_poster_order', obj=poster, group=Group.objects.get(name='proKom')) 102 103 title = unicode(poster) 104 105 # The great sending of emails 106 subject = '[ProKom] Ny bestilling for %s' % title 107 email_message = '%(message)s%(signature)s' % { 108 'message': _(''' 109 Det har blitt registrert en ny %(order_type)sbestilling pa Online sine nettsider. Dette er bestilling nummer %(id)s. 110 \n 111 Antall og type: %(num)s * %(order_type)s\n 112 Arrangement: %(event_name)s\n 113 Bestilt av: %(ordered_by)s i %(ordered_by_committee)s\n 114 Bestilt dato: %(ordered_date)s\n 115 \n 116 For mer informasjon, sjekk ut bestillingen her: %(absolute_url)s 117 ''' 118 % { 119 'site': '', 120 'order_type': type_name.lower(), 121 'num': poster.amount, 122 'ordered_by': poster.ordered_by, 123 'ordered_by_committee': poster.ordered_committee, 124 'id': poster.id, 125 'event_name': title, 126 'ordered_date': poster.ordered_date, 127 'absolute_url': request.build_absolute_uri(poster.get_dashboard_url()) 128 } 129 ), 130 'signature': _('\n\nVennlig hilsen Linjeforeningen Online') 131 } 132 from_email = settings.EMAIL_PROKOM 133 to_emails = [settings.EMAIL_PROKOM, request.user.get_email()] 134 135 try: 136 email_sent = EmailMessage(unicode(subject), unicode(message), from_email, to_emails, []).send() 137 except: 138 email_sent = False 139 140 if email_sent: 141 messages.success(request, 'Opprettet bestilling') 142 else: 143 messages.error(request, 'Klarte ikke å sende epost, men bestillingen din ble fortsatt opprettet') 144 145 return redirect(poster.get_absolute_url()) 146 else: 147 context['form'] = form 148 return render(request, 'posters/dashboard/add.html', context) 149 150 context["order_type_name"] = type_name 151 context['order_type'] = order_type 152 context['can_edit'] = True # request.user.has_perm('posters.view_poster') 153 154 if order_type == 1: 155 form = AddPosterForm() 156 elif order_type == 2: 157 form = AddBongForm() 158 elif order_type == 3: 159 form = AddOtherForm() 160 161 forms = (AddPosterForm(), AddBongForm(), AddOtherForm()) 162 163 context['form'] = forms[order_type-1] 164 165 return render(request, 'posters/dashboard/add.html', context) 166 167 168 @ensure_csrf_cookie 169 @login_required 170 def edit(request, order_id=None): 171 context = get_base_context(request) 172 context['add_poster_form'] = EditPosterForm() 173 174 if order_id: 175 poster = get_object_or_404(Poster, pk=order_id) 176 177 if request.user != poster.ordered_by and 'proKom' not in request.user.groups: 178 raise PermissionDenied 179 180 if request.POST: 181 form = AddForm(request.POST, instance=poster) 182 if form.is_valid(): 183 form.save() 184 # redirect_url = redirect(poster.get_absolute_url()) 185 return HttpResponseRedirect("../detail/"+str(poster.id)) 186 187 else: 188 context["form"] = AddForm(instance=poster) 189 190 return render(request, 'posters/dashboard/add.html', context) 191 192 193 @ensure_csrf_cookie 194 @login_required 195 @permission_required('view_poster_order', (Poster, 'pk', 'order_id'), return_403=True) 196 def detail(request, order_id=None): 197 if request.is_ajax(): 198 do_ajax_shit = True 199 200 if not order_id: 201 return HttpResponse(status=400) 202 203 context = get_base_context(request) 204 poster = get_object_or_404(Poster, pk=order_id) 205 context['poster'] = poster 206 207 if not has_view_perms(request.user, poster): 208 raise PermissionDenied 209 210 order_type = poster.order_type 211 type_names = ("Plakat", "Bong", "Generell ") 212 type_name = type_names[order_type-1] 213 context["order_type_name"] = type_name 214 215 if request.method == 'POST': 216 poster_status = request.POST.get('completed') 217 if poster_status == 'true' or poster_status == 'false': 218 poster.toggle_finished() 219 220 return render(request, 'posters/dashboard/details.html', context) 221 222 223 # Ajax 224 225 226 # @ensure_csrf_cookie 227 @login_required 228 def assign_person(request): 229 if request.is_ajax(): 230 if request.method == 'POST': 231 order_id = request.POST.get('order_id') 232 orders = Poster.objects.filter(pk=order_id) 233 assign_to_id = request.POST.get('assign_to_id') 234 assign_to = User.objects.get(pk=assign_to_id) 235 236 if orders.count() == 0: 237 response_text = json.dumps({'message': _( 238 u"""Kan ikke finne en ordre med denne IDen (%s). 239 Om feilen vedvarer etter en refresh, kontakt [email protected].""") % order_id}) 240 return HttpResponse(status=412, content=response_text) 241 242 order = orders[0] 243 244 if order.finished or order.assigned_to is not None: 245 response_text = json.dumps({'message': _(u"Denne ordren er allerede behandlet.")}) 246 return HttpResponse(status=412, content=response_text) 247 248 order.assigned_to = assign_to 249 order.save() 250 251 return HttpResponse(status=200) 252 [end of apps/posters/dashboard/views.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/apps/posters/dashboard/views.py b/apps/posters/dashboard/views.py --- a/apps/posters/dashboard/views.py +++ b/apps/posters/dashboard/views.py @@ -106,14 +106,14 @@ subject = '[ProKom] Ny bestilling for %s' % title email_message = '%(message)s%(signature)s' % { 'message': _(''' - Det har blitt registrert en ny %(order_type)sbestilling pa Online sine nettsider. Dette er bestilling nummer %(id)s. - \n - Antall og type: %(num)s * %(order_type)s\n - Arrangement: %(event_name)s\n - Bestilt av: %(ordered_by)s i %(ordered_by_committee)s\n - Bestilt dato: %(ordered_date)s\n - \n - For mer informasjon, sjekk ut bestillingen her: %(absolute_url)s +Det har blitt registrert en ny %(order_type)sbestilling pa Online sine nettsider. Dette er bestilling nummer %(id)s. +\n +Antall og type: %(num)s * %(order_type)s\n +Arrangement: %(event_name)s\n +Bestilt av: %(ordered_by)s i %(ordered_by_committee)s\n +Bestilt dato: %(ordered_date)s\n +\n +For mer informasjon, sjekk ut bestillingen her: %(absolute_url)s ''' % { 'site': '', @@ -130,13 +130,9 @@ 'signature': _('\n\nVennlig hilsen Linjeforeningen Online') } from_email = settings.EMAIL_PROKOM - to_emails = [settings.EMAIL_PROKOM, request.user.get_email()] - - try: - email_sent = EmailMessage(unicode(subject), unicode(message), from_email, to_emails, []).send() - except: - email_sent = False + to_emails = [settings.EMAIL_PROKOM, request.user.get_email().email] + email_sent = EmailMessage(unicode(subject), unicode(email_message), from_email, to_emails, []).send() if email_sent: messages.success(request, 'Opprettet bestilling') else:
{"golden_diff": "diff --git a/apps/posters/dashboard/views.py b/apps/posters/dashboard/views.py\n--- a/apps/posters/dashboard/views.py\n+++ b/apps/posters/dashboard/views.py\n@@ -106,14 +106,14 @@\n subject = '[ProKom] Ny bestilling for %s' % title\n email_message = '%(message)s%(signature)s' % {\n 'message': _('''\n- Det har blitt registrert en ny %(order_type)sbestilling pa Online sine nettsider. Dette er bestilling nummer %(id)s.\n- \\n\n- Antall og type: %(num)s * %(order_type)s\\n\n- Arrangement: %(event_name)s\\n\n- Bestilt av: %(ordered_by)s i %(ordered_by_committee)s\\n\n- Bestilt dato: %(ordered_date)s\\n\n- \\n\n- For mer informasjon, sjekk ut bestillingen her: %(absolute_url)s\n+Det har blitt registrert en ny %(order_type)sbestilling pa Online sine nettsider. Dette er bestilling nummer %(id)s.\n+\\n\n+Antall og type: %(num)s * %(order_type)s\\n\n+Arrangement: %(event_name)s\\n\n+Bestilt av: %(ordered_by)s i %(ordered_by_committee)s\\n\n+Bestilt dato: %(ordered_date)s\\n\n+\\n\n+For mer informasjon, sjekk ut bestillingen her: %(absolute_url)s\n '''\n % {\n 'site': '',\n@@ -130,13 +130,9 @@\n 'signature': _('\\n\\nVennlig hilsen Linjeforeningen Online')\n }\n from_email = settings.EMAIL_PROKOM\n- to_emails = [settings.EMAIL_PROKOM, request.user.get_email()]\n-\n- try:\n- email_sent = EmailMessage(unicode(subject), unicode(message), from_email, to_emails, []).send()\n- except:\n- email_sent = False\n+ to_emails = [settings.EMAIL_PROKOM, request.user.get_email().email]\n \n+ email_sent = EmailMessage(unicode(subject), unicode(email_message), from_email, to_emails, []).send()\n if email_sent:\n messages.success(request, 'Opprettet bestilling')\n else:\n", "issue": "Posters issues\n- [ ] Does not send mail [src](https://github.com/dotKom/onlineweb4/blob/f4c45b9d07bd2c20533dec840a3bdcb3b8d50dd3/apps/posters/dashboard/views.py#L106)\n- [x] Group is set to \"Komiteer\" [src](https://github.com/dotKom/onlineweb4/blob/f4c45b9d07bd2c20533dec840a3bdcb3b8d50dd3/apps/posters/dashboard/views.py#L93)\n\n", "before_files": [{"content": "# -*- encoding: utf-8 -*-\n\nimport json\n\nfrom django.conf import settings\nfrom django.contrib import messages\nfrom django.contrib.auth.decorators import login_required\nfrom django.contrib.auth.models import Group\nfrom django.core.mail import EmailMessage\nfrom django.core.exceptions import PermissionDenied, ObjectDoesNotExist\nfrom django.core.urlresolvers import reverse\nfrom django.forms.models import model_to_dict\nfrom django.shortcuts import render, redirect, get_object_or_404, HttpResponseRedirect, HttpResponse\nfrom django.utils import timezone\nfrom django.utils.translation import ugettext as _\nfrom django.views.decorators.csrf import ensure_csrf_cookie\n\nfrom guardian.decorators import permission_required\nfrom guardian.models import UserObjectPermission, GroupObjectPermission\n# from guardian.core import ObjectPermissionChecker\nimport guardian\n\nfrom datetime import datetime, timedelta\n\nfrom apps.authentication.models import OnlineUser as User\nfrom apps.dashboard.tools import has_access, get_base_context\nfrom apps.posters.models import Poster\nfrom apps.posters.forms import AddForm, AddPosterForm, AddBongForm, AddOtherForm, EditPosterForm\n# from apps.dashboard.posters.models import PosterForm\nfrom apps.companyprofile.models import Company\nfrom apps.posters.models import Poster, OrderMixin\nfrom apps.posters.permissions import has_view_perms, has_view_all_perms\n\n\n@ensure_csrf_cookie\n@login_required\n@permission_required('posters.overview_poster_order', return_403=True)\ndef index(request):\n if request.is_ajax():\n do_ajax_shit = True\n\n # The group with members who should populate the dropdownlist\n group = Group.objects.get(name='proKom')\n users_to_populate = group.user_set.all()\n\n context = get_base_context(request)\n\n # View to show if user not in committee, but wanting to see own orders\n if not has_view_all_perms(request.user):\n context['your_orders'] = [x for x in Poster.objects.filter(ordered_by=request.user) if request.user.has_perm('view_poster_order', x)]\n return render(request, 'posters/dashboard/index.html', context)\n\n orders = Poster.objects.all()\n\n context['new_orders'] = orders.filter(assigned_to=None)\n context['active_orders'] = orders.filter(finished=False).exclude(assigned_to=None)\n context['hanging_orders'] = orders.filter(finished=True,\n display_to__lte=datetime.now()+timedelta(days=3))\n\n context['workers'] = User.objects.filter(groups=Group.objects.get(name='proKom'))\n\n return render(request, 'posters/dashboard/index.html', context)\n\n\n@ensure_csrf_cookie\n@login_required\n@permission_required('posters.add_poster_order', return_403=True)\ndef add(request, order_type=0):\n order_type = int(order_type)\n context = get_base_context(request)\n type_names = (\"Plakat\", \"Bong\", \"Generell \")\n type_name = type_names[order_type-1]\n\n poster = Poster()\n form = None\n\n if request.method == 'POST':\n if order_type == 1:\n form = AddPosterForm(data=request.POST, instance=poster)\n elif order_type == 2:\n form = AddBongForm(data=request.POST, instance=poster)\n elif order_type == 3:\n # poster = GeneralOrder()\n form = AddOtherForm(data=request.POST, instance=poster)\n\n if form.is_valid():\n poster = form.save(commit=False)\n if request.POST.get('company'):\n poster.company = Company.objects.get(pk=request.POST.get('company'))\n poster.ordered_by = request.user\n # Should look for a more kosher solution\n poster.ordered_committee = request.user.groups.exclude(name='Komiteer').filter(name__contains=\"Kom\")[0]\n poster.order_type = order_type\n\n poster.save()\n\n # for b in poster.__class__.__bases__:\n # poster_mixin = OrderMixin.objects.get(id=poster.id)\n # Let this user have permissions to show this order\n UserObjectPermission.objects.assign_perm('view_poster_order', obj=poster, user=request.user)\n GroupObjectPermission.objects.assign_perm('view_poster_order', obj=poster, group=Group.objects.get(name='proKom'))\n\n title = unicode(poster)\n\n # The great sending of emails\n subject = '[ProKom] Ny bestilling for %s' % title\n email_message = '%(message)s%(signature)s' % {\n 'message': _('''\n Det har blitt registrert en ny %(order_type)sbestilling pa Online sine nettsider. Dette er bestilling nummer %(id)s.\n \\n\n Antall og type: %(num)s * %(order_type)s\\n\n Arrangement: %(event_name)s\\n\n Bestilt av: %(ordered_by)s i %(ordered_by_committee)s\\n\n Bestilt dato: %(ordered_date)s\\n\n \\n\n For mer informasjon, sjekk ut bestillingen her: %(absolute_url)s\n '''\n % {\n 'site': '',\n 'order_type': type_name.lower(),\n 'num': poster.amount,\n 'ordered_by': poster.ordered_by,\n 'ordered_by_committee': poster.ordered_committee,\n 'id': poster.id,\n 'event_name': title,\n 'ordered_date': poster.ordered_date,\n 'absolute_url': request.build_absolute_uri(poster.get_dashboard_url())\n }\n ),\n 'signature': _('\\n\\nVennlig hilsen Linjeforeningen Online')\n }\n from_email = settings.EMAIL_PROKOM\n to_emails = [settings.EMAIL_PROKOM, request.user.get_email()]\n\n try:\n email_sent = EmailMessage(unicode(subject), unicode(message), from_email, to_emails, []).send()\n except:\n email_sent = False\n\n if email_sent:\n messages.success(request, 'Opprettet bestilling')\n else:\n messages.error(request, 'Klarte ikke \u00e5 sende epost, men bestillingen din ble fortsatt opprettet')\n\n return redirect(poster.get_absolute_url())\n else:\n context['form'] = form\n return render(request, 'posters/dashboard/add.html', context)\n\n context[\"order_type_name\"] = type_name\n context['order_type'] = order_type\n context['can_edit'] = True # request.user.has_perm('posters.view_poster')\n\n if order_type == 1:\n form = AddPosterForm()\n elif order_type == 2:\n form = AddBongForm()\n elif order_type == 3:\n form = AddOtherForm()\n\n forms = (AddPosterForm(), AddBongForm(), AddOtherForm())\n\n context['form'] = forms[order_type-1]\n\n return render(request, 'posters/dashboard/add.html', context)\n\n\n@ensure_csrf_cookie\n@login_required\ndef edit(request, order_id=None):\n context = get_base_context(request)\n context['add_poster_form'] = EditPosterForm()\n\n if order_id:\n poster = get_object_or_404(Poster, pk=order_id)\n\n if request.user != poster.ordered_by and 'proKom' not in request.user.groups:\n raise PermissionDenied\n\n if request.POST:\n form = AddForm(request.POST, instance=poster)\n if form.is_valid():\n form.save()\n # redirect_url = redirect(poster.get_absolute_url())\n return HttpResponseRedirect(\"../detail/\"+str(poster.id))\n\n else:\n context[\"form\"] = AddForm(instance=poster)\n\n return render(request, 'posters/dashboard/add.html', context)\n\n\n@ensure_csrf_cookie\n@login_required\n@permission_required('view_poster_order', (Poster, 'pk', 'order_id'), return_403=True)\ndef detail(request, order_id=None):\n if request.is_ajax():\n do_ajax_shit = True\n\n if not order_id:\n return HttpResponse(status=400)\n\n context = get_base_context(request)\n poster = get_object_or_404(Poster, pk=order_id)\n context['poster'] = poster\n\n if not has_view_perms(request.user, poster):\n raise PermissionDenied\n\n order_type = poster.order_type\n type_names = (\"Plakat\", \"Bong\", \"Generell \")\n type_name = type_names[order_type-1]\n context[\"order_type_name\"] = type_name\n\n if request.method == 'POST':\n poster_status = request.POST.get('completed')\n if poster_status == 'true' or poster_status == 'false':\n poster.toggle_finished()\n\n return render(request, 'posters/dashboard/details.html', context)\n\n\n# Ajax\n\n\n# @ensure_csrf_cookie\n@login_required\ndef assign_person(request):\n if request.is_ajax():\n if request.method == 'POST':\n order_id = request.POST.get('order_id')\n orders = Poster.objects.filter(pk=order_id)\n assign_to_id = request.POST.get('assign_to_id')\n assign_to = User.objects.get(pk=assign_to_id)\n\n if orders.count() == 0:\n response_text = json.dumps({'message': _(\n u\"\"\"Kan ikke finne en ordre med denne IDen (%s).\nOm feilen vedvarer etter en refresh, kontakt [email protected].\"\"\") % order_id})\n return HttpResponse(status=412, content=response_text)\n\n order = orders[0]\n\n if order.finished or order.assigned_to is not None:\n response_text = json.dumps({'message': _(u\"Denne ordren er allerede behandlet.\")})\n return HttpResponse(status=412, content=response_text)\n\n order.assigned_to = assign_to\n order.save()\n\n return HttpResponse(status=200)\n", "path": "apps/posters/dashboard/views.py"}]}
3,451
504
gh_patches_debug_49032
rasdani/github-patches
git_diff
pulp__pulpcore-5465
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> Replication assumes every plugin supports Publications The ```Replicator``` class, which is being subclassed inside plugins to support the replication feature, assumes that every plugin supports publications as it tries to access ```upstream_distribution["publication"]``` inside one of its methods (for some plugins, the dictionary simply doesn't contain the "publication" key so an exception gets raised). This forces certain subclasses of ```Replicator``` to create workarounds or rewrite the given method. I propose making the method more general, removing such assumptions. Relevant code: https://github.com/pulp/pulpcore/blob/c31a39c186fd734e9e0409321735d7d6509a12ee/pulpcore/app/replica.py#L97C9-L97C96 </issue> <code> [start of pulpcore/app/replica.py] 1 from django.db.models import Model 2 import logging 3 4 from pulp_glue.common.context import PulpContext 5 from pulpcore.tasking.tasks import dispatch 6 from pulpcore.app.tasks.base import ( 7 general_update, 8 general_create, 9 general_multi_delete, 10 ) 11 from pulpcore.plugin.util import get_url, get_domain 12 13 _logger = logging.getLogger(__name__) 14 15 16 class ReplicaContext(PulpContext): 17 def __init__(self, **kwargs): 18 super().__init__(**kwargs) 19 self.out_buf = "" 20 self.err_buf = "" 21 22 def echo(self, message: str, nl: bool = True, err: bool = False) -> None: 23 if err: 24 self.err_buf += message 25 if nl: 26 _logger.warn("{}", self.err_buf) 27 self.err_buf = "" 28 else: 29 self.out_buf += message 30 if nl: 31 _logger.info("{}", self.out_buf) 32 self.out_buf = "" 33 34 35 class Replicator: 36 distribution_ctx_cls = None 37 repository_ctx_cls = None 38 publication_ctx_cls = None 39 remote_model_cls = None 40 repository_model_cls = None 41 distribution_model_cls = None 42 distribution_serializer_name = None 43 repository_serializer_name = None 44 remote_serializer_name = None 45 app_label = None 46 sync_task = None 47 48 def __init__(self, pulp_ctx, task_group, tls_settings): 49 """ 50 :param pulp_ctx: PulpReplicaContext 51 :param task_group: TaskGroup 52 :param ca_cert: str 53 """ 54 self.pulp_ctx = pulp_ctx 55 self.task_group = task_group 56 self.tls_settings = tls_settings 57 self.domain = get_domain() 58 self.distros_uris = [f"pdrn:{self.domain.pulp_id}:distributions"] 59 60 @staticmethod 61 def needs_update(fields_dict, model_instance): 62 """ 63 Compares a Model instance's attributes against a dictionary where keys are attribute 64 names and values are expected values. 65 """ 66 needs_update = False 67 for field_name, value in fields_dict.items(): 68 if isinstance(getattr(model_instance, field_name), Model): 69 if get_url(getattr(model_instance, field_name)) != value: 70 needs_update = True 71 elif getattr(model_instance, field_name) != value: 72 needs_update = True 73 return needs_update 74 75 def upstream_distributions(self, labels=None): 76 if labels: 77 params = {"pulp_label_select": labels} 78 else: 79 params = {} 80 offset = 0 81 list_size = 100 82 while True: 83 distributions = self.distribution_ctx_cls(self.pulp_ctx).list(list_size, offset, params) 84 for distro in distributions: 85 yield distro 86 if len(distributions) < list_size: 87 break 88 offset += list_size 89 90 def url(self, upstream_distribution): 91 return upstream_distribution["base_url"] 92 93 def remote_extra_fields(self, upstream_distribution): 94 return {} 95 96 def create_or_update_remote(self, upstream_distribution): 97 if not upstream_distribution["repository"] and not upstream_distribution["publication"]: 98 return None 99 url = self.url(upstream_distribution) 100 remote_fields_dict = {"url": url} 101 remote_fields_dict.update(self.tls_settings) 102 remote_fields_dict.update(self.remote_extra_fields(upstream_distribution)) 103 104 # Check if there is a remote pointing to this distribution 105 try: 106 remote = self.remote_model_cls.objects.get( 107 name=upstream_distribution["name"], pulp_domain=self.domain 108 ) 109 needs_update = self.needs_update(remote_fields_dict, remote) 110 if needs_update: 111 dispatch( 112 general_update, 113 task_group=self.task_group, 114 exclusive_resources=[remote], 115 args=(remote.pk, self.app_label, self.remote_serializer_name), 116 kwargs={"data": remote_fields_dict, "partial": True}, 117 ) 118 except self.remote_model_cls.DoesNotExist: 119 # Create the remote 120 remote = self.remote_model_cls(name=upstream_distribution["name"], **remote_fields_dict) 121 remote.save() 122 123 return remote 124 125 def repository_extra_fields(self, remote): 126 return {} 127 128 def create_or_update_repository(self, remote): 129 try: 130 repository = self.repository_model_cls.objects.get( 131 name=remote.name, pulp_domain=self.domain 132 ) 133 repo_fields_dict = self.repository_extra_fields(remote) 134 needs_update = self.needs_update(repo_fields_dict, repository) 135 if needs_update: 136 dispatch( 137 general_update, 138 task_group=self.task_group, 139 exclusive_resources=[repository], 140 args=(repository.pk, self.app_label, self.repository_serializer_name), 141 kwargs={"data": repo_fields_dict, "partial": True}, 142 ) 143 except self.repository_model_cls.DoesNotExist: 144 repository = self.repository_model_cls( 145 name=remote.name, **self.repository_extra_fields(remote) 146 ) 147 repository.save() 148 return repository 149 150 def distribution_data(self, repository, upstream_distribution): 151 """ 152 Return the fields that need to be updated/cleared on distributions for idempotence. 153 """ 154 return { 155 "repository": get_url(repository), 156 "publication": None, 157 "base_path": upstream_distribution["base_path"], 158 } 159 160 def create_or_update_distribution(self, repository, upstream_distribution): 161 distribution_data = self.distribution_data(repository, upstream_distribution) 162 try: 163 distro = self.distribution_model_cls.objects.get( 164 name=upstream_distribution["name"], pulp_domain=self.domain 165 ) 166 # Check that the distribution has the right repository associated 167 needs_update = self.needs_update(distribution_data, distro) 168 if needs_update: 169 # Update the distribution 170 dispatch( 171 general_update, 172 task_group=self.task_group, 173 shared_resources=[repository], 174 exclusive_resources=self.distros_uris, 175 args=(distro.pk, self.app_label, self.distribution_serializer_name), 176 kwargs={ 177 "data": distribution_data, 178 "partial": True, 179 }, 180 ) 181 except self.distribution_model_cls.DoesNotExist: 182 # Dispatch a task to create the distribution 183 distribution_data["name"] = upstream_distribution["name"] 184 dispatch( 185 general_create, 186 task_group=self.task_group, 187 shared_resources=[repository], 188 exclusive_resources=self.distros_uris, 189 args=(self.app_label, self.distribution_serializer_name), 190 kwargs={"data": distribution_data}, 191 ) 192 193 def sync_params(self, repository, remote): 194 """This method returns a dict that will be passed as kwargs to the sync task.""" 195 raise NotImplementedError("Each replicator must supply its own sync params.") 196 197 def sync(self, repository, remote): 198 dispatch( 199 self.sync_task, 200 task_group=self.task_group, 201 shared_resources=[remote], 202 exclusive_resources=[repository], 203 kwargs=self.sync_params(repository, remote), 204 ) 205 206 def remove_missing(self, names): 207 # Remove all distributions with names not present in the list of names 208 # Perform this in an extra task, because we hold a big lock here. 209 distribution_ids = [ 210 (distribution.pk, self.app_label, self.distribution_serializer_name) 211 for distribution in self.distribution_model_cls.objects.filter( 212 pulp_domain=self.domain 213 ).exclude(name__in=names) 214 ] 215 if distribution_ids: 216 dispatch( 217 general_multi_delete, 218 task_group=self.task_group, 219 exclusive_resources=self.distros_uris, 220 args=(distribution_ids,), 221 ) 222 223 # Remove all the repositories and remotes of the missing distributions 224 repositories = list( 225 self.repository_model_cls.objects.filter( 226 pulp_domain=self.domain, user_hidden=False 227 ).exclude(name__in=names) 228 ) 229 repository_ids = [ 230 (repo.pk, self.app_label, self.repository_serializer_name) for repo in repositories 231 ] 232 233 remotes = list( 234 self.remote_model_cls.objects.filter(pulp_domain=self.domain).exclude(name__in=names) 235 ) 236 remote_ids = [ 237 (remote.pk, self.app_label, self.remote_serializer_name) for remote in remotes 238 ] 239 240 if repository_ids or remote_ids: 241 dispatch( 242 general_multi_delete, 243 task_group=self.task_group, 244 exclusive_resources=repositories + remotes, 245 args=(repository_ids + remote_ids,), 246 ) 247 [end of pulpcore/app/replica.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/pulpcore/app/replica.py b/pulpcore/app/replica.py --- a/pulpcore/app/replica.py +++ b/pulpcore/app/replica.py @@ -94,7 +94,9 @@ return {} def create_or_update_remote(self, upstream_distribution): - if not upstream_distribution["repository"] and not upstream_distribution["publication"]: + if not upstream_distribution.get("repository") and not upstream_distribution.get( + "publication" + ): return None url = self.url(upstream_distribution) remote_fields_dict = {"url": url}
{"golden_diff": "diff --git a/pulpcore/app/replica.py b/pulpcore/app/replica.py\n--- a/pulpcore/app/replica.py\n+++ b/pulpcore/app/replica.py\n@@ -94,7 +94,9 @@\n return {}\n \n def create_or_update_remote(self, upstream_distribution):\n- if not upstream_distribution[\"repository\"] and not upstream_distribution[\"publication\"]:\n+ if not upstream_distribution.get(\"repository\") and not upstream_distribution.get(\n+ \"publication\"\n+ ):\n return None\n url = self.url(upstream_distribution)\n remote_fields_dict = {\"url\": url}\n", "issue": "Replication assumes every plugin supports Publications\nThe ```Replicator``` class, which is being subclassed inside plugins to support the replication feature, assumes that every plugin supports publications as it tries to access ```upstream_distribution[\"publication\"]``` inside one of its methods (for some plugins, the dictionary simply doesn't contain the \"publication\" key so an exception gets raised). This forces certain subclasses of ```Replicator``` to create workarounds or rewrite the given method.\r\n\r\nI propose making the method more general, removing such assumptions.\r\n\r\nRelevant code: https://github.com/pulp/pulpcore/blob/c31a39c186fd734e9e0409321735d7d6509a12ee/pulpcore/app/replica.py#L97C9-L97C96\r\n\n", "before_files": [{"content": "from django.db.models import Model\nimport logging\n\nfrom pulp_glue.common.context import PulpContext\nfrom pulpcore.tasking.tasks import dispatch\nfrom pulpcore.app.tasks.base import (\n general_update,\n general_create,\n general_multi_delete,\n)\nfrom pulpcore.plugin.util import get_url, get_domain\n\n_logger = logging.getLogger(__name__)\n\n\nclass ReplicaContext(PulpContext):\n def __init__(self, **kwargs):\n super().__init__(**kwargs)\n self.out_buf = \"\"\n self.err_buf = \"\"\n\n def echo(self, message: str, nl: bool = True, err: bool = False) -> None:\n if err:\n self.err_buf += message\n if nl:\n _logger.warn(\"{}\", self.err_buf)\n self.err_buf = \"\"\n else:\n self.out_buf += message\n if nl:\n _logger.info(\"{}\", self.out_buf)\n self.out_buf = \"\"\n\n\nclass Replicator:\n distribution_ctx_cls = None\n repository_ctx_cls = None\n publication_ctx_cls = None\n remote_model_cls = None\n repository_model_cls = None\n distribution_model_cls = None\n distribution_serializer_name = None\n repository_serializer_name = None\n remote_serializer_name = None\n app_label = None\n sync_task = None\n\n def __init__(self, pulp_ctx, task_group, tls_settings):\n \"\"\"\n :param pulp_ctx: PulpReplicaContext\n :param task_group: TaskGroup\n :param ca_cert: str\n \"\"\"\n self.pulp_ctx = pulp_ctx\n self.task_group = task_group\n self.tls_settings = tls_settings\n self.domain = get_domain()\n self.distros_uris = [f\"pdrn:{self.domain.pulp_id}:distributions\"]\n\n @staticmethod\n def needs_update(fields_dict, model_instance):\n \"\"\"\n Compares a Model instance's attributes against a dictionary where keys are attribute\n names and values are expected values.\n \"\"\"\n needs_update = False\n for field_name, value in fields_dict.items():\n if isinstance(getattr(model_instance, field_name), Model):\n if get_url(getattr(model_instance, field_name)) != value:\n needs_update = True\n elif getattr(model_instance, field_name) != value:\n needs_update = True\n return needs_update\n\n def upstream_distributions(self, labels=None):\n if labels:\n params = {\"pulp_label_select\": labels}\n else:\n params = {}\n offset = 0\n list_size = 100\n while True:\n distributions = self.distribution_ctx_cls(self.pulp_ctx).list(list_size, offset, params)\n for distro in distributions:\n yield distro\n if len(distributions) < list_size:\n break\n offset += list_size\n\n def url(self, upstream_distribution):\n return upstream_distribution[\"base_url\"]\n\n def remote_extra_fields(self, upstream_distribution):\n return {}\n\n def create_or_update_remote(self, upstream_distribution):\n if not upstream_distribution[\"repository\"] and not upstream_distribution[\"publication\"]:\n return None\n url = self.url(upstream_distribution)\n remote_fields_dict = {\"url\": url}\n remote_fields_dict.update(self.tls_settings)\n remote_fields_dict.update(self.remote_extra_fields(upstream_distribution))\n\n # Check if there is a remote pointing to this distribution\n try:\n remote = self.remote_model_cls.objects.get(\n name=upstream_distribution[\"name\"], pulp_domain=self.domain\n )\n needs_update = self.needs_update(remote_fields_dict, remote)\n if needs_update:\n dispatch(\n general_update,\n task_group=self.task_group,\n exclusive_resources=[remote],\n args=(remote.pk, self.app_label, self.remote_serializer_name),\n kwargs={\"data\": remote_fields_dict, \"partial\": True},\n )\n except self.remote_model_cls.DoesNotExist:\n # Create the remote\n remote = self.remote_model_cls(name=upstream_distribution[\"name\"], **remote_fields_dict)\n remote.save()\n\n return remote\n\n def repository_extra_fields(self, remote):\n return {}\n\n def create_or_update_repository(self, remote):\n try:\n repository = self.repository_model_cls.objects.get(\n name=remote.name, pulp_domain=self.domain\n )\n repo_fields_dict = self.repository_extra_fields(remote)\n needs_update = self.needs_update(repo_fields_dict, repository)\n if needs_update:\n dispatch(\n general_update,\n task_group=self.task_group,\n exclusive_resources=[repository],\n args=(repository.pk, self.app_label, self.repository_serializer_name),\n kwargs={\"data\": repo_fields_dict, \"partial\": True},\n )\n except self.repository_model_cls.DoesNotExist:\n repository = self.repository_model_cls(\n name=remote.name, **self.repository_extra_fields(remote)\n )\n repository.save()\n return repository\n\n def distribution_data(self, repository, upstream_distribution):\n \"\"\"\n Return the fields that need to be updated/cleared on distributions for idempotence.\n \"\"\"\n return {\n \"repository\": get_url(repository),\n \"publication\": None,\n \"base_path\": upstream_distribution[\"base_path\"],\n }\n\n def create_or_update_distribution(self, repository, upstream_distribution):\n distribution_data = self.distribution_data(repository, upstream_distribution)\n try:\n distro = self.distribution_model_cls.objects.get(\n name=upstream_distribution[\"name\"], pulp_domain=self.domain\n )\n # Check that the distribution has the right repository associated\n needs_update = self.needs_update(distribution_data, distro)\n if needs_update:\n # Update the distribution\n dispatch(\n general_update,\n task_group=self.task_group,\n shared_resources=[repository],\n exclusive_resources=self.distros_uris,\n args=(distro.pk, self.app_label, self.distribution_serializer_name),\n kwargs={\n \"data\": distribution_data,\n \"partial\": True,\n },\n )\n except self.distribution_model_cls.DoesNotExist:\n # Dispatch a task to create the distribution\n distribution_data[\"name\"] = upstream_distribution[\"name\"]\n dispatch(\n general_create,\n task_group=self.task_group,\n shared_resources=[repository],\n exclusive_resources=self.distros_uris,\n args=(self.app_label, self.distribution_serializer_name),\n kwargs={\"data\": distribution_data},\n )\n\n def sync_params(self, repository, remote):\n \"\"\"This method returns a dict that will be passed as kwargs to the sync task.\"\"\"\n raise NotImplementedError(\"Each replicator must supply its own sync params.\")\n\n def sync(self, repository, remote):\n dispatch(\n self.sync_task,\n task_group=self.task_group,\n shared_resources=[remote],\n exclusive_resources=[repository],\n kwargs=self.sync_params(repository, remote),\n )\n\n def remove_missing(self, names):\n # Remove all distributions with names not present in the list of names\n # Perform this in an extra task, because we hold a big lock here.\n distribution_ids = [\n (distribution.pk, self.app_label, self.distribution_serializer_name)\n for distribution in self.distribution_model_cls.objects.filter(\n pulp_domain=self.domain\n ).exclude(name__in=names)\n ]\n if distribution_ids:\n dispatch(\n general_multi_delete,\n task_group=self.task_group,\n exclusive_resources=self.distros_uris,\n args=(distribution_ids,),\n )\n\n # Remove all the repositories and remotes of the missing distributions\n repositories = list(\n self.repository_model_cls.objects.filter(\n pulp_domain=self.domain, user_hidden=False\n ).exclude(name__in=names)\n )\n repository_ids = [\n (repo.pk, self.app_label, self.repository_serializer_name) for repo in repositories\n ]\n\n remotes = list(\n self.remote_model_cls.objects.filter(pulp_domain=self.domain).exclude(name__in=names)\n )\n remote_ids = [\n (remote.pk, self.app_label, self.remote_serializer_name) for remote in remotes\n ]\n\n if repository_ids or remote_ids:\n dispatch(\n general_multi_delete,\n task_group=self.task_group,\n exclusive_resources=repositories + remotes,\n args=(repository_ids + remote_ids,),\n )\n", "path": "pulpcore/app/replica.py"}]}
3,094
134
gh_patches_debug_20975
rasdani/github-patches
git_diff
holoviz__panel-1318
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> DataFrame widget raises exception if two columns have the same names Panel: 0.7 #### Description of expected behavior and the observed behavior Normally I don't have the same column names for my dataframes. But I was trying to reproduce this dashboard https://getbootstrap.com/docs/4.3/examples/dashboard/ and they have the same column names in multiple columns. I can see that the panel.widgets.DataFrame does not support this. I would have expected it to support this situation. #### Complete, minimal, self-contained example code that reproduces the issue ```python import panel as pn import pandas as pd table_data = pd.DataFrame([(1001, "Lorem"),], columns=["Header", "Header"]) pn.widgets.DataFrame(table_data).servable() ``` #### Stack traceback and/or browser JavaScript console output ```bash $ python -m panel serve 'scripts\issue2.py' --dev 2019-11-29 09:00:56,834 Starting Bokeh server version 1.4.0 (running on Tornado 6.0.3) 2019-11-29 09:00:56,836 User authentication hooks NOT provided (default user enabled) 2019-11-29 09:00:56,840 Bokeh app running at: http://localhost:5006/issue2 2019-11-29 09:00:56,840 Starting Bokeh server with process id: 10992 2019-11-29 09:00:58,603 Error running application handler <bokeh.application.handlers.script.ScriptHandler object at 0x00000028FF59DD48>: 'DataFrame' object has no attribute 'dtype' File "generic.py", line 5179, in __getattr__: return object.__getattribute__(self, name) Traceback (most recent call last): File "C:\repos\private\awesome-panel\.venv\lib\site-packages\bokeh\application\handlers\code_runner.py", line 179, in run exec(self._code, module.__dict__) File "C:\repos\private\awesome-panel\scripts\issue2.py", line 5, in <module> pn.widgets.DataFrame(table_data).servable() File "C:\repos\private\awesome-panel\.venv\lib\site-packages\panel\viewable.py", line 525, in servable self.server_doc(title=title) File "C:\repos\private\awesome-panel\.venv\lib\site-packages\panel\viewable.py", line 498, in server_doc model = self.get_root(doc) File "C:\repos\private\awesome-panel\.venv\lib\site-packages\panel\viewable.py", line 434, in get_root root = self._get_model(doc, comm=comm) File "C:\repos\private\awesome-panel\.venv\lib\site-packages\panel\widgets\tables.py", line 112, in _get_model model = DataTable(**self._get_properties()) File "C:\repos\private\awesome-panel\.venv\lib\site-packages\panel\widgets\tables.py", line 99, in _get_properties props['columns'] = self._get_columns() File "C:\repos\private\awesome-panel\.venv\lib\site-packages\panel\widgets\tables.py", line 61, in _get_columns kind = data.dtype.kind File "C:\repos\private\awesome-panel\.venv\lib\site-packages\pandas\core\generic.py", line 5179, in __getattr__ return object.__getattribute__(self, name) AttributeError: 'DataFrame' object has no attribute 'dtype' 2019-11-29 09:00:58,990 WebSocket connection opened 2019-11-29 09:00:58,991 ServerConnection created ``` </issue> <code> [start of panel/widgets/tables.py] 1 from __future__ import absolute_import, division, unicode_literals 2 3 import numpy as np 4 import param 5 6 from bokeh.models import ColumnDataSource 7 from bokeh.models.widgets import ( 8 DataTable, TableColumn, NumberEditor, NumberFormatter, 9 DateFormatter, DateEditor, StringFormatter, StringEditor, IntEditor 10 ) 11 12 from ..viewable import Layoutable 13 from ..util import isdatetime 14 from .base import Widget 15 16 17 class DataFrame(Widget): 18 19 editors = param.Dict(default={}, doc=""" 20 Bokeh CellEditor to use for a particular column 21 (overrides the default chosen based on the type).""") 22 23 formatters = param.Dict(default={}, doc=""" 24 Bokeh CellFormatter to use for a particular column 25 (overrides the default chosen based on the type).""") 26 27 fit_columns = param.Boolean(default=True, doc=""" 28 Whether columns should expand to the available width. This 29 results in no horizontal scrollbar showing up, but data can 30 get unreadable if there is no enough space available.""") 31 32 selection = param.List(default=[], doc=""" 33 The currently selected rows of the table.""") 34 35 row_height = param.Integer(default=25, doc=""" 36 The height of each table row.""") 37 38 widths = param.Dict(default={}, doc=""" 39 A mapping from column name to column width.""") 40 41 value = param.Parameter(default=None) 42 43 _rename = {'editors': None, 'formatters': None, 'widths': None, 44 'disabled': None} 45 46 _manual_params = ['value', 'editors', 'formatters', 'selection', 'widths'] 47 48 def __init__(self, value=None, **params): 49 super(DataFrame, self).__init__(value=value, **params) 50 self._renamed_cols = {} 51 52 def _get_columns(self): 53 if self.value is None: 54 return [] 55 56 index = [self.value.index.name or 'index'] 57 col_names = index + list(self.value.columns) 58 columns = [] 59 for col in col_names: 60 if col in self.value.columns: 61 data = self.value[col] 62 else: 63 data = self.value.index 64 kind = data.dtype.kind 65 if kind == 'i': 66 formatter = NumberFormatter() 67 editor = IntEditor() 68 elif kind == 'f': 69 formatter = NumberFormatter(format='0,0.0[00000]') 70 editor = NumberEditor() 71 elif isdatetime(data) or kind == 'M': 72 formatter = DateFormatter(format='%Y-%m-%d %H:%M:%S') 73 editor = DateEditor() 74 else: 75 formatter = StringFormatter() 76 editor = StringEditor() 77 78 if col in self.editors: 79 editor = self.editors[col] 80 if col in self.formatters: 81 formatter = self.formatters[col] 82 if str(col) != col: 83 self._renamed_cols[str(col)] = col 84 width = self.widths.get(str(col)) 85 column = TableColumn(field=str(col), title=str(col), 86 editor=editor, formatter=formatter, 87 width=width) 88 columns.append(column) 89 return columns 90 91 def _get_properties(self): 92 props = {p : getattr(self, p) for p in list(Layoutable.param) 93 if getattr(self, p) is not None} 94 if self.value is None: 95 data = {} 96 else: 97 data = {k if isinstance(k, str) else str(k): v 98 for k, v in ColumnDataSource.from_df(self.value).items()} 99 if props.get('height', None) is None: 100 length = max([len(v) for v in data.values()]) if data else 0 101 props['height'] = length * self.row_height + 30 102 props['source'] = ColumnDataSource(data=data) 103 props['columns'] = self._get_columns() 104 props['index_position'] = None 105 props['fit_columns'] = self.fit_columns 106 props['row_height'] = self.row_height 107 props['editable'] = not self.disabled 108 return props 109 110 def _process_param_change(self, msg): 111 if 'disabled' in msg: 112 msg['editable'] = not msg.pop('disabled') 113 return super(DataFrame, self)._process_param_change(msg) 114 115 def _get_model(self, doc, root=None, parent=None, comm=None): 116 model = DataTable(**self._get_properties()) 117 if root is None: 118 root = model 119 self._link_props(model.source, ['data', ('patching', 'data')], doc, root, comm) 120 self._link_props(model.source.selected, ['indices'], doc, root, comm) 121 self._models[root.ref['id']] = (model, parent) 122 return model 123 124 def _manual_update(self, events, model, doc, root, parent, comm): 125 for event in events: 126 if event.name == 'value': 127 cds = model.source 128 data = {k if isinstance(k, str) else str(k): v 129 for k, v in ColumnDataSource.from_df(self.value).items()} 130 cds.data = data 131 model.columns = self._get_columns() 132 elif event.name == 'selection': 133 model.source.selected.indices = self.selection 134 else: 135 for col in model.columns: 136 if col.name in self.editors: 137 col.editor = self.editors[col.name] 138 if col.name in self.formatters: 139 col.formatter = self.formatters[col.name] 140 if col.name in self.widths: 141 col.width = self.widths[col.name] 142 143 def _process_events(self, events): 144 if 'data' in events: 145 data = events.pop('data') 146 updated = False 147 for k, v in data.items(): 148 if k == 'index': 149 continue 150 k = self._renamed_cols.get(k, k) 151 if isinstance(v, dict): 152 v = [v for _, v in sorted(v.items(), key=lambda it: int(it[0]))] 153 try: 154 isequal = (self.value[k].values == np.asarray(v)).all() 155 except Exception: 156 isequal = False 157 if not isequal: 158 self.value[k] = v 159 updated = True 160 if updated: 161 self.param.trigger('value') 162 if 'indices' in events: 163 self.selection = events.pop('indices') 164 super(DataFrame, self)._process_events(events) 165 166 @property 167 def selected_dataframe(self): 168 """ 169 Returns a DataFrame of the currently selected rows. 170 """ 171 if not self.selection: 172 return self.value 173 return self.value.iloc[self.selection] 174 [end of panel/widgets/tables.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/panel/widgets/tables.py b/panel/widgets/tables.py --- a/panel/widgets/tables.py +++ b/panel/widgets/tables.py @@ -47,8 +47,18 @@ def __init__(self, value=None, **params): super(DataFrame, self).__init__(value=value, **params) + self.param.watch(self._validate, 'value') + self._validate(None) self._renamed_cols = {} + def _validate(self, event): + if self.value is None: + return + cols = self.value.columns + if len(cols) != len(cols.drop_duplicates()): + raise ValueError('Cannot display a pandas.DataFrame with ' + 'duplicate column names.') + def _get_columns(self): if self.value is None: return [] @@ -122,6 +132,7 @@ return model def _manual_update(self, events, model, doc, root, parent, comm): + self._validate(None) for event in events: if event.name == 'value': cds = model.source
{"golden_diff": "diff --git a/panel/widgets/tables.py b/panel/widgets/tables.py\n--- a/panel/widgets/tables.py\n+++ b/panel/widgets/tables.py\n@@ -47,8 +47,18 @@\n \n def __init__(self, value=None, **params):\n super(DataFrame, self).__init__(value=value, **params)\n+ self.param.watch(self._validate, 'value')\n+ self._validate(None)\n self._renamed_cols = {}\n \n+ def _validate(self, event):\n+ if self.value is None:\n+ return\n+ cols = self.value.columns\n+ if len(cols) != len(cols.drop_duplicates()):\n+ raise ValueError('Cannot display a pandas.DataFrame with '\n+ 'duplicate column names.')\n+\n def _get_columns(self):\n if self.value is None:\n return []\n@@ -122,6 +132,7 @@\n return model\n \n def _manual_update(self, events, model, doc, root, parent, comm):\n+ self._validate(None)\n for event in events:\n if event.name == 'value':\n cds = model.source\n", "issue": "DataFrame widget raises exception if two columns have the same names\nPanel: 0.7\r\n\r\n#### Description of expected behavior and the observed behavior\r\n\r\nNormally I don't have the same column names for my dataframes. But I was trying to reproduce this dashboard https://getbootstrap.com/docs/4.3/examples/dashboard/ and they have the same column names in multiple columns.\r\n\r\nI can see that the panel.widgets.DataFrame does not support this. I would have expected it to support this situation.\r\n\r\n#### Complete, minimal, self-contained example code that reproduces the issue\r\n\r\n```python\r\nimport panel as pn\r\nimport pandas as pd\r\n\r\ntable_data = pd.DataFrame([(1001, \"Lorem\"),], columns=[\"Header\", \"Header\"])\r\npn.widgets.DataFrame(table_data).servable()\r\n```\r\n#### Stack traceback and/or browser JavaScript console output\r\n\r\n```bash\r\n$ python -m panel serve 'scripts\\issue2.py' --dev\r\n2019-11-29 09:00:56,834 Starting Bokeh server version 1.4.0 (running on Tornado 6.0.3)\r\n2019-11-29 09:00:56,836 User authentication hooks NOT provided (default user enabled)\r\n2019-11-29 09:00:56,840 Bokeh app running at: http://localhost:5006/issue2\r\n2019-11-29 09:00:56,840 Starting Bokeh server with process id: 10992\r\n2019-11-29 09:00:58,603 Error running application handler <bokeh.application.handlers.script.ScriptHandler object at 0x00000028FF59DD48>: 'DataFrame' object has no attribute 'dtype'\r\nFile \"generic.py\", line 5179, in __getattr__:\r\nreturn object.__getattribute__(self, name) Traceback (most recent call last):\r\n File \"C:\\repos\\private\\awesome-panel\\.venv\\lib\\site-packages\\bokeh\\application\\handlers\\code_runner.py\", line 179, in run\r\n exec(self._code, module.__dict__)\r\n File \"C:\\repos\\private\\awesome-panel\\scripts\\issue2.py\", line 5, in <module>\r\n pn.widgets.DataFrame(table_data).servable()\r\n File \"C:\\repos\\private\\awesome-panel\\.venv\\lib\\site-packages\\panel\\viewable.py\", line 525, in servable\r\n self.server_doc(title=title)\r\n File \"C:\\repos\\private\\awesome-panel\\.venv\\lib\\site-packages\\panel\\viewable.py\", line 498, in server_doc\r\n model = self.get_root(doc)\r\n File \"C:\\repos\\private\\awesome-panel\\.venv\\lib\\site-packages\\panel\\viewable.py\", line 434, in get_root\r\n root = self._get_model(doc, comm=comm)\r\n File \"C:\\repos\\private\\awesome-panel\\.venv\\lib\\site-packages\\panel\\widgets\\tables.py\", line 112, in _get_model\r\n model = DataTable(**self._get_properties())\r\n File \"C:\\repos\\private\\awesome-panel\\.venv\\lib\\site-packages\\panel\\widgets\\tables.py\", line 99, in _get_properties\r\n props['columns'] = self._get_columns()\r\n File \"C:\\repos\\private\\awesome-panel\\.venv\\lib\\site-packages\\panel\\widgets\\tables.py\", line 61, in _get_columns\r\n kind = data.dtype.kind\r\n File \"C:\\repos\\private\\awesome-panel\\.venv\\lib\\site-packages\\pandas\\core\\generic.py\", line 5179, in __getattr__\r\n return object.__getattribute__(self, name)\r\nAttributeError: 'DataFrame' object has no attribute 'dtype'\r\n\r\n2019-11-29 09:00:58,990 WebSocket connection opened\r\n2019-11-29 09:00:58,991 ServerConnection created\r\n```\n", "before_files": [{"content": "from __future__ import absolute_import, division, unicode_literals\n\nimport numpy as np\nimport param\n\nfrom bokeh.models import ColumnDataSource\nfrom bokeh.models.widgets import (\n DataTable, TableColumn, NumberEditor, NumberFormatter,\n DateFormatter, DateEditor, StringFormatter, StringEditor, IntEditor\n)\n\nfrom ..viewable import Layoutable\nfrom ..util import isdatetime\nfrom .base import Widget\n\n\nclass DataFrame(Widget):\n\n editors = param.Dict(default={}, doc=\"\"\"\n Bokeh CellEditor to use for a particular column\n (overrides the default chosen based on the type).\"\"\")\n\n formatters = param.Dict(default={}, doc=\"\"\"\n Bokeh CellFormatter to use for a particular column\n (overrides the default chosen based on the type).\"\"\")\n\n fit_columns = param.Boolean(default=True, doc=\"\"\"\n Whether columns should expand to the available width. This\n results in no horizontal scrollbar showing up, but data can\n get unreadable if there is no enough space available.\"\"\")\n\n selection = param.List(default=[], doc=\"\"\"\n The currently selected rows of the table.\"\"\")\n\n row_height = param.Integer(default=25, doc=\"\"\"\n The height of each table row.\"\"\")\n\n widths = param.Dict(default={}, doc=\"\"\"\n A mapping from column name to column width.\"\"\")\n\n value = param.Parameter(default=None)\n\n _rename = {'editors': None, 'formatters': None, 'widths': None,\n 'disabled': None}\n\n _manual_params = ['value', 'editors', 'formatters', 'selection', 'widths']\n\n def __init__(self, value=None, **params):\n super(DataFrame, self).__init__(value=value, **params)\n self._renamed_cols = {}\n\n def _get_columns(self):\n if self.value is None:\n return []\n\n index = [self.value.index.name or 'index']\n col_names = index + list(self.value.columns)\n columns = []\n for col in col_names:\n if col in self.value.columns:\n data = self.value[col]\n else:\n data = self.value.index\n kind = data.dtype.kind\n if kind == 'i':\n formatter = NumberFormatter()\n editor = IntEditor()\n elif kind == 'f':\n formatter = NumberFormatter(format='0,0.0[00000]')\n editor = NumberEditor()\n elif isdatetime(data) or kind == 'M':\n formatter = DateFormatter(format='%Y-%m-%d %H:%M:%S')\n editor = DateEditor()\n else:\n formatter = StringFormatter()\n editor = StringEditor()\n\n if col in self.editors:\n editor = self.editors[col]\n if col in self.formatters:\n formatter = self.formatters[col]\n if str(col) != col:\n self._renamed_cols[str(col)] = col\n width = self.widths.get(str(col))\n column = TableColumn(field=str(col), title=str(col),\n editor=editor, formatter=formatter,\n width=width)\n columns.append(column)\n return columns\n\n def _get_properties(self):\n props = {p : getattr(self, p) for p in list(Layoutable.param)\n if getattr(self, p) is not None}\n if self.value is None:\n data = {}\n else:\n data = {k if isinstance(k, str) else str(k): v\n for k, v in ColumnDataSource.from_df(self.value).items()}\n if props.get('height', None) is None:\n length = max([len(v) for v in data.values()]) if data else 0\n props['height'] = length * self.row_height + 30\n props['source'] = ColumnDataSource(data=data)\n props['columns'] = self._get_columns()\n props['index_position'] = None\n props['fit_columns'] = self.fit_columns\n props['row_height'] = self.row_height\n props['editable'] = not self.disabled\n return props\n\n def _process_param_change(self, msg):\n if 'disabled' in msg:\n msg['editable'] = not msg.pop('disabled')\n return super(DataFrame, self)._process_param_change(msg)\n \n def _get_model(self, doc, root=None, parent=None, comm=None):\n model = DataTable(**self._get_properties())\n if root is None:\n root = model\n self._link_props(model.source, ['data', ('patching', 'data')], doc, root, comm)\n self._link_props(model.source.selected, ['indices'], doc, root, comm)\n self._models[root.ref['id']] = (model, parent)\n return model\n\n def _manual_update(self, events, model, doc, root, parent, comm):\n for event in events:\n if event.name == 'value':\n cds = model.source\n data = {k if isinstance(k, str) else str(k): v\n for k, v in ColumnDataSource.from_df(self.value).items()}\n cds.data = data\n model.columns = self._get_columns()\n elif event.name == 'selection':\n model.source.selected.indices = self.selection\n else:\n for col in model.columns:\n if col.name in self.editors:\n col.editor = self.editors[col.name]\n if col.name in self.formatters:\n col.formatter = self.formatters[col.name]\n if col.name in self.widths:\n col.width = self.widths[col.name]\n\n def _process_events(self, events):\n if 'data' in events:\n data = events.pop('data')\n updated = False\n for k, v in data.items():\n if k == 'index':\n continue\n k = self._renamed_cols.get(k, k)\n if isinstance(v, dict):\n v = [v for _, v in sorted(v.items(), key=lambda it: int(it[0]))]\n try:\n isequal = (self.value[k].values == np.asarray(v)).all()\n except Exception:\n isequal = False\n if not isequal:\n self.value[k] = v\n updated = True\n if updated:\n self.param.trigger('value')\n if 'indices' in events:\n self.selection = events.pop('indices')\n super(DataFrame, self)._process_events(events)\n\n @property\n def selected_dataframe(self):\n \"\"\"\n Returns a DataFrame of the currently selected rows.\n \"\"\"\n if not self.selection:\n return self.value\n return self.value.iloc[self.selection]\n", "path": "panel/widgets/tables.py"}]}
3,282
248
gh_patches_debug_42532
rasdani/github-patches
git_diff
kserve__kserve-1343
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> Support cloud events for KFServing custom framework /kind feature **Describe the solution you'd like** As a user when doing feature transformation the payload can be protobuf/avro or some other binary format delivered via HTTP or KAFKA in cloudevent binary format(https://github.com/cloudevents/spec/blob/v1.0.1/kafka-protocol-binding.md#32-binary-content-mode) but currently KFServing custom framework [only supports JSON](https://github.com/kubeflow/kfserving/blob/master/python/kfserving/kfserving/handlers/http.py#L51). - I'd like to send cloud event to KFServer via KAFKA in following format and KFServer pass the decoded cloud event to `preprocess` or `predict` handler ``` ------------------ Message ------------------- Topic Name: mytopic ------------------- key ---------------------- Key: mykey ------------------ headers ------------------- ce_specversion: "1.0" ce_type: "com.example.someevent" ce_source: "/mycontext/subcontext" ce_id: "1234-1234-1234" ce_time: "2018-04-05T03:56:24Z" content-type: application/protobuf ------------------- value -------------------- ... application data encoded in protobuf ... ----------------------------------------------- consume http(cloudevent) KAFKA ----------> Kafka Event Source -----------------> InferenceService ``` ```python def has_ce_headers(headers: typing.Dict[str, str]) -> bool: return ( "ce-specversion" in headers and "ce-source" in headers and "ce-type" in headers and "ce-id" in headers ) if has_ce_headers(self.request.headers): event = from_http(self.request.headers, self.request.body) body = event['data'] else: try: body = json.loads(self.request.body) except json.decoder.JSONDecodeError as e: raise tornado.web.HTTPError( status_code=HTTPStatus.BAD_REQUEST, reason="Unrecognized request format: %s" % e) ``` - Currently HTTP headers are not exposed to the `preprocess` or `predict` handler, if a cloud event is sent to KFServer it is not possible for user to check the http headers which can be useful for tracing and examining the original event source(kafka or http). `async def predict(self, request: Dict, headers: Dict) -> Dict:` `async def preprocess(self, request: Dict, headers: Dict) -> Dict:` - Also it would be nice if KFServing can generate cloud event response so that it works with knative eventing channels seamlessly. **Anything else you would like to add:** [Miscellaneous information that will assist in solving the issue.] </issue> <code> [start of python/kfserving/kfserving/handlers/http.py] 1 # Copyright 2020 kubeflow.org. 2 # 3 # Licensed under the Apache License, Version 2.0 (the "License"); 4 # you may not use this file except in compliance with the License. 5 # You may obtain a copy of the License at 6 # 7 # http://www.apache.org/licenses/LICENSE-2.0 8 # 9 # Unless required by applicable law or agreed to in writing, software 10 # distributed under the License is distributed on an "AS IS" BASIS, 11 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 12 # See the License for the specific language governing permissions and 13 # limitations under the License. 14 15 import inspect 16 import tornado.web 17 import json 18 from http import HTTPStatus 19 from kfserving.kfmodel_repository import KFModelRepository 20 21 22 class HTTPHandler(tornado.web.RequestHandler): 23 def initialize(self, models: KFModelRepository): 24 self.models = models # pylint:disable=attribute-defined-outside-init 25 26 def get_model(self, name: str): 27 model = self.models.get_model(name) 28 if model is None: 29 raise tornado.web.HTTPError( 30 status_code=HTTPStatus.NOT_FOUND, 31 reason="Model with name %s does not exist." % name 32 ) 33 if not model.ready: 34 model.load() 35 return model 36 37 def validate(self, request): 38 if ("instances" in request and not isinstance(request["instances"], list)) or \ 39 ("inputs" in request and not isinstance(request["inputs"], list)): 40 raise tornado.web.HTTPError( 41 status_code=HTTPStatus.BAD_REQUEST, 42 reason="Expected \"instances\" or \"inputs\" to be a list" 43 ) 44 return request 45 46 47 class PredictHandler(HTTPHandler): 48 async def post(self, name: str): 49 model = self.get_model(name) 50 try: 51 body = json.loads(self.request.body) 52 except json.decoder.JSONDecodeError as e: 53 raise tornado.web.HTTPError( 54 status_code=HTTPStatus.BAD_REQUEST, 55 reason="Unrecognized request format: %s" % e 56 ) 57 request = model.preprocess(body) 58 request = self.validate(request) 59 response = (await model.predict(request)) if inspect.iscoroutinefunction(model.predict) else model.predict(request) 60 response = model.postprocess(response) 61 self.write(response) 62 63 64 class ExplainHandler(HTTPHandler): 65 async def post(self, name: str): 66 model = self.get_model(name) 67 try: 68 body = json.loads(self.request.body) 69 except json.decoder.JSONDecodeError as e: 70 raise tornado.web.HTTPError( 71 status_code=HTTPStatus.BAD_REQUEST, 72 reason="Unrecognized request format: %s" % e 73 ) 74 request = model.preprocess(body) 75 request = self.validate(request) 76 response = (await model.explain(request)) if inspect.iscoroutinefunction(model.explain) else model.explain(request) 77 response = model.postprocess(response) 78 self.write(response) 79 [end of python/kfserving/kfserving/handlers/http.py] [start of python/kfserving/kfserving/kfmodel.py] 1 # Copyright 2020 kubeflow.org. 2 # 3 # Licensed under the Apache License, Version 2.0 (the "License"); 4 # you may not use this file except in compliance with the License. 5 # You may obtain a copy of the License at 6 # 7 # http://www.apache.org/licenses/LICENSE-2.0 8 # 9 # Unless required by applicable law or agreed to in writing, software 10 # distributed under the License is distributed on an "AS IS" BASIS, 11 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 12 # See the License for the specific language governing permissions and 13 # limitations under the License. 14 15 from typing import Dict 16 import sys 17 18 import json 19 import tornado.web 20 from tornado.httpclient import AsyncHTTPClient 21 22 PREDICTOR_URL_FORMAT = "http://{0}/v1/models/{1}:predict" 23 EXPLAINER_URL_FORMAT = "http://{0}/v1/models/{1}:explain" 24 PREDICTOR_V2_URL_FORMAT = "http://{0}/v2/models/{1}/infer" 25 EXPLAINER_V2_URL_FORMAT = "http://{0}/v2/models/{1}/explain" 26 27 28 # KFModel is intended to be subclassed by various components within KFServing. 29 class KFModel: 30 31 def __init__(self, name: str): 32 self.name = name 33 self.ready = False 34 self.protocol = "v1" 35 self.predictor_host = None 36 self.explainer_host = None 37 # The timeout matches what is set in generated Istio resources. 38 # We generally don't want things to time out at the request level here, 39 # timeouts should be handled elsewhere in the system. 40 self.timeout = 600 41 self._http_client_instance = None 42 43 @property 44 def _http_client(self): 45 if self._http_client_instance is None: 46 self._http_client_instance = AsyncHTTPClient(max_clients=sys.maxsize) 47 return self._http_client_instance 48 49 def load(self) -> bool: 50 self.ready = True 51 return self.ready 52 53 def preprocess(self, request: Dict) -> Dict: 54 return request 55 56 def postprocess(self, request: Dict) -> Dict: 57 return request 58 59 async def predict(self, request: Dict) -> Dict: 60 if not self.predictor_host: 61 raise NotImplementedError 62 predict_url = PREDICTOR_URL_FORMAT.format(self.predictor_host, self.name) 63 if self.protocol == "v2": 64 predict_url = PREDICTOR_V2_URL_FORMAT.format(self.predictor_host, self.name) 65 response = await self._http_client.fetch( 66 predict_url, 67 method='POST', 68 request_timeout=self.timeout, 69 body=json.dumps(request) 70 ) 71 if response.code != 200: 72 raise tornado.web.HTTPError( 73 status_code=response.code, 74 reason=response.body) 75 return json.loads(response.body) 76 77 async def explain(self, request: Dict) -> Dict: 78 if self.explainer_host is None: 79 raise NotImplementedError 80 explain_url = EXPLAINER_URL_FORMAT.format(self.predictor_host, self.name) 81 if self.protocol == "v2": 82 explain_url = EXPLAINER_V2_URL_FORMAT.format(self.predictor_host, self.name) 83 response = await self._http_client.fetch( 84 url=explain_url, 85 method='POST', 86 request_timeout=self.timeout, 87 body=json.dumps(request) 88 ) 89 if response.code != 200: 90 raise tornado.web.HTTPError( 91 status_code=response.code, 92 reason=response.body) 93 return json.loads(response.body) 94 [end of python/kfserving/kfserving/kfmodel.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/python/kfserving/kfserving/handlers/http.py b/python/kfserving/kfserving/handlers/http.py --- a/python/kfserving/kfserving/handlers/http.py +++ b/python/kfserving/kfserving/handlers/http.py @@ -14,9 +14,15 @@ import inspect import tornado.web +import typing import json +import pytz +import cloudevents.exceptions as ce +from cloudevents.http import CloudEvent, from_http, is_binary, is_structured, to_binary, to_structured +from cloudevents.sdk.converters.util import has_binary_headers from http import HTTPStatus from kfserving.kfmodel_repository import KFModelRepository +from datetime import datetime class HTTPHandler(tornado.web.RequestHandler): @@ -43,21 +49,52 @@ ) return request - class PredictHandler(HTTPHandler): async def post(self, name: str): + if has_binary_headers(self.request.headers): + try: + #Use default unmarshaller if contenttype is set in header + if "ce-contenttype" in self.request.headers: + body = from_http(self.request.headers, self.request.body) + else: + body = from_http(self.request.headers, self.request.body, lambda x: x) + except (ce.MissingRequiredFields, ce.InvalidRequiredFields, ce.InvalidStructuredJSON, ce.InvalidHeadersFormat, ce.DataMarshallerError, ce.DataUnmarshallerError) as e: + raise tornado.web.HTTPError( + status_code=HTTPStatus.BAD_REQUEST, + reason="Cloud Event Exceptions: %s" % e + ) + else: + try: + body = json.loads(self.request.body) + except json.decoder.JSONDecodeError as e: + raise tornado.web.HTTPError( + status_code=HTTPStatus.BAD_REQUEST, + reason="Unrecognized request format: %s" % e + ) + model = self.get_model(name) - try: - body = json.loads(self.request.body) - except json.decoder.JSONDecodeError as e: - raise tornado.web.HTTPError( - status_code=HTTPStatus.BAD_REQUEST, - reason="Unrecognized request format: %s" % e - ) request = model.preprocess(body) request = self.validate(request) response = (await model.predict(request)) if inspect.iscoroutinefunction(model.predict) else model.predict(request) response = model.postprocess(response) + + if has_binary_headers(self.request.headers): + event = CloudEvent(body._attributes, response) + if is_binary(self.request.headers): + eventheader, eventbody = to_binary(event) + elif is_structured(self.request.headers): + eventheader, eventbody = to_structured(event) + for k, v in eventheader.items(): + if k != "ce-time": + self.set_header(k, v) + else: #utc now() timestamp + self.set_header('ce-time', datetime.utcnow().replace(tzinfo=pytz.utc).strftime('%Y-%m-%dT%H:%M:%S.%f%z')) + + if isinstance(eventbody, (bytes, bytearray)): + response = eventbody + else: + response = eventbody.data + self.write(response) diff --git a/python/kfserving/kfserving/kfmodel.py b/python/kfserving/kfserving/kfmodel.py --- a/python/kfserving/kfserving/kfmodel.py +++ b/python/kfserving/kfserving/kfmodel.py @@ -51,7 +51,16 @@ return self.ready def preprocess(self, request: Dict) -> Dict: - return request + # If cloudevent dict, then parse 'data' field. Otherwise, pass through. + if "data" in request \ + and "time" in request \ + and "type" in request \ + and "source" in request \ + and "id" in request \ + and "specversion" in request: + return request["data"] + else: + return request def postprocess(self, request: Dict) -> Dict: return request @@ -91,3 +100,4 @@ status_code=response.code, reason=response.body) return json.loads(response.body) +
{"golden_diff": "diff --git a/python/kfserving/kfserving/handlers/http.py b/python/kfserving/kfserving/handlers/http.py\n--- a/python/kfserving/kfserving/handlers/http.py\n+++ b/python/kfserving/kfserving/handlers/http.py\n@@ -14,9 +14,15 @@\n \n import inspect\n import tornado.web\n+import typing\n import json\n+import pytz\n+import cloudevents.exceptions as ce\n+from cloudevents.http import CloudEvent, from_http, is_binary, is_structured, to_binary, to_structured\n+from cloudevents.sdk.converters.util import has_binary_headers\n from http import HTTPStatus\n from kfserving.kfmodel_repository import KFModelRepository\n+from datetime import datetime\n \n \n class HTTPHandler(tornado.web.RequestHandler):\n@@ -43,21 +49,52 @@\n )\n return request\n \n-\n class PredictHandler(HTTPHandler):\n async def post(self, name: str):\n+ if has_binary_headers(self.request.headers): \n+ try:\n+ #Use default unmarshaller if contenttype is set in header\n+ if \"ce-contenttype\" in self.request.headers:\n+ body = from_http(self.request.headers, self.request.body)\n+ else:\n+ body = from_http(self.request.headers, self.request.body, lambda x: x)\n+ except (ce.MissingRequiredFields, ce.InvalidRequiredFields, ce.InvalidStructuredJSON, ce.InvalidHeadersFormat, ce.DataMarshallerError, ce.DataUnmarshallerError) as e:\n+ raise tornado.web.HTTPError(\n+ status_code=HTTPStatus.BAD_REQUEST,\n+ reason=\"Cloud Event Exceptions: %s\" % e\n+ )\n+ else:\n+ try:\n+ body = json.loads(self.request.body)\n+ except json.decoder.JSONDecodeError as e:\n+ raise tornado.web.HTTPError(\n+ status_code=HTTPStatus.BAD_REQUEST,\n+ reason=\"Unrecognized request format: %s\" % e\n+ )\n+\n model = self.get_model(name)\n- try:\n- body = json.loads(self.request.body)\n- except json.decoder.JSONDecodeError as e:\n- raise tornado.web.HTTPError(\n- status_code=HTTPStatus.BAD_REQUEST,\n- reason=\"Unrecognized request format: %s\" % e\n- )\n request = model.preprocess(body)\n request = self.validate(request)\n response = (await model.predict(request)) if inspect.iscoroutinefunction(model.predict) else model.predict(request)\n response = model.postprocess(response)\n+\n+ if has_binary_headers(self.request.headers):\n+ event = CloudEvent(body._attributes, response)\n+ if is_binary(self.request.headers):\n+ eventheader, eventbody = to_binary(event)\n+ elif is_structured(self.request.headers):\n+ eventheader, eventbody = to_structured(event)\n+ for k, v in eventheader.items():\n+ if k != \"ce-time\":\n+ self.set_header(k, v)\n+ else: #utc now() timestamp\n+ self.set_header('ce-time', datetime.utcnow().replace(tzinfo=pytz.utc).strftime('%Y-%m-%dT%H:%M:%S.%f%z'))\n+\n+ if isinstance(eventbody, (bytes, bytearray)):\n+ response = eventbody\n+ else:\n+ response = eventbody.data\n+\n self.write(response)\n \n \ndiff --git a/python/kfserving/kfserving/kfmodel.py b/python/kfserving/kfserving/kfmodel.py\n--- a/python/kfserving/kfserving/kfmodel.py\n+++ b/python/kfserving/kfserving/kfmodel.py\n@@ -51,7 +51,16 @@\n return self.ready\n \n def preprocess(self, request: Dict) -> Dict:\n- return request\n+ # If cloudevent dict, then parse 'data' field. Otherwise, pass through.\n+ if \"data\" in request \\\n+ and \"time\" in request \\\n+ and \"type\" in request \\\n+ and \"source\" in request \\\n+ and \"id\" in request \\\n+ and \"specversion\" in request:\n+ return request[\"data\"]\n+ else:\n+ return request\n \n def postprocess(self, request: Dict) -> Dict:\n return request\n@@ -91,3 +100,4 @@\n status_code=response.code,\n reason=response.body)\n return json.loads(response.body)\n+\n", "issue": "Support cloud events for KFServing custom framework \n/kind feature\r\n\r\n**Describe the solution you'd like**\r\nAs a user when doing feature transformation the payload can be protobuf/avro or some other binary format delivered via HTTP or KAFKA in cloudevent binary format(https://github.com/cloudevents/spec/blob/v1.0.1/kafka-protocol-binding.md#32-binary-content-mode) but currently KFServing custom framework [only supports JSON](https://github.com/kubeflow/kfserving/blob/master/python/kfserving/kfserving/handlers/http.py#L51).\r\n\r\n- I'd like to send cloud event to KFServer via KAFKA in following format and KFServer pass the decoded cloud event to `preprocess` or `predict` handler\r\n \r\n```\r\n------------------ Message -------------------\r\nTopic Name: mytopic\r\n------------------- key ----------------------\r\nKey: mykey\r\n------------------ headers -------------------\r\nce_specversion: \"1.0\"\r\nce_type: \"com.example.someevent\"\r\nce_source: \"/mycontext/subcontext\"\r\nce_id: \"1234-1234-1234\"\r\nce_time: \"2018-04-05T03:56:24Z\"\r\ncontent-type: application/protobuf\r\n------------------- value --------------------\r\n ... application data encoded in protobuf ...\r\n-----------------------------------------------\r\n\r\n consume http(cloudevent)\r\nKAFKA ----------> Kafka Event Source -----------------> InferenceService\r\n\r\n```\r\n\r\n\r\n\r\n\r\n```python\r\ndef has_ce_headers(headers: typing.Dict[str, str]) -> bool:\r\n return (\r\n \"ce-specversion\" in headers\r\n and \"ce-source\" in headers\r\n and \"ce-type\" in headers\r\n and \"ce-id\" in headers\r\n )\r\n \r\n\r\n if has_ce_headers(self.request.headers):\r\n event = from_http(self.request.headers, self.request.body)\r\n body = event['data']\r\n else:\r\n try:\r\n body = json.loads(self.request.body)\r\n except json.decoder.JSONDecodeError as e:\r\n raise tornado.web.HTTPError(\r\n status_code=HTTPStatus.BAD_REQUEST,\r\n reason=\"Unrecognized request format: %s\" % e)\r\n ``` \r\n \r\n- Currently HTTP headers are not exposed to the `preprocess` or `predict` handler, if a cloud event is sent to KFServer it is not possible for user to check the http headers which can be useful for tracing and examining the original event source(kafka or http).\r\n`async def predict(self, request: Dict, headers: Dict) -> Dict:`\r\n`async def preprocess(self, request: Dict, headers: Dict) -> Dict:`\r\n\r\n- Also it would be nice if KFServing can generate cloud event response so that it works with knative eventing channels seamlessly.\r\n\r\n**Anything else you would like to add:**\r\n[Miscellaneous information that will assist in solving the issue.]\r\n\n", "before_files": [{"content": "# Copyright 2020 kubeflow.org.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nimport inspect\nimport tornado.web\nimport json\nfrom http import HTTPStatus\nfrom kfserving.kfmodel_repository import KFModelRepository\n\n\nclass HTTPHandler(tornado.web.RequestHandler):\n def initialize(self, models: KFModelRepository):\n self.models = models # pylint:disable=attribute-defined-outside-init\n\n def get_model(self, name: str):\n model = self.models.get_model(name)\n if model is None:\n raise tornado.web.HTTPError(\n status_code=HTTPStatus.NOT_FOUND,\n reason=\"Model with name %s does not exist.\" % name\n )\n if not model.ready:\n model.load()\n return model\n\n def validate(self, request):\n if (\"instances\" in request and not isinstance(request[\"instances\"], list)) or \\\n (\"inputs\" in request and not isinstance(request[\"inputs\"], list)):\n raise tornado.web.HTTPError(\n status_code=HTTPStatus.BAD_REQUEST,\n reason=\"Expected \\\"instances\\\" or \\\"inputs\\\" to be a list\"\n )\n return request\n\n\nclass PredictHandler(HTTPHandler):\n async def post(self, name: str):\n model = self.get_model(name)\n try:\n body = json.loads(self.request.body)\n except json.decoder.JSONDecodeError as e:\n raise tornado.web.HTTPError(\n status_code=HTTPStatus.BAD_REQUEST,\n reason=\"Unrecognized request format: %s\" % e\n )\n request = model.preprocess(body)\n request = self.validate(request)\n response = (await model.predict(request)) if inspect.iscoroutinefunction(model.predict) else model.predict(request)\n response = model.postprocess(response)\n self.write(response)\n\n\nclass ExplainHandler(HTTPHandler):\n async def post(self, name: str):\n model = self.get_model(name)\n try:\n body = json.loads(self.request.body)\n except json.decoder.JSONDecodeError as e:\n raise tornado.web.HTTPError(\n status_code=HTTPStatus.BAD_REQUEST,\n reason=\"Unrecognized request format: %s\" % e\n )\n request = model.preprocess(body)\n request = self.validate(request)\n response = (await model.explain(request)) if inspect.iscoroutinefunction(model.explain) else model.explain(request)\n response = model.postprocess(response)\n self.write(response)\n", "path": "python/kfserving/kfserving/handlers/http.py"}, {"content": "# Copyright 2020 kubeflow.org.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom typing import Dict\nimport sys\n\nimport json\nimport tornado.web\nfrom tornado.httpclient import AsyncHTTPClient\n\nPREDICTOR_URL_FORMAT = \"http://{0}/v1/models/{1}:predict\"\nEXPLAINER_URL_FORMAT = \"http://{0}/v1/models/{1}:explain\"\nPREDICTOR_V2_URL_FORMAT = \"http://{0}/v2/models/{1}/infer\"\nEXPLAINER_V2_URL_FORMAT = \"http://{0}/v2/models/{1}/explain\"\n\n\n# KFModel is intended to be subclassed by various components within KFServing.\nclass KFModel:\n\n def __init__(self, name: str):\n self.name = name\n self.ready = False\n self.protocol = \"v1\"\n self.predictor_host = None\n self.explainer_host = None\n # The timeout matches what is set in generated Istio resources.\n # We generally don't want things to time out at the request level here,\n # timeouts should be handled elsewhere in the system.\n self.timeout = 600\n self._http_client_instance = None\n\n @property\n def _http_client(self):\n if self._http_client_instance is None:\n self._http_client_instance = AsyncHTTPClient(max_clients=sys.maxsize)\n return self._http_client_instance\n\n def load(self) -> bool:\n self.ready = True\n return self.ready\n\n def preprocess(self, request: Dict) -> Dict:\n return request\n\n def postprocess(self, request: Dict) -> Dict:\n return request\n\n async def predict(self, request: Dict) -> Dict:\n if not self.predictor_host:\n raise NotImplementedError\n predict_url = PREDICTOR_URL_FORMAT.format(self.predictor_host, self.name)\n if self.protocol == \"v2\":\n predict_url = PREDICTOR_V2_URL_FORMAT.format(self.predictor_host, self.name)\n response = await self._http_client.fetch(\n predict_url,\n method='POST',\n request_timeout=self.timeout,\n body=json.dumps(request)\n )\n if response.code != 200:\n raise tornado.web.HTTPError(\n status_code=response.code,\n reason=response.body)\n return json.loads(response.body)\n\n async def explain(self, request: Dict) -> Dict:\n if self.explainer_host is None:\n raise NotImplementedError\n explain_url = EXPLAINER_URL_FORMAT.format(self.predictor_host, self.name)\n if self.protocol == \"v2\":\n explain_url = EXPLAINER_V2_URL_FORMAT.format(self.predictor_host, self.name)\n response = await self._http_client.fetch(\n url=explain_url,\n method='POST',\n request_timeout=self.timeout,\n body=json.dumps(request)\n )\n if response.code != 200:\n raise tornado.web.HTTPError(\n status_code=response.code,\n reason=response.body)\n return json.loads(response.body)\n", "path": "python/kfserving/kfserving/kfmodel.py"}]}
2,899
971
gh_patches_debug_8100
rasdani/github-patches
git_diff
WeblateOrg__weblate-11568
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> Time to use `build` from `setuptools` instead of `distutils`? ### Describe the problem The following feature in setuptools has been deprecated for almost 2 years and is about to be removed: https://github.com/pypa/setuptools/blob/1ed759173983656734c3606e9c97a348895e5e0c/setuptools/command/build.py#L13-L27 It might be a good idea to import `build` directly from setuptools for the following code: https://github.com/WeblateOrg/weblate/blob/47f9f2870c4ed9fd5429eebfacc61d2267a5bb31/setup.py#L9 https://github.com/WeblateOrg/weblate/blob/47f9f2870c4ed9fd5429eebfacc61d2267a5bb31/setup.py#L51-L58 (`build` is available directly from setuptools, starting on version v62.4.0) ### Describe the solution you would like Whenever possible, it might be a good idea to import from setuptools (and minimise imports to `distutils` to the minimum viable). ### Describe alternatives you have considered _No response_ ### Screenshots _No response_ ### Additional context _No response_ Time to use `build` from `setuptools` instead of `distutils`? ### Describe the problem The following feature in setuptools has been deprecated for almost 2 years and is about to be removed: https://github.com/pypa/setuptools/blob/1ed759173983656734c3606e9c97a348895e5e0c/setuptools/command/build.py#L13-L27 It might be a good idea to import `build` directly from setuptools for the following code: https://github.com/WeblateOrg/weblate/blob/47f9f2870c4ed9fd5429eebfacc61d2267a5bb31/setup.py#L9 https://github.com/WeblateOrg/weblate/blob/47f9f2870c4ed9fd5429eebfacc61d2267a5bb31/setup.py#L51-L58 (`build` is available directly from setuptools, starting on version v62.4.0) ### Describe the solution you would like Whenever possible, it might be a good idea to import from setuptools (and minimise imports to `distutils` to the minimum viable). ### Describe alternatives you have considered _No response_ ### Screenshots _No response_ ### Additional context _No response_ </issue> <code> [start of setup.py] 1 #!/usr/bin/env python3 2 3 # Copyright © Michal Čihař <[email protected]> 4 # 5 # SPDX-License-Identifier: GPL-3.0-or-later 6 7 import os 8 from distutils import log 9 from distutils.command.build import build 10 from distutils.core import Command 11 from glob import glob 12 from itertools import chain 13 14 from setuptools import setup 15 from setuptools.command.build_py import build_py 16 from setuptools.modified import newer 17 from translate.tools.pocompile import convertmo 18 19 LOCALE_MASKS = [ 20 "weblate/locale/*/LC_MESSAGES/*.po", 21 ] 22 23 24 class WeblateBuildPy(build_py): 25 def find_package_modules(self, package, package_dir): 26 """Filter settings.py from built module.""" 27 result = super().find_package_modules(package, package_dir) 28 return [item for item in result if item[2] != "weblate/settings.py"] 29 30 31 class BuildMo(Command): 32 description = "update MO files to match PO" 33 user_options = [] 34 35 def initialize_options(self) -> None: 36 self.build_base = None 37 38 def finalize_options(self) -> None: 39 self.set_undefined_options("build", ("build_base", "build_base")) 40 41 def run(self) -> None: 42 for name in chain.from_iterable(glob(mask) for mask in LOCALE_MASKS): 43 output = os.path.splitext(name)[0] + ".mo" 44 if not newer(name, output): 45 continue 46 self.announce(f"compiling {name} -> {output}", level=log.INFO) 47 with open(name, "rb") as pofile, open(output, "wb") as mofile: 48 convertmo(pofile, mofile, None) 49 50 51 class WeblateBuild(build): 52 """Override the default build with new subcommands.""" 53 54 # The build_mo has to be before build_data 55 sub_commands = [ 56 ("build_mo", lambda self: True), # noqa: ARG005 57 *build.sub_commands, 58 ] 59 60 61 setup( 62 cmdclass={"build_py": WeblateBuildPy, "build_mo": BuildMo, "build": WeblateBuild}, 63 ) 64 [end of setup.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/setup.py b/setup.py --- a/setup.py +++ b/setup.py @@ -6,12 +6,12 @@ import os from distutils import log -from distutils.command.build import build from distutils.core import Command from glob import glob from itertools import chain from setuptools import setup +from setuptools.command.build import build from setuptools.command.build_py import build_py from setuptools.modified import newer from translate.tools.pocompile import convertmo
{"golden_diff": "diff --git a/setup.py b/setup.py\n--- a/setup.py\n+++ b/setup.py\n@@ -6,12 +6,12 @@\n \n import os\n from distutils import log\n-from distutils.command.build import build\n from distutils.core import Command\n from glob import glob\n from itertools import chain\n \n from setuptools import setup\n+from setuptools.command.build import build\n from setuptools.command.build_py import build_py\n from setuptools.modified import newer\n from translate.tools.pocompile import convertmo\n", "issue": "Time to use `build` from `setuptools` instead of `distutils`?\n### Describe the problem\n\nThe following feature in setuptools has been deprecated for almost 2 years and is about to be removed:\r\n\r\nhttps://github.com/pypa/setuptools/blob/1ed759173983656734c3606e9c97a348895e5e0c/setuptools/command/build.py#L13-L27\r\n\r\nIt might be a good idea to import `build` directly from setuptools for the following code:\r\n\r\nhttps://github.com/WeblateOrg/weblate/blob/47f9f2870c4ed9fd5429eebfacc61d2267a5bb31/setup.py#L9\r\nhttps://github.com/WeblateOrg/weblate/blob/47f9f2870c4ed9fd5429eebfacc61d2267a5bb31/setup.py#L51-L58\r\n\r\n(`build` is available directly from setuptools, starting on version v62.4.0)\n\n### Describe the solution you would like\n\nWhenever possible, it might be a good idea to import from setuptools (and minimise imports to `distutils` to the minimum viable).\n\n### Describe alternatives you have considered\n\n_No response_\n\n### Screenshots\n\n_No response_\n\n### Additional context\n\n_No response_\nTime to use `build` from `setuptools` instead of `distutils`?\n### Describe the problem\n\nThe following feature in setuptools has been deprecated for almost 2 years and is about to be removed:\r\n\r\nhttps://github.com/pypa/setuptools/blob/1ed759173983656734c3606e9c97a348895e5e0c/setuptools/command/build.py#L13-L27\r\n\r\nIt might be a good idea to import `build` directly from setuptools for the following code:\r\n\r\nhttps://github.com/WeblateOrg/weblate/blob/47f9f2870c4ed9fd5429eebfacc61d2267a5bb31/setup.py#L9\r\nhttps://github.com/WeblateOrg/weblate/blob/47f9f2870c4ed9fd5429eebfacc61d2267a5bb31/setup.py#L51-L58\r\n\r\n(`build` is available directly from setuptools, starting on version v62.4.0)\n\n### Describe the solution you would like\n\nWhenever possible, it might be a good idea to import from setuptools (and minimise imports to `distutils` to the minimum viable).\n\n### Describe alternatives you have considered\n\n_No response_\n\n### Screenshots\n\n_No response_\n\n### Additional context\n\n_No response_\n", "before_files": [{"content": "#!/usr/bin/env python3\n\n# Copyright \u00a9 Michal \u010ciha\u0159 <[email protected]>\n#\n# SPDX-License-Identifier: GPL-3.0-or-later\n\nimport os\nfrom distutils import log\nfrom distutils.command.build import build\nfrom distutils.core import Command\nfrom glob import glob\nfrom itertools import chain\n\nfrom setuptools import setup\nfrom setuptools.command.build_py import build_py\nfrom setuptools.modified import newer\nfrom translate.tools.pocompile import convertmo\n\nLOCALE_MASKS = [\n \"weblate/locale/*/LC_MESSAGES/*.po\",\n]\n\n\nclass WeblateBuildPy(build_py):\n def find_package_modules(self, package, package_dir):\n \"\"\"Filter settings.py from built module.\"\"\"\n result = super().find_package_modules(package, package_dir)\n return [item for item in result if item[2] != \"weblate/settings.py\"]\n\n\nclass BuildMo(Command):\n description = \"update MO files to match PO\"\n user_options = []\n\n def initialize_options(self) -> None:\n self.build_base = None\n\n def finalize_options(self) -> None:\n self.set_undefined_options(\"build\", (\"build_base\", \"build_base\"))\n\n def run(self) -> None:\n for name in chain.from_iterable(glob(mask) for mask in LOCALE_MASKS):\n output = os.path.splitext(name)[0] + \".mo\"\n if not newer(name, output):\n continue\n self.announce(f\"compiling {name} -> {output}\", level=log.INFO)\n with open(name, \"rb\") as pofile, open(output, \"wb\") as mofile:\n convertmo(pofile, mofile, None)\n\n\nclass WeblateBuild(build):\n \"\"\"Override the default build with new subcommands.\"\"\"\n\n # The build_mo has to be before build_data\n sub_commands = [\n (\"build_mo\", lambda self: True), # noqa: ARG005\n *build.sub_commands,\n ]\n\n\nsetup(\n cmdclass={\"build_py\": WeblateBuildPy, \"build_mo\": BuildMo, \"build\": WeblateBuild},\n)\n", "path": "setup.py"}]}
1,743
105
gh_patches_debug_8391
rasdani/github-patches
git_diff
scrapy__scrapy-3377
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> scrapy check exit code on exception When I run `scrapy check` and a contract raises unhandled exception, the command exits with successful code `0`. ``` $ scrapy check $SPIDER_NAME ... some stuff ... Ran 0 contracts in 0.000s OK $ echo $? 0 ``` Is this intended, or should I fix it? </issue> <code> [start of scrapy/contracts/__init__.py] 1 import sys 2 import re 3 from functools import wraps 4 from inspect import getmembers 5 from unittest import TestCase 6 7 from scrapy.http import Request 8 from scrapy.utils.spider import iterate_spider_output 9 from scrapy.utils.python import get_spec 10 11 12 class ContractsManager(object): 13 contracts = {} 14 15 def __init__(self, contracts): 16 for contract in contracts: 17 self.contracts[contract.name] = contract 18 19 def tested_methods_from_spidercls(self, spidercls): 20 methods = [] 21 for key, value in getmembers(spidercls): 22 if (callable(value) and value.__doc__ and 23 re.search(r'^\s*@', value.__doc__, re.MULTILINE)): 24 methods.append(key) 25 26 return methods 27 28 def extract_contracts(self, method): 29 contracts = [] 30 for line in method.__doc__.split('\n'): 31 line = line.strip() 32 33 if line.startswith('@'): 34 name, args = re.match(r'@(\w+)\s*(.*)', line).groups() 35 args = re.split(r'\s+', args) 36 37 contracts.append(self.contracts[name](method, *args)) 38 39 return contracts 40 41 def from_spider(self, spider, results): 42 requests = [] 43 for method in self.tested_methods_from_spidercls(type(spider)): 44 bound_method = spider.__getattribute__(method) 45 requests.append(self.from_method(bound_method, results)) 46 47 return requests 48 49 def from_method(self, method, results): 50 contracts = self.extract_contracts(method) 51 if contracts: 52 request_cls = Request 53 for contract in contracts: 54 if contract.request_cls is not None: 55 request_cls = contract.request_cls 56 57 # calculate request args 58 args, kwargs = get_spec(request_cls.__init__) 59 kwargs['callback'] = method 60 for contract in contracts: 61 kwargs = contract.adjust_request_args(kwargs) 62 63 args.remove('self') 64 65 # check if all positional arguments are defined in kwargs 66 if set(args).issubset(set(kwargs)): 67 request = request_cls(**kwargs) 68 69 # execute pre and post hooks in order 70 for contract in reversed(contracts): 71 request = contract.add_pre_hook(request, results) 72 for contract in contracts: 73 request = contract.add_post_hook(request, results) 74 75 self._clean_req(request, method, results) 76 return request 77 78 def _clean_req(self, request, method, results): 79 """ stop the request from returning objects and records any errors """ 80 81 cb = request.callback 82 83 @wraps(cb) 84 def cb_wrapper(response): 85 try: 86 output = cb(response) 87 output = list(iterate_spider_output(output)) 88 except: 89 case = _create_testcase(method, 'callback') 90 results.addError(case, sys.exc_info()) 91 92 def eb_wrapper(failure): 93 case = _create_testcase(method, 'errback') 94 exc_info = failure.type, failure.value, failure.getTracebackObject() 95 results.addError(case, exc_info) 96 97 request.callback = cb_wrapper 98 request.errback = eb_wrapper 99 100 101 class Contract(object): 102 """ Abstract class for contracts """ 103 request_cls = None 104 105 def __init__(self, method, *args): 106 self.testcase_pre = _create_testcase(method, '@%s pre-hook' % self.name) 107 self.testcase_post = _create_testcase(method, '@%s post-hook' % self.name) 108 self.args = args 109 110 def add_pre_hook(self, request, results): 111 if hasattr(self, 'pre_process'): 112 cb = request.callback 113 114 @wraps(cb) 115 def wrapper(response): 116 try: 117 results.startTest(self.testcase_pre) 118 self.pre_process(response) 119 results.stopTest(self.testcase_pre) 120 except AssertionError: 121 results.addFailure(self.testcase_pre, sys.exc_info()) 122 except Exception: 123 results.addError(self.testcase_pre, sys.exc_info()) 124 else: 125 results.addSuccess(self.testcase_pre) 126 finally: 127 return list(iterate_spider_output(cb(response))) 128 129 request.callback = wrapper 130 131 return request 132 133 def add_post_hook(self, request, results): 134 if hasattr(self, 'post_process'): 135 cb = request.callback 136 137 @wraps(cb) 138 def wrapper(response): 139 output = list(iterate_spider_output(cb(response))) 140 try: 141 results.startTest(self.testcase_post) 142 self.post_process(output) 143 results.stopTest(self.testcase_post) 144 except AssertionError: 145 results.addFailure(self.testcase_post, sys.exc_info()) 146 except Exception: 147 results.addError(self.testcase_post, sys.exc_info()) 148 else: 149 results.addSuccess(self.testcase_post) 150 finally: 151 return output 152 153 request.callback = wrapper 154 155 return request 156 157 def adjust_request_args(self, args): 158 return args 159 160 161 def _create_testcase(method, desc): 162 spider = method.__self__.name 163 164 class ContractTestCase(TestCase): 165 def __str__(_self): 166 return "[%s] %s (%s)" % (spider, method.__name__, desc) 167 168 name = '%s_%s' % (spider, method.__name__) 169 setattr(ContractTestCase, name, lambda x: x) 170 return ContractTestCase(name) 171 [end of scrapy/contracts/__init__.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/scrapy/contracts/__init__.py b/scrapy/contracts/__init__.py --- a/scrapy/contracts/__init__.py +++ b/scrapy/contracts/__init__.py @@ -42,7 +42,11 @@ requests = [] for method in self.tested_methods_from_spidercls(type(spider)): bound_method = spider.__getattribute__(method) - requests.append(self.from_method(bound_method, results)) + try: + requests.append(self.from_method(bound_method, results)) + except Exception: + case = _create_testcase(bound_method, 'contract') + results.addError(case, sys.exc_info()) return requests
{"golden_diff": "diff --git a/scrapy/contracts/__init__.py b/scrapy/contracts/__init__.py\n--- a/scrapy/contracts/__init__.py\n+++ b/scrapy/contracts/__init__.py\n@@ -42,7 +42,11 @@\n requests = []\n for method in self.tested_methods_from_spidercls(type(spider)):\n bound_method = spider.__getattribute__(method)\n- requests.append(self.from_method(bound_method, results))\n+ try:\n+ requests.append(self.from_method(bound_method, results))\n+ except Exception:\n+ case = _create_testcase(bound_method, 'contract')\n+ results.addError(case, sys.exc_info())\n \n return requests\n", "issue": "scrapy check exit code on exception\nWhen I run `scrapy check` and a contract raises unhandled exception, the command exits with successful code `0`.\r\n\r\n```\r\n$ scrapy check $SPIDER_NAME\r\n... some stuff ...\r\nRan 0 contracts in 0.000s\r\nOK\r\n$ echo $?\r\n0\r\n```\r\n\r\nIs this intended, or should I fix it?\n", "before_files": [{"content": "import sys\nimport re\nfrom functools import wraps\nfrom inspect import getmembers\nfrom unittest import TestCase\n\nfrom scrapy.http import Request\nfrom scrapy.utils.spider import iterate_spider_output\nfrom scrapy.utils.python import get_spec\n\n\nclass ContractsManager(object):\n contracts = {}\n\n def __init__(self, contracts):\n for contract in contracts:\n self.contracts[contract.name] = contract\n\n def tested_methods_from_spidercls(self, spidercls):\n methods = []\n for key, value in getmembers(spidercls):\n if (callable(value) and value.__doc__ and\n re.search(r'^\\s*@', value.__doc__, re.MULTILINE)):\n methods.append(key)\n\n return methods\n\n def extract_contracts(self, method):\n contracts = []\n for line in method.__doc__.split('\\n'):\n line = line.strip()\n\n if line.startswith('@'):\n name, args = re.match(r'@(\\w+)\\s*(.*)', line).groups()\n args = re.split(r'\\s+', args)\n\n contracts.append(self.contracts[name](method, *args))\n\n return contracts\n\n def from_spider(self, spider, results):\n requests = []\n for method in self.tested_methods_from_spidercls(type(spider)):\n bound_method = spider.__getattribute__(method)\n requests.append(self.from_method(bound_method, results))\n\n return requests\n\n def from_method(self, method, results):\n contracts = self.extract_contracts(method)\n if contracts:\n request_cls = Request\n for contract in contracts:\n if contract.request_cls is not None:\n request_cls = contract.request_cls\n\n # calculate request args\n args, kwargs = get_spec(request_cls.__init__)\n kwargs['callback'] = method\n for contract in contracts:\n kwargs = contract.adjust_request_args(kwargs)\n\n args.remove('self')\n\n # check if all positional arguments are defined in kwargs\n if set(args).issubset(set(kwargs)):\n request = request_cls(**kwargs)\n\n # execute pre and post hooks in order\n for contract in reversed(contracts):\n request = contract.add_pre_hook(request, results)\n for contract in contracts:\n request = contract.add_post_hook(request, results)\n\n self._clean_req(request, method, results)\n return request\n\n def _clean_req(self, request, method, results):\n \"\"\" stop the request from returning objects and records any errors \"\"\"\n\n cb = request.callback\n\n @wraps(cb)\n def cb_wrapper(response):\n try:\n output = cb(response)\n output = list(iterate_spider_output(output))\n except:\n case = _create_testcase(method, 'callback')\n results.addError(case, sys.exc_info())\n\n def eb_wrapper(failure):\n case = _create_testcase(method, 'errback')\n exc_info = failure.type, failure.value, failure.getTracebackObject()\n results.addError(case, exc_info)\n\n request.callback = cb_wrapper\n request.errback = eb_wrapper\n\n\nclass Contract(object):\n \"\"\" Abstract class for contracts \"\"\"\n request_cls = None\n\n def __init__(self, method, *args):\n self.testcase_pre = _create_testcase(method, '@%s pre-hook' % self.name)\n self.testcase_post = _create_testcase(method, '@%s post-hook' % self.name)\n self.args = args\n\n def add_pre_hook(self, request, results):\n if hasattr(self, 'pre_process'):\n cb = request.callback\n\n @wraps(cb)\n def wrapper(response):\n try:\n results.startTest(self.testcase_pre)\n self.pre_process(response)\n results.stopTest(self.testcase_pre)\n except AssertionError:\n results.addFailure(self.testcase_pre, sys.exc_info())\n except Exception:\n results.addError(self.testcase_pre, sys.exc_info())\n else:\n results.addSuccess(self.testcase_pre)\n finally:\n return list(iterate_spider_output(cb(response)))\n\n request.callback = wrapper\n\n return request\n\n def add_post_hook(self, request, results):\n if hasattr(self, 'post_process'):\n cb = request.callback\n\n @wraps(cb)\n def wrapper(response):\n output = list(iterate_spider_output(cb(response)))\n try:\n results.startTest(self.testcase_post)\n self.post_process(output)\n results.stopTest(self.testcase_post)\n except AssertionError:\n results.addFailure(self.testcase_post, sys.exc_info())\n except Exception:\n results.addError(self.testcase_post, sys.exc_info())\n else:\n results.addSuccess(self.testcase_post)\n finally:\n return output\n\n request.callback = wrapper\n\n return request\n\n def adjust_request_args(self, args):\n return args\n\n\ndef _create_testcase(method, desc):\n spider = method.__self__.name\n\n class ContractTestCase(TestCase):\n def __str__(_self):\n return \"[%s] %s (%s)\" % (spider, method.__name__, desc)\n\n name = '%s_%s' % (spider, method.__name__)\n setattr(ContractTestCase, name, lambda x: x)\n return ContractTestCase(name)\n", "path": "scrapy/contracts/__init__.py"}]}
2,153
154
gh_patches_debug_7395
rasdani/github-patches
git_diff
svthalia__concrexit-1513
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> URL's from Events API are not always full URL's ### Describe the bug The `google_maps_url` from the Events API is not always a full URL. ### How to reproduce Steps to reproduce the behaviour: 1. On staging, go to `https://staging.thalia.nu/api/v1/events/147/` and see that the `google_maps_url` is not a full URL but a relative one. ### Expected behaviour The `google_maps_url` should always be a full URL. </issue> <code> [start of website/utils/snippets.py] 1 """Provides various utilities that are useful across the project.""" 2 import datetime 3 import hmac 4 from base64 import urlsafe_b64decode, urlsafe_b64encode 5 from collections import namedtuple 6 from _sha1 import sha1 7 8 from django.conf import settings 9 from django.template.defaultfilters import urlencode 10 from django.templatetags.static import static 11 from django.utils import timezone, dateparse 12 from rest_framework.exceptions import ParseError 13 14 15 def dict2obj(d, name="Object"): 16 return namedtuple(name, d.keys())(*d.values()) 17 18 19 def datetime_to_lectureyear(date): 20 """Convert a :class:`~datetime.date` to the start of the lectureyear. 21 22 >>> from datetime import date, datetime, timezone 23 >>> nov_23 = date(1990, 11, 7) 24 >>> datetime_to_lectureyear(nov_23) 25 1990 26 >>> mar_2 = date(1993, 3, 2) 27 >>> datetime_to_lectureyear(mar_2) 28 1992 29 30 Also works on :class:`~datetime.datetime`, but they need to be tz-aware: 31 32 >>> new_year = datetime(2000, 1, 1, tzinfo=timezone.utc) 33 >>> datetime_to_lectureyear(new_year) 34 1999 35 """ 36 if isinstance(date, timezone.datetime): 37 date = timezone.localtime(date).date() 38 sept_1 = timezone.make_aware(timezone.datetime(date.year, 9, 1)) 39 if date < sept_1.date(): 40 return date.year - 1 41 return date.year 42 43 44 def create_google_maps_url(location, zoom, size): 45 if location.lower().strip() == "online": 46 return static("img/locations/online.png") 47 if location.lower().strip() == "discord": 48 return static("img/locations/discord.png") 49 50 maps_url = ( 51 f"/maps/api/staticmap?" 52 f"center={ urlencode(location) }&" 53 f"zoom={ zoom }&size={ size }&" 54 f"markers={ urlencode(location) }&" 55 f"key={ settings.GOOGLE_MAPS_API_KEY }" 56 ) 57 58 decoded_key = urlsafe_b64decode(settings.GOOGLE_MAPS_API_SECRET) 59 60 signature = hmac.new(decoded_key, maps_url.encode(), sha1) 61 62 encoded_signature = urlsafe_b64encode(signature.digest()) 63 64 maps_url += f"&signature={encoded_signature.decode('utf-8')}" 65 66 return "https://maps.googleapis.com" + maps_url 67 68 69 def _extract_date(param): 70 """Extract the date from an arbitrary string.""" 71 if param is None: 72 return None 73 try: 74 return dateparse.parse_datetime(param) 75 except ValueError: 76 return dateparse.parse_date(param) 77 78 79 def extract_date_range(request, allow_empty=False): 80 """Extract a date range from an arbitrary string.""" 81 default_value = None 82 83 start = request.query_params.get("start", default_value) 84 if start or not allow_empty: 85 try: 86 start = dateparse.parse_datetime(start) 87 if not timezone.is_aware(start): 88 start = timezone.make_aware(start) 89 except (ValueError, AttributeError, TypeError) as e: 90 raise ParseError(detail="start query parameter invalid") from e 91 92 end = request.query_params.get("end", default_value) 93 if end or not allow_empty: 94 try: 95 end = dateparse.parse_datetime(end) 96 if not timezone.is_aware(end): 97 end = timezone.make_aware(end) 98 except (ValueError, AttributeError, TypeError) as e: 99 raise ParseError(detail="end query parameter invalid") from e 100 101 return start, end 102 103 104 def overlaps(check, others, can_equal=True): 105 """Check for overlapping date ranges. 106 107 This works by checking the maximum of the two `since` times, and the minimum of 108 the two `until` times. Because there are no infinite dates, the value date_max 109 is created for when the `until` value is None; this signifies a timespan that 110 has not ended yet and is the maximum possible date in Python's datetime. 111 112 The ranges overlap when the maximum start time is smaller than the minimum 113 end time, as can be seen in this example of two integer ranges: 114 115 check: . . . .[4]. . . . 9 116 other: . . 2 . .[5]. . . . 117 118 check: . . . .[4]. . . . 9 119 other: . . 2 . . . . . . . [date_max] 120 121 And when non overlapping: 122 check: . . . . . .[6] . . 9 123 other: . . 2 . .[5]. . . . 124 125 4 < 5 == True so these intervals overlap, while 6 < 5 == False so these intervals 126 don't overlap 127 128 The can_equal argument is used for boards, where the end date can't be the same 129 as the start date. 130 131 >>> overlaps( \ 132 dict2obj({ \ 133 'pk': 1 \ 134 , 'since': datetime.date(2018, 12, 1) \ 135 , 'until': datetime.date(2019, 1, 1) \ 136 }) \ 137 , [dict2obj({ \ 138 'pk': 2 \ 139 , 'since': datetime.date(2019, 1, 1) \ 140 , 'until': datetime.date(2019, 1, 31) \ 141 })]) 142 False 143 144 >>> overlaps( \ 145 dict2obj({ \ 146 'pk': 1 \ 147 , 'since': datetime.date(2018, 12, 1) \ 148 , 'until': datetime.date(2019, 1, 1) \ 149 }) \ 150 , [dict2obj({ \ 151 'pk': 2 \ 152 , 'since': datetime.date(2019, 1, 1) \ 153 , 'until': datetime.date(2019, 1, 31) \ 154 })], False) 155 True 156 157 >>> overlaps( \ 158 dict2obj({ \ 159 'pk': 1 \ 160 , 'since': datetime.date(2018, 12, 1) \ 161 , 'until': datetime.date(2019, 1, 2) \ 162 }) \ 163 , [dict2obj({ \ 164 'pk': 2 \ 165 , 'since': datetime.date(2019, 1, 1) \ 166 , 'until': datetime.date(2019, 1, 31) \ 167 })]) 168 True 169 """ 170 date_max = datetime.date(datetime.MAXYEAR, 12, 31) 171 for other in others: 172 if check.pk == other.pk: 173 # No checks for the object we're validating 174 continue 175 176 max_start = max(check.since, other.since) 177 min_end = min(check.until or date_max, other.until or date_max) 178 179 if max_start == min_end and not can_equal: 180 return True 181 if max_start < min_end: 182 return True 183 184 return False 185 [end of website/utils/snippets.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/website/utils/snippets.py b/website/utils/snippets.py --- a/website/utils/snippets.py +++ b/website/utils/snippets.py @@ -43,9 +43,9 @@ def create_google_maps_url(location, zoom, size): if location.lower().strip() == "online": - return static("img/locations/online.png") + return settings.BASE_URL + static("img/locations/online.png") if location.lower().strip() == "discord": - return static("img/locations/discord.png") + return settings.BASE_URL + static("img/locations/discord.png") maps_url = ( f"/maps/api/staticmap?"
{"golden_diff": "diff --git a/website/utils/snippets.py b/website/utils/snippets.py\n--- a/website/utils/snippets.py\n+++ b/website/utils/snippets.py\n@@ -43,9 +43,9 @@\n \n def create_google_maps_url(location, zoom, size):\n if location.lower().strip() == \"online\":\n- return static(\"img/locations/online.png\")\n+ return settings.BASE_URL + static(\"img/locations/online.png\")\n if location.lower().strip() == \"discord\":\n- return static(\"img/locations/discord.png\")\n+ return settings.BASE_URL + static(\"img/locations/discord.png\")\n \n maps_url = (\n f\"/maps/api/staticmap?\"\n", "issue": "URL's from Events API are not always full URL's\n### Describe the bug\r\nThe `google_maps_url` from the Events API is not always a full URL.\r\n\r\n### How to reproduce\r\nSteps to reproduce the behaviour:\r\n1. On staging, go to `https://staging.thalia.nu/api/v1/events/147/` and see that the `google_maps_url` is not a full URL but a relative one.\r\n\r\n### Expected behaviour\r\nThe `google_maps_url` should always be a full URL.\r\n\r\n\n", "before_files": [{"content": "\"\"\"Provides various utilities that are useful across the project.\"\"\"\nimport datetime\nimport hmac\nfrom base64 import urlsafe_b64decode, urlsafe_b64encode\nfrom collections import namedtuple\nfrom _sha1 import sha1\n\nfrom django.conf import settings\nfrom django.template.defaultfilters import urlencode\nfrom django.templatetags.static import static\nfrom django.utils import timezone, dateparse\nfrom rest_framework.exceptions import ParseError\n\n\ndef dict2obj(d, name=\"Object\"):\n return namedtuple(name, d.keys())(*d.values())\n\n\ndef datetime_to_lectureyear(date):\n \"\"\"Convert a :class:`~datetime.date` to the start of the lectureyear.\n\n >>> from datetime import date, datetime, timezone\n >>> nov_23 = date(1990, 11, 7)\n >>> datetime_to_lectureyear(nov_23)\n 1990\n >>> mar_2 = date(1993, 3, 2)\n >>> datetime_to_lectureyear(mar_2)\n 1992\n\n Also works on :class:`~datetime.datetime`, but they need to be tz-aware:\n\n >>> new_year = datetime(2000, 1, 1, tzinfo=timezone.utc)\n >>> datetime_to_lectureyear(new_year)\n 1999\n \"\"\"\n if isinstance(date, timezone.datetime):\n date = timezone.localtime(date).date()\n sept_1 = timezone.make_aware(timezone.datetime(date.year, 9, 1))\n if date < sept_1.date():\n return date.year - 1\n return date.year\n\n\ndef create_google_maps_url(location, zoom, size):\n if location.lower().strip() == \"online\":\n return static(\"img/locations/online.png\")\n if location.lower().strip() == \"discord\":\n return static(\"img/locations/discord.png\")\n\n maps_url = (\n f\"/maps/api/staticmap?\"\n f\"center={ urlencode(location) }&\"\n f\"zoom={ zoom }&size={ size }&\"\n f\"markers={ urlencode(location) }&\"\n f\"key={ settings.GOOGLE_MAPS_API_KEY }\"\n )\n\n decoded_key = urlsafe_b64decode(settings.GOOGLE_MAPS_API_SECRET)\n\n signature = hmac.new(decoded_key, maps_url.encode(), sha1)\n\n encoded_signature = urlsafe_b64encode(signature.digest())\n\n maps_url += f\"&signature={encoded_signature.decode('utf-8')}\"\n\n return \"https://maps.googleapis.com\" + maps_url\n\n\ndef _extract_date(param):\n \"\"\"Extract the date from an arbitrary string.\"\"\"\n if param is None:\n return None\n try:\n return dateparse.parse_datetime(param)\n except ValueError:\n return dateparse.parse_date(param)\n\n\ndef extract_date_range(request, allow_empty=False):\n \"\"\"Extract a date range from an arbitrary string.\"\"\"\n default_value = None\n\n start = request.query_params.get(\"start\", default_value)\n if start or not allow_empty:\n try:\n start = dateparse.parse_datetime(start)\n if not timezone.is_aware(start):\n start = timezone.make_aware(start)\n except (ValueError, AttributeError, TypeError) as e:\n raise ParseError(detail=\"start query parameter invalid\") from e\n\n end = request.query_params.get(\"end\", default_value)\n if end or not allow_empty:\n try:\n end = dateparse.parse_datetime(end)\n if not timezone.is_aware(end):\n end = timezone.make_aware(end)\n except (ValueError, AttributeError, TypeError) as e:\n raise ParseError(detail=\"end query parameter invalid\") from e\n\n return start, end\n\n\ndef overlaps(check, others, can_equal=True):\n \"\"\"Check for overlapping date ranges.\n\n This works by checking the maximum of the two `since` times, and the minimum of\n the two `until` times. Because there are no infinite dates, the value date_max\n is created for when the `until` value is None; this signifies a timespan that\n has not ended yet and is the maximum possible date in Python's datetime.\n\n The ranges overlap when the maximum start time is smaller than the minimum\n end time, as can be seen in this example of two integer ranges:\n\n check: . . . .[4]. . . . 9\n other: . . 2 . .[5]. . . .\n\n check: . . . .[4]. . . . 9\n other: . . 2 . . . . . . . [date_max]\n\n And when non overlapping:\n check: . . . . . .[6] . . 9\n other: . . 2 . .[5]. . . .\n\n 4 < 5 == True so these intervals overlap, while 6 < 5 == False so these intervals\n don't overlap\n\n The can_equal argument is used for boards, where the end date can't be the same\n as the start date.\n\n >>> overlaps( \\\n dict2obj({ \\\n 'pk': 1 \\\n , 'since': datetime.date(2018, 12, 1) \\\n , 'until': datetime.date(2019, 1, 1) \\\n }) \\\n , [dict2obj({ \\\n 'pk': 2 \\\n , 'since': datetime.date(2019, 1, 1) \\\n , 'until': datetime.date(2019, 1, 31) \\\n })])\n False\n\n >>> overlaps( \\\n dict2obj({ \\\n 'pk': 1 \\\n , 'since': datetime.date(2018, 12, 1) \\\n , 'until': datetime.date(2019, 1, 1) \\\n }) \\\n , [dict2obj({ \\\n 'pk': 2 \\\n , 'since': datetime.date(2019, 1, 1) \\\n , 'until': datetime.date(2019, 1, 31) \\\n })], False)\n True\n\n >>> overlaps( \\\n dict2obj({ \\\n 'pk': 1 \\\n , 'since': datetime.date(2018, 12, 1) \\\n , 'until': datetime.date(2019, 1, 2) \\\n }) \\\n , [dict2obj({ \\\n 'pk': 2 \\\n , 'since': datetime.date(2019, 1, 1) \\\n , 'until': datetime.date(2019, 1, 31) \\\n })])\n True\n \"\"\"\n date_max = datetime.date(datetime.MAXYEAR, 12, 31)\n for other in others:\n if check.pk == other.pk:\n # No checks for the object we're validating\n continue\n\n max_start = max(check.since, other.since)\n min_end = min(check.until or date_max, other.until or date_max)\n\n if max_start == min_end and not can_equal:\n return True\n if max_start < min_end:\n return True\n\n return False\n", "path": "website/utils/snippets.py"}]}
2,681
152
gh_patches_debug_16724
rasdani/github-patches
git_diff
jazzband__pip-tools-828
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> log which python version to use/was used eg add this to the header: ``` # This file is autogenerated by pip-compile on: # Python 2.7.15+ (default, Oct 2 2018, 22:12:08) # [GCC 8.2.0] on linux2 # To update, run: # # pip-compile --generate-hashes --output-file=requirements.txt requirements.in ``` </issue> <code> [start of piptools/writer.py] 1 import os 2 import re 3 from itertools import chain 4 from typing import BinaryIO, Dict, Iterable, Iterator, List, Optional, Set, Tuple 5 6 from click import unstyle 7 from click.core import Context 8 from pip._internal.models.format_control import FormatControl 9 from pip._internal.req.req_install import InstallRequirement 10 from pip._vendor.packaging.markers import Marker 11 12 from .logging import log 13 from .utils import ( 14 UNSAFE_PACKAGES, 15 comment, 16 dedup, 17 format_requirement, 18 get_compile_command, 19 key_from_ireq, 20 ) 21 22 MESSAGE_UNHASHED_PACKAGE = comment( 23 "# WARNING: pip install will require the following package to be hashed." 24 "\n# Consider using a hashable URL like " 25 "https://github.com/jazzband/pip-tools/archive/SOMECOMMIT.zip" 26 ) 27 28 MESSAGE_UNSAFE_PACKAGES_UNPINNED = comment( 29 "# WARNING: The following packages were not pinned, but pip requires them to be" 30 "\n# pinned when the requirements file includes hashes. " 31 "Consider using the --allow-unsafe flag." 32 ) 33 34 MESSAGE_UNSAFE_PACKAGES = comment( 35 "# The following packages are considered to be unsafe in a requirements file:" 36 ) 37 38 MESSAGE_UNINSTALLABLE = ( 39 "The generated requirements file may be rejected by pip install. " 40 "See # WARNING lines for details." 41 ) 42 43 44 strip_comes_from_line_re = re.compile(r" \(line \d+\)$") 45 46 47 def _comes_from_as_string(ireq: InstallRequirement) -> str: 48 if isinstance(ireq.comes_from, str): 49 return strip_comes_from_line_re.sub("", ireq.comes_from) 50 return key_from_ireq(ireq.comes_from) 51 52 53 class OutputWriter: 54 def __init__( 55 self, 56 dst_file: BinaryIO, 57 click_ctx: Context, 58 dry_run: bool, 59 emit_header: bool, 60 emit_index_url: bool, 61 emit_trusted_host: bool, 62 annotate: bool, 63 generate_hashes: bool, 64 default_index_url: str, 65 index_urls: Iterable[str], 66 trusted_hosts: Iterable[str], 67 format_control: FormatControl, 68 allow_unsafe: bool, 69 find_links: List[str], 70 emit_find_links: bool, 71 ) -> None: 72 self.dst_file = dst_file 73 self.click_ctx = click_ctx 74 self.dry_run = dry_run 75 self.emit_header = emit_header 76 self.emit_index_url = emit_index_url 77 self.emit_trusted_host = emit_trusted_host 78 self.annotate = annotate 79 self.generate_hashes = generate_hashes 80 self.default_index_url = default_index_url 81 self.index_urls = index_urls 82 self.trusted_hosts = trusted_hosts 83 self.format_control = format_control 84 self.allow_unsafe = allow_unsafe 85 self.find_links = find_links 86 self.emit_find_links = emit_find_links 87 88 def _sort_key(self, ireq: InstallRequirement) -> Tuple[bool, str]: 89 return (not ireq.editable, str(ireq.req).lower()) 90 91 def write_header(self) -> Iterator[str]: 92 if self.emit_header: 93 yield comment("#") 94 yield comment("# This file is autogenerated by pip-compile") 95 yield comment("# To update, run:") 96 yield comment("#") 97 compile_command = os.environ.get( 98 "CUSTOM_COMPILE_COMMAND" 99 ) or get_compile_command(self.click_ctx) 100 yield comment(f"# {compile_command}") 101 yield comment("#") 102 103 def write_index_options(self) -> Iterator[str]: 104 if self.emit_index_url: 105 for index, index_url in enumerate(dedup(self.index_urls)): 106 if index == 0 and index_url.rstrip("/") == self.default_index_url: 107 continue 108 flag = "--index-url" if index == 0 else "--extra-index-url" 109 yield f"{flag} {index_url}" 110 111 def write_trusted_hosts(self) -> Iterator[str]: 112 if self.emit_trusted_host: 113 for trusted_host in dedup(self.trusted_hosts): 114 yield f"--trusted-host {trusted_host}" 115 116 def write_format_controls(self) -> Iterator[str]: 117 for nb in dedup(sorted(self.format_control.no_binary)): 118 yield f"--no-binary {nb}" 119 for ob in dedup(sorted(self.format_control.only_binary)): 120 yield f"--only-binary {ob}" 121 122 def write_find_links(self) -> Iterator[str]: 123 if self.emit_find_links: 124 for find_link in dedup(self.find_links): 125 yield f"--find-links {find_link}" 126 127 def write_flags(self) -> Iterator[str]: 128 emitted = False 129 for line in chain( 130 self.write_index_options(), 131 self.write_find_links(), 132 self.write_trusted_hosts(), 133 self.write_format_controls(), 134 ): 135 emitted = True 136 yield line 137 if emitted: 138 yield "" 139 140 def _iter_lines( 141 self, 142 results: Set[InstallRequirement], 143 unsafe_requirements: Optional[Set[InstallRequirement]] = None, 144 markers: Optional[Dict[str, Marker]] = None, 145 hashes: Optional[Dict[InstallRequirement, Set[str]]] = None, 146 ) -> Iterator[str]: 147 # default values 148 unsafe_requirements = unsafe_requirements or set() 149 markers = markers or {} 150 hashes = hashes or {} 151 152 # Check for unhashed or unpinned packages if at least one package does have 153 # hashes, which will trigger pip install's --require-hashes mode. 154 warn_uninstallable = False 155 has_hashes = hashes and any(hash for hash in hashes.values()) 156 157 yielded = False 158 159 for line in self.write_header(): 160 yield line 161 yielded = True 162 for line in self.write_flags(): 163 yield line 164 yielded = True 165 166 unsafe_requirements = ( 167 {r for r in results if r.name in UNSAFE_PACKAGES} 168 if not unsafe_requirements 169 else unsafe_requirements 170 ) 171 packages = {r for r in results if r.name not in UNSAFE_PACKAGES} 172 173 if packages: 174 for ireq in sorted(packages, key=self._sort_key): 175 if has_hashes and not hashes.get(ireq): 176 yield MESSAGE_UNHASHED_PACKAGE 177 warn_uninstallable = True 178 line = self._format_requirement( 179 ireq, markers.get(key_from_ireq(ireq)), hashes=hashes 180 ) 181 yield line 182 yielded = True 183 184 if unsafe_requirements: 185 yield "" 186 yielded = True 187 if has_hashes and not self.allow_unsafe: 188 yield MESSAGE_UNSAFE_PACKAGES_UNPINNED 189 warn_uninstallable = True 190 else: 191 yield MESSAGE_UNSAFE_PACKAGES 192 193 for ireq in sorted(unsafe_requirements, key=self._sort_key): 194 ireq_key = key_from_ireq(ireq) 195 if not self.allow_unsafe: 196 yield comment(f"# {ireq_key}") 197 else: 198 line = self._format_requirement( 199 ireq, marker=markers.get(ireq_key), hashes=hashes 200 ) 201 yield line 202 203 # Yield even when there's no real content, so that blank files are written 204 if not yielded: 205 yield "" 206 207 if warn_uninstallable: 208 log.warning(MESSAGE_UNINSTALLABLE) 209 210 def write( 211 self, 212 results: Set[InstallRequirement], 213 unsafe_requirements: Set[InstallRequirement], 214 markers: Dict[str, Marker], 215 hashes: Optional[Dict[InstallRequirement, Set[str]]], 216 ) -> None: 217 218 for line in self._iter_lines(results, unsafe_requirements, markers, hashes): 219 log.info(line) 220 if not self.dry_run: 221 self.dst_file.write(unstyle(line).encode()) 222 self.dst_file.write(os.linesep.encode()) 223 224 def _format_requirement( 225 self, 226 ireq: InstallRequirement, 227 marker: Optional[Marker] = None, 228 hashes: Optional[Dict[InstallRequirement, Set[str]]] = None, 229 ) -> str: 230 ireq_hashes = (hashes if hashes is not None else {}).get(ireq) 231 232 line = format_requirement(ireq, marker=marker, hashes=ireq_hashes) 233 234 if not self.annotate: 235 return line 236 237 # Annotate what packages or reqs-ins this package is required by 238 required_by = set() 239 if hasattr(ireq, "_source_ireqs"): 240 required_by |= { 241 _comes_from_as_string(src_ireq) 242 for src_ireq in ireq._source_ireqs 243 if src_ireq.comes_from 244 } 245 elif ireq.comes_from: 246 required_by.add(_comes_from_as_string(ireq)) 247 248 if required_by: 249 sorted_required_by = sorted(required_by) 250 if len(sorted_required_by) == 1: 251 source = sorted_required_by[0] 252 annotation = " # via " + source 253 else: 254 annotation_lines = [" # via"] 255 for source in sorted_required_by: 256 annotation_lines.append(" # " + source) 257 annotation = "\n".join(annotation_lines) 258 line = f"{line}\n{comment(annotation)}" 259 260 return line 261 [end of piptools/writer.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/piptools/writer.py b/piptools/writer.py --- a/piptools/writer.py +++ b/piptools/writer.py @@ -1,5 +1,6 @@ import os import re +import sys from itertools import chain from typing import BinaryIO, Dict, Iterable, Iterator, List, Optional, Set, Tuple @@ -91,7 +92,10 @@ def write_header(self) -> Iterator[str]: if self.emit_header: yield comment("#") - yield comment("# This file is autogenerated by pip-compile") + yield comment( + "# This file is autogenerated by pip-compile with python " + f"{sys.version_info.major}.{sys.version_info.minor}" + ) yield comment("# To update, run:") yield comment("#") compile_command = os.environ.get(
{"golden_diff": "diff --git a/piptools/writer.py b/piptools/writer.py\n--- a/piptools/writer.py\n+++ b/piptools/writer.py\n@@ -1,5 +1,6 @@\n import os\n import re\n+import sys\n from itertools import chain\n from typing import BinaryIO, Dict, Iterable, Iterator, List, Optional, Set, Tuple\n \n@@ -91,7 +92,10 @@\n def write_header(self) -> Iterator[str]:\n if self.emit_header:\n yield comment(\"#\")\n- yield comment(\"# This file is autogenerated by pip-compile\")\n+ yield comment(\n+ \"# This file is autogenerated by pip-compile with python \"\n+ f\"{sys.version_info.major}.{sys.version_info.minor}\"\n+ )\n yield comment(\"# To update, run:\")\n yield comment(\"#\")\n compile_command = os.environ.get(\n", "issue": "log which python version to use/was used\neg add this to the header:\r\n\r\n```\r\n # This file is autogenerated by pip-compile on:\r\n # Python 2.7.15+ (default, Oct 2 2018, 22:12:08) \r\n # [GCC 8.2.0] on linux2\r\n # To update, run:\r\n #\r\n # pip-compile --generate-hashes --output-file=requirements.txt requirements.in\r\n```\n", "before_files": [{"content": "import os\nimport re\nfrom itertools import chain\nfrom typing import BinaryIO, Dict, Iterable, Iterator, List, Optional, Set, Tuple\n\nfrom click import unstyle\nfrom click.core import Context\nfrom pip._internal.models.format_control import FormatControl\nfrom pip._internal.req.req_install import InstallRequirement\nfrom pip._vendor.packaging.markers import Marker\n\nfrom .logging import log\nfrom .utils import (\n UNSAFE_PACKAGES,\n comment,\n dedup,\n format_requirement,\n get_compile_command,\n key_from_ireq,\n)\n\nMESSAGE_UNHASHED_PACKAGE = comment(\n \"# WARNING: pip install will require the following package to be hashed.\"\n \"\\n# Consider using a hashable URL like \"\n \"https://github.com/jazzband/pip-tools/archive/SOMECOMMIT.zip\"\n)\n\nMESSAGE_UNSAFE_PACKAGES_UNPINNED = comment(\n \"# WARNING: The following packages were not pinned, but pip requires them to be\"\n \"\\n# pinned when the requirements file includes hashes. \"\n \"Consider using the --allow-unsafe flag.\"\n)\n\nMESSAGE_UNSAFE_PACKAGES = comment(\n \"# The following packages are considered to be unsafe in a requirements file:\"\n)\n\nMESSAGE_UNINSTALLABLE = (\n \"The generated requirements file may be rejected by pip install. \"\n \"See # WARNING lines for details.\"\n)\n\n\nstrip_comes_from_line_re = re.compile(r\" \\(line \\d+\\)$\")\n\n\ndef _comes_from_as_string(ireq: InstallRequirement) -> str:\n if isinstance(ireq.comes_from, str):\n return strip_comes_from_line_re.sub(\"\", ireq.comes_from)\n return key_from_ireq(ireq.comes_from)\n\n\nclass OutputWriter:\n def __init__(\n self,\n dst_file: BinaryIO,\n click_ctx: Context,\n dry_run: bool,\n emit_header: bool,\n emit_index_url: bool,\n emit_trusted_host: bool,\n annotate: bool,\n generate_hashes: bool,\n default_index_url: str,\n index_urls: Iterable[str],\n trusted_hosts: Iterable[str],\n format_control: FormatControl,\n allow_unsafe: bool,\n find_links: List[str],\n emit_find_links: bool,\n ) -> None:\n self.dst_file = dst_file\n self.click_ctx = click_ctx\n self.dry_run = dry_run\n self.emit_header = emit_header\n self.emit_index_url = emit_index_url\n self.emit_trusted_host = emit_trusted_host\n self.annotate = annotate\n self.generate_hashes = generate_hashes\n self.default_index_url = default_index_url\n self.index_urls = index_urls\n self.trusted_hosts = trusted_hosts\n self.format_control = format_control\n self.allow_unsafe = allow_unsafe\n self.find_links = find_links\n self.emit_find_links = emit_find_links\n\n def _sort_key(self, ireq: InstallRequirement) -> Tuple[bool, str]:\n return (not ireq.editable, str(ireq.req).lower())\n\n def write_header(self) -> Iterator[str]:\n if self.emit_header:\n yield comment(\"#\")\n yield comment(\"# This file is autogenerated by pip-compile\")\n yield comment(\"# To update, run:\")\n yield comment(\"#\")\n compile_command = os.environ.get(\n \"CUSTOM_COMPILE_COMMAND\"\n ) or get_compile_command(self.click_ctx)\n yield comment(f\"# {compile_command}\")\n yield comment(\"#\")\n\n def write_index_options(self) -> Iterator[str]:\n if self.emit_index_url:\n for index, index_url in enumerate(dedup(self.index_urls)):\n if index == 0 and index_url.rstrip(\"/\") == self.default_index_url:\n continue\n flag = \"--index-url\" if index == 0 else \"--extra-index-url\"\n yield f\"{flag} {index_url}\"\n\n def write_trusted_hosts(self) -> Iterator[str]:\n if self.emit_trusted_host:\n for trusted_host in dedup(self.trusted_hosts):\n yield f\"--trusted-host {trusted_host}\"\n\n def write_format_controls(self) -> Iterator[str]:\n for nb in dedup(sorted(self.format_control.no_binary)):\n yield f\"--no-binary {nb}\"\n for ob in dedup(sorted(self.format_control.only_binary)):\n yield f\"--only-binary {ob}\"\n\n def write_find_links(self) -> Iterator[str]:\n if self.emit_find_links:\n for find_link in dedup(self.find_links):\n yield f\"--find-links {find_link}\"\n\n def write_flags(self) -> Iterator[str]:\n emitted = False\n for line in chain(\n self.write_index_options(),\n self.write_find_links(),\n self.write_trusted_hosts(),\n self.write_format_controls(),\n ):\n emitted = True\n yield line\n if emitted:\n yield \"\"\n\n def _iter_lines(\n self,\n results: Set[InstallRequirement],\n unsafe_requirements: Optional[Set[InstallRequirement]] = None,\n markers: Optional[Dict[str, Marker]] = None,\n hashes: Optional[Dict[InstallRequirement, Set[str]]] = None,\n ) -> Iterator[str]:\n # default values\n unsafe_requirements = unsafe_requirements or set()\n markers = markers or {}\n hashes = hashes or {}\n\n # Check for unhashed or unpinned packages if at least one package does have\n # hashes, which will trigger pip install's --require-hashes mode.\n warn_uninstallable = False\n has_hashes = hashes and any(hash for hash in hashes.values())\n\n yielded = False\n\n for line in self.write_header():\n yield line\n yielded = True\n for line in self.write_flags():\n yield line\n yielded = True\n\n unsafe_requirements = (\n {r for r in results if r.name in UNSAFE_PACKAGES}\n if not unsafe_requirements\n else unsafe_requirements\n )\n packages = {r for r in results if r.name not in UNSAFE_PACKAGES}\n\n if packages:\n for ireq in sorted(packages, key=self._sort_key):\n if has_hashes and not hashes.get(ireq):\n yield MESSAGE_UNHASHED_PACKAGE\n warn_uninstallable = True\n line = self._format_requirement(\n ireq, markers.get(key_from_ireq(ireq)), hashes=hashes\n )\n yield line\n yielded = True\n\n if unsafe_requirements:\n yield \"\"\n yielded = True\n if has_hashes and not self.allow_unsafe:\n yield MESSAGE_UNSAFE_PACKAGES_UNPINNED\n warn_uninstallable = True\n else:\n yield MESSAGE_UNSAFE_PACKAGES\n\n for ireq in sorted(unsafe_requirements, key=self._sort_key):\n ireq_key = key_from_ireq(ireq)\n if not self.allow_unsafe:\n yield comment(f\"# {ireq_key}\")\n else:\n line = self._format_requirement(\n ireq, marker=markers.get(ireq_key), hashes=hashes\n )\n yield line\n\n # Yield even when there's no real content, so that blank files are written\n if not yielded:\n yield \"\"\n\n if warn_uninstallable:\n log.warning(MESSAGE_UNINSTALLABLE)\n\n def write(\n self,\n results: Set[InstallRequirement],\n unsafe_requirements: Set[InstallRequirement],\n markers: Dict[str, Marker],\n hashes: Optional[Dict[InstallRequirement, Set[str]]],\n ) -> None:\n\n for line in self._iter_lines(results, unsafe_requirements, markers, hashes):\n log.info(line)\n if not self.dry_run:\n self.dst_file.write(unstyle(line).encode())\n self.dst_file.write(os.linesep.encode())\n\n def _format_requirement(\n self,\n ireq: InstallRequirement,\n marker: Optional[Marker] = None,\n hashes: Optional[Dict[InstallRequirement, Set[str]]] = None,\n ) -> str:\n ireq_hashes = (hashes if hashes is not None else {}).get(ireq)\n\n line = format_requirement(ireq, marker=marker, hashes=ireq_hashes)\n\n if not self.annotate:\n return line\n\n # Annotate what packages or reqs-ins this package is required by\n required_by = set()\n if hasattr(ireq, \"_source_ireqs\"):\n required_by |= {\n _comes_from_as_string(src_ireq)\n for src_ireq in ireq._source_ireqs\n if src_ireq.comes_from\n }\n elif ireq.comes_from:\n required_by.add(_comes_from_as_string(ireq))\n\n if required_by:\n sorted_required_by = sorted(required_by)\n if len(sorted_required_by) == 1:\n source = sorted_required_by[0]\n annotation = \" # via \" + source\n else:\n annotation_lines = [\" # via\"]\n for source in sorted_required_by:\n annotation_lines.append(\" # \" + source)\n annotation = \"\\n\".join(annotation_lines)\n line = f\"{line}\\n{comment(annotation)}\"\n\n return line\n", "path": "piptools/writer.py"}]}
3,280
188
gh_patches_debug_5037
rasdani/github-patches
git_diff
facebookresearch__hydra-793
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> [Bug] MISSING for Object Conf cls argument # 🐛 Bug OmegaConf cls argument should not be a mandatory value if target is defined. Can we change this to be an optional value with None being the default? ** Stack trace/error message ** ``` omegaconf.errors.MissingMandatoryValue: Missing mandatory value: scheduler.cls full_key: scheduler.cls reference_type=ObjectConf object_type=ObjectConf ``` ## System information - **Hydra Version** : 1.0.0rc2 - **Python version** : 3.7.7 </issue> <code> [start of hydra/types.py] 1 # Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved 2 from dataclasses import dataclass, field 3 from enum import Enum 4 from typing import Any, Callable, Dict 5 6 from omegaconf import MISSING 7 8 TaskFunction = Callable[[Any], Any] 9 10 11 @dataclass 12 # This extends Dict[str, Any] to allow for the deprecated "class" field. 13 # Once support for class field removed this can stop extending Dict. 14 class ObjectConf(Dict[str, Any]): 15 # class, class method or function name 16 target: str = MISSING 17 18 # parameters to pass to cls when calling it 19 params: Any = field(default_factory=dict) 20 21 # cls is deprecated, use target, cls will be removed in Hydra 1.1 22 cls: str = MISSING 23 24 # class is deprecated, use target, class will be removed in Hydra 1.1 25 # (class is Python keyword and is only supported through DictConfig) 26 # class: str = MISSING 27 28 29 class RunMode(Enum): 30 RUN = 1 31 MULTIRUN = 2 32 [end of hydra/types.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/hydra/types.py b/hydra/types.py --- a/hydra/types.py +++ b/hydra/types.py @@ -18,13 +18,6 @@ # parameters to pass to cls when calling it params: Any = field(default_factory=dict) - # cls is deprecated, use target, cls will be removed in Hydra 1.1 - cls: str = MISSING - - # class is deprecated, use target, class will be removed in Hydra 1.1 - # (class is Python keyword and is only supported through DictConfig) - # class: str = MISSING - class RunMode(Enum): RUN = 1
{"golden_diff": "diff --git a/hydra/types.py b/hydra/types.py\n--- a/hydra/types.py\n+++ b/hydra/types.py\n@@ -18,13 +18,6 @@\n # parameters to pass to cls when calling it\n params: Any = field(default_factory=dict)\n \n- # cls is deprecated, use target, cls will be removed in Hydra 1.1\n- cls: str = MISSING\n-\n- # class is deprecated, use target, class will be removed in Hydra 1.1\n- # (class is Python keyword and is only supported through DictConfig)\n- # class: str = MISSING\n-\n \n class RunMode(Enum):\n RUN = 1\n", "issue": "[Bug] MISSING for Object Conf cls argument\n# \ud83d\udc1b Bug\r\n\r\nOmegaConf cls argument should not be a mandatory value if target is defined. Can we change this to be an optional value with None being the default?\r\n \r\n** Stack trace/error message **\r\n```\r\nomegaconf.errors.MissingMandatoryValue: Missing mandatory value: scheduler.cls\r\n full_key: scheduler.cls\r\n reference_type=ObjectConf\r\n object_type=ObjectConf\r\n```\r\n\r\n\r\n## System information\r\n- **Hydra Version** : 1.0.0rc2\r\n- **Python version** : 3.7.7\r\n\n", "before_files": [{"content": "# Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved\nfrom dataclasses import dataclass, field\nfrom enum import Enum\nfrom typing import Any, Callable, Dict\n\nfrom omegaconf import MISSING\n\nTaskFunction = Callable[[Any], Any]\n\n\n@dataclass\n# This extends Dict[str, Any] to allow for the deprecated \"class\" field.\n# Once support for class field removed this can stop extending Dict.\nclass ObjectConf(Dict[str, Any]):\n # class, class method or function name\n target: str = MISSING\n\n # parameters to pass to cls when calling it\n params: Any = field(default_factory=dict)\n\n # cls is deprecated, use target, cls will be removed in Hydra 1.1\n cls: str = MISSING\n\n # class is deprecated, use target, class will be removed in Hydra 1.1\n # (class is Python keyword and is only supported through DictConfig)\n # class: str = MISSING\n\n\nclass RunMode(Enum):\n RUN = 1\n MULTIRUN = 2\n", "path": "hydra/types.py"}]}
953
156
gh_patches_debug_245
rasdani/github-patches
git_diff
mlcommons__GaNDLF-809
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> `gdown` does not seem to be working **Describe the bug** Current CI seems to be broken. **To Reproduce** Steps to reproduce the behavior: 1. Run any CI test 2. See error: ```python-traceback [SNIP!] if gdrive_file_id and is_gdrive_download_link: content_disposition = six.moves.urllib_parse.unquote( res.headers["Content-Disposition"] ) m = re.search(r"filename\*=UTF-8''(.*)", content_disposition) > filename_from_url = m.groups()[0] E AttributeError: 'NoneType' object has no attribute 'groups' ``` Example: https://github.com/mlcommons/GaNDLF/actions/runs/7489779631/job/20387346791?pr=764#step:9:219 **Expected behavior** The sample data file download should work. **Screenshots** N.A. **GaNDLF Version** Current master **Desktop (please complete the following information):** N.A. **Additional context** Basically, it is this error: https://github.com/wkentaro/gdown/issues/291 </issue> <code> [start of setup.py] 1 #!/usr/bin/env python 2 3 """The setup script.""" 4 5 6 import sys, re, os 7 from setuptools import setup, find_packages 8 from setuptools.command.install import install 9 from setuptools.command.develop import develop 10 from setuptools.command.egg_info import egg_info 11 12 try: 13 with open("README.md") as readme_file: 14 readme = readme_file.read() 15 except Exception as error: 16 readme = "No README information found." 17 sys.stderr.write( 18 "Warning: Could not open '%s' due %s\n" % ("README.md", error) 19 ) 20 21 22 class CustomInstallCommand(install): 23 def run(self): 24 install.run(self) 25 26 27 class CustomDevelopCommand(develop): 28 def run(self): 29 develop.run(self) 30 31 32 class CustomEggInfoCommand(egg_info): 33 def run(self): 34 egg_info.run(self) 35 36 37 try: 38 filepath = "GANDLF/version.py" 39 version_file = open(filepath) 40 (__version__,) = re.findall('__version__ = "(.*)"', version_file.read()) 41 42 except Exception as error: 43 __version__ = "0.0.1" 44 sys.stderr.write( 45 "Warning: Could not open '%s' due %s\n" % (filepath, error) 46 ) 47 48 # Handle cases where specific files need to be bundled into the final package as installed via PyPI 49 dockerfiles = [ 50 item 51 for item in os.listdir(os.path.dirname(os.path.abspath(__file__))) 52 if (os.path.isfile(item) and item.startswith("Dockerfile-")) 53 ] 54 entrypoint_files = [ 55 item 56 for item in os.listdir(os.path.dirname(os.path.abspath(__file__))) 57 if (os.path.isfile(item) and item.startswith("gandlf_")) 58 ] 59 setup_files = ["setup.py", ".dockerignore", "pyproject.toml", "MANIFEST.in"] 60 all_extra_files = dockerfiles + entrypoint_files + setup_files 61 all_extra_files_pathcorrected = [ 62 os.path.join("../", item) for item in all_extra_files 63 ] 64 # find_packages should only ever find these as subpackages of gandlf, not as top-level packages 65 # generate this dynamically? 66 # GANDLF.GANDLF is needed to prevent recursion madness in deployments 67 toplevel_package_excludes = [ 68 "GANDLF.GANDLF", 69 "anonymize", 70 "cli", 71 "compute", 72 "data", 73 "grad_clipping", 74 "losses", 75 "metrics", 76 "models", 77 "optimizers", 78 "schedulers", 79 "utils", 80 ] 81 82 83 requirements = [ 84 "torch==2.1.2", 85 "black==23.11.0", 86 "numpy==1.25.0", 87 "scipy", 88 "SimpleITK!=2.0.*", 89 "SimpleITK!=2.2.1", # https://github.com/mlcommons/GaNDLF/issues/536 90 "torchvision", 91 "tqdm", 92 "torchio==0.19.5", 93 "pandas>=2.0.0", 94 "scikit-learn>=0.23.2", 95 "scikit-image>=0.19.1", 96 "setuptools", 97 "seaborn", 98 "pyyaml", 99 "tiffslide", 100 "matplotlib", 101 "gdown==4.6.3", 102 "pytest", 103 "coverage", 104 "pytest-cov", 105 "psutil", 106 "medcam", 107 "opencv-python", 108 "torchmetrics==1.1.2", 109 "zarr==2.10.3", 110 "pydicom", 111 "onnx", 112 "torchinfo==1.7.0", 113 "segmentation-models-pytorch==0.3.3", 114 "ACSConv==0.1.1", 115 "docker", 116 "dicom-anonymizer", 117 "twine", 118 "zarr", 119 "keyring", 120 "monai==1.3.0", 121 ] 122 123 if __name__ == "__main__": 124 setup( 125 name="GANDLF", 126 version=__version__, 127 author="MLCommons", 128 author_email="[email protected]", 129 python_requires=">3.8, <3.12", 130 packages=find_packages( 131 where=os.path.dirname(os.path.abspath(__file__)), 132 exclude=toplevel_package_excludes, 133 ), 134 cmdclass={ 135 "install": CustomInstallCommand, 136 "develop": CustomDevelopCommand, 137 "egg_info": CustomEggInfoCommand, 138 }, 139 scripts=[ 140 "gandlf_run", 141 "gandlf_constructCSV", 142 "gandlf_collectStats", 143 "gandlf_patchMiner", 144 "gandlf_preprocess", 145 "gandlf_anonymizer", 146 "gandlf_verifyInstall", 147 "gandlf_configGenerator", 148 "gandlf_recoverConfig", 149 "gandlf_deploy", 150 "gandlf_optimizeModel", 151 "gandlf_generateMetrics", 152 ], 153 classifiers=[ 154 "Development Status :: 3 - Alpha", 155 "Intended Audience :: Science/Research", 156 "License :: OSI Approved :: Apache Software License", 157 "Natural Language :: English", 158 "Operating System :: OS Independent", 159 "Programming Language :: Python :: 3.9", 160 "Programming Language :: Python :: 3.10", 161 "Programming Language :: Python :: 3.11", 162 "Topic :: Scientific/Engineering :: Medical Science Apps.", 163 ], 164 description=( 165 "PyTorch-based framework that handles segmentation/regression/classification using various DL architectures for medical imaging." 166 ), 167 install_requires=requirements, 168 license="Apache-2.0", 169 long_description=readme, 170 long_description_content_type="text/markdown", 171 include_package_data=True, 172 package_data={"GANDLF": all_extra_files_pathcorrected}, 173 keywords="semantic, segmentation, regression, classification, data-augmentation, medical-imaging, clinical-workflows, deep-learning, pytorch", 174 zip_safe=False, 175 ) 176 [end of setup.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/setup.py b/setup.py --- a/setup.py +++ b/setup.py @@ -98,7 +98,7 @@ "pyyaml", "tiffslide", "matplotlib", - "gdown==4.6.3", + "gdown==5.1.0", "pytest", "coverage", "pytest-cov",
{"golden_diff": "diff --git a/setup.py b/setup.py\n--- a/setup.py\n+++ b/setup.py\n@@ -98,7 +98,7 @@\n \"pyyaml\",\n \"tiffslide\",\n \"matplotlib\",\n- \"gdown==4.6.3\",\n+ \"gdown==5.1.0\",\n \"pytest\",\n \"coverage\",\n \"pytest-cov\",\n", "issue": "`gdown` does not seem to be working\n**Describe the bug**\r\nCurrent CI seems to be broken.\r\n\r\n**To Reproduce**\r\nSteps to reproduce the behavior:\r\n1. Run any CI test\r\n2. See error:\r\n```python-traceback\r\n[SNIP!]\r\n if gdrive_file_id and is_gdrive_download_link:\r\n content_disposition = six.moves.urllib_parse.unquote(\r\n res.headers[\"Content-Disposition\"]\r\n )\r\n m = re.search(r\"filename\\*=UTF-8''(.*)\", content_disposition)\r\n> filename_from_url = m.groups()[0]\r\nE AttributeError: 'NoneType' object has no attribute 'groups'\r\n```\r\nExample: https://github.com/mlcommons/GaNDLF/actions/runs/7489779631/job/20387346791?pr=764#step:9:219\r\n\r\n**Expected behavior**\r\nThe sample data file download should work.\r\n\r\n**Screenshots**\r\nN.A.\r\n\r\n**GaNDLF Version**\r\nCurrent master\r\n\r\n**Desktop (please complete the following information):**\r\nN.A.\r\n\r\n**Additional context**\r\nBasically, it is this error: https://github.com/wkentaro/gdown/issues/291\r\n\n", "before_files": [{"content": "#!/usr/bin/env python\n\n\"\"\"The setup script.\"\"\"\n\n\nimport sys, re, os\nfrom setuptools import setup, find_packages\nfrom setuptools.command.install import install\nfrom setuptools.command.develop import develop\nfrom setuptools.command.egg_info import egg_info\n\ntry:\n with open(\"README.md\") as readme_file:\n readme = readme_file.read()\nexcept Exception as error:\n readme = \"No README information found.\"\n sys.stderr.write(\n \"Warning: Could not open '%s' due %s\\n\" % (\"README.md\", error)\n )\n\n\nclass CustomInstallCommand(install):\n def run(self):\n install.run(self)\n\n\nclass CustomDevelopCommand(develop):\n def run(self):\n develop.run(self)\n\n\nclass CustomEggInfoCommand(egg_info):\n def run(self):\n egg_info.run(self)\n\n\ntry:\n filepath = \"GANDLF/version.py\"\n version_file = open(filepath)\n (__version__,) = re.findall('__version__ = \"(.*)\"', version_file.read())\n\nexcept Exception as error:\n __version__ = \"0.0.1\"\n sys.stderr.write(\n \"Warning: Could not open '%s' due %s\\n\" % (filepath, error)\n )\n\n# Handle cases where specific files need to be bundled into the final package as installed via PyPI\ndockerfiles = [\n item\n for item in os.listdir(os.path.dirname(os.path.abspath(__file__)))\n if (os.path.isfile(item) and item.startswith(\"Dockerfile-\"))\n]\nentrypoint_files = [\n item\n for item in os.listdir(os.path.dirname(os.path.abspath(__file__)))\n if (os.path.isfile(item) and item.startswith(\"gandlf_\"))\n]\nsetup_files = [\"setup.py\", \".dockerignore\", \"pyproject.toml\", \"MANIFEST.in\"]\nall_extra_files = dockerfiles + entrypoint_files + setup_files\nall_extra_files_pathcorrected = [\n os.path.join(\"../\", item) for item in all_extra_files\n]\n# find_packages should only ever find these as subpackages of gandlf, not as top-level packages\n# generate this dynamically?\n# GANDLF.GANDLF is needed to prevent recursion madness in deployments\ntoplevel_package_excludes = [\n \"GANDLF.GANDLF\",\n \"anonymize\",\n \"cli\",\n \"compute\",\n \"data\",\n \"grad_clipping\",\n \"losses\",\n \"metrics\",\n \"models\",\n \"optimizers\",\n \"schedulers\",\n \"utils\",\n]\n\n\nrequirements = [\n \"torch==2.1.2\",\n \"black==23.11.0\",\n \"numpy==1.25.0\",\n \"scipy\",\n \"SimpleITK!=2.0.*\",\n \"SimpleITK!=2.2.1\", # https://github.com/mlcommons/GaNDLF/issues/536\n \"torchvision\",\n \"tqdm\",\n \"torchio==0.19.5\",\n \"pandas>=2.0.0\",\n \"scikit-learn>=0.23.2\",\n \"scikit-image>=0.19.1\",\n \"setuptools\",\n \"seaborn\",\n \"pyyaml\",\n \"tiffslide\",\n \"matplotlib\",\n \"gdown==4.6.3\",\n \"pytest\",\n \"coverage\",\n \"pytest-cov\",\n \"psutil\",\n \"medcam\",\n \"opencv-python\",\n \"torchmetrics==1.1.2\",\n \"zarr==2.10.3\",\n \"pydicom\",\n \"onnx\",\n \"torchinfo==1.7.0\",\n \"segmentation-models-pytorch==0.3.3\",\n \"ACSConv==0.1.1\",\n \"docker\",\n \"dicom-anonymizer\",\n \"twine\",\n \"zarr\",\n \"keyring\",\n \"monai==1.3.0\",\n]\n\nif __name__ == \"__main__\":\n setup(\n name=\"GANDLF\",\n version=__version__,\n author=\"MLCommons\",\n author_email=\"[email protected]\",\n python_requires=\">3.8, <3.12\",\n packages=find_packages(\n where=os.path.dirname(os.path.abspath(__file__)),\n exclude=toplevel_package_excludes,\n ),\n cmdclass={\n \"install\": CustomInstallCommand,\n \"develop\": CustomDevelopCommand,\n \"egg_info\": CustomEggInfoCommand,\n },\n scripts=[\n \"gandlf_run\",\n \"gandlf_constructCSV\",\n \"gandlf_collectStats\",\n \"gandlf_patchMiner\",\n \"gandlf_preprocess\",\n \"gandlf_anonymizer\",\n \"gandlf_verifyInstall\",\n \"gandlf_configGenerator\",\n \"gandlf_recoverConfig\",\n \"gandlf_deploy\",\n \"gandlf_optimizeModel\",\n \"gandlf_generateMetrics\",\n ],\n classifiers=[\n \"Development Status :: 3 - Alpha\",\n \"Intended Audience :: Science/Research\",\n \"License :: OSI Approved :: Apache Software License\",\n \"Natural Language :: English\",\n \"Operating System :: OS Independent\",\n \"Programming Language :: Python :: 3.9\",\n \"Programming Language :: Python :: 3.10\",\n \"Programming Language :: Python :: 3.11\",\n \"Topic :: Scientific/Engineering :: Medical Science Apps.\",\n ],\n description=(\n \"PyTorch-based framework that handles segmentation/regression/classification using various DL architectures for medical imaging.\"\n ),\n install_requires=requirements,\n license=\"Apache-2.0\",\n long_description=readme,\n long_description_content_type=\"text/markdown\",\n include_package_data=True,\n package_data={\"GANDLF\": all_extra_files_pathcorrected},\n keywords=\"semantic, segmentation, regression, classification, data-augmentation, medical-imaging, clinical-workflows, deep-learning, pytorch\",\n zip_safe=False,\n )\n", "path": "setup.py"}]}
2,514
85
gh_patches_debug_9165
rasdani/github-patches
git_diff
microsoft__DeepSpeed-1921
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> [BUG] No module named 'fcntl' while importing the package Hey, Working on windows 11, Python 3.7 and tried importing the package. Got the error in the title, is there a way around it since it's exclusive to Linux? ![fcntl](https://user-images.githubusercontent.com/61480454/164317686-5303f264-5fc7-42c0-89b5-d16a125ae11f.png) </issue> <code> [start of deepspeed/utils/debug.py] 1 """ debug utils """ 2 3 import fcntl 4 5 # for debug purposes map module and param objects to their fully qualified names 6 module_names = {} 7 param_names = {} 8 9 10 def debug_extract_module_and_param_names(model): 11 # extract the fully qualified names as soon as the model is acquired 12 global module_names 13 global param_names 14 # XXX: can probably make a map of param2module and vice-versa 15 module_names = {module: name for name, module in model.named_modules()} 16 param_names = {param: name for name, param in model.named_parameters()} 17 18 19 def debug_module2name(module): 20 if module in module_names: 21 return module_names[module] 22 else: 23 return "unknown" 24 25 26 def debug_module2name_id(module): 27 return f"name={debug_module2name(module)} id={module.id}" 28 29 30 def debug_module2name_class(module): 31 return f"name={debug_module2name(module)} {module.__class__.__name__}" 32 33 34 def debug_param2name(param): 35 if param in param_names: 36 return param_names[param] 37 else: 38 return "unknown" 39 40 41 def debug_param2name_id(param): 42 return f"name={debug_param2name(param)} id={param.ds_id}" 43 44 45 def debug_param2name_id_shape(param): 46 return f"name={debug_param2name(param)} id={param.ds_id} shape={param.data.shape}" 47 48 49 def debug_param2name_id_shape_device(param): 50 return f"name={debug_param2name(param)} id={param.ds_id} shape={param.data.shape} device={param.device}" 51 52 53 def debug_param2name_id_numel(param): 54 return f"name={debug_param2name(param)} id={param.ds_id} numel={param.numel()}" 55 56 57 def debug_param2name_id_shape_status(param): 58 return f"name={debug_param2name(param)} id={param.ds_id} shape={param.data.shape} status={param.ds_status}" 59 60 61 def printflock(*msgs): 62 """ 63 64 For printing messages for all concurrent gpus w/o getting interleaved text. 65 66 This is useful when debugging issues where multi-gpus don't sync. 67 68 1. Enable the force debug in say partitioning and zero3 files 69 2. Override the usual versions with :: 70 71 def print_rank_0(message, debug=False, force=False): 72 rank = torch.distributed.get_rank() 73 printflock(f"[{rank}] {message}") 74 3. run the program and you get both logs non-interleaved 75 76 But this makes it very difficult to make sense of the output, so the ``log_rank_file`` helper 77 function might be more useful, as it's easier to send each log stream into a separate file and 78 then compare those. 79 80 """ 81 82 with open(__file__, "r") as fh: 83 fcntl.flock(fh, fcntl.LOCK_EX) 84 try: 85 print(*msgs) 86 finally: 87 fcntl.flock(fh, fcntl.LOCK_UN) 88 89 90 fh = None 91 92 93 def log_rank_file(rank, *msgs): 94 """ 95 Print to a log file of the given rank 96 97 This is useful for debugging hanging in sync processes. Here is a possible workflow: 98 99 1. Enable the force debug in say partitioning and zero3 files 100 2. Override the usual versions of print_rank_0 in those files with :: 101 102 def print_rank_0(message, debug=False, force=False): 103 rank = torch.distributed.get_rank() 104 log_rank_file(rank, message) 105 106 3. run the program 107 4. fix up the expected differences, e.g. different cuda numbers :: 108 109 perl -pi -e 's|cuda:1|cuda:0|' log_rank_* 110 111 5. now diff and see where names and ids diverge - you will find where the gpus don't do the same 112 work (e.g. when some layers get conditionally skipped on one gpu but not all) 113 114 diff -u log_rank_0.txt log_rank_1.txt | less 115 116 """ 117 global fh 118 if fh is None: 119 fh = open(f"log_rank_{rank}.txt", "w") 120 for m in msgs: 121 fh.write(f"{m}\n") 122 fh.flush() 123 124 125 def print_backward_tensors(tensor): 126 def _print_bwd_tensors(grad_fn): 127 print(f"Backward tensors in {grad_fn}") 128 for funcs in grad_fn.next_functions: 129 if funcs[0]: 130 try: 131 tensor = getattr(funcs[0], 'variable') 132 print(funcs[0]) 133 print( 134 f"Tensor - id: {id(tensor)}, shape: {tensor.shape}, data: {tensor}, grad: {tensor.grad}" 135 ) 136 except AttributeError as e: 137 _print_bwd_tensors(funcs[0]) 138 139 if hasattr(tensor, 'grad_fn'): 140 _print_bwd_tensors(tensor.grad_fn) 141 [end of deepspeed/utils/debug.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/deepspeed/utils/debug.py b/deepspeed/utils/debug.py --- a/deepspeed/utils/debug.py +++ b/deepspeed/utils/debug.py @@ -1,6 +1,7 @@ """ debug utils """ -import fcntl +# For lazy import with printflock() +fcntl = None # for debug purposes map module and param objects to their fully qualified names module_names = {} @@ -78,6 +79,9 @@ then compare those. """ + global fcntl + if fcntl == None: + import fcntl with open(__file__, "r") as fh: fcntl.flock(fh, fcntl.LOCK_EX)
{"golden_diff": "diff --git a/deepspeed/utils/debug.py b/deepspeed/utils/debug.py\n--- a/deepspeed/utils/debug.py\n+++ b/deepspeed/utils/debug.py\n@@ -1,6 +1,7 @@\n \"\"\" debug utils \"\"\"\n \n-import fcntl\n+# For lazy import with printflock()\n+fcntl = None\n \n # for debug purposes map module and param objects to their fully qualified names\n module_names = {}\n@@ -78,6 +79,9 @@\n then compare those.\n \n \"\"\"\n+ global fcntl\n+ if fcntl == None:\n+ import fcntl\n \n with open(__file__, \"r\") as fh:\n fcntl.flock(fh, fcntl.LOCK_EX)\n", "issue": "[BUG] No module named 'fcntl' while importing the package\nHey,\r\nWorking on windows 11, Python 3.7 and tried importing the package.\r\nGot the error in the title, is there a way around it since it's exclusive to Linux?\r\n\r\n![fcntl](https://user-images.githubusercontent.com/61480454/164317686-5303f264-5fc7-42c0-89b5-d16a125ae11f.png)\r\n \n", "before_files": [{"content": "\"\"\" debug utils \"\"\"\n\nimport fcntl\n\n# for debug purposes map module and param objects to their fully qualified names\nmodule_names = {}\nparam_names = {}\n\n\ndef debug_extract_module_and_param_names(model):\n # extract the fully qualified names as soon as the model is acquired\n global module_names\n global param_names\n # XXX: can probably make a map of param2module and vice-versa\n module_names = {module: name for name, module in model.named_modules()}\n param_names = {param: name for name, param in model.named_parameters()}\n\n\ndef debug_module2name(module):\n if module in module_names:\n return module_names[module]\n else:\n return \"unknown\"\n\n\ndef debug_module2name_id(module):\n return f\"name={debug_module2name(module)} id={module.id}\"\n\n\ndef debug_module2name_class(module):\n return f\"name={debug_module2name(module)} {module.__class__.__name__}\"\n\n\ndef debug_param2name(param):\n if param in param_names:\n return param_names[param]\n else:\n return \"unknown\"\n\n\ndef debug_param2name_id(param):\n return f\"name={debug_param2name(param)} id={param.ds_id}\"\n\n\ndef debug_param2name_id_shape(param):\n return f\"name={debug_param2name(param)} id={param.ds_id} shape={param.data.shape}\"\n\n\ndef debug_param2name_id_shape_device(param):\n return f\"name={debug_param2name(param)} id={param.ds_id} shape={param.data.shape} device={param.device}\"\n\n\ndef debug_param2name_id_numel(param):\n return f\"name={debug_param2name(param)} id={param.ds_id} numel={param.numel()}\"\n\n\ndef debug_param2name_id_shape_status(param):\n return f\"name={debug_param2name(param)} id={param.ds_id} shape={param.data.shape} status={param.ds_status}\"\n\n\ndef printflock(*msgs):\n \"\"\"\n\n For printing messages for all concurrent gpus w/o getting interleaved text.\n\n This is useful when debugging issues where multi-gpus don't sync.\n\n 1. Enable the force debug in say partitioning and zero3 files\n 2. Override the usual versions with ::\n\n def print_rank_0(message, debug=False, force=False):\n rank = torch.distributed.get_rank()\n printflock(f\"[{rank}] {message}\")\n 3. run the program and you get both logs non-interleaved\n\n But this makes it very difficult to make sense of the output, so the ``log_rank_file`` helper\n function might be more useful, as it's easier to send each log stream into a separate file and\n then compare those.\n\n \"\"\"\n\n with open(__file__, \"r\") as fh:\n fcntl.flock(fh, fcntl.LOCK_EX)\n try:\n print(*msgs)\n finally:\n fcntl.flock(fh, fcntl.LOCK_UN)\n\n\nfh = None\n\n\ndef log_rank_file(rank, *msgs):\n \"\"\"\n Print to a log file of the given rank\n\n This is useful for debugging hanging in sync processes. Here is a possible workflow:\n\n 1. Enable the force debug in say partitioning and zero3 files\n 2. Override the usual versions of print_rank_0 in those files with ::\n\n def print_rank_0(message, debug=False, force=False):\n rank = torch.distributed.get_rank()\n log_rank_file(rank, message)\n\n 3. run the program\n 4. fix up the expected differences, e.g. different cuda numbers ::\n\n perl -pi -e 's|cuda:1|cuda:0|' log_rank_*\n\n 5. now diff and see where names and ids diverge - you will find where the gpus don't do the same\n work (e.g. when some layers get conditionally skipped on one gpu but not all)\n\n diff -u log_rank_0.txt log_rank_1.txt | less\n\n \"\"\"\n global fh\n if fh is None:\n fh = open(f\"log_rank_{rank}.txt\", \"w\")\n for m in msgs:\n fh.write(f\"{m}\\n\")\n fh.flush()\n\n\ndef print_backward_tensors(tensor):\n def _print_bwd_tensors(grad_fn):\n print(f\"Backward tensors in {grad_fn}\")\n for funcs in grad_fn.next_functions:\n if funcs[0]:\n try:\n tensor = getattr(funcs[0], 'variable')\n print(funcs[0])\n print(\n f\"Tensor - id: {id(tensor)}, shape: {tensor.shape}, data: {tensor}, grad: {tensor.grad}\"\n )\n except AttributeError as e:\n _print_bwd_tensors(funcs[0])\n\n if hasattr(tensor, 'grad_fn'):\n _print_bwd_tensors(tensor.grad_fn)\n", "path": "deepspeed/utils/debug.py"}]}
2,028
152
gh_patches_debug_31923
rasdani/github-patches
git_diff
alpa-projects__alpa-511
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> Adding the `pjit` in the comparison Some people are more familiar with using model parallel via [`pjit`](https://github.com/google/jax/blob/main/jax/experimental/pjit.py). What about adding one more rows [here](https://github.com/alpa-projects/alpa/blob/main/docs/gallery/tutorials/alpa_vs_pmap.py#L46-L52)? </issue> <code> [start of docs/gallery/tutorials/alpa_vs_pmap.py] 1 """ 2 Differences between alpa.parallelize and jax.pmap 3 ================================================= 4 5 The most common tool for parallelization or distributed computing in jax is 6 `pmap <https://jax.readthedocs.io/en/latest/jax.html#parallelization-pmap>`_. 7 With several lines of code change, we can use ``pmap`` for data parallel 8 training. However, we cannot use ``pmap`` for model parallel training, 9 which is required for training large models with billions of parameters. 10 11 On the contrary, ``alpa.parallelize`` supports both data parallelism and 12 model parallelism in an automatic way. ``alpa.parallelize`` analyzes the 13 jax computational graph and picks the best strategy. 14 If data parallelism is more suitable, ``alpa.parallelize`` achieves the same 15 performance as ``pmap`` but with less code change. 16 If model parallelism is more suitable, ``alpa.parallelize`` achieves better performance 17 and uses less memory than ``pmap``. 18 19 In this tutorial, we are going to compare ``alpa.parallelize`` and ``pmap`` on two 20 workloads. A more detailed comparison among ``alpa.parallelize``, ``pmap``, and ``xmap`` 21 is also attached at the end of the article. 22 """ 23 24 ################################################################################ 25 # When data parallelism is prefered 26 # --------------------------------- 27 28 # TODO 29 30 ################################################################################ 31 # When model parallelism is prefered 32 # ---------------------------------- 33 34 # TODO 35 36 ################################################################################ 37 # Comparing ``alpa.parallelize``, ``pmap``, and ``xmap`` 38 # ------------------------------------------------------ 39 # Besides ``pmap``, jax also provides 40 # `xmap <https://jax.readthedocs.io/en/latest/notebooks/xmap_tutorial.html>`_ 41 # for more advanced parallelization. 42 # The table below compares the features of ``alpa.parallelize``, ``pmap``, and ``xmap``. 43 # In summary, ``alpa.parallelize`` supports more parallelism techniques in a 44 # more automatic way. 45 # 46 # ================ ================ ==================== ==================== ========= 47 # Transformation Data Parallelism Operator Parallelism Pipeline Parallelism Automated 48 # ================ ================ ==================== ==================== ========= 49 # alpa.parallelize yes yes yes yes 50 # pmap yes no no no 51 # xmap yes yes no no 52 # ================ ================ ==================== ==================== ========= 53 # 54 # .. note:: 55 # Operator parallelism and pipeline parallelism are two forms of model parallelism. 56 # Operator parallelism partitions the work in a single operator and assigns them 57 # to different devices. Pipeline parallelism partitions the computational 58 # graphs and assigns different operators to different devices. 59 [end of docs/gallery/tutorials/alpa_vs_pmap.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/docs/gallery/tutorials/alpa_vs_pmap.py b/docs/gallery/tutorials/alpa_vs_pmap.py --- a/docs/gallery/tutorials/alpa_vs_pmap.py +++ b/docs/gallery/tutorials/alpa_vs_pmap.py @@ -34,14 +34,15 @@ # TODO ################################################################################ -# Comparing ``alpa.parallelize``, ``pmap``, and ``xmap`` -# ------------------------------------------------------ +# Comparing ``alpa.parallelize``, ``pmap``, ``xmap``, and ``pjit`` +# ----------------------------------------------------------------- # Besides ``pmap``, jax also provides -# `xmap <https://jax.readthedocs.io/en/latest/notebooks/xmap_tutorial.html>`_ +# `xmap <https://jax.readthedocs.io/en/latest/notebooks/xmap_tutorial.html>`_ and +# `pjit <https://jax.readthedocs.io/en/latest/jax-101/08-pjit.html>`_ # for more advanced parallelization. -# The table below compares the features of ``alpa.parallelize``, ``pmap``, and ``xmap``. -# In summary, ``alpa.parallelize`` supports more parallelism techniques in a -# more automatic way. +# The table below compares the features of ``alpa.parallelize``, ``pmap``, ``xmap`` +# and ``pjit``. In summary, ``alpa.parallelize`` supports more parallelism +# techniques in a more automatic way. # # ================ ================ ==================== ==================== ========= # Transformation Data Parallelism Operator Parallelism Pipeline Parallelism Automated @@ -49,6 +50,7 @@ # alpa.parallelize yes yes yes yes # pmap yes no no no # xmap yes yes no no +# pjit yes yes no no # ================ ================ ==================== ==================== ========= # # .. note::
{"golden_diff": "diff --git a/docs/gallery/tutorials/alpa_vs_pmap.py b/docs/gallery/tutorials/alpa_vs_pmap.py\n--- a/docs/gallery/tutorials/alpa_vs_pmap.py\n+++ b/docs/gallery/tutorials/alpa_vs_pmap.py\n@@ -34,14 +34,15 @@\n # TODO\n \n ################################################################################\n-# Comparing ``alpa.parallelize``, ``pmap``, and ``xmap``\n-# ------------------------------------------------------\n+# Comparing ``alpa.parallelize``, ``pmap``, ``xmap``, and ``pjit``\n+# -----------------------------------------------------------------\n # Besides ``pmap``, jax also provides\n-# `xmap <https://jax.readthedocs.io/en/latest/notebooks/xmap_tutorial.html>`_\n+# `xmap <https://jax.readthedocs.io/en/latest/notebooks/xmap_tutorial.html>`_ and \n+# `pjit <https://jax.readthedocs.io/en/latest/jax-101/08-pjit.html>`_\n # for more advanced parallelization.\n-# The table below compares the features of ``alpa.parallelize``, ``pmap``, and ``xmap``.\n-# In summary, ``alpa.parallelize`` supports more parallelism techniques in a\n-# more automatic way.\n+# The table below compares the features of ``alpa.parallelize``, ``pmap``, ``xmap`` \n+# and ``pjit``. In summary, ``alpa.parallelize`` supports more parallelism \n+# techniques in a more automatic way.\n #\n # ================ ================ ==================== ==================== =========\n # Transformation Data Parallelism Operator Parallelism Pipeline Parallelism Automated\n@@ -49,6 +50,7 @@\n # alpa.parallelize yes yes yes yes\n # pmap yes no no no\n # xmap yes yes no no\n+# pjit yes yes no no\n # ================ ================ ==================== ==================== =========\n #\n # .. note::\n", "issue": "Adding the `pjit` in the comparison\nSome people are more familiar with using model parallel via [`pjit`](https://github.com/google/jax/blob/main/jax/experimental/pjit.py). What about adding one more rows [here](https://github.com/alpa-projects/alpa/blob/main/docs/gallery/tutorials/alpa_vs_pmap.py#L46-L52)?\n", "before_files": [{"content": "\"\"\"\nDifferences between alpa.parallelize and jax.pmap\n=================================================\n\nThe most common tool for parallelization or distributed computing in jax is\n`pmap <https://jax.readthedocs.io/en/latest/jax.html#parallelization-pmap>`_.\nWith several lines of code change, we can use ``pmap`` for data parallel\ntraining. However, we cannot use ``pmap`` for model parallel training,\nwhich is required for training large models with billions of parameters.\n\nOn the contrary, ``alpa.parallelize`` supports both data parallelism and\nmodel parallelism in an automatic way. ``alpa.parallelize`` analyzes the\njax computational graph and picks the best strategy.\nIf data parallelism is more suitable, ``alpa.parallelize`` achieves the same\nperformance as ``pmap`` but with less code change.\nIf model parallelism is more suitable, ``alpa.parallelize`` achieves better performance\nand uses less memory than ``pmap``.\n\nIn this tutorial, we are going to compare ``alpa.parallelize`` and ``pmap`` on two\nworkloads. A more detailed comparison among ``alpa.parallelize``, ``pmap``, and ``xmap``\nis also attached at the end of the article.\n\"\"\"\n\n################################################################################\n# When data parallelism is prefered\n# ---------------------------------\n\n# TODO\n\n################################################################################\n# When model parallelism is prefered\n# ----------------------------------\n\n# TODO\n\n################################################################################\n# Comparing ``alpa.parallelize``, ``pmap``, and ``xmap``\n# ------------------------------------------------------\n# Besides ``pmap``, jax also provides\n# `xmap <https://jax.readthedocs.io/en/latest/notebooks/xmap_tutorial.html>`_\n# for more advanced parallelization.\n# The table below compares the features of ``alpa.parallelize``, ``pmap``, and ``xmap``.\n# In summary, ``alpa.parallelize`` supports more parallelism techniques in a\n# more automatic way.\n#\n# ================ ================ ==================== ==================== =========\n# Transformation Data Parallelism Operator Parallelism Pipeline Parallelism Automated\n# ================ ================ ==================== ==================== =========\n# alpa.parallelize yes yes yes yes\n# pmap yes no no no\n# xmap yes yes no no\n# ================ ================ ==================== ==================== =========\n#\n# .. note::\n# Operator parallelism and pipeline parallelism are two forms of model parallelism.\n# Operator parallelism partitions the work in a single operator and assigns them\n# to different devices. Pipeline parallelism partitions the computational\n# graphs and assigns different operators to different devices.\n", "path": "docs/gallery/tutorials/alpa_vs_pmap.py"}]}
1,298
424
gh_patches_debug_19670
rasdani/github-patches
git_diff
google__flax-2842
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> lm1b sampler update is out of bound? Hi, _Disclaimer: I'm not confident if this is some intended behavior._ I think in the lm1b example, the prediction sampler could write to indices out of bound of the result. The particular offending line is: https://github.com/google/flax/blob/master/examples/lm1b/temperature_sampler.py#L109. Since `i <= max_decode_len` according to https://github.com/google/flax/blob/master/examples/lm1b/temperature_sampler.py#L75, and since `prompt_inputs.shape[1] == max_decode_len`, the last iteration (or two iterations?) of while loop would overwrite to `prompt_inputs[..., max_decode_len - 1]` (due to XLA semantics). It's either a subtle bug (it won't raise cause error), or there's something I'm not understanding about the sampling procedure. </issue> <code> [start of examples/lm1b/temperature_sampler.py] 1 # Copyright 2022 The Flax Authors. 2 # 3 # Licensed under the Apache License, Version 2.0 (the "License"); 4 # you may not use this file except in compliance with the License. 5 # You may obtain a copy of the License at 6 # 7 # http://www.apache.org/licenses/LICENSE-2.0 8 # 9 # Unless required by applicable law or agreed to in writing, software 10 # distributed under the License is distributed on an "AS IS" BASIS, 11 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 12 # See the License for the specific language governing permissions and 13 # limitations under the License. 14 15 """Fast decoding routines for inference from a trained language model.""" 16 17 from jax import lax 18 from jax import random 19 import jax.numpy as jnp 20 21 22 # Constants 23 # The default End-of-Sentence token id is 2 (SentencePiece). 24 EOS_ID = 2 25 26 27 def temperature_sample(prompt_inputs, 28 init_cache, 29 tokens_to_logits, 30 prng_key, 31 temperature=1.0, 32 topk=20, 33 eos_token=EOS_ID): 34 """Temperature sampling for language model generation. 35 36 Args: 37 prompt_inputs: array: [batch_size, max_decode_len] int32 sequence of tokens. 38 init_cache: flax attention cache. 39 tokens_to_logits: fast autoregressive decoder function taking single token 40 slices and cache and returning next-token logits and updated cache. 41 prng_key: JAX PRNGKey. 42 temperature: float: sampling temperature factor. As it approaches 43 zero this becomes equivalent to greedy sampling. 44 topk: integer: if nonzero only use the top-k logits to sample next token, 45 if zero don't use any cutoff and sample from full logits over vocabulary. 46 eos_token: int: end-of-sentence token for target vocabulary. 47 48 Returns: 49 Array of sampled sequences: [batch_size, max_decode_len] 50 """ 51 batch_size = prompt_inputs.shape[0] 52 max_decode_len = prompt_inputs.shape[1] 53 end_marker = jnp.array(eos_token) 54 temperature = jnp.array(temperature) 55 56 # Initialize sampling loop state. 57 # initial loop PRNGKey 58 rng0 = prng_key 59 # loop position counter. 60 i0 = jnp.array(0) 61 # per batch-item holding current token in loop. 62 token0 = jnp.zeros((batch_size, 1), dtype=jnp.int32) 63 # per batch-item state bit indicating if sentence has finished. 64 ended0 = jnp.zeros((batch_size, 1), dtype=jnp.bool_) 65 # (batch, length) array containing prefix prompt tokens for sampling loop 66 # as well as the generated output of newly sampled tokens. 67 sequences0 = prompt_inputs 68 # Sampling loop state is stored in a simple tuple. 69 sampling_loop_init_state = (i0, sequences0, init_cache, token0, ended0, rng0) 70 71 def sampling_loop_cond_fn(state): 72 """Sampling loop termination condition.""" 73 (i, _, _, _, ended, _) = state 74 # Have we reached max decoding length? 75 not_at_end = (i < max_decode_len) 76 # Have all sampled sequences reached an end marker? 77 all_sequences_ended = jnp.all(ended) 78 return not_at_end & (~all_sequences_ended) 79 80 def sampling_loop_body_fn(state): 81 """Sampling loop state update.""" 82 i, sequences, cache, cur_token, ended, rng = state 83 # Split RNG for sampling. 84 rng1, rng2 = random.split(rng) 85 # Call fast-decoder model on current tokens to get next-position logits. 86 logits, new_cache = tokens_to_logits(cur_token, cache) 87 # Sample next token from logits. 88 # TODO(levskaya): add top-p "nucleus" sampling option. 89 if topk: 90 # Get top-k logits and their indices, sample within these top-k tokens. 91 topk_logits, topk_idxs = lax.top_k(logits, topk) 92 topk_token = jnp.expand_dims(random.categorical( 93 rng1, topk_logits / temperature).astype(jnp.int32), axis=-1) 94 # Return the original indices corresponding to the sampled top-k tokens. 95 next_token = jnp.squeeze( 96 jnp.take_along_axis(topk_idxs, topk_token, axis=-1), axis=-1) 97 else: 98 next_token = random.categorical( 99 rng1, logits / temperature).astype(jnp.int32) 100 # Only use sampled tokens if we're past provided prefix tokens. 101 out_of_prompt = (sequences[:, i+1] == 0) 102 next_token = (next_token * out_of_prompt + 103 sequences[:, i+1] * ~out_of_prompt) 104 # If end-marker reached for batch item, only emit padding tokens. 105 next_token_or_endpad = (next_token[None] * ~ended) 106 ended |= (next_token_or_endpad == end_marker) 107 # Add current sampled tokens to recorded sequences. 108 new_sequences = lax.dynamic_update_slice( 109 sequences, next_token_or_endpad, (0, i+1)) 110 return (i+1, new_sequences, new_cache, next_token_or_endpad, ended, rng2) 111 112 # Run sampling loop and collect final state. 113 final_state = lax.while_loop(sampling_loop_cond_fn, 114 sampling_loop_body_fn, 115 sampling_loop_init_state) 116 117 # Pick part of the state corresponding to the sampled sequences. 118 final_sequences = final_state[1] 119 return final_sequences 120 [end of examples/lm1b/temperature_sampler.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/examples/lm1b/temperature_sampler.py b/examples/lm1b/temperature_sampler.py --- a/examples/lm1b/temperature_sampler.py +++ b/examples/lm1b/temperature_sampler.py @@ -57,7 +57,7 @@ # initial loop PRNGKey rng0 = prng_key # loop position counter. - i0 = jnp.array(0) + i0 = jnp.array(-1) # per batch-item holding current token in loop. token0 = jnp.zeros((batch_size, 1), dtype=jnp.int32) # per batch-item state bit indicating if sentence has finished. @@ -72,7 +72,7 @@ """Sampling loop termination condition.""" (i, _, _, _, ended, _) = state # Have we reached max decoding length? - not_at_end = (i < max_decode_len) + not_at_end = (i < max_decode_len - 1) # Have all sampled sequences reached an end marker? all_sequences_ended = jnp.all(ended) return not_at_end & (~all_sequences_ended)
{"golden_diff": "diff --git a/examples/lm1b/temperature_sampler.py b/examples/lm1b/temperature_sampler.py\n--- a/examples/lm1b/temperature_sampler.py\n+++ b/examples/lm1b/temperature_sampler.py\n@@ -57,7 +57,7 @@\n # initial loop PRNGKey\n rng0 = prng_key\n # loop position counter.\n- i0 = jnp.array(0)\n+ i0 = jnp.array(-1)\n # per batch-item holding current token in loop.\n token0 = jnp.zeros((batch_size, 1), dtype=jnp.int32)\n # per batch-item state bit indicating if sentence has finished.\n@@ -72,7 +72,7 @@\n \"\"\"Sampling loop termination condition.\"\"\"\n (i, _, _, _, ended, _) = state\n # Have we reached max decoding length?\n- not_at_end = (i < max_decode_len)\n+ not_at_end = (i < max_decode_len - 1)\n # Have all sampled sequences reached an end marker?\n all_sequences_ended = jnp.all(ended)\n return not_at_end & (~all_sequences_ended)\n", "issue": "lm1b sampler update is out of bound?\nHi,\r\n\r\n_Disclaimer: I'm not confident if this is some intended behavior._\r\n\r\nI think in the lm1b example, the prediction sampler could write to indices out of bound of the result.\r\nThe particular offending line is: https://github.com/google/flax/blob/master/examples/lm1b/temperature_sampler.py#L109.\r\n\r\nSince `i <= max_decode_len` according to https://github.com/google/flax/blob/master/examples/lm1b/temperature_sampler.py#L75, and since `prompt_inputs.shape[1] == max_decode_len`, the last iteration (or two iterations?) of while loop would overwrite to `prompt_inputs[..., max_decode_len - 1]` (due to XLA semantics). \r\n\r\nIt's either a subtle bug (it won't raise cause error), or there's something I'm not understanding about the sampling procedure.\n", "before_files": [{"content": "# Copyright 2022 The Flax Authors.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\n\"\"\"Fast decoding routines for inference from a trained language model.\"\"\"\n\nfrom jax import lax\nfrom jax import random\nimport jax.numpy as jnp\n\n\n# Constants\n# The default End-of-Sentence token id is 2 (SentencePiece).\nEOS_ID = 2\n\n\ndef temperature_sample(prompt_inputs,\n init_cache,\n tokens_to_logits,\n prng_key,\n temperature=1.0,\n topk=20,\n eos_token=EOS_ID):\n \"\"\"Temperature sampling for language model generation.\n\n Args:\n prompt_inputs: array: [batch_size, max_decode_len] int32 sequence of tokens.\n init_cache: flax attention cache.\n tokens_to_logits: fast autoregressive decoder function taking single token\n slices and cache and returning next-token logits and updated cache.\n prng_key: JAX PRNGKey.\n temperature: float: sampling temperature factor. As it approaches\n zero this becomes equivalent to greedy sampling.\n topk: integer: if nonzero only use the top-k logits to sample next token,\n if zero don't use any cutoff and sample from full logits over vocabulary.\n eos_token: int: end-of-sentence token for target vocabulary.\n\n Returns:\n Array of sampled sequences: [batch_size, max_decode_len]\n \"\"\"\n batch_size = prompt_inputs.shape[0]\n max_decode_len = prompt_inputs.shape[1]\n end_marker = jnp.array(eos_token)\n temperature = jnp.array(temperature)\n\n # Initialize sampling loop state.\n # initial loop PRNGKey\n rng0 = prng_key\n # loop position counter.\n i0 = jnp.array(0)\n # per batch-item holding current token in loop.\n token0 = jnp.zeros((batch_size, 1), dtype=jnp.int32)\n # per batch-item state bit indicating if sentence has finished.\n ended0 = jnp.zeros((batch_size, 1), dtype=jnp.bool_)\n # (batch, length) array containing prefix prompt tokens for sampling loop\n # as well as the generated output of newly sampled tokens.\n sequences0 = prompt_inputs\n # Sampling loop state is stored in a simple tuple.\n sampling_loop_init_state = (i0, sequences0, init_cache, token0, ended0, rng0)\n\n def sampling_loop_cond_fn(state):\n \"\"\"Sampling loop termination condition.\"\"\"\n (i, _, _, _, ended, _) = state\n # Have we reached max decoding length?\n not_at_end = (i < max_decode_len)\n # Have all sampled sequences reached an end marker?\n all_sequences_ended = jnp.all(ended)\n return not_at_end & (~all_sequences_ended)\n\n def sampling_loop_body_fn(state):\n \"\"\"Sampling loop state update.\"\"\"\n i, sequences, cache, cur_token, ended, rng = state\n # Split RNG for sampling.\n rng1, rng2 = random.split(rng)\n # Call fast-decoder model on current tokens to get next-position logits.\n logits, new_cache = tokens_to_logits(cur_token, cache)\n # Sample next token from logits.\n # TODO(levskaya): add top-p \"nucleus\" sampling option.\n if topk:\n # Get top-k logits and their indices, sample within these top-k tokens.\n topk_logits, topk_idxs = lax.top_k(logits, topk)\n topk_token = jnp.expand_dims(random.categorical(\n rng1, topk_logits / temperature).astype(jnp.int32), axis=-1)\n # Return the original indices corresponding to the sampled top-k tokens.\n next_token = jnp.squeeze(\n jnp.take_along_axis(topk_idxs, topk_token, axis=-1), axis=-1)\n else:\n next_token = random.categorical(\n rng1, logits / temperature).astype(jnp.int32)\n # Only use sampled tokens if we're past provided prefix tokens.\n out_of_prompt = (sequences[:, i+1] == 0)\n next_token = (next_token * out_of_prompt +\n sequences[:, i+1] * ~out_of_prompt)\n # If end-marker reached for batch item, only emit padding tokens.\n next_token_or_endpad = (next_token[None] * ~ended)\n ended |= (next_token_or_endpad == end_marker)\n # Add current sampled tokens to recorded sequences.\n new_sequences = lax.dynamic_update_slice(\n sequences, next_token_or_endpad, (0, i+1))\n return (i+1, new_sequences, new_cache, next_token_or_endpad, ended, rng2)\n\n # Run sampling loop and collect final state.\n final_state = lax.while_loop(sampling_loop_cond_fn,\n sampling_loop_body_fn,\n sampling_loop_init_state)\n\n # Pick part of the state corresponding to the sampled sequences.\n final_sequences = final_state[1]\n return final_sequences\n", "path": "examples/lm1b/temperature_sampler.py"}]}
2,172
254
gh_patches_debug_12520
rasdani/github-patches
git_diff
rucio__rucio-1573
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> Problems with accessing the REST API and the WebUI in the development docker container Motivation ---------- It is not possible to access the REST API under 'https://localhost/proxy' and to access the Web UI under 'https://localhost/ui/' within the docker container build with the Dockerfile under /etc/docker/dev. Modification ------------ A few modifications can be copied from the demo environment. </issue> <code> [start of lib/rucio/core/lifetime_exception.py] 1 # Copyright European Organization for Nuclear Research (CERN) 2 # 3 # Licensed under the Apache License, Version 2.0 (the "License"); 4 # You may not use this file except in compliance with the License. 5 # You may obtain a copy of the License at 6 # http://www.apache.org/licenses/LICENSE-2.0 7 # 8 # Authors: 9 # - Cedric Serfon, <[email protected]>, 2016-2018 10 11 from re import match 12 from datetime import datetime, timedelta 13 14 from sqlalchemy import or_ 15 from sqlalchemy.exc import IntegrityError 16 from sqlalchemy.orm.exc import NoResultFound 17 18 from rucio.common.exception import ConfigNotFound, RucioException, LifetimeExceptionDuplicate, LifetimeExceptionNotFound, UnsupportedOperation 19 from rucio.common.utils import generate_uuid, str_to_date 20 import rucio.common.policy 21 from rucio.core.config import get 22 from rucio.core.message import add_message 23 from rucio.core.rse import list_rse_attributes 24 25 from rucio.db.sqla import models 26 from rucio.db.sqla.constants import DIDType, LifetimeExceptionsState 27 from rucio.db.sqla.session import transactional_session, stream_session, read_session 28 29 30 @stream_session 31 def list_exceptions(exception_id, states, session=None): 32 """ 33 List exceptions to Lifetime Model. 34 35 :param exception_id: The id of the exception 36 :param states: The states to filter 37 :param session: The database session in use. 38 """ 39 40 state_clause = [] 41 if states: 42 state_clause = [models.LifetimeExceptions.state == state for state in states] 43 44 query = session.query(models.LifetimeExceptions.id, 45 models.LifetimeExceptions.scope, models.LifetimeExceptions.name, 46 models.LifetimeExceptions.did_type, 47 models.LifetimeExceptions.account, 48 models.LifetimeExceptions.pattern, 49 models.LifetimeExceptions.comments, 50 models.LifetimeExceptions.state, 51 models.LifetimeExceptions.expires_at, 52 models.LifetimeExceptions.created_at) 53 if state_clause != []: 54 query = query.filter(or_(*state_clause)) 55 if exception_id: 56 query = query.filter(id=exception_id) 57 58 for exception in query.yield_per(5): 59 yield {'id': exception.id, 'scope': exception.scope, 'name': exception.name, 60 'did_type': exception.did_type, 'account': exception.account, 61 'pattern': exception.pattern, 'comments': exception.comments, 62 'state': exception.state, 'created_at': exception.created_at, 63 'expires_at': exception.expires_at} 64 65 66 @transactional_session 67 def add_exception(dids, account, pattern, comments, expires_at, session=None): 68 """ 69 Add exceptions to Lifetime Model. 70 71 :param dids: The list of dids 72 :param account: The account of the requester. 73 :param pattern: The account. 74 :param comments: The comments associated to the exception. 75 :param expires_at: The expiration date of the exception. 76 :param session: The database session in use. 77 78 returns: The id of the exception. 79 """ 80 exception_id = generate_uuid() 81 text = 'Account %s requested a lifetime extension for a list of DIDs that can be found below\n' % account 82 reason = comments 83 volume = None 84 lifetime = None 85 if comments.find('||||') > -1: 86 reason, volume = comments.split('||||') 87 text += 'The reason for the extension is "%s"\n' % reason 88 text += 'It represents %s datasets\n' % len(dids) 89 if volume: 90 text += 'The estimated physical volume is %s\n' % volume 91 if expires_at and (isinstance(expires_at, str) or isinstance(expires_at, unicode)): 92 lifetime = str_to_date(expires_at) 93 text += 'The lifetime exception should expires on %s\n' % str(expires_at) 94 elif isinstance(expires_at, datetime): 95 lifetime = expires_at 96 text += 'The lifetime exception should expires on %s\n' % str(expires_at) 97 text += 'Link to approve or reject this request can be found at the end of the mail\n' 98 text += '\n' 99 text += 'DIDTYPE SCOPE NAME\n' 100 text += '\n' 101 truncated_message = False 102 for did in dids: 103 did_type = None 104 if 'did_type' in did: 105 if isinstance(did['did_type'], str) or isinstance(did['did_type'], unicode): 106 did_type = DIDType.from_sym(did['did_type']) 107 else: 108 did_type = did['did_type'] 109 new_exception = models.LifetimeExceptions(id=exception_id, scope=did['scope'], name=did['name'], did_type=did_type, 110 account=account, pattern=pattern, comments=reason, state=LifetimeExceptionsState.WAITING, expires_at=lifetime) 111 if len(text) < 3000: 112 text += '%s %s %s\n' % (str(did_type), did['scope'], did['name']) 113 else: 114 truncated_message = True 115 try: 116 new_exception.save(session=session, flush=False) 117 except IntegrityError as error: 118 if match('.*ORA-00001.*', str(error.args[0]))\ 119 or match('.*IntegrityError.*UNIQUE constraint failed.*', str(error.args[0]))\ 120 or match('.*1062.*Duplicate entry.*for key.*', str(error.args[0]))\ 121 or match('.*sqlite3.IntegrityError.*are not unique.*', error.args[0]): 122 raise LifetimeExceptionDuplicate() 123 raise RucioException(error.args[0]) 124 if truncated_message: 125 text += '...\n' 126 text += 'List too long. Truncated\n' 127 text += '\n' 128 text += 'Approve: https://rucio-ui.cern.ch/lifetime_exception?id=%s&action=approve\n' % str(exception_id) 129 text += 'Deny: https://rucio-ui.cern.ch/lifetime_exception?id=%s&action=deny\n' % str(exception_id) 130 approvers_email = [] 131 try: 132 approvers_email = get('lifetime_model', 'approvers_email', session=session) 133 approvers_email = approvers_email.split(',') # pylint: disable=no-member 134 except ConfigNotFound: 135 approvers_email = [] 136 137 add_message(event_type='email', 138 payload={'body': text, 'to': approvers_email, 139 'subject': '[RUCIO] Request to approve lifetime exception %s' % str(exception_id)}, 140 session=session) 141 return exception_id 142 143 144 @transactional_session 145 def update_exception(exception_id, state, session=None): 146 """ 147 Update exceptions state to Lifetime Model. 148 149 :param exception_id: The id of the exception 150 :param state: The states to filter 151 :param session: The database session in use. 152 """ 153 query = session.query(models.LifetimeExceptions).filter_by(id=exception_id) 154 try: 155 query.one() 156 except NoResultFound: 157 raise LifetimeExceptionNotFound 158 159 if state in [LifetimeExceptionsState.APPROVED, LifetimeExceptionsState.REJECTED]: 160 query.update({'state': state, 'updated_at': datetime.utcnow()}, synchronize_session=False) 161 else: 162 raise UnsupportedOperation 163 164 165 @read_session 166 def define_eol(scope, name, rses, session=None): 167 """ 168 ATLAS policy for rules on SCRATCHDISK 169 170 :param scope: Scope of the DID. 171 :param name: Name of the DID. 172 :param rses: List of RSEs. 173 :param session: The database session in use. 174 """ 175 policy = rucio.common.policy.get_policy() 176 if policy != 'atlas': 177 return None 178 179 # Check if on ATLAS managed space 180 if [rse for rse in rses if list_rse_attributes(rse=None, rse_id=rse['id'], session=session).get('type') in ['LOCALGROUPDISK', 'LOCALGROUPTAPE', 'GROUPDISK', 'GROUPTAPE']]: 181 return None 182 # Now check the lifetime policy 183 try: 184 did = session.query(models.DataIdentifier).filter(models.DataIdentifier.scope == scope, 185 models.DataIdentifier.name == name).one() 186 except NoResultFound: 187 return None 188 policy_dict = rucio.common.policy.get_lifetime_policy() 189 did_type = 'other' 190 if scope.startswith('mc'): 191 did_type = 'mc' 192 elif scope.startswith('data'): 193 did_type = 'data' 194 elif scope.startswith('valid'): 195 did_type = 'valid' 196 else: 197 did_type = 'other' 198 for policy in policy_dict[did_type]: 199 if 'exclude' in policy: 200 to_exclude = False 201 for key in policy['exclude']: 202 meta_key = None 203 if key not in ['datatype', 'project', ]: 204 if key == 'stream': 205 meta_key = 'stream_name' 206 elif key == 'tags': 207 meta_key = 'version' 208 else: 209 meta_key = key 210 values = policy['exclude'][key] 211 for value in values: 212 value = value.replace('%', '.*') 213 if meta_key and did[meta_key] and value and match(value, did[meta_key]): 214 to_exclude = True 215 break 216 if to_exclude: 217 break 218 if to_exclude: 219 continue 220 if 'include' in policy: 221 match_policy = True 222 for key in policy['include']: 223 meta_key = None 224 if key not in ['datatype', 'project', ]: 225 if key == 'stream': 226 meta_key = 'stream_name' 227 elif key == 'tags': 228 meta_key = 'version' 229 else: 230 continue 231 else: 232 meta_key = key 233 values = policy['include'][key] 234 to_keep = False 235 for value in values: 236 value = value.replace('%', '.*') 237 if meta_key and did[meta_key] and value and match(value, did[meta_key]): 238 to_keep = True 239 break 240 match_policy = match_policy and to_keep 241 if not to_keep: 242 match_policy = False 243 break 244 if match_policy: 245 if int(policy['age']) >= 12: 246 years = int(int(policy['age']) / 12) 247 months = int(policy['age']) - years * 12 248 lifetime_value = 365 * years + 30 * months 249 else: 250 lifetime_value = int(policy['age']) * 30 251 if int(policy['extension']) >= 12: 252 years = int(int(policy['extension']) / 12) 253 months = int(policy['extension']) - years * 12 254 extension = 365 * years + 30 * months 255 else: 256 extension = int(policy['extension']) * 30 257 258 default_eol_at = did.created_at + timedelta(days=lifetime_value) 259 eol_at = default_eol_at 260 if did.accessed_at: 261 eol_at = did.accessed_at + timedelta(days=extension) 262 if eol_at < default_eol_at: 263 eol_at = default_eol_at 264 return eol_at 265 return None 266 [end of lib/rucio/core/lifetime_exception.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/lib/rucio/core/lifetime_exception.py b/lib/rucio/core/lifetime_exception.py --- a/lib/rucio/core/lifetime_exception.py +++ b/lib/rucio/core/lifetime_exception.py @@ -7,6 +7,7 @@ # # Authors: # - Cedric Serfon, <[email protected]>, 2016-2018 +# - Dimitrios Christidis, <[email protected]> 2018 from re import match from datetime import datetime, timedelta @@ -152,7 +153,7 @@ """ query = session.query(models.LifetimeExceptions).filter_by(id=exception_id) try: - query.one() + query.first() except NoResultFound: raise LifetimeExceptionNotFound
{"golden_diff": "diff --git a/lib/rucio/core/lifetime_exception.py b/lib/rucio/core/lifetime_exception.py\n--- a/lib/rucio/core/lifetime_exception.py\n+++ b/lib/rucio/core/lifetime_exception.py\n@@ -7,6 +7,7 @@\n #\n # Authors:\n # - Cedric Serfon, <[email protected]>, 2016-2018\n+# - Dimitrios Christidis, <[email protected]> 2018\n \n from re import match\n from datetime import datetime, timedelta\n@@ -152,7 +153,7 @@\n \"\"\"\n query = session.query(models.LifetimeExceptions).filter_by(id=exception_id)\n try:\n- query.one()\n+ query.first()\n except NoResultFound:\n raise LifetimeExceptionNotFound\n", "issue": "Problems with accessing the REST API and the WebUI in the development docker container\nMotivation\r\n----------\r\nIt is not possible to access the REST API under 'https://localhost/proxy' and to access the Web UI under 'https://localhost/ui/' within the docker container build with the Dockerfile under /etc/docker/dev.\r\n\r\nModification\r\n------------\r\nA few modifications can be copied from the demo environment.\n", "before_files": [{"content": "# Copyright European Organization for Nuclear Research (CERN)\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# You may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Authors:\n# - Cedric Serfon, <[email protected]>, 2016-2018\n\nfrom re import match\nfrom datetime import datetime, timedelta\n\nfrom sqlalchemy import or_\nfrom sqlalchemy.exc import IntegrityError\nfrom sqlalchemy.orm.exc import NoResultFound\n\nfrom rucio.common.exception import ConfigNotFound, RucioException, LifetimeExceptionDuplicate, LifetimeExceptionNotFound, UnsupportedOperation\nfrom rucio.common.utils import generate_uuid, str_to_date\nimport rucio.common.policy\nfrom rucio.core.config import get\nfrom rucio.core.message import add_message\nfrom rucio.core.rse import list_rse_attributes\n\nfrom rucio.db.sqla import models\nfrom rucio.db.sqla.constants import DIDType, LifetimeExceptionsState\nfrom rucio.db.sqla.session import transactional_session, stream_session, read_session\n\n\n@stream_session\ndef list_exceptions(exception_id, states, session=None):\n \"\"\"\n List exceptions to Lifetime Model.\n\n :param exception_id: The id of the exception\n :param states: The states to filter\n :param session: The database session in use.\n \"\"\"\n\n state_clause = []\n if states:\n state_clause = [models.LifetimeExceptions.state == state for state in states]\n\n query = session.query(models.LifetimeExceptions.id,\n models.LifetimeExceptions.scope, models.LifetimeExceptions.name,\n models.LifetimeExceptions.did_type,\n models.LifetimeExceptions.account,\n models.LifetimeExceptions.pattern,\n models.LifetimeExceptions.comments,\n models.LifetimeExceptions.state,\n models.LifetimeExceptions.expires_at,\n models.LifetimeExceptions.created_at)\n if state_clause != []:\n query = query.filter(or_(*state_clause))\n if exception_id:\n query = query.filter(id=exception_id)\n\n for exception in query.yield_per(5):\n yield {'id': exception.id, 'scope': exception.scope, 'name': exception.name,\n 'did_type': exception.did_type, 'account': exception.account,\n 'pattern': exception.pattern, 'comments': exception.comments,\n 'state': exception.state, 'created_at': exception.created_at,\n 'expires_at': exception.expires_at}\n\n\n@transactional_session\ndef add_exception(dids, account, pattern, comments, expires_at, session=None):\n \"\"\"\n Add exceptions to Lifetime Model.\n\n :param dids: The list of dids\n :param account: The account of the requester.\n :param pattern: The account.\n :param comments: The comments associated to the exception.\n :param expires_at: The expiration date of the exception.\n :param session: The database session in use.\n\n returns: The id of the exception.\n \"\"\"\n exception_id = generate_uuid()\n text = 'Account %s requested a lifetime extension for a list of DIDs that can be found below\\n' % account\n reason = comments\n volume = None\n lifetime = None\n if comments.find('||||') > -1:\n reason, volume = comments.split('||||')\n text += 'The reason for the extension is \"%s\"\\n' % reason\n text += 'It represents %s datasets\\n' % len(dids)\n if volume:\n text += 'The estimated physical volume is %s\\n' % volume\n if expires_at and (isinstance(expires_at, str) or isinstance(expires_at, unicode)):\n lifetime = str_to_date(expires_at)\n text += 'The lifetime exception should expires on %s\\n' % str(expires_at)\n elif isinstance(expires_at, datetime):\n lifetime = expires_at\n text += 'The lifetime exception should expires on %s\\n' % str(expires_at)\n text += 'Link to approve or reject this request can be found at the end of the mail\\n'\n text += '\\n'\n text += 'DIDTYPE SCOPE NAME\\n'\n text += '\\n'\n truncated_message = False\n for did in dids:\n did_type = None\n if 'did_type' in did:\n if isinstance(did['did_type'], str) or isinstance(did['did_type'], unicode):\n did_type = DIDType.from_sym(did['did_type'])\n else:\n did_type = did['did_type']\n new_exception = models.LifetimeExceptions(id=exception_id, scope=did['scope'], name=did['name'], did_type=did_type,\n account=account, pattern=pattern, comments=reason, state=LifetimeExceptionsState.WAITING, expires_at=lifetime)\n if len(text) < 3000:\n text += '%s %s %s\\n' % (str(did_type), did['scope'], did['name'])\n else:\n truncated_message = True\n try:\n new_exception.save(session=session, flush=False)\n except IntegrityError as error:\n if match('.*ORA-00001.*', str(error.args[0]))\\\n or match('.*IntegrityError.*UNIQUE constraint failed.*', str(error.args[0]))\\\n or match('.*1062.*Duplicate entry.*for key.*', str(error.args[0]))\\\n or match('.*sqlite3.IntegrityError.*are not unique.*', error.args[0]):\n raise LifetimeExceptionDuplicate()\n raise RucioException(error.args[0])\n if truncated_message:\n text += '...\\n'\n text += 'List too long. Truncated\\n'\n text += '\\n'\n text += 'Approve: https://rucio-ui.cern.ch/lifetime_exception?id=%s&action=approve\\n' % str(exception_id)\n text += 'Deny: https://rucio-ui.cern.ch/lifetime_exception?id=%s&action=deny\\n' % str(exception_id)\n approvers_email = []\n try:\n approvers_email = get('lifetime_model', 'approvers_email', session=session)\n approvers_email = approvers_email.split(',') # pylint: disable=no-member\n except ConfigNotFound:\n approvers_email = []\n\n add_message(event_type='email',\n payload={'body': text, 'to': approvers_email,\n 'subject': '[RUCIO] Request to approve lifetime exception %s' % str(exception_id)},\n session=session)\n return exception_id\n\n\n@transactional_session\ndef update_exception(exception_id, state, session=None):\n \"\"\"\n Update exceptions state to Lifetime Model.\n\n :param exception_id: The id of the exception\n :param state: The states to filter\n :param session: The database session in use.\n \"\"\"\n query = session.query(models.LifetimeExceptions).filter_by(id=exception_id)\n try:\n query.one()\n except NoResultFound:\n raise LifetimeExceptionNotFound\n\n if state in [LifetimeExceptionsState.APPROVED, LifetimeExceptionsState.REJECTED]:\n query.update({'state': state, 'updated_at': datetime.utcnow()}, synchronize_session=False)\n else:\n raise UnsupportedOperation\n\n\n@read_session\ndef define_eol(scope, name, rses, session=None):\n \"\"\"\n ATLAS policy for rules on SCRATCHDISK\n\n :param scope: Scope of the DID.\n :param name: Name of the DID.\n :param rses: List of RSEs.\n :param session: The database session in use.\n \"\"\"\n policy = rucio.common.policy.get_policy()\n if policy != 'atlas':\n return None\n\n # Check if on ATLAS managed space\n if [rse for rse in rses if list_rse_attributes(rse=None, rse_id=rse['id'], session=session).get('type') in ['LOCALGROUPDISK', 'LOCALGROUPTAPE', 'GROUPDISK', 'GROUPTAPE']]:\n return None\n # Now check the lifetime policy\n try:\n did = session.query(models.DataIdentifier).filter(models.DataIdentifier.scope == scope,\n models.DataIdentifier.name == name).one()\n except NoResultFound:\n return None\n policy_dict = rucio.common.policy.get_lifetime_policy()\n did_type = 'other'\n if scope.startswith('mc'):\n did_type = 'mc'\n elif scope.startswith('data'):\n did_type = 'data'\n elif scope.startswith('valid'):\n did_type = 'valid'\n else:\n did_type = 'other'\n for policy in policy_dict[did_type]:\n if 'exclude' in policy:\n to_exclude = False\n for key in policy['exclude']:\n meta_key = None\n if key not in ['datatype', 'project', ]:\n if key == 'stream':\n meta_key = 'stream_name'\n elif key == 'tags':\n meta_key = 'version'\n else:\n meta_key = key\n values = policy['exclude'][key]\n for value in values:\n value = value.replace('%', '.*')\n if meta_key and did[meta_key] and value and match(value, did[meta_key]):\n to_exclude = True\n break\n if to_exclude:\n break\n if to_exclude:\n continue\n if 'include' in policy:\n match_policy = True\n for key in policy['include']:\n meta_key = None\n if key not in ['datatype', 'project', ]:\n if key == 'stream':\n meta_key = 'stream_name'\n elif key == 'tags':\n meta_key = 'version'\n else:\n continue\n else:\n meta_key = key\n values = policy['include'][key]\n to_keep = False\n for value in values:\n value = value.replace('%', '.*')\n if meta_key and did[meta_key] and value and match(value, did[meta_key]):\n to_keep = True\n break\n match_policy = match_policy and to_keep\n if not to_keep:\n match_policy = False\n break\n if match_policy:\n if int(policy['age']) >= 12:\n years = int(int(policy['age']) / 12)\n months = int(policy['age']) - years * 12\n lifetime_value = 365 * years + 30 * months\n else:\n lifetime_value = int(policy['age']) * 30\n if int(policy['extension']) >= 12:\n years = int(int(policy['extension']) / 12)\n months = int(policy['extension']) - years * 12\n extension = 365 * years + 30 * months\n else:\n extension = int(policy['extension']) * 30\n\n default_eol_at = did.created_at + timedelta(days=lifetime_value)\n eol_at = default_eol_at\n if did.accessed_at:\n eol_at = did.accessed_at + timedelta(days=extension)\n if eol_at < default_eol_at:\n eol_at = default_eol_at\n return eol_at\n return None\n", "path": "lib/rucio/core/lifetime_exception.py"}]}
3,769
186
gh_patches_debug_4479
rasdani/github-patches
git_diff
freedomofpress__securedrop-3709
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> [functional testing] Fix staging CI job on tbb-0.9.0 We removed the application/functional test run from the staging environment in #3697. We should also update the testinfra test references and remove the application test run from CI, otherwise we get a few testinfra test failures due to pip deps, and an error when we attempt to run the application tests in CI: ``` TASK [Run application tests] *************************************************** Friday 10 August 2018 19:28:17 +0000 (0:00:00.037) 0:01:08.223 ********* fatal: [app-staging]: FAILED! => {"changed": true, "msg": "non-zero return code", "rc": 127, "stderr": "Shared connection to 52.36.194.59 closed.\r\n", "stdout": "/home/sdrop/.ansible/tmp/ansible-tmp-1533929297.62-93522333058246/app-tests.sh: line 13: pytest: command not found\r\n", "stdout_lines": ["/home/sdrop/.ansible/tmp/ansible-tmp-1533929297.62-93522333058246/app-tests.sh: line 13: pytest: command not found"]} ...ignoring ``` </issue> <code> [start of securedrop/create-dev-data.py] 1 #!/usr/bin/env python 2 # -*- coding: utf-8 -*- 3 4 import datetime 5 import os 6 import sys 7 import argparse 8 from sqlalchemy.exc import IntegrityError 9 10 os.environ["SECUREDROP_ENV"] = "dev" # noqa 11 import journalist_app 12 from sdconfig import config 13 from db import db 14 from models import Journalist, Source, Submission 15 16 17 def add_test_user(username, password, otp_secret, is_admin=False): 18 context = journalist_app.create_app(config).app_context() 19 context.push() 20 21 try: 22 user = Journalist(username=username, 23 password=password, 24 is_admin=is_admin) 25 user.otp_secret = otp_secret 26 db.session.add(user) 27 db.session.commit() 28 print('Test user successfully added: ' 29 'username={}, password={}, otp_secret={}, is_admin={}' 30 ''.format(username, password, otp_secret, is_admin)) 31 except IntegrityError: 32 print("Test user already added") 33 db.session.rollback() 34 35 context.pop() 36 37 38 def create_source_and_submissions(num_submissions=2): 39 app = journalist_app.create_app(config) 40 41 with app.app_context(): 42 # Store source in database 43 codename = app.crypto_util.genrandomid() 44 filesystem_id = app.crypto_util.hash_codename(codename) 45 journalist_designation = app.crypto_util.display_id() 46 source = Source(filesystem_id, journalist_designation) 47 source.pending = False 48 db.session.add(source) 49 db.session.commit() 50 51 # Generate submissions directory and generate source key 52 os.mkdir(app.storage.path(source.filesystem_id)) 53 app.crypto_util.genkeypair(source.filesystem_id, codename) 54 55 # Generate some test submissions 56 for _ in range(num_submissions): 57 source.interaction_count += 1 58 fpath = app.storage.save_message_submission( 59 source.filesystem_id, 60 source.interaction_count, 61 source.journalist_filename, 62 'test submission!' 63 ) 64 source.last_updated = datetime.datetime.utcnow() 65 submission = Submission(source, fpath) 66 db.session.add(submission) 67 68 db.session.commit() 69 print("Test source '{}' added with {} submissions".format( 70 journalist_designation, num_submissions) 71 ) 72 73 74 if __name__ == "__main__": # pragma: no cover 75 # Add two test users 76 test_password = "correct horse battery staple profanity oil chewy" 77 test_otp_secret = "JHCOGO7VCER3EJ4L" 78 79 parser = argparse.ArgumentParser() 80 parser.add_argument("--staging", help="Adding user for staging tests.", 81 action="store_true") 82 args = parser.parse_args() 83 add_test_user("journalist", 84 test_password, 85 test_otp_secret, 86 is_admin=True) 87 88 # If staging, we only need the journalist user (admin) 89 if args.staging: 90 sys.exit(0) 91 92 add_test_user("dellsberg", 93 test_password, 94 test_otp_secret, 95 is_admin=False) 96 97 # Add test sources and submissions 98 num_sources = 2 99 for _ in range(num_sources): 100 create_source_and_submissions() 101 [end of securedrop/create-dev-data.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/securedrop/create-dev-data.py b/securedrop/create-dev-data.py --- a/securedrop/create-dev-data.py +++ b/securedrop/create-dev-data.py @@ -78,7 +78,7 @@ parser = argparse.ArgumentParser() parser.add_argument("--staging", help="Adding user for staging tests.", - action="store_true") + action="store_true") args = parser.parse_args() add_test_user("journalist", test_password,
{"golden_diff": "diff --git a/securedrop/create-dev-data.py b/securedrop/create-dev-data.py\n--- a/securedrop/create-dev-data.py\n+++ b/securedrop/create-dev-data.py\n@@ -78,7 +78,7 @@\n \n parser = argparse.ArgumentParser()\n parser.add_argument(\"--staging\", help=\"Adding user for staging tests.\",\n- action=\"store_true\")\n+ action=\"store_true\")\n args = parser.parse_args()\n add_test_user(\"journalist\",\n test_password,\n", "issue": "[functional testing] Fix staging CI job on tbb-0.9.0\nWe removed the application/functional test run from the staging environment in #3697. We should also update the testinfra test references and remove the application test run from CI, otherwise we get a few testinfra test failures due to pip deps, and an error when we attempt to run the application tests in CI: \r\n\r\n```\r\nTASK [Run application tests] ***************************************************\r\n Friday 10 August 2018 19:28:17 +0000 (0:00:00.037) 0:01:08.223 *********\r\n fatal: [app-staging]: FAILED! => {\"changed\": true, \"msg\": \"non-zero return code\", \"rc\": 127, \"stderr\": \"Shared connection to 52.36.194.59 closed.\\r\\n\", \"stdout\": \"/home/sdrop/.ansible/tmp/ansible-tmp-1533929297.62-93522333058246/app-tests.sh: line 13: pytest: command not found\\r\\n\", \"stdout_lines\": [\"/home/sdrop/.ansible/tmp/ansible-tmp-1533929297.62-93522333058246/app-tests.sh: line 13: pytest: command not found\"]}\r\n ...ignoring\r\n```\n", "before_files": [{"content": "#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\nimport datetime\nimport os\nimport sys\nimport argparse\nfrom sqlalchemy.exc import IntegrityError\n\nos.environ[\"SECUREDROP_ENV\"] = \"dev\" # noqa\nimport journalist_app\nfrom sdconfig import config\nfrom db import db\nfrom models import Journalist, Source, Submission\n\n\ndef add_test_user(username, password, otp_secret, is_admin=False):\n context = journalist_app.create_app(config).app_context()\n context.push()\n\n try:\n user = Journalist(username=username,\n password=password,\n is_admin=is_admin)\n user.otp_secret = otp_secret\n db.session.add(user)\n db.session.commit()\n print('Test user successfully added: '\n 'username={}, password={}, otp_secret={}, is_admin={}'\n ''.format(username, password, otp_secret, is_admin))\n except IntegrityError:\n print(\"Test user already added\")\n db.session.rollback()\n\n context.pop()\n\n\ndef create_source_and_submissions(num_submissions=2):\n app = journalist_app.create_app(config)\n\n with app.app_context():\n # Store source in database\n codename = app.crypto_util.genrandomid()\n filesystem_id = app.crypto_util.hash_codename(codename)\n journalist_designation = app.crypto_util.display_id()\n source = Source(filesystem_id, journalist_designation)\n source.pending = False\n db.session.add(source)\n db.session.commit()\n\n # Generate submissions directory and generate source key\n os.mkdir(app.storage.path(source.filesystem_id))\n app.crypto_util.genkeypair(source.filesystem_id, codename)\n\n # Generate some test submissions\n for _ in range(num_submissions):\n source.interaction_count += 1\n fpath = app.storage.save_message_submission(\n source.filesystem_id,\n source.interaction_count,\n source.journalist_filename,\n 'test submission!'\n )\n source.last_updated = datetime.datetime.utcnow()\n submission = Submission(source, fpath)\n db.session.add(submission)\n\n db.session.commit()\n print(\"Test source '{}' added with {} submissions\".format(\n journalist_designation, num_submissions)\n )\n\n\nif __name__ == \"__main__\": # pragma: no cover\n # Add two test users\n test_password = \"correct horse battery staple profanity oil chewy\"\n test_otp_secret = \"JHCOGO7VCER3EJ4L\"\n\n parser = argparse.ArgumentParser()\n parser.add_argument(\"--staging\", help=\"Adding user for staging tests.\",\n action=\"store_true\")\n args = parser.parse_args()\n add_test_user(\"journalist\",\n test_password,\n test_otp_secret,\n is_admin=True)\n\n # If staging, we only need the journalist user (admin)\n if args.staging:\n sys.exit(0)\n\n add_test_user(\"dellsberg\",\n test_password,\n test_otp_secret,\n is_admin=False)\n\n # Add test sources and submissions\n num_sources = 2\n for _ in range(num_sources):\n create_source_and_submissions()\n", "path": "securedrop/create-dev-data.py"}]}
1,735
108
gh_patches_debug_3856
rasdani/github-patches
git_diff
googleapis__google-auth-library-python-385
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> Docs: compute_engine.credentials doesn't render note for service_account_email On this doc item: https://google-auth.readthedocs.io/en/latest/reference/google.auth.compute_engine.credentials.html#google.auth.compute_engine.credentials.Credentials.service_account_email The rendered docs say simply `The service account email.` However, the [docstring](https://google-auth.readthedocs.io/en/latest/_modules/google/auth/compute_engine/credentials.html#Credentials.refresh) has more information: ``` """The service account email. .. note: This is not guaranteed to be set until :meth`refresh` has been called. """ ``` Why is this important note not rendered? </issue> <code> [start of google/auth/compute_engine/credentials.py] 1 # Copyright 2016 Google Inc. 2 # 3 # Licensed under the Apache License, Version 2.0 (the "License"); 4 # you may not use this file except in compliance with the License. 5 # You may obtain a copy of the License at 6 # 7 # http://www.apache.org/licenses/LICENSE-2.0 8 # 9 # Unless required by applicable law or agreed to in writing, software 10 # distributed under the License is distributed on an "AS IS" BASIS, 11 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 12 # See the License for the specific language governing permissions and 13 # limitations under the License. 14 15 """Google Compute Engine credentials. 16 17 This module provides authentication for application running on Google Compute 18 Engine using the Compute Engine metadata server. 19 20 """ 21 22 import datetime 23 24 import six 25 26 from google.auth import _helpers 27 from google.auth import credentials 28 from google.auth import exceptions 29 from google.auth import iam 30 from google.auth import jwt 31 from google.auth.compute_engine import _metadata 32 from google.oauth2 import _client 33 34 35 class Credentials(credentials.ReadOnlyScoped, credentials.Credentials): 36 """Compute Engine Credentials. 37 38 These credentials use the Google Compute Engine metadata server to obtain 39 OAuth 2.0 access tokens associated with the instance's service account. 40 41 For more information about Compute Engine authentication, including how 42 to configure scopes, see the `Compute Engine authentication 43 documentation`_. 44 45 .. note:: Compute Engine instances can be created with scopes and therefore 46 these credentials are considered to be 'scoped'. However, you can 47 not use :meth:`~google.auth.credentials.ScopedCredentials.with_scopes` 48 because it is not possible to change the scopes that the instance 49 has. Also note that 50 :meth:`~google.auth.credentials.ScopedCredentials.has_scopes` will not 51 work until the credentials have been refreshed. 52 53 .. _Compute Engine authentication documentation: 54 https://cloud.google.com/compute/docs/authentication#using 55 """ 56 57 def __init__(self, service_account_email="default"): 58 """ 59 Args: 60 service_account_email (str): The service account email to use, or 61 'default'. A Compute Engine instance may have multiple service 62 accounts. 63 """ 64 super(Credentials, self).__init__() 65 self._service_account_email = service_account_email 66 67 def _retrieve_info(self, request): 68 """Retrieve information about the service account. 69 70 Updates the scopes and retrieves the full service account email. 71 72 Args: 73 request (google.auth.transport.Request): The object used to make 74 HTTP requests. 75 """ 76 info = _metadata.get_service_account_info( 77 request, service_account=self._service_account_email 78 ) 79 80 self._service_account_email = info["email"] 81 self._scopes = info["scopes"] 82 83 def refresh(self, request): 84 """Refresh the access token and scopes. 85 86 Args: 87 request (google.auth.transport.Request): The object used to make 88 HTTP requests. 89 90 Raises: 91 google.auth.exceptions.RefreshError: If the Compute Engine metadata 92 service can't be reached if if the instance has not 93 credentials. 94 """ 95 try: 96 self._retrieve_info(request) 97 self.token, self.expiry = _metadata.get_service_account_token( 98 request, service_account=self._service_account_email 99 ) 100 except exceptions.TransportError as caught_exc: 101 new_exc = exceptions.RefreshError(caught_exc) 102 six.raise_from(new_exc, caught_exc) 103 104 @property 105 def service_account_email(self): 106 """The service account email. 107 108 .. note: This is not guaranteed to be set until :meth`refresh` has been 109 called. 110 """ 111 return self._service_account_email 112 113 @property 114 def requires_scopes(self): 115 """False: Compute Engine credentials can not be scoped.""" 116 return False 117 118 119 _DEFAULT_TOKEN_LIFETIME_SECS = 3600 # 1 hour in seconds 120 _DEFAULT_TOKEN_URI = "https://www.googleapis.com/oauth2/v4/token" 121 122 123 class IDTokenCredentials(credentials.Credentials, credentials.Signing): 124 """Open ID Connect ID Token-based service account credentials. 125 126 These credentials relies on the default service account of a GCE instance. 127 128 In order for this to work, the GCE instance must have been started with 129 a service account that has access to the IAM Cloud API. 130 """ 131 132 def __init__( 133 self, 134 request, 135 target_audience, 136 token_uri=_DEFAULT_TOKEN_URI, 137 additional_claims=None, 138 service_account_email=None, 139 ): 140 """ 141 Args: 142 request (google.auth.transport.Request): The object used to make 143 HTTP requests. 144 target_audience (str): The intended audience for these credentials, 145 used when requesting the ID Token. The ID Token's ``aud`` claim 146 will be set to this string. 147 token_uri (str): The OAuth 2.0 Token URI. 148 additional_claims (Mapping[str, str]): Any additional claims for 149 the JWT assertion used in the authorization grant. 150 service_account_email (str): Optional explicit service account to 151 use to sign JWT tokens. 152 By default, this is the default GCE service account. 153 """ 154 super(IDTokenCredentials, self).__init__() 155 156 if service_account_email is None: 157 sa_info = _metadata.get_service_account_info(request) 158 service_account_email = sa_info["email"] 159 self._service_account_email = service_account_email 160 161 self._signer = iam.Signer( 162 request=request, 163 credentials=Credentials(), 164 service_account_email=service_account_email, 165 ) 166 167 self._token_uri = token_uri 168 self._target_audience = target_audience 169 170 if additional_claims is not None: 171 self._additional_claims = additional_claims 172 else: 173 self._additional_claims = {} 174 175 def with_target_audience(self, target_audience): 176 """Create a copy of these credentials with the specified target 177 audience. 178 Args: 179 target_audience (str): The intended audience for these credentials, 180 used when requesting the ID Token. 181 Returns: 182 google.auth.service_account.IDTokenCredentials: A new credentials 183 instance. 184 """ 185 return self.__class__( 186 self._signer, 187 service_account_email=self._service_account_email, 188 token_uri=self._token_uri, 189 target_audience=target_audience, 190 additional_claims=self._additional_claims.copy(), 191 ) 192 193 def _make_authorization_grant_assertion(self): 194 """Create the OAuth 2.0 assertion. 195 This assertion is used during the OAuth 2.0 grant to acquire an 196 ID token. 197 Returns: 198 bytes: The authorization grant assertion. 199 """ 200 now = _helpers.utcnow() 201 lifetime = datetime.timedelta(seconds=_DEFAULT_TOKEN_LIFETIME_SECS) 202 expiry = now + lifetime 203 204 payload = { 205 "iat": _helpers.datetime_to_secs(now), 206 "exp": _helpers.datetime_to_secs(expiry), 207 # The issuer must be the service account email. 208 "iss": self.service_account_email, 209 # The audience must be the auth token endpoint's URI 210 "aud": self._token_uri, 211 # The target audience specifies which service the ID token is 212 # intended for. 213 "target_audience": self._target_audience, 214 } 215 216 payload.update(self._additional_claims) 217 218 token = jwt.encode(self._signer, payload) 219 220 return token 221 222 @_helpers.copy_docstring(credentials.Credentials) 223 def refresh(self, request): 224 assertion = self._make_authorization_grant_assertion() 225 access_token, expiry, _ = _client.id_token_jwt_grant( 226 request, self._token_uri, assertion 227 ) 228 self.token = access_token 229 self.expiry = expiry 230 231 @property 232 @_helpers.copy_docstring(credentials.Signing) 233 def signer(self): 234 return self._signer 235 236 @_helpers.copy_docstring(credentials.Signing) 237 def sign_bytes(self, message): 238 return self._signer.sign(message) 239 240 @property 241 def service_account_email(self): 242 """The service account email.""" 243 return self._service_account_email 244 245 @property 246 def signer_email(self): 247 return self._service_account_email 248 [end of google/auth/compute_engine/credentials.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/google/auth/compute_engine/credentials.py b/google/auth/compute_engine/credentials.py --- a/google/auth/compute_engine/credentials.py +++ b/google/auth/compute_engine/credentials.py @@ -105,7 +105,7 @@ def service_account_email(self): """The service account email. - .. note: This is not guaranteed to be set until :meth`refresh` has been + .. note:: This is not guaranteed to be set until :meth:`refresh` has been called. """ return self._service_account_email
{"golden_diff": "diff --git a/google/auth/compute_engine/credentials.py b/google/auth/compute_engine/credentials.py\n--- a/google/auth/compute_engine/credentials.py\n+++ b/google/auth/compute_engine/credentials.py\n@@ -105,7 +105,7 @@\n def service_account_email(self):\n \"\"\"The service account email.\n \n- .. note: This is not guaranteed to be set until :meth`refresh` has been\n+ .. note:: This is not guaranteed to be set until :meth:`refresh` has been\n called.\n \"\"\"\n return self._service_account_email\n", "issue": "Docs: compute_engine.credentials doesn't render note for service_account_email\nOn this doc item: https://google-auth.readthedocs.io/en/latest/reference/google.auth.compute_engine.credentials.html#google.auth.compute_engine.credentials.Credentials.service_account_email\r\n\r\nThe rendered docs say simply `The service account email.`\r\n\r\nHowever, the [docstring](https://google-auth.readthedocs.io/en/latest/_modules/google/auth/compute_engine/credentials.html#Credentials.refresh) has more information:\r\n\r\n```\r\n \"\"\"The service account email.\r\n\r\n .. note: This is not guaranteed to be set until :meth`refresh` has been\r\n called.\r\n \"\"\"\r\n```\r\n\r\nWhy is this important note not rendered?\n", "before_files": [{"content": "# Copyright 2016 Google Inc.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\n\"\"\"Google Compute Engine credentials.\n\nThis module provides authentication for application running on Google Compute\nEngine using the Compute Engine metadata server.\n\n\"\"\"\n\nimport datetime\n\nimport six\n\nfrom google.auth import _helpers\nfrom google.auth import credentials\nfrom google.auth import exceptions\nfrom google.auth import iam\nfrom google.auth import jwt\nfrom google.auth.compute_engine import _metadata\nfrom google.oauth2 import _client\n\n\nclass Credentials(credentials.ReadOnlyScoped, credentials.Credentials):\n \"\"\"Compute Engine Credentials.\n\n These credentials use the Google Compute Engine metadata server to obtain\n OAuth 2.0 access tokens associated with the instance's service account.\n\n For more information about Compute Engine authentication, including how\n to configure scopes, see the `Compute Engine authentication\n documentation`_.\n\n .. note:: Compute Engine instances can be created with scopes and therefore\n these credentials are considered to be 'scoped'. However, you can\n not use :meth:`~google.auth.credentials.ScopedCredentials.with_scopes`\n because it is not possible to change the scopes that the instance\n has. Also note that\n :meth:`~google.auth.credentials.ScopedCredentials.has_scopes` will not\n work until the credentials have been refreshed.\n\n .. _Compute Engine authentication documentation:\n https://cloud.google.com/compute/docs/authentication#using\n \"\"\"\n\n def __init__(self, service_account_email=\"default\"):\n \"\"\"\n Args:\n service_account_email (str): The service account email to use, or\n 'default'. A Compute Engine instance may have multiple service\n accounts.\n \"\"\"\n super(Credentials, self).__init__()\n self._service_account_email = service_account_email\n\n def _retrieve_info(self, request):\n \"\"\"Retrieve information about the service account.\n\n Updates the scopes and retrieves the full service account email.\n\n Args:\n request (google.auth.transport.Request): The object used to make\n HTTP requests.\n \"\"\"\n info = _metadata.get_service_account_info(\n request, service_account=self._service_account_email\n )\n\n self._service_account_email = info[\"email\"]\n self._scopes = info[\"scopes\"]\n\n def refresh(self, request):\n \"\"\"Refresh the access token and scopes.\n\n Args:\n request (google.auth.transport.Request): The object used to make\n HTTP requests.\n\n Raises:\n google.auth.exceptions.RefreshError: If the Compute Engine metadata\n service can't be reached if if the instance has not\n credentials.\n \"\"\"\n try:\n self._retrieve_info(request)\n self.token, self.expiry = _metadata.get_service_account_token(\n request, service_account=self._service_account_email\n )\n except exceptions.TransportError as caught_exc:\n new_exc = exceptions.RefreshError(caught_exc)\n six.raise_from(new_exc, caught_exc)\n\n @property\n def service_account_email(self):\n \"\"\"The service account email.\n\n .. note: This is not guaranteed to be set until :meth`refresh` has been\n called.\n \"\"\"\n return self._service_account_email\n\n @property\n def requires_scopes(self):\n \"\"\"False: Compute Engine credentials can not be scoped.\"\"\"\n return False\n\n\n_DEFAULT_TOKEN_LIFETIME_SECS = 3600 # 1 hour in seconds\n_DEFAULT_TOKEN_URI = \"https://www.googleapis.com/oauth2/v4/token\"\n\n\nclass IDTokenCredentials(credentials.Credentials, credentials.Signing):\n \"\"\"Open ID Connect ID Token-based service account credentials.\n\n These credentials relies on the default service account of a GCE instance.\n\n In order for this to work, the GCE instance must have been started with\n a service account that has access to the IAM Cloud API.\n \"\"\"\n\n def __init__(\n self,\n request,\n target_audience,\n token_uri=_DEFAULT_TOKEN_URI,\n additional_claims=None,\n service_account_email=None,\n ):\n \"\"\"\n Args:\n request (google.auth.transport.Request): The object used to make\n HTTP requests.\n target_audience (str): The intended audience for these credentials,\n used when requesting the ID Token. The ID Token's ``aud`` claim\n will be set to this string.\n token_uri (str): The OAuth 2.0 Token URI.\n additional_claims (Mapping[str, str]): Any additional claims for\n the JWT assertion used in the authorization grant.\n service_account_email (str): Optional explicit service account to\n use to sign JWT tokens.\n By default, this is the default GCE service account.\n \"\"\"\n super(IDTokenCredentials, self).__init__()\n\n if service_account_email is None:\n sa_info = _metadata.get_service_account_info(request)\n service_account_email = sa_info[\"email\"]\n self._service_account_email = service_account_email\n\n self._signer = iam.Signer(\n request=request,\n credentials=Credentials(),\n service_account_email=service_account_email,\n )\n\n self._token_uri = token_uri\n self._target_audience = target_audience\n\n if additional_claims is not None:\n self._additional_claims = additional_claims\n else:\n self._additional_claims = {}\n\n def with_target_audience(self, target_audience):\n \"\"\"Create a copy of these credentials with the specified target\n audience.\n Args:\n target_audience (str): The intended audience for these credentials,\n used when requesting the ID Token.\n Returns:\n google.auth.service_account.IDTokenCredentials: A new credentials\n instance.\n \"\"\"\n return self.__class__(\n self._signer,\n service_account_email=self._service_account_email,\n token_uri=self._token_uri,\n target_audience=target_audience,\n additional_claims=self._additional_claims.copy(),\n )\n\n def _make_authorization_grant_assertion(self):\n \"\"\"Create the OAuth 2.0 assertion.\n This assertion is used during the OAuth 2.0 grant to acquire an\n ID token.\n Returns:\n bytes: The authorization grant assertion.\n \"\"\"\n now = _helpers.utcnow()\n lifetime = datetime.timedelta(seconds=_DEFAULT_TOKEN_LIFETIME_SECS)\n expiry = now + lifetime\n\n payload = {\n \"iat\": _helpers.datetime_to_secs(now),\n \"exp\": _helpers.datetime_to_secs(expiry),\n # The issuer must be the service account email.\n \"iss\": self.service_account_email,\n # The audience must be the auth token endpoint's URI\n \"aud\": self._token_uri,\n # The target audience specifies which service the ID token is\n # intended for.\n \"target_audience\": self._target_audience,\n }\n\n payload.update(self._additional_claims)\n\n token = jwt.encode(self._signer, payload)\n\n return token\n\n @_helpers.copy_docstring(credentials.Credentials)\n def refresh(self, request):\n assertion = self._make_authorization_grant_assertion()\n access_token, expiry, _ = _client.id_token_jwt_grant(\n request, self._token_uri, assertion\n )\n self.token = access_token\n self.expiry = expiry\n\n @property\n @_helpers.copy_docstring(credentials.Signing)\n def signer(self):\n return self._signer\n\n @_helpers.copy_docstring(credentials.Signing)\n def sign_bytes(self, message):\n return self._signer.sign(message)\n\n @property\n def service_account_email(self):\n \"\"\"The service account email.\"\"\"\n return self._service_account_email\n\n @property\n def signer_email(self):\n return self._service_account_email\n", "path": "google/auth/compute_engine/credentials.py"}]}
3,081
128
gh_patches_debug_57973
rasdani/github-patches
git_diff
pyjanitor-devs__pyjanitor-1191
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> [INF/CI] Add `--cov-append` for `pytest` <!-- Thank you for your PR! BEFORE YOU CONTINUE! Please add the appropriate three-letter abbreviation to your title. The abbreviations can be: - [DOC]: Documentation fixes. - [ENH]: Code contributions and new features. - [TST]: Test-related contributions. - [INF]: Infrastructure-related contributions. Also, do not forget to tag the relevant issue here as well. Finally, as commits come in, don't forget to regularly rebase! --> # PR Description Please describe the changes proposed in the pull request: > Another reason code coverage failed is that pytest doesn't add `--cov-append` option. `--cov-append` can get a sum coverage. I'll add this option in the next PR. First let us merge `codecov.yml` into `tests.yml`. Keep the same test logic for the dev branch or a PR. _Originally posted by @Zeroto521 in https://github.com/pyjanitor-devs/pyjanitor/issues/1185#issuecomment-1296479926_ <!-- Doing so provides maintainers with context on what the PR is, and can help us more effectively review your PR. --> <!-- Please also identify below which issue that has been raised that you are going to close. --> <!-- As you go down the PR template, please feel free to delete sections that are irrelevant. --> # PR Checklist <!-- This checklist exists for newcomers who are not yet familiar with our requirements. If you are experienced with the project, please feel free to delete this section. --> Please ensure that you have done the following: 1. [x] PR in from a fork off your branch. Do not PR from `<your_username>`:`dev`, but rather from `<your_username>`:`<feature-branch_name>`. <!-- Doing this helps us keep the commit history much cleaner than it would otherwise be. --> 2. [x] If you're not on the contributors list, add yourself to `AUTHORS.md`. <!-- We'd like to acknowledge your contributions! --> 3. [x] Add a line to `CHANGELOG.md` under the latest version header (i.e. the one that is "on deck") describing the contribution. - Do use some discretion here; if there are multiple PRs that are related, keep them in a single line. # Automatic checks There will be automatic checks run on the PR. These include: - Building a preview of the docs on Netlify - Automatically linting the code - Making sure the code is documented - Making sure that all tests are passed - Making sure that code coverage doesn't go down. # Relevant Reviewers <!-- Finally, please tag relevant maintainers to review. --> Please tag maintainers to review. - @ericmjl </issue> <code> [start of janitor/accessors/__init__.py] 1 """Miscellaneous mathematical operators. 2 3 Lazy loading used here to speed up imports. 4 """ 5 6 import warnings 7 from typing import Tuple 8 9 10 import lazy_loader as lazy 11 12 scipy_special = lazy.load("scipy.special") 13 ss = lazy.load("scipy.stats") 14 pf = lazy.load("pandas_flavor") 15 pd = lazy.load("pandas") 16 np = lazy.load("numpy") 17 pdtypes = lazy.load("pandas.api.types") 18 [end of janitor/accessors/__init__.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/janitor/accessors/__init__.py b/janitor/accessors/__init__.py --- a/janitor/accessors/__init__.py +++ b/janitor/accessors/__init__.py @@ -1,17 +1,3 @@ -"""Miscellaneous mathematical operators. +"""Miscellaneous mathematical operators.""" -Lazy loading used here to speed up imports. -""" - -import warnings -from typing import Tuple - - -import lazy_loader as lazy - -scipy_special = lazy.load("scipy.special") -ss = lazy.load("scipy.stats") -pf = lazy.load("pandas_flavor") -pd = lazy.load("pandas") -np = lazy.load("numpy") -pdtypes = lazy.load("pandas.api.types") +from janitor.accessors.data_description import DataDescription # noqa: F401
{"golden_diff": "diff --git a/janitor/accessors/__init__.py b/janitor/accessors/__init__.py\n--- a/janitor/accessors/__init__.py\n+++ b/janitor/accessors/__init__.py\n@@ -1,17 +1,3 @@\n-\"\"\"Miscellaneous mathematical operators.\n+\"\"\"Miscellaneous mathematical operators.\"\"\"\n \n-Lazy loading used here to speed up imports.\n-\"\"\"\n-\n-import warnings\n-from typing import Tuple\n-\n-\n-import lazy_loader as lazy\n-\n-scipy_special = lazy.load(\"scipy.special\")\n-ss = lazy.load(\"scipy.stats\")\n-pf = lazy.load(\"pandas_flavor\")\n-pd = lazy.load(\"pandas\")\n-np = lazy.load(\"numpy\")\n-pdtypes = lazy.load(\"pandas.api.types\")\n+from janitor.accessors.data_description import DataDescription # noqa: F401\n", "issue": "[INF/CI] Add `--cov-append` for `pytest`\n<!-- Thank you for your PR!\r\n\r\nBEFORE YOU CONTINUE! Please add the appropriate three-letter abbreviation to your title.\r\n\r\nThe abbreviations can be:\r\n- [DOC]: Documentation fixes.\r\n- [ENH]: Code contributions and new features.\r\n- [TST]: Test-related contributions.\r\n- [INF]: Infrastructure-related contributions.\r\n\r\nAlso, do not forget to tag the relevant issue here as well.\r\n\r\nFinally, as commits come in, don't forget to regularly rebase!\r\n-->\r\n\r\n# PR Description\r\n\r\nPlease describe the changes proposed in the pull request:\r\n\r\n> Another reason code coverage failed is that pytest doesn't add `--cov-append` option.\r\n`--cov-append` can get a sum coverage. I'll add this option in the next PR.\r\nFirst let us merge `codecov.yml` into `tests.yml`. Keep the same test logic for the dev branch or a PR.\r\n\r\n_Originally posted by @Zeroto521 in https://github.com/pyjanitor-devs/pyjanitor/issues/1185#issuecomment-1296479926_\r\n\r\n<!-- Doing so provides maintainers with context on what the PR is, and can help us more effectively review your PR. -->\r\n\r\n<!-- Please also identify below which issue that has been raised that you are going to close. -->\r\n\r\n<!-- As you go down the PR template, please feel free to delete sections that are irrelevant. -->\r\n\r\n# PR Checklist\r\n\r\n<!-- This checklist exists for newcomers who are not yet familiar with our requirements. If you are experienced with\r\nthe project, please feel free to delete this section. -->\r\n\r\nPlease ensure that you have done the following:\r\n\r\n1. [x] PR in from a fork off your branch. Do not PR from `<your_username>`:`dev`, but rather from `<your_username>`:`<feature-branch_name>`.\r\n<!-- Doing this helps us keep the commit history much cleaner than it would otherwise be. -->\r\n2. [x] If you're not on the contributors list, add yourself to `AUTHORS.md`.\r\n<!-- We'd like to acknowledge your contributions! -->\r\n3. [x] Add a line to `CHANGELOG.md` under the latest version header (i.e. the one that is \"on deck\") describing the contribution.\r\n - Do use some discretion here; if there are multiple PRs that are related, keep them in a single line.\r\n\r\n# Automatic checks\r\n\r\nThere will be automatic checks run on the PR. These include:\r\n\r\n- Building a preview of the docs on Netlify\r\n- Automatically linting the code\r\n- Making sure the code is documented\r\n- Making sure that all tests are passed\r\n- Making sure that code coverage doesn't go down.\r\n\r\n# Relevant Reviewers\r\n\r\n<!-- Finally, please tag relevant maintainers to review. -->\r\n\r\nPlease tag maintainers to review.\r\n\r\n- @ericmjl\r\n\n", "before_files": [{"content": "\"\"\"Miscellaneous mathematical operators.\n\nLazy loading used here to speed up imports.\n\"\"\"\n\nimport warnings\nfrom typing import Tuple\n\n\nimport lazy_loader as lazy\n\nscipy_special = lazy.load(\"scipy.special\")\nss = lazy.load(\"scipy.stats\")\npf = lazy.load(\"pandas_flavor\")\npd = lazy.load(\"pandas\")\nnp = lazy.load(\"numpy\")\npdtypes = lazy.load(\"pandas.api.types\")\n", "path": "janitor/accessors/__init__.py"}]}
1,251
185
gh_patches_debug_38001
rasdani/github-patches
git_diff
biolab__orange3-1907
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> TypeError: OneClassSVMLearner doesn't support multiple class variables https://sentry.io/biolab/orange3/issues/202498152/ ``` TypeError: OneClassSVMLearner doesn't support multiple class variables Module "Orange.widgets.data.owoutliers", line 141, in commit y_pred = self.detect_outliers() Module "Orange.widgets.data.owoutliers", line 169, in detect_outliers model = learner(self.data) Module "Orange.base", line 253, in __call__ m = super().__call__(data) Module "Orange.base", line 52, in __call__ self.__class__.__name__) ``` </issue> <code> [start of Orange/widgets/data/owoutliers.py] 1 import sys 2 3 import numpy as np 4 from AnyQt.QtWidgets import QLayout 5 6 from Orange.base import SklLearner 7 from Orange.classification import OneClassSVMLearner, EllipticEnvelopeLearner 8 from Orange.data import Table, Domain, ContinuousVariable 9 from Orange.widgets import widget, gui 10 from Orange.widgets.settings import Setting 11 from Orange.widgets.utils.sql import check_sql_input 12 13 14 class OWOutliers(widget.OWWidget): 15 name = "Outliers" 16 description = "Detect outliers." 17 icon = "icons/Outliers.svg" 18 priority = 3000 19 category = "Data" 20 keywords = ["data", "outlier", "inlier"] 21 22 inputs = [("Data", Table, "set_data")] 23 outputs = [("Inliers", Table), ("Outliers", Table)] 24 25 want_main_area = False 26 27 OneClassSVM, Covariance = range(2) 28 29 outlier_method = Setting(OneClassSVM) 30 nu = Setting(50) 31 gamma = Setting(0.01) 32 cont = Setting(10) 33 empirical_covariance = Setting(False) 34 support_fraction = Setting(1) 35 36 data_info_default = 'No data on input.' 37 in_out_info_default = ' ' 38 39 def __init__(self): 40 super().__init__() 41 self.data = None 42 self.n_inliers = self.n_outliers = None 43 44 box = gui.vBox(self.controlArea, "Information") 45 self.data_info_label = gui.widgetLabel(box, self.data_info_default) 46 self.in_out_info_label = gui.widgetLabel(box, 47 self.in_out_info_default) 48 49 box = gui.vBox(self.controlArea, "Outlier Detection Method") 50 detection = gui.radioButtons(box, self, "outlier_method") 51 52 gui.appendRadioButton(detection, 53 "One class SVM with non-linear kernel (RBF)") 54 ibox = gui.indentedBox(detection) 55 tooltip = "An upper bound on the fraction of training errors and a " \ 56 "lower bound of the fraction of support vectors" 57 gui.widgetLabel(ibox, 'Nu:', tooltip=tooltip) 58 self.nu_slider = gui.hSlider( 59 ibox, self, "nu", minValue=1, maxValue=100, ticks=10, 60 labelFormat="%d %%", callback=self.nu_changed, tooltip=tooltip) 61 self.gamma_spin = gui.spin( 62 ibox, self, "gamma", label="Kernel coefficient:", step=1e-2, 63 spinType=float, minv=0.01, maxv=10, callback=self.gamma_changed) 64 gui.separator(detection, 12) 65 66 self.rb_cov = gui.appendRadioButton(detection, "Covariance estimator") 67 ibox = gui.indentedBox(detection) 68 self.l_cov = gui.widgetLabel(ibox, 'Contamination:') 69 self.cont_slider = gui.hSlider( 70 ibox, self, "cont", minValue=0, maxValue=100, ticks=10, 71 labelFormat="%d %%", callback=self.cont_changed) 72 73 ebox = gui.hBox(ibox) 74 self.cb_emp_cov = gui.checkBox( 75 ebox, self, "empirical_covariance", 76 "Support fraction:", callback=self.empirical_changed) 77 self.support_fraction_spin = gui.spin( 78 ebox, self, "support_fraction", step=1e-1, spinType=float, 79 minv=0.1, maxv=10, callback=self.support_fraction_changed) 80 81 gui.separator(detection, 12) 82 83 gui.button(self.buttonsArea, self, "Detect Outliers", 84 callback=self.commit) 85 self.layout().setSizeConstraint(QLayout.SetFixedSize) 86 87 def nu_changed(self): 88 self.outlier_method = self.OneClassSVM 89 90 def gamma_changed(self): 91 self.outlier_method = self.OneClassSVM 92 93 def cont_changed(self): 94 self.outlier_method = self.Covariance 95 96 def support_fraction_changed(self): 97 self.outlier_method = self.Covariance 98 99 def empirical_changed(self): 100 self.outlier_method = self.Covariance 101 102 def disable_covariance(self): 103 self.outlier_method = self.OneClassSVM 104 self.rb_cov.setDisabled(True) 105 self.l_cov.setDisabled(True) 106 self.cont_slider.setDisabled(True) 107 self.cb_emp_cov.setDisabled(True) 108 self.support_fraction_spin.setDisabled(True) 109 self.warning('Too many features for covariance estimation.') 110 111 def enable_covariance(self): 112 self.rb_cov.setDisabled(False) 113 self.l_cov.setDisabled(False) 114 self.cont_slider.setDisabled(False) 115 self.cb_emp_cov.setDisabled(False) 116 self.support_fraction_spin.setDisabled(False) 117 self.warning() 118 119 @check_sql_input 120 def set_data(self, dataset): 121 self.data = dataset 122 if self.data is None: 123 self.data_info_label.setText(self.data_info_default) 124 self.in_out_info_label.setText(self.in_out_info_default) 125 else: 126 self.data_info_label.setText('%d instances' % len(self.data)) 127 self.in_out_info_label.setText(' ') 128 129 self.enable_covariance() 130 if self.data and len(self.data.domain.attributes) > 1500: 131 self.disable_covariance() 132 133 self.commit() 134 135 def commit(self): 136 self.error() 137 inliers = outliers = None 138 self.n_inliers = self.n_outliers = None 139 if self.data is not None and len(self.data) > 0: 140 try: 141 y_pred = self.detect_outliers() 142 except ValueError: 143 self.error("Singular covariance matrix.") 144 self.in_out_info_label.setText(self.in_out_info_default) 145 else: 146 inliers_ind = np.where(y_pred == 1)[0] 147 outliers_ind = np.where(y_pred == -1)[0] 148 inliers = Table(self.new_domain, self.new_data, inliers_ind) 149 outliers = Table(self.new_domain, 150 self.new_data, outliers_ind) 151 self.in_out_info_label.setText('%d inliers, %d outliers' % 152 (len(inliers), len(outliers))) 153 self.n_inliers = len(inliers) 154 self.n_outliers = len(outliers) 155 156 self.send("Inliers", inliers) 157 self.send("Outliers", outliers) 158 159 def detect_outliers(self): 160 if self.outlier_method == self.OneClassSVM: 161 learner = OneClassSVMLearner( 162 gamma=self.gamma, nu=self.nu / 100, 163 preprocessors=SklLearner.preprocessors) 164 else: 165 learner = EllipticEnvelopeLearner( 166 support_fraction=self.support_fraction 167 if self.empirical_covariance else None, 168 contamination=self.cont / 100.) 169 model = learner(self.data) 170 y_pred = model(self.data) 171 self.add_metas(model) 172 return np.array(y_pred) 173 174 def add_metas(self, model): 175 if self.outlier_method == self.Covariance: 176 mahal = model.mahalanobis(self.data.X) 177 mahal = mahal.reshape(len(self.data), 1) 178 attrs = self.data.domain.attributes 179 classes = self.data.domain.class_vars 180 new_metas = list(self.data.domain.metas) + \ 181 [ContinuousVariable(name="Mahalanobis")] 182 self.new_domain = Domain(attrs, classes, new_metas) 183 self.new_data = Table(self.new_domain, self.data) 184 self.new_data.metas = np.hstack((self.data.metas, mahal)) 185 else: 186 self.new_domain = self.data.domain 187 self.new_data = self.data 188 189 def send_report(self): 190 if self.n_outliers is None or self.n_inliers is None: 191 return 192 self.report_items("Data", 193 (("Input instances", len(self.data)), 194 ("Inliers", self.n_inliers), 195 ("Outliers", self.n_outliers))) 196 if self.outlier_method == 0: 197 self.report_items( 198 "Detection", 199 (("Detection method", 200 "One class SVM with non-linear kernel (RBF)"), 201 ("Regularization (nu)", self.nu), 202 ("Kernel coefficient", self.gamma))) 203 else: 204 self.report_items( 205 "Detection", 206 (("Detection method", "Covariance estimator"), 207 ("Contamination", self.cont), 208 ("Support fraction", self.support_fraction))) 209 210 def test_main(): 211 from AnyQt.QtWidgets import QApplication 212 app = QApplication([]) 213 data = Table("iris") 214 w = OWOutliers() 215 w.set_data(data) 216 w.commit() 217 w.show() 218 return app.exec_() 219 220 221 if __name__ == "__main__": 222 sys.exit(test_main()) 223 [end of Orange/widgets/data/owoutliers.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/Orange/widgets/data/owoutliers.py b/Orange/widgets/data/owoutliers.py --- a/Orange/widgets/data/owoutliers.py +++ b/Orange/widgets/data/owoutliers.py @@ -8,6 +8,7 @@ from Orange.data import Table, Domain, ContinuousVariable from Orange.widgets import widget, gui from Orange.widgets.settings import Setting +from Orange.widgets.widget import Msg from Orange.widgets.utils.sql import check_sql_input @@ -36,6 +37,10 @@ data_info_default = 'No data on input.' in_out_info_default = ' ' + class Error(widget.OWWidget.Error): + singular_cov = Msg("Singular covariance matrix.") + multiclass_error = Msg("Multiple class data is not supported") + def __init__(self): super().__init__() self.data = None @@ -133,25 +138,29 @@ self.commit() def commit(self): - self.error() + self.clear_messages() inliers = outliers = None self.n_inliers = self.n_outliers = None if self.data is not None and len(self.data) > 0: - try: - y_pred = self.detect_outliers() - except ValueError: - self.error("Singular covariance matrix.") - self.in_out_info_label.setText(self.in_out_info_default) + if self.data.Y.ndim > 1: + self.Error.multiclass_error() else: - inliers_ind = np.where(y_pred == 1)[0] - outliers_ind = np.where(y_pred == -1)[0] - inliers = Table(self.new_domain, self.new_data, inliers_ind) - outliers = Table(self.new_domain, - self.new_data, outliers_ind) - self.in_out_info_label.setText('%d inliers, %d outliers' % - (len(inliers), len(outliers))) - self.n_inliers = len(inliers) - self.n_outliers = len(outliers) + try: + y_pred = self.detect_outliers() + except ValueError: + self.Error.singular_cov() + self.in_out_info_label.setText(self.in_out_info_default) + else: + inliers_ind = np.where(y_pred == 1)[0] + outliers_ind = np.where(y_pred == -1)[0] + inliers = Table(self.new_domain, self.new_data, inliers_ind) + outliers = Table(self.new_domain, + self.new_data, outliers_ind) + self.in_out_info_label.setText( + "{} inliers, {} outliers".format(len(inliers), + len(outliers))) + self.n_inliers = len(inliers) + self.n_outliers = len(outliers) self.send("Inliers", inliers) self.send("Outliers", outliers)
{"golden_diff": "diff --git a/Orange/widgets/data/owoutliers.py b/Orange/widgets/data/owoutliers.py\n--- a/Orange/widgets/data/owoutliers.py\n+++ b/Orange/widgets/data/owoutliers.py\n@@ -8,6 +8,7 @@\n from Orange.data import Table, Domain, ContinuousVariable\n from Orange.widgets import widget, gui\n from Orange.widgets.settings import Setting\n+from Orange.widgets.widget import Msg\n from Orange.widgets.utils.sql import check_sql_input\n \n \n@@ -36,6 +37,10 @@\n data_info_default = 'No data on input.'\n in_out_info_default = ' '\n \n+ class Error(widget.OWWidget.Error):\n+ singular_cov = Msg(\"Singular covariance matrix.\")\n+ multiclass_error = Msg(\"Multiple class data is not supported\")\n+\n def __init__(self):\n super().__init__()\n self.data = None\n@@ -133,25 +138,29 @@\n self.commit()\n \n def commit(self):\n- self.error()\n+ self.clear_messages()\n inliers = outliers = None\n self.n_inliers = self.n_outliers = None\n if self.data is not None and len(self.data) > 0:\n- try:\n- y_pred = self.detect_outliers()\n- except ValueError:\n- self.error(\"Singular covariance matrix.\")\n- self.in_out_info_label.setText(self.in_out_info_default)\n+ if self.data.Y.ndim > 1:\n+ self.Error.multiclass_error()\n else:\n- inliers_ind = np.where(y_pred == 1)[0]\n- outliers_ind = np.where(y_pred == -1)[0]\n- inliers = Table(self.new_domain, self.new_data, inliers_ind)\n- outliers = Table(self.new_domain,\n- self.new_data, outliers_ind)\n- self.in_out_info_label.setText('%d inliers, %d outliers' %\n- (len(inliers), len(outliers)))\n- self.n_inliers = len(inliers)\n- self.n_outliers = len(outliers)\n+ try:\n+ y_pred = self.detect_outliers()\n+ except ValueError:\n+ self.Error.singular_cov()\n+ self.in_out_info_label.setText(self.in_out_info_default)\n+ else:\n+ inliers_ind = np.where(y_pred == 1)[0]\n+ outliers_ind = np.where(y_pred == -1)[0]\n+ inliers = Table(self.new_domain, self.new_data, inliers_ind)\n+ outliers = Table(self.new_domain,\n+ self.new_data, outliers_ind)\n+ self.in_out_info_label.setText(\n+ \"{} inliers, {} outliers\".format(len(inliers),\n+ len(outliers)))\n+ self.n_inliers = len(inliers)\n+ self.n_outliers = len(outliers)\n \n self.send(\"Inliers\", inliers)\n self.send(\"Outliers\", outliers)\n", "issue": "TypeError: OneClassSVMLearner doesn't support multiple class variables\nhttps://sentry.io/biolab/orange3/issues/202498152/\n\n```\nTypeError: OneClassSVMLearner doesn't support multiple class variables\n Module \"Orange.widgets.data.owoutliers\", line 141, in commit\n\n y_pred = self.detect_outliers()\n Module \"Orange.widgets.data.owoutliers\", line 169, in detect_outliers\n\n model = learner(self.data)\n Module \"Orange.base\", line 253, in __call__\n\n m = super().__call__(data)\n Module \"Orange.base\", line 52, in __call__\n\n self.__class__.__name__)\n```\n", "before_files": [{"content": "import sys\n\nimport numpy as np\nfrom AnyQt.QtWidgets import QLayout\n\nfrom Orange.base import SklLearner\nfrom Orange.classification import OneClassSVMLearner, EllipticEnvelopeLearner\nfrom Orange.data import Table, Domain, ContinuousVariable\nfrom Orange.widgets import widget, gui\nfrom Orange.widgets.settings import Setting\nfrom Orange.widgets.utils.sql import check_sql_input\n\n\nclass OWOutliers(widget.OWWidget):\n name = \"Outliers\"\n description = \"Detect outliers.\"\n icon = \"icons/Outliers.svg\"\n priority = 3000\n category = \"Data\"\n keywords = [\"data\", \"outlier\", \"inlier\"]\n\n inputs = [(\"Data\", Table, \"set_data\")]\n outputs = [(\"Inliers\", Table), (\"Outliers\", Table)]\n\n want_main_area = False\n\n OneClassSVM, Covariance = range(2)\n\n outlier_method = Setting(OneClassSVM)\n nu = Setting(50)\n gamma = Setting(0.01)\n cont = Setting(10)\n empirical_covariance = Setting(False)\n support_fraction = Setting(1)\n\n data_info_default = 'No data on input.'\n in_out_info_default = ' '\n\n def __init__(self):\n super().__init__()\n self.data = None\n self.n_inliers = self.n_outliers = None\n\n box = gui.vBox(self.controlArea, \"Information\")\n self.data_info_label = gui.widgetLabel(box, self.data_info_default)\n self.in_out_info_label = gui.widgetLabel(box,\n self.in_out_info_default)\n\n box = gui.vBox(self.controlArea, \"Outlier Detection Method\")\n detection = gui.radioButtons(box, self, \"outlier_method\")\n\n gui.appendRadioButton(detection,\n \"One class SVM with non-linear kernel (RBF)\")\n ibox = gui.indentedBox(detection)\n tooltip = \"An upper bound on the fraction of training errors and a \" \\\n \"lower bound of the fraction of support vectors\"\n gui.widgetLabel(ibox, 'Nu:', tooltip=tooltip)\n self.nu_slider = gui.hSlider(\n ibox, self, \"nu\", minValue=1, maxValue=100, ticks=10,\n labelFormat=\"%d %%\", callback=self.nu_changed, tooltip=tooltip)\n self.gamma_spin = gui.spin(\n ibox, self, \"gamma\", label=\"Kernel coefficient:\", step=1e-2,\n spinType=float, minv=0.01, maxv=10, callback=self.gamma_changed)\n gui.separator(detection, 12)\n\n self.rb_cov = gui.appendRadioButton(detection, \"Covariance estimator\")\n ibox = gui.indentedBox(detection)\n self.l_cov = gui.widgetLabel(ibox, 'Contamination:')\n self.cont_slider = gui.hSlider(\n ibox, self, \"cont\", minValue=0, maxValue=100, ticks=10,\n labelFormat=\"%d %%\", callback=self.cont_changed)\n\n ebox = gui.hBox(ibox)\n self.cb_emp_cov = gui.checkBox(\n ebox, self, \"empirical_covariance\",\n \"Support fraction:\", callback=self.empirical_changed)\n self.support_fraction_spin = gui.spin(\n ebox, self, \"support_fraction\", step=1e-1, spinType=float,\n minv=0.1, maxv=10, callback=self.support_fraction_changed)\n\n gui.separator(detection, 12)\n\n gui.button(self.buttonsArea, self, \"Detect Outliers\",\n callback=self.commit)\n self.layout().setSizeConstraint(QLayout.SetFixedSize)\n\n def nu_changed(self):\n self.outlier_method = self.OneClassSVM\n\n def gamma_changed(self):\n self.outlier_method = self.OneClassSVM\n\n def cont_changed(self):\n self.outlier_method = self.Covariance\n\n def support_fraction_changed(self):\n self.outlier_method = self.Covariance\n\n def empirical_changed(self):\n self.outlier_method = self.Covariance\n\n def disable_covariance(self):\n self.outlier_method = self.OneClassSVM\n self.rb_cov.setDisabled(True)\n self.l_cov.setDisabled(True)\n self.cont_slider.setDisabled(True)\n self.cb_emp_cov.setDisabled(True)\n self.support_fraction_spin.setDisabled(True)\n self.warning('Too many features for covariance estimation.')\n\n def enable_covariance(self):\n self.rb_cov.setDisabled(False)\n self.l_cov.setDisabled(False)\n self.cont_slider.setDisabled(False)\n self.cb_emp_cov.setDisabled(False)\n self.support_fraction_spin.setDisabled(False)\n self.warning()\n\n @check_sql_input\n def set_data(self, dataset):\n self.data = dataset\n if self.data is None:\n self.data_info_label.setText(self.data_info_default)\n self.in_out_info_label.setText(self.in_out_info_default)\n else:\n self.data_info_label.setText('%d instances' % len(self.data))\n self.in_out_info_label.setText(' ')\n\n self.enable_covariance()\n if self.data and len(self.data.domain.attributes) > 1500:\n self.disable_covariance()\n\n self.commit()\n\n def commit(self):\n self.error()\n inliers = outliers = None\n self.n_inliers = self.n_outliers = None\n if self.data is not None and len(self.data) > 0:\n try:\n y_pred = self.detect_outliers()\n except ValueError:\n self.error(\"Singular covariance matrix.\")\n self.in_out_info_label.setText(self.in_out_info_default)\n else:\n inliers_ind = np.where(y_pred == 1)[0]\n outliers_ind = np.where(y_pred == -1)[0]\n inliers = Table(self.new_domain, self.new_data, inliers_ind)\n outliers = Table(self.new_domain,\n self.new_data, outliers_ind)\n self.in_out_info_label.setText('%d inliers, %d outliers' %\n (len(inliers), len(outliers)))\n self.n_inliers = len(inliers)\n self.n_outliers = len(outliers)\n\n self.send(\"Inliers\", inliers)\n self.send(\"Outliers\", outliers)\n\n def detect_outliers(self):\n if self.outlier_method == self.OneClassSVM:\n learner = OneClassSVMLearner(\n gamma=self.gamma, nu=self.nu / 100,\n preprocessors=SklLearner.preprocessors)\n else:\n learner = EllipticEnvelopeLearner(\n support_fraction=self.support_fraction\n if self.empirical_covariance else None,\n contamination=self.cont / 100.)\n model = learner(self.data)\n y_pred = model(self.data)\n self.add_metas(model)\n return np.array(y_pred)\n\n def add_metas(self, model):\n if self.outlier_method == self.Covariance:\n mahal = model.mahalanobis(self.data.X)\n mahal = mahal.reshape(len(self.data), 1)\n attrs = self.data.domain.attributes\n classes = self.data.domain.class_vars\n new_metas = list(self.data.domain.metas) + \\\n [ContinuousVariable(name=\"Mahalanobis\")]\n self.new_domain = Domain(attrs, classes, new_metas)\n self.new_data = Table(self.new_domain, self.data)\n self.new_data.metas = np.hstack((self.data.metas, mahal))\n else:\n self.new_domain = self.data.domain\n self.new_data = self.data\n\n def send_report(self):\n if self.n_outliers is None or self.n_inliers is None:\n return\n self.report_items(\"Data\",\n ((\"Input instances\", len(self.data)),\n (\"Inliers\", self.n_inliers),\n (\"Outliers\", self.n_outliers)))\n if self.outlier_method == 0:\n self.report_items(\n \"Detection\",\n ((\"Detection method\",\n \"One class SVM with non-linear kernel (RBF)\"),\n (\"Regularization (nu)\", self.nu),\n (\"Kernel coefficient\", self.gamma)))\n else:\n self.report_items(\n \"Detection\",\n ((\"Detection method\", \"Covariance estimator\"),\n (\"Contamination\", self.cont),\n (\"Support fraction\", self.support_fraction)))\n\ndef test_main():\n from AnyQt.QtWidgets import QApplication\n app = QApplication([])\n data = Table(\"iris\")\n w = OWOutliers()\n w.set_data(data)\n w.commit()\n w.show()\n return app.exec_()\n\n\nif __name__ == \"__main__\":\n sys.exit(test_main())\n", "path": "Orange/widgets/data/owoutliers.py"}]}
3,127
632
gh_patches_debug_6944
rasdani/github-patches
git_diff
OpenNMT__OpenNMT-py-1805
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> Logger is not passed to translator When building the translator, there is a logger created but not passed to the translator: https://github.com/OpenNMT/OpenNMT-py/blob/35cf4f0ae774a4aa500318879a1a4d53408ac129/onmt/bin/translate.py#L18 This results in a log file that only contains a single entry: https://github.com/OpenNMT/OpenNMT-py/blob/35cf4f0ae774a4aa500318879a1a4d53408ac129/onmt/bin/translate.py#L24 </issue> <code> [start of onmt/bin/translate.py] 1 #!/usr/bin/env python 2 # -*- coding: utf-8 -*- 3 4 from __future__ import unicode_literals 5 6 from onmt.utils.logging import init_logger 7 from onmt.utils.misc import split_corpus 8 from onmt.translate.translator import build_translator 9 10 import onmt.opts as opts 11 from onmt.utils.parse import ArgumentParser 12 13 14 def translate(opt): 15 ArgumentParser.validate_translate_opts(opt) 16 logger = init_logger(opt.log_file) 17 18 translator = build_translator(opt, report_score=True) 19 src_shards = split_corpus(opt.src, opt.shard_size) 20 tgt_shards = split_corpus(opt.tgt, opt.shard_size) 21 shard_pairs = zip(src_shards, tgt_shards) 22 23 for i, (src_shard, tgt_shard) in enumerate(shard_pairs): 24 logger.info("Translating shard %d." % i) 25 translator.translate( 26 src=src_shard, 27 tgt=tgt_shard, 28 src_dir=opt.src_dir, 29 batch_size=opt.batch_size, 30 batch_type=opt.batch_type, 31 attn_debug=opt.attn_debug, 32 align_debug=opt.align_debug 33 ) 34 35 36 def _get_parser(): 37 parser = ArgumentParser(description='translate.py') 38 39 opts.config_opts(parser) 40 opts.translate_opts(parser) 41 return parser 42 43 44 def main(): 45 parser = _get_parser() 46 47 opt = parser.parse_args() 48 translate(opt) 49 50 51 if __name__ == "__main__": 52 main() 53 [end of onmt/bin/translate.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/onmt/bin/translate.py b/onmt/bin/translate.py --- a/onmt/bin/translate.py +++ b/onmt/bin/translate.py @@ -15,7 +15,7 @@ ArgumentParser.validate_translate_opts(opt) logger = init_logger(opt.log_file) - translator = build_translator(opt, report_score=True) + translator = build_translator(opt, logger=logger, report_score=True) src_shards = split_corpus(opt.src, opt.shard_size) tgt_shards = split_corpus(opt.tgt, opt.shard_size) shard_pairs = zip(src_shards, tgt_shards)
{"golden_diff": "diff --git a/onmt/bin/translate.py b/onmt/bin/translate.py\n--- a/onmt/bin/translate.py\n+++ b/onmt/bin/translate.py\n@@ -15,7 +15,7 @@\n ArgumentParser.validate_translate_opts(opt)\n logger = init_logger(opt.log_file)\n \n- translator = build_translator(opt, report_score=True)\n+ translator = build_translator(opt, logger=logger, report_score=True)\n src_shards = split_corpus(opt.src, opt.shard_size)\n tgt_shards = split_corpus(opt.tgt, opt.shard_size)\n shard_pairs = zip(src_shards, tgt_shards)\n", "issue": "Logger is not passed to translator\nWhen building the translator, there is a logger created but not passed to the translator:\r\nhttps://github.com/OpenNMT/OpenNMT-py/blob/35cf4f0ae774a4aa500318879a1a4d53408ac129/onmt/bin/translate.py#L18\r\nThis results in a log file that only contains a single entry:\r\nhttps://github.com/OpenNMT/OpenNMT-py/blob/35cf4f0ae774a4aa500318879a1a4d53408ac129/onmt/bin/translate.py#L24\n", "before_files": [{"content": "#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\nfrom __future__ import unicode_literals\n\nfrom onmt.utils.logging import init_logger\nfrom onmt.utils.misc import split_corpus\nfrom onmt.translate.translator import build_translator\n\nimport onmt.opts as opts\nfrom onmt.utils.parse import ArgumentParser\n\n\ndef translate(opt):\n ArgumentParser.validate_translate_opts(opt)\n logger = init_logger(opt.log_file)\n\n translator = build_translator(opt, report_score=True)\n src_shards = split_corpus(opt.src, opt.shard_size)\n tgt_shards = split_corpus(opt.tgt, opt.shard_size)\n shard_pairs = zip(src_shards, tgt_shards)\n\n for i, (src_shard, tgt_shard) in enumerate(shard_pairs):\n logger.info(\"Translating shard %d.\" % i)\n translator.translate(\n src=src_shard,\n tgt=tgt_shard,\n src_dir=opt.src_dir,\n batch_size=opt.batch_size,\n batch_type=opt.batch_type,\n attn_debug=opt.attn_debug,\n align_debug=opt.align_debug\n )\n\n\ndef _get_parser():\n parser = ArgumentParser(description='translate.py')\n\n opts.config_opts(parser)\n opts.translate_opts(parser)\n return parser\n\n\ndef main():\n parser = _get_parser()\n\n opt = parser.parse_args()\n translate(opt)\n\n\nif __name__ == \"__main__\":\n main()\n", "path": "onmt/bin/translate.py"}]}
1,110
140
gh_patches_debug_15331
rasdani/github-patches
git_diff
napari__napari-1578
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> Plugin Errors not work in develop ## 🐛 Bug <!-- A clear and concise description of what the bug is. --> $ NAPARI_CATCH_ERRORS=0 python -m napari 16:11:17 ERROR Unhandled exception: Traceback (most recent call last): File "/home/czaki/Dokumenty/projekty/napari/napari/_qt/qt_main_window.py", line 300, in _show_plugin_err_reporter plugin_sorter = QtPluginErrReporter(parent=self._qt_window) File "/home/czaki/Dokumenty/projekty/napari/napari/_qt/dialogs/qt_plugin_report.py", line 62, in __init__ from ..plugins import plugin_manager as _pm ModuleNotFoundError: No module named 'napari._qt.plugins' ## To Reproduce Steps to reproduce the behavior: 1. install last napari from github 2. try open "Plugin Errors" dialog <!-- If you have a code sample, error messages, stack traces, please provide it here as well --> ## Expected behavior <!-- A clear and concise description of what you expected to happen. --> ## Environment - Please copy and paste the information at napari info option in help menubar here: napari: 0.3.7.dev30+gf101ef7 Platform: Linux-4.15.0-112-generic-x86_64-with-debian-buster-sid Python: 3.7.6 (default, Feb 20 2020, 10:04:45) [GCC 7.4.0] Qt: 5.14.2 PyQt5: 5.14.2 NumPy: 1.18.5 SciPy: 1.4.1 Dask: 2.17.2 VisPy: 0.6.4 GL version: 4.6.0 NVIDIA 440.100 MAX_TEXTURE_SIZE: 32768 Plugins: - napari-plugin-engine: 0.1.5 - svg: 0.1.3 - Any other relevant information: ## Additional context <!-- Add any other context about the problem here. --> @sofroniewn @jni </issue> <code> [start of napari/_qt/dialogs/qt_plugin_report.py] 1 """Provides a QtPluginErrReporter that allows the user report plugin errors. 2 """ 3 from typing import Optional 4 5 from napari_plugin_engine import PluginManager, standard_metadata 6 from qtpy.QtCore import Qt 7 from qtpy.QtGui import QGuiApplication 8 from qtpy.QtWidgets import ( 9 QComboBox, 10 QDialog, 11 QHBoxLayout, 12 QLabel, 13 QPushButton, 14 QTextEdit, 15 QVBoxLayout, 16 QWidget, 17 ) 18 19 from ...plugins.exceptions import format_exceptions 20 21 22 class QtPluginErrReporter(QDialog): 23 """Dialog that allows users to review and report PluginError tracebacks. 24 25 Parameters 26 ---------- 27 parent : QWidget, optional 28 Optional parent widget for this widget. 29 initial_plugin : str, optional 30 If provided, errors from ``initial_plugin`` will be shown when the 31 dialog is created, by default None 32 33 Attributes 34 ---------- 35 text_area : qtpy.QtWidgets.QTextEdit 36 The text area where traceback information will be shown. 37 plugin_combo : qtpy.QtWidgets.QComboBox 38 The dropdown menu used to select the current plugin 39 github_button : qtpy.QtWidgets.QPushButton 40 A button that, when pressed, will open an issue at the current plugin's 41 github issue tracker, prepopulated with a formatted traceback. Button 42 is only visible if a github URL is detected in the package metadata for 43 the current plugin. 44 clipboard_button : qtpy.QtWidgets.QPushButton 45 A button that, when pressed, copies the current traceback information 46 to the clipboard. (HTML tags are removed in the copied text.) 47 plugin_meta : qtpy.QtWidgets.QLabel 48 A label that will show available plugin metadata (such as home page). 49 """ 50 51 NULL_OPTION = 'select plugin... ' 52 53 def __init__( 54 self, 55 plugin_manager: Optional[PluginManager] = None, 56 *, 57 parent: Optional[QWidget] = None, 58 initial_plugin: Optional[str] = None, 59 ) -> None: 60 super().__init__(parent) 61 if not plugin_manager: 62 from ..plugins import plugin_manager as _pm 63 64 self.plugin_manager = _pm 65 else: 66 self.plugin_manager = plugin_manager 67 68 self.setWindowTitle('Recorded Plugin Exceptions') 69 self.setWindowModality(Qt.NonModal) 70 self.layout = QVBoxLayout() 71 self.layout.setSpacing(0) 72 self.layout.setContentsMargins(10, 10, 10, 10) 73 self.setLayout(self.layout) 74 75 self.text_area = QTextEdit() 76 self.text_area.setTextInteractionFlags(Qt.TextSelectableByMouse) 77 self.text_area.setMinimumWidth(360) 78 79 # Create plugin dropdown menu 80 self.plugin_combo = QComboBox() 81 self.plugin_combo.addItem(self.NULL_OPTION) 82 bad_plugins = [e.plugin_name for e in self.plugin_manager.get_errors()] 83 self.plugin_combo.addItems(list(sorted(set(bad_plugins)))) 84 self.plugin_combo.currentTextChanged.connect(self.set_plugin) 85 self.plugin_combo.setCurrentText(self.NULL_OPTION) 86 87 # create github button (gets connected in self.set_plugin) 88 self.github_button = QPushButton('Open issue on GitHub', self) 89 self.github_button.setToolTip( 90 "Open a web browser to submit this error log\n" 91 "to the developer's GitHub issue tracker" 92 ) 93 self.github_button.hide() 94 95 # create copy to clipboard button 96 self.clipboard_button = QPushButton() 97 self.clipboard_button.hide() 98 self.clipboard_button.setObjectName("QtCopyToClipboardButton") 99 self.clipboard_button.setToolTip("Copy error log to clipboard") 100 self.clipboard_button.clicked.connect(self.copyToClipboard) 101 102 # plugin_meta contains a URL to the home page, (and/or other details) 103 self.plugin_meta = QLabel('', parent=self) 104 self.plugin_meta.setObjectName("pluginInfo") 105 self.plugin_meta.setTextFormat(Qt.RichText) 106 self.plugin_meta.setTextInteractionFlags(Qt.TextBrowserInteraction) 107 self.plugin_meta.setOpenExternalLinks(True) 108 self.plugin_meta.setAlignment(Qt.AlignRight) 109 110 # make layout 111 row_1_layout = QHBoxLayout() 112 row_1_layout.setContentsMargins(11, 5, 10, 0) 113 row_1_layout.addStretch(1) 114 row_1_layout.addWidget(self.plugin_meta) 115 row_2_layout = QHBoxLayout() 116 row_2_layout.setContentsMargins(11, 5, 10, 0) 117 row_2_layout.addWidget(self.plugin_combo) 118 row_2_layout.addStretch(1) 119 row_2_layout.addWidget(self.github_button) 120 row_2_layout.addWidget(self.clipboard_button) 121 row_2_layout.setSpacing(5) 122 self.layout.addLayout(row_1_layout) 123 self.layout.addLayout(row_2_layout) 124 self.layout.addWidget(self.text_area, 1) 125 self.setMinimumWidth(750) 126 self.setMinimumHeight(600) 127 128 if initial_plugin: 129 self.set_plugin(initial_plugin) 130 131 def set_plugin(self, plugin: str) -> None: 132 """Set the current plugin shown in the dropdown and text area. 133 134 Parameters 135 ---------- 136 plugin : str 137 name of a plugin that has created an error this session. 138 """ 139 self.github_button.hide() 140 self.clipboard_button.hide() 141 try: 142 self.github_button.clicked.disconnect() 143 # when disconnecting a non-existent signal 144 # PySide2 raises runtimeError, PyQt5 raises TypeError 145 except (RuntimeError, TypeError): 146 pass 147 148 if not plugin or (plugin == self.NULL_OPTION): 149 self.plugin_meta.setText('') 150 self.text_area.setHtml('') 151 return 152 153 if not self.plugin_manager.get_errors(plugin): 154 raise ValueError(f"No errors reported for plugin '{plugin}'") 155 self.plugin_combo.setCurrentText(plugin) 156 157 err_string = format_exceptions(plugin, as_html=True) 158 self.text_area.setHtml(err_string) 159 self.clipboard_button.show() 160 161 # set metadata and outbound links/buttons 162 err0 = self.plugin_manager.get_errors(plugin)[0] 163 meta = standard_metadata(err0.plugin) if err0.plugin else {} 164 meta_text = '' 165 if not meta: 166 self.plugin_meta.setText(meta_text) 167 return 168 169 url = meta.get('url') 170 if url: 171 meta_text += ( 172 '<span style="color:#999;">plugin home page:&nbsp;&nbsp;' 173 f'</span><a href="{url}" style="color:#999">{url}</a>' 174 ) 175 if 'github.com' in url: 176 177 def onclick(): 178 import webbrowser 179 180 err = format_exceptions(plugin, as_html=False) 181 err = ( 182 "<!--Provide detail on the error here-->\n\n\n\n" 183 "<details>\n<summary>Traceback from napari</summary>" 184 f"\n\n```\n{err}\n```\n</details>" 185 ) 186 url = f'{meta.get("url")}/issues/new?&body={err}' 187 webbrowser.open(url, new=2) 188 189 self.github_button.clicked.connect(onclick) 190 self.github_button.show() 191 self.plugin_meta.setText(meta_text) 192 193 def copyToClipboard(self) -> None: 194 """Copy current plugin traceback info to clipboard as plain text.""" 195 plugin = self.plugin_combo.currentText() 196 err_string = format_exceptions(plugin, as_html=False) 197 cb = QGuiApplication.clipboard() 198 cb.setText(err_string) 199 [end of napari/_qt/dialogs/qt_plugin_report.py] [start of napari/_qt/dialogs/qt_plugin_table.py] 1 from qtpy.QtCore import Qt 2 from qtpy.QtWidgets import QAbstractItemView, QDialog, QLabel, QVBoxLayout 3 4 from ..widgets.qt_dict_table import QtDictTable 5 6 7 class QtPluginTable(QDialog): 8 def __init__(self, parent, plugin_manager=None): 9 super().__init__(parent) 10 if not plugin_manager: 11 from ..plugins import plugin_manager 12 13 self.setMaximumHeight(800) 14 self.setMaximumWidth(1280) 15 layout = QVBoxLayout() 16 # maybe someday add a search bar here? 17 title = QLabel("Installed Plugins") 18 title.setObjectName("h2") 19 layout.addWidget(title) 20 # get metadata for successfully registered plugins 21 plugin_manager.discover() 22 data = plugin_manager.list_plugin_metadata() 23 data = list(filter(lambda x: x['plugin_name'] != 'builtins', data)) 24 # create a table for it 25 self.table = QtDictTable( 26 parent, 27 data, 28 headers=[ 29 'plugin_name', 30 'package', 31 'version', 32 'url', 33 'author', 34 'license', 35 ], 36 min_section_width=60, 37 ) 38 self.table.setObjectName("pluginTable") 39 self.table.horizontalHeader().setObjectName("pluginTableHeader") 40 self.table.verticalHeader().setObjectName("pluginTableHeader") 41 self.table.setGridStyle(Qt.NoPen) 42 # prevent editing of table 43 self.table.setEditTriggers(QAbstractItemView.NoEditTriggers) 44 layout.addWidget(self.table) 45 self.setLayout(layout) 46 self.setAttribute(Qt.WA_DeleteOnClose) 47 [end of napari/_qt/dialogs/qt_plugin_table.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/napari/_qt/dialogs/qt_plugin_report.py b/napari/_qt/dialogs/qt_plugin_report.py --- a/napari/_qt/dialogs/qt_plugin_report.py +++ b/napari/_qt/dialogs/qt_plugin_report.py @@ -59,7 +59,7 @@ ) -> None: super().__init__(parent) if not plugin_manager: - from ..plugins import plugin_manager as _pm + from ...plugins import plugin_manager as _pm self.plugin_manager = _pm else: diff --git a/napari/_qt/dialogs/qt_plugin_table.py b/napari/_qt/dialogs/qt_plugin_table.py --- a/napari/_qt/dialogs/qt_plugin_table.py +++ b/napari/_qt/dialogs/qt_plugin_table.py @@ -8,7 +8,7 @@ def __init__(self, parent, plugin_manager=None): super().__init__(parent) if not plugin_manager: - from ..plugins import plugin_manager + from ...plugins import plugin_manager self.setMaximumHeight(800) self.setMaximumWidth(1280)
{"golden_diff": "diff --git a/napari/_qt/dialogs/qt_plugin_report.py b/napari/_qt/dialogs/qt_plugin_report.py\n--- a/napari/_qt/dialogs/qt_plugin_report.py\n+++ b/napari/_qt/dialogs/qt_plugin_report.py\n@@ -59,7 +59,7 @@\n ) -> None:\n super().__init__(parent)\n if not plugin_manager:\n- from ..plugins import plugin_manager as _pm\n+ from ...plugins import plugin_manager as _pm\n \n self.plugin_manager = _pm\n else:\ndiff --git a/napari/_qt/dialogs/qt_plugin_table.py b/napari/_qt/dialogs/qt_plugin_table.py\n--- a/napari/_qt/dialogs/qt_plugin_table.py\n+++ b/napari/_qt/dialogs/qt_plugin_table.py\n@@ -8,7 +8,7 @@\n def __init__(self, parent, plugin_manager=None):\n super().__init__(parent)\n if not plugin_manager:\n- from ..plugins import plugin_manager\n+ from ...plugins import plugin_manager\n \n self.setMaximumHeight(800)\n self.setMaximumWidth(1280)\n", "issue": "Plugin Errors not work in develop \n## \ud83d\udc1b Bug\r\n\r\n<!-- A clear and concise description of what the bug is. -->\r\n\r\n$ NAPARI_CATCH_ERRORS=0 python -m napari\r\n16:11:17 ERROR Unhandled exception:\r\nTraceback (most recent call last):\r\n File \"/home/czaki/Dokumenty/projekty/napari/napari/_qt/qt_main_window.py\", line 300, in _show_plugin_err_reporter\r\n plugin_sorter = QtPluginErrReporter(parent=self._qt_window)\r\n File \"/home/czaki/Dokumenty/projekty/napari/napari/_qt/dialogs/qt_plugin_report.py\", line 62, in __init__\r\n from ..plugins import plugin_manager as _pm\r\nModuleNotFoundError: No module named 'napari._qt.plugins'\r\n\r\n\r\n## To Reproduce\r\n\r\nSteps to reproduce the behavior:\r\n\r\n1. install last napari from github\r\n2. try open \"Plugin Errors\" dialog\r\n\r\n\r\n<!-- If you have a code sample, error messages, stack traces, please provide it here as well -->\r\n\r\n## Expected behavior\r\n\r\n<!-- A clear and concise description of what you expected to happen. -->\r\n\r\n## Environment\r\n\r\n - Please copy and paste the information at napari info option in help menubar here:\r\n\r\nnapari: 0.3.7.dev30+gf101ef7\r\nPlatform: Linux-4.15.0-112-generic-x86_64-with-debian-buster-sid\r\nPython: 3.7.6 (default, Feb 20 2020, 10:04:45) [GCC 7.4.0]\r\nQt: 5.14.2\r\nPyQt5: 5.14.2\r\nNumPy: 1.18.5\r\nSciPy: 1.4.1\r\nDask: 2.17.2\r\nVisPy: 0.6.4\r\n\r\nGL version: 4.6.0 NVIDIA 440.100\r\nMAX_TEXTURE_SIZE: 32768\r\n\r\nPlugins:\r\n- napari-plugin-engine: 0.1.5\r\n- svg: 0.1.3\r\n\r\n - Any other relevant information:\r\n\r\n## Additional context\r\n\r\n<!-- Add any other context about the problem here. -->\r\n@sofroniewn @jni \n", "before_files": [{"content": "\"\"\"Provides a QtPluginErrReporter that allows the user report plugin errors.\n\"\"\"\nfrom typing import Optional\n\nfrom napari_plugin_engine import PluginManager, standard_metadata\nfrom qtpy.QtCore import Qt\nfrom qtpy.QtGui import QGuiApplication\nfrom qtpy.QtWidgets import (\n QComboBox,\n QDialog,\n QHBoxLayout,\n QLabel,\n QPushButton,\n QTextEdit,\n QVBoxLayout,\n QWidget,\n)\n\nfrom ...plugins.exceptions import format_exceptions\n\n\nclass QtPluginErrReporter(QDialog):\n \"\"\"Dialog that allows users to review and report PluginError tracebacks.\n\n Parameters\n ----------\n parent : QWidget, optional\n Optional parent widget for this widget.\n initial_plugin : str, optional\n If provided, errors from ``initial_plugin`` will be shown when the\n dialog is created, by default None\n\n Attributes\n ----------\n text_area : qtpy.QtWidgets.QTextEdit\n The text area where traceback information will be shown.\n plugin_combo : qtpy.QtWidgets.QComboBox\n The dropdown menu used to select the current plugin\n github_button : qtpy.QtWidgets.QPushButton\n A button that, when pressed, will open an issue at the current plugin's\n github issue tracker, prepopulated with a formatted traceback. Button\n is only visible if a github URL is detected in the package metadata for\n the current plugin.\n clipboard_button : qtpy.QtWidgets.QPushButton\n A button that, when pressed, copies the current traceback information\n to the clipboard. (HTML tags are removed in the copied text.)\n plugin_meta : qtpy.QtWidgets.QLabel\n A label that will show available plugin metadata (such as home page).\n \"\"\"\n\n NULL_OPTION = 'select plugin... '\n\n def __init__(\n self,\n plugin_manager: Optional[PluginManager] = None,\n *,\n parent: Optional[QWidget] = None,\n initial_plugin: Optional[str] = None,\n ) -> None:\n super().__init__(parent)\n if not plugin_manager:\n from ..plugins import plugin_manager as _pm\n\n self.plugin_manager = _pm\n else:\n self.plugin_manager = plugin_manager\n\n self.setWindowTitle('Recorded Plugin Exceptions')\n self.setWindowModality(Qt.NonModal)\n self.layout = QVBoxLayout()\n self.layout.setSpacing(0)\n self.layout.setContentsMargins(10, 10, 10, 10)\n self.setLayout(self.layout)\n\n self.text_area = QTextEdit()\n self.text_area.setTextInteractionFlags(Qt.TextSelectableByMouse)\n self.text_area.setMinimumWidth(360)\n\n # Create plugin dropdown menu\n self.plugin_combo = QComboBox()\n self.plugin_combo.addItem(self.NULL_OPTION)\n bad_plugins = [e.plugin_name for e in self.plugin_manager.get_errors()]\n self.plugin_combo.addItems(list(sorted(set(bad_plugins))))\n self.plugin_combo.currentTextChanged.connect(self.set_plugin)\n self.plugin_combo.setCurrentText(self.NULL_OPTION)\n\n # create github button (gets connected in self.set_plugin)\n self.github_button = QPushButton('Open issue on GitHub', self)\n self.github_button.setToolTip(\n \"Open a web browser to submit this error log\\n\"\n \"to the developer's GitHub issue tracker\"\n )\n self.github_button.hide()\n\n # create copy to clipboard button\n self.clipboard_button = QPushButton()\n self.clipboard_button.hide()\n self.clipboard_button.setObjectName(\"QtCopyToClipboardButton\")\n self.clipboard_button.setToolTip(\"Copy error log to clipboard\")\n self.clipboard_button.clicked.connect(self.copyToClipboard)\n\n # plugin_meta contains a URL to the home page, (and/or other details)\n self.plugin_meta = QLabel('', parent=self)\n self.plugin_meta.setObjectName(\"pluginInfo\")\n self.plugin_meta.setTextFormat(Qt.RichText)\n self.plugin_meta.setTextInteractionFlags(Qt.TextBrowserInteraction)\n self.plugin_meta.setOpenExternalLinks(True)\n self.plugin_meta.setAlignment(Qt.AlignRight)\n\n # make layout\n row_1_layout = QHBoxLayout()\n row_1_layout.setContentsMargins(11, 5, 10, 0)\n row_1_layout.addStretch(1)\n row_1_layout.addWidget(self.plugin_meta)\n row_2_layout = QHBoxLayout()\n row_2_layout.setContentsMargins(11, 5, 10, 0)\n row_2_layout.addWidget(self.plugin_combo)\n row_2_layout.addStretch(1)\n row_2_layout.addWidget(self.github_button)\n row_2_layout.addWidget(self.clipboard_button)\n row_2_layout.setSpacing(5)\n self.layout.addLayout(row_1_layout)\n self.layout.addLayout(row_2_layout)\n self.layout.addWidget(self.text_area, 1)\n self.setMinimumWidth(750)\n self.setMinimumHeight(600)\n\n if initial_plugin:\n self.set_plugin(initial_plugin)\n\n def set_plugin(self, plugin: str) -> None:\n \"\"\"Set the current plugin shown in the dropdown and text area.\n\n Parameters\n ----------\n plugin : str\n name of a plugin that has created an error this session.\n \"\"\"\n self.github_button.hide()\n self.clipboard_button.hide()\n try:\n self.github_button.clicked.disconnect()\n # when disconnecting a non-existent signal\n # PySide2 raises runtimeError, PyQt5 raises TypeError\n except (RuntimeError, TypeError):\n pass\n\n if not plugin or (plugin == self.NULL_OPTION):\n self.plugin_meta.setText('')\n self.text_area.setHtml('')\n return\n\n if not self.plugin_manager.get_errors(plugin):\n raise ValueError(f\"No errors reported for plugin '{plugin}'\")\n self.plugin_combo.setCurrentText(plugin)\n\n err_string = format_exceptions(plugin, as_html=True)\n self.text_area.setHtml(err_string)\n self.clipboard_button.show()\n\n # set metadata and outbound links/buttons\n err0 = self.plugin_manager.get_errors(plugin)[0]\n meta = standard_metadata(err0.plugin) if err0.plugin else {}\n meta_text = ''\n if not meta:\n self.plugin_meta.setText(meta_text)\n return\n\n url = meta.get('url')\n if url:\n meta_text += (\n '<span style=\"color:#999;\">plugin home page:&nbsp;&nbsp;'\n f'</span><a href=\"{url}\" style=\"color:#999\">{url}</a>'\n )\n if 'github.com' in url:\n\n def onclick():\n import webbrowser\n\n err = format_exceptions(plugin, as_html=False)\n err = (\n \"<!--Provide detail on the error here-->\\n\\n\\n\\n\"\n \"<details>\\n<summary>Traceback from napari</summary>\"\n f\"\\n\\n```\\n{err}\\n```\\n</details>\"\n )\n url = f'{meta.get(\"url\")}/issues/new?&body={err}'\n webbrowser.open(url, new=2)\n\n self.github_button.clicked.connect(onclick)\n self.github_button.show()\n self.plugin_meta.setText(meta_text)\n\n def copyToClipboard(self) -> None:\n \"\"\"Copy current plugin traceback info to clipboard as plain text.\"\"\"\n plugin = self.plugin_combo.currentText()\n err_string = format_exceptions(plugin, as_html=False)\n cb = QGuiApplication.clipboard()\n cb.setText(err_string)\n", "path": "napari/_qt/dialogs/qt_plugin_report.py"}, {"content": "from qtpy.QtCore import Qt\nfrom qtpy.QtWidgets import QAbstractItemView, QDialog, QLabel, QVBoxLayout\n\nfrom ..widgets.qt_dict_table import QtDictTable\n\n\nclass QtPluginTable(QDialog):\n def __init__(self, parent, plugin_manager=None):\n super().__init__(parent)\n if not plugin_manager:\n from ..plugins import plugin_manager\n\n self.setMaximumHeight(800)\n self.setMaximumWidth(1280)\n layout = QVBoxLayout()\n # maybe someday add a search bar here?\n title = QLabel(\"Installed Plugins\")\n title.setObjectName(\"h2\")\n layout.addWidget(title)\n # get metadata for successfully registered plugins\n plugin_manager.discover()\n data = plugin_manager.list_plugin_metadata()\n data = list(filter(lambda x: x['plugin_name'] != 'builtins', data))\n # create a table for it\n self.table = QtDictTable(\n parent,\n data,\n headers=[\n 'plugin_name',\n 'package',\n 'version',\n 'url',\n 'author',\n 'license',\n ],\n min_section_width=60,\n )\n self.table.setObjectName(\"pluginTable\")\n self.table.horizontalHeader().setObjectName(\"pluginTableHeader\")\n self.table.verticalHeader().setObjectName(\"pluginTableHeader\")\n self.table.setGridStyle(Qt.NoPen)\n # prevent editing of table\n self.table.setEditTriggers(QAbstractItemView.NoEditTriggers)\n layout.addWidget(self.table)\n self.setLayout(layout)\n self.setAttribute(Qt.WA_DeleteOnClose)\n", "path": "napari/_qt/dialogs/qt_plugin_table.py"}]}
3,555
250
gh_patches_debug_17769
rasdani/github-patches
git_diff
kivy__python-for-android-2340
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> Webview app crashes when trying to request permissions ### Checklist - [X] the issue is indeed a bug and not a support request - [X] issue doesn't already exist: https://github.com/kivy/python-for-android/issues - [X] I have a short, runnable example that reproduces the issue - [X] I reproduced the problem with the latest development version (`p4a.branch = develop`) - [X] I used the grave accent (aka backticks) to format code or logs when appropriated ### Versions - Python: 3.6.6 - OS: Ubuntu 20.04 - python-for-android: 2020.6.2 ### Description I'm building with: ``` HERE=$(pwd) cd app export ANDROIDSDK="$HERE/android_sdks" # Have also tried with # export ANDROIDNDK="$HERE/android_sdks/android-ndk-r19c" export ANDROIDNDK="$HERE/android_sdks/android-ndk-r21d" export ANDROIDAPI="27" # Target API version of your application export NDKAPI="21" # Minimum supported API version of your application p4a apk --private "." --package=org.anarres.pdfreader --name "pdf reader" --version 0.1 --bootstrap=webview --requirements=python3,flask --port=5000 --orientation=sensor --blacklist="../blacklist.txt" --permission READ_EXTERNAL_STORAGE --permission WRITE_EXTERNAL_STORAGE ``` My python code is just: ``` from android.permissions import request_permissions, Permission request_permissions([Permission.READ_EXTERNAL_STORAGE, Permission.WRITE_EXTERNAL_STORAGE]) from flask import Flask, render_template app = Flask(__name__) @app.route('/') def index(): return render_template('index.html') if __name__ == '__main__': app.run(port=5000) ``` If I comment out: ``` from android.permissions import request_permissions, Permission request_permissions([Permission.READ_EXTERNAL_STORAGE, Permission.WRITE_EXTERNAL_STORAGE]) ``` I don't see the error, but then of course I can't access the external storage. ### Logs ``` ImportError: dlopen failed: cannot locate symbol "SDL_ANDROID_GetJNIEnv" referenced by "/data/data/org.anarres.pdfreader/files/app/_python_bundle/site-packages/android/_android.so"... ``` </issue> <code> [start of pythonforandroid/recipes/android/src/setup.py] 1 from distutils.core import setup, Extension 2 import os 3 4 library_dirs = ['libs/' + os.environ['ARCH']] 5 lib_dict = { 6 'sdl2': ['SDL2', 'SDL2_image', 'SDL2_mixer', 'SDL2_ttf'] 7 } 8 sdl_libs = lib_dict.get(os.environ['BOOTSTRAP'], []) 9 10 modules = [Extension('android._android', 11 ['android/_android.c', 'android/_android_jni.c'], 12 libraries=sdl_libs + ['log'], 13 library_dirs=library_dirs), 14 Extension('android._android_billing', 15 ['android/_android_billing.c', 'android/_android_billing_jni.c'], 16 libraries=['log'], 17 library_dirs=library_dirs)] 18 19 setup(name='android', 20 version='1.0', 21 packages=['android'], 22 package_dir={'android': 'android'}, 23 ext_modules=modules 24 ) 25 [end of pythonforandroid/recipes/android/src/setup.py] [start of pythonforandroid/recipes/android/__init__.py] 1 from pythonforandroid.recipe import CythonRecipe, IncludedFilesBehaviour 2 from pythonforandroid.util import current_directory 3 from pythonforandroid import logger 4 5 from os.path import join 6 7 8 class AndroidRecipe(IncludedFilesBehaviour, CythonRecipe): 9 # name = 'android' 10 version = None 11 url = None 12 13 src_filename = 'src' 14 15 depends = [('sdl2', 'genericndkbuild'), 'pyjnius'] 16 17 config_env = {} 18 19 def get_recipe_env(self, arch): 20 env = super().get_recipe_env(arch) 21 env.update(self.config_env) 22 return env 23 24 def prebuild_arch(self, arch): 25 super().prebuild_arch(arch) 26 ctx_bootstrap = self.ctx.bootstrap.name 27 28 # define macros for Cython, C, Python 29 tpxi = 'DEF {} = {}\n' 30 th = '#define {} {}\n' 31 tpy = '{} = {}\n' 32 33 # make sure bootstrap name is in unicode 34 if isinstance(ctx_bootstrap, bytes): 35 ctx_bootstrap = ctx_bootstrap.decode('utf-8') 36 bootstrap = bootstrap_name = ctx_bootstrap 37 is_sdl2 = (bootstrap_name == "sdl2") 38 if bootstrap_name in ["sdl2", "webview", "service_only", "service_library"]: 39 java_ns = u'org.kivy.android' 40 jni_ns = u'org/kivy/android' 41 else: 42 logger.error(( 43 'unsupported bootstrap for android recipe: {}' 44 ''.format(bootstrap_name) 45 )) 46 exit(1) 47 48 config = { 49 'BOOTSTRAP': bootstrap, 50 'IS_SDL2': int(is_sdl2), 51 'PY2': 0, 52 'JAVA_NAMESPACE': java_ns, 53 'JNI_NAMESPACE': jni_ns, 54 'ACTIVITY_CLASS_NAME': self.ctx.activity_class_name, 55 'ACTIVITY_CLASS_NAMESPACE': self.ctx.activity_class_name.replace('.', '/'), 56 } 57 58 # create config files for Cython, C and Python 59 with ( 60 current_directory(self.get_build_dir(arch.arch))), ( 61 open(join('android', 'config.pxi'), 'w')) as fpxi, ( 62 open(join('android', 'config.h'), 'w')) as fh, ( 63 open(join('android', 'config.py'), 'w')) as fpy: 64 65 for key, value in config.items(): 66 fpxi.write(tpxi.format(key, repr(value))) 67 fpy.write(tpy.format(key, repr(value))) 68 69 fh.write(th.format( 70 key, 71 value if isinstance(value, int) else '"{}"'.format(value) 72 )) 73 self.config_env[key] = str(value) 74 75 if is_sdl2: 76 fh.write('JNIEnv *SDL_AndroidGetJNIEnv(void);\n') 77 fh.write( 78 '#define SDL_ANDROID_GetJNIEnv SDL_AndroidGetJNIEnv\n' 79 ) 80 81 82 recipe = AndroidRecipe() 83 [end of pythonforandroid/recipes/android/__init__.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/pythonforandroid/recipes/android/__init__.py b/pythonforandroid/recipes/android/__init__.py --- a/pythonforandroid/recipes/android/__init__.py +++ b/pythonforandroid/recipes/android/__init__.py @@ -77,6 +77,11 @@ fh.write( '#define SDL_ANDROID_GetJNIEnv SDL_AndroidGetJNIEnv\n' ) + else: + fh.write('JNIEnv *WebView_AndroidGetJNIEnv(void);\n') + fh.write( + '#define SDL_ANDROID_GetJNIEnv WebView_AndroidGetJNIEnv\n' + ) recipe = AndroidRecipe() diff --git a/pythonforandroid/recipes/android/src/setup.py b/pythonforandroid/recipes/android/src/setup.py --- a/pythonforandroid/recipes/android/src/setup.py +++ b/pythonforandroid/recipes/android/src/setup.py @@ -5,7 +5,7 @@ lib_dict = { 'sdl2': ['SDL2', 'SDL2_image', 'SDL2_mixer', 'SDL2_ttf'] } -sdl_libs = lib_dict.get(os.environ['BOOTSTRAP'], []) +sdl_libs = lib_dict.get(os.environ['BOOTSTRAP'], ['main']) modules = [Extension('android._android', ['android/_android.c', 'android/_android_jni.c'],
{"golden_diff": "diff --git a/pythonforandroid/recipes/android/__init__.py b/pythonforandroid/recipes/android/__init__.py\n--- a/pythonforandroid/recipes/android/__init__.py\n+++ b/pythonforandroid/recipes/android/__init__.py\n@@ -77,6 +77,11 @@\n fh.write(\n '#define SDL_ANDROID_GetJNIEnv SDL_AndroidGetJNIEnv\\n'\n )\n+ else:\n+ fh.write('JNIEnv *WebView_AndroidGetJNIEnv(void);\\n')\n+ fh.write(\n+ '#define SDL_ANDROID_GetJNIEnv WebView_AndroidGetJNIEnv\\n'\n+ )\n \n \n recipe = AndroidRecipe()\ndiff --git a/pythonforandroid/recipes/android/src/setup.py b/pythonforandroid/recipes/android/src/setup.py\n--- a/pythonforandroid/recipes/android/src/setup.py\n+++ b/pythonforandroid/recipes/android/src/setup.py\n@@ -5,7 +5,7 @@\n lib_dict = {\n 'sdl2': ['SDL2', 'SDL2_image', 'SDL2_mixer', 'SDL2_ttf']\n }\n-sdl_libs = lib_dict.get(os.environ['BOOTSTRAP'], [])\n+sdl_libs = lib_dict.get(os.environ['BOOTSTRAP'], ['main'])\n \n modules = [Extension('android._android',\n ['android/_android.c', 'android/_android_jni.c'],\n", "issue": "Webview app crashes when trying to request permissions\n### Checklist\r\n\r\n- [X] the issue is indeed a bug and not a support request\r\n- [X] issue doesn't already exist: https://github.com/kivy/python-for-android/issues\r\n- [X] I have a short, runnable example that reproduces the issue\r\n- [X] I reproduced the problem with the latest development version (`p4a.branch = develop`)\r\n- [X] I used the grave accent (aka backticks) to format code or logs when appropriated\r\n\r\n### Versions\r\n\r\n- Python: 3.6.6\r\n- OS: Ubuntu 20.04\r\n- python-for-android: 2020.6.2\r\n\r\n### Description\r\n\r\nI'm building with:\r\n\r\n```\r\nHERE=$(pwd)\r\ncd app\r\n\r\nexport ANDROIDSDK=\"$HERE/android_sdks\"\r\n# Have also tried with\r\n# export ANDROIDNDK=\"$HERE/android_sdks/android-ndk-r19c\"\r\nexport ANDROIDNDK=\"$HERE/android_sdks/android-ndk-r21d\"\r\nexport ANDROIDAPI=\"27\" # Target API version of your application\r\nexport NDKAPI=\"21\" # Minimum supported API version of your application\r\n\r\np4a apk --private \".\" --package=org.anarres.pdfreader --name \"pdf reader\" --version 0.1 --bootstrap=webview --requirements=python3,flask --port=5000 --orientation=sensor --blacklist=\"../blacklist.txt\" --permission READ_EXTERNAL_STORAGE --permission WRITE_EXTERNAL_STORAGE\r\n```\r\n\r\nMy python code is just:\r\n\r\n```\r\nfrom android.permissions import request_permissions, Permission\r\nrequest_permissions([Permission.READ_EXTERNAL_STORAGE, Permission.WRITE_EXTERNAL_STORAGE])\r\n\r\nfrom flask import Flask, render_template\r\n\r\napp = Flask(__name__)\r\n\r\[email protected]('/')\r\ndef index():\r\n return render_template('index.html')\r\n\r\nif __name__ == '__main__':\r\n app.run(port=5000)\r\n```\r\n\r\nIf I comment out:\r\n\r\n```\r\nfrom android.permissions import request_permissions, Permission\r\nrequest_permissions([Permission.READ_EXTERNAL_STORAGE, Permission.WRITE_EXTERNAL_STORAGE])\r\n```\r\n\r\nI don't see the error, but then of course I can't access the external storage.\r\n\r\n### Logs\r\n\r\n```\r\nImportError: dlopen failed: cannot locate symbol \"SDL_ANDROID_GetJNIEnv\" referenced by \"/data/data/org.anarres.pdfreader/files/app/_python_bundle/site-packages/android/_android.so\"...\r\n```\r\n\n", "before_files": [{"content": "from distutils.core import setup, Extension\nimport os\n\nlibrary_dirs = ['libs/' + os.environ['ARCH']]\nlib_dict = {\n 'sdl2': ['SDL2', 'SDL2_image', 'SDL2_mixer', 'SDL2_ttf']\n}\nsdl_libs = lib_dict.get(os.environ['BOOTSTRAP'], [])\n\nmodules = [Extension('android._android',\n ['android/_android.c', 'android/_android_jni.c'],\n libraries=sdl_libs + ['log'],\n library_dirs=library_dirs),\n Extension('android._android_billing',\n ['android/_android_billing.c', 'android/_android_billing_jni.c'],\n libraries=['log'],\n library_dirs=library_dirs)]\n\nsetup(name='android',\n version='1.0',\n packages=['android'],\n package_dir={'android': 'android'},\n ext_modules=modules\n )\n", "path": "pythonforandroid/recipes/android/src/setup.py"}, {"content": "from pythonforandroid.recipe import CythonRecipe, IncludedFilesBehaviour\nfrom pythonforandroid.util import current_directory\nfrom pythonforandroid import logger\n\nfrom os.path import join\n\n\nclass AndroidRecipe(IncludedFilesBehaviour, CythonRecipe):\n # name = 'android'\n version = None\n url = None\n\n src_filename = 'src'\n\n depends = [('sdl2', 'genericndkbuild'), 'pyjnius']\n\n config_env = {}\n\n def get_recipe_env(self, arch):\n env = super().get_recipe_env(arch)\n env.update(self.config_env)\n return env\n\n def prebuild_arch(self, arch):\n super().prebuild_arch(arch)\n ctx_bootstrap = self.ctx.bootstrap.name\n\n # define macros for Cython, C, Python\n tpxi = 'DEF {} = {}\\n'\n th = '#define {} {}\\n'\n tpy = '{} = {}\\n'\n\n # make sure bootstrap name is in unicode\n if isinstance(ctx_bootstrap, bytes):\n ctx_bootstrap = ctx_bootstrap.decode('utf-8')\n bootstrap = bootstrap_name = ctx_bootstrap\n is_sdl2 = (bootstrap_name == \"sdl2\")\n if bootstrap_name in [\"sdl2\", \"webview\", \"service_only\", \"service_library\"]:\n java_ns = u'org.kivy.android'\n jni_ns = u'org/kivy/android'\n else:\n logger.error((\n 'unsupported bootstrap for android recipe: {}'\n ''.format(bootstrap_name)\n ))\n exit(1)\n\n config = {\n 'BOOTSTRAP': bootstrap,\n 'IS_SDL2': int(is_sdl2),\n 'PY2': 0,\n 'JAVA_NAMESPACE': java_ns,\n 'JNI_NAMESPACE': jni_ns,\n 'ACTIVITY_CLASS_NAME': self.ctx.activity_class_name,\n 'ACTIVITY_CLASS_NAMESPACE': self.ctx.activity_class_name.replace('.', '/'),\n }\n\n # create config files for Cython, C and Python\n with (\n current_directory(self.get_build_dir(arch.arch))), (\n open(join('android', 'config.pxi'), 'w')) as fpxi, (\n open(join('android', 'config.h'), 'w')) as fh, (\n open(join('android', 'config.py'), 'w')) as fpy:\n\n for key, value in config.items():\n fpxi.write(tpxi.format(key, repr(value)))\n fpy.write(tpy.format(key, repr(value)))\n\n fh.write(th.format(\n key,\n value if isinstance(value, int) else '\"{}\"'.format(value)\n ))\n self.config_env[key] = str(value)\n\n if is_sdl2:\n fh.write('JNIEnv *SDL_AndroidGetJNIEnv(void);\\n')\n fh.write(\n '#define SDL_ANDROID_GetJNIEnv SDL_AndroidGetJNIEnv\\n'\n )\n\n\nrecipe = AndroidRecipe()\n", "path": "pythonforandroid/recipes/android/__init__.py"}]}
2,086
283
gh_patches_debug_40385
rasdani/github-patches
git_diff
learningequality__kolibri-8277
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> Make learner polling indicator conditional for only SoUD For the new polling indicator on the dropdown menu, this should be updated to appear and poll only when the user is a learner who is logged in on a subset of users device. <img width="270" alt="Screen Shot 2021-08-05 at 10 58 50 AM" src="https://user-images.githubusercontent.com/17235236/128372343-da85c951-2258-4d5e-ade3-a15b753e6806.png"> </issue> <code> [start of kolibri/plugins/facility/kolibri_plugin.py] 1 from __future__ import absolute_import 2 from __future__ import print_function 3 from __future__ import unicode_literals 4 5 from kolibri.core.auth.constants.user_kinds import ADMIN 6 from kolibri.core.hooks import NavigationHook 7 from kolibri.core.hooks import RoleBasedRedirectHook 8 from kolibri.core.webpack.hooks import WebpackBundleHook 9 from kolibri.plugins import KolibriPluginBase 10 from kolibri.plugins.hooks import register_hook 11 12 13 class FacilityManagementPlugin(KolibriPluginBase): 14 translated_view_urls = "urls" 15 untranslated_view_urls = "api_urls" 16 17 18 @register_hook 19 class FacilityManagementAsset(WebpackBundleHook): 20 bundle_id = "app" 21 22 23 @register_hook 24 class FacilityRedirect(RoleBasedRedirectHook): 25 roles = (ADMIN,) 26 27 @property 28 def url(self): 29 return self.plugin_url(FacilityManagementPlugin, "facility_management") 30 31 32 @register_hook 33 class FacilityManagementNavItem(NavigationHook): 34 bundle_id = "side_nav" 35 [end of kolibri/plugins/facility/kolibri_plugin.py] [start of kolibri/plugins/coach/kolibri_plugin.py] 1 from __future__ import absolute_import 2 from __future__ import print_function 3 from __future__ import unicode_literals 4 5 from kolibri.core.auth.constants.user_kinds import COACH 6 from kolibri.core.hooks import NavigationHook 7 from kolibri.core.hooks import RoleBasedRedirectHook 8 from kolibri.core.webpack import hooks as webpack_hooks 9 from kolibri.plugins import KolibriPluginBase 10 from kolibri.plugins.hooks import register_hook 11 12 13 class Coach(KolibriPluginBase): 14 untranslated_view_urls = "api_urls" 15 translated_view_urls = "urls" 16 17 18 @register_hook 19 class CoachRedirect(RoleBasedRedirectHook): 20 roles = (COACH,) 21 22 @property 23 def url(self): 24 return self.plugin_url(Coach, "coach") 25 26 27 @register_hook 28 class CoachNavItem(NavigationHook): 29 bundle_id = "side_nav" 30 31 32 @register_hook 33 class CoachAsset(webpack_hooks.WebpackBundleHook): 34 bundle_id = "app" 35 [end of kolibri/plugins/coach/kolibri_plugin.py] [start of kolibri/core/kolibri_plugin.py] 1 from __future__ import absolute_import 2 from __future__ import print_function 3 from __future__ import unicode_literals 4 5 from django.conf import settings 6 from django.contrib.staticfiles.templatetags.staticfiles import static 7 from django.core.urlresolvers import get_resolver 8 from django.core.urlresolvers import reverse 9 from django.template.loader import render_to_string 10 from django.utils.html import mark_safe 11 from django.utils.translation import get_language 12 from django.utils.translation import get_language_bidi 13 from django.utils.translation import get_language_info 14 from django_js_reverse.core import _safe_json 15 from django_js_reverse.core import generate_json 16 from django_js_reverse.rjsmin import jsmin 17 18 import kolibri 19 from kolibri.core.content.utils.paths import get_content_storage_url 20 from kolibri.core.content.utils.paths import get_hashi_path 21 from kolibri.core.content.utils.paths import get_zip_content_base_path 22 from kolibri.core.content.utils.paths import get_zip_content_config 23 from kolibri.core.device.models import ContentCacheKey 24 from kolibri.core.device.utils import allow_other_browsers_to_connect 25 from kolibri.core.hooks import NavigationHook 26 from kolibri.core.oidc_provider_hook import OIDCProviderHook 27 from kolibri.core.theme_hook import ThemeHook 28 from kolibri.core.webpack.hooks import WebpackBundleHook 29 from kolibri.plugins.app.utils import interface 30 from kolibri.plugins.hooks import register_hook 31 from kolibri.utils import i18n 32 from kolibri.utils.conf import OPTIONS 33 34 35 @register_hook 36 class FrontEndCoreAppAssetHook(WebpackBundleHook): 37 bundle_id = "default_frontend" 38 39 def url_tag(self): 40 # Modified from: 41 # https://github.com/ierror/django-js-reverse/blob/master/django_js_reverse/core.py#L101 42 js_name = "window.kolibriPluginDataGlobal['{bundle}'].urls".format( 43 bundle=self.unique_id 44 ) 45 default_urlresolver = get_resolver(None) 46 47 data = generate_json(default_urlresolver) 48 49 # Generate the JS that exposes functions to reverse all Django URLs 50 # in the frontend. 51 js = render_to_string( 52 "django_js_reverse/urls_js.tpl", 53 {"data": _safe_json(data), "js_name": "__placeholder__"}, 54 # For some reason the js_name gets escaped going into the template 55 # so this was the easiest way to inject it. 56 ).replace("__placeholder__", js_name) 57 zip_content_origin, zip_content_port = get_zip_content_config() 58 return [ 59 mark_safe( 60 """<script type="text/javascript">""" 61 # Minify the generated Javascript 62 + jsmin(js) 63 # Add URL references for our base static URL, the Django media URL 64 # and our content storage URL - this allows us to calculate 65 # the path at which to access a local file on the frontend if needed. 66 + """ 67 {js_name}.__staticUrl = '{static_url}'; 68 {js_name}.__mediaUrl = '{media_url}'; 69 {js_name}.__contentUrl = '{content_url}'; 70 {js_name}.__zipContentUrl = '{zip_content_url}'; 71 {js_name}.__hashiUrl = '{hashi_url}'; 72 {js_name}.__zipContentOrigin = '{zip_content_origin}'; 73 {js_name}.__zipContentPort = {zip_content_port}; 74 </script> 75 """.format( 76 js_name=js_name, 77 static_url=settings.STATIC_URL, 78 media_url=settings.MEDIA_URL, 79 content_url=get_content_storage_url( 80 baseurl=OPTIONS["Deployment"]["URL_PATH_PREFIX"] 81 ), 82 zip_content_url=get_zip_content_base_path(), 83 hashi_url=get_hashi_path(), 84 zip_content_origin=zip_content_origin, 85 zip_content_port=zip_content_port, 86 ) 87 ) 88 ] 89 90 def navigation_tags(self): 91 return [ 92 hook.render_to_page_load_sync_html() 93 for hook in NavigationHook.registered_hooks 94 ] 95 96 def render_to_page_load_sync_html(self): 97 """ 98 Don't render the frontend message files in the usual way 99 as the global object to register them does not exist yet. 100 Instead they are loaded through plugin data. 101 """ 102 tags = ( 103 self.plugin_data_tag() 104 + self.url_tag() 105 + list(self.js_and_css_tags()) 106 + self.navigation_tags() 107 ) 108 109 return mark_safe("\n".join(tags)) 110 111 @property 112 def plugin_data(self): 113 language_code = get_language() 114 static_root = static("assets/fonts/noto-full") 115 full_file = "{}.{}.{}.css?v={}" 116 return { 117 "fullCSSFileModern": full_file.format( 118 static_root, language_code, "modern", kolibri.__version__ 119 ), 120 "fullCSSFileBasic": full_file.format( 121 static_root, language_code, "basic", kolibri.__version__ 122 ), 123 "allowRemoteAccess": allow_other_browsers_to_connect() 124 or not interface.enabled, 125 "appCapabilities": interface.capabilities, 126 "contentCacheKey": ContentCacheKey.get_cache_key(), 127 "languageGlobals": self.language_globals(), 128 "oidcProviderEnabled": OIDCProviderHook.is_enabled(), 129 "kolibriTheme": ThemeHook.get_theme(), 130 } 131 132 def language_globals(self): 133 language_code = get_language() 134 lang_dir = "rtl" if get_language_bidi() else "ltr" 135 136 languages = {} 137 for code, language_name in settings.LANGUAGES: 138 lang_info = i18n.KOLIBRI_LANGUAGE_INFO[code] 139 languages[code] = { 140 # Format to match the schema of the content Language model 141 "id": code, 142 "lang_name": language_name, 143 "english_name": lang_info["english_name"] 144 if lang_info 145 else get_language_info(code)["name"], 146 "lang_direction": get_language_info(code)["bidi"], 147 } 148 return { 149 "coreLanguageMessages": self.frontend_messages(), 150 "languageCode": language_code, 151 "languageDir": lang_dir, 152 "languages": languages, 153 } 154 155 156 @register_hook 157 class FrontendHeadAssetsHook(WebpackBundleHook): 158 """ 159 Render these assets in the <head> tag of base.html, before other JS and assets. 160 """ 161 162 bundle_id = "frontend_head_assets" 163 inline = True 164 165 def render_to_page_load_sync_html(self): 166 """ 167 Add in the extra language font file tags needed 168 for preloading our custom font files. 169 """ 170 tags = ( 171 self.plugin_data_tag() 172 + self.language_font_file_tags() 173 + self.frontend_message_tag() 174 + list(self.js_and_css_tags()) 175 ) 176 177 return mark_safe("\n".join(tags)) 178 179 def language_font_file_tags(self): 180 language_code = get_language() 181 common_file = static("assets/fonts/noto-common.css") 182 subset_file = static("assets/fonts/noto-subset.{}.css".format(language_code)) 183 return [ 184 '<link type="text/css" href="{common_css_file}?v={version}" rel="stylesheet"/>'.format( 185 common_css_file=common_file, version=kolibri.__version__ 186 ), 187 '<link type="text/css" href="{subset_css_file}?v={version}" rel="stylesheet"/>'.format( 188 subset_css_file=subset_file, version=kolibri.__version__ 189 ), 190 ] 191 192 @property 193 def plugin_data(self): 194 return {"unsupportedUrl": reverse("kolibri:core:unsupported")} 195 [end of kolibri/core/kolibri_plugin.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/kolibri/core/kolibri_plugin.py b/kolibri/core/kolibri_plugin.py --- a/kolibri/core/kolibri_plugin.py +++ b/kolibri/core/kolibri_plugin.py @@ -22,6 +22,7 @@ from kolibri.core.content.utils.paths import get_zip_content_config from kolibri.core.device.models import ContentCacheKey from kolibri.core.device.utils import allow_other_browsers_to_connect +from kolibri.core.device.utils import get_device_setting from kolibri.core.hooks import NavigationHook from kolibri.core.oidc_provider_hook import OIDCProviderHook from kolibri.core.theme_hook import ThemeHook @@ -127,6 +128,7 @@ "languageGlobals": self.language_globals(), "oidcProviderEnabled": OIDCProviderHook.is_enabled(), "kolibriTheme": ThemeHook.get_theme(), + "isSubsetOfUsersDevice": get_device_setting("subset_of_users_device"), } def language_globals(self): diff --git a/kolibri/plugins/coach/kolibri_plugin.py b/kolibri/plugins/coach/kolibri_plugin.py --- a/kolibri/plugins/coach/kolibri_plugin.py +++ b/kolibri/plugins/coach/kolibri_plugin.py @@ -3,6 +3,7 @@ from __future__ import unicode_literals from kolibri.core.auth.constants.user_kinds import COACH +from kolibri.core.device.utils import get_device_setting from kolibri.core.hooks import NavigationHook from kolibri.core.hooks import RoleBasedRedirectHook from kolibri.core.webpack import hooks as webpack_hooks @@ -12,7 +13,14 @@ class Coach(KolibriPluginBase): untranslated_view_urls = "api_urls" - translated_view_urls = "urls" + + @property + def translated_view_urls(self): + # On an SoUD this plugin should be disabled. In lieu of properly + # disabling the plugin, we will just not register any urls for now + if not get_device_setting("subset_of_users_device", False): + return "urls" + return None @register_hook diff --git a/kolibri/plugins/facility/kolibri_plugin.py b/kolibri/plugins/facility/kolibri_plugin.py --- a/kolibri/plugins/facility/kolibri_plugin.py +++ b/kolibri/plugins/facility/kolibri_plugin.py @@ -3,6 +3,7 @@ from __future__ import unicode_literals from kolibri.core.auth.constants.user_kinds import ADMIN +from kolibri.core.device.utils import get_device_setting from kolibri.core.hooks import NavigationHook from kolibri.core.hooks import RoleBasedRedirectHook from kolibri.core.webpack.hooks import WebpackBundleHook @@ -11,9 +12,16 @@ class FacilityManagementPlugin(KolibriPluginBase): - translated_view_urls = "urls" untranslated_view_urls = "api_urls" + @property + def translated_view_urls(self): + # On an SoUD this plugin should be disabled. In lieu of properly + # disabling the plugin, we will just not register any urls for now + if not get_device_setting("subset_of_users_device", False): + return "urls" + return None + @register_hook class FacilityManagementAsset(WebpackBundleHook):
{"golden_diff": "diff --git a/kolibri/core/kolibri_plugin.py b/kolibri/core/kolibri_plugin.py\n--- a/kolibri/core/kolibri_plugin.py\n+++ b/kolibri/core/kolibri_plugin.py\n@@ -22,6 +22,7 @@\n from kolibri.core.content.utils.paths import get_zip_content_config\n from kolibri.core.device.models import ContentCacheKey\n from kolibri.core.device.utils import allow_other_browsers_to_connect\n+from kolibri.core.device.utils import get_device_setting\n from kolibri.core.hooks import NavigationHook\n from kolibri.core.oidc_provider_hook import OIDCProviderHook\n from kolibri.core.theme_hook import ThemeHook\n@@ -127,6 +128,7 @@\n \"languageGlobals\": self.language_globals(),\n \"oidcProviderEnabled\": OIDCProviderHook.is_enabled(),\n \"kolibriTheme\": ThemeHook.get_theme(),\n+ \"isSubsetOfUsersDevice\": get_device_setting(\"subset_of_users_device\"),\n }\n \n def language_globals(self):\ndiff --git a/kolibri/plugins/coach/kolibri_plugin.py b/kolibri/plugins/coach/kolibri_plugin.py\n--- a/kolibri/plugins/coach/kolibri_plugin.py\n+++ b/kolibri/plugins/coach/kolibri_plugin.py\n@@ -3,6 +3,7 @@\n from __future__ import unicode_literals\n \n from kolibri.core.auth.constants.user_kinds import COACH\n+from kolibri.core.device.utils import get_device_setting\n from kolibri.core.hooks import NavigationHook\n from kolibri.core.hooks import RoleBasedRedirectHook\n from kolibri.core.webpack import hooks as webpack_hooks\n@@ -12,7 +13,14 @@\n \n class Coach(KolibriPluginBase):\n untranslated_view_urls = \"api_urls\"\n- translated_view_urls = \"urls\"\n+\n+ @property\n+ def translated_view_urls(self):\n+ # On an SoUD this plugin should be disabled. In lieu of properly\n+ # disabling the plugin, we will just not register any urls for now\n+ if not get_device_setting(\"subset_of_users_device\", False):\n+ return \"urls\"\n+ return None\n \n \n @register_hook\ndiff --git a/kolibri/plugins/facility/kolibri_plugin.py b/kolibri/plugins/facility/kolibri_plugin.py\n--- a/kolibri/plugins/facility/kolibri_plugin.py\n+++ b/kolibri/plugins/facility/kolibri_plugin.py\n@@ -3,6 +3,7 @@\n from __future__ import unicode_literals\n \n from kolibri.core.auth.constants.user_kinds import ADMIN\n+from kolibri.core.device.utils import get_device_setting\n from kolibri.core.hooks import NavigationHook\n from kolibri.core.hooks import RoleBasedRedirectHook\n from kolibri.core.webpack.hooks import WebpackBundleHook\n@@ -11,9 +12,16 @@\n \n \n class FacilityManagementPlugin(KolibriPluginBase):\n- translated_view_urls = \"urls\"\n untranslated_view_urls = \"api_urls\"\n \n+ @property\n+ def translated_view_urls(self):\n+ # On an SoUD this plugin should be disabled. In lieu of properly\n+ # disabling the plugin, we will just not register any urls for now\n+ if not get_device_setting(\"subset_of_users_device\", False):\n+ return \"urls\"\n+ return None\n+\n \n @register_hook\n class FacilityManagementAsset(WebpackBundleHook):\n", "issue": "Make learner polling indicator conditional for only SoUD\nFor the new polling indicator on the dropdown menu, this should be updated to appear and poll only when the user is a learner who is logged in on a subset of users device. \r\n\r\n<img width=\"270\" alt=\"Screen Shot 2021-08-05 at 10 58 50 AM\" src=\"https://user-images.githubusercontent.com/17235236/128372343-da85c951-2258-4d5e-ade3-a15b753e6806.png\">\r\n\n", "before_files": [{"content": "from __future__ import absolute_import\nfrom __future__ import print_function\nfrom __future__ import unicode_literals\n\nfrom kolibri.core.auth.constants.user_kinds import ADMIN\nfrom kolibri.core.hooks import NavigationHook\nfrom kolibri.core.hooks import RoleBasedRedirectHook\nfrom kolibri.core.webpack.hooks import WebpackBundleHook\nfrom kolibri.plugins import KolibriPluginBase\nfrom kolibri.plugins.hooks import register_hook\n\n\nclass FacilityManagementPlugin(KolibriPluginBase):\n translated_view_urls = \"urls\"\n untranslated_view_urls = \"api_urls\"\n\n\n@register_hook\nclass FacilityManagementAsset(WebpackBundleHook):\n bundle_id = \"app\"\n\n\n@register_hook\nclass FacilityRedirect(RoleBasedRedirectHook):\n roles = (ADMIN,)\n\n @property\n def url(self):\n return self.plugin_url(FacilityManagementPlugin, \"facility_management\")\n\n\n@register_hook\nclass FacilityManagementNavItem(NavigationHook):\n bundle_id = \"side_nav\"\n", "path": "kolibri/plugins/facility/kolibri_plugin.py"}, {"content": "from __future__ import absolute_import\nfrom __future__ import print_function\nfrom __future__ import unicode_literals\n\nfrom kolibri.core.auth.constants.user_kinds import COACH\nfrom kolibri.core.hooks import NavigationHook\nfrom kolibri.core.hooks import RoleBasedRedirectHook\nfrom kolibri.core.webpack import hooks as webpack_hooks\nfrom kolibri.plugins import KolibriPluginBase\nfrom kolibri.plugins.hooks import register_hook\n\n\nclass Coach(KolibriPluginBase):\n untranslated_view_urls = \"api_urls\"\n translated_view_urls = \"urls\"\n\n\n@register_hook\nclass CoachRedirect(RoleBasedRedirectHook):\n roles = (COACH,)\n\n @property\n def url(self):\n return self.plugin_url(Coach, \"coach\")\n\n\n@register_hook\nclass CoachNavItem(NavigationHook):\n bundle_id = \"side_nav\"\n\n\n@register_hook\nclass CoachAsset(webpack_hooks.WebpackBundleHook):\n bundle_id = \"app\"\n", "path": "kolibri/plugins/coach/kolibri_plugin.py"}, {"content": "from __future__ import absolute_import\nfrom __future__ import print_function\nfrom __future__ import unicode_literals\n\nfrom django.conf import settings\nfrom django.contrib.staticfiles.templatetags.staticfiles import static\nfrom django.core.urlresolvers import get_resolver\nfrom django.core.urlresolvers import reverse\nfrom django.template.loader import render_to_string\nfrom django.utils.html import mark_safe\nfrom django.utils.translation import get_language\nfrom django.utils.translation import get_language_bidi\nfrom django.utils.translation import get_language_info\nfrom django_js_reverse.core import _safe_json\nfrom django_js_reverse.core import generate_json\nfrom django_js_reverse.rjsmin import jsmin\n\nimport kolibri\nfrom kolibri.core.content.utils.paths import get_content_storage_url\nfrom kolibri.core.content.utils.paths import get_hashi_path\nfrom kolibri.core.content.utils.paths import get_zip_content_base_path\nfrom kolibri.core.content.utils.paths import get_zip_content_config\nfrom kolibri.core.device.models import ContentCacheKey\nfrom kolibri.core.device.utils import allow_other_browsers_to_connect\nfrom kolibri.core.hooks import NavigationHook\nfrom kolibri.core.oidc_provider_hook import OIDCProviderHook\nfrom kolibri.core.theme_hook import ThemeHook\nfrom kolibri.core.webpack.hooks import WebpackBundleHook\nfrom kolibri.plugins.app.utils import interface\nfrom kolibri.plugins.hooks import register_hook\nfrom kolibri.utils import i18n\nfrom kolibri.utils.conf import OPTIONS\n\n\n@register_hook\nclass FrontEndCoreAppAssetHook(WebpackBundleHook):\n bundle_id = \"default_frontend\"\n\n def url_tag(self):\n # Modified from:\n # https://github.com/ierror/django-js-reverse/blob/master/django_js_reverse/core.py#L101\n js_name = \"window.kolibriPluginDataGlobal['{bundle}'].urls\".format(\n bundle=self.unique_id\n )\n default_urlresolver = get_resolver(None)\n\n data = generate_json(default_urlresolver)\n\n # Generate the JS that exposes functions to reverse all Django URLs\n # in the frontend.\n js = render_to_string(\n \"django_js_reverse/urls_js.tpl\",\n {\"data\": _safe_json(data), \"js_name\": \"__placeholder__\"},\n # For some reason the js_name gets escaped going into the template\n # so this was the easiest way to inject it.\n ).replace(\"__placeholder__\", js_name)\n zip_content_origin, zip_content_port = get_zip_content_config()\n return [\n mark_safe(\n \"\"\"<script type=\"text/javascript\">\"\"\"\n # Minify the generated Javascript\n + jsmin(js)\n # Add URL references for our base static URL, the Django media URL\n # and our content storage URL - this allows us to calculate\n # the path at which to access a local file on the frontend if needed.\n + \"\"\"\n {js_name}.__staticUrl = '{static_url}';\n {js_name}.__mediaUrl = '{media_url}';\n {js_name}.__contentUrl = '{content_url}';\n {js_name}.__zipContentUrl = '{zip_content_url}';\n {js_name}.__hashiUrl = '{hashi_url}';\n {js_name}.__zipContentOrigin = '{zip_content_origin}';\n {js_name}.__zipContentPort = {zip_content_port};\n </script>\n \"\"\".format(\n js_name=js_name,\n static_url=settings.STATIC_URL,\n media_url=settings.MEDIA_URL,\n content_url=get_content_storage_url(\n baseurl=OPTIONS[\"Deployment\"][\"URL_PATH_PREFIX\"]\n ),\n zip_content_url=get_zip_content_base_path(),\n hashi_url=get_hashi_path(),\n zip_content_origin=zip_content_origin,\n zip_content_port=zip_content_port,\n )\n )\n ]\n\n def navigation_tags(self):\n return [\n hook.render_to_page_load_sync_html()\n for hook in NavigationHook.registered_hooks\n ]\n\n def render_to_page_load_sync_html(self):\n \"\"\"\n Don't render the frontend message files in the usual way\n as the global object to register them does not exist yet.\n Instead they are loaded through plugin data.\n \"\"\"\n tags = (\n self.plugin_data_tag()\n + self.url_tag()\n + list(self.js_and_css_tags())\n + self.navigation_tags()\n )\n\n return mark_safe(\"\\n\".join(tags))\n\n @property\n def plugin_data(self):\n language_code = get_language()\n static_root = static(\"assets/fonts/noto-full\")\n full_file = \"{}.{}.{}.css?v={}\"\n return {\n \"fullCSSFileModern\": full_file.format(\n static_root, language_code, \"modern\", kolibri.__version__\n ),\n \"fullCSSFileBasic\": full_file.format(\n static_root, language_code, \"basic\", kolibri.__version__\n ),\n \"allowRemoteAccess\": allow_other_browsers_to_connect()\n or not interface.enabled,\n \"appCapabilities\": interface.capabilities,\n \"contentCacheKey\": ContentCacheKey.get_cache_key(),\n \"languageGlobals\": self.language_globals(),\n \"oidcProviderEnabled\": OIDCProviderHook.is_enabled(),\n \"kolibriTheme\": ThemeHook.get_theme(),\n }\n\n def language_globals(self):\n language_code = get_language()\n lang_dir = \"rtl\" if get_language_bidi() else \"ltr\"\n\n languages = {}\n for code, language_name in settings.LANGUAGES:\n lang_info = i18n.KOLIBRI_LANGUAGE_INFO[code]\n languages[code] = {\n # Format to match the schema of the content Language model\n \"id\": code,\n \"lang_name\": language_name,\n \"english_name\": lang_info[\"english_name\"]\n if lang_info\n else get_language_info(code)[\"name\"],\n \"lang_direction\": get_language_info(code)[\"bidi\"],\n }\n return {\n \"coreLanguageMessages\": self.frontend_messages(),\n \"languageCode\": language_code,\n \"languageDir\": lang_dir,\n \"languages\": languages,\n }\n\n\n@register_hook\nclass FrontendHeadAssetsHook(WebpackBundleHook):\n \"\"\"\n Render these assets in the <head> tag of base.html, before other JS and assets.\n \"\"\"\n\n bundle_id = \"frontend_head_assets\"\n inline = True\n\n def render_to_page_load_sync_html(self):\n \"\"\"\n Add in the extra language font file tags needed\n for preloading our custom font files.\n \"\"\"\n tags = (\n self.plugin_data_tag()\n + self.language_font_file_tags()\n + self.frontend_message_tag()\n + list(self.js_and_css_tags())\n )\n\n return mark_safe(\"\\n\".join(tags))\n\n def language_font_file_tags(self):\n language_code = get_language()\n common_file = static(\"assets/fonts/noto-common.css\")\n subset_file = static(\"assets/fonts/noto-subset.{}.css\".format(language_code))\n return [\n '<link type=\"text/css\" href=\"{common_css_file}?v={version}\" rel=\"stylesheet\"/>'.format(\n common_css_file=common_file, version=kolibri.__version__\n ),\n '<link type=\"text/css\" href=\"{subset_css_file}?v={version}\" rel=\"stylesheet\"/>'.format(\n subset_css_file=subset_file, version=kolibri.__version__\n ),\n ]\n\n @property\n def plugin_data(self):\n return {\"unsupportedUrl\": reverse(\"kolibri:core:unsupported\")}\n", "path": "kolibri/core/kolibri_plugin.py"}]}
3,342
743
gh_patches_debug_25408
rasdani/github-patches
git_diff
liberapay__liberapay.com-1241
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> Paypal email is not verified. Although I verified it. I see you added Paypal as a payment option which is great. However when I add my paypal email - which is different from liberapay email - it still says it is not verified although I received an email and clicked "verify". Is this a bug? </issue> <code> [start of liberapay/payin/paypal.py] 1 from __future__ import absolute_import, division, print_function, unicode_literals 2 3 import logging 4 5 import requests 6 7 from ..exceptions import PaymentError 8 from ..utils.currencies import Money 9 from ..website import website 10 from .common import update_payin, update_payin_transfer 11 12 13 PAYMENT_STATES_MAP = { 14 'approved': 'succeeded', 15 'created': 'pending', 16 'failed': 'failed', 17 } 18 SALE_STATES_MAP = { 19 'completed': 'succeeded', 20 'denied': 'failed', 21 'pending': 'pending', 22 } 23 24 logger = logging.Logger('paypal') 25 26 session = requests.Session() 27 28 29 def _init_session(): 30 # TODO switch to bearer tokens to reduce the risk of exposing the long-lived secret 31 if 'Authentication' in session.headers: 32 return session 33 from base64 import b64encode 34 session.headers.update({ 35 'Authorization': 'Basic ' + b64encode(( 36 '%s:%s' % (website.app_conf.paypal_id, website.app_conf.paypal_secret) 37 ).encode('ascii')).decode('ascii'), 38 }) 39 return session 40 41 42 def create_payment(db, payin, payer, return_url, state): 43 """Create a Payment. 44 45 Doc: https://developer.paypal.com/docs/api/payments/v1/#payment_create 46 47 Note: even though the API expects a list of transactions it rejects the 48 request if the list contains more than one transaction. 49 """ 50 transfers = db.all(""" 51 SELECT pt.* 52 , recipient.username AS recipient_username 53 , team.username AS team_name 54 , a.id AS merchant_id 55 FROM payin_transfers pt 56 JOIN participants recipient ON recipient.id = pt.recipient 57 LEFT JOIN participants team ON team.id = pt.team 58 JOIN payment_accounts a ON a.pk = pt.destination 59 WHERE pt.payin = %s 60 ORDER BY pt.id 61 """, (payin.id,)) 62 assert transfers 63 locale, _ = state['locale'], state['_'] 64 data = { 65 "intent": "sale", 66 "application_context": { 67 "brand_name": "Liberapay", 68 "locale": locale.language, 69 "landing_page": "Login", 70 "shipping_preference": "NO_SHIPPING", 71 "user_action": "commit", 72 }, 73 "payer": { 74 "payment_method": "paypal" 75 }, 76 "transactions": [{ 77 "amount": { 78 "total": str(pt.amount.amount), 79 "currency": pt.amount.currency 80 }, 81 "description": ( 82 _("donation to {0} for their role in the {1} team", 83 pt.recipient_username, pt.team_name) 84 if pt.team_name else 85 _("donation to {0}", pt.recipient_username) 86 ), 87 "invoice_number": str(pt.id), 88 "note_to_payee": ( 89 "donation via Liberapay for your role in the %s team" % pt.team_name 90 if pt.team_name else 91 "donation via Liberapay" 92 ), 93 "payee": { 94 "email": pt.merchant_id, 95 }, 96 "payment_options": { 97 "allowed_payment_method": "UNRESTRICTED" 98 }, 99 "soft_descriptor": "Liberapay", 100 "item_list": { 101 "items": [{ 102 "name": 103 _("One week of the donation") if pt.period == 'weekly' else 104 _("One month of the donation") if pt.period == 'monthly' else 105 _("One year of the donation"), 106 "quantity": pt.n_units, 107 "price": str(pt.unit_amount.amount), 108 "currency": pt.unit_amount.currency 109 }] 110 } 111 } for pt in transfers], 112 "redirect_urls": { 113 "return_url": return_url, 114 "cancel_url": return_url 115 } 116 } 117 url = 'https://api.%s/v1/payments/payment' % website.app_conf.paypal_domain 118 headers = { 119 'PayPal-Request-Id': 'payin_%i' % payin.id 120 } 121 response = _init_session().post(url, json=data, headers=headers) 122 if response.status_code != 201: 123 error = response.text # for Sentry 124 logger.debug(error) 125 raise PaymentError('PayPal', error) 126 payment = response.json() 127 status = PAYMENT_STATES_MAP[payment['state']] 128 error = payment.get('failure_reason') 129 payin = update_payin(db, payin.id, payment['id'], status, error) 130 if payin.status == 'pending': 131 redirect_url = [l['href'] for l in payment['links'] if l['rel'] == 'approval_url'][0] 132 raise state['response'].redirect(redirect_url) 133 return payin 134 135 136 def execute_payment(db, payin, payer_id): 137 """Execute a previously approved payment. 138 139 Doc: https://developer.paypal.com/docs/api/payments/v1/#payment_execute 140 """ 141 url = 'https://api.%s/v1/payments/payment/%s/execute' % ( 142 website.app_conf.paypal_domain, payin.remote_id 143 ) 144 headers = {'PayPal-Request-Id': 'payin_execute_%i' % payin.id} 145 data = {"payer_id": payer_id} 146 response = _init_session().post(url, json=data, headers=headers) 147 if response.status_code != 200: 148 error = response.text # for Sentry 149 logger.debug(error) 150 raise PaymentError('PayPal') 151 payment = response.json() 152 153 # Update the payin 154 status = PAYMENT_STATES_MAP[payment['state']] 155 error = payment.get('failure_reason') 156 payin = update_payin(db, payin.id, payment['id'], status, error) 157 158 # Update the payin transfers 159 for tr in payment['transactions']: 160 sale = tr.get('related_resources', [{}])[0].get('sale') 161 if sale: 162 pt_id = tr['invoice_number'] 163 pt_remote_id = sale['id'] 164 pt_status = SALE_STATES_MAP[sale['state']] 165 pt_error = sale.get('reason_code') 166 pt_fee = Money(sale['transaction_fee']['value'], sale['transaction_fee']['currency']) 167 charge_amount = Money(sale['amount']['total'], sale['amount']['currency']) 168 net_amount = charge_amount - pt_fee 169 update_payin_transfer( 170 db, pt_id, pt_remote_id, pt_status, pt_error, 171 amount=net_amount, fee=pt_fee 172 ) 173 174 return payin 175 176 177 def sync_payment(db, payin): 178 """Fetch the payment's data and update our database. 179 180 Doc: https://developer.paypal.com/docs/api/payments/v1/#payment_get 181 """ 182 url = 'https://api.%s/v1/payments/payment/%s' % ( 183 website.app_conf.paypal_domain, payin.remote_id 184 ) 185 response = _init_session().get(url) 186 if response.status_code != 200: 187 error = response.text # for Sentry 188 logger.debug(error) 189 raise PaymentError('PayPal') 190 payment = response.json() 191 status = PAYMENT_STATES_MAP[payment['state']] 192 error = payment.get('failure_reason') 193 return update_payin(db, payin.id, payment['id'], status, error) 194 [end of liberapay/payin/paypal.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/liberapay/payin/paypal.py b/liberapay/payin/paypal.py --- a/liberapay/payin/paypal.py +++ b/liberapay/payin/paypal.py @@ -107,7 +107,7 @@ "price": str(pt.unit_amount.amount), "currency": pt.unit_amount.currency }] - } + } if pt.n_units >= 1 else {} } for pt in transfers], "redirect_urls": { "return_url": return_url, @@ -163,9 +163,11 @@ pt_remote_id = sale['id'] pt_status = SALE_STATES_MAP[sale['state']] pt_error = sale.get('reason_code') - pt_fee = Money(sale['transaction_fee']['value'], sale['transaction_fee']['currency']) + pt_fee = sale.get('transaction_fee') + if pt_fee: + pt_fee = Money(pt_fee['value'], pt_fee['currency']) charge_amount = Money(sale['amount']['total'], sale['amount']['currency']) - net_amount = charge_amount - pt_fee + net_amount = charge_amount - (pt_fee or 0) update_payin_transfer( db, pt_id, pt_remote_id, pt_status, pt_error, amount=net_amount, fee=pt_fee
{"golden_diff": "diff --git a/liberapay/payin/paypal.py b/liberapay/payin/paypal.py\n--- a/liberapay/payin/paypal.py\n+++ b/liberapay/payin/paypal.py\n@@ -107,7 +107,7 @@\n \"price\": str(pt.unit_amount.amount),\n \"currency\": pt.unit_amount.currency\n }]\n- }\n+ } if pt.n_units >= 1 else {}\n } for pt in transfers],\n \"redirect_urls\": {\n \"return_url\": return_url,\n@@ -163,9 +163,11 @@\n pt_remote_id = sale['id']\n pt_status = SALE_STATES_MAP[sale['state']]\n pt_error = sale.get('reason_code')\n- pt_fee = Money(sale['transaction_fee']['value'], sale['transaction_fee']['currency'])\n+ pt_fee = sale.get('transaction_fee')\n+ if pt_fee:\n+ pt_fee = Money(pt_fee['value'], pt_fee['currency'])\n charge_amount = Money(sale['amount']['total'], sale['amount']['currency'])\n- net_amount = charge_amount - pt_fee\n+ net_amount = charge_amount - (pt_fee or 0)\n update_payin_transfer(\n db, pt_id, pt_remote_id, pt_status, pt_error,\n amount=net_amount, fee=pt_fee\n", "issue": "Paypal email is not verified. Although I verified it.\nI see you added Paypal as a payment option which is great. However when I add my paypal email - which is different from liberapay email - it still says it is not verified although I received an email and clicked \"verify\". Is this a bug?\n", "before_files": [{"content": "from __future__ import absolute_import, division, print_function, unicode_literals\n\nimport logging\n\nimport requests\n\nfrom ..exceptions import PaymentError\nfrom ..utils.currencies import Money\nfrom ..website import website\nfrom .common import update_payin, update_payin_transfer\n\n\nPAYMENT_STATES_MAP = {\n 'approved': 'succeeded',\n 'created': 'pending',\n 'failed': 'failed',\n}\nSALE_STATES_MAP = {\n 'completed': 'succeeded',\n 'denied': 'failed',\n 'pending': 'pending',\n}\n\nlogger = logging.Logger('paypal')\n\nsession = requests.Session()\n\n\ndef _init_session():\n # TODO switch to bearer tokens to reduce the risk of exposing the long-lived secret\n if 'Authentication' in session.headers:\n return session\n from base64 import b64encode\n session.headers.update({\n 'Authorization': 'Basic ' + b64encode((\n '%s:%s' % (website.app_conf.paypal_id, website.app_conf.paypal_secret)\n ).encode('ascii')).decode('ascii'),\n })\n return session\n\n\ndef create_payment(db, payin, payer, return_url, state):\n \"\"\"Create a Payment.\n\n Doc: https://developer.paypal.com/docs/api/payments/v1/#payment_create\n\n Note: even though the API expects a list of transactions it rejects the\n request if the list contains more than one transaction.\n \"\"\"\n transfers = db.all(\"\"\"\n SELECT pt.*\n , recipient.username AS recipient_username\n , team.username AS team_name\n , a.id AS merchant_id\n FROM payin_transfers pt\n JOIN participants recipient ON recipient.id = pt.recipient\n LEFT JOIN participants team ON team.id = pt.team\n JOIN payment_accounts a ON a.pk = pt.destination\n WHERE pt.payin = %s\n ORDER BY pt.id\n \"\"\", (payin.id,))\n assert transfers\n locale, _ = state['locale'], state['_']\n data = {\n \"intent\": \"sale\",\n \"application_context\": {\n \"brand_name\": \"Liberapay\",\n \"locale\": locale.language,\n \"landing_page\": \"Login\",\n \"shipping_preference\": \"NO_SHIPPING\",\n \"user_action\": \"commit\",\n },\n \"payer\": {\n \"payment_method\": \"paypal\"\n },\n \"transactions\": [{\n \"amount\": {\n \"total\": str(pt.amount.amount),\n \"currency\": pt.amount.currency\n },\n \"description\": (\n _(\"donation to {0} for their role in the {1} team\",\n pt.recipient_username, pt.team_name)\n if pt.team_name else\n _(\"donation to {0}\", pt.recipient_username)\n ),\n \"invoice_number\": str(pt.id),\n \"note_to_payee\": (\n \"donation via Liberapay for your role in the %s team\" % pt.team_name\n if pt.team_name else\n \"donation via Liberapay\"\n ),\n \"payee\": {\n \"email\": pt.merchant_id,\n },\n \"payment_options\": {\n \"allowed_payment_method\": \"UNRESTRICTED\"\n },\n \"soft_descriptor\": \"Liberapay\",\n \"item_list\": {\n \"items\": [{\n \"name\":\n _(\"One week of the donation\") if pt.period == 'weekly' else\n _(\"One month of the donation\") if pt.period == 'monthly' else\n _(\"One year of the donation\"),\n \"quantity\": pt.n_units,\n \"price\": str(pt.unit_amount.amount),\n \"currency\": pt.unit_amount.currency\n }]\n }\n } for pt in transfers],\n \"redirect_urls\": {\n \"return_url\": return_url,\n \"cancel_url\": return_url\n }\n }\n url = 'https://api.%s/v1/payments/payment' % website.app_conf.paypal_domain\n headers = {\n 'PayPal-Request-Id': 'payin_%i' % payin.id\n }\n response = _init_session().post(url, json=data, headers=headers)\n if response.status_code != 201:\n error = response.text # for Sentry\n logger.debug(error)\n raise PaymentError('PayPal', error)\n payment = response.json()\n status = PAYMENT_STATES_MAP[payment['state']]\n error = payment.get('failure_reason')\n payin = update_payin(db, payin.id, payment['id'], status, error)\n if payin.status == 'pending':\n redirect_url = [l['href'] for l in payment['links'] if l['rel'] == 'approval_url'][0]\n raise state['response'].redirect(redirect_url)\n return payin\n\n\ndef execute_payment(db, payin, payer_id):\n \"\"\"Execute a previously approved payment.\n\n Doc: https://developer.paypal.com/docs/api/payments/v1/#payment_execute\n \"\"\"\n url = 'https://api.%s/v1/payments/payment/%s/execute' % (\n website.app_conf.paypal_domain, payin.remote_id\n )\n headers = {'PayPal-Request-Id': 'payin_execute_%i' % payin.id}\n data = {\"payer_id\": payer_id}\n response = _init_session().post(url, json=data, headers=headers)\n if response.status_code != 200:\n error = response.text # for Sentry\n logger.debug(error)\n raise PaymentError('PayPal')\n payment = response.json()\n\n # Update the payin\n status = PAYMENT_STATES_MAP[payment['state']]\n error = payment.get('failure_reason')\n payin = update_payin(db, payin.id, payment['id'], status, error)\n\n # Update the payin transfers\n for tr in payment['transactions']:\n sale = tr.get('related_resources', [{}])[0].get('sale')\n if sale:\n pt_id = tr['invoice_number']\n pt_remote_id = sale['id']\n pt_status = SALE_STATES_MAP[sale['state']]\n pt_error = sale.get('reason_code')\n pt_fee = Money(sale['transaction_fee']['value'], sale['transaction_fee']['currency'])\n charge_amount = Money(sale['amount']['total'], sale['amount']['currency'])\n net_amount = charge_amount - pt_fee\n update_payin_transfer(\n db, pt_id, pt_remote_id, pt_status, pt_error,\n amount=net_amount, fee=pt_fee\n )\n\n return payin\n\n\ndef sync_payment(db, payin):\n \"\"\"Fetch the payment's data and update our database.\n\n Doc: https://developer.paypal.com/docs/api/payments/v1/#payment_get\n \"\"\"\n url = 'https://api.%s/v1/payments/payment/%s' % (\n website.app_conf.paypal_domain, payin.remote_id\n )\n response = _init_session().get(url)\n if response.status_code != 200:\n error = response.text # for Sentry\n logger.debug(error)\n raise PaymentError('PayPal')\n payment = response.json()\n status = PAYMENT_STATES_MAP[payment['state']]\n error = payment.get('failure_reason')\n return update_payin(db, payin.id, payment['id'], status, error)\n", "path": "liberapay/payin/paypal.py"}]}
2,651
296
gh_patches_debug_19739
rasdani/github-patches
git_diff
cocotb__cocotb-278
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> getvalue() in handle.py should configure the returned Binary value When getting a value from a handle a "BinaryValue()" object is created. In my case, I tried to slice that value to get the upper bits, but the call failed because the returned value is not configured. I believe at a minimum, the Handle should set the number of bits, but ideally it would just configure everything the BinaryValue() object needs before returning it. </issue> <code> [start of cocotb/handle.py] 1 #!/usr/bin/env python 2 3 ''' Copyright (c) 2013 Potential Ventures Ltd 4 Copyright (c) 2013 SolarFlare Communications Inc 5 All rights reserved. 6 7 Redistribution and use in source and binary forms, with or without 8 modification, are permitted provided that the following conditions are met: 9 * Redistributions of source code must retain the above copyright 10 notice, this list of conditions and the following disclaimer. 11 * Redistributions in binary form must reproduce the above copyright 12 notice, this list of conditions and the following disclaimer in the 13 documentation and/or other materials provided with the distribution. 14 * Neither the name of Potential Ventures Ltd, 15 SolarFlare Communications Inc nor the 16 names of its contributors may be used to endorse or promote products 17 derived from this software without specific prior written permission. 18 19 THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND 20 ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED 21 WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE 22 DISCLAIMED. IN NO EVENT SHALL POTENTIAL VENTURES LTD BE LIABLE FOR ANY 23 DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES 24 (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; 25 LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND 26 ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT 27 (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS 28 SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. ''' 29 30 # -*- coding: utf-8 -*- 31 32 import logging 33 import ctypes 34 import traceback 35 import sys 36 # from StringIO import StringIO 37 38 from io import StringIO, BytesIO 39 40 import os 41 42 # For autodocumentation don't need the extension modules 43 if "SPHINX_BUILD" in os.environ: 44 simulator = None 45 else: 46 import simulator 47 48 import cocotb 49 from cocotb.binary import BinaryValue 50 from cocotb.log import SimLog 51 from cocotb.result import TestError 52 from cocotb.triggers import _RisingEdge, _FallingEdge 53 from cocotb.utils import get_python_integer_types 54 55 56 class SimHandle(object): 57 58 def __init__(self, handle): 59 """ 60 Args: 61 _handle [integer] : vpi/vhpi handle to the simulator object 62 """ 63 self._handle = handle # handle used for future simulator transactions 64 self._sub_handles = {} # Dict. of SimHandle objects created by getattr 65 self._len = None 66 67 self.name = simulator.get_name_string(self._handle) 68 self.fullname = '%s(%s)' % (self.name, 69 simulator.get_type_string(self._handle)) 70 self.log = SimLog('cocotb.' + self.name) 71 self.log.debug("Created!") 72 self._r_edge = _RisingEdge(self) 73 self._f_edge = _FallingEdge(self) 74 75 def __hash__(self): 76 return self._handle 77 78 def __str__(self): 79 return "%s @0x%x" % (self.name, self._handle) 80 81 def __getattr__(self, name): 82 """ Query the simulator for a object with the specified name 83 and cache the result to build a tree 84 """ 85 # python's builtin dir and IPython's dir2 search for these, 86 # raise an AttributeError to avoid incorrect calls to _raise_testerror 87 if name in ["__methods__","__members__","trait_names","_getAttributeNames"]: 88 raise AttributeError(name) 89 if name in self._sub_handles: 90 return self._sub_handles[name] 91 new_handle = simulator.get_handle_by_name(self._handle, name) 92 if not new_handle: 93 self._raise_testerror("%s contains no object named %s" % 94 (self.name, name)) 95 self._sub_handles[name] = SimHandle(new_handle) 96 return self._sub_handles[name] 97 98 def _raise_testerror(self, msg): 99 lastframe = sys._getframe(2) 100 if sys.version_info[0] >= 3: 101 buff = StringIO() 102 traceback.print_stack(lastframe, file=buff) 103 else: 104 buff_bytes = BytesIO() 105 traceback.print_stack(lastframe, file=buff_bytes) 106 buff = StringIO(buff_bytes.getvalue().decode("UTF8")) 107 self.log.error("%s\n%s" % (msg, buff.getvalue())) 108 exception = TestError(msg) 109 exception.stderr.write(buff.getvalue()) 110 buff.close() 111 raise exception 112 113 def __setattr__(self, name, value): 114 """Provide transparent access to signals""" 115 if (not name.startswith('_') and 116 not name in ["name", "fullname", "log", "value"] and 117 self.__hasattr__(name)): 118 getattr(self, name).setcachedvalue(value) 119 return 120 object.__setattr__(self, name, value) 121 122 def __hasattr__(self, name): 123 """ 124 Since calling hasattr(handle, "something") will print out a 125 backtrace to the log since usually attempting to access a 126 non-existent member is an error we provide a 'peek function 127 128 We still add the found handle to our dictionary to prevent leaking 129 handles. 130 """ 131 if name in self._sub_handles: 132 return self._sub_handles[name] 133 new_handle = simulator.get_handle_by_name(self._handle, name) 134 if new_handle: 135 self._sub_handles[name] = SimHandle(new_handle) 136 return new_handle 137 138 def __getitem__(self, index): 139 if index in self._sub_handles: 140 return self._sub_handles[index] 141 new_handle = simulator.get_handle_by_index(self._handle, index) 142 if not new_handle: 143 self._raise_testerror("%s contains no object at index %d" % 144 (self.name, index)) 145 self._sub_handles[index] = SimHandle(new_handle) 146 return self._sub_handles[index] 147 148 def __setitem__(self, index, value): 149 """Provide transparent assignment to bit index""" 150 self.__getitem__(index).setcachedvalue(value) 151 152 def getvalue(self): 153 result = BinaryValue() 154 result.binstr = self._get_value_str() 155 return result 156 157 def setimmediatevalue(self, value): 158 """ 159 Set the value of the underlying simulation object to value. 160 161 Args: 162 value (ctypes.Structure, cocotb.binary.BinaryValue, int) 163 The value to drive onto the simulator object 164 165 Raises: 166 TypeError 167 168 This operation will fail unless the handle refers to a modifiable 169 object eg net, signal or variable. 170 171 We determine the library call to make based on the type of the value 172 173 Assigning integers less than 32-bits is faster 174 """ 175 if (isinstance(value, get_python_integer_types()) and 176 value < 0x7fffffff): 177 simulator.set_signal_val(self._handle, value) 178 return 179 180 if isinstance(value, ctypes.Structure): 181 value = BinaryValue(value=cocotb.utils.pack(value), bits=len(self)) 182 elif isinstance(value, get_python_integer_types()): 183 value = BinaryValue(value=value, bits=len(self), bigEndian=False) 184 elif not isinstance(value, BinaryValue): 185 self.log.critical("Unsupported type for value assignment: %s (%s)" 186 % (type(value), repr(value))) 187 raise TypeError("Unable to set simulator value with type %s" % 188 (type(value))) 189 190 simulator.set_signal_val_str(self._handle, value.binstr) 191 192 def setcachedvalue(self, value): 193 """Intercept the store of a value and hold in cache. 194 195 This operation is to enable all of the scheduled callbacks to completed 196 with the same read data and for the writes to occour on the next 197 sim time""" 198 cocotb.scheduler.save_write(self, value) 199 200 # We want to maintain compatability with python 2.5 so we can't use 201 # @property with a setter 202 value = property(getvalue, setcachedvalue, None, 203 "A reference to the value") 204 205 def _get_value_str(self): 206 return simulator.get_signal_val(self._handle) 207 208 def __le__(self, value): 209 """Overload the less than or equal to operator to 210 provide an hdl-like shortcut 211 module.signal <= 2 212 """ 213 self.value = value 214 215 def __len__(self): 216 """Returns the 'length' of the underlying object. 217 218 For vectors this is the number of bits. 219 220 TODO: Handle other types (loops, generate etc) 221 """ 222 if self._len is None: 223 self._len = len(self._get_value_str()) 224 return self._len 225 226 def __cmp__(self, other): 227 228 # Permits comparison of handles i.e. if clk == dut.clk 229 if isinstance(other, SimHandle): 230 if self._handle == other._handle: 231 return 0 232 return 1 233 234 # Use the comparison method of the other object against our value 235 return self.value.__cmp__(other) 236 237 def __iter__(self): 238 """Iterates over all known types defined by simulator module""" 239 for handle_type in [simulator.MODULE, 240 simulator.PARAMETER, 241 simulator.REG, 242 simulator.NET, 243 simulator.NETARRAY]: 244 iterator = simulator.iterate(handle_type, self._handle) 245 while True: 246 try: 247 thing = simulator.next(iterator) 248 except StopIteration: 249 break 250 hdl = SimHandle(thing) 251 self._sub_handles[hdl.name] = hdl 252 yield hdl 253 254 def __int__(self): 255 return int(self.value) 256 [end of cocotb/handle.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/cocotb/handle.py b/cocotb/handle.py --- a/cocotb/handle.py +++ b/cocotb/handle.py @@ -82,7 +82,7 @@ """ Query the simulator for a object with the specified name and cache the result to build a tree """ - # python's builtin dir and IPython's dir2 search for these, + # python's builtin dir and IPython's dir2 search for these, # raise an AttributeError to avoid incorrect calls to _raise_testerror if name in ["__methods__","__members__","trait_names","_getAttributeNames"]: raise AttributeError(name) @@ -150,7 +150,7 @@ self.__getitem__(index).setcachedvalue(value) def getvalue(self): - result = BinaryValue() + result = BinaryValue(bits=len(self)) result.binstr = self._get_value_str() return result
{"golden_diff": "diff --git a/cocotb/handle.py b/cocotb/handle.py\n--- a/cocotb/handle.py\n+++ b/cocotb/handle.py\n@@ -82,7 +82,7 @@\n \"\"\" Query the simulator for a object with the specified name\n and cache the result to build a tree\n \"\"\"\n- # python's builtin dir and IPython's dir2 search for these, \n+ # python's builtin dir and IPython's dir2 search for these,\n # raise an AttributeError to avoid incorrect calls to _raise_testerror\n if name in [\"__methods__\",\"__members__\",\"trait_names\",\"_getAttributeNames\"]:\n raise AttributeError(name)\n@@ -150,7 +150,7 @@\n self.__getitem__(index).setcachedvalue(value)\n \n def getvalue(self):\n- result = BinaryValue()\n+ result = BinaryValue(bits=len(self))\n result.binstr = self._get_value_str()\n return result\n", "issue": "getvalue() in handle.py should configure the returned Binary value\nWhen getting a value from a handle a \"BinaryValue()\" object is created. In my case, I tried to slice that value to get the upper bits, but the call failed because the returned value is not configured.\n\nI believe at a minimum, the Handle should set the number of bits, but ideally it would just configure everything the BinaryValue() object needs before returning it.\n\n", "before_files": [{"content": "#!/usr/bin/env python\n\n''' Copyright (c) 2013 Potential Ventures Ltd\nCopyright (c) 2013 SolarFlare Communications Inc\nAll rights reserved.\n\nRedistribution and use in source and binary forms, with or without\nmodification, are permitted provided that the following conditions are met:\n * Redistributions of source code must retain the above copyright\n notice, this list of conditions and the following disclaimer.\n * Redistributions in binary form must reproduce the above copyright\n notice, this list of conditions and the following disclaimer in the\n documentation and/or other materials provided with the distribution.\n * Neither the name of Potential Ventures Ltd,\n SolarFlare Communications Inc nor the\n names of its contributors may be used to endorse or promote products\n derived from this software without specific prior written permission.\n\nTHIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS \"AS IS\" AND\nANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED\nWARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE\nDISCLAIMED. IN NO EVENT SHALL POTENTIAL VENTURES LTD BE LIABLE FOR ANY\nDIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES\n(INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;\nLOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND\nON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT\n(INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS\nSOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. '''\n\n# -*- coding: utf-8 -*-\n\nimport logging\nimport ctypes\nimport traceback\nimport sys\n# from StringIO import StringIO\n\nfrom io import StringIO, BytesIO\n\nimport os\n\n# For autodocumentation don't need the extension modules\nif \"SPHINX_BUILD\" in os.environ:\n simulator = None\nelse:\n import simulator\n\nimport cocotb\nfrom cocotb.binary import BinaryValue\nfrom cocotb.log import SimLog\nfrom cocotb.result import TestError\nfrom cocotb.triggers import _RisingEdge, _FallingEdge\nfrom cocotb.utils import get_python_integer_types\n\n\nclass SimHandle(object):\n\n def __init__(self, handle):\n \"\"\"\n Args:\n _handle [integer] : vpi/vhpi handle to the simulator object\n \"\"\"\n self._handle = handle # handle used for future simulator transactions\n self._sub_handles = {} # Dict. of SimHandle objects created by getattr\n self._len = None\n\n self.name = simulator.get_name_string(self._handle)\n self.fullname = '%s(%s)' % (self.name,\n simulator.get_type_string(self._handle))\n self.log = SimLog('cocotb.' + self.name)\n self.log.debug(\"Created!\")\n self._r_edge = _RisingEdge(self)\n self._f_edge = _FallingEdge(self)\n\n def __hash__(self):\n return self._handle\n\n def __str__(self):\n return \"%s @0x%x\" % (self.name, self._handle)\n\n def __getattr__(self, name):\n \"\"\" Query the simulator for a object with the specified name\n and cache the result to build a tree\n \"\"\"\n # python's builtin dir and IPython's dir2 search for these, \n # raise an AttributeError to avoid incorrect calls to _raise_testerror\n if name in [\"__methods__\",\"__members__\",\"trait_names\",\"_getAttributeNames\"]:\n raise AttributeError(name)\n if name in self._sub_handles:\n return self._sub_handles[name]\n new_handle = simulator.get_handle_by_name(self._handle, name)\n if not new_handle:\n self._raise_testerror(\"%s contains no object named %s\" %\n (self.name, name))\n self._sub_handles[name] = SimHandle(new_handle)\n return self._sub_handles[name]\n\n def _raise_testerror(self, msg):\n lastframe = sys._getframe(2)\n if sys.version_info[0] >= 3:\n buff = StringIO()\n traceback.print_stack(lastframe, file=buff)\n else:\n buff_bytes = BytesIO()\n traceback.print_stack(lastframe, file=buff_bytes)\n buff = StringIO(buff_bytes.getvalue().decode(\"UTF8\"))\n self.log.error(\"%s\\n%s\" % (msg, buff.getvalue()))\n exception = TestError(msg)\n exception.stderr.write(buff.getvalue())\n buff.close()\n raise exception\n\n def __setattr__(self, name, value):\n \"\"\"Provide transparent access to signals\"\"\"\n if (not name.startswith('_') and\n not name in [\"name\", \"fullname\", \"log\", \"value\"] and\n self.__hasattr__(name)):\n getattr(self, name).setcachedvalue(value)\n return\n object.__setattr__(self, name, value)\n\n def __hasattr__(self, name):\n \"\"\"\n Since calling hasattr(handle, \"something\") will print out a\n backtrace to the log since usually attempting to access a\n non-existent member is an error we provide a 'peek function\n\n We still add the found handle to our dictionary to prevent leaking\n handles.\n \"\"\"\n if name in self._sub_handles:\n return self._sub_handles[name]\n new_handle = simulator.get_handle_by_name(self._handle, name)\n if new_handle:\n self._sub_handles[name] = SimHandle(new_handle)\n return new_handle\n\n def __getitem__(self, index):\n if index in self._sub_handles:\n return self._sub_handles[index]\n new_handle = simulator.get_handle_by_index(self._handle, index)\n if not new_handle:\n self._raise_testerror(\"%s contains no object at index %d\" %\n (self.name, index))\n self._sub_handles[index] = SimHandle(new_handle)\n return self._sub_handles[index]\n\n def __setitem__(self, index, value):\n \"\"\"Provide transparent assignment to bit index\"\"\"\n self.__getitem__(index).setcachedvalue(value)\n\n def getvalue(self):\n result = BinaryValue()\n result.binstr = self._get_value_str()\n return result\n\n def setimmediatevalue(self, value):\n \"\"\"\n Set the value of the underlying simulation object to value.\n\n Args:\n value (ctypes.Structure, cocotb.binary.BinaryValue, int)\n The value to drive onto the simulator object\n\n Raises:\n TypeError\n\n This operation will fail unless the handle refers to a modifiable\n object eg net, signal or variable.\n\n We determine the library call to make based on the type of the value\n\n Assigning integers less than 32-bits is faster\n \"\"\"\n if (isinstance(value, get_python_integer_types()) and\n value < 0x7fffffff):\n simulator.set_signal_val(self._handle, value)\n return\n\n if isinstance(value, ctypes.Structure):\n value = BinaryValue(value=cocotb.utils.pack(value), bits=len(self))\n elif isinstance(value, get_python_integer_types()):\n value = BinaryValue(value=value, bits=len(self), bigEndian=False)\n elif not isinstance(value, BinaryValue):\n self.log.critical(\"Unsupported type for value assignment: %s (%s)\"\n % (type(value), repr(value)))\n raise TypeError(\"Unable to set simulator value with type %s\" %\n (type(value)))\n\n simulator.set_signal_val_str(self._handle, value.binstr)\n\n def setcachedvalue(self, value):\n \"\"\"Intercept the store of a value and hold in cache.\n\n This operation is to enable all of the scheduled callbacks to completed\n with the same read data and for the writes to occour on the next\n sim time\"\"\"\n cocotb.scheduler.save_write(self, value)\n\n # We want to maintain compatability with python 2.5 so we can't use\n # @property with a setter\n value = property(getvalue, setcachedvalue, None,\n \"A reference to the value\")\n\n def _get_value_str(self):\n return simulator.get_signal_val(self._handle)\n\n def __le__(self, value):\n \"\"\"Overload the less than or equal to operator to\n provide an hdl-like shortcut\n module.signal <= 2\n \"\"\"\n self.value = value\n\n def __len__(self):\n \"\"\"Returns the 'length' of the underlying object.\n\n For vectors this is the number of bits.\n\n TODO: Handle other types (loops, generate etc)\n \"\"\"\n if self._len is None:\n self._len = len(self._get_value_str())\n return self._len\n\n def __cmp__(self, other):\n\n # Permits comparison of handles i.e. if clk == dut.clk\n if isinstance(other, SimHandle):\n if self._handle == other._handle:\n return 0\n return 1\n\n # Use the comparison method of the other object against our value\n return self.value.__cmp__(other)\n\n def __iter__(self):\n \"\"\"Iterates over all known types defined by simulator module\"\"\"\n for handle_type in [simulator.MODULE,\n simulator.PARAMETER,\n simulator.REG,\n simulator.NET,\n simulator.NETARRAY]:\n iterator = simulator.iterate(handle_type, self._handle)\n while True:\n try:\n thing = simulator.next(iterator)\n except StopIteration:\n break\n hdl = SimHandle(thing)\n self._sub_handles[hdl.name] = hdl\n yield hdl\n\n def __int__(self):\n return int(self.value)\n", "path": "cocotb/handle.py"}]}
3,365
217
gh_patches_debug_20978
rasdani/github-patches
git_diff
python-poetry__poetry-1635
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> 1.0.0b5 removes my private source from the lockfile <!-- Hi there! Thank you for discovering and submitting an issue. Before you submit this; let's make sure of a few things. Please make sure the following boxes are ticked if they are correct. If not, please try and fulfill these first. --> <!-- Checked checkbox should look like this: [x] --> - [x] I am on the [latest](https://github.com/sdispater/poetry/releases/latest) Poetry version. - [x] I have searched the [issues](https://github.com/sdispater/poetry/issues) of this repo and believe that this is not a duplicate. - [x] If an exception occurs when executing a command, I executed it again in debug mode (`-vvv` option). <!-- Once those are done, if you're able to fill in the following list with your information, it'd be very helpful to whoever handles the issue. --> - **OS version and name**: macOS 10.14.6 (18G103) - **Poetry version**: <!-- 1.0.0b5 - **Link of a [Gist](https://gist.github.com/) with the contents of your pyproject.toml file**: <!-- Gist Link Here --> https://gist.github.com/JBKahn/4bbb040c0181d974e0b71407ad764b6c ## Issue When moving from `1.0.0b4` to `1.0.0b5` `poetry lock` is removing the ``` [package.source] reference = "private" type = "legacy" url = "https://private.url" ``` </issue> <code> [start of poetry/packages/locker.py] 1 import json 2 import re 3 4 from hashlib import sha256 5 from typing import List 6 7 from tomlkit import document 8 from tomlkit import inline_table 9 from tomlkit import item 10 from tomlkit import table 11 from tomlkit.exceptions import TOMLKitError 12 13 import poetry.packages 14 import poetry.repositories 15 16 from poetry.utils._compat import Path 17 from poetry.utils.toml_file import TomlFile 18 from poetry.version.markers import parse_marker 19 20 21 class Locker(object): 22 23 _relevant_keys = ["dependencies", "dev-dependencies", "source", "extras"] 24 25 def __init__(self, lock, local_config): # type: (Path, dict) -> None 26 self._lock = TomlFile(lock) 27 self._local_config = local_config 28 self._lock_data = None 29 self._content_hash = self._get_content_hash() 30 31 @property 32 def lock(self): # type: () -> TomlFile 33 return self._lock 34 35 @property 36 def lock_data(self): 37 if self._lock_data is None: 38 self._lock_data = self._get_lock_data() 39 40 return self._lock_data 41 42 def is_locked(self): # type: () -> bool 43 """ 44 Checks whether the locker has been locked (lockfile found). 45 """ 46 if not self._lock.exists(): 47 return False 48 49 return "package" in self.lock_data 50 51 def is_fresh(self): # type: () -> bool 52 """ 53 Checks whether the lock file is still up to date with the current hash. 54 """ 55 lock = self._lock.read() 56 metadata = lock.get("metadata", {}) 57 58 if "content-hash" in metadata: 59 return self._content_hash == lock["metadata"]["content-hash"] 60 61 return False 62 63 def locked_repository( 64 self, with_dev_reqs=False 65 ): # type: (bool) -> poetry.repositories.Repository 66 """ 67 Searches and returns a repository of locked packages. 68 """ 69 if not self.is_locked(): 70 return poetry.repositories.Repository() 71 72 lock_data = self.lock_data 73 packages = poetry.repositories.Repository() 74 75 if with_dev_reqs: 76 locked_packages = lock_data["package"] 77 else: 78 locked_packages = [ 79 p for p in lock_data["package"] if p["category"] == "main" 80 ] 81 82 if not locked_packages: 83 return packages 84 85 for info in locked_packages: 86 package = poetry.packages.Package( 87 info["name"], info["version"], info["version"] 88 ) 89 package.description = info.get("description", "") 90 package.category = info["category"] 91 package.optional = info["optional"] 92 if "hashes" in lock_data["metadata"]: 93 # Old lock so we create dummy files from the hashes 94 package.files = [ 95 {"name": h, "hash": h} 96 for h in lock_data["metadata"]["hashes"][info["name"]] 97 ] 98 else: 99 package.files = lock_data["metadata"]["files"][info["name"]] 100 101 package.python_versions = info["python-versions"] 102 extras = info.get("extras", {}) 103 if extras: 104 for name, deps in extras.items(): 105 package.extras[name] = [] 106 107 for dep in deps: 108 m = re.match(r"^(.+?)(?:\s+\((.+)\))?$", dep) 109 dep_name = m.group(1) 110 constraint = m.group(2) or "*" 111 112 package.extras[name].append( 113 poetry.packages.Dependency(dep_name, constraint) 114 ) 115 116 if "marker" in info: 117 package.marker = parse_marker(info["marker"]) 118 else: 119 # Compatibility for old locks 120 if "requirements" in info: 121 dep = poetry.packages.Dependency("foo", "0.0.0") 122 for name, value in info["requirements"].items(): 123 if name == "python": 124 dep.python_versions = value 125 elif name == "platform": 126 dep.platform = value 127 128 split_dep = dep.to_pep_508(False).split(";") 129 if len(split_dep) > 1: 130 package.marker = parse_marker(split_dep[1].strip()) 131 132 for dep_name, constraint in info.get("dependencies", {}).items(): 133 if isinstance(constraint, list): 134 for c in constraint: 135 package.add_dependency(dep_name, c) 136 137 continue 138 139 package.add_dependency(dep_name, constraint) 140 141 if "source" in info: 142 package.source_type = info["source"]["type"] 143 package.source_url = info["source"]["url"] 144 package.source_reference = info["source"]["reference"] 145 146 packages.add_package(package) 147 148 return packages 149 150 def set_lock_data(self, root, packages): # type: (...) -> bool 151 files = table() 152 packages = self._lock_packages(packages) 153 # Retrieving hashes 154 for package in packages: 155 if package["name"] not in files: 156 files[package["name"]] = [] 157 158 for f in package["files"]: 159 file_metadata = inline_table() 160 for k, v in sorted(f.items()): 161 file_metadata[k] = v 162 163 files[package["name"]].append(file_metadata) 164 165 if files[package["name"]]: 166 files[package["name"]] = item(files[package["name"]]).multiline(True) 167 168 del package["files"] 169 170 lock = document() 171 lock["package"] = packages 172 173 if root.extras: 174 lock["extras"] = { 175 extra: [dep.pretty_name for dep in deps] 176 for extra, deps in root.extras.items() 177 } 178 179 lock["metadata"] = { 180 "python-versions": root.python_versions, 181 "content-hash": self._content_hash, 182 "files": files, 183 } 184 185 if not self.is_locked() or lock != self.lock_data: 186 self._write_lock_data(lock) 187 188 return True 189 190 return False 191 192 def _write_lock_data(self, data): 193 self.lock.write(data) 194 195 # Checking lock file data consistency 196 if data != self.lock.read(): 197 raise RuntimeError("Inconsistent lock file data.") 198 199 self._lock_data = None 200 201 def _get_content_hash(self): # type: () -> str 202 """ 203 Returns the sha256 hash of the sorted content of the pyproject file. 204 """ 205 content = self._local_config 206 207 relevant_content = {} 208 for key in self._relevant_keys: 209 relevant_content[key] = content.get(key) 210 211 content_hash = sha256( 212 json.dumps(relevant_content, sort_keys=True).encode() 213 ).hexdigest() 214 215 return content_hash 216 217 def _get_lock_data(self): # type: () -> dict 218 if not self._lock.exists(): 219 raise RuntimeError("No lockfile found. Unable to read locked packages") 220 221 try: 222 return self._lock.read() 223 except TOMLKitError as e: 224 raise RuntimeError("Unable to read the lock file ({}).".format(e)) 225 226 def _lock_packages( 227 self, packages 228 ): # type: (List['poetry.packages.Package']) -> list 229 locked = [] 230 231 for package in sorted(packages, key=lambda x: x.name): 232 spec = self._dump_package(package) 233 234 locked.append(spec) 235 236 return locked 237 238 def _dump_package(self, package): # type: (poetry.packages.Package) -> dict 239 dependencies = {} 240 for dependency in sorted(package.requires, key=lambda d: d.name): 241 if dependency.is_optional() and not dependency.is_activated(): 242 continue 243 244 if dependency.pretty_name not in dependencies: 245 dependencies[dependency.pretty_name] = [] 246 247 constraint = {"version": str(dependency.pretty_constraint)} 248 249 if dependency.extras: 250 constraint["extras"] = dependency.extras 251 252 if dependency.is_optional(): 253 constraint["optional"] = True 254 255 if not dependency.python_constraint.is_any(): 256 constraint["python"] = str(dependency.python_constraint) 257 258 dependencies[dependency.pretty_name].append(constraint) 259 260 # All the constraints should have the same type, 261 # but we want to simplify them if it's possible 262 for dependency, constraints in tuple(dependencies.items()): 263 if all(len(constraint) == 1 for constraint in constraints): 264 dependencies[dependency] = [ 265 constraint["version"] for constraint in constraints 266 ] 267 268 data = { 269 "name": package.pretty_name, 270 "version": package.pretty_version, 271 "description": package.description or "", 272 "category": package.category, 273 "optional": package.optional, 274 "python-versions": package.python_versions, 275 "files": sorted(package.files, key=lambda x: x["file"]), 276 } 277 if not package.marker.is_any(): 278 data["marker"] = str(package.marker) 279 280 if package.extras: 281 extras = {} 282 for name, deps in package.extras.items(): 283 extras[name] = [ 284 str(dep) if not dep.constraint.is_any() else dep.name 285 for dep in deps 286 ] 287 288 data["extras"] = extras 289 290 if dependencies: 291 for k, constraints in dependencies.items(): 292 if len(constraints) == 1: 293 dependencies[k] = constraints[0] 294 295 data["dependencies"] = dependencies 296 297 if package.source_type: 298 data["source"] = { 299 "type": package.source_type, 300 "url": package.source_url, 301 "reference": package.source_reference, 302 } 303 304 return data 305 [end of poetry/packages/locker.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/poetry/packages/locker.py b/poetry/packages/locker.py --- a/poetry/packages/locker.py +++ b/poetry/packages/locker.py @@ -139,7 +139,7 @@ package.add_dependency(dep_name, constraint) if "source" in info: - package.source_type = info["source"]["type"] + package.source_type = info["source"].get("type", "") package.source_url = info["source"]["url"] package.source_reference = info["source"]["reference"] @@ -294,11 +294,12 @@ data["dependencies"] = dependencies - if package.source_type: + if package.source_url: data["source"] = { - "type": package.source_type, "url": package.source_url, "reference": package.source_reference, } + if package.source_type: + data["source"]["type"] = package.source_type return data
{"golden_diff": "diff --git a/poetry/packages/locker.py b/poetry/packages/locker.py\n--- a/poetry/packages/locker.py\n+++ b/poetry/packages/locker.py\n@@ -139,7 +139,7 @@\n package.add_dependency(dep_name, constraint)\n \n if \"source\" in info:\n- package.source_type = info[\"source\"][\"type\"]\n+ package.source_type = info[\"source\"].get(\"type\", \"\")\n package.source_url = info[\"source\"][\"url\"]\n package.source_reference = info[\"source\"][\"reference\"]\n \n@@ -294,11 +294,12 @@\n \n data[\"dependencies\"] = dependencies\n \n- if package.source_type:\n+ if package.source_url:\n data[\"source\"] = {\n- \"type\": package.source_type,\n \"url\": package.source_url,\n \"reference\": package.source_reference,\n }\n+ if package.source_type:\n+ data[\"source\"][\"type\"] = package.source_type\n \n return data\n", "issue": "1.0.0b5 removes my private source from the lockfile\n<!--\r\n Hi there! Thank you for discovering and submitting an issue.\r\n\r\n Before you submit this; let's make sure of a few things.\r\n Please make sure the following boxes are ticked if they are correct.\r\n If not, please try and fulfill these first.\r\n-->\r\n\r\n<!-- Checked checkbox should look like this: [x] -->\r\n- [x] I am on the [latest](https://github.com/sdispater/poetry/releases/latest) Poetry version.\r\n- [x] I have searched the [issues](https://github.com/sdispater/poetry/issues) of this repo and believe that this is not a duplicate.\r\n- [x] If an exception occurs when executing a command, I executed it again in debug mode (`-vvv` option).\r\n\r\n<!--\r\n Once those are done, if you're able to fill in the following list with your information,\r\n it'd be very helpful to whoever handles the issue.\r\n-->\r\n\r\n- **OS version and name**: macOS 10.14.6 (18G103)\r\n- **Poetry version**: <!-- 1.0.0b5\r\n- **Link of a [Gist](https://gist.github.com/) with the contents of your pyproject.toml file**: <!-- Gist Link Here --> https://gist.github.com/JBKahn/4bbb040c0181d974e0b71407ad764b6c\r\n\r\n\r\n## Issue\r\nWhen moving from `1.0.0b4` to `1.0.0b5` `poetry lock` is removing the \r\n```\r\n[package.source]\r\nreference = \"private\"\r\ntype = \"legacy\"\r\nurl = \"https://private.url\"\r\n```\r\n\n", "before_files": [{"content": "import json\nimport re\n\nfrom hashlib import sha256\nfrom typing import List\n\nfrom tomlkit import document\nfrom tomlkit import inline_table\nfrom tomlkit import item\nfrom tomlkit import table\nfrom tomlkit.exceptions import TOMLKitError\n\nimport poetry.packages\nimport poetry.repositories\n\nfrom poetry.utils._compat import Path\nfrom poetry.utils.toml_file import TomlFile\nfrom poetry.version.markers import parse_marker\n\n\nclass Locker(object):\n\n _relevant_keys = [\"dependencies\", \"dev-dependencies\", \"source\", \"extras\"]\n\n def __init__(self, lock, local_config): # type: (Path, dict) -> None\n self._lock = TomlFile(lock)\n self._local_config = local_config\n self._lock_data = None\n self._content_hash = self._get_content_hash()\n\n @property\n def lock(self): # type: () -> TomlFile\n return self._lock\n\n @property\n def lock_data(self):\n if self._lock_data is None:\n self._lock_data = self._get_lock_data()\n\n return self._lock_data\n\n def is_locked(self): # type: () -> bool\n \"\"\"\n Checks whether the locker has been locked (lockfile found).\n \"\"\"\n if not self._lock.exists():\n return False\n\n return \"package\" in self.lock_data\n\n def is_fresh(self): # type: () -> bool\n \"\"\"\n Checks whether the lock file is still up to date with the current hash.\n \"\"\"\n lock = self._lock.read()\n metadata = lock.get(\"metadata\", {})\n\n if \"content-hash\" in metadata:\n return self._content_hash == lock[\"metadata\"][\"content-hash\"]\n\n return False\n\n def locked_repository(\n self, with_dev_reqs=False\n ): # type: (bool) -> poetry.repositories.Repository\n \"\"\"\n Searches and returns a repository of locked packages.\n \"\"\"\n if not self.is_locked():\n return poetry.repositories.Repository()\n\n lock_data = self.lock_data\n packages = poetry.repositories.Repository()\n\n if with_dev_reqs:\n locked_packages = lock_data[\"package\"]\n else:\n locked_packages = [\n p for p in lock_data[\"package\"] if p[\"category\"] == \"main\"\n ]\n\n if not locked_packages:\n return packages\n\n for info in locked_packages:\n package = poetry.packages.Package(\n info[\"name\"], info[\"version\"], info[\"version\"]\n )\n package.description = info.get(\"description\", \"\")\n package.category = info[\"category\"]\n package.optional = info[\"optional\"]\n if \"hashes\" in lock_data[\"metadata\"]:\n # Old lock so we create dummy files from the hashes\n package.files = [\n {\"name\": h, \"hash\": h}\n for h in lock_data[\"metadata\"][\"hashes\"][info[\"name\"]]\n ]\n else:\n package.files = lock_data[\"metadata\"][\"files\"][info[\"name\"]]\n\n package.python_versions = info[\"python-versions\"]\n extras = info.get(\"extras\", {})\n if extras:\n for name, deps in extras.items():\n package.extras[name] = []\n\n for dep in deps:\n m = re.match(r\"^(.+?)(?:\\s+\\((.+)\\))?$\", dep)\n dep_name = m.group(1)\n constraint = m.group(2) or \"*\"\n\n package.extras[name].append(\n poetry.packages.Dependency(dep_name, constraint)\n )\n\n if \"marker\" in info:\n package.marker = parse_marker(info[\"marker\"])\n else:\n # Compatibility for old locks\n if \"requirements\" in info:\n dep = poetry.packages.Dependency(\"foo\", \"0.0.0\")\n for name, value in info[\"requirements\"].items():\n if name == \"python\":\n dep.python_versions = value\n elif name == \"platform\":\n dep.platform = value\n\n split_dep = dep.to_pep_508(False).split(\";\")\n if len(split_dep) > 1:\n package.marker = parse_marker(split_dep[1].strip())\n\n for dep_name, constraint in info.get(\"dependencies\", {}).items():\n if isinstance(constraint, list):\n for c in constraint:\n package.add_dependency(dep_name, c)\n\n continue\n\n package.add_dependency(dep_name, constraint)\n\n if \"source\" in info:\n package.source_type = info[\"source\"][\"type\"]\n package.source_url = info[\"source\"][\"url\"]\n package.source_reference = info[\"source\"][\"reference\"]\n\n packages.add_package(package)\n\n return packages\n\n def set_lock_data(self, root, packages): # type: (...) -> bool\n files = table()\n packages = self._lock_packages(packages)\n # Retrieving hashes\n for package in packages:\n if package[\"name\"] not in files:\n files[package[\"name\"]] = []\n\n for f in package[\"files\"]:\n file_metadata = inline_table()\n for k, v in sorted(f.items()):\n file_metadata[k] = v\n\n files[package[\"name\"]].append(file_metadata)\n\n if files[package[\"name\"]]:\n files[package[\"name\"]] = item(files[package[\"name\"]]).multiline(True)\n\n del package[\"files\"]\n\n lock = document()\n lock[\"package\"] = packages\n\n if root.extras:\n lock[\"extras\"] = {\n extra: [dep.pretty_name for dep in deps]\n for extra, deps in root.extras.items()\n }\n\n lock[\"metadata\"] = {\n \"python-versions\": root.python_versions,\n \"content-hash\": self._content_hash,\n \"files\": files,\n }\n\n if not self.is_locked() or lock != self.lock_data:\n self._write_lock_data(lock)\n\n return True\n\n return False\n\n def _write_lock_data(self, data):\n self.lock.write(data)\n\n # Checking lock file data consistency\n if data != self.lock.read():\n raise RuntimeError(\"Inconsistent lock file data.\")\n\n self._lock_data = None\n\n def _get_content_hash(self): # type: () -> str\n \"\"\"\n Returns the sha256 hash of the sorted content of the pyproject file.\n \"\"\"\n content = self._local_config\n\n relevant_content = {}\n for key in self._relevant_keys:\n relevant_content[key] = content.get(key)\n\n content_hash = sha256(\n json.dumps(relevant_content, sort_keys=True).encode()\n ).hexdigest()\n\n return content_hash\n\n def _get_lock_data(self): # type: () -> dict\n if not self._lock.exists():\n raise RuntimeError(\"No lockfile found. Unable to read locked packages\")\n\n try:\n return self._lock.read()\n except TOMLKitError as e:\n raise RuntimeError(\"Unable to read the lock file ({}).\".format(e))\n\n def _lock_packages(\n self, packages\n ): # type: (List['poetry.packages.Package']) -> list\n locked = []\n\n for package in sorted(packages, key=lambda x: x.name):\n spec = self._dump_package(package)\n\n locked.append(spec)\n\n return locked\n\n def _dump_package(self, package): # type: (poetry.packages.Package) -> dict\n dependencies = {}\n for dependency in sorted(package.requires, key=lambda d: d.name):\n if dependency.is_optional() and not dependency.is_activated():\n continue\n\n if dependency.pretty_name not in dependencies:\n dependencies[dependency.pretty_name] = []\n\n constraint = {\"version\": str(dependency.pretty_constraint)}\n\n if dependency.extras:\n constraint[\"extras\"] = dependency.extras\n\n if dependency.is_optional():\n constraint[\"optional\"] = True\n\n if not dependency.python_constraint.is_any():\n constraint[\"python\"] = str(dependency.python_constraint)\n\n dependencies[dependency.pretty_name].append(constraint)\n\n # All the constraints should have the same type,\n # but we want to simplify them if it's possible\n for dependency, constraints in tuple(dependencies.items()):\n if all(len(constraint) == 1 for constraint in constraints):\n dependencies[dependency] = [\n constraint[\"version\"] for constraint in constraints\n ]\n\n data = {\n \"name\": package.pretty_name,\n \"version\": package.pretty_version,\n \"description\": package.description or \"\",\n \"category\": package.category,\n \"optional\": package.optional,\n \"python-versions\": package.python_versions,\n \"files\": sorted(package.files, key=lambda x: x[\"file\"]),\n }\n if not package.marker.is_any():\n data[\"marker\"] = str(package.marker)\n\n if package.extras:\n extras = {}\n for name, deps in package.extras.items():\n extras[name] = [\n str(dep) if not dep.constraint.is_any() else dep.name\n for dep in deps\n ]\n\n data[\"extras\"] = extras\n\n if dependencies:\n for k, constraints in dependencies.items():\n if len(constraints) == 1:\n dependencies[k] = constraints[0]\n\n data[\"dependencies\"] = dependencies\n\n if package.source_type:\n data[\"source\"] = {\n \"type\": package.source_type,\n \"url\": package.source_url,\n \"reference\": package.source_reference,\n }\n\n return data\n", "path": "poetry/packages/locker.py"}]}
3,775
219
gh_patches_debug_42509
rasdani/github-patches
git_diff
napari__napari-1749
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> Track vertex shader does not handle blending/opacity correctly ## 🐛 Bug Two somewhat related bugs in `_vispy_tracks_shader.py` in the `Tracks` layer: 1. The opacity slider only changes the vispy node opacity, which has no effect on the vertex shader used for rendering the tracks or graph. 2. The translucent mode renders parts of the track, which should not be rendered, as dark lines: ![bug](https://user-images.githubusercontent.com/8217795/96697072-64ef9b00-1383-11eb-9648-54c92319ae84.gif) ## To Reproduce Steps to reproduce the behavior: 1. run `examples\tracks_3d.py` 2. Change the blending mode to translucent in the `Tracks` layer 3. Change the opacity slider ## Expected behavior `opacity` should change the track opacity, and the `translucent` blending should not render parts of the track which should not be visible. ## Environment napari: 0.3.9.dev32+g911e095.d20201007 Platform: Linux-4.15.0-1100-oem-x86_64-with-debian-buster-sid Python: 3.7.7 (default, May 7 2020, 21:25:33) [GCC 7.3.0] Qt: 5.14.2 PySide2: 5.14.2.2 NumPy: 1.19.1 SciPy: 1.4.1 Dask: 2.17.2 VisPy: 0.6.4 GL version: 4.6.0 NVIDIA 450.80.02 MAX_TEXTURE_SIZE: 32768 Plugins: - napari-plugin-engine: 0.1.5 - svg: 0.1.3 ## Additional context This relates to the fact that the shader manipulates the alpha values of the vertices directly, late in the rendering pipeline. </issue> <code> [start of napari/_vispy/vispy_tracks_layer.py] 1 from vispy.scene.visuals import Compound, Line, Text 2 3 from ._vispy_tracks_shader import TrackShader 4 from .vispy_base_layer import VispyBaseLayer 5 6 7 class VispyTracksLayer(VispyBaseLayer): 8 """VispyTracksLayer 9 10 Track layer for visualizing tracks. 11 12 Components: 13 - Track lines (vispy.LineVisual) 14 - Track IDs (vispy.TextVisual) 15 - Graph edges (vispy.LineVisual) 16 17 """ 18 19 def __init__(self, layer): 20 node = Compound([Line(), Text(), Line()]) 21 super().__init__(layer, node) 22 23 self.layer.events.tail_width.connect(self._on_appearance_change) 24 self.layer.events.tail_length.connect(self._on_appearance_change) 25 self.layer.events.display_id.connect(self._on_appearance_change) 26 self.layer.events.display_tail.connect(self._on_appearance_change) 27 self.layer.events.display_graph.connect(self._on_appearance_change) 28 29 self.layer.events.color_by.connect(self._on_appearance_change) 30 self.layer.events.colormap.connect(self._on_appearance_change) 31 32 # these events are fired when changes occur to the tracks or the 33 # graph - as the vertex buffer of the shader needs to be updated 34 # alongside the actual vertex data 35 self.layer.events.rebuild_tracks.connect(self._on_tracks_change) 36 self.layer.events.rebuild_graph.connect(self._on_graph_change) 37 38 # build and attach the shader to the track 39 self.track_shader = TrackShader() 40 self.graph_shader = TrackShader() 41 node._subvisuals[0].attach(self.track_shader) 42 node._subvisuals[2].attach(self.graph_shader) 43 44 # text label properties 45 self.node._subvisuals[1].color = 'white' 46 self.node._subvisuals[1].font_size = 8 47 48 self._reset_base() 49 50 self._on_data_change() 51 self._on_appearance_change() 52 53 def _on_data_change(self, event=None): 54 """ update the display """ 55 56 # update the shaders 57 self.track_shader.current_time = self.layer.current_time 58 self.graph_shader.current_time = self.layer.current_time 59 60 # add text labels if they're visible 61 if self.node._subvisuals[1].visible: 62 labels_text, labels_pos = self.layer.track_labels 63 self.node._subvisuals[1].text = labels_text 64 self.node._subvisuals[1].pos = labels_pos 65 66 self.node.update() 67 # Call to update order of translation values with new dims: 68 self._on_matrix_change() 69 70 def _on_appearance_change(self, event=None): 71 """ change the appearance of the data """ 72 73 # update shader properties related to appearance 74 self.track_shader.use_fade = self.layer.use_fade 75 self.track_shader.tail_length = self.layer.tail_length 76 self.graph_shader.use_fade = self.layer.use_fade 77 self.graph_shader.tail_length = self.layer.tail_length 78 79 # set visibility of subvisuals 80 self.node._subvisuals[0].visible = self.layer.display_tail 81 self.node._subvisuals[1].visible = self.layer.display_id 82 self.node._subvisuals[2].visible = self.layer.display_graph 83 84 # set the width of the track tails 85 self.node._subvisuals[0].set_data( 86 width=self.layer.tail_width, color=self.layer.track_colors, 87 ) 88 self.node._subvisuals[2].set_data(width=self.layer.tail_width,) 89 90 def _on_tracks_change(self, event=None): 91 """ update the shader when the track data changes """ 92 93 self.track_shader.use_fade = self.layer.use_fade 94 self.track_shader.tail_length = self.layer.tail_length 95 self.track_shader.vertex_time = self.layer.track_times 96 97 # change the data to the vispy line visual 98 self.node._subvisuals[0].set_data( 99 pos=self.layer._view_data, 100 connect=self.layer.track_connex, 101 width=self.layer.tail_width, 102 color=self.layer.track_colors, 103 ) 104 105 # Call to update order of translation values with new dims: 106 self._on_matrix_change() 107 108 def _on_graph_change(self, event=None): 109 """ update the shader when the graph data changes """ 110 111 self.graph_shader.use_fade = self.layer.use_fade 112 self.graph_shader.tail_length = self.layer.tail_length 113 self.graph_shader.vertex_time = self.layer.graph_times 114 115 # if the user clears a graph after it has been created, vispy offers 116 # no method to clear the data, therefore, we need to set private 117 # attributes to None to prevent errors 118 if self.layer._view_graph is None: 119 self.node._subvisuals[2]._pos = None 120 self.node._subvisuals[2]._connect = None 121 self.node.update() 122 return 123 124 self.node._subvisuals[2].set_data( 125 pos=self.layer._view_graph, 126 connect=self.layer.graph_connex, 127 width=self.layer.tail_width, 128 color='white', 129 ) 130 131 # Call to update order of translation values with new dims: 132 self._on_matrix_change() 133 [end of napari/_vispy/vispy_tracks_layer.py] [start of napari/_vispy/_vispy_tracks_shader.py] 1 from typing import List, Union 2 3 import numpy as np 4 from vispy.gloo import VertexBuffer 5 from vispy.visuals.filters.base_filter import Filter 6 7 8 class TrackShader(Filter): 9 """ TrackShader 10 11 Custom vertex and fragment shaders for visualizing tracks quickly with 12 vispy. The central assumption is that the tracks are rendered as a 13 continuous vispy Line segment, with connections and colors defined when 14 the visual is created. 15 16 The shader simply changes the visibility and/or fading of the data according 17 to the current_time and the associate time metadata for each vertex. This 18 is scaled according to the tail length. Points ahead of the current time 19 are rendered with alpha set to zero. 20 21 Parameters 22 ---------- 23 current_time : int, float 24 the current time, which is typically the frame index, although this 25 can be an arbitrary float 26 tail_length : int, float 27 the upper limit on length of the 'tail' 28 use_fade : bool 29 this will enable/disable tail fading with time 30 vertex_time : 1D array, list 31 a vector describing the time associated with each vertex 32 33 34 TODO 35 ---- 36 - the track is still displayed, albeit with fading, once the track has 37 finished but is still within the 'tail_length' window. Should it 38 disappear? 39 - check the shader positioning within the GL pipeline, currently 40 overrides layer opacity settings 41 42 vertex_mask: 1D array, list 43 a vector describing whether to mask each vertex 44 45 """ 46 47 VERT_SHADER = """ 48 varying vec4 v_track_color; 49 void apply_track_shading() { 50 51 float alpha; 52 53 if ($a_vertex_time > $current_time) { 54 // this is a hack to minimize the frag shader rendering ahead 55 // of the current time point due to interpolation 56 if ($a_vertex_time <= $current_time + 1){ 57 alpha = -100.; 58 } else { 59 alpha = 0.; 60 } 61 } else { 62 // fade the track into the temporal distance, scaled by the 63 // maximum tail length from the gui 64 float fade = ($current_time - $a_vertex_time) / $tail_length; 65 alpha = clamp(1.0-fade, 0.0, 1.0); 66 } 67 68 // when use_fade is disabled, the entire track is visible 69 if ($use_fade == 0) { 70 alpha = 1.0; 71 } 72 73 // set the vertex alpha according to the fade 74 v_track_color.a = alpha; 75 } 76 """ 77 78 FRAG_SHADER = """ 79 varying vec4 v_track_color; 80 void apply_track_shading() { 81 // interpolate 82 gl_FragColor.a = clamp(v_track_color.a, 0.0, 1.0); 83 } 84 """ 85 86 def __init__( 87 self, 88 current_time=0, 89 tail_length=30, 90 use_fade: bool = True, 91 vertex_time: Union[List, np.ndarray] = None, 92 ): 93 94 super().__init__( 95 vcode=self.VERT_SHADER, vpos=3, fcode=self.FRAG_SHADER, fpos=9 96 ) 97 98 self.current_time = current_time 99 self.tail_length = tail_length 100 self.use_fade = use_fade 101 self.vertex_time = vertex_time 102 103 @property 104 def current_time(self) -> Union[int, float]: 105 return self._current_time 106 107 @current_time.setter 108 def current_time(self, n: Union[int, float]): 109 self._current_time = n 110 if isinstance(n, slice): 111 n = np.max(self._vertex_time) 112 self.vshader['current_time'] = float(n) 113 114 @property 115 def use_fade(self) -> bool: 116 return self._use_fade 117 118 @use_fade.setter 119 def use_fade(self, value: bool): 120 self._use_fade = value 121 self.vshader['use_fade'] = float(self._use_fade) 122 123 @property 124 def tail_length(self) -> Union[int, float]: 125 return self._tail_length 126 127 @tail_length.setter 128 def tail_length(self, tail_length: Union[int, float]): 129 self._tail_length = tail_length 130 self.vshader['tail_length'] = float(self._tail_length) 131 132 def _attach(self, visual): 133 super()._attach(visual) 134 135 @property 136 def vertex_time(self): 137 return self._vertex_time 138 139 @vertex_time.setter 140 def vertex_time(self, v_time): 141 self._vertex_time = np.array(v_time).reshape(-1, 1).astype(np.float32) 142 self.vshader['a_vertex_time'] = VertexBuffer(self.vertex_time) 143 [end of napari/_vispy/_vispy_tracks_shader.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/napari/_vispy/_vispy_tracks_shader.py b/napari/_vispy/_vispy_tracks_shader.py --- a/napari/_vispy/_vispy_tracks_shader.py +++ b/napari/_vispy/_vispy_tracks_shader.py @@ -6,7 +6,7 @@ class TrackShader(Filter): - """ TrackShader + """TrackShader. Custom vertex and fragment shaders for visualizing tracks quickly with vispy. The central assumption is that the tracks are rendered as a @@ -30,17 +30,11 @@ vertex_time : 1D array, list a vector describing the time associated with each vertex - TODO ---- - the track is still displayed, albeit with fading, once the track has finished but is still within the 'tail_length' window. Should it disappear? - - check the shader positioning within the GL pipeline, currently - overrides layer opacity settings - - vertex_mask: 1D array, list - a vector describing whether to mask each vertex """ @@ -78,15 +72,21 @@ FRAG_SHADER = """ varying vec4 v_track_color; void apply_track_shading() { + + // if the alpha is below the threshold, discard the fragment + if( v_track_color.a <= 0.0 ) { + discard; + } + // interpolate - gl_FragColor.a = clamp(v_track_color.a, 0.0, 1.0); + gl_FragColor.a = clamp(v_track_color.a * gl_FragColor.a, 0.0, 1.0); } """ def __init__( self, - current_time=0, - tail_length=30, + current_time: Union[int, float] = 0, + tail_length: Union[int, float] = 30, use_fade: bool = True, vertex_time: Union[List, np.ndarray] = None, ): diff --git a/napari/_vispy/vispy_tracks_layer.py b/napari/_vispy/vispy_tracks_layer.py --- a/napari/_vispy/vispy_tracks_layer.py +++ b/napari/_vispy/vispy_tracks_layer.py @@ -51,7 +51,7 @@ self._on_appearance_change() def _on_data_change(self, event=None): - """ update the display """ + """Update the display.""" # update the shaders self.track_shader.current_time = self.layer.current_time @@ -68,7 +68,7 @@ self._on_matrix_change() def _on_appearance_change(self, event=None): - """ change the appearance of the data """ + """Change the appearance of the data.""" # update shader properties related to appearance self.track_shader.use_fade = self.layer.use_fade @@ -88,7 +88,7 @@ self.node._subvisuals[2].set_data(width=self.layer.tail_width,) def _on_tracks_change(self, event=None): - """ update the shader when the track data changes """ + """Update the shader when the track data changes.""" self.track_shader.use_fade = self.layer.use_fade self.track_shader.tail_length = self.layer.tail_length @@ -106,7 +106,7 @@ self._on_matrix_change() def _on_graph_change(self, event=None): - """ update the shader when the graph data changes """ + """Update the shader when the graph data changes.""" self.graph_shader.use_fade = self.layer.use_fade self.graph_shader.tail_length = self.layer.tail_length
{"golden_diff": "diff --git a/napari/_vispy/_vispy_tracks_shader.py b/napari/_vispy/_vispy_tracks_shader.py\n--- a/napari/_vispy/_vispy_tracks_shader.py\n+++ b/napari/_vispy/_vispy_tracks_shader.py\n@@ -6,7 +6,7 @@\n \n \n class TrackShader(Filter):\n- \"\"\" TrackShader\n+ \"\"\"TrackShader.\n \n Custom vertex and fragment shaders for visualizing tracks quickly with\n vispy. The central assumption is that the tracks are rendered as a\n@@ -30,17 +30,11 @@\n vertex_time : 1D array, list\n a vector describing the time associated with each vertex\n \n-\n TODO\n ----\n - the track is still displayed, albeit with fading, once the track has\n finished but is still within the 'tail_length' window. Should it\n disappear?\n- - check the shader positioning within the GL pipeline, currently\n- overrides layer opacity settings\n-\n- vertex_mask: 1D array, list\n- a vector describing whether to mask each vertex\n \n \"\"\"\n \n@@ -78,15 +72,21 @@\n FRAG_SHADER = \"\"\"\n varying vec4 v_track_color;\n void apply_track_shading() {\n+\n+ // if the alpha is below the threshold, discard the fragment\n+ if( v_track_color.a <= 0.0 ) {\n+ discard;\n+ }\n+\n // interpolate\n- gl_FragColor.a = clamp(v_track_color.a, 0.0, 1.0);\n+ gl_FragColor.a = clamp(v_track_color.a * gl_FragColor.a, 0.0, 1.0);\n }\n \"\"\"\n \n def __init__(\n self,\n- current_time=0,\n- tail_length=30,\n+ current_time: Union[int, float] = 0,\n+ tail_length: Union[int, float] = 30,\n use_fade: bool = True,\n vertex_time: Union[List, np.ndarray] = None,\n ):\ndiff --git a/napari/_vispy/vispy_tracks_layer.py b/napari/_vispy/vispy_tracks_layer.py\n--- a/napari/_vispy/vispy_tracks_layer.py\n+++ b/napari/_vispy/vispy_tracks_layer.py\n@@ -51,7 +51,7 @@\n self._on_appearance_change()\n \n def _on_data_change(self, event=None):\n- \"\"\" update the display \"\"\"\n+ \"\"\"Update the display.\"\"\"\n \n # update the shaders\n self.track_shader.current_time = self.layer.current_time\n@@ -68,7 +68,7 @@\n self._on_matrix_change()\n \n def _on_appearance_change(self, event=None):\n- \"\"\" change the appearance of the data \"\"\"\n+ \"\"\"Change the appearance of the data.\"\"\"\n \n # update shader properties related to appearance\n self.track_shader.use_fade = self.layer.use_fade\n@@ -88,7 +88,7 @@\n self.node._subvisuals[2].set_data(width=self.layer.tail_width,)\n \n def _on_tracks_change(self, event=None):\n- \"\"\" update the shader when the track data changes \"\"\"\n+ \"\"\"Update the shader when the track data changes.\"\"\"\n \n self.track_shader.use_fade = self.layer.use_fade\n self.track_shader.tail_length = self.layer.tail_length\n@@ -106,7 +106,7 @@\n self._on_matrix_change()\n \n def _on_graph_change(self, event=None):\n- \"\"\" update the shader when the graph data changes \"\"\"\n+ \"\"\"Update the shader when the graph data changes.\"\"\"\n \n self.graph_shader.use_fade = self.layer.use_fade\n self.graph_shader.tail_length = self.layer.tail_length\n", "issue": "Track vertex shader does not handle blending/opacity correctly\n## \ud83d\udc1b Bug\r\n\r\nTwo somewhat related bugs in `_vispy_tracks_shader.py` in the `Tracks` layer:\r\n1. The opacity slider only changes the vispy node opacity, which has no effect on the vertex shader used for rendering the tracks or graph.\r\n2. The translucent mode renders parts of the track, which should not be rendered, as dark lines:\r\n\r\n![bug](https://user-images.githubusercontent.com/8217795/96697072-64ef9b00-1383-11eb-9648-54c92319ae84.gif)\r\n\r\n## To Reproduce\r\n\r\nSteps to reproduce the behavior:\r\n\r\n1. run `examples\\tracks_3d.py`\r\n2. Change the blending mode to translucent in the `Tracks` layer\r\n3. Change the opacity slider \r\n\r\n## Expected behavior\r\n\r\n`opacity` should change the track opacity, and the `translucent` blending should not render parts of the track which should not be visible.\r\n\r\n## Environment\r\n\r\nnapari: 0.3.9.dev32+g911e095.d20201007\r\nPlatform: Linux-4.15.0-1100-oem-x86_64-with-debian-buster-sid\r\nPython: 3.7.7 (default, May 7 2020, 21:25:33) [GCC 7.3.0]\r\nQt: 5.14.2\r\nPySide2: 5.14.2.2\r\nNumPy: 1.19.1\r\nSciPy: 1.4.1\r\nDask: 2.17.2\r\nVisPy: 0.6.4\r\n\r\nGL version: 4.6.0 NVIDIA 450.80.02\r\nMAX_TEXTURE_SIZE: 32768\r\n\r\nPlugins:\r\n- napari-plugin-engine: 0.1.5\r\n- svg: 0.1.3\r\n\r\n## Additional context\r\nThis relates to the fact that the shader manipulates the alpha values of the vertices directly, late in the rendering pipeline.\n", "before_files": [{"content": "from vispy.scene.visuals import Compound, Line, Text\n\nfrom ._vispy_tracks_shader import TrackShader\nfrom .vispy_base_layer import VispyBaseLayer\n\n\nclass VispyTracksLayer(VispyBaseLayer):\n \"\"\"VispyTracksLayer\n\n Track layer for visualizing tracks.\n\n Components:\n - Track lines (vispy.LineVisual)\n - Track IDs (vispy.TextVisual)\n - Graph edges (vispy.LineVisual)\n\n \"\"\"\n\n def __init__(self, layer):\n node = Compound([Line(), Text(), Line()])\n super().__init__(layer, node)\n\n self.layer.events.tail_width.connect(self._on_appearance_change)\n self.layer.events.tail_length.connect(self._on_appearance_change)\n self.layer.events.display_id.connect(self._on_appearance_change)\n self.layer.events.display_tail.connect(self._on_appearance_change)\n self.layer.events.display_graph.connect(self._on_appearance_change)\n\n self.layer.events.color_by.connect(self._on_appearance_change)\n self.layer.events.colormap.connect(self._on_appearance_change)\n\n # these events are fired when changes occur to the tracks or the\n # graph - as the vertex buffer of the shader needs to be updated\n # alongside the actual vertex data\n self.layer.events.rebuild_tracks.connect(self._on_tracks_change)\n self.layer.events.rebuild_graph.connect(self._on_graph_change)\n\n # build and attach the shader to the track\n self.track_shader = TrackShader()\n self.graph_shader = TrackShader()\n node._subvisuals[0].attach(self.track_shader)\n node._subvisuals[2].attach(self.graph_shader)\n\n # text label properties\n self.node._subvisuals[1].color = 'white'\n self.node._subvisuals[1].font_size = 8\n\n self._reset_base()\n\n self._on_data_change()\n self._on_appearance_change()\n\n def _on_data_change(self, event=None):\n \"\"\" update the display \"\"\"\n\n # update the shaders\n self.track_shader.current_time = self.layer.current_time\n self.graph_shader.current_time = self.layer.current_time\n\n # add text labels if they're visible\n if self.node._subvisuals[1].visible:\n labels_text, labels_pos = self.layer.track_labels\n self.node._subvisuals[1].text = labels_text\n self.node._subvisuals[1].pos = labels_pos\n\n self.node.update()\n # Call to update order of translation values with new dims:\n self._on_matrix_change()\n\n def _on_appearance_change(self, event=None):\n \"\"\" change the appearance of the data \"\"\"\n\n # update shader properties related to appearance\n self.track_shader.use_fade = self.layer.use_fade\n self.track_shader.tail_length = self.layer.tail_length\n self.graph_shader.use_fade = self.layer.use_fade\n self.graph_shader.tail_length = self.layer.tail_length\n\n # set visibility of subvisuals\n self.node._subvisuals[0].visible = self.layer.display_tail\n self.node._subvisuals[1].visible = self.layer.display_id\n self.node._subvisuals[2].visible = self.layer.display_graph\n\n # set the width of the track tails\n self.node._subvisuals[0].set_data(\n width=self.layer.tail_width, color=self.layer.track_colors,\n )\n self.node._subvisuals[2].set_data(width=self.layer.tail_width,)\n\n def _on_tracks_change(self, event=None):\n \"\"\" update the shader when the track data changes \"\"\"\n\n self.track_shader.use_fade = self.layer.use_fade\n self.track_shader.tail_length = self.layer.tail_length\n self.track_shader.vertex_time = self.layer.track_times\n\n # change the data to the vispy line visual\n self.node._subvisuals[0].set_data(\n pos=self.layer._view_data,\n connect=self.layer.track_connex,\n width=self.layer.tail_width,\n color=self.layer.track_colors,\n )\n\n # Call to update order of translation values with new dims:\n self._on_matrix_change()\n\n def _on_graph_change(self, event=None):\n \"\"\" update the shader when the graph data changes \"\"\"\n\n self.graph_shader.use_fade = self.layer.use_fade\n self.graph_shader.tail_length = self.layer.tail_length\n self.graph_shader.vertex_time = self.layer.graph_times\n\n # if the user clears a graph after it has been created, vispy offers\n # no method to clear the data, therefore, we need to set private\n # attributes to None to prevent errors\n if self.layer._view_graph is None:\n self.node._subvisuals[2]._pos = None\n self.node._subvisuals[2]._connect = None\n self.node.update()\n return\n\n self.node._subvisuals[2].set_data(\n pos=self.layer._view_graph,\n connect=self.layer.graph_connex,\n width=self.layer.tail_width,\n color='white',\n )\n\n # Call to update order of translation values with new dims:\n self._on_matrix_change()\n", "path": "napari/_vispy/vispy_tracks_layer.py"}, {"content": "from typing import List, Union\n\nimport numpy as np\nfrom vispy.gloo import VertexBuffer\nfrom vispy.visuals.filters.base_filter import Filter\n\n\nclass TrackShader(Filter):\n \"\"\" TrackShader\n\n Custom vertex and fragment shaders for visualizing tracks quickly with\n vispy. The central assumption is that the tracks are rendered as a\n continuous vispy Line segment, with connections and colors defined when\n the visual is created.\n\n The shader simply changes the visibility and/or fading of the data according\n to the current_time and the associate time metadata for each vertex. This\n is scaled according to the tail length. Points ahead of the current time\n are rendered with alpha set to zero.\n\n Parameters\n ----------\n current_time : int, float\n the current time, which is typically the frame index, although this\n can be an arbitrary float\n tail_length : int, float\n the upper limit on length of the 'tail'\n use_fade : bool\n this will enable/disable tail fading with time\n vertex_time : 1D array, list\n a vector describing the time associated with each vertex\n\n\n TODO\n ----\n - the track is still displayed, albeit with fading, once the track has\n finished but is still within the 'tail_length' window. Should it\n disappear?\n - check the shader positioning within the GL pipeline, currently\n overrides layer opacity settings\n\n vertex_mask: 1D array, list\n a vector describing whether to mask each vertex\n\n \"\"\"\n\n VERT_SHADER = \"\"\"\n varying vec4 v_track_color;\n void apply_track_shading() {\n\n float alpha;\n\n if ($a_vertex_time > $current_time) {\n // this is a hack to minimize the frag shader rendering ahead\n // of the current time point due to interpolation\n if ($a_vertex_time <= $current_time + 1){\n alpha = -100.;\n } else {\n alpha = 0.;\n }\n } else {\n // fade the track into the temporal distance, scaled by the\n // maximum tail length from the gui\n float fade = ($current_time - $a_vertex_time) / $tail_length;\n alpha = clamp(1.0-fade, 0.0, 1.0);\n }\n\n // when use_fade is disabled, the entire track is visible\n if ($use_fade == 0) {\n alpha = 1.0;\n }\n\n // set the vertex alpha according to the fade\n v_track_color.a = alpha;\n }\n \"\"\"\n\n FRAG_SHADER = \"\"\"\n varying vec4 v_track_color;\n void apply_track_shading() {\n // interpolate\n gl_FragColor.a = clamp(v_track_color.a, 0.0, 1.0);\n }\n \"\"\"\n\n def __init__(\n self,\n current_time=0,\n tail_length=30,\n use_fade: bool = True,\n vertex_time: Union[List, np.ndarray] = None,\n ):\n\n super().__init__(\n vcode=self.VERT_SHADER, vpos=3, fcode=self.FRAG_SHADER, fpos=9\n )\n\n self.current_time = current_time\n self.tail_length = tail_length\n self.use_fade = use_fade\n self.vertex_time = vertex_time\n\n @property\n def current_time(self) -> Union[int, float]:\n return self._current_time\n\n @current_time.setter\n def current_time(self, n: Union[int, float]):\n self._current_time = n\n if isinstance(n, slice):\n n = np.max(self._vertex_time)\n self.vshader['current_time'] = float(n)\n\n @property\n def use_fade(self) -> bool:\n return self._use_fade\n\n @use_fade.setter\n def use_fade(self, value: bool):\n self._use_fade = value\n self.vshader['use_fade'] = float(self._use_fade)\n\n @property\n def tail_length(self) -> Union[int, float]:\n return self._tail_length\n\n @tail_length.setter\n def tail_length(self, tail_length: Union[int, float]):\n self._tail_length = tail_length\n self.vshader['tail_length'] = float(self._tail_length)\n\n def _attach(self, visual):\n super()._attach(visual)\n\n @property\n def vertex_time(self):\n return self._vertex_time\n\n @vertex_time.setter\n def vertex_time(self, v_time):\n self._vertex_time = np.array(v_time).reshape(-1, 1).astype(np.float32)\n self.vshader['a_vertex_time'] = VertexBuffer(self.vertex_time)\n", "path": "napari/_vispy/_vispy_tracks_shader.py"}]}
3,827
837
gh_patches_debug_32652
rasdani/github-patches
git_diff
getnikola__nikola-1267
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> Unable to use Tidy on html Hi, I enabled the tidy filter on html file: from nikola import filters FILTERS = { ".html": [filters.tidy], } and now NIkola refuses to build my site: TaskError - taskid:render_tags:../categories/index.html PythonAction Error Traceback (most recent call last): File "/usr/lib64/python3.3/site-packages/doit/action.py", line 360, in execute returned_value = self.py_callable(_self.args, *_kwargs) File "/usr/lib64/python3.3/site-packages/nikola/utils.py", line 529, in unlessLink action(target) File "/usr/lib64/python3.3/site-packages/nikola/filters.py", line 144, in tidy for line in output.split("\n"): TypeError: Type str doesn't support the buffer AP Is there a problem with tidy ? Can I use another thing as a replacement ? Unable to use Tidy on html Hi, I enabled the tidy filter on html file: from nikola import filters FILTERS = { ".html": [filters.tidy], } and now NIkola refuses to build my site: TaskError - taskid:render_tags:../categories/index.html PythonAction Error Traceback (most recent call last): File "/usr/lib64/python3.3/site-packages/doit/action.py", line 360, in execute returned_value = self.py_callable(_self.args, *_kwargs) File "/usr/lib64/python3.3/site-packages/nikola/utils.py", line 529, in unlessLink action(target) File "/usr/lib64/python3.3/site-packages/nikola/filters.py", line 144, in tidy for line in output.split("\n"): TypeError: Type str doesn't support the buffer AP Is there a problem with tidy ? Can I use another thing as a replacement ? </issue> <code> [start of nikola/filters.py] 1 # -*- coding: utf-8 -*- 2 3 # Copyright © 2012-2014 Roberto Alsina and others. 4 5 # Permission is hereby granted, free of charge, to any 6 # person obtaining a copy of this software and associated 7 # documentation files (the "Software"), to deal in the 8 # Software without restriction, including without limitation 9 # the rights to use, copy, modify, merge, publish, 10 # distribute, sublicense, and/or sell copies of the 11 # Software, and to permit persons to whom the Software is 12 # furnished to do so, subject to the following conditions: 13 # 14 # The above copyright notice and this permission notice 15 # shall be included in all copies or substantial portions of 16 # the Software. 17 # 18 # THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY 19 # KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE 20 # WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR 21 # PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS 22 # OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR 23 # OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR 24 # OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE 25 # SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. 26 27 """Utility functions to help you run filters on files.""" 28 29 from .utils import req_missing 30 from functools import wraps 31 import os 32 import codecs 33 import re 34 import shutil 35 import subprocess 36 import tempfile 37 import shlex 38 39 try: 40 import typogrify.filters as typo 41 except ImportError: 42 typo = None # NOQA 43 44 45 def apply_to_binary_file(f): 46 """Take a function f that transforms a data argument, and returns 47 a function that takes a filename and applies f to the contents, 48 in place. Reads files in binary mode.""" 49 @wraps(f) 50 def f_in_file(fname): 51 with open(fname, 'rb') as inf: 52 data = inf.read() 53 data = f(data) 54 with open(fname, 'wb+') as outf: 55 outf.write(data) 56 57 return f_in_file 58 59 60 def apply_to_text_file(f): 61 """Take a function f that transforms a data argument, and returns 62 a function that takes a filename and applies f to the contents, 63 in place. Reads files in UTF-8.""" 64 @wraps(f) 65 def f_in_file(fname): 66 with codecs.open(fname, 'r', 'utf-8') as inf: 67 data = inf.read() 68 data = f(data) 69 with codecs.open(fname, 'w+', 'utf-8') as outf: 70 outf.write(data) 71 72 return f_in_file 73 74 75 def list_replace(the_list, find, replacement): 76 "Replace all occurrences of ``find`` with ``replacement`` in ``the_list``" 77 for i, v in enumerate(the_list): 78 if v == find: 79 the_list[i] = replacement 80 81 82 def runinplace(command, infile): 83 """Run a command in-place on a file. 84 85 command is a string of the form: "commandname %1 %2" and 86 it will be execed with infile as %1 and a temporary file 87 as %2. Then, that temporary file will be moved over %1. 88 89 Example usage: 90 91 runinplace("yui-compressor %1 -o %2", "myfile.css") 92 93 That will replace myfile.css with a minified version. 94 95 You can also supply command as a list. 96 """ 97 98 if not isinstance(command, list): 99 command = shlex.split(command) 100 101 tmpdir = None 102 103 if "%2" in command: 104 tmpdir = tempfile.mkdtemp(prefix="nikola") 105 tmpfname = os.path.join(tmpdir, os.path.basename(infile)) 106 107 try: 108 list_replace(command, "%1", infile) 109 if tmpdir: 110 list_replace(command, "%2", tmpfname) 111 112 subprocess.check_call(command) 113 114 if tmpdir: 115 shutil.move(tmpfname, infile) 116 finally: 117 if tmpdir: 118 shutil.rmtree(tmpdir) 119 120 121 def yui_compressor(infile): 122 yuicompressor = False 123 try: 124 subprocess.call('yui-compressor', stdout=open(os.devnull, 'w'), stderr=open(os.devnull, 'w')) 125 yuicompressor = 'yui-compressor' 126 except Exception: 127 pass 128 if not yuicompressor: 129 try: 130 subprocess.call('yuicompressor', stdout=open(os.devnull, 'w'), stderr=open(os.devnull, 'w')) 131 yuicompressor = 'yuicompressor' 132 except: 133 raise Exception("yui-compressor is not installed.") 134 return False 135 136 return runinplace(r'{} --nomunge %1 -o %2'.format(yuicompressor), infile) 137 138 139 def optipng(infile): 140 return runinplace(r"optipng -preserve -o2 -quiet %1", infile) 141 142 143 def jpegoptim(infile): 144 return runinplace(r"jpegoptim -p --strip-all -q %1", infile) 145 146 147 def tidy(inplace): 148 # Google site verifcation files are not HTML 149 if re.match(r"google[a-f0-9]+.html", os.path.basename(inplace)) \ 150 and open(inplace).readline().startswith( 151 "google-site-verification:"): 152 return 153 154 # Tidy will give error exits, that we will ignore. 155 output = subprocess.check_output( 156 "tidy -m -w 90 --indent no --quote-marks" 157 "no --keep-time yes --tidy-mark no " 158 "--force-output yes '{0}'; exit 0".format(inplace), stderr=subprocess.STDOUT, shell=True) 159 160 for line in output.split("\n"): 161 if "Warning:" in line: 162 if '<meta> proprietary attribute "charset"' in line: 163 # We want to set it though. 164 continue 165 elif '<meta> lacks "content" attribute' in line: 166 # False alarm to me. 167 continue 168 elif '<div> anchor' in line and 'already defined' in line: 169 # Some seeming problem with JavaScript terminators. 170 continue 171 elif '<img> lacks "alt" attribute' in line: 172 # Happens in gallery code, probably can be tolerated. 173 continue 174 elif '<table> lacks "summary" attribute' in line: 175 # Happens for tables, TODO: Check this is normal. 176 continue 177 elif 'proprietary attribute "data-toggle"' in line or \ 178 'proprietary attribute "data-target"': 179 # Some of our own tricks 180 continue 181 else: 182 assert False, (inplace, line) 183 elif "Error:" in line: 184 if '<time> is not recognized' in line: 185 # False alarm, time is proper HTML5. 186 continue 187 else: 188 assert False, line 189 190 191 @apply_to_text_file 192 def typogrify(data): 193 if typo is None: 194 req_missing(['typogrify'], 'use the typogrify filter') 195 196 data = typo.amp(data) 197 data = typo.widont(data) 198 data = typo.smartypants(data) 199 # Disabled because of typogrify bug where it breaks <title> 200 # data = typo.caps(data) 201 data = typo.initial_quotes(data) 202 return data 203 [end of nikola/filters.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/nikola/filters.py b/nikola/filters.py --- a/nikola/filters.py +++ b/nikola/filters.py @@ -30,7 +30,6 @@ from functools import wraps import os import codecs -import re import shutil import subprocess import tempfile @@ -144,50 +143,6 @@ return runinplace(r"jpegoptim -p --strip-all -q %1", infile) -def tidy(inplace): - # Google site verifcation files are not HTML - if re.match(r"google[a-f0-9]+.html", os.path.basename(inplace)) \ - and open(inplace).readline().startswith( - "google-site-verification:"): - return - - # Tidy will give error exits, that we will ignore. - output = subprocess.check_output( - "tidy -m -w 90 --indent no --quote-marks" - "no --keep-time yes --tidy-mark no " - "--force-output yes '{0}'; exit 0".format(inplace), stderr=subprocess.STDOUT, shell=True) - - for line in output.split("\n"): - if "Warning:" in line: - if '<meta> proprietary attribute "charset"' in line: - # We want to set it though. - continue - elif '<meta> lacks "content" attribute' in line: - # False alarm to me. - continue - elif '<div> anchor' in line and 'already defined' in line: - # Some seeming problem with JavaScript terminators. - continue - elif '<img> lacks "alt" attribute' in line: - # Happens in gallery code, probably can be tolerated. - continue - elif '<table> lacks "summary" attribute' in line: - # Happens for tables, TODO: Check this is normal. - continue - elif 'proprietary attribute "data-toggle"' in line or \ - 'proprietary attribute "data-target"': - # Some of our own tricks - continue - else: - assert False, (inplace, line) - elif "Error:" in line: - if '<time> is not recognized' in line: - # False alarm, time is proper HTML5. - continue - else: - assert False, line - - @apply_to_text_file def typogrify(data): if typo is None:
{"golden_diff": "diff --git a/nikola/filters.py b/nikola/filters.py\n--- a/nikola/filters.py\n+++ b/nikola/filters.py\n@@ -30,7 +30,6 @@\n from functools import wraps\n import os\n import codecs\n-import re\n import shutil\n import subprocess\n import tempfile\n@@ -144,50 +143,6 @@\n return runinplace(r\"jpegoptim -p --strip-all -q %1\", infile)\n \n \n-def tidy(inplace):\n- # Google site verifcation files are not HTML\n- if re.match(r\"google[a-f0-9]+.html\", os.path.basename(inplace)) \\\n- and open(inplace).readline().startswith(\n- \"google-site-verification:\"):\n- return\n-\n- # Tidy will give error exits, that we will ignore.\n- output = subprocess.check_output(\n- \"tidy -m -w 90 --indent no --quote-marks\"\n- \"no --keep-time yes --tidy-mark no \"\n- \"--force-output yes '{0}'; exit 0\".format(inplace), stderr=subprocess.STDOUT, shell=True)\n-\n- for line in output.split(\"\\n\"):\n- if \"Warning:\" in line:\n- if '<meta> proprietary attribute \"charset\"' in line:\n- # We want to set it though.\n- continue\n- elif '<meta> lacks \"content\" attribute' in line:\n- # False alarm to me.\n- continue\n- elif '<div> anchor' in line and 'already defined' in line:\n- # Some seeming problem with JavaScript terminators.\n- continue\n- elif '<img> lacks \"alt\" attribute' in line:\n- # Happens in gallery code, probably can be tolerated.\n- continue\n- elif '<table> lacks \"summary\" attribute' in line:\n- # Happens for tables, TODO: Check this is normal.\n- continue\n- elif 'proprietary attribute \"data-toggle\"' in line or \\\n- 'proprietary attribute \"data-target\"':\n- # Some of our own tricks\n- continue\n- else:\n- assert False, (inplace, line)\n- elif \"Error:\" in line:\n- if '<time> is not recognized' in line:\n- # False alarm, time is proper HTML5.\n- continue\n- else:\n- assert False, line\n-\n-\n @apply_to_text_file\n def typogrify(data):\n if typo is None:\n", "issue": "Unable to use Tidy on html\nHi, \n\nI enabled the tidy filter on html file:\n\nfrom nikola import filters\nFILTERS = {\n \".html\": [filters.tidy],\n }\n\nand now NIkola refuses to build my site: \n\nTaskError - taskid:render_tags:../categories/index.html\nPythonAction Error\nTraceback (most recent call last):\n File \"/usr/lib64/python3.3/site-packages/doit/action.py\", line 360, in execute\n returned_value = self.py_callable(_self.args, *_kwargs)\n File \"/usr/lib64/python3.3/site-packages/nikola/utils.py\", line 529, in unlessLink\n action(target)\n File \"/usr/lib64/python3.3/site-packages/nikola/filters.py\", line 144, in tidy\n for line in output.split(\"\\n\"):\nTypeError: Type str doesn't support the buffer AP\n\nIs there a problem with tidy ? Can I use another thing as a replacement ?\n\nUnable to use Tidy on html\nHi, \n\nI enabled the tidy filter on html file:\n\nfrom nikola import filters\nFILTERS = {\n \".html\": [filters.tidy],\n }\n\nand now NIkola refuses to build my site: \n\nTaskError - taskid:render_tags:../categories/index.html\nPythonAction Error\nTraceback (most recent call last):\n File \"/usr/lib64/python3.3/site-packages/doit/action.py\", line 360, in execute\n returned_value = self.py_callable(_self.args, *_kwargs)\n File \"/usr/lib64/python3.3/site-packages/nikola/utils.py\", line 529, in unlessLink\n action(target)\n File \"/usr/lib64/python3.3/site-packages/nikola/filters.py\", line 144, in tidy\n for line in output.split(\"\\n\"):\nTypeError: Type str doesn't support the buffer AP\n\nIs there a problem with tidy ? Can I use another thing as a replacement ?\n\n", "before_files": [{"content": "# -*- coding: utf-8 -*-\n\n# Copyright \u00a9 2012-2014 Roberto Alsina and others.\n\n# Permission is hereby granted, free of charge, to any\n# person obtaining a copy of this software and associated\n# documentation files (the \"Software\"), to deal in the\n# Software without restriction, including without limitation\n# the rights to use, copy, modify, merge, publish,\n# distribute, sublicense, and/or sell copies of the\n# Software, and to permit persons to whom the Software is\n# furnished to do so, subject to the following conditions:\n#\n# The above copyright notice and this permission notice\n# shall be included in all copies or substantial portions of\n# the Software.\n#\n# THE SOFTWARE IS PROVIDED \"AS IS\", WITHOUT WARRANTY OF ANY\n# KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE\n# WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR\n# PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS\n# OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR\n# OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR\n# OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE\n# SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.\n\n\"\"\"Utility functions to help you run filters on files.\"\"\"\n\nfrom .utils import req_missing\nfrom functools import wraps\nimport os\nimport codecs\nimport re\nimport shutil\nimport subprocess\nimport tempfile\nimport shlex\n\ntry:\n import typogrify.filters as typo\nexcept ImportError:\n typo = None # NOQA\n\n\ndef apply_to_binary_file(f):\n \"\"\"Take a function f that transforms a data argument, and returns\n a function that takes a filename and applies f to the contents,\n in place. Reads files in binary mode.\"\"\"\n @wraps(f)\n def f_in_file(fname):\n with open(fname, 'rb') as inf:\n data = inf.read()\n data = f(data)\n with open(fname, 'wb+') as outf:\n outf.write(data)\n\n return f_in_file\n\n\ndef apply_to_text_file(f):\n \"\"\"Take a function f that transforms a data argument, and returns\n a function that takes a filename and applies f to the contents,\n in place. Reads files in UTF-8.\"\"\"\n @wraps(f)\n def f_in_file(fname):\n with codecs.open(fname, 'r', 'utf-8') as inf:\n data = inf.read()\n data = f(data)\n with codecs.open(fname, 'w+', 'utf-8') as outf:\n outf.write(data)\n\n return f_in_file\n\n\ndef list_replace(the_list, find, replacement):\n \"Replace all occurrences of ``find`` with ``replacement`` in ``the_list``\"\n for i, v in enumerate(the_list):\n if v == find:\n the_list[i] = replacement\n\n\ndef runinplace(command, infile):\n \"\"\"Run a command in-place on a file.\n\n command is a string of the form: \"commandname %1 %2\" and\n it will be execed with infile as %1 and a temporary file\n as %2. Then, that temporary file will be moved over %1.\n\n Example usage:\n\n runinplace(\"yui-compressor %1 -o %2\", \"myfile.css\")\n\n That will replace myfile.css with a minified version.\n\n You can also supply command as a list.\n \"\"\"\n\n if not isinstance(command, list):\n command = shlex.split(command)\n\n tmpdir = None\n\n if \"%2\" in command:\n tmpdir = tempfile.mkdtemp(prefix=\"nikola\")\n tmpfname = os.path.join(tmpdir, os.path.basename(infile))\n\n try:\n list_replace(command, \"%1\", infile)\n if tmpdir:\n list_replace(command, \"%2\", tmpfname)\n\n subprocess.check_call(command)\n\n if tmpdir:\n shutil.move(tmpfname, infile)\n finally:\n if tmpdir:\n shutil.rmtree(tmpdir)\n\n\ndef yui_compressor(infile):\n yuicompressor = False\n try:\n subprocess.call('yui-compressor', stdout=open(os.devnull, 'w'), stderr=open(os.devnull, 'w'))\n yuicompressor = 'yui-compressor'\n except Exception:\n pass\n if not yuicompressor:\n try:\n subprocess.call('yuicompressor', stdout=open(os.devnull, 'w'), stderr=open(os.devnull, 'w'))\n yuicompressor = 'yuicompressor'\n except:\n raise Exception(\"yui-compressor is not installed.\")\n return False\n\n return runinplace(r'{} --nomunge %1 -o %2'.format(yuicompressor), infile)\n\n\ndef optipng(infile):\n return runinplace(r\"optipng -preserve -o2 -quiet %1\", infile)\n\n\ndef jpegoptim(infile):\n return runinplace(r\"jpegoptim -p --strip-all -q %1\", infile)\n\n\ndef tidy(inplace):\n # Google site verifcation files are not HTML\n if re.match(r\"google[a-f0-9]+.html\", os.path.basename(inplace)) \\\n and open(inplace).readline().startswith(\n \"google-site-verification:\"):\n return\n\n # Tidy will give error exits, that we will ignore.\n output = subprocess.check_output(\n \"tidy -m -w 90 --indent no --quote-marks\"\n \"no --keep-time yes --tidy-mark no \"\n \"--force-output yes '{0}'; exit 0\".format(inplace), stderr=subprocess.STDOUT, shell=True)\n\n for line in output.split(\"\\n\"):\n if \"Warning:\" in line:\n if '<meta> proprietary attribute \"charset\"' in line:\n # We want to set it though.\n continue\n elif '<meta> lacks \"content\" attribute' in line:\n # False alarm to me.\n continue\n elif '<div> anchor' in line and 'already defined' in line:\n # Some seeming problem with JavaScript terminators.\n continue\n elif '<img> lacks \"alt\" attribute' in line:\n # Happens in gallery code, probably can be tolerated.\n continue\n elif '<table> lacks \"summary\" attribute' in line:\n # Happens for tables, TODO: Check this is normal.\n continue\n elif 'proprietary attribute \"data-toggle\"' in line or \\\n 'proprietary attribute \"data-target\"':\n # Some of our own tricks\n continue\n else:\n assert False, (inplace, line)\n elif \"Error:\" in line:\n if '<time> is not recognized' in line:\n # False alarm, time is proper HTML5.\n continue\n else:\n assert False, line\n\n\n@apply_to_text_file\ndef typogrify(data):\n if typo is None:\n req_missing(['typogrify'], 'use the typogrify filter')\n\n data = typo.amp(data)\n data = typo.widont(data)\n data = typo.smartypants(data)\n # Disabled because of typogrify bug where it breaks <title>\n # data = typo.caps(data)\n data = typo.initial_quotes(data)\n return data\n", "path": "nikola/filters.py"}]}
3,043
554
gh_patches_debug_20428
rasdani/github-patches
git_diff
pyodide__pyodide-1231
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> Minor fix: add parameter for shallowCopyToJavascript Follow #1202 </issue> <code> [start of conftest.py] 1 """ 2 Various common utilities for testing. 3 """ 4 5 import contextlib 6 import multiprocessing 7 import textwrap 8 import tempfile 9 import time 10 import os 11 import pathlib 12 import queue 13 import sys 14 import shutil 15 16 ROOT_PATH = pathlib.Path(__file__).parents[0].resolve() 17 TEST_PATH = ROOT_PATH / "src" / "tests" 18 BUILD_PATH = ROOT_PATH / "build" 19 20 sys.path.append(str(ROOT_PATH)) 21 22 from pyodide_build._fixes import _selenium_is_connectable # noqa: E402 23 import selenium.webdriver.common.utils # noqa: E402 24 25 # XXX: Temporary fix for ConnectionError in selenium 26 27 selenium.webdriver.common.utils.is_connectable = _selenium_is_connectable 28 29 try: 30 import pytest 31 32 def pytest_addoption(parser): 33 group = parser.getgroup("general") 34 group.addoption( 35 "--build-dir", 36 action="store", 37 default=BUILD_PATH, 38 help="Path to the build directory", 39 ) 40 group.addoption( 41 "--run-xfail", 42 action="store_true", 43 help="If provided, tests marked as xfail will be run", 44 ) 45 46 47 except ImportError: 48 pytest = None # type: ignore 49 50 51 class JavascriptException(Exception): 52 def __init__(self, msg, stack): 53 self.msg = msg 54 self.stack = stack 55 # In chrome the stack contains the message 56 if self.stack and self.stack.startswith(self.msg): 57 self.msg = "" 58 59 def __str__(self): 60 return "\n\n".join(x for x in [self.msg, self.stack] if x) 61 62 63 class SeleniumWrapper: 64 JavascriptException = JavascriptException 65 66 def __init__( 67 self, server_port, server_hostname="127.0.0.1", server_log=None, build_dir=None 68 ): 69 if build_dir is None: 70 build_dir = BUILD_PATH 71 72 self.driver = self.get_driver() 73 self.server_port = server_port 74 self.server_hostname = server_hostname 75 self.server_log = server_log 76 77 if not (pathlib.Path(build_dir) / "test.html").exists(): 78 # selenium does not expose HTTP response codes 79 raise ValueError( 80 f"{(build_dir / 'test.html').resolve()} " f"does not exist!" 81 ) 82 self.driver.get(f"http://{server_hostname}:{server_port}/test.html") 83 self.run_js("Error.stackTraceLimit = Infinity") 84 self.run_js("await languagePluginLoader") 85 86 @property 87 def logs(self): 88 logs = self.driver.execute_script("return window.logs") 89 if logs is not None: 90 return "\n".join(str(x) for x in logs) 91 else: 92 return "" 93 94 def clean_logs(self): 95 self.driver.execute_script("window.logs = []") 96 97 def run(self, code): 98 return self.run_js( 99 f""" 100 let result = pyodide.runPython({code!r}); 101 if(result && result.deepCopyToJavascript){{ 102 let converted_result = result.deepCopyToJavascript(); 103 result.destroy(); 104 return converted_result; 105 }} 106 return result; 107 """ 108 ) 109 110 def run_async(self, code): 111 return self.run_js( 112 f""" 113 let result = await pyodide.runPythonAsync({code!r}); 114 if(result && result.deepCopyToJavascript){{ 115 let converted_result = result.deepCopyToJavascript(); 116 result.destroy(); 117 return converted_result; 118 }} 119 return result; 120 """ 121 ) 122 123 def run_js(self, code): 124 if isinstance(code, str) and code.startswith("\n"): 125 # we have a multiline string, fix indentation 126 code = textwrap.dedent(code) 127 128 wrapper = """ 129 let cb = arguments[arguments.length - 1]; 130 let run = async () => { %s } 131 (async () => { 132 try { 133 let result = await run(); 134 if(pyodide && pyodide._module && pyodide._module._PyErr_Occurred()){ 135 try { 136 pyodide._module._pythonexc2js(); 137 } catch(e){ 138 console.error(`Python exited with error flag set! Error was:\n{e.message}`); 139 // Don't put original error message in new one: we want 140 // "pytest.raises(xxx, match=msg)" to fail 141 throw new Error(`Python exited with error flag set!`); 142 } 143 } 144 cb([0, result]); 145 } catch (e) { 146 cb([1, e.toString(), e.stack]); 147 } 148 })() 149 """ 150 151 retval = self.driver.execute_async_script(wrapper % code) 152 153 if retval[0] == 0: 154 return retval[1] 155 else: 156 raise JavascriptException(retval[1], retval[2]) 157 158 def run_webworker(self, code): 159 if isinstance(code, str) and code.startswith("\n"): 160 # we have a multiline string, fix indentation 161 code = textwrap.dedent(code) 162 163 return self.run_js( 164 """ 165 let worker = new Worker( '{}' ); 166 worker.postMessage({{ python: {!r} }}); 167 return new Promise((res, rej) => {{ 168 worker.onerror = e => rej(e); 169 worker.onmessage = e => {{ 170 if (e.data.results) {{ 171 res(e.data.results); 172 }} else {{ 173 rej(e.data.error); 174 }} 175 }}; 176 }}) 177 """.format( 178 f"http://{self.server_hostname}:{self.server_port}/webworker_dev.js", 179 code, 180 ) 181 ) 182 183 def load_package(self, packages): 184 self.run_js("await pyodide.loadPackage({!r})".format(packages)) 185 186 @property 187 def urls(self): 188 for handle in self.driver.window_handles: 189 self.driver.switch_to.window(handle) 190 yield self.driver.current_url 191 192 193 class FirefoxWrapper(SeleniumWrapper): 194 195 browser = "firefox" 196 197 def get_driver(self): 198 from selenium.webdriver import Firefox 199 from selenium.webdriver.firefox.options import Options 200 201 options = Options() 202 options.add_argument("-headless") 203 204 return Firefox(executable_path="geckodriver", options=options) 205 206 207 class ChromeWrapper(SeleniumWrapper): 208 209 browser = "chrome" 210 211 def get_driver(self): 212 from selenium.webdriver import Chrome 213 from selenium.webdriver.chrome.options import Options 214 215 options = Options() 216 options.add_argument("--headless") 217 options.add_argument("--no-sandbox") 218 219 return Chrome(options=options) 220 221 222 if pytest is not None: 223 224 @pytest.fixture(params=["firefox", "chrome"]) 225 def selenium_standalone(request, web_server_main): 226 server_hostname, server_port, server_log = web_server_main 227 if request.param == "firefox": 228 cls = FirefoxWrapper 229 elif request.param == "chrome": 230 cls = ChromeWrapper 231 selenium = cls( 232 build_dir=request.config.option.build_dir, 233 server_port=server_port, 234 server_hostname=server_hostname, 235 server_log=server_log, 236 ) 237 try: 238 yield selenium 239 finally: 240 print(selenium.logs) 241 selenium.driver.quit() 242 243 @pytest.fixture(params=["firefox", "chrome"], scope="module") 244 def _selenium_cached(request, web_server_main): 245 # Cached selenium instance. This is a copy-paste of 246 # selenium_standalone to avoid fixture scope issues 247 server_hostname, server_port, server_log = web_server_main 248 if request.param == "firefox": 249 cls = FirefoxWrapper 250 elif request.param == "chrome": 251 cls = ChromeWrapper 252 selenium = cls( 253 build_dir=request.config.option.build_dir, 254 server_port=server_port, 255 server_hostname=server_hostname, 256 server_log=server_log, 257 ) 258 try: 259 yield selenium 260 finally: 261 selenium.driver.quit() 262 263 @pytest.fixture 264 def selenium(_selenium_cached): 265 # selenium instance cached at the module level 266 try: 267 _selenium_cached.clean_logs() 268 yield _selenium_cached 269 finally: 270 print(_selenium_cached.logs) 271 272 273 @pytest.fixture(scope="session") 274 def web_server_main(request): 275 """Web server that serves files in the build/ directory""" 276 with spawn_web_server(request.config.option.build_dir) as output: 277 yield output 278 279 280 @pytest.fixture(scope="session") 281 def web_server_secondary(request): 282 """Secondary web server that serves files build/ directory""" 283 with spawn_web_server(request.config.option.build_dir) as output: 284 yield output 285 286 287 @pytest.fixture(scope="session") 288 def web_server_tst_data(request): 289 """Web server that serves files in the src/tests/data/ directory""" 290 with spawn_web_server(TEST_PATH / "data") as output: 291 yield output 292 293 294 @contextlib.contextmanager 295 def spawn_web_server(build_dir=None): 296 297 if build_dir is None: 298 build_dir = BUILD_PATH 299 300 tmp_dir = tempfile.mkdtemp() 301 log_path = pathlib.Path(tmp_dir) / "http-server.log" 302 q = multiprocessing.Queue() 303 p = multiprocessing.Process(target=run_web_server, args=(q, log_path, build_dir)) 304 305 try: 306 p.start() 307 port = q.get() 308 hostname = "127.0.0.1" 309 310 print( 311 f"Spawning webserver at http://{hostname}:{port} " 312 f"(see logs in {log_path})" 313 ) 314 yield hostname, port, log_path 315 finally: 316 q.put("TERMINATE") 317 p.join() 318 shutil.rmtree(tmp_dir) 319 320 321 def run_web_server(q, log_filepath, build_dir): 322 """Start the HTTP web server 323 324 Parameters 325 ---------- 326 q : Queue 327 communication queue 328 log_path : pathlib.Path 329 path to the file where to store the logs 330 """ 331 import http.server 332 import socketserver 333 334 os.chdir(build_dir) 335 336 log_fh = log_filepath.open("w", buffering=1) 337 sys.stdout = log_fh 338 sys.stderr = log_fh 339 340 test_prefix = "/src/tests/" 341 342 class Handler(http.server.SimpleHTTPRequestHandler): 343 def log_message(self, format_, *args): 344 print( 345 "[%s] source: %s:%s - %s" 346 % (self.log_date_time_string(), *self.client_address, format_ % args) 347 ) 348 349 def end_headers(self): 350 # Enable Cross-Origin Resource Sharing (CORS) 351 self.send_header("Access-Control-Allow-Origin", "*") 352 super().end_headers() 353 354 with socketserver.TCPServer(("", 0), Handler) as httpd: 355 host, port = httpd.server_address 356 print(f"Starting webserver at http://{host}:{port}") 357 httpd.server_name = "test-server" 358 httpd.server_port = port 359 q.put(port) 360 361 def service_actions(): 362 try: 363 if q.get(False) == "TERMINATE": 364 print("Stopping server...") 365 sys.exit(0) 366 except queue.Empty: 367 pass 368 369 httpd.service_actions = service_actions 370 httpd.serve_forever() 371 372 373 if ( 374 __name__ == "__main__" 375 and multiprocessing.current_process().name == "MainProcess" 376 and not hasattr(sys, "_pytest_session") 377 ): 378 with spawn_web_server(): 379 # run forever 380 while True: 381 time.sleep(1) 382 [end of conftest.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/conftest.py b/conftest.py --- a/conftest.py +++ b/conftest.py @@ -98,8 +98,8 @@ return self.run_js( f""" let result = pyodide.runPython({code!r}); - if(result && result.deepCopyToJavascript){{ - let converted_result = result.deepCopyToJavascript(); + if(result && result.toJs){{ + let converted_result = result.toJs(); result.destroy(); return converted_result; }} @@ -111,8 +111,8 @@ return self.run_js( f""" let result = await pyodide.runPythonAsync({code!r}); - if(result && result.deepCopyToJavascript){{ - let converted_result = result.deepCopyToJavascript(); + if(result && result.toJs){{ + let converted_result = result.toJs(); result.destroy(); return converted_result; }}
{"golden_diff": "diff --git a/conftest.py b/conftest.py\n--- a/conftest.py\n+++ b/conftest.py\n@@ -98,8 +98,8 @@\n return self.run_js(\n f\"\"\"\n let result = pyodide.runPython({code!r});\n- if(result && result.deepCopyToJavascript){{\n- let converted_result = result.deepCopyToJavascript();\n+ if(result && result.toJs){{\n+ let converted_result = result.toJs();\n result.destroy();\n return converted_result;\n }}\n@@ -111,8 +111,8 @@\n return self.run_js(\n f\"\"\"\n let result = await pyodide.runPythonAsync({code!r});\n- if(result && result.deepCopyToJavascript){{\n- let converted_result = result.deepCopyToJavascript();\n+ if(result && result.toJs){{\n+ let converted_result = result.toJs();\n result.destroy();\n return converted_result;\n }}\n", "issue": "Minor fix: add parameter for shallowCopyToJavascript \nFollow #1202 \r\n\r\n\n", "before_files": [{"content": "\"\"\"\nVarious common utilities for testing.\n\"\"\"\n\nimport contextlib\nimport multiprocessing\nimport textwrap\nimport tempfile\nimport time\nimport os\nimport pathlib\nimport queue\nimport sys\nimport shutil\n\nROOT_PATH = pathlib.Path(__file__).parents[0].resolve()\nTEST_PATH = ROOT_PATH / \"src\" / \"tests\"\nBUILD_PATH = ROOT_PATH / \"build\"\n\nsys.path.append(str(ROOT_PATH))\n\nfrom pyodide_build._fixes import _selenium_is_connectable # noqa: E402\nimport selenium.webdriver.common.utils # noqa: E402\n\n# XXX: Temporary fix for ConnectionError in selenium\n\nselenium.webdriver.common.utils.is_connectable = _selenium_is_connectable\n\ntry:\n import pytest\n\n def pytest_addoption(parser):\n group = parser.getgroup(\"general\")\n group.addoption(\n \"--build-dir\",\n action=\"store\",\n default=BUILD_PATH,\n help=\"Path to the build directory\",\n )\n group.addoption(\n \"--run-xfail\",\n action=\"store_true\",\n help=\"If provided, tests marked as xfail will be run\",\n )\n\n\nexcept ImportError:\n pytest = None # type: ignore\n\n\nclass JavascriptException(Exception):\n def __init__(self, msg, stack):\n self.msg = msg\n self.stack = stack\n # In chrome the stack contains the message\n if self.stack and self.stack.startswith(self.msg):\n self.msg = \"\"\n\n def __str__(self):\n return \"\\n\\n\".join(x for x in [self.msg, self.stack] if x)\n\n\nclass SeleniumWrapper:\n JavascriptException = JavascriptException\n\n def __init__(\n self, server_port, server_hostname=\"127.0.0.1\", server_log=None, build_dir=None\n ):\n if build_dir is None:\n build_dir = BUILD_PATH\n\n self.driver = self.get_driver()\n self.server_port = server_port\n self.server_hostname = server_hostname\n self.server_log = server_log\n\n if not (pathlib.Path(build_dir) / \"test.html\").exists():\n # selenium does not expose HTTP response codes\n raise ValueError(\n f\"{(build_dir / 'test.html').resolve()} \" f\"does not exist!\"\n )\n self.driver.get(f\"http://{server_hostname}:{server_port}/test.html\")\n self.run_js(\"Error.stackTraceLimit = Infinity\")\n self.run_js(\"await languagePluginLoader\")\n\n @property\n def logs(self):\n logs = self.driver.execute_script(\"return window.logs\")\n if logs is not None:\n return \"\\n\".join(str(x) for x in logs)\n else:\n return \"\"\n\n def clean_logs(self):\n self.driver.execute_script(\"window.logs = []\")\n\n def run(self, code):\n return self.run_js(\n f\"\"\"\n let result = pyodide.runPython({code!r});\n if(result && result.deepCopyToJavascript){{\n let converted_result = result.deepCopyToJavascript();\n result.destroy();\n return converted_result;\n }}\n return result;\n \"\"\"\n )\n\n def run_async(self, code):\n return self.run_js(\n f\"\"\"\n let result = await pyodide.runPythonAsync({code!r});\n if(result && result.deepCopyToJavascript){{\n let converted_result = result.deepCopyToJavascript();\n result.destroy();\n return converted_result;\n }}\n return result;\n \"\"\"\n )\n\n def run_js(self, code):\n if isinstance(code, str) and code.startswith(\"\\n\"):\n # we have a multiline string, fix indentation\n code = textwrap.dedent(code)\n\n wrapper = \"\"\"\n let cb = arguments[arguments.length - 1];\n let run = async () => { %s }\n (async () => {\n try {\n let result = await run();\n if(pyodide && pyodide._module && pyodide._module._PyErr_Occurred()){\n try {\n pyodide._module._pythonexc2js();\n } catch(e){\n console.error(`Python exited with error flag set! Error was:\\n{e.message}`);\n // Don't put original error message in new one: we want\n // \"pytest.raises(xxx, match=msg)\" to fail\n throw new Error(`Python exited with error flag set!`);\n }\n }\n cb([0, result]);\n } catch (e) {\n cb([1, e.toString(), e.stack]);\n }\n })()\n \"\"\"\n\n retval = self.driver.execute_async_script(wrapper % code)\n\n if retval[0] == 0:\n return retval[1]\n else:\n raise JavascriptException(retval[1], retval[2])\n\n def run_webworker(self, code):\n if isinstance(code, str) and code.startswith(\"\\n\"):\n # we have a multiline string, fix indentation\n code = textwrap.dedent(code)\n\n return self.run_js(\n \"\"\"\n let worker = new Worker( '{}' );\n worker.postMessage({{ python: {!r} }});\n return new Promise((res, rej) => {{\n worker.onerror = e => rej(e);\n worker.onmessage = e => {{\n if (e.data.results) {{\n res(e.data.results);\n }} else {{\n rej(e.data.error);\n }}\n }};\n }})\n \"\"\".format(\n f\"http://{self.server_hostname}:{self.server_port}/webworker_dev.js\",\n code,\n )\n )\n\n def load_package(self, packages):\n self.run_js(\"await pyodide.loadPackage({!r})\".format(packages))\n\n @property\n def urls(self):\n for handle in self.driver.window_handles:\n self.driver.switch_to.window(handle)\n yield self.driver.current_url\n\n\nclass FirefoxWrapper(SeleniumWrapper):\n\n browser = \"firefox\"\n\n def get_driver(self):\n from selenium.webdriver import Firefox\n from selenium.webdriver.firefox.options import Options\n\n options = Options()\n options.add_argument(\"-headless\")\n\n return Firefox(executable_path=\"geckodriver\", options=options)\n\n\nclass ChromeWrapper(SeleniumWrapper):\n\n browser = \"chrome\"\n\n def get_driver(self):\n from selenium.webdriver import Chrome\n from selenium.webdriver.chrome.options import Options\n\n options = Options()\n options.add_argument(\"--headless\")\n options.add_argument(\"--no-sandbox\")\n\n return Chrome(options=options)\n\n\nif pytest is not None:\n\n @pytest.fixture(params=[\"firefox\", \"chrome\"])\n def selenium_standalone(request, web_server_main):\n server_hostname, server_port, server_log = web_server_main\n if request.param == \"firefox\":\n cls = FirefoxWrapper\n elif request.param == \"chrome\":\n cls = ChromeWrapper\n selenium = cls(\n build_dir=request.config.option.build_dir,\n server_port=server_port,\n server_hostname=server_hostname,\n server_log=server_log,\n )\n try:\n yield selenium\n finally:\n print(selenium.logs)\n selenium.driver.quit()\n\n @pytest.fixture(params=[\"firefox\", \"chrome\"], scope=\"module\")\n def _selenium_cached(request, web_server_main):\n # Cached selenium instance. This is a copy-paste of\n # selenium_standalone to avoid fixture scope issues\n server_hostname, server_port, server_log = web_server_main\n if request.param == \"firefox\":\n cls = FirefoxWrapper\n elif request.param == \"chrome\":\n cls = ChromeWrapper\n selenium = cls(\n build_dir=request.config.option.build_dir,\n server_port=server_port,\n server_hostname=server_hostname,\n server_log=server_log,\n )\n try:\n yield selenium\n finally:\n selenium.driver.quit()\n\n @pytest.fixture\n def selenium(_selenium_cached):\n # selenium instance cached at the module level\n try:\n _selenium_cached.clean_logs()\n yield _selenium_cached\n finally:\n print(_selenium_cached.logs)\n\n\[email protected](scope=\"session\")\ndef web_server_main(request):\n \"\"\"Web server that serves files in the build/ directory\"\"\"\n with spawn_web_server(request.config.option.build_dir) as output:\n yield output\n\n\[email protected](scope=\"session\")\ndef web_server_secondary(request):\n \"\"\"Secondary web server that serves files build/ directory\"\"\"\n with spawn_web_server(request.config.option.build_dir) as output:\n yield output\n\n\[email protected](scope=\"session\")\ndef web_server_tst_data(request):\n \"\"\"Web server that serves files in the src/tests/data/ directory\"\"\"\n with spawn_web_server(TEST_PATH / \"data\") as output:\n yield output\n\n\[email protected]\ndef spawn_web_server(build_dir=None):\n\n if build_dir is None:\n build_dir = BUILD_PATH\n\n tmp_dir = tempfile.mkdtemp()\n log_path = pathlib.Path(tmp_dir) / \"http-server.log\"\n q = multiprocessing.Queue()\n p = multiprocessing.Process(target=run_web_server, args=(q, log_path, build_dir))\n\n try:\n p.start()\n port = q.get()\n hostname = \"127.0.0.1\"\n\n print(\n f\"Spawning webserver at http://{hostname}:{port} \"\n f\"(see logs in {log_path})\"\n )\n yield hostname, port, log_path\n finally:\n q.put(\"TERMINATE\")\n p.join()\n shutil.rmtree(tmp_dir)\n\n\ndef run_web_server(q, log_filepath, build_dir):\n \"\"\"Start the HTTP web server\n\n Parameters\n ----------\n q : Queue\n communication queue\n log_path : pathlib.Path\n path to the file where to store the logs\n \"\"\"\n import http.server\n import socketserver\n\n os.chdir(build_dir)\n\n log_fh = log_filepath.open(\"w\", buffering=1)\n sys.stdout = log_fh\n sys.stderr = log_fh\n\n test_prefix = \"/src/tests/\"\n\n class Handler(http.server.SimpleHTTPRequestHandler):\n def log_message(self, format_, *args):\n print(\n \"[%s] source: %s:%s - %s\"\n % (self.log_date_time_string(), *self.client_address, format_ % args)\n )\n\n def end_headers(self):\n # Enable Cross-Origin Resource Sharing (CORS)\n self.send_header(\"Access-Control-Allow-Origin\", \"*\")\n super().end_headers()\n\n with socketserver.TCPServer((\"\", 0), Handler) as httpd:\n host, port = httpd.server_address\n print(f\"Starting webserver at http://{host}:{port}\")\n httpd.server_name = \"test-server\"\n httpd.server_port = port\n q.put(port)\n\n def service_actions():\n try:\n if q.get(False) == \"TERMINATE\":\n print(\"Stopping server...\")\n sys.exit(0)\n except queue.Empty:\n pass\n\n httpd.service_actions = service_actions\n httpd.serve_forever()\n\n\nif (\n __name__ == \"__main__\"\n and multiprocessing.current_process().name == \"MainProcess\"\n and not hasattr(sys, \"_pytest_session\")\n):\n with spawn_web_server():\n # run forever\n while True:\n time.sleep(1)\n", "path": "conftest.py"}]}
4,006
212
gh_patches_debug_19727
rasdani/github-patches
git_diff
facebookresearch__hydra-1424
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> Upgrade to OmegaConf 2.1 OmegaConf 2.1 is adding many important new features. For example: * Powerful interpolation grammar supporting nested interpolations * Relative interpolations * And many many bug fixes Release notes: [omegaconf==2.1.0.rc1](https://github.com/omry/omegaconf/releases/tag/v2.1.0.rc1). </issue> <code> [start of plugins/hydra_rq_launcher/hydra_plugins/hydra_rq_launcher/config.py] 1 # Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved 2 from dataclasses import dataclass 3 from typing import Optional 4 5 from hydra.core.config_store import ConfigStore 6 from omegaconf import II 7 8 9 @dataclass 10 class RedisConf: 11 # host address via REDIS_HOST environment variable, default: localhost 12 host: str = II("env:REDIS_HOST,localhost") 13 # port via REDIS_PORT environment variable, default: 6379 14 port: int = II("env:REDIS_PORT,6379") 15 # database via REDIS_DB environment variable, default: 0 16 db: Optional[str] = II("env:REDIS_DB,0") 17 # password via REDIS_PASSWORD environment variable, default: no password 18 password: str = II("env:REDIS_PASSWORD,") 19 # switch to run without redis server in single thread, for testing purposes only 20 mock: bool = II("env:REDIS_MOCK,False") 21 22 23 @dataclass 24 class EnqueueConf: 25 # maximum runtime of the job before it's killed (e.g. "1d" for 1 day, units: d/h/m/s), default: no limit 26 job_timeout: Optional[str] = None 27 # maximum queued time before the job before is discarded (e.g. "1d" for 1 day, units: d/h/m/s), default: no limit 28 ttl: Optional[str] = None 29 # how long successful jobs and their results are kept (e.g. "1d" for 1 day, units: d/h/m/s), default: no limit 30 result_ttl: Optional[str] = None 31 # specifies how long failed jobs are kept (e.g. "1d" for 1 day, units: d/h/m/s), default: no limit 32 failure_ttl: Optional[str] = None 33 # place job at the front of the queue, instead of the back 34 at_front: bool = False 35 # job id, will be overidden automatically by a uuid unless specified explicitly 36 job_id: Optional[str] = None 37 # description, will be overidden automatically unless specified explicitly 38 description: Optional[str] = None 39 40 41 @dataclass 42 class RQLauncherConf: 43 _target_: str = "hydra_plugins.hydra_rq_launcher.rq_launcher.RQLauncher" 44 # enqueue configuration 45 enqueue: EnqueueConf = EnqueueConf() 46 # queue name 47 queue: str = "default" 48 # redis configuration 49 redis: RedisConf = RedisConf() 50 # stop after enqueueing by raising custom exception 51 stop_after_enqueue: bool = False 52 # wait time in seconds when polling results 53 wait_polling: float = 1.0 54 55 56 ConfigStore.instance().store( 57 group="hydra/launcher", name="rq", node=RQLauncherConf, provider="rq_launcher" 58 ) 59 [end of plugins/hydra_rq_launcher/hydra_plugins/hydra_rq_launcher/config.py] [start of plugins/hydra_ax_sweeper/setup.py] 1 # Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved 2 # type: ignore 3 from setuptools import find_namespace_packages, setup 4 5 with open("README.md", "r") as fh: 6 LONG_DESC = fh.read() 7 setup( 8 name="hydra-ax-sweeper", 9 version="1.1.0rc1", 10 author="Omry Yadan, Shagun Sodhani", 11 author_email="[email protected], [email protected]", 12 description="Hydra Ax Sweeper plugin", 13 long_description=LONG_DESC, 14 long_description_content_type="text/markdown", 15 url="https://github.com/facebookresearch/hydra/", 16 packages=find_namespace_packages(include=["hydra_plugins.*"]), 17 classifiers=[ 18 "License :: OSI Approved :: MIT License", 19 "Programming Language :: Python :: 3.7", 20 "Programming Language :: Python :: 3.8", 21 "Programming Language :: Python :: 3.9", 22 "Operating System :: POSIX :: Linux", 23 "Operating System :: MacOS", 24 "Development Status :: 4 - Beta", 25 ], 26 install_requires=[ 27 "hydra-core>=1.0.0", 28 "ax-platform>=0.1.13", 29 "numpy<1.20.0", # remove once ax is upgraded to support numpy 1.20 30 ], 31 include_package_data=True, 32 ) 33 [end of plugins/hydra_ax_sweeper/setup.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/plugins/hydra_ax_sweeper/setup.py b/plugins/hydra_ax_sweeper/setup.py --- a/plugins/hydra_ax_sweeper/setup.py +++ b/plugins/hydra_ax_sweeper/setup.py @@ -25,8 +25,7 @@ ], install_requires=[ "hydra-core>=1.0.0", - "ax-platform>=0.1.13", - "numpy<1.20.0", # remove once ax is upgraded to support numpy 1.20 + "ax-platform>=0.1.20", ], include_package_data=True, ) diff --git a/plugins/hydra_rq_launcher/hydra_plugins/hydra_rq_launcher/config.py b/plugins/hydra_rq_launcher/hydra_plugins/hydra_rq_launcher/config.py --- a/plugins/hydra_rq_launcher/hydra_plugins/hydra_rq_launcher/config.py +++ b/plugins/hydra_rq_launcher/hydra_plugins/hydra_rq_launcher/config.py @@ -15,7 +15,7 @@ # database via REDIS_DB environment variable, default: 0 db: Optional[str] = II("env:REDIS_DB,0") # password via REDIS_PASSWORD environment variable, default: no password - password: str = II("env:REDIS_PASSWORD,") + password: str = II("env:REDIS_PASSWORD") # switch to run without redis server in single thread, for testing purposes only mock: bool = II("env:REDIS_MOCK,False")
{"golden_diff": "diff --git a/plugins/hydra_ax_sweeper/setup.py b/plugins/hydra_ax_sweeper/setup.py\n--- a/plugins/hydra_ax_sweeper/setup.py\n+++ b/plugins/hydra_ax_sweeper/setup.py\n@@ -25,8 +25,7 @@\n ],\n install_requires=[\n \"hydra-core>=1.0.0\",\n- \"ax-platform>=0.1.13\",\n- \"numpy<1.20.0\", # remove once ax is upgraded to support numpy 1.20\n+ \"ax-platform>=0.1.20\",\n ],\n include_package_data=True,\n )\ndiff --git a/plugins/hydra_rq_launcher/hydra_plugins/hydra_rq_launcher/config.py b/plugins/hydra_rq_launcher/hydra_plugins/hydra_rq_launcher/config.py\n--- a/plugins/hydra_rq_launcher/hydra_plugins/hydra_rq_launcher/config.py\n+++ b/plugins/hydra_rq_launcher/hydra_plugins/hydra_rq_launcher/config.py\n@@ -15,7 +15,7 @@\n # database via REDIS_DB environment variable, default: 0\n db: Optional[str] = II(\"env:REDIS_DB,0\")\n # password via REDIS_PASSWORD environment variable, default: no password\n- password: str = II(\"env:REDIS_PASSWORD,\")\n+ password: str = II(\"env:REDIS_PASSWORD\")\n # switch to run without redis server in single thread, for testing purposes only\n mock: bool = II(\"env:REDIS_MOCK,False\")\n", "issue": "Upgrade to OmegaConf 2.1\nOmegaConf 2.1 is adding many important new features.\r\nFor example:\r\n* Powerful interpolation grammar supporting nested interpolations\r\n* Relative interpolations\r\n* And many many bug fixes\r\n\r\nRelease notes: [omegaconf==2.1.0.rc1](https://github.com/omry/omegaconf/releases/tag/v2.1.0.rc1).\n", "before_files": [{"content": "# Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved\nfrom dataclasses import dataclass\nfrom typing import Optional\n\nfrom hydra.core.config_store import ConfigStore\nfrom omegaconf import II\n\n\n@dataclass\nclass RedisConf:\n # host address via REDIS_HOST environment variable, default: localhost\n host: str = II(\"env:REDIS_HOST,localhost\")\n # port via REDIS_PORT environment variable, default: 6379\n port: int = II(\"env:REDIS_PORT,6379\")\n # database via REDIS_DB environment variable, default: 0\n db: Optional[str] = II(\"env:REDIS_DB,0\")\n # password via REDIS_PASSWORD environment variable, default: no password\n password: str = II(\"env:REDIS_PASSWORD,\")\n # switch to run without redis server in single thread, for testing purposes only\n mock: bool = II(\"env:REDIS_MOCK,False\")\n\n\n@dataclass\nclass EnqueueConf:\n # maximum runtime of the job before it's killed (e.g. \"1d\" for 1 day, units: d/h/m/s), default: no limit\n job_timeout: Optional[str] = None\n # maximum queued time before the job before is discarded (e.g. \"1d\" for 1 day, units: d/h/m/s), default: no limit\n ttl: Optional[str] = None\n # how long successful jobs and their results are kept (e.g. \"1d\" for 1 day, units: d/h/m/s), default: no limit\n result_ttl: Optional[str] = None\n # specifies how long failed jobs are kept (e.g. \"1d\" for 1 day, units: d/h/m/s), default: no limit\n failure_ttl: Optional[str] = None\n # place job at the front of the queue, instead of the back\n at_front: bool = False\n # job id, will be overidden automatically by a uuid unless specified explicitly\n job_id: Optional[str] = None\n # description, will be overidden automatically unless specified explicitly\n description: Optional[str] = None\n\n\n@dataclass\nclass RQLauncherConf:\n _target_: str = \"hydra_plugins.hydra_rq_launcher.rq_launcher.RQLauncher\"\n # enqueue configuration\n enqueue: EnqueueConf = EnqueueConf()\n # queue name\n queue: str = \"default\"\n # redis configuration\n redis: RedisConf = RedisConf()\n # stop after enqueueing by raising custom exception\n stop_after_enqueue: bool = False\n # wait time in seconds when polling results\n wait_polling: float = 1.0\n\n\nConfigStore.instance().store(\n group=\"hydra/launcher\", name=\"rq\", node=RQLauncherConf, provider=\"rq_launcher\"\n)\n", "path": "plugins/hydra_rq_launcher/hydra_plugins/hydra_rq_launcher/config.py"}, {"content": "# Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved\n# type: ignore\nfrom setuptools import find_namespace_packages, setup\n\nwith open(\"README.md\", \"r\") as fh:\n LONG_DESC = fh.read()\n setup(\n name=\"hydra-ax-sweeper\",\n version=\"1.1.0rc1\",\n author=\"Omry Yadan, Shagun Sodhani\",\n author_email=\"[email protected], [email protected]\",\n description=\"Hydra Ax Sweeper plugin\",\n long_description=LONG_DESC,\n long_description_content_type=\"text/markdown\",\n url=\"https://github.com/facebookresearch/hydra/\",\n packages=find_namespace_packages(include=[\"hydra_plugins.*\"]),\n classifiers=[\n \"License :: OSI Approved :: MIT License\",\n \"Programming Language :: Python :: 3.7\",\n \"Programming Language :: Python :: 3.8\",\n \"Programming Language :: Python :: 3.9\",\n \"Operating System :: POSIX :: Linux\",\n \"Operating System :: MacOS\",\n \"Development Status :: 4 - Beta\",\n ],\n install_requires=[\n \"hydra-core>=1.0.0\",\n \"ax-platform>=0.1.13\",\n \"numpy<1.20.0\", # remove once ax is upgraded to support numpy 1.20\n ],\n include_package_data=True,\n )\n", "path": "plugins/hydra_ax_sweeper/setup.py"}]}
1,765
342
gh_patches_debug_31077
rasdani/github-patches
git_diff
sopel-irc__sopel-1441
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> imdb module not working anymore I just noticed that anytime you make a call to imdb now the bot responds: > [MOVIE] No API key provided. I know it used to work, not sure how recently. Maybe it can be switched to a different database that doesn't require an API key? </issue> <code> [start of sopel/modules/movie.py] 1 # coding=utf-8 2 """ 3 imdb.py - Sopel Movie Information Module 4 Copyright © 2012-2013, Elad Alfassa, <[email protected]> 5 Licensed under the Eiffel Forum License 2. 6 7 This module relies on omdbapi.com 8 """ 9 from __future__ import unicode_literals, absolute_import, print_function, division 10 11 import requests 12 import sopel.module 13 from sopel.logger import get_logger 14 15 LOGGER = get_logger(__name__) 16 17 18 @sopel.module.commands('movie', 'imdb') 19 @sopel.module.example('.movie ThisTitleDoesNotExist', '[MOVIE] Movie not found!') 20 @sopel.module.example('.movie Citizen Kane', '[MOVIE] Title: Citizen Kane | Year: 1941 | Rating: 8.4 | Genre: Drama, Mystery | IMDB Link: http://imdb.com/title/tt0033467') 21 def movie(bot, trigger): 22 """ 23 Returns some information about a movie, like Title, Year, Rating, Genre and IMDB Link. 24 """ 25 if not trigger.group(2): 26 return 27 word = trigger.group(2).rstrip() 28 uri = "http://www.omdbapi.com/" 29 data = requests.get(uri, params={'t': word}, timeout=30, 30 verify=bot.config.core.verify_ssl).json() 31 if data['Response'] == 'False': 32 if 'Error' in data: 33 message = '[MOVIE] %s' % data['Error'] 34 else: 35 LOGGER.warning( 36 'Got an error from the OMDb api, search phrase was %s; data was %s', 37 word, str(data)) 38 message = '[MOVIE] Got an error from OMDbapi' 39 else: 40 message = '[MOVIE] Title: ' + data['Title'] + \ 41 ' | Year: ' + data['Year'] + \ 42 ' | Rating: ' + data['imdbRating'] + \ 43 ' | Genre: ' + data['Genre'] + \ 44 ' | IMDB Link: http://imdb.com/title/' + data['imdbID'] 45 bot.say(message) 46 47 48 if __name__ == "__main__": 49 from sopel.test_tools import run_example_tests 50 run_example_tests(__file__) 51 [end of sopel/modules/movie.py] [start of conftest.py] 1 # This file lists files which should be ignored by pytest 2 collect_ignore = ["setup.py", "sopel.py", "sopel/modules/ipython.py", "sopel/modules/movie.py"] 3 [end of conftest.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/conftest.py b/conftest.py --- a/conftest.py +++ b/conftest.py @@ -1,2 +1,2 @@ # This file lists files which should be ignored by pytest -collect_ignore = ["setup.py", "sopel.py", "sopel/modules/ipython.py", "sopel/modules/movie.py"] +collect_ignore = ["setup.py", "sopel.py", "sopel/modules/ipython.py"] diff --git a/sopel/modules/movie.py b/sopel/modules/movie.py deleted file mode 100644 --- a/sopel/modules/movie.py +++ /dev/null @@ -1,50 +0,0 @@ -# coding=utf-8 -""" -imdb.py - Sopel Movie Information Module -Copyright © 2012-2013, Elad Alfassa, <[email protected]> -Licensed under the Eiffel Forum License 2. - -This module relies on omdbapi.com -""" -from __future__ import unicode_literals, absolute_import, print_function, division - -import requests -import sopel.module -from sopel.logger import get_logger - -LOGGER = get_logger(__name__) - - [email protected]('movie', 'imdb') [email protected]('.movie ThisTitleDoesNotExist', '[MOVIE] Movie not found!') [email protected]('.movie Citizen Kane', '[MOVIE] Title: Citizen Kane | Year: 1941 | Rating: 8.4 | Genre: Drama, Mystery | IMDB Link: http://imdb.com/title/tt0033467') -def movie(bot, trigger): - """ - Returns some information about a movie, like Title, Year, Rating, Genre and IMDB Link. - """ - if not trigger.group(2): - return - word = trigger.group(2).rstrip() - uri = "http://www.omdbapi.com/" - data = requests.get(uri, params={'t': word}, timeout=30, - verify=bot.config.core.verify_ssl).json() - if data['Response'] == 'False': - if 'Error' in data: - message = '[MOVIE] %s' % data['Error'] - else: - LOGGER.warning( - 'Got an error from the OMDb api, search phrase was %s; data was %s', - word, str(data)) - message = '[MOVIE] Got an error from OMDbapi' - else: - message = '[MOVIE] Title: ' + data['Title'] + \ - ' | Year: ' + data['Year'] + \ - ' | Rating: ' + data['imdbRating'] + \ - ' | Genre: ' + data['Genre'] + \ - ' | IMDB Link: http://imdb.com/title/' + data['imdbID'] - bot.say(message) - - -if __name__ == "__main__": - from sopel.test_tools import run_example_tests - run_example_tests(__file__)
{"golden_diff": "diff --git a/conftest.py b/conftest.py\n--- a/conftest.py\n+++ b/conftest.py\n@@ -1,2 +1,2 @@\n # This file lists files which should be ignored by pytest\n-collect_ignore = [\"setup.py\", \"sopel.py\", \"sopel/modules/ipython.py\", \"sopel/modules/movie.py\"]\n+collect_ignore = [\"setup.py\", \"sopel.py\", \"sopel/modules/ipython.py\"]\ndiff --git a/sopel/modules/movie.py b/sopel/modules/movie.py\ndeleted file mode 100644\n--- a/sopel/modules/movie.py\n+++ /dev/null\n@@ -1,50 +0,0 @@\n-# coding=utf-8\n-\"\"\"\n-imdb.py - Sopel Movie Information Module\n-Copyright \u00a9 2012-2013, Elad Alfassa, <[email protected]>\n-Licensed under the Eiffel Forum License 2.\n-\n-This module relies on omdbapi.com\n-\"\"\"\n-from __future__ import unicode_literals, absolute_import, print_function, division\n-\n-import requests\n-import sopel.module\n-from sopel.logger import get_logger\n-\n-LOGGER = get_logger(__name__)\n-\n-\[email protected]('movie', 'imdb')\[email protected]('.movie ThisTitleDoesNotExist', '[MOVIE] Movie not found!')\[email protected]('.movie Citizen Kane', '[MOVIE] Title: Citizen Kane | Year: 1941 | Rating: 8.4 | Genre: Drama, Mystery | IMDB Link: http://imdb.com/title/tt0033467')\n-def movie(bot, trigger):\n- \"\"\"\n- Returns some information about a movie, like Title, Year, Rating, Genre and IMDB Link.\n- \"\"\"\n- if not trigger.group(2):\n- return\n- word = trigger.group(2).rstrip()\n- uri = \"http://www.omdbapi.com/\"\n- data = requests.get(uri, params={'t': word}, timeout=30,\n- verify=bot.config.core.verify_ssl).json()\n- if data['Response'] == 'False':\n- if 'Error' in data:\n- message = '[MOVIE] %s' % data['Error']\n- else:\n- LOGGER.warning(\n- 'Got an error from the OMDb api, search phrase was %s; data was %s',\n- word, str(data))\n- message = '[MOVIE] Got an error from OMDbapi'\n- else:\n- message = '[MOVIE] Title: ' + data['Title'] + \\\n- ' | Year: ' + data['Year'] + \\\n- ' | Rating: ' + data['imdbRating'] + \\\n- ' | Genre: ' + data['Genre'] + \\\n- ' | IMDB Link: http://imdb.com/title/' + data['imdbID']\n- bot.say(message)\n-\n-\n-if __name__ == \"__main__\":\n- from sopel.test_tools import run_example_tests\n- run_example_tests(__file__)\n", "issue": "imdb module not working anymore\nI just noticed that anytime you make a call to imdb now the bot responds: \r\n\r\n> [MOVIE] No API key provided.\r\n\r\nI know it used to work, not sure how recently. Maybe it can be switched to a different database that doesn't require an API key?\n", "before_files": [{"content": "# coding=utf-8\n\"\"\"\nimdb.py - Sopel Movie Information Module\nCopyright \u00a9 2012-2013, Elad Alfassa, <[email protected]>\nLicensed under the Eiffel Forum License 2.\n\nThis module relies on omdbapi.com\n\"\"\"\nfrom __future__ import unicode_literals, absolute_import, print_function, division\n\nimport requests\nimport sopel.module\nfrom sopel.logger import get_logger\n\nLOGGER = get_logger(__name__)\n\n\[email protected]('movie', 'imdb')\[email protected]('.movie ThisTitleDoesNotExist', '[MOVIE] Movie not found!')\[email protected]('.movie Citizen Kane', '[MOVIE] Title: Citizen Kane | Year: 1941 | Rating: 8.4 | Genre: Drama, Mystery | IMDB Link: http://imdb.com/title/tt0033467')\ndef movie(bot, trigger):\n \"\"\"\n Returns some information about a movie, like Title, Year, Rating, Genre and IMDB Link.\n \"\"\"\n if not trigger.group(2):\n return\n word = trigger.group(2).rstrip()\n uri = \"http://www.omdbapi.com/\"\n data = requests.get(uri, params={'t': word}, timeout=30,\n verify=bot.config.core.verify_ssl).json()\n if data['Response'] == 'False':\n if 'Error' in data:\n message = '[MOVIE] %s' % data['Error']\n else:\n LOGGER.warning(\n 'Got an error from the OMDb api, search phrase was %s; data was %s',\n word, str(data))\n message = '[MOVIE] Got an error from OMDbapi'\n else:\n message = '[MOVIE] Title: ' + data['Title'] + \\\n ' | Year: ' + data['Year'] + \\\n ' | Rating: ' + data['imdbRating'] + \\\n ' | Genre: ' + data['Genre'] + \\\n ' | IMDB Link: http://imdb.com/title/' + data['imdbID']\n bot.say(message)\n\n\nif __name__ == \"__main__\":\n from sopel.test_tools import run_example_tests\n run_example_tests(__file__)\n", "path": "sopel/modules/movie.py"}, {"content": "# This file lists files which should be ignored by pytest\ncollect_ignore = [\"setup.py\", \"sopel.py\", \"sopel/modules/ipython.py\", \"sopel/modules/movie.py\"]\n", "path": "conftest.py"}]}
1,250
695
gh_patches_debug_19487
rasdani/github-patches
git_diff
Lightning-AI__torchmetrics-1011
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> Compatibility with future PT 1.12 ## 🐛 Bug ``` > distance = x @ y.T E RuntimeError: "addmm_impl_cpu_" not implemented for 'Half' ``` ### To Reproduce https://github.com/PyTorchLightning/metrics/runs/6275393755?check_suite_focus=true #### Code sample ```py distance = x @ y.T ``` ### Expected behavior <!-- A clear and concise description of what you expected to happen. --> ### Environment - OS (e.g., Linux): linux - Python & PyTorch Version (e.g., 1.0): py3.8 & pt1.12 ### Additional context context see https://github.com/pytorch/pytorch/pull/75647 slack: https://pytorch.slack.com/archives/C3PDTEV8E/p1651742487294399 tl;dr: We don't have correct and fast linalg algorithms for half in CPU as none of the backend libraries that we use support them so it's better not to support these via half-baked implementations, and simply make the users cast their inputs if they want to use these methods > but for GPU it shall be supported, correct? >> Yes note as well that this just applies to Half on CPU and for linalg functios (think matmul and conv). Pointwise functions on Half on CPU will still be available, and Half on CUDA will still have full support > it was implemented up till 1.11 but there was no real speed-up, correct? Not only it was slower, but it was not numerically stable, so it was pretty much a bug (hence the removal without deprecation) </issue> <code> [start of torchmetrics/functional/pairwise/cosine.py] 1 # Copyright The PyTorch Lightning team. 2 # 3 # Licensed under the Apache License, Version 2.0 (the "License"); 4 # you may not use this file except in compliance with the License. 5 # You may obtain a copy of the License at 6 # 7 # http://www.apache.org/licenses/LICENSE-2.0 8 # 9 # Unless required by applicable law or agreed to in writing, software 10 # distributed under the License is distributed on an "AS IS" BASIS, 11 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 12 # See the License for the specific language governing permissions and 13 # limitations under the License. 14 from typing import Optional 15 16 import torch 17 from torch import Tensor 18 from typing_extensions import Literal 19 20 from torchmetrics.functional.pairwise.helpers import _check_input, _reduce_distance_matrix 21 22 23 def _pairwise_cosine_similarity_update( 24 x: Tensor, y: Optional[Tensor] = None, zero_diagonal: Optional[bool] = None 25 ) -> Tensor: 26 """Calculates the pairwise cosine similarity matrix. 27 28 Args: 29 x: tensor of shape ``[N,d]`` 30 y: tensor of shape ``[M,d]`` 31 zero_diagonal: determines if the diagonal of the distance matrix should be set to zero 32 """ 33 x, y, zero_diagonal = _check_input(x, y, zero_diagonal) 34 35 norm = torch.norm(x, p=2, dim=1) 36 x /= norm.unsqueeze(1) 37 norm = torch.norm(y, p=2, dim=1) 38 y /= norm.unsqueeze(1) 39 40 distance = x @ y.T 41 if zero_diagonal: 42 distance.fill_diagonal_(0) 43 return distance 44 45 46 def pairwise_cosine_similarity( 47 x: Tensor, 48 y: Optional[Tensor] = None, 49 reduction: Literal["mean", "sum", "none", None] = None, 50 zero_diagonal: Optional[bool] = None, 51 ) -> Tensor: 52 r"""Calculates pairwise cosine similarity: 53 54 .. math:: 55 s_{cos}(x,y) = \frac{<x,y>}{||x|| \cdot ||y||} 56 = \frac{\sum_{d=1}^D x_d \cdot y_d }{\sqrt{\sum_{d=1}^D x_i^2} \cdot \sqrt{\sum_{d=1}^D x_i^2}} 57 58 If both :math:`x` and :math:`y` are passed in, the calculation will be performed pairwise 59 between the rows of :math:`x` and :math:`y`. 60 If only :math:`x` is passed in, the calculation will be performed between the rows of :math:`x`. 61 62 Args: 63 x: Tensor with shape ``[N, d]`` 64 y: Tensor with shape ``[M, d]``, optional 65 reduction: reduction to apply along the last dimension. Choose between `'mean'`, `'sum'` 66 (applied along column dimension) or `'none'`, `None` for no reduction 67 zero_diagonal: if the diagonal of the distance matrix should be set to 0. If only :math:`x` is given 68 this defaults to ``True`` else if :math:`y` is also given it defaults to ``False`` 69 70 Returns: 71 A ``[N,N]`` matrix of distances if only ``x`` is given, else a ``[N,M]`` matrix 72 73 Example: 74 >>> import torch 75 >>> from torchmetrics.functional import pairwise_cosine_similarity 76 >>> x = torch.tensor([[2, 3], [3, 5], [5, 8]], dtype=torch.float32) 77 >>> y = torch.tensor([[1, 0], [2, 1]], dtype=torch.float32) 78 >>> pairwise_cosine_similarity(x, y) 79 tensor([[0.5547, 0.8682], 80 [0.5145, 0.8437], 81 [0.5300, 0.8533]]) 82 >>> pairwise_cosine_similarity(x) 83 tensor([[0.0000, 0.9989, 0.9996], 84 [0.9989, 0.0000, 0.9998], 85 [0.9996, 0.9998, 0.0000]]) 86 87 """ 88 distance = _pairwise_cosine_similarity_update(x, y, zero_diagonal) 89 return _reduce_distance_matrix(distance, reduction) 90 [end of torchmetrics/functional/pairwise/cosine.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/torchmetrics/functional/pairwise/cosine.py b/torchmetrics/functional/pairwise/cosine.py --- a/torchmetrics/functional/pairwise/cosine.py +++ b/torchmetrics/functional/pairwise/cosine.py @@ -20,6 +20,16 @@ from torchmetrics.functional.pairwise.helpers import _check_input, _reduce_distance_matrix +def _safe_matmul(x: Tensor, y: Tensor) -> Tensor: + """Safe calculation of matrix multiplication. + + If input is float16, will cast to float32 for computation and back again. + """ + if x.dtype == torch.float16 or y.dtype == torch.float16: + return (x.float() @ y.T.float()).half() + return x @ y.T + + def _pairwise_cosine_similarity_update( x: Tensor, y: Optional[Tensor] = None, zero_diagonal: Optional[bool] = None ) -> Tensor: @@ -37,7 +47,7 @@ norm = torch.norm(y, p=2, dim=1) y /= norm.unsqueeze(1) - distance = x @ y.T + distance = _safe_matmul(x, y) if zero_diagonal: distance.fill_diagonal_(0) return distance
{"golden_diff": "diff --git a/torchmetrics/functional/pairwise/cosine.py b/torchmetrics/functional/pairwise/cosine.py\n--- a/torchmetrics/functional/pairwise/cosine.py\n+++ b/torchmetrics/functional/pairwise/cosine.py\n@@ -20,6 +20,16 @@\n from torchmetrics.functional.pairwise.helpers import _check_input, _reduce_distance_matrix\n \n \n+def _safe_matmul(x: Tensor, y: Tensor) -> Tensor:\n+ \"\"\"Safe calculation of matrix multiplication.\n+\n+ If input is float16, will cast to float32 for computation and back again.\n+ \"\"\"\n+ if x.dtype == torch.float16 or y.dtype == torch.float16:\n+ return (x.float() @ y.T.float()).half()\n+ return x @ y.T\n+\n+\n def _pairwise_cosine_similarity_update(\n x: Tensor, y: Optional[Tensor] = None, zero_diagonal: Optional[bool] = None\n ) -> Tensor:\n@@ -37,7 +47,7 @@\n norm = torch.norm(y, p=2, dim=1)\n y /= norm.unsqueeze(1)\n \n- distance = x @ y.T\n+ distance = _safe_matmul(x, y)\n if zero_diagonal:\n distance.fill_diagonal_(0)\n return distance\n", "issue": "Compatibility with future PT 1.12\n## \ud83d\udc1b Bug\r\n\r\n```\r\n> distance = x @ y.T\r\nE RuntimeError: \"addmm_impl_cpu_\" not implemented for 'Half'\r\n```\r\n\r\n### To Reproduce\r\n\r\nhttps://github.com/PyTorchLightning/metrics/runs/6275393755?check_suite_focus=true\r\n\r\n#### Code sample\r\n\r\n```py\r\ndistance = x @ y.T\r\n```\r\n\r\n### Expected behavior\r\n\r\n<!-- A clear and concise description of what you expected to happen. -->\r\n\r\n### Environment\r\n\r\n- OS (e.g., Linux): linux\r\n- Python & PyTorch Version (e.g., 1.0): py3.8 & pt1.12\r\n\r\n### Additional context\r\n\r\ncontext see https://github.com/pytorch/pytorch/pull/75647\r\nslack: https://pytorch.slack.com/archives/C3PDTEV8E/p1651742487294399\r\n\r\ntl;dr: We don't have correct and fast linalg algorithms for half in CPU as none of the backend libraries that we use support them so it's better not to support these via half-baked implementations, and simply make the users cast their inputs if they want to use these methods\r\n\r\n> but for GPU it shall be supported, correct? >> Yes\r\n\r\nnote as well that this just applies to Half on CPU and for linalg functios (think matmul and conv). Pointwise functions on Half on CPU will still be available, and Half on CUDA will still have full support\r\n\r\n> it was implemented up till 1.11 but there was no real speed-up, correct?\r\n\r\nNot only it was slower, but it was not numerically stable, so it was pretty much a bug (hence the removal without deprecation)\n", "before_files": [{"content": "# Copyright The PyTorch Lightning team.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\nfrom typing import Optional\n\nimport torch\nfrom torch import Tensor\nfrom typing_extensions import Literal\n\nfrom torchmetrics.functional.pairwise.helpers import _check_input, _reduce_distance_matrix\n\n\ndef _pairwise_cosine_similarity_update(\n x: Tensor, y: Optional[Tensor] = None, zero_diagonal: Optional[bool] = None\n) -> Tensor:\n \"\"\"Calculates the pairwise cosine similarity matrix.\n\n Args:\n x: tensor of shape ``[N,d]``\n y: tensor of shape ``[M,d]``\n zero_diagonal: determines if the diagonal of the distance matrix should be set to zero\n \"\"\"\n x, y, zero_diagonal = _check_input(x, y, zero_diagonal)\n\n norm = torch.norm(x, p=2, dim=1)\n x /= norm.unsqueeze(1)\n norm = torch.norm(y, p=2, dim=1)\n y /= norm.unsqueeze(1)\n\n distance = x @ y.T\n if zero_diagonal:\n distance.fill_diagonal_(0)\n return distance\n\n\ndef pairwise_cosine_similarity(\n x: Tensor,\n y: Optional[Tensor] = None,\n reduction: Literal[\"mean\", \"sum\", \"none\", None] = None,\n zero_diagonal: Optional[bool] = None,\n) -> Tensor:\n r\"\"\"Calculates pairwise cosine similarity:\n\n .. math::\n s_{cos}(x,y) = \\frac{<x,y>}{||x|| \\cdot ||y||}\n = \\frac{\\sum_{d=1}^D x_d \\cdot y_d }{\\sqrt{\\sum_{d=1}^D x_i^2} \\cdot \\sqrt{\\sum_{d=1}^D x_i^2}}\n\n If both :math:`x` and :math:`y` are passed in, the calculation will be performed pairwise\n between the rows of :math:`x` and :math:`y`.\n If only :math:`x` is passed in, the calculation will be performed between the rows of :math:`x`.\n\n Args:\n x: Tensor with shape ``[N, d]``\n y: Tensor with shape ``[M, d]``, optional\n reduction: reduction to apply along the last dimension. Choose between `'mean'`, `'sum'`\n (applied along column dimension) or `'none'`, `None` for no reduction\n zero_diagonal: if the diagonal of the distance matrix should be set to 0. If only :math:`x` is given\n this defaults to ``True`` else if :math:`y` is also given it defaults to ``False``\n\n Returns:\n A ``[N,N]`` matrix of distances if only ``x`` is given, else a ``[N,M]`` matrix\n\n Example:\n >>> import torch\n >>> from torchmetrics.functional import pairwise_cosine_similarity\n >>> x = torch.tensor([[2, 3], [3, 5], [5, 8]], dtype=torch.float32)\n >>> y = torch.tensor([[1, 0], [2, 1]], dtype=torch.float32)\n >>> pairwise_cosine_similarity(x, y)\n tensor([[0.5547, 0.8682],\n [0.5145, 0.8437],\n [0.5300, 0.8533]])\n >>> pairwise_cosine_similarity(x)\n tensor([[0.0000, 0.9989, 0.9996],\n [0.9989, 0.0000, 0.9998],\n [0.9996, 0.9998, 0.0000]])\n\n \"\"\"\n distance = _pairwise_cosine_similarity_update(x, y, zero_diagonal)\n return _reduce_distance_matrix(distance, reduction)\n", "path": "torchmetrics/functional/pairwise/cosine.py"}]}
2,093
298
gh_patches_debug_38729
rasdani/github-patches
git_diff
kivy__kivy-1348
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> Kivy 1.8.0 dev: touch.is_double_tap is set allways to true on Windows Hello, On MS Windows with Kivy 1.8.0 dev: If you press a button twice, even after a greater time gap between the touches, you always get touch.is_double_tap == True on the second tap. Does not appear with 1.7.1 on Android. Greetings Carsten </issue> <code> [start of kivy/input/postproc/doubletap.py] 1 ''' 2 Double Tap 3 ========== 4 5 Search touch for a double tap 6 ''' 7 8 __all__ = ('InputPostprocDoubleTap', ) 9 10 from kivy.config import Config 11 from kivy.vector import Vector 12 from kivy.clock import Clock 13 14 15 class InputPostprocDoubleTap(object): 16 ''' 17 InputPostProcDoubleTap is a post-processor to check if 18 a touch is a double tap or not. 19 Double tap can be configured in the Kivy config file:: 20 21 [postproc] 22 double_tap_time = 250 23 double_tap_distance = 20 24 25 Distance parameter is in 0-1000, and time is in millisecond. 26 ''' 27 28 def __init__(self): 29 dist = Config.getint('postproc', 'double_tap_distance') 30 self.double_tap_distance = dist / 1000.0 31 time = Config.getint('postproc', 'double_tap_time') 32 self.double_tap_time = time / 1000.0 33 self.touches = {} 34 35 def find_double_tap(self, ref): 36 '''Find a double tap touch within self.touches. 37 The touch must be not a previous double tap, and the distance 38 must be ok, also, the touch profile must be compared so the kind 39 of touch is the same 40 ''' 41 ref_button = None 42 if 'button' in ref.profile: 43 ref_button = ref.button 44 45 for touchid in self.touches: 46 if ref.uid == touchid: 47 continue 48 etype, touch = self.touches[touchid] 49 if etype != 'end': 50 continue 51 if touch.is_double_tap: 52 continue 53 distance = Vector.distance( 54 Vector(ref.sx, ref.sy), 55 Vector(touch.osx, touch.osy)) 56 if distance > self.double_tap_distance: 57 continue 58 if touch.is_mouse_scrolling or ref.is_mouse_scrolling: 59 continue 60 touch_button = None 61 if 'button' in touch.profile: 62 touch_button = touch.button 63 if touch_button != ref_button: 64 continue 65 touch.double_tap_distance = distance 66 return touch 67 return None 68 69 def process(self, events): 70 if self.double_tap_distance == 0 or self.double_tap_time == 0: 71 return events 72 # first, check if a touch down have a double tap 73 for etype, touch in events: 74 if not touch.is_touch: 75 continue 76 if etype == 'begin': 77 double_tap = self.find_double_tap(touch) 78 if double_tap: 79 touch.is_double_tap = True 80 time = touch.time_start - double_tap.time_start 81 touch.double_tap_time = time 82 distance = double_tap.double_tap_distance 83 touch.double_tap_distance = distance 84 85 # add the touch internaly 86 self.touches[touch.uid] = (etype, touch) 87 88 # second, check if up-touch is timeout for double tap 89 time_current = Clock.get_time() 90 to_delete = [] 91 for touchid in self.touches.keys(): 92 etype, touch = self.touches[touchid] 93 if etype != 'end': 94 continue 95 if time_current - touch.time_start < self.double_tap_time: 96 continue 97 to_delete.append(touchid) 98 99 for touchid in to_delete: 100 del self.touches[touchid] 101 102 return events 103 [end of kivy/input/postproc/doubletap.py] [start of kivy/input/postproc/tripletap.py] 1 ''' 2 Triple Tap 3 ========== 4 5 .. versionadded:: 1.7.0 6 7 Search touch for a triple tap 8 ''' 9 10 __all__ = ('InputPostprocTripleTap', ) 11 12 from kivy.config import Config 13 from kivy.vector import Vector 14 from kivy.clock import Clock 15 16 17 class InputPostprocTripleTap(object): 18 ''' 19 InputPostProcTripleTap is a post-processor to check if 20 a touch is a triple tap or not. 21 Triple tap can be configured in the Kivy config file:: 22 23 [postproc] 24 triple_tap_time = 250 25 triple_tap_distance = 20 26 27 Distance parameter is in 0-1000, and time is in millisecond. 28 ''' 29 30 def __init__(self): 31 dist = Config.getint('postproc', 'triple_tap_distance') 32 self.triple_tap_distance = dist / 1000.0 33 time = Config.getint('postproc', 'triple_tap_time') 34 self.triple_tap_time = time / 1000.0 35 self.touches = {} 36 37 def find_triple_tap(self, ref): 38 '''Find a triple tap touch within self.touches. 39 The touch must be not a previous triple tap, and the distance 40 must be ok, also, the touch profile must be compared so the kind 41 of touch is the same 42 ''' 43 ref_button = None 44 if 'button' in ref.profile: 45 ref_button = ref.button 46 47 for touchid in self.touches: 48 if ref.uid == touchid: 49 continue 50 etype, touch = self.touches[touchid] 51 if not touch.is_double_tap: 52 continue 53 if etype != 'end': 54 continue 55 if touch.is_triple_tap: 56 continue 57 distance = Vector.distance( 58 Vector(ref.sx, ref.sy), 59 Vector(touch.osx, touch.osy)) 60 if distance > self.triple_tap_distance: 61 continue 62 if touch.is_mouse_scrolling or ref.is_mouse_scrolling: 63 continue 64 touch_button = None 65 if 'button' in touch.profile: 66 touch_button = touch.button 67 if touch_button != ref_button: 68 continue 69 touch.triple_tap_distance = distance 70 return touch 71 return None 72 73 def process(self, events): 74 if self.triple_tap_distance == 0 or self.triple_tap_time == 0: 75 return events 76 # first, check if a touch down have a triple tap 77 for etype, touch in events: 78 if not touch.is_touch: 79 continue 80 if etype == 'begin': 81 triple_tap = self.find_triple_tap(touch) 82 if triple_tap: 83 touch.is_double_tap = False 84 touch.is_triple_tap = True 85 time = touch.time_start - triple_tap.time_start 86 touch.triple_tap_time = time 87 distance = triple_tap.triple_tap_distance 88 touch.triple_tap_distance = distance 89 90 # add the touch internaly 91 self.touches[touch.uid] = (etype, touch) 92 93 # second, check if up-touch is timeout for triple tap 94 time_current = Clock.get_time() 95 to_delete = [] 96 for touchid in self.touches.keys(): 97 etype, touch = self.touches[touchid] 98 if etype != 'end': 99 continue 100 if time_current - touch.time_start < self.triple_tap_time: 101 continue 102 to_delete.append(touchid) 103 104 for touchid in to_delete: 105 del self.touches[touchid] 106 107 return events 108 [end of kivy/input/postproc/tripletap.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/kivy/input/postproc/doubletap.py b/kivy/input/postproc/doubletap.py --- a/kivy/input/postproc/doubletap.py +++ b/kivy/input/postproc/doubletap.py @@ -7,6 +7,7 @@ __all__ = ('InputPostprocDoubleTap', ) +from time import time from kivy.config import Config from kivy.vector import Vector from kivy.clock import Clock @@ -77,8 +78,8 @@ double_tap = self.find_double_tap(touch) if double_tap: touch.is_double_tap = True - time = touch.time_start - double_tap.time_start - touch.double_tap_time = time + tap_time = touch.time_start - double_tap.time_start + touch.double_tap_time = tap_time distance = double_tap.double_tap_distance touch.double_tap_distance = distance @@ -86,7 +87,7 @@ self.touches[touch.uid] = (etype, touch) # second, check if up-touch is timeout for double tap - time_current = Clock.get_time() + time_current = time() to_delete = [] for touchid in self.touches.keys(): etype, touch = self.touches[touchid] diff --git a/kivy/input/postproc/tripletap.py b/kivy/input/postproc/tripletap.py --- a/kivy/input/postproc/tripletap.py +++ b/kivy/input/postproc/tripletap.py @@ -9,6 +9,7 @@ __all__ = ('InputPostprocTripleTap', ) +from time import time from kivy.config import Config from kivy.vector import Vector from kivy.clock import Clock @@ -82,8 +83,8 @@ if triple_tap: touch.is_double_tap = False touch.is_triple_tap = True - time = touch.time_start - triple_tap.time_start - touch.triple_tap_time = time + tap_time = touch.time_start - triple_tap.time_start + touch.triple_tap_time = tap_time distance = triple_tap.triple_tap_distance touch.triple_tap_distance = distance @@ -91,7 +92,7 @@ self.touches[touch.uid] = (etype, touch) # second, check if up-touch is timeout for triple tap - time_current = Clock.get_time() + time_current = time() to_delete = [] for touchid in self.touches.keys(): etype, touch = self.touches[touchid]
{"golden_diff": "diff --git a/kivy/input/postproc/doubletap.py b/kivy/input/postproc/doubletap.py\n--- a/kivy/input/postproc/doubletap.py\n+++ b/kivy/input/postproc/doubletap.py\n@@ -7,6 +7,7 @@\n \n __all__ = ('InputPostprocDoubleTap', )\n \n+from time import time\n from kivy.config import Config\n from kivy.vector import Vector\n from kivy.clock import Clock\n@@ -77,8 +78,8 @@\n double_tap = self.find_double_tap(touch)\n if double_tap:\n touch.is_double_tap = True\n- time = touch.time_start - double_tap.time_start\n- touch.double_tap_time = time\n+ tap_time = touch.time_start - double_tap.time_start\n+ touch.double_tap_time = tap_time\n distance = double_tap.double_tap_distance\n touch.double_tap_distance = distance\n \n@@ -86,7 +87,7 @@\n self.touches[touch.uid] = (etype, touch)\n \n # second, check if up-touch is timeout for double tap\n- time_current = Clock.get_time()\n+ time_current = time()\n to_delete = []\n for touchid in self.touches.keys():\n etype, touch = self.touches[touchid]\ndiff --git a/kivy/input/postproc/tripletap.py b/kivy/input/postproc/tripletap.py\n--- a/kivy/input/postproc/tripletap.py\n+++ b/kivy/input/postproc/tripletap.py\n@@ -9,6 +9,7 @@\n \n __all__ = ('InputPostprocTripleTap', )\n \n+from time import time\n from kivy.config import Config\n from kivy.vector import Vector\n from kivy.clock import Clock\n@@ -82,8 +83,8 @@\n if triple_tap:\n touch.is_double_tap = False\n touch.is_triple_tap = True\n- time = touch.time_start - triple_tap.time_start\n- touch.triple_tap_time = time\n+ tap_time = touch.time_start - triple_tap.time_start\n+ touch.triple_tap_time = tap_time\n distance = triple_tap.triple_tap_distance\n touch.triple_tap_distance = distance\n \n@@ -91,7 +92,7 @@\n self.touches[touch.uid] = (etype, touch)\n \n # second, check if up-touch is timeout for triple tap\n- time_current = Clock.get_time()\n+ time_current = time()\n to_delete = []\n for touchid in self.touches.keys():\n etype, touch = self.touches[touchid]\n", "issue": "Kivy 1.8.0 dev: touch.is_double_tap is set allways to true on Windows\nHello,\nOn MS Windows with Kivy 1.8.0 dev: If you press a button twice, even after a greater time gap between the touches, you always get touch.is_double_tap == True on the second tap. \n\nDoes not appear with 1.7.1 on Android.\n\nGreetings\n\nCarsten\n\n", "before_files": [{"content": "'''\nDouble Tap\n==========\n\nSearch touch for a double tap\n'''\n\n__all__ = ('InputPostprocDoubleTap', )\n\nfrom kivy.config import Config\nfrom kivy.vector import Vector\nfrom kivy.clock import Clock\n\n\nclass InputPostprocDoubleTap(object):\n '''\n InputPostProcDoubleTap is a post-processor to check if\n a touch is a double tap or not.\n Double tap can be configured in the Kivy config file::\n\n [postproc]\n double_tap_time = 250\n double_tap_distance = 20\n\n Distance parameter is in 0-1000, and time is in millisecond.\n '''\n\n def __init__(self):\n dist = Config.getint('postproc', 'double_tap_distance')\n self.double_tap_distance = dist / 1000.0\n time = Config.getint('postproc', 'double_tap_time')\n self.double_tap_time = time / 1000.0\n self.touches = {}\n\n def find_double_tap(self, ref):\n '''Find a double tap touch within self.touches.\n The touch must be not a previous double tap, and the distance\n must be ok, also, the touch profile must be compared so the kind\n of touch is the same\n '''\n ref_button = None\n if 'button' in ref.profile:\n ref_button = ref.button\n\n for touchid in self.touches:\n if ref.uid == touchid:\n continue\n etype, touch = self.touches[touchid]\n if etype != 'end':\n continue\n if touch.is_double_tap:\n continue\n distance = Vector.distance(\n Vector(ref.sx, ref.sy),\n Vector(touch.osx, touch.osy))\n if distance > self.double_tap_distance:\n continue\n if touch.is_mouse_scrolling or ref.is_mouse_scrolling:\n continue\n touch_button = None\n if 'button' in touch.profile:\n touch_button = touch.button\n if touch_button != ref_button:\n continue\n touch.double_tap_distance = distance\n return touch\n return None\n\n def process(self, events):\n if self.double_tap_distance == 0 or self.double_tap_time == 0:\n return events\n # first, check if a touch down have a double tap\n for etype, touch in events:\n if not touch.is_touch:\n continue\n if etype == 'begin':\n double_tap = self.find_double_tap(touch)\n if double_tap:\n touch.is_double_tap = True\n time = touch.time_start - double_tap.time_start\n touch.double_tap_time = time\n distance = double_tap.double_tap_distance\n touch.double_tap_distance = distance\n\n # add the touch internaly\n self.touches[touch.uid] = (etype, touch)\n\n # second, check if up-touch is timeout for double tap\n time_current = Clock.get_time()\n to_delete = []\n for touchid in self.touches.keys():\n etype, touch = self.touches[touchid]\n if etype != 'end':\n continue\n if time_current - touch.time_start < self.double_tap_time:\n continue\n to_delete.append(touchid)\n\n for touchid in to_delete:\n del self.touches[touchid]\n\n return events\n", "path": "kivy/input/postproc/doubletap.py"}, {"content": "'''\nTriple Tap\n==========\n\n.. versionadded:: 1.7.0\n\nSearch touch for a triple tap\n'''\n\n__all__ = ('InputPostprocTripleTap', )\n\nfrom kivy.config import Config\nfrom kivy.vector import Vector\nfrom kivy.clock import Clock\n\n\nclass InputPostprocTripleTap(object):\n '''\n InputPostProcTripleTap is a post-processor to check if\n a touch is a triple tap or not.\n Triple tap can be configured in the Kivy config file::\n\n [postproc]\n triple_tap_time = 250\n triple_tap_distance = 20\n\n Distance parameter is in 0-1000, and time is in millisecond.\n '''\n\n def __init__(self):\n dist = Config.getint('postproc', 'triple_tap_distance')\n self.triple_tap_distance = dist / 1000.0\n time = Config.getint('postproc', 'triple_tap_time')\n self.triple_tap_time = time / 1000.0\n self.touches = {}\n\n def find_triple_tap(self, ref):\n '''Find a triple tap touch within self.touches.\n The touch must be not a previous triple tap, and the distance\n must be ok, also, the touch profile must be compared so the kind\n of touch is the same\n '''\n ref_button = None\n if 'button' in ref.profile:\n ref_button = ref.button\n\n for touchid in self.touches:\n if ref.uid == touchid:\n continue\n etype, touch = self.touches[touchid]\n if not touch.is_double_tap:\n continue\n if etype != 'end':\n continue\n if touch.is_triple_tap:\n continue\n distance = Vector.distance(\n Vector(ref.sx, ref.sy),\n Vector(touch.osx, touch.osy))\n if distance > self.triple_tap_distance:\n continue\n if touch.is_mouse_scrolling or ref.is_mouse_scrolling:\n continue\n touch_button = None\n if 'button' in touch.profile:\n touch_button = touch.button\n if touch_button != ref_button:\n continue\n touch.triple_tap_distance = distance\n return touch\n return None\n\n def process(self, events):\n if self.triple_tap_distance == 0 or self.triple_tap_time == 0:\n return events\n # first, check if a touch down have a triple tap\n for etype, touch in events:\n if not touch.is_touch:\n continue\n if etype == 'begin':\n triple_tap = self.find_triple_tap(touch)\n if triple_tap:\n touch.is_double_tap = False\n touch.is_triple_tap = True\n time = touch.time_start - triple_tap.time_start\n touch.triple_tap_time = time\n distance = triple_tap.triple_tap_distance\n touch.triple_tap_distance = distance\n\n # add the touch internaly\n self.touches[touch.uid] = (etype, touch)\n\n # second, check if up-touch is timeout for triple tap\n time_current = Clock.get_time()\n to_delete = []\n for touchid in self.touches.keys():\n etype, touch = self.touches[touchid]\n if etype != 'end':\n continue\n if time_current - touch.time_start < self.triple_tap_time:\n continue\n to_delete.append(touchid)\n\n for touchid in to_delete:\n del self.touches[touchid]\n\n return events\n", "path": "kivy/input/postproc/tripletap.py"}]}
2,615
586
gh_patches_debug_37251
rasdani/github-patches
git_diff
databricks__koalas-189
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> Document all the methods in Metadata There are a bunch of methods like index_info, index_fields. It's pretty difficult to figure out what they do. We should just add some basic docstring comments. @ueshin you are probably the best person to take this since you created the file. </issue> <code> [start of databricks/koalas/metadata.py] 1 # 2 # Copyright (C) 2019 Databricks, Inc. 3 # 4 # Licensed under the Apache License, Version 2.0 (the "License"); 5 # you may not use this file except in compliance with the License. 6 # You may obtain a copy of the License at 7 # 8 # http://www.apache.org/licenses/LICENSE-2.0 9 # 10 # Unless required by applicable law or agreed to in writing, software 11 # distributed under the License is distributed on an "AS IS" BASIS, 12 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 13 # See the License for the specific language governing permissions and 14 # limitations under the License. 15 # 16 17 """ 18 A metadata to manage indexes. 19 """ 20 import pandas as pd 21 22 from databricks.koalas.dask.compatibility import string_types 23 24 25 class Metadata(object): 26 """ 27 Manages column names and index information 28 """ 29 30 def __init__(self, column_fields, index_info=None): 31 """ Create a new metadata to manage column fields and index fields and names. 32 33 :param column_fields: list of string 34 Field names to appear as columns. 35 :param index_info: list of string pair 36 Each pair holds the index field name which exists in Spark fields, 37 and the index name. 38 """ 39 assert all(isinstance(col, string_types) for col in column_fields) 40 assert index_info is None \ 41 or all(isinstance(index_field, string_types) 42 and (index_name is None or isinstance(index_name, string_types)) 43 for index_field, index_name in index_info) 44 self._column_fields = column_fields 45 self._index_info = index_info or [] 46 47 @property 48 def column_fields(self): 49 return self._column_fields 50 51 @property 52 def index_info(self): 53 return self._index_info 54 55 @property 56 def index_fields(self): 57 return [index_field for index_field, _ in self._index_info] 58 59 @property 60 def index_names(self): 61 return [name for _, name in self._index_info] 62 63 @property 64 def all_fields(self): 65 index_fields = self.index_fields 66 return index_fields + [field for field in self._column_fields 67 if field not in index_fields] 68 69 def copy(self, column_fields=None, index_info=None): 70 if column_fields is None: 71 column_fields = self._column_fields 72 if index_info is None: 73 index_info = self._index_info 74 return Metadata(column_fields=column_fields.copy(), index_info=index_info.copy()) 75 76 @staticmethod 77 def from_pandas(pdf): 78 column_fields = [str(col) for col in pdf.columns] 79 index = pdf.index 80 if isinstance(index, pd.MultiIndex): 81 if index.names is None: 82 index_info = [('__index_level_{}__'.format(i), None) 83 for i in range(len(index.levels))] 84 else: 85 index_info = [('__index_level_{}__'.format(i) if name is None else name, name) 86 for i, name in enumerate(index.names)] 87 else: 88 index_info = [(index.name 89 if index.name is not None else '__index_level_0__', index.name)] 90 91 return Metadata(column_fields=column_fields, index_info=index_info) 92 [end of databricks/koalas/metadata.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/databricks/koalas/metadata.py b/databricks/koalas/metadata.py --- a/databricks/koalas/metadata.py +++ b/databricks/koalas/metadata.py @@ -24,7 +24,11 @@ class Metadata(object): """ - Manages column names and index information + Manages column names and index information. + + :ivar _column_fields: list of the Spark field names to be seen as columns in Koalas DataFrame. + :ivar _index_info: list of pair holding the Spark field names for indexes, + and the index name to be seen in Koalas DataFrame. """ def __init__(self, column_fields, index_info=None): @@ -46,27 +50,38 @@ @property def column_fields(self): + """ Returns the managed column field names. """ return self._column_fields @property def index_info(self): + """ Return the managed index information. """ return self._index_info @property def index_fields(self): + """ Returns the managed index field names. """ return [index_field for index_field, _ in self._index_info] @property def index_names(self): + """ Return the managed index names. """ return [name for _, name in self._index_info] @property def all_fields(self): + """ Return all the field names including index field names. """ index_fields = self.index_fields return index_fields + [field for field in self._column_fields if field not in index_fields] def copy(self, column_fields=None, index_info=None): + """ Copy the metadata. + + :param column_fields: the new column field names. If None, then the original ones are used. + :param index_info: the new index information. If None, then the original one is used. + :return: the copied metadata. + """ if column_fields is None: column_fields = self._column_fields if index_info is None: @@ -75,6 +90,11 @@ @staticmethod def from_pandas(pdf): + """ Create a metadata from pandas DataFrame. + + :param pdf: :class:`pd.DataFrame` + :return: the created metadata + """ column_fields = [str(col) for col in pdf.columns] index = pdf.index if isinstance(index, pd.MultiIndex):
{"golden_diff": "diff --git a/databricks/koalas/metadata.py b/databricks/koalas/metadata.py\n--- a/databricks/koalas/metadata.py\n+++ b/databricks/koalas/metadata.py\n@@ -24,7 +24,11 @@\n \n class Metadata(object):\n \"\"\"\n- Manages column names and index information\n+ Manages column names and index information.\n+\n+ :ivar _column_fields: list of the Spark field names to be seen as columns in Koalas DataFrame.\n+ :ivar _index_info: list of pair holding the Spark field names for indexes,\n+ and the index name to be seen in Koalas DataFrame.\n \"\"\"\n \n def __init__(self, column_fields, index_info=None):\n@@ -46,27 +50,38 @@\n \n @property\n def column_fields(self):\n+ \"\"\" Returns the managed column field names. \"\"\"\n return self._column_fields\n \n @property\n def index_info(self):\n+ \"\"\" Return the managed index information. \"\"\"\n return self._index_info\n \n @property\n def index_fields(self):\n+ \"\"\" Returns the managed index field names. \"\"\"\n return [index_field for index_field, _ in self._index_info]\n \n @property\n def index_names(self):\n+ \"\"\" Return the managed index names. \"\"\"\n return [name for _, name in self._index_info]\n \n @property\n def all_fields(self):\n+ \"\"\" Return all the field names including index field names. \"\"\"\n index_fields = self.index_fields\n return index_fields + [field for field in self._column_fields\n if field not in index_fields]\n \n def copy(self, column_fields=None, index_info=None):\n+ \"\"\" Copy the metadata.\n+\n+ :param column_fields: the new column field names. If None, then the original ones are used.\n+ :param index_info: the new index information. If None, then the original one is used.\n+ :return: the copied metadata.\n+ \"\"\"\n if column_fields is None:\n column_fields = self._column_fields\n if index_info is None:\n@@ -75,6 +90,11 @@\n \n @staticmethod\n def from_pandas(pdf):\n+ \"\"\" Create a metadata from pandas DataFrame.\n+\n+ :param pdf: :class:`pd.DataFrame`\n+ :return: the created metadata\n+ \"\"\"\n column_fields = [str(col) for col in pdf.columns]\n index = pdf.index\n if isinstance(index, pd.MultiIndex):\n", "issue": "Document all the methods in Metadata\nThere are a bunch of methods like index_info, index_fields. It's pretty difficult to figure out what they do. We should just add some basic docstring comments.\r\n\r\n@ueshin you are probably the best person to take this since you created the file.\r\n\n", "before_files": [{"content": "#\n# Copyright (C) 2019 Databricks, Inc.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n#\n\n\"\"\"\nA metadata to manage indexes.\n\"\"\"\nimport pandas as pd\n\nfrom databricks.koalas.dask.compatibility import string_types\n\n\nclass Metadata(object):\n \"\"\"\n Manages column names and index information\n \"\"\"\n\n def __init__(self, column_fields, index_info=None):\n \"\"\" Create a new metadata to manage column fields and index fields and names.\n\n :param column_fields: list of string\n Field names to appear as columns.\n :param index_info: list of string pair\n Each pair holds the index field name which exists in Spark fields,\n and the index name.\n \"\"\"\n assert all(isinstance(col, string_types) for col in column_fields)\n assert index_info is None \\\n or all(isinstance(index_field, string_types)\n and (index_name is None or isinstance(index_name, string_types))\n for index_field, index_name in index_info)\n self._column_fields = column_fields\n self._index_info = index_info or []\n\n @property\n def column_fields(self):\n return self._column_fields\n\n @property\n def index_info(self):\n return self._index_info\n\n @property\n def index_fields(self):\n return [index_field for index_field, _ in self._index_info]\n\n @property\n def index_names(self):\n return [name for _, name in self._index_info]\n\n @property\n def all_fields(self):\n index_fields = self.index_fields\n return index_fields + [field for field in self._column_fields\n if field not in index_fields]\n\n def copy(self, column_fields=None, index_info=None):\n if column_fields is None:\n column_fields = self._column_fields\n if index_info is None:\n index_info = self._index_info\n return Metadata(column_fields=column_fields.copy(), index_info=index_info.copy())\n\n @staticmethod\n def from_pandas(pdf):\n column_fields = [str(col) for col in pdf.columns]\n index = pdf.index\n if isinstance(index, pd.MultiIndex):\n if index.names is None:\n index_info = [('__index_level_{}__'.format(i), None)\n for i in range(len(index.levels))]\n else:\n index_info = [('__index_level_{}__'.format(i) if name is None else name, name)\n for i, name in enumerate(index.names)]\n else:\n index_info = [(index.name\n if index.name is not None else '__index_level_0__', index.name)]\n\n return Metadata(column_fields=column_fields, index_info=index_info)\n", "path": "databricks/koalas/metadata.py"}]}
1,471
557
gh_patches_debug_5870
rasdani/github-patches
git_diff
mitmproxy__mitmproxy-1513
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> Wait for new cryptography release For the next mitmproxy release, we want to recommend Python 3.5 by default. This requires a cryptography release which includes https://github.com/pyca/cryptography/pull/3063. </issue> <code> [start of setup.py] 1 from setuptools import setup, find_packages 2 from codecs import open 3 import os 4 5 from netlib import version 6 7 # Based on https://github.com/pypa/sampleproject/blob/master/setup.py 8 # and https://python-packaging-user-guide.readthedocs.org/ 9 10 here = os.path.abspath(os.path.dirname(__file__)) 11 12 with open(os.path.join(here, 'README.rst'), encoding='utf-8') as f: 13 long_description = f.read() 14 15 setup( 16 name="mitmproxy", 17 version=version.VERSION, 18 description="An interactive, SSL-capable, man-in-the-middle HTTP proxy for penetration testers and software developers.", 19 long_description=long_description, 20 url="http://mitmproxy.org", 21 author="Aldo Cortesi", 22 author_email="[email protected]", 23 license="MIT", 24 classifiers=[ 25 "License :: OSI Approved :: MIT License", 26 "Development Status :: 5 - Production/Stable", 27 "Environment :: Console", 28 "Environment :: Console :: Curses", 29 "Operating System :: MacOS :: MacOS X", 30 "Operating System :: POSIX", 31 "Operating System :: Microsoft :: Windows", 32 "Programming Language :: Python", 33 "Programming Language :: Python :: 2", 34 "Programming Language :: Python :: 2.7", 35 "Programming Language :: Python :: 3", 36 "Programming Language :: Python :: 3.5", 37 "Programming Language :: Python :: Implementation :: CPython", 38 "Programming Language :: Python :: Implementation :: PyPy", 39 "Topic :: Security", 40 "Topic :: Internet", 41 "Topic :: Internet :: WWW/HTTP", 42 "Topic :: Internet :: Proxy Servers", 43 "Topic :: Software Development :: Testing" 44 ], 45 packages=find_packages(include=[ 46 "mitmproxy", "mitmproxy.*", 47 "pathod", "pathod.*", 48 "netlib", "netlib.*" 49 ]), 50 include_package_data=True, 51 entry_points={ 52 'console_scripts': [ 53 "mitmproxy = mitmproxy.main:mitmproxy", 54 "mitmdump = mitmproxy.main:mitmdump", 55 "mitmweb = mitmproxy.main:mitmweb", 56 "pathod = pathod.pathod_cmdline:go_pathod", 57 "pathoc = pathod.pathoc_cmdline:go_pathoc" 58 ] 59 }, 60 # https://packaging.python.org/en/latest/requirements/#install-requires 61 # It is not considered best practice to use install_requires to pin dependencies to specific versions. 62 install_requires=[ 63 "backports.ssl_match_hostname>=3.5.0.1, <3.6", 64 "blinker>=1.4, <1.5", 65 "click>=6.2, <7.0", 66 "certifi>=2015.11.20.1", # no semver here - this should always be on the last release! 67 "configargparse>=0.10, <0.11", 68 "construct>=2.5.2, <2.6", 69 "cryptography>=1.3, <1.5", 70 "cssutils>=1.0.1, <1.1", 71 "Flask>=0.10.1, <0.12", 72 "h2>=2.4.1, <3", 73 "html2text>=2016.1.8, <=2016.5.29", 74 "hyperframe>=4.0.1, <5", 75 "jsbeautifier>=1.6.3, <1.7", 76 "lxml>=3.5.0, <=3.6.0", # no wheels for 3.6.1 yet. 77 "Pillow>=3.2, <3.4", 78 "passlib>=1.6.5, <1.7", 79 "pyasn1>=0.1.9, <0.2", 80 "pyOpenSSL>=16.0, <17.0", 81 "pyparsing>=2.1.3, <2.2", 82 "pyperclip>=1.5.22, <1.6", 83 "requests>=2.9.1, <2.12", 84 "six>=1.10, <1.11", 85 "tornado>=4.3, <4.5", 86 "urwid>=1.3.1, <1.4", 87 "watchdog>=0.8.3, <0.9", 88 "brotlipy>=0.3.0, <0.5", 89 ], 90 extras_require={ 91 ':sys_platform == "win32"': [ 92 "pydivert>=0.0.7, <0.1", 93 ], 94 ':sys_platform != "win32"': [ 95 ], 96 # Do not use a range operator here: https://bitbucket.org/pypa/setuptools/issues/380 97 # Ubuntu Trusty and other still ship with setuptools < 17.1 98 ':python_version == "2.7"': [ 99 "enum34>=1.0.4, <2", 100 "ipaddress>=1.0.15, <1.1", 101 "typing==3.5.2.2", 102 ], 103 'dev': [ 104 "tox>=2.3, <3", 105 "mock>=2.0, <2.1", 106 "pytest>=2.8.7, <3", 107 "pytest-cov>=2.2.1, <3", 108 "pytest-timeout>=1.0.0, <2", 109 "pytest-xdist>=1.14, <2", 110 "sphinx>=1.3.5, <1.5", 111 "sphinx-autobuild>=0.5.2, <0.7", 112 "sphinxcontrib-documentedlist>=0.4.0, <0.5", 113 "sphinx_rtd_theme>=0.1.9, <0.2", 114 ], 115 'contentviews': [ 116 # TODO: Find Python 3 replacements 117 # "protobuf>=2.6.1, <2.7", 118 # "pyamf>=0.8.0, <0.9", 119 ], 120 'examples': [ 121 "beautifulsoup4>=4.4.1, <4.6", 122 "pytz>=2015.07.0, <=2016.6.1", 123 ] 124 } 125 ) 126 [end of setup.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/setup.py b/setup.py --- a/setup.py +++ b/setup.py @@ -66,7 +66,7 @@ "certifi>=2015.11.20.1", # no semver here - this should always be on the last release! "configargparse>=0.10, <0.11", "construct>=2.5.2, <2.6", - "cryptography>=1.3, <1.5", + "cryptography>=1.3, <1.6", "cssutils>=1.0.1, <1.1", "Flask>=0.10.1, <0.12", "h2>=2.4.1, <3",
{"golden_diff": "diff --git a/setup.py b/setup.py\n--- a/setup.py\n+++ b/setup.py\n@@ -66,7 +66,7 @@\n \"certifi>=2015.11.20.1\", # no semver here - this should always be on the last release!\n \"configargparse>=0.10, <0.11\",\n \"construct>=2.5.2, <2.6\",\n- \"cryptography>=1.3, <1.5\",\n+ \"cryptography>=1.3, <1.6\",\n \"cssutils>=1.0.1, <1.1\",\n \"Flask>=0.10.1, <0.12\",\n \"h2>=2.4.1, <3\",\n", "issue": "Wait for new cryptography release\nFor the next mitmproxy release, we want to recommend Python 3.5 by default. This requires a cryptography release which includes https://github.com/pyca/cryptography/pull/3063.\n\n", "before_files": [{"content": "from setuptools import setup, find_packages\nfrom codecs import open\nimport os\n\nfrom netlib import version\n\n# Based on https://github.com/pypa/sampleproject/blob/master/setup.py\n# and https://python-packaging-user-guide.readthedocs.org/\n\nhere = os.path.abspath(os.path.dirname(__file__))\n\nwith open(os.path.join(here, 'README.rst'), encoding='utf-8') as f:\n long_description = f.read()\n\nsetup(\n name=\"mitmproxy\",\n version=version.VERSION,\n description=\"An interactive, SSL-capable, man-in-the-middle HTTP proxy for penetration testers and software developers.\",\n long_description=long_description,\n url=\"http://mitmproxy.org\",\n author=\"Aldo Cortesi\",\n author_email=\"[email protected]\",\n license=\"MIT\",\n classifiers=[\n \"License :: OSI Approved :: MIT License\",\n \"Development Status :: 5 - Production/Stable\",\n \"Environment :: Console\",\n \"Environment :: Console :: Curses\",\n \"Operating System :: MacOS :: MacOS X\",\n \"Operating System :: POSIX\",\n \"Operating System :: Microsoft :: Windows\",\n \"Programming Language :: Python\",\n \"Programming Language :: Python :: 2\",\n \"Programming Language :: Python :: 2.7\",\n \"Programming Language :: Python :: 3\",\n \"Programming Language :: Python :: 3.5\",\n \"Programming Language :: Python :: Implementation :: CPython\",\n \"Programming Language :: Python :: Implementation :: PyPy\",\n \"Topic :: Security\",\n \"Topic :: Internet\",\n \"Topic :: Internet :: WWW/HTTP\",\n \"Topic :: Internet :: Proxy Servers\",\n \"Topic :: Software Development :: Testing\"\n ],\n packages=find_packages(include=[\n \"mitmproxy\", \"mitmproxy.*\",\n \"pathod\", \"pathod.*\",\n \"netlib\", \"netlib.*\"\n ]),\n include_package_data=True,\n entry_points={\n 'console_scripts': [\n \"mitmproxy = mitmproxy.main:mitmproxy\",\n \"mitmdump = mitmproxy.main:mitmdump\",\n \"mitmweb = mitmproxy.main:mitmweb\",\n \"pathod = pathod.pathod_cmdline:go_pathod\",\n \"pathoc = pathod.pathoc_cmdline:go_pathoc\"\n ]\n },\n # https://packaging.python.org/en/latest/requirements/#install-requires\n # It is not considered best practice to use install_requires to pin dependencies to specific versions.\n install_requires=[\n \"backports.ssl_match_hostname>=3.5.0.1, <3.6\",\n \"blinker>=1.4, <1.5\",\n \"click>=6.2, <7.0\",\n \"certifi>=2015.11.20.1\", # no semver here - this should always be on the last release!\n \"configargparse>=0.10, <0.11\",\n \"construct>=2.5.2, <2.6\",\n \"cryptography>=1.3, <1.5\",\n \"cssutils>=1.0.1, <1.1\",\n \"Flask>=0.10.1, <0.12\",\n \"h2>=2.4.1, <3\",\n \"html2text>=2016.1.8, <=2016.5.29\",\n \"hyperframe>=4.0.1, <5\",\n \"jsbeautifier>=1.6.3, <1.7\",\n \"lxml>=3.5.0, <=3.6.0\", # no wheels for 3.6.1 yet.\n \"Pillow>=3.2, <3.4\",\n \"passlib>=1.6.5, <1.7\",\n \"pyasn1>=0.1.9, <0.2\",\n \"pyOpenSSL>=16.0, <17.0\",\n \"pyparsing>=2.1.3, <2.2\",\n \"pyperclip>=1.5.22, <1.6\",\n \"requests>=2.9.1, <2.12\",\n \"six>=1.10, <1.11\",\n \"tornado>=4.3, <4.5\",\n \"urwid>=1.3.1, <1.4\",\n \"watchdog>=0.8.3, <0.9\",\n \"brotlipy>=0.3.0, <0.5\",\n ],\n extras_require={\n ':sys_platform == \"win32\"': [\n \"pydivert>=0.0.7, <0.1\",\n ],\n ':sys_platform != \"win32\"': [\n ],\n # Do not use a range operator here: https://bitbucket.org/pypa/setuptools/issues/380\n # Ubuntu Trusty and other still ship with setuptools < 17.1\n ':python_version == \"2.7\"': [\n \"enum34>=1.0.4, <2\",\n \"ipaddress>=1.0.15, <1.1\",\n \"typing==3.5.2.2\",\n ],\n 'dev': [\n \"tox>=2.3, <3\",\n \"mock>=2.0, <2.1\",\n \"pytest>=2.8.7, <3\",\n \"pytest-cov>=2.2.1, <3\",\n \"pytest-timeout>=1.0.0, <2\",\n \"pytest-xdist>=1.14, <2\",\n \"sphinx>=1.3.5, <1.5\",\n \"sphinx-autobuild>=0.5.2, <0.7\",\n \"sphinxcontrib-documentedlist>=0.4.0, <0.5\",\n \"sphinx_rtd_theme>=0.1.9, <0.2\",\n ],\n 'contentviews': [\n # TODO: Find Python 3 replacements\n # \"protobuf>=2.6.1, <2.7\",\n # \"pyamf>=0.8.0, <0.9\",\n ],\n 'examples': [\n \"beautifulsoup4>=4.4.1, <4.6\",\n \"pytz>=2015.07.0, <=2016.6.1\",\n ]\n }\n)\n", "path": "setup.py"}]}
2,263
176
gh_patches_debug_63301
rasdani/github-patches
git_diff
scikit-hep__pyhf-372
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> Update tensorflow-probability to the next release that includes continuous approximations # Description This is a follow up to #302. As the bug is fixed in upstream tensorflow-probability, we just need to wait for a new release to be shipped. This bug was because of a change in the API to get rid of the continuous approximation to the Poisson pmf which broke our tests. ### Describe the solution you'd like Unfix tensorflow-probability to `0.3.0` and bump to the next available release post-0.4.0. Update Tensorflow to TensorFlow 1.12.0 release # Description [TensorFlow 1.12.0 has been released](https://github.com/tensorflow/tensorflow/releases/tag/v1.12.0) and it has breaking changes. Most notably > Remove `tf.contrib.linalg`. `tf.linalg` should be used instead. Once there is a new release of TensorFlow probability (`v0.5.0` &mdash; c.f. Issue #360 and #330) that upgrades to `v1.12.0` then we can follow them in upgrading. </issue> <code> [start of setup.py] 1 #!/usr/bin/env python 2 3 from setuptools import setup, find_packages 4 from os import path 5 import sys 6 7 this_directory = path.abspath(path.dirname(__file__)) 8 if sys.version_info.major < 3: 9 from io import open 10 with open(path.join(this_directory, 'README.md'), encoding='utf-8') as readme_md: 11 long_description = readme_md.read() 12 13 extras_require = { 14 'tensorflow': [ 15 'tensorflow<1.12.0,>=1.10.0', 16 'tensorflow-probability==0.3.0', 17 'numpy<=1.14.5,>=1.14.0', # Lower of 1.14.0 instead of 1.13.3 to ensure doctest pass 18 'setuptools<=39.1.0', 19 ], 20 'torch': ['torch>=0.4.0'], 21 'mxnet': [ 22 'mxnet>=1.0.0', 23 'requests<2.19.0,>=2.18.4', 24 'numpy<1.15.0,>=1.8.2', 25 'requests<2.19.0,>=2.18.4', 26 ], 27 # 'dask': [ 28 # 'dask[array]' 29 # ], 30 'xmlimport': ['uproot'], 31 'minuit': ['iminuit'], 32 'develop': [ 33 'pyflakes', 34 'pytest<4.0.0,>=3.5.1', 35 'pytest-cov>=2.5.1', 36 'pytest-mock', 37 'pytest-benchmark[histogram]', 38 'pytest-console-scripts', 39 'python-coveralls', 40 'coverage>=4.0', # coveralls 41 'matplotlib', 42 'jupyter', 43 'nbdime', 44 'uproot>=3.0.0', 45 'papermill>=0.16.0', 46 'graphviz', 47 'bumpversion', 48 'sphinx', 49 'sphinxcontrib-bibtex', 50 'sphinxcontrib-napoleon', 51 'sphinx_rtd_theme', 52 'nbsphinx', 53 'sphinx-issues', 54 'm2r', 55 'jsonpatch', 56 'ipython<7', # jupyter_console and ipython clash in dependency requirement -- downgrade ipython for now 57 'pre-commit', 58 'black;python_version>="3.6"', # Black is Python3 only 59 'twine', 60 ], 61 } 62 extras_require['complete'] = sorted(set(sum(extras_require.values(), []))) 63 64 setup( 65 name='pyhf', 66 version='0.0.15', 67 description='(partial) pure python histfactory implementation', 68 long_description=long_description, 69 long_description_content_type='text/markdown', 70 url='https://github.com/diana-hep/pyhf', 71 author='Lukas Heinrich', 72 author_email='[email protected]', 73 license='Apache', 74 keywords='physics fitting numpy scipy tensorflow pytorch mxnet dask', 75 classifiers=[ 76 "Programming Language :: Python :: 2", 77 "Programming Language :: Python :: 2.7", 78 "Programming Language :: Python :: 3", 79 "Programming Language :: Python :: 3.6", 80 ], 81 packages=find_packages(), 82 include_package_data=True, 83 python_requires=">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*, !=3.5.*", 84 install_requires=[ 85 'scipy', # requires numpy, which is required by pyhf, tensorflow, and mxnet 86 'click>=6.0', # for console scripts, 87 'tqdm', # for readxml 88 'six', # for modifiers 89 'jsonschema>=v3.0.0a2', # for utils, alpha-release for draft 6 90 'jsonpatch', 91 ], 92 extras_require=extras_require, 93 entry_points={'console_scripts': ['pyhf=pyhf.commandline:pyhf']}, 94 dependency_links=[], 95 ) 96 [end of setup.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/setup.py b/setup.py --- a/setup.py +++ b/setup.py @@ -12,8 +12,8 @@ extras_require = { 'tensorflow': [ - 'tensorflow<1.12.0,>=1.10.0', - 'tensorflow-probability==0.3.0', + 'tensorflow>=1.12.0', + 'tensorflow-probability>=0.5.0', 'numpy<=1.14.5,>=1.14.0', # Lower of 1.14.0 instead of 1.13.3 to ensure doctest pass 'setuptools<=39.1.0', ],
{"golden_diff": "diff --git a/setup.py b/setup.py\n--- a/setup.py\n+++ b/setup.py\n@@ -12,8 +12,8 @@\n \n extras_require = {\n 'tensorflow': [\n- 'tensorflow<1.12.0,>=1.10.0',\n- 'tensorflow-probability==0.3.0',\n+ 'tensorflow>=1.12.0',\n+ 'tensorflow-probability>=0.5.0',\n 'numpy<=1.14.5,>=1.14.0', # Lower of 1.14.0 instead of 1.13.3 to ensure doctest pass\n 'setuptools<=39.1.0',\n ],\n", "issue": "Update tensorflow-probability to the next release that includes continuous approximations\n# Description\r\n\r\nThis is a follow up to #302. As the bug is fixed in upstream tensorflow-probability, we just need to wait for a new release to be shipped.\r\n\r\nThis bug was because of a change in the API to get rid of the continuous approximation to the Poisson pmf which broke our tests.\r\n\r\n### Describe the solution you'd like\r\n\r\nUnfix tensorflow-probability to `0.3.0` and bump to the next available release post-0.4.0.\nUpdate Tensorflow to TensorFlow 1.12.0 release\n# Description\r\n\r\n[TensorFlow 1.12.0 has been released](https://github.com/tensorflow/tensorflow/releases/tag/v1.12.0) and it has breaking changes. Most notably\r\n\r\n> Remove `tf.contrib.linalg`. `tf.linalg` should be used instead. \r\n\r\nOnce there is a new release of TensorFlow probability (`v0.5.0` &mdash; c.f. Issue #360 and #330) that upgrades to `v1.12.0` then we can follow them in upgrading.\r\n\n", "before_files": [{"content": "#!/usr/bin/env python\n\nfrom setuptools import setup, find_packages\nfrom os import path\nimport sys\n\nthis_directory = path.abspath(path.dirname(__file__))\nif sys.version_info.major < 3:\n from io import open\nwith open(path.join(this_directory, 'README.md'), encoding='utf-8') as readme_md:\n long_description = readme_md.read()\n\nextras_require = {\n 'tensorflow': [\n 'tensorflow<1.12.0,>=1.10.0',\n 'tensorflow-probability==0.3.0',\n 'numpy<=1.14.5,>=1.14.0', # Lower of 1.14.0 instead of 1.13.3 to ensure doctest pass\n 'setuptools<=39.1.0',\n ],\n 'torch': ['torch>=0.4.0'],\n 'mxnet': [\n 'mxnet>=1.0.0',\n 'requests<2.19.0,>=2.18.4',\n 'numpy<1.15.0,>=1.8.2',\n 'requests<2.19.0,>=2.18.4',\n ],\n # 'dask': [\n # 'dask[array]'\n # ],\n 'xmlimport': ['uproot'],\n 'minuit': ['iminuit'],\n 'develop': [\n 'pyflakes',\n 'pytest<4.0.0,>=3.5.1',\n 'pytest-cov>=2.5.1',\n 'pytest-mock',\n 'pytest-benchmark[histogram]',\n 'pytest-console-scripts',\n 'python-coveralls',\n 'coverage>=4.0', # coveralls\n 'matplotlib',\n 'jupyter',\n 'nbdime',\n 'uproot>=3.0.0',\n 'papermill>=0.16.0',\n 'graphviz',\n 'bumpversion',\n 'sphinx',\n 'sphinxcontrib-bibtex',\n 'sphinxcontrib-napoleon',\n 'sphinx_rtd_theme',\n 'nbsphinx',\n 'sphinx-issues',\n 'm2r',\n 'jsonpatch',\n 'ipython<7', # jupyter_console and ipython clash in dependency requirement -- downgrade ipython for now\n 'pre-commit',\n 'black;python_version>=\"3.6\"', # Black is Python3 only\n 'twine',\n ],\n}\nextras_require['complete'] = sorted(set(sum(extras_require.values(), [])))\n\nsetup(\n name='pyhf',\n version='0.0.15',\n description='(partial) pure python histfactory implementation',\n long_description=long_description,\n long_description_content_type='text/markdown',\n url='https://github.com/diana-hep/pyhf',\n author='Lukas Heinrich',\n author_email='[email protected]',\n license='Apache',\n keywords='physics fitting numpy scipy tensorflow pytorch mxnet dask',\n classifiers=[\n \"Programming Language :: Python :: 2\",\n \"Programming Language :: Python :: 2.7\",\n \"Programming Language :: Python :: 3\",\n \"Programming Language :: Python :: 3.6\",\n ],\n packages=find_packages(),\n include_package_data=True,\n python_requires=\">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*, !=3.5.*\",\n install_requires=[\n 'scipy', # requires numpy, which is required by pyhf, tensorflow, and mxnet\n 'click>=6.0', # for console scripts,\n 'tqdm', # for readxml\n 'six', # for modifiers\n 'jsonschema>=v3.0.0a2', # for utils, alpha-release for draft 6\n 'jsonpatch',\n ],\n extras_require=extras_require,\n entry_points={'console_scripts': ['pyhf=pyhf.commandline:pyhf']},\n dependency_links=[],\n)\n", "path": "setup.py"}]}
1,860
161
gh_patches_debug_33301
rasdani/github-patches
git_diff
ansible__ansible-lint-1747
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> command-instead-of-module "yum clean" revisited <!--- Verify first that your issue is not already reported on GitHub --> <!--- Also test if the latest release and main branch are affected too --> ##### Summary <!--- Explain the problem briefly below --> The lint error `command-instead-of-module` is raised for command `yum clean`. As the `yum` module documentation states there is no support for doing `yum clean` operation in the module. So the only option is to run a command. This issue has been raised in the past as #369 and `warn: false` was recommended. But that is now deprecated in 2.11. So shouldn't ansible-lint detect if `yum clean` is being run and not raise `command-instead-of-module`? If there is positive feedback and nobody else wants to do it then I can volunteer to make the patch - it should be a simple change, no? ##### Issue Type - Bug Report ##### Ansible and Ansible Lint details <!--- Paste verbatim output between triple backticks --> ```console (paste below) ansible --version ansible [core 2.11.5] config file = /etc/ansible/ansible.cfg configured module search path = ['/usr/share/ansible/plugins/modules'] ansible python module location = /opt/ansible/lib64/python3.6/site-packages/ansible ansible collection location = /usr/share/ansible/collections executable location = /opt/ansible/bin/ansible python version = 3.6.8 (default, Nov 16 2020, 16:55:22) [GCC 4.8.5 20150623 (Red Hat 4.8.5-44)] jinja version = 3.0.2 libyaml = True ansible-lint --version ansible-lint 5.2.0 using ansible 2.11.5 ``` - ansible installation method: pip - ansible-lint installation method: pip ##### OS / ENVIRONMENT <!--- Provide all relevant information below, e.g. target OS versions, network device firmware, etc. --> EL7 ##### STEPS TO REPRODUCE <!--- Describe exactly how to reproduce the problem, using a minimal test-case --> Have a command that runs `yum clean`. <!--- Paste example playbooks or commands between triple backticks below --> ```yaml command: cmd: yum clean all ``` <!--- HINT: You can paste gist.github.com links for larger files --> ##### Desired Behaviour <!--- Describe what you expected to happen when running the steps above --> `command-instead-of-module` should not be raised for `yum clean` which cannot be done using the `yum` module. Possible security bugs should be reported via email to `[email protected]` ##### Actual Behaviour <!--- Describe what actually happened. If possible run with extra verbosity (-vvvv) --> `command-instead-of-module` is raised even though the module cannot do this action. <!--- Paste verbatim command output between triple backticks --> ```paste below ``` [minimum complete verifiable example]: http://stackoverflow.com/help/mcve </issue> <code> [start of src/ansiblelint/rules/CommandsInsteadOfModulesRule.py] 1 # Copyright (c) 2013-2014 Will Thames <[email protected]> 2 # 3 # Permission is hereby granted, free of charge, to any person obtaining a copy 4 # of this software and associated documentation files (the "Software"), to deal 5 # in the Software without restriction, including without limitation the rights 6 # to use, copy, modify, merge, publish, distribute, sublicense, and/or sell 7 # copies of the Software, and to permit persons to whom the Software is 8 # furnished to do so, subject to the following conditions: 9 # 10 # The above copyright notice and this permission notice shall be included in 11 # all copies or substantial portions of the Software. 12 # 13 # THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR 14 # IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, 15 # FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE 16 # AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER 17 # LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, 18 # OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN 19 # THE SOFTWARE. 20 21 import os 22 import sys 23 from typing import TYPE_CHECKING, Any, Dict, Union 24 25 from ansiblelint.rules import AnsibleLintRule 26 from ansiblelint.utils import convert_to_boolean, get_first_cmd_arg, get_second_cmd_arg 27 28 if TYPE_CHECKING: 29 from typing import Optional 30 31 from ansiblelint.file_utils import Lintable 32 33 34 class CommandsInsteadOfModulesRule(AnsibleLintRule): 35 id = 'command-instead-of-module' 36 shortdesc = 'Using command rather than module' 37 description = ( 38 'Executing a command when there is an Ansible module is generally a bad idea' 39 ) 40 severity = 'HIGH' 41 tags = ['command-shell', 'idiom'] 42 version_added = 'historic' 43 44 _commands = ['command', 'shell'] 45 _modules = { 46 'apt-get': 'apt-get', 47 'chkconfig': 'service', 48 'curl': 'get_url or uri', 49 'git': 'git', 50 'hg': 'hg', 51 'letsencrypt': 'acme_certificate', 52 'mktemp': 'tempfile', 53 'mount': 'mount', 54 'patch': 'patch', 55 'rpm': 'yum or rpm_key', 56 'rsync': 'synchronize', 57 'sed': 'template, replace or lineinfile', 58 'service': 'service', 59 'supervisorctl': 'supervisorctl', 60 'svn': 'subversion', 61 'systemctl': 'systemd', 62 'tar': 'unarchive', 63 'unzip': 'unarchive', 64 'wget': 'get_url or uri', 65 'yum': 'yum', 66 } 67 68 _executable_options = { 69 'git': ['branch', 'log'], 70 'systemctl': ['set-default', 'show-environment', 'status'], 71 } 72 73 def matchtask( 74 self, task: Dict[str, Any], file: 'Optional[Lintable]' = None 75 ) -> Union[bool, str]: 76 77 if task['action']['__ansible_module__'] not in self._commands: 78 return False 79 80 first_cmd_arg = get_first_cmd_arg(task) 81 second_cmd_arg = get_second_cmd_arg(task) 82 83 if not first_cmd_arg: 84 return False 85 86 executable = os.path.basename(first_cmd_arg) 87 88 if ( 89 second_cmd_arg 90 and executable in self._executable_options 91 and second_cmd_arg in self._executable_options[executable] 92 ): 93 return False 94 95 if executable in self._modules and convert_to_boolean( 96 task['action'].get('warn', True) 97 ): 98 message = '{0} used in place of {1} module' 99 return message.format(executable, self._modules[executable]) 100 return False 101 102 103 if "pytest" in sys.modules: 104 import pytest 105 106 from ansiblelint.testing import RunFromText # pylint: disable=ungrouped-imports 107 108 APT_GET = ''' 109 - hosts: all 110 tasks: 111 - name: run apt-get update 112 command: apt-get update 113 ''' 114 115 GIT_BRANCH = ''' 116 - hosts: all 117 tasks: 118 - name: print current git branch 119 command: git branch 120 ''' 121 122 GIT_LOG = ''' 123 - hosts: all 124 tasks: 125 - name: print git log 126 command: git log 127 ''' 128 129 RESTART_SSHD = ''' 130 - hosts: all 131 tasks: 132 - name: restart sshd 133 command: systemctl restart sshd 134 ''' 135 136 SYSTEMCTL_STATUS = ''' 137 - hosts: all 138 tasks: 139 - name: show systemctl service status 140 command: systemctl status systemd-timesyncd 141 ''' 142 143 SYSTEMD_ENVIRONMENT = ''' 144 - hosts: all 145 tasks: 146 - name: show systemd environment 147 command: systemctl show-environment 148 ''' 149 150 SYSTEMD_RUNLEVEL = ''' 151 - hosts: all 152 tasks: 153 - name: set systemd runlevel 154 command: systemctl set-default multi-user.target 155 ''' 156 157 @pytest.mark.parametrize( 158 'rule_runner', (CommandsInsteadOfModulesRule,), indirect=['rule_runner'] 159 ) 160 def test_apt_get(rule_runner: RunFromText) -> None: 161 """The apt module supports update.""" 162 results = rule_runner.run_playbook(APT_GET) 163 assert len(results) == 1 164 165 @pytest.mark.parametrize( 166 'rule_runner', (CommandsInsteadOfModulesRule,), indirect=['rule_runner'] 167 ) 168 def test_restart_sshd(rule_runner: RunFromText) -> None: 169 """Restarting services is supported by the systemd module.""" 170 results = rule_runner.run_playbook(RESTART_SSHD) 171 assert len(results) == 1 172 173 @pytest.mark.parametrize( 174 'rule_runner', (CommandsInsteadOfModulesRule,), indirect=['rule_runner'] 175 ) 176 def test_git_log(rule_runner: RunFromText) -> None: 177 """The git log command is not supported by the git module.""" 178 results = rule_runner.run_playbook(GIT_LOG) 179 assert len(results) == 0 180 181 @pytest.mark.parametrize( 182 'rule_runner', (CommandsInsteadOfModulesRule,), indirect=['rule_runner'] 183 ) 184 def test_git_branch(rule_runner: RunFromText) -> None: 185 """The git branch command is not supported by the git module.""" 186 results = rule_runner.run_playbook(GIT_BRANCH) 187 assert len(results) == 0 188 189 @pytest.mark.parametrize( 190 'rule_runner', (CommandsInsteadOfModulesRule,), indirect=['rule_runner'] 191 ) 192 def test_systemd_status(rule_runner: RunFromText) -> None: 193 """Set-default is not supported by the systemd module.""" 194 results = rule_runner.run_playbook(SYSTEMCTL_STATUS) 195 assert len(results) == 0 196 197 @pytest.mark.parametrize( 198 'rule_runner', (CommandsInsteadOfModulesRule,), indirect=['rule_runner'] 199 ) 200 def test_systemd_environment(rule_runner: RunFromText) -> None: 201 """Showing the environment is not supported by the systemd module.""" 202 results = rule_runner.run_playbook(SYSTEMD_ENVIRONMENT) 203 assert len(results) == 0 204 205 @pytest.mark.parametrize( 206 'rule_runner', (CommandsInsteadOfModulesRule,), indirect=['rule_runner'] 207 ) 208 def test_systemd_runlevel(rule_runner: RunFromText) -> None: 209 """Set-default is not supported by the systemd module.""" 210 results = rule_runner.run_playbook(SYSTEMD_RUNLEVEL) 211 assert len(results) == 0 212 [end of src/ansiblelint/rules/CommandsInsteadOfModulesRule.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/src/ansiblelint/rules/CommandsInsteadOfModulesRule.py b/src/ansiblelint/rules/CommandsInsteadOfModulesRule.py --- a/src/ansiblelint/rules/CommandsInsteadOfModulesRule.py +++ b/src/ansiblelint/rules/CommandsInsteadOfModulesRule.py @@ -68,6 +68,7 @@ _executable_options = { 'git': ['branch', 'log'], 'systemctl': ['set-default', 'show-environment', 'status'], + 'yum': ['clean'], } def matchtask( @@ -100,7 +101,7 @@ return False -if "pytest" in sys.modules: +if "pytest" in sys.modules: # noqa: C901 import pytest from ansiblelint.testing import RunFromText # pylint: disable=ungrouped-imports @@ -154,6 +155,20 @@ command: systemctl set-default multi-user.target ''' + YUM_UPDATE = ''' +- hosts: all + tasks: + - name: run yum update + command: yum update +''' + + YUM_CLEAN = ''' +- hosts: all + tasks: + - name: clear yum cache + command: yum clean all +''' + @pytest.mark.parametrize( 'rule_runner', (CommandsInsteadOfModulesRule,), indirect=['rule_runner'] ) @@ -209,3 +224,19 @@ """Set-default is not supported by the systemd module.""" results = rule_runner.run_playbook(SYSTEMD_RUNLEVEL) assert len(results) == 0 + + @pytest.mark.parametrize( + 'rule_runner', (CommandsInsteadOfModulesRule,), indirect=['rule_runner'] + ) + def test_yum_update(rule_runner: RunFromText) -> None: + """Using yum update should fail.""" + results = rule_runner.run_playbook(YUM_UPDATE) + assert len(results) == 1 + + @pytest.mark.parametrize( + 'rule_runner', (CommandsInsteadOfModulesRule,), indirect=['rule_runner'] + ) + def test_yum_clean(rule_runner: RunFromText) -> None: + """The yum module does not support clearing yum cache.""" + results = rule_runner.run_playbook(YUM_CLEAN) + assert len(results) == 0
{"golden_diff": "diff --git a/src/ansiblelint/rules/CommandsInsteadOfModulesRule.py b/src/ansiblelint/rules/CommandsInsteadOfModulesRule.py\n--- a/src/ansiblelint/rules/CommandsInsteadOfModulesRule.py\n+++ b/src/ansiblelint/rules/CommandsInsteadOfModulesRule.py\n@@ -68,6 +68,7 @@\n _executable_options = {\n 'git': ['branch', 'log'],\n 'systemctl': ['set-default', 'show-environment', 'status'],\n+ 'yum': ['clean'],\n }\n \n def matchtask(\n@@ -100,7 +101,7 @@\n return False\n \n \n-if \"pytest\" in sys.modules:\n+if \"pytest\" in sys.modules: # noqa: C901\n import pytest\n \n from ansiblelint.testing import RunFromText # pylint: disable=ungrouped-imports\n@@ -154,6 +155,20 @@\n command: systemctl set-default multi-user.target\n '''\n \n+ YUM_UPDATE = '''\n+- hosts: all\n+ tasks:\n+ - name: run yum update\n+ command: yum update\n+'''\n+\n+ YUM_CLEAN = '''\n+- hosts: all\n+ tasks:\n+ - name: clear yum cache\n+ command: yum clean all\n+'''\n+\n @pytest.mark.parametrize(\n 'rule_runner', (CommandsInsteadOfModulesRule,), indirect=['rule_runner']\n )\n@@ -209,3 +224,19 @@\n \"\"\"Set-default is not supported by the systemd module.\"\"\"\n results = rule_runner.run_playbook(SYSTEMD_RUNLEVEL)\n assert len(results) == 0\n+\n+ @pytest.mark.parametrize(\n+ 'rule_runner', (CommandsInsteadOfModulesRule,), indirect=['rule_runner']\n+ )\n+ def test_yum_update(rule_runner: RunFromText) -> None:\n+ \"\"\"Using yum update should fail.\"\"\"\n+ results = rule_runner.run_playbook(YUM_UPDATE)\n+ assert len(results) == 1\n+\n+ @pytest.mark.parametrize(\n+ 'rule_runner', (CommandsInsteadOfModulesRule,), indirect=['rule_runner']\n+ )\n+ def test_yum_clean(rule_runner: RunFromText) -> None:\n+ \"\"\"The yum module does not support clearing yum cache.\"\"\"\n+ results = rule_runner.run_playbook(YUM_CLEAN)\n+ assert len(results) == 0\n", "issue": "command-instead-of-module \"yum clean\" revisited\n<!--- Verify first that your issue is not already reported on GitHub -->\r\n<!--- Also test if the latest release and main branch are affected too -->\r\n\r\n##### Summary\r\n<!--- Explain the problem briefly below -->\r\nThe lint error `command-instead-of-module` is raised for command `yum clean`. As the `yum` module documentation states there is no support for doing `yum clean` operation in the module. So the only option is to run a command.\r\n\r\nThis issue has been raised in the past as #369 and `warn: false` was recommended. But that is now deprecated in 2.11.\r\n\r\nSo shouldn't ansible-lint detect if `yum clean` is being run and not raise `command-instead-of-module`?\r\n\r\nIf there is positive feedback and nobody else wants to do it then I can volunteer to make the patch - it should be a simple change, no?\r\n\r\n##### Issue Type\r\n\r\n- Bug Report\r\n\r\n##### Ansible and Ansible Lint details\r\n<!--- Paste verbatim output between triple backticks -->\r\n```console (paste below)\r\nansible --version\r\nansible [core 2.11.5]\r\n config file = /etc/ansible/ansible.cfg\r\n configured module search path = ['/usr/share/ansible/plugins/modules']\r\n ansible python module location = /opt/ansible/lib64/python3.6/site-packages/ansible\r\n ansible collection location = /usr/share/ansible/collections\r\n executable location = /opt/ansible/bin/ansible\r\n python version = 3.6.8 (default, Nov 16 2020, 16:55:22) [GCC 4.8.5 20150623 (Red Hat 4.8.5-44)]\r\n jinja version = 3.0.2\r\n libyaml = True\r\n\r\nansible-lint --version\r\nansible-lint 5.2.0 using ansible 2.11.5\r\n\r\n```\r\n\r\n- ansible installation method: pip\r\n- ansible-lint installation method: pip\r\n\r\n##### OS / ENVIRONMENT\r\n<!--- Provide all relevant information below, e.g. target OS versions, network device firmware, etc. -->\r\nEL7\r\n\r\n##### STEPS TO REPRODUCE\r\n<!--- Describe exactly how to reproduce the problem, using a minimal test-case -->\r\nHave a command that runs `yum clean`.\r\n\r\n<!--- Paste example playbooks or commands between triple backticks below -->\r\n```yaml\r\n command:\r\n cmd: yum clean all\r\n```\r\n\r\n<!--- HINT: You can paste gist.github.com links for larger files -->\r\n\r\n##### Desired Behaviour\r\n<!--- Describe what you expected to happen when running the steps above -->\r\n`command-instead-of-module` should not be raised for `yum clean` which cannot be done using the `yum` module.\r\n\r\nPossible security bugs should be reported via email to `[email protected]`\r\n\r\n##### Actual Behaviour\r\n<!--- Describe what actually happened. If possible run with extra verbosity (-vvvv) -->\r\n`command-instead-of-module` is raised even though the module cannot do this action.\r\n\r\n<!--- Paste verbatim command output between triple backticks -->\r\n```paste below\r\n\r\n```\r\n\r\n\r\n[minimum complete verifiable example]: http://stackoverflow.com/help/mcve\r\n\n", "before_files": [{"content": "# Copyright (c) 2013-2014 Will Thames <[email protected]>\n#\n# Permission is hereby granted, free of charge, to any person obtaining a copy\n# of this software and associated documentation files (the \"Software\"), to deal\n# in the Software without restriction, including without limitation the rights\n# to use, copy, modify, merge, publish, distribute, sublicense, and/or sell\n# copies of the Software, and to permit persons to whom the Software is\n# furnished to do so, subject to the following conditions:\n#\n# The above copyright notice and this permission notice shall be included in\n# all copies or substantial portions of the Software.\n#\n# THE SOFTWARE IS PROVIDED \"AS IS\", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR\n# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,\n# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE\n# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER\n# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,\n# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN\n# THE SOFTWARE.\n\nimport os\nimport sys\nfrom typing import TYPE_CHECKING, Any, Dict, Union\n\nfrom ansiblelint.rules import AnsibleLintRule\nfrom ansiblelint.utils import convert_to_boolean, get_first_cmd_arg, get_second_cmd_arg\n\nif TYPE_CHECKING:\n from typing import Optional\n\n from ansiblelint.file_utils import Lintable\n\n\nclass CommandsInsteadOfModulesRule(AnsibleLintRule):\n id = 'command-instead-of-module'\n shortdesc = 'Using command rather than module'\n description = (\n 'Executing a command when there is an Ansible module is generally a bad idea'\n )\n severity = 'HIGH'\n tags = ['command-shell', 'idiom']\n version_added = 'historic'\n\n _commands = ['command', 'shell']\n _modules = {\n 'apt-get': 'apt-get',\n 'chkconfig': 'service',\n 'curl': 'get_url or uri',\n 'git': 'git',\n 'hg': 'hg',\n 'letsencrypt': 'acme_certificate',\n 'mktemp': 'tempfile',\n 'mount': 'mount',\n 'patch': 'patch',\n 'rpm': 'yum or rpm_key',\n 'rsync': 'synchronize',\n 'sed': 'template, replace or lineinfile',\n 'service': 'service',\n 'supervisorctl': 'supervisorctl',\n 'svn': 'subversion',\n 'systemctl': 'systemd',\n 'tar': 'unarchive',\n 'unzip': 'unarchive',\n 'wget': 'get_url or uri',\n 'yum': 'yum',\n }\n\n _executable_options = {\n 'git': ['branch', 'log'],\n 'systemctl': ['set-default', 'show-environment', 'status'],\n }\n\n def matchtask(\n self, task: Dict[str, Any], file: 'Optional[Lintable]' = None\n ) -> Union[bool, str]:\n\n if task['action']['__ansible_module__'] not in self._commands:\n return False\n\n first_cmd_arg = get_first_cmd_arg(task)\n second_cmd_arg = get_second_cmd_arg(task)\n\n if not first_cmd_arg:\n return False\n\n executable = os.path.basename(first_cmd_arg)\n\n if (\n second_cmd_arg\n and executable in self._executable_options\n and second_cmd_arg in self._executable_options[executable]\n ):\n return False\n\n if executable in self._modules and convert_to_boolean(\n task['action'].get('warn', True)\n ):\n message = '{0} used in place of {1} module'\n return message.format(executable, self._modules[executable])\n return False\n\n\nif \"pytest\" in sys.modules:\n import pytest\n\n from ansiblelint.testing import RunFromText # pylint: disable=ungrouped-imports\n\n APT_GET = '''\n- hosts: all\n tasks:\n - name: run apt-get update\n command: apt-get update\n'''\n\n GIT_BRANCH = '''\n- hosts: all\n tasks:\n - name: print current git branch\n command: git branch\n'''\n\n GIT_LOG = '''\n- hosts: all\n tasks:\n - name: print git log\n command: git log\n'''\n\n RESTART_SSHD = '''\n- hosts: all\n tasks:\n - name: restart sshd\n command: systemctl restart sshd\n'''\n\n SYSTEMCTL_STATUS = '''\n- hosts: all\n tasks:\n - name: show systemctl service status\n command: systemctl status systemd-timesyncd\n'''\n\n SYSTEMD_ENVIRONMENT = '''\n- hosts: all\n tasks:\n - name: show systemd environment\n command: systemctl show-environment\n'''\n\n SYSTEMD_RUNLEVEL = '''\n- hosts: all\n tasks:\n - name: set systemd runlevel\n command: systemctl set-default multi-user.target\n'''\n\n @pytest.mark.parametrize(\n 'rule_runner', (CommandsInsteadOfModulesRule,), indirect=['rule_runner']\n )\n def test_apt_get(rule_runner: RunFromText) -> None:\n \"\"\"The apt module supports update.\"\"\"\n results = rule_runner.run_playbook(APT_GET)\n assert len(results) == 1\n\n @pytest.mark.parametrize(\n 'rule_runner', (CommandsInsteadOfModulesRule,), indirect=['rule_runner']\n )\n def test_restart_sshd(rule_runner: RunFromText) -> None:\n \"\"\"Restarting services is supported by the systemd module.\"\"\"\n results = rule_runner.run_playbook(RESTART_SSHD)\n assert len(results) == 1\n\n @pytest.mark.parametrize(\n 'rule_runner', (CommandsInsteadOfModulesRule,), indirect=['rule_runner']\n )\n def test_git_log(rule_runner: RunFromText) -> None:\n \"\"\"The git log command is not supported by the git module.\"\"\"\n results = rule_runner.run_playbook(GIT_LOG)\n assert len(results) == 0\n\n @pytest.mark.parametrize(\n 'rule_runner', (CommandsInsteadOfModulesRule,), indirect=['rule_runner']\n )\n def test_git_branch(rule_runner: RunFromText) -> None:\n \"\"\"The git branch command is not supported by the git module.\"\"\"\n results = rule_runner.run_playbook(GIT_BRANCH)\n assert len(results) == 0\n\n @pytest.mark.parametrize(\n 'rule_runner', (CommandsInsteadOfModulesRule,), indirect=['rule_runner']\n )\n def test_systemd_status(rule_runner: RunFromText) -> None:\n \"\"\"Set-default is not supported by the systemd module.\"\"\"\n results = rule_runner.run_playbook(SYSTEMCTL_STATUS)\n assert len(results) == 0\n\n @pytest.mark.parametrize(\n 'rule_runner', (CommandsInsteadOfModulesRule,), indirect=['rule_runner']\n )\n def test_systemd_environment(rule_runner: RunFromText) -> None:\n \"\"\"Showing the environment is not supported by the systemd module.\"\"\"\n results = rule_runner.run_playbook(SYSTEMD_ENVIRONMENT)\n assert len(results) == 0\n\n @pytest.mark.parametrize(\n 'rule_runner', (CommandsInsteadOfModulesRule,), indirect=['rule_runner']\n )\n def test_systemd_runlevel(rule_runner: RunFromText) -> None:\n \"\"\"Set-default is not supported by the systemd module.\"\"\"\n results = rule_runner.run_playbook(SYSTEMD_RUNLEVEL)\n assert len(results) == 0\n", "path": "src/ansiblelint/rules/CommandsInsteadOfModulesRule.py"}]}
3,392
524
gh_patches_debug_2778
rasdani/github-patches
git_diff
pytorch__ignite-1324
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> Create a logo for Ignite ### 🚀 Feature @vfdev-5 asked me to work on a logo to represent Ignite and its relation to main PyTorch, so I’m opening this issue to post iterations and provide a place for discussion. All feedback is welcome! I'm looking forward to hearing the community's thoughts. These first iterations use the same orange as PyTorch (`#ee4c2c`)and maintain similar line weight (and sometimes the small circle) in order to show a visual relationship between the projects. The potential directions for Ignite’s logomark are meant to highlight Ignite’s strengths of simplicity and speed as well as the kick starting nature of some of its “out-of-the-box” features. ### Option 1-1 Compared to the current PyTorch logo, Option 1 is meant to be more a energetic spark than a steady, symmetrical flame. <img width="502" alt="1-1" src="https://user-images.githubusercontent.com/50221806/88605197-19e91480-d02e-11ea-9d97-5fc07286dc95.png"> <img width="502" alt="1-1color" src="https://user-images.githubusercontent.com/50221806/88605198-1a81ab00-d02e-11ea-8771-2f2713077df9.png"> ### Option 2-1 Option 2 is the furthest from the current PyTorch logo and revolves around the idea of a matchstick. The flame shape is closest to the current PyTorch logo of the different options. <img width="502" alt="2-1" src="https://user-images.githubusercontent.com/50221806/88605212-22d9e600-d02e-11ea-8678-eb0785fcd140.png"> <img width="502" alt="2-1color" src="https://user-images.githubusercontent.com/50221806/88605215-23727c80-d02e-11ea-89cf-b5ac451839ac.png"> ### Option 3-1 Another idea around a spark originating from a single point. Different flame shape and direction. <img width="502" alt="3-1" src="https://user-images.githubusercontent.com/50221806/88605221-2a998a80-d02e-11ea-85d0-8e9ac7b711c3.png"> <img width="502" alt="3-1color" src="https://user-images.githubusercontent.com/50221806/88605222-2a998a80-d02e-11ea-8372-87e1aa4eaca7.png"> ### Option 4-1 This developed as a variation of Option 3. While it is my least favorite, I wanted to provide an option that feels sharper and more explosive. <img width="502" alt="4-1" src="https://user-images.githubusercontent.com/50221806/88605232-2ff6d500-d02e-11ea-8078-e4d25c48d10b.png"> <img width="502" alt="4-1color" src="https://user-images.githubusercontent.com/50221806/88605234-308f6b80-d02e-11ea-99c9-8ba077afe688.png"> Update Circle CI docker image to pytorch 1.6 ## 🚀 Feature Following the release of 1.6 version of PyTorch, let's update Circle CI docker image : - https://github.com/pytorch/ignite/blob/master/.circleci/config.yml and add 1.5.1 version to pytorch compatibilty versions checking : - https://github.com/pytorch/ignite/blob/master/.github/workflows/pytorch-version-tests.yml </issue> <code> [start of docs/source/conf.py] 1 # -*- coding: utf-8 -*- 2 # 3 # Configuration file for the Sphinx documentation builder. 4 # 5 # This file does only contain a selection of the most common options. For a 6 # full list see the documentation: 7 # http://www.sphinx-doc.org/en/stable/config 8 9 # -- Path setup -------------------------------------------------------------- 10 11 # If extensions (or modules to document with autodoc) are in another directory, 12 # add these directories to sys.path here. If the directory is relative to the 13 # documentation root, use os.path.abspath to make it absolute, like shown here. 14 # 15 import os 16 import sys 17 18 sys.path.insert(0, os.path.abspath("../..")) 19 import ignite 20 import pytorch_sphinx_theme 21 22 # -- Project information ----------------------------------------------------- 23 24 project = "ignite" 25 copyright = "2020, PyTorch-Ignite Contributors" 26 author = "PyTorch-Ignite Contributors" 27 28 # The short X.Y version 29 try: 30 version = os.environ["code_version"] 31 if "master" in version: 32 version = "master (" + ignite.__version__ + ")" 33 else: 34 version = version.replace("v", "") 35 except KeyError: 36 version = ignite.__version__ 37 38 # The full version, including alpha/beta/rc tags 39 release = "master" 40 41 42 # -- General configuration --------------------------------------------------- 43 44 # If your documentation needs a minimal Sphinx version, state it here. 45 # 46 # needs_sphinx = '1.0' 47 48 # Add any Sphinx extension module names here, as strings. They can be 49 # extensions coming with Sphinx (named 'sphinx.ext.*') or your custom 50 # ones. 51 extensions = [ 52 "sphinx.ext.autosummary", 53 "sphinx.ext.doctest", 54 "sphinx.ext.intersphinx", 55 "sphinx.ext.todo", 56 "sphinx.ext.coverage", 57 "sphinx.ext.mathjax", 58 "sphinx.ext.napoleon", 59 "sphinx.ext.viewcode", 60 "sphinx.ext.autosectionlabel", 61 ] 62 63 # Add any paths that contain templates here, relative to this directory. 64 templates_path = ["_templates"] 65 66 # The suffix(es) of source filenames. 67 # You can specify multiple suffix as a list of string: 68 # 69 # source_suffix = ['.rst', '.md'] 70 source_suffix = ".rst" 71 72 # The master toctree document. 73 master_doc = "index" 74 75 # The language for content autogenerated by Sphinx. Refer to documentation 76 # for a list of supported languages. 77 # 78 # This is also used if you do content translation via gettext catalogs. 79 # Usually you set "language" from the command line for these cases. 80 language = None 81 82 # List of patterns, relative to source directory, that match files and 83 # directories to ignore when looking for source files. 84 # This pattern also affects html_static_path and html_extra_path . 85 exclude_patterns = [] 86 87 # The name of the Pygments (syntax highlighting) style to use. 88 pygments_style = "sphinx" 89 90 91 # -- Options for HTML output ------------------------------------------------- 92 93 # The theme to use for HTML and HTML Help pages. See the documentation for 94 # a list of builtin themes. 95 # 96 html_theme = "pytorch_sphinx_theme" 97 html_theme_path = [pytorch_sphinx_theme.get_html_theme_path()] 98 99 html_theme_options = { 100 "canonical_url": "https://pytorch.org/ignite/index.html", 101 "collapse_navigation": False, 102 "display_version": True, 103 "logo_only": True, 104 } 105 106 html_logo = "_static/img/ignite-logo-dark.svg" 107 108 # Theme options are theme-specific and customize the look and feel of a theme 109 # further. For a list of options available for each theme, see the 110 # documentation. 111 # 112 # html_theme_options = {} 113 114 # Add any paths that contain custom static files (such as style sheets) here, 115 # relative to this directory. They are copied after the builtin static files, 116 # so a file named "default.css" will overwrite the builtin "default.css". 117 html_static_path = ["_static", "_templates/_static"] 118 119 html_context = { 120 "css_files": [ 121 # 'https://fonts.googleapis.com/css?family=Lato', 122 # '_static/css/pytorch_theme.css' 123 "_static/css/ignite_theme.css" 124 ], 125 } 126 127 128 # -- Options for HTMLHelp output --------------------------------------------- 129 130 # Output file base name for HTML help builder. 131 htmlhelp_basename = "ignitedoc" 132 133 134 # -- Options for LaTeX output ------------------------------------------------ 135 136 latex_elements = { 137 # The paper size ('letterpaper' or 'a4paper'). 138 # 139 # 'papersize': 'letterpaper', 140 # The font size ('10pt', '11pt' or '12pt'). 141 # 142 # 'pointsize': '10pt', 143 # Additional stuff for the LaTeX preamble. 144 # 145 # 'preamble': '', 146 # Latex figure (float) alignment 147 # 148 # 'figure_align': 'htbp', 149 } 150 151 # Grouping the document tree into LaTeX files. List of tuples 152 # (source start file, target name, title, 153 # author, documentclass [howto, manual, or own class]). 154 latex_documents = [ 155 (master_doc, "ignite.tex", "ignite Documentation", "Torch Contributors", "manual"), 156 ] 157 158 159 # -- Options for manual page output ------------------------------------------ 160 161 # One entry per manual page. List of tuples 162 # (source start file, name, description, authors, manual section). 163 man_pages = [(master_doc, "ignite", "ignite Documentation", [author], 1)] 164 165 166 # -- Options for Texinfo output ---------------------------------------------- 167 168 # Grouping the document tree into Texinfo files. List of tuples 169 # (source start file, target name, title, author, 170 # dir menu entry, description, category) 171 texinfo_documents = [ 172 ( 173 master_doc, 174 "ignite", 175 "ignite Documentation", 176 author, 177 "ignite", 178 "One line description of project.", 179 "Miscellaneous", 180 ), 181 ] 182 183 184 # -- Extension configuration ------------------------------------------------- 185 186 # -- Options for intersphinx extension --------------------------------------- 187 188 # Example configuration for intersphinx: refer to the Python standard library. 189 intersphinx_mapping = {"https://docs.python.org/": None} 190 191 # -- Options for todo extension ---------------------------------------------- 192 193 # If true, `todo` and `todoList` produce output, else they produce nothing. 194 todo_include_todos = True 195 [end of docs/source/conf.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/docs/source/conf.py b/docs/source/conf.py --- a/docs/source/conf.py +++ b/docs/source/conf.py @@ -103,7 +103,7 @@ "logo_only": True, } -html_logo = "_static/img/ignite-logo-dark.svg" +html_logo = "_static/img/ignite_logo.svg" # Theme options are theme-specific and customize the look and feel of a theme # further. For a list of options available for each theme, see the
{"golden_diff": "diff --git a/docs/source/conf.py b/docs/source/conf.py\n--- a/docs/source/conf.py\n+++ b/docs/source/conf.py\n@@ -103,7 +103,7 @@\n \"logo_only\": True,\n }\n \n-html_logo = \"_static/img/ignite-logo-dark.svg\"\n+html_logo = \"_static/img/ignite_logo.svg\"\n \n # Theme options are theme-specific and customize the look and feel of a theme\n # further. For a list of options available for each theme, see the\n", "issue": "Create a logo for Ignite\n### \ud83d\ude80 Feature\r\n\r\n@vfdev-5 asked me to work on a logo to represent Ignite and its relation to main PyTorch, so I\u2019m opening this issue to post iterations and provide a place for discussion. \r\n\r\nAll feedback is welcome! I'm looking forward to hearing the community's thoughts.\r\n\r\nThese first iterations use the same orange as PyTorch (`#ee4c2c`)and maintain similar line weight (and sometimes the small circle) in order to show a visual relationship between the projects. The potential directions for Ignite\u2019s logomark are meant to highlight Ignite\u2019s strengths of simplicity and speed as well as the kick starting nature of some of its \u201cout-of-the-box\u201d features.\r\n\r\n### Option 1-1\r\nCompared to the current PyTorch logo, Option 1 is meant to be more a energetic spark than a steady, symmetrical flame.\r\n<img width=\"502\" alt=\"1-1\" src=\"https://user-images.githubusercontent.com/50221806/88605197-19e91480-d02e-11ea-9d97-5fc07286dc95.png\">\r\n<img width=\"502\" alt=\"1-1color\" src=\"https://user-images.githubusercontent.com/50221806/88605198-1a81ab00-d02e-11ea-8771-2f2713077df9.png\">\r\n\r\n### Option 2-1\r\nOption 2 is the furthest from the current PyTorch logo and revolves around the idea of a matchstick. The flame shape is closest to the current PyTorch logo of the different options.\r\n<img width=\"502\" alt=\"2-1\" src=\"https://user-images.githubusercontent.com/50221806/88605212-22d9e600-d02e-11ea-8678-eb0785fcd140.png\">\r\n<img width=\"502\" alt=\"2-1color\" src=\"https://user-images.githubusercontent.com/50221806/88605215-23727c80-d02e-11ea-89cf-b5ac451839ac.png\">\r\n\r\n### Option 3-1\r\nAnother idea around a spark originating from a single point. Different flame shape and direction.\r\n<img width=\"502\" alt=\"3-1\" src=\"https://user-images.githubusercontent.com/50221806/88605221-2a998a80-d02e-11ea-85d0-8e9ac7b711c3.png\">\r\n<img width=\"502\" alt=\"3-1color\" src=\"https://user-images.githubusercontent.com/50221806/88605222-2a998a80-d02e-11ea-8372-87e1aa4eaca7.png\">\r\n\r\n### Option 4-1\r\nThis developed as a variation of Option 3. While it is my least favorite, I wanted to provide an option that feels sharper and more explosive.\r\n<img width=\"502\" alt=\"4-1\" src=\"https://user-images.githubusercontent.com/50221806/88605232-2ff6d500-d02e-11ea-8078-e4d25c48d10b.png\">\r\n<img width=\"502\" alt=\"4-1color\" src=\"https://user-images.githubusercontent.com/50221806/88605234-308f6b80-d02e-11ea-99c9-8ba077afe688.png\">\nUpdate Circle CI docker image to pytorch 1.6\n## \ud83d\ude80 Feature\r\n\r\nFollowing the release of 1.6 version of PyTorch, let's update Circle CI docker image :\r\n- https://github.com/pytorch/ignite/blob/master/.circleci/config.yml\r\nand add 1.5.1 version to pytorch compatibilty versions checking :\r\n- https://github.com/pytorch/ignite/blob/master/.github/workflows/pytorch-version-tests.yml\n", "before_files": [{"content": "# -*- coding: utf-8 -*-\n#\n# Configuration file for the Sphinx documentation builder.\n#\n# This file does only contain a selection of the most common options. For a\n# full list see the documentation:\n# http://www.sphinx-doc.org/en/stable/config\n\n# -- Path setup --------------------------------------------------------------\n\n# If extensions (or modules to document with autodoc) are in another directory,\n# add these directories to sys.path here. If the directory is relative to the\n# documentation root, use os.path.abspath to make it absolute, like shown here.\n#\nimport os\nimport sys\n\nsys.path.insert(0, os.path.abspath(\"../..\"))\nimport ignite\nimport pytorch_sphinx_theme\n\n# -- Project information -----------------------------------------------------\n\nproject = \"ignite\"\ncopyright = \"2020, PyTorch-Ignite Contributors\"\nauthor = \"PyTorch-Ignite Contributors\"\n\n# The short X.Y version\ntry:\n version = os.environ[\"code_version\"]\n if \"master\" in version:\n version = \"master (\" + ignite.__version__ + \")\"\n else:\n version = version.replace(\"v\", \"\")\nexcept KeyError:\n version = ignite.__version__\n\n# The full version, including alpha/beta/rc tags\nrelease = \"master\"\n\n\n# -- General configuration ---------------------------------------------------\n\n# If your documentation needs a minimal Sphinx version, state it here.\n#\n# needs_sphinx = '1.0'\n\n# Add any Sphinx extension module names here, as strings. They can be\n# extensions coming with Sphinx (named 'sphinx.ext.*') or your custom\n# ones.\nextensions = [\n \"sphinx.ext.autosummary\",\n \"sphinx.ext.doctest\",\n \"sphinx.ext.intersphinx\",\n \"sphinx.ext.todo\",\n \"sphinx.ext.coverage\",\n \"sphinx.ext.mathjax\",\n \"sphinx.ext.napoleon\",\n \"sphinx.ext.viewcode\",\n \"sphinx.ext.autosectionlabel\",\n]\n\n# Add any paths that contain templates here, relative to this directory.\ntemplates_path = [\"_templates\"]\n\n# The suffix(es) of source filenames.\n# You can specify multiple suffix as a list of string:\n#\n# source_suffix = ['.rst', '.md']\nsource_suffix = \".rst\"\n\n# The master toctree document.\nmaster_doc = \"index\"\n\n# The language for content autogenerated by Sphinx. Refer to documentation\n# for a list of supported languages.\n#\n# This is also used if you do content translation via gettext catalogs.\n# Usually you set \"language\" from the command line for these cases.\nlanguage = None\n\n# List of patterns, relative to source directory, that match files and\n# directories to ignore when looking for source files.\n# This pattern also affects html_static_path and html_extra_path .\nexclude_patterns = []\n\n# The name of the Pygments (syntax highlighting) style to use.\npygments_style = \"sphinx\"\n\n\n# -- Options for HTML output -------------------------------------------------\n\n# The theme to use for HTML and HTML Help pages. See the documentation for\n# a list of builtin themes.\n#\nhtml_theme = \"pytorch_sphinx_theme\"\nhtml_theme_path = [pytorch_sphinx_theme.get_html_theme_path()]\n\nhtml_theme_options = {\n \"canonical_url\": \"https://pytorch.org/ignite/index.html\",\n \"collapse_navigation\": False,\n \"display_version\": True,\n \"logo_only\": True,\n}\n\nhtml_logo = \"_static/img/ignite-logo-dark.svg\"\n\n# Theme options are theme-specific and customize the look and feel of a theme\n# further. For a list of options available for each theme, see the\n# documentation.\n#\n# html_theme_options = {}\n\n# Add any paths that contain custom static files (such as style sheets) here,\n# relative to this directory. They are copied after the builtin static files,\n# so a file named \"default.css\" will overwrite the builtin \"default.css\".\nhtml_static_path = [\"_static\", \"_templates/_static\"]\n\nhtml_context = {\n \"css_files\": [\n # 'https://fonts.googleapis.com/css?family=Lato',\n # '_static/css/pytorch_theme.css'\n \"_static/css/ignite_theme.css\"\n ],\n}\n\n\n# -- Options for HTMLHelp output ---------------------------------------------\n\n# Output file base name for HTML help builder.\nhtmlhelp_basename = \"ignitedoc\"\n\n\n# -- Options for LaTeX output ------------------------------------------------\n\nlatex_elements = {\n # The paper size ('letterpaper' or 'a4paper').\n #\n # 'papersize': 'letterpaper',\n # The font size ('10pt', '11pt' or '12pt').\n #\n # 'pointsize': '10pt',\n # Additional stuff for the LaTeX preamble.\n #\n # 'preamble': '',\n # Latex figure (float) alignment\n #\n # 'figure_align': 'htbp',\n}\n\n# Grouping the document tree into LaTeX files. List of tuples\n# (source start file, target name, title,\n# author, documentclass [howto, manual, or own class]).\nlatex_documents = [\n (master_doc, \"ignite.tex\", \"ignite Documentation\", \"Torch Contributors\", \"manual\"),\n]\n\n\n# -- Options for manual page output ------------------------------------------\n\n# One entry per manual page. List of tuples\n# (source start file, name, description, authors, manual section).\nman_pages = [(master_doc, \"ignite\", \"ignite Documentation\", [author], 1)]\n\n\n# -- Options for Texinfo output ----------------------------------------------\n\n# Grouping the document tree into Texinfo files. List of tuples\n# (source start file, target name, title, author,\n# dir menu entry, description, category)\ntexinfo_documents = [\n (\n master_doc,\n \"ignite\",\n \"ignite Documentation\",\n author,\n \"ignite\",\n \"One line description of project.\",\n \"Miscellaneous\",\n ),\n]\n\n\n# -- Extension configuration -------------------------------------------------\n\n# -- Options for intersphinx extension ---------------------------------------\n\n# Example configuration for intersphinx: refer to the Python standard library.\nintersphinx_mapping = {\"https://docs.python.org/\": None}\n\n# -- Options for todo extension ----------------------------------------------\n\n# If true, `todo` and `todoList` produce output, else they produce nothing.\ntodo_include_todos = True\n", "path": "docs/source/conf.py"}]}
3,340
109
gh_patches_debug_947
rasdani/github-patches
git_diff
databricks__koalas-105
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> How to call dev/_make_missing_functions.py? ``` rxin @ C02XT0W6JGH5 : ~/workspace/spark-pandas (master) > dev/_make_missing_functions.py Traceback (most recent call last): File "dev/_make_missing_functions.py", line 22, in <module> from databricks.koala.frame import PandasLikeDataFrame ImportError: No module named databricks.koala.frame ``` Do I need to install koala first? We should add documentation to CONTRIBUTING.md. It'd also be best if this runs against the existing code base, rather than a system-wide installed Koala. </issue> <code> [start of dev/_make_missing_functions.py] 1 #!/usr/bin/env python 2 # 3 # Copyright (C) 2019 Databricks, Inc. 4 # 5 # Licensed under the Apache License, Version 2.0 (the "License"); 6 # you may not use this file except in compliance with the License. 7 # You may obtain a copy of the License at 8 # 9 # http://www.apache.org/licenses/LICENSE-2.0 10 # 11 # Unless required by applicable law or agreed to in writing, software 12 # distributed under the License is distributed on an "AS IS" BASIS, 13 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 14 # See the License for the specific language governing permissions and 15 # limitations under the License. 16 # 17 18 import inspect 19 20 import pandas as pd 21 22 from databricks.koalas.frame import PandasLikeDataFrame 23 from databricks.koalas.missing.frame import _MissingPandasLikeDataFrame 24 from databricks.koalas.missing.series import _MissingPandasLikeSeries 25 from databricks.koalas.series import PandasLikeSeries 26 27 28 INDENT_LEN = 4 29 LINE_LEN_LIMIT = 100 30 31 32 def inspect_missing_functions(original_type, target_type, missing_type): 33 """ 34 Find functions which exist in original_type but not in target_type, 35 or the signature is modified. 36 37 :return: the tuple of the missing function name and its signature, 38 and the name of the functions the signature of which is different 39 and its original and modified signature. 40 """ 41 missing = [] 42 modified = [] 43 44 already_in_missing = set([(name, inspect.signature(func)) for name, func 45 in inspect.getmembers(missing_type, inspect.isfunction)]) 46 for name, func in inspect.getmembers(original_type, inspect.isfunction): 47 # Skip the private attributes 48 if name.startswith('_'): 49 continue 50 51 original_signature = inspect.signature(func, follow_wrapped=True) 52 53 if hasattr(target_type, name): 54 f = getattr(target_type, name) 55 if inspect.isfunction(f): 56 target_signature = inspect.signature(f) 57 if (name, target_signature) in already_in_missing: 58 missing.append((name, original_signature)) 59 elif str(original_signature) != str(target_signature): 60 modified.append((name, original_signature, target_signature)) 61 continue 62 63 missing.append((name, original_signature)) 64 65 return missing, modified 66 67 68 def format_arguments(arguments, prefix_len, suffix_len): 69 """Format arguments not to break pydocstyle. 70 71 :param arguments: the argument list 72 :param prefix_len: the prefix length when the argument string needs line break 73 :param suffix_len: the suffix length to check the line length exceeds the limit 74 :return: the formatted argument string 75 """ 76 lines = [''] 77 78 def append_arg(arg): 79 if prefix_len + len(lines[-1]) + len(', ') + len(arg) + suffix_len > LINE_LEN_LIMIT: 80 lines.append('') 81 append_arg(arg) 82 else: 83 if len(lines[-1]) > 0: 84 arg = ', {}'.format(arg) 85 lines[-1] += arg 86 87 for arg in arguments: 88 append_arg(arg) 89 90 return (',\n' + (' ' * prefix_len)).join(lines) 91 92 93 def format_method_arguments(name, signature): 94 """Format the method arguments from its name and signature. 95 96 :return: the formatted argument string 97 """ 98 arguments = [] 99 100 for param in signature.parameters.values(): 101 if param.default is not inspect.Signature.empty and isinstance(param.default, type): 102 arguments.append('{}={}'.format(param.name, param.default.__name__)) 103 elif param.default is not inspect.Signature.empty and repr(param.default) == 'nan': 104 arguments.append('{}={}'.format(param.name, 'np.nan')) 105 else: 106 arguments.append(str(param)) 107 108 prefix_len = INDENT_LEN + len('def {}('.format(name)) 109 suffix_len = len('):') 110 return format_arguments(arguments, prefix_len, suffix_len) 111 112 113 def format_derived_from(original_type, unavailable_arguments, signature): 114 """Format `@derived_from` decorator. 115 116 :param original_type: the original type to be derived 117 :param unavailable_arguments: the arguments Koalas does not support yet 118 :param signature: the method signature 119 :return: the formatted `@derived_from` decorator 120 """ 121 if len(unavailable_arguments) == 0: 122 return '@derived_from(pd.{})'.format(original_type.__name__) 123 124 arguments = [] 125 126 for arg in unavailable_arguments: 127 param = signature.parameters[arg] 128 if param.default == inspect.Parameter.empty or \ 129 param.kind == inspect.Parameter.VAR_POSITIONAL or \ 130 param.kind == inspect.Parameter.VAR_KEYWORD: 131 continue 132 arguments.append(repr(arg)) 133 134 prefix = '@derived_from(pd.{}, ua_args=['.format(original_type.__name__) 135 suffix = '])' 136 prefix_len = INDENT_LEN + len(prefix) 137 suffix_len = len(suffix) 138 return '{}{}{}'.format(prefix, format_arguments(arguments, prefix_len, suffix_len), suffix) 139 140 141 def format_raise_errors(original_type, name, unavailable_arguments, signature): 142 """ 143 Format raise error statements for unavailable arguments when specified the different value 144 from the default value. 145 146 :return: the formatted raise error statements 147 """ 148 raise_errors = '' 149 150 for arg in unavailable_arguments: 151 param = signature.parameters[arg] 152 if param.default == inspect.Parameter.empty or \ 153 param.kind == inspect.Parameter.VAR_POSITIONAL or \ 154 param.kind == inspect.Parameter.VAR_KEYWORD: 155 continue 156 if repr(param.default) == 'nan': 157 not_equal = 'not np.isnan({})'.format(arg) 158 elif isinstance(param.default, type): 159 not_equal = '{} is not {}'.format(arg, param.default.__name__) 160 elif param.default is None or \ 161 param.default is True or param.default is False: 162 not_equal = '{} is not {}'.format(arg, repr(param.default)) 163 else: 164 not_equal = '{} != {}'.format(arg, repr(param.default)) 165 166 raise_error_prefix = 'raise PandasNotImplementedError(' 167 raise_error_suffix = ')' 168 arguments = format_arguments( 169 arguments=["class_name='pd.{}'".format(original_type.__name__), 170 "method_name='{}'".format(name), 171 "arg_name='{}'".format(arg)], 172 prefix_len=(INDENT_LEN * 3 + len(raise_error_prefix)), 173 suffix_len=len(raise_error_suffix)) 174 raise_errors += (""" 175 if {0}: 176 {1}{2}{3}""".format(not_equal, raise_error_prefix, arguments, raise_error_suffix)) 177 178 return raise_errors 179 180 181 def make_misssing_function(original_type, name, signature): 182 """Make a missing functions stub. 183 184 :return: the stub definition for the missing function 185 """ 186 arguments = format_method_arguments(name, signature) 187 error_argument = format_arguments( 188 arguments=["class_name='pd.{}'".format(original_type.__name__), 189 "method_name='{}'".format(name)], 190 prefix_len=(8 + len('raise PandasNotImplementedError(')), 191 suffix_len=len(')')) 192 193 return (""" 194 def {0}({1}): 195 \"""A stub for the equivalent method to `pd.{2}.{0}()`. 196 197 The method `pd.{2}.{0}()` is not implemented yet. 198 \""" 199 raise PandasNotImplementedError({3})""" 200 .format(name, arguments, original_type.__name__, error_argument)) 201 202 203 def make_modified_function_def(original_type, name, original, target): 204 """Make the modified function definition. 205 206 :return: the definition for the modified function 207 """ 208 arguments = format_method_arguments(name, original) 209 argument_names = set(target.parameters) 210 unavailable_arguments = [p for p in original.parameters if p not in argument_names] 211 derived_from = format_derived_from(original_type, unavailable_arguments, original) 212 raise_error = format_raise_errors(original_type, name, unavailable_arguments, original) 213 return (""" 214 {0} 215 def {1}({2}):{3}""".format(derived_from, name, arguments, raise_error)) 216 217 218 def _main(): 219 for original_type, target_type, missing_type in \ 220 [(pd.DataFrame, PandasLikeDataFrame, _MissingPandasLikeDataFrame), 221 (pd.Series, PandasLikeSeries, _MissingPandasLikeSeries)]: 222 missing, modified = inspect_missing_functions(original_type, target_type, missing_type) 223 224 print('MISSING functions for {}'.format(original_type.__name__)) 225 for name, signature in missing: 226 # print(make_misssing_function(original_type, name, signature)) 227 print(""" {0} = unsupported_function('{0}')""".format(name)) 228 229 print() 230 print('MODIFIED functions for {}'.format(original_type.__name__)) 231 for name, original, target in modified: 232 print(make_modified_function_def(original_type, name, original, target)) 233 234 235 if __name__ == '__main__': 236 _main() 237 [end of dev/_make_missing_functions.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/dev/_make_missing_functions.py b/dev/_make_missing_functions.py --- a/dev/_make_missing_functions.py +++ b/dev/_make_missing_functions.py @@ -15,6 +15,12 @@ # limitations under the License. # +""" +A script to generate the missing function stubs. Before running this, +make sure you install koalas from the current checkout by running: +pip install -e . +""" + import inspect import pandas as pd
{"golden_diff": "diff --git a/dev/_make_missing_functions.py b/dev/_make_missing_functions.py\n--- a/dev/_make_missing_functions.py\n+++ b/dev/_make_missing_functions.py\n@@ -15,6 +15,12 @@\n # limitations under the License.\n #\n \n+\"\"\"\n+A script to generate the missing function stubs. Before running this,\n+make sure you install koalas from the current checkout by running:\n+pip install -e .\n+\"\"\"\n+\n import inspect\n \n import pandas as pd\n", "issue": "How to call dev/_make_missing_functions.py?\n```\r\nrxin @ C02XT0W6JGH5 : ~/workspace/spark-pandas (master) \r\n> dev/_make_missing_functions.py \r\nTraceback (most recent call last):\r\n File \"dev/_make_missing_functions.py\", line 22, in <module>\r\n from databricks.koala.frame import PandasLikeDataFrame\r\nImportError: No module named databricks.koala.frame\r\n```\r\n\r\nDo I need to install koala first? We should add documentation to CONTRIBUTING.md. It'd also be best if this runs against the existing code base, rather than a system-wide installed Koala.\r\n\n", "before_files": [{"content": "#!/usr/bin/env python\n#\n# Copyright (C) 2019 Databricks, Inc.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n#\n\nimport inspect\n\nimport pandas as pd\n\nfrom databricks.koalas.frame import PandasLikeDataFrame\nfrom databricks.koalas.missing.frame import _MissingPandasLikeDataFrame\nfrom databricks.koalas.missing.series import _MissingPandasLikeSeries\nfrom databricks.koalas.series import PandasLikeSeries\n\n\nINDENT_LEN = 4\nLINE_LEN_LIMIT = 100\n\n\ndef inspect_missing_functions(original_type, target_type, missing_type):\n \"\"\"\n Find functions which exist in original_type but not in target_type,\n or the signature is modified.\n\n :return: the tuple of the missing function name and its signature,\n and the name of the functions the signature of which is different\n and its original and modified signature.\n \"\"\"\n missing = []\n modified = []\n\n already_in_missing = set([(name, inspect.signature(func)) for name, func\n in inspect.getmembers(missing_type, inspect.isfunction)])\n for name, func in inspect.getmembers(original_type, inspect.isfunction):\n # Skip the private attributes\n if name.startswith('_'):\n continue\n\n original_signature = inspect.signature(func, follow_wrapped=True)\n\n if hasattr(target_type, name):\n f = getattr(target_type, name)\n if inspect.isfunction(f):\n target_signature = inspect.signature(f)\n if (name, target_signature) in already_in_missing:\n missing.append((name, original_signature))\n elif str(original_signature) != str(target_signature):\n modified.append((name, original_signature, target_signature))\n continue\n\n missing.append((name, original_signature))\n\n return missing, modified\n\n\ndef format_arguments(arguments, prefix_len, suffix_len):\n \"\"\"Format arguments not to break pydocstyle.\n\n :param arguments: the argument list\n :param prefix_len: the prefix length when the argument string needs line break\n :param suffix_len: the suffix length to check the line length exceeds the limit\n :return: the formatted argument string\n \"\"\"\n lines = ['']\n\n def append_arg(arg):\n if prefix_len + len(lines[-1]) + len(', ') + len(arg) + suffix_len > LINE_LEN_LIMIT:\n lines.append('')\n append_arg(arg)\n else:\n if len(lines[-1]) > 0:\n arg = ', {}'.format(arg)\n lines[-1] += arg\n\n for arg in arguments:\n append_arg(arg)\n\n return (',\\n' + (' ' * prefix_len)).join(lines)\n\n\ndef format_method_arguments(name, signature):\n \"\"\"Format the method arguments from its name and signature.\n\n :return: the formatted argument string\n \"\"\"\n arguments = []\n\n for param in signature.parameters.values():\n if param.default is not inspect.Signature.empty and isinstance(param.default, type):\n arguments.append('{}={}'.format(param.name, param.default.__name__))\n elif param.default is not inspect.Signature.empty and repr(param.default) == 'nan':\n arguments.append('{}={}'.format(param.name, 'np.nan'))\n else:\n arguments.append(str(param))\n\n prefix_len = INDENT_LEN + len('def {}('.format(name))\n suffix_len = len('):')\n return format_arguments(arguments, prefix_len, suffix_len)\n\n\ndef format_derived_from(original_type, unavailable_arguments, signature):\n \"\"\"Format `@derived_from` decorator.\n\n :param original_type: the original type to be derived\n :param unavailable_arguments: the arguments Koalas does not support yet\n :param signature: the method signature\n :return: the formatted `@derived_from` decorator\n \"\"\"\n if len(unavailable_arguments) == 0:\n return '@derived_from(pd.{})'.format(original_type.__name__)\n\n arguments = []\n\n for arg in unavailable_arguments:\n param = signature.parameters[arg]\n if param.default == inspect.Parameter.empty or \\\n param.kind == inspect.Parameter.VAR_POSITIONAL or \\\n param.kind == inspect.Parameter.VAR_KEYWORD:\n continue\n arguments.append(repr(arg))\n\n prefix = '@derived_from(pd.{}, ua_args=['.format(original_type.__name__)\n suffix = '])'\n prefix_len = INDENT_LEN + len(prefix)\n suffix_len = len(suffix)\n return '{}{}{}'.format(prefix, format_arguments(arguments, prefix_len, suffix_len), suffix)\n\n\ndef format_raise_errors(original_type, name, unavailable_arguments, signature):\n \"\"\"\n Format raise error statements for unavailable arguments when specified the different value\n from the default value.\n\n :return: the formatted raise error statements\n \"\"\"\n raise_errors = ''\n\n for arg in unavailable_arguments:\n param = signature.parameters[arg]\n if param.default == inspect.Parameter.empty or \\\n param.kind == inspect.Parameter.VAR_POSITIONAL or \\\n param.kind == inspect.Parameter.VAR_KEYWORD:\n continue\n if repr(param.default) == 'nan':\n not_equal = 'not np.isnan({})'.format(arg)\n elif isinstance(param.default, type):\n not_equal = '{} is not {}'.format(arg, param.default.__name__)\n elif param.default is None or \\\n param.default is True or param.default is False:\n not_equal = '{} is not {}'.format(arg, repr(param.default))\n else:\n not_equal = '{} != {}'.format(arg, repr(param.default))\n\n raise_error_prefix = 'raise PandasNotImplementedError('\n raise_error_suffix = ')'\n arguments = format_arguments(\n arguments=[\"class_name='pd.{}'\".format(original_type.__name__),\n \"method_name='{}'\".format(name),\n \"arg_name='{}'\".format(arg)],\n prefix_len=(INDENT_LEN * 3 + len(raise_error_prefix)),\n suffix_len=len(raise_error_suffix))\n raise_errors += (\"\"\"\n if {0}:\n {1}{2}{3}\"\"\".format(not_equal, raise_error_prefix, arguments, raise_error_suffix))\n\n return raise_errors\n\n\ndef make_misssing_function(original_type, name, signature):\n \"\"\"Make a missing functions stub.\n\n :return: the stub definition for the missing function\n \"\"\"\n arguments = format_method_arguments(name, signature)\n error_argument = format_arguments(\n arguments=[\"class_name='pd.{}'\".format(original_type.__name__),\n \"method_name='{}'\".format(name)],\n prefix_len=(8 + len('raise PandasNotImplementedError(')),\n suffix_len=len(')'))\n\n return (\"\"\"\n def {0}({1}):\n \\\"\"\"A stub for the equivalent method to `pd.{2}.{0}()`.\n\n The method `pd.{2}.{0}()` is not implemented yet.\n \\\"\"\"\n raise PandasNotImplementedError({3})\"\"\"\n .format(name, arguments, original_type.__name__, error_argument))\n\n\ndef make_modified_function_def(original_type, name, original, target):\n \"\"\"Make the modified function definition.\n\n :return: the definition for the modified function\n \"\"\"\n arguments = format_method_arguments(name, original)\n argument_names = set(target.parameters)\n unavailable_arguments = [p for p in original.parameters if p not in argument_names]\n derived_from = format_derived_from(original_type, unavailable_arguments, original)\n raise_error = format_raise_errors(original_type, name, unavailable_arguments, original)\n return (\"\"\"\n {0}\n def {1}({2}):{3}\"\"\".format(derived_from, name, arguments, raise_error))\n\n\ndef _main():\n for original_type, target_type, missing_type in \\\n [(pd.DataFrame, PandasLikeDataFrame, _MissingPandasLikeDataFrame),\n (pd.Series, PandasLikeSeries, _MissingPandasLikeSeries)]:\n missing, modified = inspect_missing_functions(original_type, target_type, missing_type)\n\n print('MISSING functions for {}'.format(original_type.__name__))\n for name, signature in missing:\n # print(make_misssing_function(original_type, name, signature))\n print(\"\"\" {0} = unsupported_function('{0}')\"\"\".format(name))\n\n print()\n print('MODIFIED functions for {}'.format(original_type.__name__))\n for name, original, target in modified:\n print(make_modified_function_def(original_type, name, original, target))\n\n\nif __name__ == '__main__':\n _main()\n", "path": "dev/_make_missing_functions.py"}]}
3,208
105
gh_patches_debug_14775
rasdani/github-patches
git_diff
hylang__hy-1122
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> Hy sets are really broken. ``` Hy (env-hy) C:\Users\ME\Code>hy hy 0.11.0 using CPython(v3.4.3:9b73f1c3e601) 3.4.3 on Windows => #{:a 'a} Traceback (most recent call last): File "C:\Python34\Scripts\env-hy\Scripts\hy-script.py", line 9, in <module> load_entry_point('hy==0.11.0', 'console_scripts', 'hy')() File "C:\Python34\Scripts\env-hy\lib\site-packages\hy\cmdline.py", line 341, in hy_main sys.exit(cmdline_handler("hy", sys.argv)) File "C:\Python34\Scripts\env-hy\lib\site-packages\hy\cmdline.py", line 336, in cmdline_handler return run_repl(spy=options.spy) File "C:\Python34\Scripts\env-hy\lib\site-packages\hy\cmdline.py", line 234, in run_repl os=platform.system() File "C:\Python34\Lib\code.py", line 234, in interact more = self.push(line) File "C:\Python34\Lib\code.py", line 256, in push more = self.runsource(source, self.filename) File "C:\Python34\Scripts\env-hy\lib\site-packages\hy\cmdline.py", line 93, in runsource tokens = tokenize(source) File "C:\Python34\Scripts\env-hy\lib\site-packages\hy\lex\__init__.py", line 33, in tokenize return parser.parse(lexer.lex(buf)) File "C:\Python34\Scripts\env-hy\lib\site-packages\rply\parser.py", line 23, in parse t, symstack, statestack, state File "C:\Python34\Scripts\env-hy\lib\site-packages\rply\parser.py", line 80, in _reduce_production value = p.func(targ) File "C:\Python34\Scripts\env-hy\lib\site-packages\hy\lex\parser.py", line 69, in wrapped ret = fun(p) File "C:\Python34\Scripts\env-hy\lib\site-packages\hy\lex\parser.py", line 214, in t_set return HySet(p[1]) File "C:\Python34\Scripts\env-hy\lib\site-packages\hy\models\set.py", line 31, in __init__ items = sorted(items) TypeError: unorderable types: HyExpression() < HyKeyword() (env-hy) C:\Users\ME\Code> ``` That is NOT supposed to happen. </issue> <code> [start of hy/models/set.py] 1 # Copyright (c) 2013 Paul Tagliamonte <[email protected]> 2 # 3 # Permission is hereby granted, free of charge, to any person obtaining a 4 # copy of this software and associated documentation files (the "Software"), 5 # to deal in the Software without restriction, including without limitation 6 # the rights to use, copy, modify, merge, publish, distribute, sublicense, 7 # and/or sell copies of the Software, and to permit persons to whom the 8 # Software is furnished to do so, subject to the following conditions: 9 # 10 # The above copyright notice and this permission notice shall be included in 11 # all copies or substantial portions of the Software. 12 # 13 # THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR 14 # IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, 15 # FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL 16 # THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER 17 # LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING 18 # FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER 19 # DEALINGS IN THE SOFTWARE. 20 21 from hy.models.list import HyList 22 from functools import reduce 23 24 25 class HySet(HyList): 26 """ 27 Hy set (actually a list that pretends to be a set) 28 """ 29 30 def __init__(self, items): 31 items = sorted(items) 32 items = list(reduce(lambda r, v: v in r and r or r+[v], items, [])) 33 super(HySet, self).__init__(items) 34 35 def __repr__(self): 36 return "#{%s}" % (" ".join([repr(x) for x in self])) 37 [end of hy/models/set.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/hy/models/set.py b/hy/models/set.py --- a/hy/models/set.py +++ b/hy/models/set.py @@ -18,19 +18,16 @@ # FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER # DEALINGS IN THE SOFTWARE. +from hy.models import _wrappers, wrap_value from hy.models.list import HyList -from functools import reduce class HySet(HyList): """ - Hy set (actually a list that pretends to be a set) + Hy set (just a representation of a set) """ - def __init__(self, items): - items = sorted(items) - items = list(reduce(lambda r, v: v in r and r or r+[v], items, [])) - super(HySet, self).__init__(items) - def __repr__(self): return "#{%s}" % (" ".join([repr(x) for x in self])) + +_wrappers[set] = lambda s: HySet(wrap_value(x) for x in s)
{"golden_diff": "diff --git a/hy/models/set.py b/hy/models/set.py\n--- a/hy/models/set.py\n+++ b/hy/models/set.py\n@@ -18,19 +18,16 @@\n # FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER\n # DEALINGS IN THE SOFTWARE.\n \n+from hy.models import _wrappers, wrap_value\n from hy.models.list import HyList\n-from functools import reduce\n \n \n class HySet(HyList):\n \"\"\"\n- Hy set (actually a list that pretends to be a set)\n+ Hy set (just a representation of a set)\n \"\"\"\n \n- def __init__(self, items):\n- items = sorted(items)\n- items = list(reduce(lambda r, v: v in r and r or r+[v], items, []))\n- super(HySet, self).__init__(items)\n-\n def __repr__(self):\n return \"#{%s}\" % (\" \".join([repr(x) for x in self]))\n+\n+_wrappers[set] = lambda s: HySet(wrap_value(x) for x in s)\n", "issue": "Hy sets are really broken.\n``` Hy\n(env-hy) C:\\Users\\ME\\Code>hy\nhy 0.11.0 using CPython(v3.4.3:9b73f1c3e601) 3.4.3 on Windows\n=> #{:a 'a}\nTraceback (most recent call last):\n File \"C:\\Python34\\Scripts\\env-hy\\Scripts\\hy-script.py\", line 9, in <module>\n load_entry_point('hy==0.11.0', 'console_scripts', 'hy')()\n File \"C:\\Python34\\Scripts\\env-hy\\lib\\site-packages\\hy\\cmdline.py\", line 341, in hy_main\n sys.exit(cmdline_handler(\"hy\", sys.argv))\n File \"C:\\Python34\\Scripts\\env-hy\\lib\\site-packages\\hy\\cmdline.py\", line 336, in cmdline_handler\n return run_repl(spy=options.spy)\n File \"C:\\Python34\\Scripts\\env-hy\\lib\\site-packages\\hy\\cmdline.py\", line 234, in run_repl\n os=platform.system()\n File \"C:\\Python34\\Lib\\code.py\", line 234, in interact\n more = self.push(line)\n File \"C:\\Python34\\Lib\\code.py\", line 256, in push\n more = self.runsource(source, self.filename)\n File \"C:\\Python34\\Scripts\\env-hy\\lib\\site-packages\\hy\\cmdline.py\", line 93, in runsource\n tokens = tokenize(source)\n File \"C:\\Python34\\Scripts\\env-hy\\lib\\site-packages\\hy\\lex\\__init__.py\", line 33, in tokenize\n return parser.parse(lexer.lex(buf))\n File \"C:\\Python34\\Scripts\\env-hy\\lib\\site-packages\\rply\\parser.py\", line 23, in parse\n t, symstack, statestack, state\n File \"C:\\Python34\\Scripts\\env-hy\\lib\\site-packages\\rply\\parser.py\", line 80, in _reduce_production\n value = p.func(targ)\n File \"C:\\Python34\\Scripts\\env-hy\\lib\\site-packages\\hy\\lex\\parser.py\", line 69, in wrapped\n ret = fun(p)\n File \"C:\\Python34\\Scripts\\env-hy\\lib\\site-packages\\hy\\lex\\parser.py\", line 214, in t_set\n return HySet(p[1])\n File \"C:\\Python34\\Scripts\\env-hy\\lib\\site-packages\\hy\\models\\set.py\", line 31, in __init__\n items = sorted(items)\nTypeError: unorderable types: HyExpression() < HyKeyword()\n\n(env-hy) C:\\Users\\ME\\Code>\n```\n\nThat is NOT supposed to happen.\n\n", "before_files": [{"content": "# Copyright (c) 2013 Paul Tagliamonte <[email protected]>\n#\n# Permission is hereby granted, free of charge, to any person obtaining a\n# copy of this software and associated documentation files (the \"Software\"),\n# to deal in the Software without restriction, including without limitation\n# the rights to use, copy, modify, merge, publish, distribute, sublicense,\n# and/or sell copies of the Software, and to permit persons to whom the\n# Software is furnished to do so, subject to the following conditions:\n#\n# The above copyright notice and this permission notice shall be included in\n# all copies or substantial portions of the Software.\n#\n# THE SOFTWARE IS PROVIDED \"AS IS\", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR\n# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,\n# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL\n# THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER\n# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING\n# FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER\n# DEALINGS IN THE SOFTWARE.\n\nfrom hy.models.list import HyList\nfrom functools import reduce\n\n\nclass HySet(HyList):\n \"\"\"\n Hy set (actually a list that pretends to be a set)\n \"\"\"\n\n def __init__(self, items):\n items = sorted(items)\n items = list(reduce(lambda r, v: v in r and r or r+[v], items, []))\n super(HySet, self).__init__(items)\n\n def __repr__(self):\n return \"#{%s}\" % (\" \".join([repr(x) for x in self]))\n", "path": "hy/models/set.py"}]}
1,627
242
gh_patches_debug_28579
rasdani/github-patches
git_diff
davanstrien__flyswot-115
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> update default for ensure_model to get latest model </issue> <code> [start of src/flyswot/inference.py] 1 """Inference functionality""" 2 import csv 3 import mimetypes 4 import time 5 from abc import ABC 6 from abc import abstractmethod 7 from dataclasses import asdict 8 from dataclasses import dataclass 9 from datetime import datetime 10 from datetime import timedelta 11 from pathlib import Path 12 from typing import Iterable 13 from typing import Iterator 14 from typing import List 15 from typing import Union 16 17 import numpy as np 18 import onnxruntime as rt # type: ignore 19 import typer 20 from PIL import Image # type: ignore 21 from rich.table import Table 22 from toolz import itertoolz # type: ignore 23 24 from flyswot import core 25 from flyswot import models 26 from flyswot.console import console 27 28 app = typer.Typer() 29 30 31 @dataclass() 32 class ImagePredictionItem: 33 """Prediction for an image. 34 35 Attributes: 36 path: The Path to the image 37 predicted_label: The predicted label i.e. the argmax value for the prediction tensor 38 condidence: The confidence for `predicted_label` i.e. the max value for prediction tensor 39 """ 40 41 path: Path 42 predicted_label: str 43 confidence: float 44 45 def __post_init__(self) -> Union[Path, None]: 46 """attempt to get absolute path""" 47 try: 48 self.path: Path = self.path.absolute() 49 except AttributeError: 50 pass 51 52 53 @dataclass 54 class PredictionBatch: 55 """Container for ImagePredictionItems""" 56 57 batch: List[ImagePredictionItem] 58 59 def __post_init__(self): 60 """Returns a list of all predicted labels in batch""" 61 self.batch_labels: Iterator[str] = (item.predicted_label for item in self.batch) 62 63 64 image_extensions = {k for k, v in mimetypes.types_map.items() if v.startswith("image/")} 65 66 67 @app.command() 68 def predict_image( 69 image: Path = typer.Argument(..., readable=True, resolve_path=True) 70 ) -> None: 71 """Predict a single image""" 72 pass # pragma: no cover 73 74 75 @app.command(name="directory") 76 def predict_directory( 77 directory: Path = typer.Argument( 78 ..., 79 readable=True, 80 resolve_path=True, 81 help="Directory to start searching for images from", 82 ), 83 csv_save_dir: Path = typer.Argument( 84 ..., 85 writable=True, 86 resolve_path=True, 87 help="Directory used to store the csv report", 88 ), 89 pattern: str = typer.Option("fse", help="Pattern used to filter image filenames"), 90 bs: int = typer.Option(16, help="Batch Size"), 91 image_format: str = typer.Option( 92 ".tif", 93 help="Image format for flyswot to use for predictions", 94 ), 95 ): 96 """Predicts against all images stored under DIRECTORY which match PATTERN in the filename. 97 98 By default searches for filenames containing 'fse'. 99 100 Creates a CSV report saved to `csv_save_dir` 101 """ 102 start_time = time.perf_counter() 103 model_dir = models.ensure_model_dir() 104 # TODO add load learner function that can be passed a model name 105 model_parts = models.ensure_model(model_dir) 106 model = model_parts.model 107 vocab = models.load_vocab(model_parts.vocab) 108 onnxinference = OnnxInferenceSession(model, vocab) 109 files = core.get_image_files_from_pattern(directory, pattern, image_format) 110 files = list(files) 111 typer.echo(f"Found {len(files)} files matching {pattern} in {directory}") 112 csv_fname = create_csv_fname(csv_save_dir) 113 create_csv_header(csv_fname) 114 with typer.progressbar(length=len(files)) as progress: 115 all_preds = [] 116 predictions = [] 117 for batch in itertoolz.partition_all(bs, files): 118 batch_predictions = onnxinference.predict_batch(batch, bs) 119 all_preds.append(batch_predictions.batch_labels) 120 predictions.append(batch_predictions) 121 progress.update(len(batch)) 122 write_batch_preds_to_csv(csv_fname, batch_predictions) 123 all_preds = list(itertoolz.concat(all_preds)) 124 typer.echo(f"CSV report stored in {csv_fname}") 125 delta = timedelta(seconds=time.perf_counter() - start_time) 126 typer.echo(f"Time taken to run: {str(delta)}") 127 print_table(all_preds) 128 129 130 def print_table(decoded) -> None: 131 """Prints table summary of predicted labels""" 132 table = Table(show_header=True, title="Prediction summary") 133 table.add_column( 134 "Class", 135 ) 136 table.add_column("Count") 137 table.add_column("Percentage") 138 total = len(decoded) 139 frequencies = itertoolz.frequencies(decoded) 140 for is_last_element, var in core.signal_last(frequencies.items()): 141 key, value = var 142 count = value 143 percentage = round((count / total) * 100, 2) 144 if is_last_element: 145 table.add_row(key, str(count), f"{percentage}", end_section=True) 146 table.add_row("Total", str(total), "") 147 else: 148 table.add_row(key, str(count), f"{percentage}") 149 console.print(table) 150 151 152 def create_csv_fname(csv_directory: Path) -> Path: 153 """Creates a csv filename""" 154 date_now = datetime.now() 155 date_now = date_now.strftime("%Y_%m_%d_%H_%M") 156 fname = Path(date_now + ".csv") 157 return Path(csv_directory / fname) 158 159 160 def create_csv_header(csv_path: Path) -> None: 161 """Creates a header for csv `csv_path`""" 162 with open(csv_path, mode="w", newline="") as csv_file: 163 field_names = ["path", "directory", "predicted_label", "confidence"] 164 writer = csv.DictWriter(csv_file, fieldnames=field_names) 165 writer.writeheader() 166 167 168 def write_batch_preds_to_csv(csv_fpath: Path, predictions: PredictionBatch) -> None: 169 """Appends `predictions` batch to `csv_path`""" 170 with open(csv_fpath, mode="a", newline="") as csv_file: 171 field_names = ["path", "directory", "predicted_label", "confidence"] 172 writer = csv.DictWriter(csv_file, fieldnames=field_names) 173 for pred in predictions.batch: 174 row = asdict(pred) 175 row["directory"] = pred.path.parent 176 writer.writerow(row) 177 178 179 class InferenceSession(ABC): 180 """Abstract class for inference sessions""" 181 182 @abstractmethod 183 def __init__(self, model: Path, vocab: List): 184 """Inference Sessions should init from a model file and vocab""" 185 self.model = model 186 self.vocab = vocab 187 188 @abstractmethod 189 def predict_image(self, image: Path): 190 """Predict a single image""" 191 pass 192 193 @abstractmethod 194 def predict_batch(self, model: Path, batch: Iterable[Path], bs: int): 195 """Predict a batch""" 196 pass 197 198 199 def softmax(x): 200 """return softmax of `x`""" 201 x = x.reshape(-1) 202 e_x = np.exp(x - np.max(x)) 203 return e_x / e_x.sum(axis=0) 204 205 206 # class FastaiInferenceModel(InferenceSession): 207 # def __init__(self, model): 208 # self.model = model 209 # self.learn = load_learner(model) 210 211 # def predict_image(self, image: Path) -> Any: 212 # return self.learn.predict(image) 213 214 # def predict_batch(self, batch: Iterable[Path], bs: int) -> PredictionBatch: 215 # test_dl = self.learn.dls.test_dl(batch, bs=bs) 216 # vocab = dict(enumerate(self.learn.dls.vocab)) 217 # with self.learn.no_bar(): 218 # fastai_preds: Any = self.learn.get_preds(dl=test_dl, with_decoded=True) 219 # prediction_tensors: Iterable[Any] = fastai_preds[0] 220 # prediction_items = [] 221 # for file, pred in zip(batch, prediction_tensors): 222 # arg_max = int(np.array(pred).argmax()) 223 # predicted_label = vocab[int(arg_max)] 224 # confidence = float(np.array(pred).max()) 225 # prediction_items.append( 226 # ImagePredictionItem(file, predicted_label, confidence) 227 # ) 228 # return PredictionBatch(prediction_items) 229 230 231 class OnnxInferenceSession(InferenceSession): 232 """onnx inference session""" 233 234 def __init__(self, model: Path, vocab: Path): 235 """Create onnx session""" 236 self.model = model 237 self.session = rt.InferenceSession(str(model)) 238 239 self.vocab = vocab 240 self.vocab_mapping = dict(enumerate(self.vocab)) 241 242 def _load_vocab(self, vocab: Path) -> List: 243 with open(vocab, "r") as f: 244 return [item.strip("\n") for item in f.readlines()] 245 246 def predict_image(self, image: Path): 247 """Predict a single image""" 248 img = self._load_image(image) 249 raw_result = self.session.run(["output"], {"image": img}) 250 pred = self._postprocess(raw_result) 251 arg_max = int(np.array(pred).argmax()) 252 predicted_label = self.vocab_mapping[int(arg_max)] 253 confidence = float(np.array(pred).max()) 254 return ImagePredictionItem(image, predicted_label, confidence) 255 256 def _preprocess(self, input_data: np.ndarray) -> np.ndarray: 257 # converts the input data into the float32 input for onnx 258 img_data = input_data.astype("float32") 259 260 # normalize 261 mean_vec = np.array([0.485, 0.456, 0.406]) 262 stddev_vec = np.array([0.229, 0.224, 0.225]) 263 norm_img_data = np.zeros(img_data.shape).astype("float32") 264 for i in range(img_data.shape[0]): 265 norm_img_data[i, :, :] = ( 266 img_data[i, :, :] / 255 - mean_vec[i] 267 ) / stddev_vec[i] 268 269 # add batch channel 270 norm_img_data = norm_img_data.reshape(1, 3, 512, 512).astype("float32") 271 return norm_img_data 272 273 def _load_image(self, file: Path) -> np.ndarray: 274 """loads image and carries out preprocessing for inference""" 275 image = Image.open(file, mode="r") 276 image = image.resize((512, 512), Image.BILINEAR) 277 image_data = np.array(image).transpose(2, 0, 1) 278 return self._preprocess(image_data) 279 280 def _postprocess(self, result: List): 281 """process results from onnx session""" 282 return softmax(np.array(result)).tolist() 283 284 def predict_batch(self, batch: Iterable[Path], bs: int): 285 """predicts a batch of images""" 286 prediction_items = [self.predict_image(file) for file in batch] 287 return PredictionBatch(prediction_items) 288 289 290 if __name__ == "__main__": 291 app() 292 [end of src/flyswot/inference.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/src/flyswot/inference.py b/src/flyswot/inference.py --- a/src/flyswot/inference.py +++ b/src/flyswot/inference.py @@ -19,7 +19,7 @@ import typer from PIL import Image # type: ignore from rich.table import Table -from toolz import itertoolz # type: ignore +from toolz import itertoolz from flyswot import core from flyswot import models @@ -89,9 +89,9 @@ pattern: str = typer.Option("fse", help="Pattern used to filter image filenames"), bs: int = typer.Option(16, help="Batch Size"), image_format: str = typer.Option( - ".tif", - help="Image format for flyswot to use for predictions", + ".tif", help="Image format for flyswot to use for predictions" ), + check_latest: bool = typer.Option(True, help="Use latest available model"), ): """Predicts against all images stored under DIRECTORY which match PATTERN in the filename. @@ -102,7 +102,7 @@ start_time = time.perf_counter() model_dir = models.ensure_model_dir() # TODO add load learner function that can be passed a model name - model_parts = models.ensure_model(model_dir) + model_parts = models.ensure_model(model_dir, check_latest) model = model_parts.model vocab = models.load_vocab(model_parts.vocab) onnxinference = OnnxInferenceSession(model, vocab)
{"golden_diff": "diff --git a/src/flyswot/inference.py b/src/flyswot/inference.py\n--- a/src/flyswot/inference.py\n+++ b/src/flyswot/inference.py\n@@ -19,7 +19,7 @@\n import typer\n from PIL import Image # type: ignore\n from rich.table import Table\n-from toolz import itertoolz # type: ignore\n+from toolz import itertoolz\n \n from flyswot import core\n from flyswot import models\n@@ -89,9 +89,9 @@\n pattern: str = typer.Option(\"fse\", help=\"Pattern used to filter image filenames\"),\n bs: int = typer.Option(16, help=\"Batch Size\"),\n image_format: str = typer.Option(\n- \".tif\",\n- help=\"Image format for flyswot to use for predictions\",\n+ \".tif\", help=\"Image format for flyswot to use for predictions\"\n ),\n+ check_latest: bool = typer.Option(True, help=\"Use latest available model\"),\n ):\n \"\"\"Predicts against all images stored under DIRECTORY which match PATTERN in the filename.\n \n@@ -102,7 +102,7 @@\n start_time = time.perf_counter()\n model_dir = models.ensure_model_dir()\n # TODO add load learner function that can be passed a model name\n- model_parts = models.ensure_model(model_dir)\n+ model_parts = models.ensure_model(model_dir, check_latest)\n model = model_parts.model\n vocab = models.load_vocab(model_parts.vocab)\n onnxinference = OnnxInferenceSession(model, vocab)\n", "issue": "update default for ensure_model to get latest model \n\n", "before_files": [{"content": "\"\"\"Inference functionality\"\"\"\nimport csv\nimport mimetypes\nimport time\nfrom abc import ABC\nfrom abc import abstractmethod\nfrom dataclasses import asdict\nfrom dataclasses import dataclass\nfrom datetime import datetime\nfrom datetime import timedelta\nfrom pathlib import Path\nfrom typing import Iterable\nfrom typing import Iterator\nfrom typing import List\nfrom typing import Union\n\nimport numpy as np\nimport onnxruntime as rt # type: ignore\nimport typer\nfrom PIL import Image # type: ignore\nfrom rich.table import Table\nfrom toolz import itertoolz # type: ignore\n\nfrom flyswot import core\nfrom flyswot import models\nfrom flyswot.console import console\n\napp = typer.Typer()\n\n\n@dataclass()\nclass ImagePredictionItem:\n \"\"\"Prediction for an image.\n\n Attributes:\n path: The Path to the image\n predicted_label: The predicted label i.e. the argmax value for the prediction tensor\n condidence: The confidence for `predicted_label` i.e. the max value for prediction tensor\n \"\"\"\n\n path: Path\n predicted_label: str\n confidence: float\n\n def __post_init__(self) -> Union[Path, None]:\n \"\"\"attempt to get absolute path\"\"\"\n try:\n self.path: Path = self.path.absolute()\n except AttributeError:\n pass\n\n\n@dataclass\nclass PredictionBatch:\n \"\"\"Container for ImagePredictionItems\"\"\"\n\n batch: List[ImagePredictionItem]\n\n def __post_init__(self):\n \"\"\"Returns a list of all predicted labels in batch\"\"\"\n self.batch_labels: Iterator[str] = (item.predicted_label for item in self.batch)\n\n\nimage_extensions = {k for k, v in mimetypes.types_map.items() if v.startswith(\"image/\")}\n\n\[email protected]()\ndef predict_image(\n image: Path = typer.Argument(..., readable=True, resolve_path=True)\n) -> None:\n \"\"\"Predict a single image\"\"\"\n pass # pragma: no cover\n\n\[email protected](name=\"directory\")\ndef predict_directory(\n directory: Path = typer.Argument(\n ...,\n readable=True,\n resolve_path=True,\n help=\"Directory to start searching for images from\",\n ),\n csv_save_dir: Path = typer.Argument(\n ...,\n writable=True,\n resolve_path=True,\n help=\"Directory used to store the csv report\",\n ),\n pattern: str = typer.Option(\"fse\", help=\"Pattern used to filter image filenames\"),\n bs: int = typer.Option(16, help=\"Batch Size\"),\n image_format: str = typer.Option(\n \".tif\",\n help=\"Image format for flyswot to use for predictions\",\n ),\n):\n \"\"\"Predicts against all images stored under DIRECTORY which match PATTERN in the filename.\n\n By default searches for filenames containing 'fse'.\n\n Creates a CSV report saved to `csv_save_dir`\n \"\"\"\n start_time = time.perf_counter()\n model_dir = models.ensure_model_dir()\n # TODO add load learner function that can be passed a model name\n model_parts = models.ensure_model(model_dir)\n model = model_parts.model\n vocab = models.load_vocab(model_parts.vocab)\n onnxinference = OnnxInferenceSession(model, vocab)\n files = core.get_image_files_from_pattern(directory, pattern, image_format)\n files = list(files)\n typer.echo(f\"Found {len(files)} files matching {pattern} in {directory}\")\n csv_fname = create_csv_fname(csv_save_dir)\n create_csv_header(csv_fname)\n with typer.progressbar(length=len(files)) as progress:\n all_preds = []\n predictions = []\n for batch in itertoolz.partition_all(bs, files):\n batch_predictions = onnxinference.predict_batch(batch, bs)\n all_preds.append(batch_predictions.batch_labels)\n predictions.append(batch_predictions)\n progress.update(len(batch))\n write_batch_preds_to_csv(csv_fname, batch_predictions)\n all_preds = list(itertoolz.concat(all_preds))\n typer.echo(f\"CSV report stored in {csv_fname}\")\n delta = timedelta(seconds=time.perf_counter() - start_time)\n typer.echo(f\"Time taken to run: {str(delta)}\")\n print_table(all_preds)\n\n\ndef print_table(decoded) -> None:\n \"\"\"Prints table summary of predicted labels\"\"\"\n table = Table(show_header=True, title=\"Prediction summary\")\n table.add_column(\n \"Class\",\n )\n table.add_column(\"Count\")\n table.add_column(\"Percentage\")\n total = len(decoded)\n frequencies = itertoolz.frequencies(decoded)\n for is_last_element, var in core.signal_last(frequencies.items()):\n key, value = var\n count = value\n percentage = round((count / total) * 100, 2)\n if is_last_element:\n table.add_row(key, str(count), f\"{percentage}\", end_section=True)\n table.add_row(\"Total\", str(total), \"\")\n else:\n table.add_row(key, str(count), f\"{percentage}\")\n console.print(table)\n\n\ndef create_csv_fname(csv_directory: Path) -> Path:\n \"\"\"Creates a csv filename\"\"\"\n date_now = datetime.now()\n date_now = date_now.strftime(\"%Y_%m_%d_%H_%M\")\n fname = Path(date_now + \".csv\")\n return Path(csv_directory / fname)\n\n\ndef create_csv_header(csv_path: Path) -> None:\n \"\"\"Creates a header for csv `csv_path`\"\"\"\n with open(csv_path, mode=\"w\", newline=\"\") as csv_file:\n field_names = [\"path\", \"directory\", \"predicted_label\", \"confidence\"]\n writer = csv.DictWriter(csv_file, fieldnames=field_names)\n writer.writeheader()\n\n\ndef write_batch_preds_to_csv(csv_fpath: Path, predictions: PredictionBatch) -> None:\n \"\"\"Appends `predictions` batch to `csv_path`\"\"\"\n with open(csv_fpath, mode=\"a\", newline=\"\") as csv_file:\n field_names = [\"path\", \"directory\", \"predicted_label\", \"confidence\"]\n writer = csv.DictWriter(csv_file, fieldnames=field_names)\n for pred in predictions.batch:\n row = asdict(pred)\n row[\"directory\"] = pred.path.parent\n writer.writerow(row)\n\n\nclass InferenceSession(ABC):\n \"\"\"Abstract class for inference sessions\"\"\"\n\n @abstractmethod\n def __init__(self, model: Path, vocab: List):\n \"\"\"Inference Sessions should init from a model file and vocab\"\"\"\n self.model = model\n self.vocab = vocab\n\n @abstractmethod\n def predict_image(self, image: Path):\n \"\"\"Predict a single image\"\"\"\n pass\n\n @abstractmethod\n def predict_batch(self, model: Path, batch: Iterable[Path], bs: int):\n \"\"\"Predict a batch\"\"\"\n pass\n\n\ndef softmax(x):\n \"\"\"return softmax of `x`\"\"\"\n x = x.reshape(-1)\n e_x = np.exp(x - np.max(x))\n return e_x / e_x.sum(axis=0)\n\n\n# class FastaiInferenceModel(InferenceSession):\n# def __init__(self, model):\n# self.model = model\n# self.learn = load_learner(model)\n\n# def predict_image(self, image: Path) -> Any:\n# return self.learn.predict(image)\n\n# def predict_batch(self, batch: Iterable[Path], bs: int) -> PredictionBatch:\n# test_dl = self.learn.dls.test_dl(batch, bs=bs)\n# vocab = dict(enumerate(self.learn.dls.vocab))\n# with self.learn.no_bar():\n# fastai_preds: Any = self.learn.get_preds(dl=test_dl, with_decoded=True)\n# prediction_tensors: Iterable[Any] = fastai_preds[0]\n# prediction_items = []\n# for file, pred in zip(batch, prediction_tensors):\n# arg_max = int(np.array(pred).argmax())\n# predicted_label = vocab[int(arg_max)]\n# confidence = float(np.array(pred).max())\n# prediction_items.append(\n# ImagePredictionItem(file, predicted_label, confidence)\n# )\n# return PredictionBatch(prediction_items)\n\n\nclass OnnxInferenceSession(InferenceSession):\n \"\"\"onnx inference session\"\"\"\n\n def __init__(self, model: Path, vocab: Path):\n \"\"\"Create onnx session\"\"\"\n self.model = model\n self.session = rt.InferenceSession(str(model))\n\n self.vocab = vocab\n self.vocab_mapping = dict(enumerate(self.vocab))\n\n def _load_vocab(self, vocab: Path) -> List:\n with open(vocab, \"r\") as f:\n return [item.strip(\"\\n\") for item in f.readlines()]\n\n def predict_image(self, image: Path):\n \"\"\"Predict a single image\"\"\"\n img = self._load_image(image)\n raw_result = self.session.run([\"output\"], {\"image\": img})\n pred = self._postprocess(raw_result)\n arg_max = int(np.array(pred).argmax())\n predicted_label = self.vocab_mapping[int(arg_max)]\n confidence = float(np.array(pred).max())\n return ImagePredictionItem(image, predicted_label, confidence)\n\n def _preprocess(self, input_data: np.ndarray) -> np.ndarray:\n # converts the input data into the float32 input for onnx\n img_data = input_data.astype(\"float32\")\n\n # normalize\n mean_vec = np.array([0.485, 0.456, 0.406])\n stddev_vec = np.array([0.229, 0.224, 0.225])\n norm_img_data = np.zeros(img_data.shape).astype(\"float32\")\n for i in range(img_data.shape[0]):\n norm_img_data[i, :, :] = (\n img_data[i, :, :] / 255 - mean_vec[i]\n ) / stddev_vec[i]\n\n # add batch channel\n norm_img_data = norm_img_data.reshape(1, 3, 512, 512).astype(\"float32\")\n return norm_img_data\n\n def _load_image(self, file: Path) -> np.ndarray:\n \"\"\"loads image and carries out preprocessing for inference\"\"\"\n image = Image.open(file, mode=\"r\")\n image = image.resize((512, 512), Image.BILINEAR)\n image_data = np.array(image).transpose(2, 0, 1)\n return self._preprocess(image_data)\n\n def _postprocess(self, result: List):\n \"\"\"process results from onnx session\"\"\"\n return softmax(np.array(result)).tolist()\n\n def predict_batch(self, batch: Iterable[Path], bs: int):\n \"\"\"predicts a batch of images\"\"\"\n prediction_items = [self.predict_image(file) for file in batch]\n return PredictionBatch(prediction_items)\n\n\nif __name__ == \"__main__\":\n app()\n", "path": "src/flyswot/inference.py"}]}
3,677
357
gh_patches_debug_1429
rasdani/github-patches
git_diff
google__turbinia-785
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> import TurbiniaException to partitions.py ``` Traceback (most recent call last): File "PATH/v2/lib/python3.8/site-packages/turbinia/workers/__init__.py", line 916, in run_wrapper self.result = self.run(evidence, self.result) File "PATH/v2/lib/python3.8/site-packages/turbinia/workers/partitions.py", line 144, in run path_specs = partitions.Enumerate(evidence) File "/PATH/v2/lib/python3.8/site-packages/turbinia/processors/partitions.py", line 49, in Enumerate raise TurbiniaException( NameError: name 'TurbiniaException' is not defined 2021-03-05 18:45:56 [ERROR] PartitionEnumerationTask Task failed with exception: [name 'TurbiniaException' is not defined] ``` </issue> <code> [start of turbinia/processors/partitions.py] 1 # -*- coding: utf-8 -*- 2 # Copyright 2021 Google LLC 3 # 4 # Licensed under the Apache License, Version 2.0 (the "License"); 5 # you may not use this file except in compliance with the License. 6 # You may obtain a copy of the License at 7 # 8 # https://www.apache.org/licenses/LICENSE-2.0 9 # 10 # Unless required by applicable law or agreed to in writing, software 11 # distributed under the License is distributed on an "AS IS" BASIS, 12 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 13 # See the License for the specific language governing permissions and 14 # limitations under the License. 15 """Evidence processor to enumerate partitions.""" 16 17 import logging 18 19 from dfvfs.helpers import volume_scanner 20 from dfvfs.lib import definitions as dfvfs_definitions 21 from dfvfs.lib import errors as dfvfs_errors 22 23 from turbinia.lib.dfvfs_classes import UnattendedVolumeScannerMediator 24 25 log = logging.getLogger('turbinia') 26 27 28 def Enumerate(evidence): 29 """Uses dfVFS to enumerate partitions in a disk / image. 30 31 Args: 32 evidence: Evidence object to be scanned. 33 34 Raises: 35 TurbiniaException if source evidence can't be scanned. 36 37 Returns: 38 list[dfVFS.path_spec]: path specs for identified partitions 39 """ 40 dfvfs_definitions.PREFERRED_GPT_BACK_END = ( 41 dfvfs_definitions.TYPE_INDICATOR_GPT) 42 mediator = UnattendedVolumeScannerMediator() 43 mediator.credentials = evidence.credentials 44 path_specs = [] 45 try: 46 scanner = volume_scanner.VolumeScanner(mediator=mediator) 47 path_specs = scanner.GetBasePathSpecs(evidence.local_path) 48 except dfvfs_errors.ScannerError as e: 49 raise TurbiniaException( 50 'Could not enumerate partitions [{0!s}]: {1!s}'.format( 51 evidence.local_path, e)) 52 53 return path_specs 54 55 56 def GetPartitionEncryptionType(path_spec): 57 """Checks a partition for encryption. 58 59 Args: 60 path_spec (dfVFS.path_spec): Partition path_spec. 61 62 Returns: 63 String representing the type of encryption, or None. 64 """ 65 encryption_type = None 66 if path_spec.parent.type_indicator == dfvfs_definitions.TYPE_INDICATOR_BDE: 67 encryption_type = 'BDE' 68 return encryption_type 69 70 71 def GetPathSpecByLocation(path_specs, location): 72 """Finds a path_spec from a list of path_specs for a given location. 73 74 Args: 75 path_specs (list[dfVFS.path_spec]): List of path_specs from volume scanner. 76 location (str): dfVFS location to search for. 77 78 Returns: 79 dfVFS.path_spec for the given location or None if not found. 80 """ 81 for path_spec in path_specs: 82 child_path_spec = path_spec 83 fs_location = getattr(path_spec, 'location', None) 84 while path_spec.HasParent(): 85 type_indicator = path_spec.type_indicator 86 if type_indicator in (dfvfs_definitions.TYPE_INDICATOR_TSK_PARTITION, 87 dfvfs_definitions.TYPE_INDICATOR_GPT): 88 if fs_location in ('\\', '/'): 89 fs_location = getattr(path_spec, 'location', None) 90 break 91 path_spec = path_spec.parent 92 if fs_location == location: 93 return child_path_spec 94 return None 95 [end of turbinia/processors/partitions.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/turbinia/processors/partitions.py b/turbinia/processors/partitions.py --- a/turbinia/processors/partitions.py +++ b/turbinia/processors/partitions.py @@ -21,6 +21,7 @@ from dfvfs.lib import errors as dfvfs_errors from turbinia.lib.dfvfs_classes import UnattendedVolumeScannerMediator +from turbinia import TurbiniaException log = logging.getLogger('turbinia')
{"golden_diff": "diff --git a/turbinia/processors/partitions.py b/turbinia/processors/partitions.py\n--- a/turbinia/processors/partitions.py\n+++ b/turbinia/processors/partitions.py\n@@ -21,6 +21,7 @@\n from dfvfs.lib import errors as dfvfs_errors\n \n from turbinia.lib.dfvfs_classes import UnattendedVolumeScannerMediator\n+from turbinia import TurbiniaException\n \n log = logging.getLogger('turbinia')\n", "issue": "import TurbiniaException to partitions.py\n```\r\nTraceback (most recent call last):\r\n File \"PATH/v2/lib/python3.8/site-packages/turbinia/workers/__init__.py\", line 916, in run_wrapper\r\n self.result = self.run(evidence, self.result)\r\n File \"PATH/v2/lib/python3.8/site-packages/turbinia/workers/partitions.py\", line 144, in run\r\n path_specs = partitions.Enumerate(evidence)\r\n File \"/PATH/v2/lib/python3.8/site-packages/turbinia/processors/partitions.py\", line 49, in Enumerate\r\n raise TurbiniaException(\r\nNameError: name 'TurbiniaException' is not defined\r\n\r\n2021-03-05 18:45:56 [ERROR] PartitionEnumerationTask Task failed with exception: [name 'TurbiniaException' is not defined]\r\n```\n", "before_files": [{"content": "# -*- coding: utf-8 -*-\n# Copyright 2021 Google LLC\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# https://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\"\"\"Evidence processor to enumerate partitions.\"\"\"\n\nimport logging\n\nfrom dfvfs.helpers import volume_scanner\nfrom dfvfs.lib import definitions as dfvfs_definitions\nfrom dfvfs.lib import errors as dfvfs_errors\n\nfrom turbinia.lib.dfvfs_classes import UnattendedVolumeScannerMediator\n\nlog = logging.getLogger('turbinia')\n\n\ndef Enumerate(evidence):\n \"\"\"Uses dfVFS to enumerate partitions in a disk / image.\n\n Args:\n evidence: Evidence object to be scanned.\n\n Raises:\n TurbiniaException if source evidence can't be scanned.\n\n Returns:\n list[dfVFS.path_spec]: path specs for identified partitions\n \"\"\"\n dfvfs_definitions.PREFERRED_GPT_BACK_END = (\n dfvfs_definitions.TYPE_INDICATOR_GPT)\n mediator = UnattendedVolumeScannerMediator()\n mediator.credentials = evidence.credentials\n path_specs = []\n try:\n scanner = volume_scanner.VolumeScanner(mediator=mediator)\n path_specs = scanner.GetBasePathSpecs(evidence.local_path)\n except dfvfs_errors.ScannerError as e:\n raise TurbiniaException(\n 'Could not enumerate partitions [{0!s}]: {1!s}'.format(\n evidence.local_path, e))\n\n return path_specs\n\n\ndef GetPartitionEncryptionType(path_spec):\n \"\"\"Checks a partition for encryption.\n\n Args:\n path_spec (dfVFS.path_spec): Partition path_spec.\n\n Returns:\n String representing the type of encryption, or None.\n \"\"\"\n encryption_type = None\n if path_spec.parent.type_indicator == dfvfs_definitions.TYPE_INDICATOR_BDE:\n encryption_type = 'BDE'\n return encryption_type\n\n\ndef GetPathSpecByLocation(path_specs, location):\n \"\"\"Finds a path_spec from a list of path_specs for a given location.\n\n Args:\n path_specs (list[dfVFS.path_spec]): List of path_specs from volume scanner.\n location (str): dfVFS location to search for.\n\n Returns:\n dfVFS.path_spec for the given location or None if not found.\n \"\"\"\n for path_spec in path_specs:\n child_path_spec = path_spec\n fs_location = getattr(path_spec, 'location', None)\n while path_spec.HasParent():\n type_indicator = path_spec.type_indicator\n if type_indicator in (dfvfs_definitions.TYPE_INDICATOR_TSK_PARTITION,\n dfvfs_definitions.TYPE_INDICATOR_GPT):\n if fs_location in ('\\\\', '/'):\n fs_location = getattr(path_spec, 'location', None)\n break\n path_spec = path_spec.parent\n if fs_location == location:\n return child_path_spec\n return None\n", "path": "turbinia/processors/partitions.py"}]}
1,636
108
gh_patches_debug_16024
rasdani/github-patches
git_diff
mindsdb__mindsdb-2390
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> Error when csv header has empty space ### Discussed in https://github.com/mindsdb/mindsdb/discussions/2343 <div type='discussions-op-text'> <sup>Originally posted by **Ashok0** June 13, 2022</sup> Hi, I have a question regarding usage of imported data with MindsDB. I registered a MindsDB account and selected "Add Data" > "Import File" and uploaded some sample CSV data to a table. I can view the data just fine as shown in the following screencap: https://imgur.com/a/X6mrzHt If I attempt to select an individual column i.e. 'sensitivity' with the following statement: **SELECT sensitivity FROM files.bedcheck**, I receive the following error message: **Binder Error: Referenced column "sensitivity" not found in FROM clause! Candidate bindings: "df_table. sensitivity" LINE 1: SELECT sensitivity AS sensitivity** Any assistance in how I can resolve this error would be greatly appreciated. Thank you!</div> </issue> <code> [start of mindsdb/integrations/handlers/file_handler/file_handler.py] 1 from io import BytesIO, StringIO 2 import os 3 import csv 4 import json 5 import codecs 6 import traceback 7 import tempfile 8 from urllib.parse import urlparse 9 10 import requests 11 import pandas as pd 12 13 from mindsdb_sql import parse_sql 14 from mindsdb_sql.parser.ast.base import ASTNode 15 from mindsdb_sql.parser.ast import DropTables, Select 16 17 from mindsdb.api.mysql.mysql_proxy.utilities.sql import query_df 18 from mindsdb.integrations.libs.base_handler import DatabaseHandler 19 from mindsdb.integrations.libs.response import ( 20 HandlerStatusResponse as StatusResponse, 21 HandlerResponse as Response, 22 RESPONSE_TYPE 23 ) 24 25 26 def clean_row(row): 27 n_row = [] 28 for cell in row: 29 if str(cell) in ['', ' ', ' ', 'NaN', 'nan', 'NA']: 30 n_row.append(None) 31 else: 32 n_row.append(cell) 33 34 return n_row 35 36 37 class FileHandler(DatabaseHandler): 38 """ 39 Handler for files 40 """ 41 type = 'files' 42 43 def __init__(self, name=None, db_store=None, fs_store=None, connection_data=None, file_controller=None): 44 super().__init__(name) 45 self.parser = parse_sql 46 self.fs_store = fs_store 47 self.custom_parser = connection_data.get('custom_parser') 48 self.clean_rows = connection_data.get('clean_rows', True) 49 self.file_controller = file_controller 50 51 def connect(self, **kwargs): 52 return 53 54 def disconnect(self, **kwargs): 55 return 56 57 def check_connection(self) -> StatusResponse: 58 return StatusResponse(True) 59 60 def query(self, query: ASTNode) -> Response: 61 if type(query) == DropTables: 62 for table_identifier in query.tables: 63 if len(table_identifier.parts) == 2 and table_identifier.parts[0] != self.name: 64 return Response( 65 RESPONSE_TYPE.ERROR, 66 error_message=f"Can't delete table from database '{table_identifier.parts[0]}'" 67 ) 68 table_name = table_identifier.parts[-1] 69 try: 70 self.file_controller.delete_file(table_name) 71 except Exception as e: 72 return Response( 73 RESPONSE_TYPE.ERROR, 74 error_message=f"Can't delete table '{table_name}': {e}" 75 ) 76 return Response(RESPONSE_TYPE.OK) 77 elif type(query) == Select: 78 table_name = query.from_table.parts[-1] 79 file_path = self.file_controller.get_file_path(table_name, company_id=None) 80 df, _columns = self._handle_source(file_path, self.clean_rows, self.custom_parser) 81 result_df = query_df(df, query) 82 return Response( 83 RESPONSE_TYPE.TABLE, 84 data_frame=result_df 85 ) 86 else: 87 return Response( 88 RESPONSE_TYPE.ERROR, 89 error_message="Only 'select' and 'drop' queries allowed for files" 90 ) 91 92 def native_query(self, query: str) -> Response: 93 ast = self.parser(query, dialect='mindsdb') 94 return self.query(ast) 95 96 @staticmethod 97 def _handle_source(file_path, clean_rows=True, custom_parser=None): 98 # get file data io, format and dialect 99 data, fmt, dialect = FileHandler._get_data_io(file_path) 100 data.seek(0) # make sure we are at 0 in file pointer 101 102 if custom_parser: 103 header, file_data = custom_parser(data, fmt) 104 105 elif fmt == 'csv': 106 csv_reader = list(csv.reader(data, dialect)) 107 header = csv_reader[0] 108 file_data = csv_reader[1:] 109 110 elif fmt in ['xlsx', 'xls']: 111 data.seek(0) 112 df = pd.read_excel(data) 113 header = df.columns.values.tolist() 114 file_data = df.values.tolist() 115 116 elif fmt == 'json': 117 data.seek(0) 118 json_doc = json.loads(data.read()) 119 df = pd.json_normalize(json_doc, max_level=0) 120 header = df.columns.values.tolist() 121 file_data = df.values.tolist() 122 123 else: 124 raise ValueError('Could not load file into any format, supported formats are csv, json, xls, xlsx') 125 126 if clean_rows: 127 file_list_data = [clean_row(row) for row in file_data] 128 else: 129 file_list_data = file_data 130 131 col_map = dict((col, col) for col in header) 132 return pd.DataFrame(file_list_data, columns=header), col_map 133 134 @staticmethod 135 def _get_data_io(file_path): 136 """ 137 This gets a file either url or local file and defiens what the format is as well as dialect 138 :param file: file path or url 139 :return: data_io, format, dialect 140 """ 141 142 ############ 143 # get file as io object 144 ############ 145 146 # file_path = self._get_file_path() 147 148 data = BytesIO() 149 150 try: 151 with open(file_path, 'rb') as fp: 152 data = BytesIO(fp.read()) 153 except Exception as e: 154 error = 'Could not load file, possible exception : {exception}'.format(exception=e) 155 print(error) 156 raise ValueError(error) 157 158 dialect = None 159 160 ############ 161 # check for file type 162 ############ 163 164 # try to guess if its an excel file 165 xlsx_sig = b'\x50\x4B\x05\06' 166 # xlsx_sig2 = b'\x50\x4B\x03\x04' 167 xls_sig = b'\x09\x08\x10\x00\x00\x06\x05\x00' 168 169 # different whence, offset, size for different types 170 excel_meta = [('xls', 0, 512, 8), ('xlsx', 2, -22, 4)] 171 172 for filename, whence, offset, size in excel_meta: 173 174 try: 175 data.seek(offset, whence) # Seek to the offset. 176 bytes = data.read(size) # Capture the specified number of bytes. 177 data.seek(0) 178 codecs.getencoder('hex')(bytes) 179 180 if bytes == xls_sig: 181 return data, 'xls', dialect 182 elif bytes == xlsx_sig: 183 return data, 'xlsx', dialect 184 185 except Exception: 186 data.seek(0) 187 188 # if not excel it can be a json file or a CSV, convert from binary to stringio 189 190 byte_str = data.read() 191 # Move it to StringIO 192 try: 193 # Handle Microsoft's BOM "special" UTF-8 encoding 194 if byte_str.startswith(codecs.BOM_UTF8): 195 data = StringIO(byte_str.decode('utf-8-sig')) 196 else: 197 data = StringIO(byte_str.decode('utf-8')) 198 199 except Exception: 200 print(traceback.format_exc()) 201 print('Could not load into string') 202 203 # see if its JSON 204 buffer = data.read(100) 205 data.seek(0) 206 text = buffer.strip() 207 # analyze first n characters 208 if len(text) > 0: 209 text = text.strip() 210 # it it looks like a json, then try to parse it 211 if text.startswith('{') or text.startswith('['): 212 try: 213 json.loads(data.read()) 214 data.seek(0) 215 return data, 'json', dialect 216 except Exception: 217 data.seek(0) 218 return data, None, dialect 219 220 # lets try to figure out if its a csv 221 try: 222 dialect = FileHandler._get_csv_dialect(file_path) 223 if dialect: 224 return data, 'csv', dialect 225 return data, None, dialect 226 except Exception: 227 data.seek(0) 228 print('Could not detect format for this file') 229 print(traceback.format_exc()) 230 # No file type identified 231 return data, None, dialect 232 233 @staticmethod 234 def _get_file_path(path) -> str: 235 try: 236 is_url = urlparse(path).scheme in ('http', 'https') 237 except Exception: 238 is_url = False 239 if is_url: 240 path = FileHandler._fetch_url(path) 241 return path 242 243 @staticmethod 244 def _get_csv_dialect(file_path) -> csv.Dialect: 245 with open(file_path, 'rt') as f: 246 try: 247 accepted_csv_delimiters = [',', '\t', ';'] 248 dialect = csv.Sniffer().sniff(f.read(128 * 1024), delimiters=accepted_csv_delimiters) 249 except csv.Error: 250 dialect = None 251 return dialect 252 253 @staticmethod 254 def _fetch_url(url: str) -> str: 255 temp_dir = tempfile.mkdtemp(prefix='mindsdb_file_url_') 256 try: 257 r = requests.get(url, stream=True) 258 if r.status_code == 200: 259 with open(os.path.join(temp_dir, 'file'), 'wb') as f: 260 for chunk in r: 261 f.write(chunk) 262 else: 263 raise Exception(f'Responce status code is {r.status_code}') 264 except Exception as e: 265 print(f'Error during getting {url}') 266 print(e) 267 raise 268 return os.path.join(temp_dir, 'file') 269 270 def get_tables(self) -> Response: 271 """ 272 List all files 273 """ 274 files_meta = self.file_controller.get_files() 275 data = [{ 276 'TABLE_NAME': x['name'], 277 'TABLE_ROWS': x['row_count'] 278 } for x in files_meta] 279 return Response( 280 RESPONSE_TYPE.TABLE, 281 data_frame=pd.DataFrame(data) 282 ) 283 284 def get_columns(self, table_name) -> Response: 285 file_meta = self.file_controller.get_file_meta(table_name) 286 result = Response( 287 RESPONSE_TYPE.TABLE, 288 data_frame=pd.DataFrame([ 289 { 290 'Field': x, 291 'Type': 'str' 292 } for x in file_meta['columns'] 293 ]) 294 ) 295 return result 296 [end of mindsdb/integrations/handlers/file_handler/file_handler.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/mindsdb/integrations/handlers/file_handler/file_handler.py b/mindsdb/integrations/handlers/file_handler/file_handler.py --- a/mindsdb/integrations/handlers/file_handler/file_handler.py +++ b/mindsdb/integrations/handlers/file_handler/file_handler.py @@ -128,6 +128,7 @@ else: file_list_data = file_data + header = [x.strip() for x in header] col_map = dict((col, col) for col in header) return pd.DataFrame(file_list_data, columns=header), col_map @@ -287,7 +288,7 @@ RESPONSE_TYPE.TABLE, data_frame=pd.DataFrame([ { - 'Field': x, + 'Field': x.strip(), 'Type': 'str' } for x in file_meta['columns'] ])
{"golden_diff": "diff --git a/mindsdb/integrations/handlers/file_handler/file_handler.py b/mindsdb/integrations/handlers/file_handler/file_handler.py\n--- a/mindsdb/integrations/handlers/file_handler/file_handler.py\n+++ b/mindsdb/integrations/handlers/file_handler/file_handler.py\n@@ -128,6 +128,7 @@\n else:\n file_list_data = file_data\n \n+ header = [x.strip() for x in header]\n col_map = dict((col, col) for col in header)\n return pd.DataFrame(file_list_data, columns=header), col_map\n \n@@ -287,7 +288,7 @@\n RESPONSE_TYPE.TABLE,\n data_frame=pd.DataFrame([\n {\n- 'Field': x,\n+ 'Field': x.strip(),\n 'Type': 'str'\n } for x in file_meta['columns']\n ])\n", "issue": "Error when csv header has empty space\n### Discussed in https://github.com/mindsdb/mindsdb/discussions/2343\r\n\r\n<div type='discussions-op-text'>\r\n\r\n<sup>Originally posted by **Ashok0** June 13, 2022</sup>\r\nHi,\r\n\r\nI have a question regarding usage of imported data with MindsDB. I registered a MindsDB account and selected \"Add Data\" > \"Import File\" and uploaded some sample CSV data to a table. I can view the data just fine as shown in the following screencap: https://imgur.com/a/X6mrzHt\r\n\r\nIf I attempt to select an individual column i.e. 'sensitivity' with the following statement: **SELECT sensitivity FROM files.bedcheck**, \r\nI receive the following error message: **Binder Error: Referenced column \"sensitivity\" not found in FROM clause! Candidate bindings: \"df_table. sensitivity\" LINE 1: SELECT sensitivity AS sensitivity**\r\n\r\nAny assistance in how I can resolve this error would be greatly appreciated. \r\n\r\nThank you!</div>\n", "before_files": [{"content": "from io import BytesIO, StringIO\nimport os\nimport csv\nimport json\nimport codecs\nimport traceback\nimport tempfile\nfrom urllib.parse import urlparse\n\nimport requests\nimport pandas as pd\n\nfrom mindsdb_sql import parse_sql\nfrom mindsdb_sql.parser.ast.base import ASTNode\nfrom mindsdb_sql.parser.ast import DropTables, Select\n\nfrom mindsdb.api.mysql.mysql_proxy.utilities.sql import query_df\nfrom mindsdb.integrations.libs.base_handler import DatabaseHandler\nfrom mindsdb.integrations.libs.response import (\n HandlerStatusResponse as StatusResponse,\n HandlerResponse as Response,\n RESPONSE_TYPE\n)\n\n\ndef clean_row(row):\n n_row = []\n for cell in row:\n if str(cell) in ['', ' ', ' ', 'NaN', 'nan', 'NA']:\n n_row.append(None)\n else:\n n_row.append(cell)\n\n return n_row\n\n\nclass FileHandler(DatabaseHandler):\n \"\"\"\n Handler for files\n \"\"\"\n type = 'files'\n\n def __init__(self, name=None, db_store=None, fs_store=None, connection_data=None, file_controller=None):\n super().__init__(name)\n self.parser = parse_sql\n self.fs_store = fs_store\n self.custom_parser = connection_data.get('custom_parser')\n self.clean_rows = connection_data.get('clean_rows', True)\n self.file_controller = file_controller\n\n def connect(self, **kwargs):\n return\n\n def disconnect(self, **kwargs):\n return\n\n def check_connection(self) -> StatusResponse:\n return StatusResponse(True)\n\n def query(self, query: ASTNode) -> Response:\n if type(query) == DropTables:\n for table_identifier in query.tables:\n if len(table_identifier.parts) == 2 and table_identifier.parts[0] != self.name:\n return Response(\n RESPONSE_TYPE.ERROR,\n error_message=f\"Can't delete table from database '{table_identifier.parts[0]}'\"\n )\n table_name = table_identifier.parts[-1]\n try:\n self.file_controller.delete_file(table_name)\n except Exception as e:\n return Response(\n RESPONSE_TYPE.ERROR,\n error_message=f\"Can't delete table '{table_name}': {e}\"\n )\n return Response(RESPONSE_TYPE.OK)\n elif type(query) == Select:\n table_name = query.from_table.parts[-1]\n file_path = self.file_controller.get_file_path(table_name, company_id=None)\n df, _columns = self._handle_source(file_path, self.clean_rows, self.custom_parser)\n result_df = query_df(df, query)\n return Response(\n RESPONSE_TYPE.TABLE,\n data_frame=result_df\n )\n else:\n return Response(\n RESPONSE_TYPE.ERROR,\n error_message=\"Only 'select' and 'drop' queries allowed for files\"\n )\n\n def native_query(self, query: str) -> Response:\n ast = self.parser(query, dialect='mindsdb')\n return self.query(ast)\n\n @staticmethod\n def _handle_source(file_path, clean_rows=True, custom_parser=None):\n # get file data io, format and dialect\n data, fmt, dialect = FileHandler._get_data_io(file_path)\n data.seek(0) # make sure we are at 0 in file pointer\n\n if custom_parser:\n header, file_data = custom_parser(data, fmt)\n\n elif fmt == 'csv':\n csv_reader = list(csv.reader(data, dialect))\n header = csv_reader[0]\n file_data = csv_reader[1:]\n\n elif fmt in ['xlsx', 'xls']:\n data.seek(0)\n df = pd.read_excel(data)\n header = df.columns.values.tolist()\n file_data = df.values.tolist()\n\n elif fmt == 'json':\n data.seek(0)\n json_doc = json.loads(data.read())\n df = pd.json_normalize(json_doc, max_level=0)\n header = df.columns.values.tolist()\n file_data = df.values.tolist()\n\n else:\n raise ValueError('Could not load file into any format, supported formats are csv, json, xls, xlsx')\n\n if clean_rows:\n file_list_data = [clean_row(row) for row in file_data]\n else:\n file_list_data = file_data\n\n col_map = dict((col, col) for col in header)\n return pd.DataFrame(file_list_data, columns=header), col_map\n\n @staticmethod\n def _get_data_io(file_path):\n \"\"\"\n This gets a file either url or local file and defiens what the format is as well as dialect\n :param file: file path or url\n :return: data_io, format, dialect\n \"\"\"\n\n ############\n # get file as io object\n ############\n\n # file_path = self._get_file_path()\n\n data = BytesIO()\n\n try:\n with open(file_path, 'rb') as fp:\n data = BytesIO(fp.read())\n except Exception as e:\n error = 'Could not load file, possible exception : {exception}'.format(exception=e)\n print(error)\n raise ValueError(error)\n\n dialect = None\n\n ############\n # check for file type\n ############\n\n # try to guess if its an excel file\n xlsx_sig = b'\\x50\\x4B\\x05\\06'\n # xlsx_sig2 = b'\\x50\\x4B\\x03\\x04'\n xls_sig = b'\\x09\\x08\\x10\\x00\\x00\\x06\\x05\\x00'\n\n # different whence, offset, size for different types\n excel_meta = [('xls', 0, 512, 8), ('xlsx', 2, -22, 4)]\n\n for filename, whence, offset, size in excel_meta:\n\n try:\n data.seek(offset, whence) # Seek to the offset.\n bytes = data.read(size) # Capture the specified number of bytes.\n data.seek(0)\n codecs.getencoder('hex')(bytes)\n\n if bytes == xls_sig:\n return data, 'xls', dialect\n elif bytes == xlsx_sig:\n return data, 'xlsx', dialect\n\n except Exception:\n data.seek(0)\n\n # if not excel it can be a json file or a CSV, convert from binary to stringio\n\n byte_str = data.read()\n # Move it to StringIO\n try:\n # Handle Microsoft's BOM \"special\" UTF-8 encoding\n if byte_str.startswith(codecs.BOM_UTF8):\n data = StringIO(byte_str.decode('utf-8-sig'))\n else:\n data = StringIO(byte_str.decode('utf-8'))\n\n except Exception:\n print(traceback.format_exc())\n print('Could not load into string')\n\n # see if its JSON\n buffer = data.read(100)\n data.seek(0)\n text = buffer.strip()\n # analyze first n characters\n if len(text) > 0:\n text = text.strip()\n # it it looks like a json, then try to parse it\n if text.startswith('{') or text.startswith('['):\n try:\n json.loads(data.read())\n data.seek(0)\n return data, 'json', dialect\n except Exception:\n data.seek(0)\n return data, None, dialect\n\n # lets try to figure out if its a csv\n try:\n dialect = FileHandler._get_csv_dialect(file_path)\n if dialect:\n return data, 'csv', dialect\n return data, None, dialect\n except Exception:\n data.seek(0)\n print('Could not detect format for this file')\n print(traceback.format_exc())\n # No file type identified\n return data, None, dialect\n\n @staticmethod\n def _get_file_path(path) -> str:\n try:\n is_url = urlparse(path).scheme in ('http', 'https')\n except Exception:\n is_url = False\n if is_url:\n path = FileHandler._fetch_url(path)\n return path\n\n @staticmethod\n def _get_csv_dialect(file_path) -> csv.Dialect:\n with open(file_path, 'rt') as f:\n try:\n accepted_csv_delimiters = [',', '\\t', ';']\n dialect = csv.Sniffer().sniff(f.read(128 * 1024), delimiters=accepted_csv_delimiters)\n except csv.Error:\n dialect = None\n return dialect\n\n @staticmethod\n def _fetch_url(url: str) -> str:\n temp_dir = tempfile.mkdtemp(prefix='mindsdb_file_url_')\n try:\n r = requests.get(url, stream=True)\n if r.status_code == 200:\n with open(os.path.join(temp_dir, 'file'), 'wb') as f:\n for chunk in r:\n f.write(chunk)\n else:\n raise Exception(f'Responce status code is {r.status_code}')\n except Exception as e:\n print(f'Error during getting {url}')\n print(e)\n raise\n return os.path.join(temp_dir, 'file')\n\n def get_tables(self) -> Response:\n \"\"\"\n List all files\n \"\"\"\n files_meta = self.file_controller.get_files()\n data = [{\n 'TABLE_NAME': x['name'],\n 'TABLE_ROWS': x['row_count']\n } for x in files_meta]\n return Response(\n RESPONSE_TYPE.TABLE,\n data_frame=pd.DataFrame(data)\n )\n\n def get_columns(self, table_name) -> Response:\n file_meta = self.file_controller.get_file_meta(table_name)\n result = Response(\n RESPONSE_TYPE.TABLE,\n data_frame=pd.DataFrame([\n {\n 'Field': x,\n 'Type': 'str'\n } for x in file_meta['columns']\n ])\n )\n return result\n", "path": "mindsdb/integrations/handlers/file_handler/file_handler.py"}]}
3,691
197
gh_patches_debug_2174
rasdani/github-patches
git_diff
fedora-infra__bodhi-1042
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> GET on /masher/ errors with a 404 response GET on `/masher` works, but GET on `/masher/` returns an HTTP 404 error code. We should add a route to allow the trailing slash on this URL. </issue> <code> [start of bodhi/server/__init__.py] 1 # This program is free software; you can redistribute it and/or 2 # modify it under the terms of the GNU General Public License 3 # as published by the Free Software Foundation; either version 2 4 # of the License, or (at your option) any later version. 5 # 6 # This program is distributed in the hope that it will be useful, 7 # but WITHOUT ANY WARRANTY; without even the implied warranty of 8 # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the 9 # GNU General Public License for more details. 10 # 11 # You should have received a copy of the GNU General Public License 12 # along with this program; if not, write to the Free Software 13 # Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. 14 15 from collections import defaultdict 16 from dogpile.cache import make_region 17 from munch import munchify 18 from sqlalchemy import engine_from_config 19 from sqlalchemy.orm import scoped_session, sessionmaker 20 from zope.sqlalchemy import ZopeTransactionExtension 21 22 from pyramid.settings import asbool 23 from pyramid.security import unauthenticated_userid 24 from pyramid.config import Configurator 25 from pyramid.authentication import AuthTktAuthenticationPolicy 26 from pyramid.authorization import ACLAuthorizationPolicy 27 from pyramid.renderers import JSONP 28 from pyramid.exceptions import HTTPForbidden 29 30 from . import buildsys 31 32 import logging 33 34 log = logging.getLogger(__name__) 35 36 37 # TODO -- someday move this externally to "fedora_flavored_markdown" 38 from bodhi.server import ffmarkdown 39 ffmarkdown.inject() 40 41 # 42 # Request methods 43 # 44 45 def get_dbsession(request): 46 engine = engine_from_config(request.registry.settings, 'sqlalchemy.') 47 Sess = scoped_session(sessionmaker(extension=ZopeTransactionExtension())) 48 Sess.configure(bind=engine) 49 session = Sess() 50 51 def cleanup(request): 52 ## No need to do rollback/commit ourselves. the zope transaction 53 ## manager takes care of that for us... 54 #if request.exception is not None: 55 # session.rollback() 56 #else: 57 # session.commit() 58 ## However, we may still want to explicitly close the session we opened 59 #log.debug("Closing session at the end of a request.") 60 session.close() 61 62 request.add_finished_callback(cleanup) 63 64 return session 65 66 67 def get_cacheregion(request): 68 region = make_region() 69 region.configure_from_config(request.registry.settings, "dogpile.cache.") 70 return region 71 72 73 def get_user(request): 74 from bodhi.server.models import User 75 userid = unauthenticated_userid(request) 76 if userid is not None: 77 user = request.db.query(User).filter_by(name=unicode(userid)).first() 78 # Why munch? https://github.com/fedora-infra/bodhi/issues/473 79 return munchify(user.__json__(request=request)) 80 81 82 def groupfinder(userid, request): 83 from bodhi.server.models import User 84 if request.user: 85 user = User.get(request.user.name, request.db) 86 return ['group:' + group.name for group in user.groups] 87 88 89 def get_koji(request): 90 return buildsys.get_session() 91 92 93 def get_buildinfo(request): 94 """ 95 A per-request cache populated by the validators and shared with the views 96 to store frequently used package-specific data, like build tags and ACLs. 97 """ 98 return defaultdict(dict) 99 100 101 def get_releases(request): 102 from bodhi.server.models import Release 103 return Release.all_releases(request.db) 104 105 # 106 # Cornice filters 107 # 108 109 def exception_filter(response, request): 110 """Log exceptions that get thrown up to cornice""" 111 if isinstance(response, Exception): 112 log.exception('Unhandled exception raised: %r' % response) 113 return response 114 115 from cornice.validators import DEFAULT_FILTERS 116 DEFAULT_FILTERS.insert(0, exception_filter) 117 118 119 # 120 # Bodhi initialization 121 # 122 123 def main(global_config, testing=None, session=None, **settings): 124 """ This function returns a WSGI application """ 125 # Setup our buildsystem 126 buildsys.setup_buildsystem(settings) 127 128 # Sessions & Caching 129 from pyramid.session import SignedCookieSessionFactory 130 session_factory = SignedCookieSessionFactory(settings['session.secret']) 131 132 # Construct a list of all groups we're interested in 133 default = ' '.join([settings.get(key, '') for key in [ 134 'important_groups', 135 'admin_packager_groups', 136 'mandatory_packager_groups', 137 'admin_groups', 138 ]]) 139 # pyramid_fas_openid looks for this setting 140 settings['openid.groups'] = settings.get('openid.groups', default).split() 141 142 config = Configurator(settings=settings, session_factory=session_factory) 143 144 # Plugins 145 config.include('pyramid_mako') 146 config.include('cornice') 147 148 # Lazy-loaded memoized request properties 149 if session: 150 config.add_request_method(lambda _: session, 'db', reify=True) 151 else: 152 config.add_request_method(get_dbsession, 'db', reify=True) 153 154 config.add_request_method(get_user, 'user', reify=True) 155 config.add_request_method(get_koji, 'koji', reify=True) 156 config.add_request_method(get_cacheregion, 'cache', reify=True) 157 config.add_request_method(get_buildinfo, 'buildinfo', reify=True) 158 config.add_request_method(get_releases, 'releases', reify=True) 159 160 # Templating 161 config.add_mako_renderer('.html', settings_prefix='mako.') 162 config.add_static_view('static', 'bodhi:server/static') 163 164 from bodhi.server.renderers import rss, jpeg 165 config.add_renderer('rss', rss) 166 config.add_renderer('jpeg', jpeg) 167 config.add_renderer('jsonp', JSONP(param_name='callback')) 168 169 # i18n 170 config.add_translation_dirs('bodhi:server/locale/') 171 172 # Authentication & Authorization 173 if testing: 174 # use a permissive security policy while running unit tests 175 config.testing_securitypolicy(userid=testing, permissive=True) 176 else: 177 config.set_authentication_policy(AuthTktAuthenticationPolicy( 178 settings['authtkt.secret'], 179 callback=groupfinder, 180 secure=asbool(settings['authtkt.secure']), 181 hashalg='sha512')) 182 config.set_authorization_policy(ACLAuthorizationPolicy()) 183 184 # Frontpage 185 config.add_route('home', '/') 186 187 # Views for creating new objects 188 config.add_route('new_update', '/updates/new') 189 config.add_route('new_override', '/overrides/new') 190 config.add_route('new_stack', '/stacks/new') 191 192 # Metrics 193 config.add_route('metrics', '/metrics') 194 config.add_route('masher_status', '/masher') 195 196 # Auto-completion search 197 config.add_route('search_packages', '/search/packages') 198 config.add_route('latest_candidates', '/latest_candidates') 199 config.add_route('latest_builds', '/latest_builds') 200 201 config.add_route('captcha_image', '/captcha/{cipherkey}/') 202 203 # pyramid.openid 204 config.add_route('login', '/login') 205 config.add_view('bodhi.server.security.login', route_name='login') 206 config.add_view('bodhi.server.security.login', context=HTTPForbidden) 207 config.add_route('logout', '/logout') 208 config.add_view('bodhi.server.security.logout', route_name='logout') 209 config.add_route('verify_openid', pattern='/dologin.html') 210 config.add_view('pyramid_fas_openid.verify_openid', route_name='verify_openid') 211 212 config.add_route('api_version', '/api_version') 213 214 # The only user preference we have. 215 config.add_route('popup_toggle', '/popup_toggle') 216 217 config.scan('bodhi.server.views') 218 config.scan('bodhi.server.services') 219 config.scan('bodhi.server.captcha') 220 config.scan('bodhi.server.events') 221 222 return config.make_wsgi_app() 223 [end of bodhi/server/__init__.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/bodhi/server/__init__.py b/bodhi/server/__init__.py --- a/bodhi/server/__init__.py +++ b/bodhi/server/__init__.py @@ -191,7 +191,7 @@ # Metrics config.add_route('metrics', '/metrics') - config.add_route('masher_status', '/masher') + config.add_route('masher_status', '/masher/') # Auto-completion search config.add_route('search_packages', '/search/packages')
{"golden_diff": "diff --git a/bodhi/server/__init__.py b/bodhi/server/__init__.py\n--- a/bodhi/server/__init__.py\n+++ b/bodhi/server/__init__.py\n@@ -191,7 +191,7 @@\n \n # Metrics\n config.add_route('metrics', '/metrics')\n- config.add_route('masher_status', '/masher')\n+ config.add_route('masher_status', '/masher/')\n \n # Auto-completion search\n config.add_route('search_packages', '/search/packages')\n", "issue": "GET on /masher/ errors with a 404 response\nGET on `/masher` works, but GET on `/masher/` returns an HTTP 404 error code. We should add a route to allow the trailing slash on this URL.\n\n", "before_files": [{"content": "# This program is free software; you can redistribute it and/or\n# modify it under the terms of the GNU General Public License\n# as published by the Free Software Foundation; either version 2\n# of the License, or (at your option) any later version.\n#\n# This program is distributed in the hope that it will be useful,\n# but WITHOUT ANY WARRANTY; without even the implied warranty of\n# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the\n# GNU General Public License for more details.\n#\n# You should have received a copy of the GNU General Public License\n# along with this program; if not, write to the Free Software\n# Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA.\n\nfrom collections import defaultdict\nfrom dogpile.cache import make_region\nfrom munch import munchify\nfrom sqlalchemy import engine_from_config\nfrom sqlalchemy.orm import scoped_session, sessionmaker\nfrom zope.sqlalchemy import ZopeTransactionExtension\n\nfrom pyramid.settings import asbool\nfrom pyramid.security import unauthenticated_userid\nfrom pyramid.config import Configurator\nfrom pyramid.authentication import AuthTktAuthenticationPolicy\nfrom pyramid.authorization import ACLAuthorizationPolicy\nfrom pyramid.renderers import JSONP\nfrom pyramid.exceptions import HTTPForbidden\n\nfrom . import buildsys\n\nimport logging\n\nlog = logging.getLogger(__name__)\n\n\n# TODO -- someday move this externally to \"fedora_flavored_markdown\"\nfrom bodhi.server import ffmarkdown\nffmarkdown.inject()\n\n#\n# Request methods\n#\n\ndef get_dbsession(request):\n engine = engine_from_config(request.registry.settings, 'sqlalchemy.')\n Sess = scoped_session(sessionmaker(extension=ZopeTransactionExtension()))\n Sess.configure(bind=engine)\n session = Sess()\n\n def cleanup(request):\n ## No need to do rollback/commit ourselves. the zope transaction\n ## manager takes care of that for us...\n #if request.exception is not None:\n # session.rollback()\n #else:\n # session.commit()\n ## However, we may still want to explicitly close the session we opened\n #log.debug(\"Closing session at the end of a request.\")\n session.close()\n\n request.add_finished_callback(cleanup)\n\n return session\n\n\ndef get_cacheregion(request):\n region = make_region()\n region.configure_from_config(request.registry.settings, \"dogpile.cache.\")\n return region\n\n\ndef get_user(request):\n from bodhi.server.models import User\n userid = unauthenticated_userid(request)\n if userid is not None:\n user = request.db.query(User).filter_by(name=unicode(userid)).first()\n # Why munch? https://github.com/fedora-infra/bodhi/issues/473\n return munchify(user.__json__(request=request))\n\n\ndef groupfinder(userid, request):\n from bodhi.server.models import User\n if request.user:\n user = User.get(request.user.name, request.db)\n return ['group:' + group.name for group in user.groups]\n\n\ndef get_koji(request):\n return buildsys.get_session()\n\n\ndef get_buildinfo(request):\n \"\"\"\n A per-request cache populated by the validators and shared with the views\n to store frequently used package-specific data, like build tags and ACLs.\n \"\"\"\n return defaultdict(dict)\n\n\ndef get_releases(request):\n from bodhi.server.models import Release\n return Release.all_releases(request.db)\n\n#\n# Cornice filters\n#\n\ndef exception_filter(response, request):\n \"\"\"Log exceptions that get thrown up to cornice\"\"\"\n if isinstance(response, Exception):\n log.exception('Unhandled exception raised: %r' % response)\n return response\n\nfrom cornice.validators import DEFAULT_FILTERS\nDEFAULT_FILTERS.insert(0, exception_filter)\n\n\n#\n# Bodhi initialization\n#\n\ndef main(global_config, testing=None, session=None, **settings):\n \"\"\" This function returns a WSGI application \"\"\"\n # Setup our buildsystem\n buildsys.setup_buildsystem(settings)\n\n # Sessions & Caching\n from pyramid.session import SignedCookieSessionFactory\n session_factory = SignedCookieSessionFactory(settings['session.secret'])\n\n # Construct a list of all groups we're interested in\n default = ' '.join([settings.get(key, '') for key in [\n 'important_groups',\n 'admin_packager_groups',\n 'mandatory_packager_groups',\n 'admin_groups',\n ]])\n # pyramid_fas_openid looks for this setting\n settings['openid.groups'] = settings.get('openid.groups', default).split()\n\n config = Configurator(settings=settings, session_factory=session_factory)\n\n # Plugins\n config.include('pyramid_mako')\n config.include('cornice')\n\n # Lazy-loaded memoized request properties\n if session:\n config.add_request_method(lambda _: session, 'db', reify=True)\n else:\n config.add_request_method(get_dbsession, 'db', reify=True)\n\n config.add_request_method(get_user, 'user', reify=True)\n config.add_request_method(get_koji, 'koji', reify=True)\n config.add_request_method(get_cacheregion, 'cache', reify=True)\n config.add_request_method(get_buildinfo, 'buildinfo', reify=True)\n config.add_request_method(get_releases, 'releases', reify=True)\n\n # Templating\n config.add_mako_renderer('.html', settings_prefix='mako.')\n config.add_static_view('static', 'bodhi:server/static')\n\n from bodhi.server.renderers import rss, jpeg\n config.add_renderer('rss', rss)\n config.add_renderer('jpeg', jpeg)\n config.add_renderer('jsonp', JSONP(param_name='callback'))\n\n # i18n\n config.add_translation_dirs('bodhi:server/locale/')\n\n # Authentication & Authorization\n if testing:\n # use a permissive security policy while running unit tests\n config.testing_securitypolicy(userid=testing, permissive=True)\n else:\n config.set_authentication_policy(AuthTktAuthenticationPolicy(\n settings['authtkt.secret'],\n callback=groupfinder,\n secure=asbool(settings['authtkt.secure']),\n hashalg='sha512'))\n config.set_authorization_policy(ACLAuthorizationPolicy())\n\n # Frontpage\n config.add_route('home', '/')\n\n # Views for creating new objects\n config.add_route('new_update', '/updates/new')\n config.add_route('new_override', '/overrides/new')\n config.add_route('new_stack', '/stacks/new')\n\n # Metrics\n config.add_route('metrics', '/metrics')\n config.add_route('masher_status', '/masher')\n\n # Auto-completion search\n config.add_route('search_packages', '/search/packages')\n config.add_route('latest_candidates', '/latest_candidates')\n config.add_route('latest_builds', '/latest_builds')\n\n config.add_route('captcha_image', '/captcha/{cipherkey}/')\n\n # pyramid.openid\n config.add_route('login', '/login')\n config.add_view('bodhi.server.security.login', route_name='login')\n config.add_view('bodhi.server.security.login', context=HTTPForbidden)\n config.add_route('logout', '/logout')\n config.add_view('bodhi.server.security.logout', route_name='logout')\n config.add_route('verify_openid', pattern='/dologin.html')\n config.add_view('pyramid_fas_openid.verify_openid', route_name='verify_openid')\n\n config.add_route('api_version', '/api_version')\n\n # The only user preference we have.\n config.add_route('popup_toggle', '/popup_toggle')\n\n config.scan('bodhi.server.views')\n config.scan('bodhi.server.services')\n config.scan('bodhi.server.captcha')\n config.scan('bodhi.server.events')\n\n return config.make_wsgi_app()\n", "path": "bodhi/server/__init__.py"}]}
2,857
121
gh_patches_debug_25484
rasdani/github-patches
git_diff
dj-stripe__dj-stripe-1259
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> DJStripeSubscriptionPermission issue returning bool This permission is not returning properly the bool. **Current behaviour** ```python class DJStripeSubscriptionPermission(BasePermission): """ A permission to be used when wanting to permit users with active subscriptions. """ def has_permission(self, request, view): """ Check if the subscriber has an active subscription. Returns false if: * a subscriber isn't passed through the request See ``utils.subscriber_has_active_subscription`` for more rules. """ try: subscriber_has_active_subscription(subscriber_request_callback(request)) except AttributeError: return False ``` Here is not returning True or False except if it falls in the exception. **Expected Behaviour** ```python class DJStripeSubscriptionPermission(BasePermission): """ A permission to be used when wanting to permit users with active subscriptions. """ def has_permission(self, request, view): """ Check if the subscriber has an active subscription. Returns false if: * a subscriber isn't passed through the request See ``utils.subscriber_has_active_subscription`` for more rules. """ try: return bool(subscriber_has_active_subscription(subscriber_request_callback(request))) except AttributeError: return False ``` Just missing a return and it solves the problem. We don't need a bool directly there, I just added just to follow the same patterns as the DRF (also being added to the other project :-)) </issue> <code> [start of djstripe/contrib/rest_framework/serializers.py] 1 """ 2 .. module:: dj-stripe.contrib.rest_framework.serializers. 3 4 :synopsis: dj-stripe - Serializers to be used with the dj-stripe REST API. 5 6 .. moduleauthor:: Philippe Luickx (@philippeluickx) 7 8 """ 9 10 from rest_framework import serializers 11 from rest_framework.serializers import ModelSerializer 12 13 from djstripe.models import Subscription 14 15 16 class SubscriptionSerializer(ModelSerializer): 17 """A serializer used for the Subscription model.""" 18 19 class Meta: 20 """Model class options.""" 21 22 model = Subscription 23 exclude = ["default_tax_rates"] 24 25 26 class CreateSubscriptionSerializer(serializers.Serializer): 27 """A serializer used to create a Subscription.""" 28 29 stripe_token = serializers.CharField(max_length=200) 30 plan = serializers.CharField(max_length=50) 31 charge_immediately = serializers.BooleanField(required=False, allow_null=True, default=None) 32 tax_percent = serializers.DecimalField( 33 required=False, max_digits=5, decimal_places=2 34 ) 35 [end of djstripe/contrib/rest_framework/serializers.py] [start of djstripe/contrib/rest_framework/permissions.py] 1 """ 2 .. module:: dj-stripe.contrib.rest_framework.permissions. 3 4 :synopsis: dj-stripe - Permissions to be used with the dj-stripe REST API. 5 6 .. moduleauthor:: @kavdev, @pydanny 7 8 """ 9 from rest_framework.permissions import BasePermission 10 11 from ...settings import subscriber_request_callback 12 from ...utils import subscriber_has_active_subscription 13 14 15 class DJStripeSubscriptionPermission(BasePermission): 16 """ 17 A permission to be used when wanting to permit users with active subscriptions. 18 """ 19 20 def has_permission(self, request, view): 21 """ 22 Check if the subscriber has an active subscription. 23 24 Returns false if: 25 * a subscriber isn't passed through the request 26 27 See ``utils.subscriber_has_active_subscription`` for more rules. 28 29 """ 30 try: 31 subscriber_has_active_subscription(subscriber_request_callback(request)) 32 except AttributeError: 33 return False 34 [end of djstripe/contrib/rest_framework/permissions.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/djstripe/contrib/rest_framework/permissions.py b/djstripe/contrib/rest_framework/permissions.py --- a/djstripe/contrib/rest_framework/permissions.py +++ b/djstripe/contrib/rest_framework/permissions.py @@ -17,7 +17,7 @@ A permission to be used when wanting to permit users with active subscriptions. """ - def has_permission(self, request, view): + def has_permission(self, request, view) -> bool: """ Check if the subscriber has an active subscription. @@ -28,6 +28,8 @@ """ try: - subscriber_has_active_subscription(subscriber_request_callback(request)) + return subscriber_has_active_subscription( + subscriber_request_callback(request) + ) except AttributeError: return False diff --git a/djstripe/contrib/rest_framework/serializers.py b/djstripe/contrib/rest_framework/serializers.py --- a/djstripe/contrib/rest_framework/serializers.py +++ b/djstripe/contrib/rest_framework/serializers.py @@ -28,7 +28,9 @@ stripe_token = serializers.CharField(max_length=200) plan = serializers.CharField(max_length=50) - charge_immediately = serializers.BooleanField(required=False, allow_null=True, default=None) + charge_immediately = serializers.BooleanField( + required=False, allow_null=True, default=None + ) tax_percent = serializers.DecimalField( required=False, max_digits=5, decimal_places=2 )
{"golden_diff": "diff --git a/djstripe/contrib/rest_framework/permissions.py b/djstripe/contrib/rest_framework/permissions.py\n--- a/djstripe/contrib/rest_framework/permissions.py\n+++ b/djstripe/contrib/rest_framework/permissions.py\n@@ -17,7 +17,7 @@\n A permission to be used when wanting to permit users with active subscriptions.\n \"\"\"\n \n- def has_permission(self, request, view):\n+ def has_permission(self, request, view) -> bool:\n \"\"\"\n Check if the subscriber has an active subscription.\n \n@@ -28,6 +28,8 @@\n \n \"\"\"\n try:\n- subscriber_has_active_subscription(subscriber_request_callback(request))\n+ return subscriber_has_active_subscription(\n+ subscriber_request_callback(request)\n+ )\n except AttributeError:\n return False\ndiff --git a/djstripe/contrib/rest_framework/serializers.py b/djstripe/contrib/rest_framework/serializers.py\n--- a/djstripe/contrib/rest_framework/serializers.py\n+++ b/djstripe/contrib/rest_framework/serializers.py\n@@ -28,7 +28,9 @@\n \n stripe_token = serializers.CharField(max_length=200)\n plan = serializers.CharField(max_length=50)\n- charge_immediately = serializers.BooleanField(required=False, allow_null=True, default=None)\n+ charge_immediately = serializers.BooleanField(\n+ required=False, allow_null=True, default=None\n+ )\n tax_percent = serializers.DecimalField(\n required=False, max_digits=5, decimal_places=2\n )\n", "issue": "DJStripeSubscriptionPermission issue returning bool\nThis permission is not returning properly the bool.\r\n\r\n**Current behaviour**\r\n\r\n```python\r\nclass DJStripeSubscriptionPermission(BasePermission):\r\n \"\"\"\r\n A permission to be used when wanting to permit users with active subscriptions.\r\n \"\"\"\r\n\r\n def has_permission(self, request, view):\r\n \"\"\"\r\n Check if the subscriber has an active subscription.\r\n\r\n Returns false if:\r\n * a subscriber isn't passed through the request\r\n\r\n See ``utils.subscriber_has_active_subscription`` for more rules.\r\n\r\n \"\"\"\r\n try:\r\n subscriber_has_active_subscription(subscriber_request_callback(request))\r\n except AttributeError:\r\n return False\r\n```\r\n\r\nHere is not returning True or False except if it falls in the exception.\r\n\r\n\r\n**Expected Behaviour**\r\n\r\n\r\n```python\r\nclass DJStripeSubscriptionPermission(BasePermission):\r\n \"\"\"\r\n A permission to be used when wanting to permit users with active subscriptions.\r\n \"\"\"\r\n\r\n def has_permission(self, request, view):\r\n \"\"\"\r\n Check if the subscriber has an active subscription.\r\n\r\n Returns false if:\r\n * a subscriber isn't passed through the request\r\n\r\n See ``utils.subscriber_has_active_subscription`` for more rules.\r\n\r\n \"\"\"\r\n try:\r\n return bool(subscriber_has_active_subscription(subscriber_request_callback(request)))\r\n except AttributeError:\r\n return False\r\n```\r\n\r\nJust missing a return and it solves the problem. We don't need a bool directly there, I just added just to follow the same patterns as the DRF (also being added to the other project :-))\n", "before_files": [{"content": "\"\"\"\n.. module:: dj-stripe.contrib.rest_framework.serializers.\n\n :synopsis: dj-stripe - Serializers to be used with the dj-stripe REST API.\n\n.. moduleauthor:: Philippe Luickx (@philippeluickx)\n\n\"\"\"\n\nfrom rest_framework import serializers\nfrom rest_framework.serializers import ModelSerializer\n\nfrom djstripe.models import Subscription\n\n\nclass SubscriptionSerializer(ModelSerializer):\n \"\"\"A serializer used for the Subscription model.\"\"\"\n\n class Meta:\n \"\"\"Model class options.\"\"\"\n\n model = Subscription\n exclude = [\"default_tax_rates\"]\n\n\nclass CreateSubscriptionSerializer(serializers.Serializer):\n \"\"\"A serializer used to create a Subscription.\"\"\"\n\n stripe_token = serializers.CharField(max_length=200)\n plan = serializers.CharField(max_length=50)\n charge_immediately = serializers.BooleanField(required=False, allow_null=True, default=None)\n tax_percent = serializers.DecimalField(\n required=False, max_digits=5, decimal_places=2\n )\n", "path": "djstripe/contrib/rest_framework/serializers.py"}, {"content": "\"\"\"\n.. module:: dj-stripe.contrib.rest_framework.permissions.\n\n :synopsis: dj-stripe - Permissions to be used with the dj-stripe REST API.\n\n.. moduleauthor:: @kavdev, @pydanny\n\n\"\"\"\nfrom rest_framework.permissions import BasePermission\n\nfrom ...settings import subscriber_request_callback\nfrom ...utils import subscriber_has_active_subscription\n\n\nclass DJStripeSubscriptionPermission(BasePermission):\n \"\"\"\n A permission to be used when wanting to permit users with active subscriptions.\n \"\"\"\n\n def has_permission(self, request, view):\n \"\"\"\n Check if the subscriber has an active subscription.\n\n Returns false if:\n * a subscriber isn't passed through the request\n\n See ``utils.subscriber_has_active_subscription`` for more rules.\n\n \"\"\"\n try:\n subscriber_has_active_subscription(subscriber_request_callback(request))\n except AttributeError:\n return False\n", "path": "djstripe/contrib/rest_framework/permissions.py"}]}
1,385
334
gh_patches_debug_9096
rasdani/github-patches
git_diff
ivy-llc__ivy-26785
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> Remove duplicate-keys in `ivy\functional\backends\numpy\__init__.py` In the file: `ivy\functional\backends\numpy\__init__.py` duplicate keys are present in the following lines https://github.com/unifyai/ivy/blob/1be0f8c2c8d916b10f06ab538b43fc506a1992a6/ivy/functional/backends/numpy/__init__.py#L25 https://github.com/unifyai/ivy/blob/1be0f8c2c8d916b10f06ab538b43fc506a1992a6/ivy/functional/backends/numpy/__init__.py#L40 https://github.com/unifyai/ivy/blob/1be0f8c2c8d916b10f06ab538b43fc506a1992a6/ivy/functional/backends/numpy/__init__.py#L31 https://github.com/unifyai/ivy/blob/1be0f8c2c8d916b10f06ab538b43fc506a1992a6/ivy/functional/backends/numpy/__init__.py#L37 They can be removed. </issue> <code> [start of ivy/functional/backends/numpy/__init__.py] 1 # global 2 import sys 3 import numpy as np 4 5 # local 6 import ivy 7 from ivy.func_wrapper import _dtype_from_version 8 9 backend_version = {"version": np.__version__} 10 11 # noinspection PyUnresolvedReferences 12 if not ivy.is_local(): 13 _module_in_memory = sys.modules[__name__] 14 else: 15 _module_in_memory = sys.modules[ivy.import_module_path].import_cache[__name__] 16 17 use = ivy.utils.backend.ContextManager(_module_in_memory) 18 19 # wrap __array_ufunc__ method of ivy.Array to prioritize Ivy array methods when using numpu backend 20 21 22 def wrap__array_ufunc__(func): 23 def rep_method(self, ufunc, method, *inputs, **kwargs): 24 methods = { 25 "not_equal": "not_equal", 26 "greater": "greater", 27 "less": "less", 28 "greater_equal": "greater_equal", 29 "less_equal": "less_equal", 30 "multiply": "multiply", 31 "divide": "divide", 32 "remainder": "remainder", 33 "equal": "equal", 34 "bitwise_and": "bitwise_and", 35 "matmul": "matmul", 36 "power": "pow", 37 "divide": "divide", 38 "subtract": "subtract", 39 "add": "add", 40 "not_equal": "not_equal", 41 } 42 if ufunc.__name__ in methods.keys(): 43 return eval("ivy." + methods[ufunc.__name__] + "(*inputs, **kwargs)") 44 return func(self, ufunc, method, *inputs, **kwargs) 45 46 return rep_method 47 48 49 ivy.Array.__array_ufunc__ = wrap__array_ufunc__(ivy.Array.__array_ufunc__) 50 51 NativeArray = np.ndarray 52 NativeDevice = str 53 NativeDtype = np.dtype 54 NativeShape = tuple 55 56 NativeSparseArray = None 57 58 59 # devices 60 valid_devices = ("cpu",) 61 62 invalid_devices = ("gpu", "tpu") 63 64 # native data types 65 native_int8 = np.dtype("int8") 66 native_int16 = np.dtype("int16") 67 native_int32 = np.dtype("int32") 68 native_int64 = np.dtype("int64") 69 native_uint8 = np.dtype("uint8") 70 native_uint16 = np.dtype("uint16") 71 native_uint32 = np.dtype("uint32") 72 native_uint64 = np.dtype("uint64") 73 native_float16 = np.dtype("float16") 74 native_float32 = np.dtype("float32") 75 native_float64 = np.dtype("float64") 76 native_complex64 = np.dtype("complex64") 77 native_complex128 = np.dtype("complex128") 78 native_double = native_float64 79 native_bool = np.dtype("bool") 80 81 # valid data types 82 # ToDo: Add complex dtypes to valid_dtypes and fix all resulting failures. 83 84 # update these to add new dtypes 85 valid_dtypes = { 86 "1.26.0 and below": ( 87 ivy.int8, 88 ivy.int16, 89 ivy.int32, 90 ivy.int64, 91 ivy.uint8, 92 ivy.uint16, 93 ivy.uint32, 94 ivy.uint64, 95 ivy.float16, 96 ivy.float32, 97 ivy.float64, 98 ivy.complex64, 99 ivy.complex128, 100 ivy.bool, 101 ) 102 } 103 valid_numeric_dtypes = { 104 "1.26.0 and below": ( 105 ivy.int8, 106 ivy.int16, 107 ivy.int32, 108 ivy.int64, 109 ivy.uint8, 110 ivy.uint16, 111 ivy.uint32, 112 ivy.uint64, 113 ivy.float16, 114 ivy.float32, 115 ivy.float64, 116 ivy.complex64, 117 ivy.complex128, 118 ) 119 } 120 valid_int_dtypes = { 121 "1.26.0 and below": ( 122 ivy.int8, 123 ivy.int16, 124 ivy.int32, 125 ivy.int64, 126 ivy.uint8, 127 ivy.uint16, 128 ivy.uint32, 129 ivy.uint64, 130 ) 131 } 132 valid_float_dtypes = {"1.26.0 and below": (ivy.float16, ivy.float32, ivy.float64)} 133 valid_uint_dtypes = { 134 "1.26.0 and below": (ivy.uint8, ivy.uint16, ivy.uint32, ivy.uint64) 135 } 136 valid_complex_dtypes = {"1.26.0 and below": (ivy.complex64, ivy.complex128)} 137 138 # leave these untouched 139 valid_dtypes = _dtype_from_version(valid_dtypes, backend_version) 140 valid_numeric_dtypes = _dtype_from_version(valid_numeric_dtypes, backend_version) 141 valid_int_dtypes = _dtype_from_version(valid_int_dtypes, backend_version) 142 valid_float_dtypes = _dtype_from_version(valid_float_dtypes, backend_version) 143 valid_uint_dtypes = _dtype_from_version(valid_uint_dtypes, backend_version) 144 valid_complex_dtypes = _dtype_from_version(valid_complex_dtypes, backend_version) 145 146 # invalid data types 147 # update these to add new dtypes 148 invalid_dtypes = {"1.26.0 and below": (ivy.bfloat16,)} 149 invalid_numeric_dtypes = {"1.26.0 and below": (ivy.bfloat16,)} 150 invalid_int_dtypes = {"1.26.0 and below": ()} 151 invalid_float_dtypes = {"1.26.0 and below": (ivy.bfloat16,)} 152 invalid_uint_dtypes = {"1.26.0 and below": ()} 153 invalid_complex_dtypes = {"1.26.0 and below": ()} 154 155 156 # leave these untouched 157 invalid_dtypes = _dtype_from_version(invalid_dtypes, backend_version) 158 invalid_numeric_dtypes = _dtype_from_version(invalid_numeric_dtypes, backend_version) 159 invalid_int_dtypes = _dtype_from_version(invalid_int_dtypes, backend_version) 160 invalid_float_dtypes = _dtype_from_version(invalid_float_dtypes, backend_version) 161 invalid_uint_dtypes = _dtype_from_version(invalid_uint_dtypes, backend_version) 162 invalid_complex_dtypes = _dtype_from_version(invalid_complex_dtypes, backend_version) 163 164 165 native_inplace_support = True 166 167 supports_gradients = False 168 169 170 def closest_valid_dtype(type=None, /, as_native=False): 171 if type is None: 172 type = ivy.default_dtype() 173 elif isinstance(type, str) and type in invalid_dtypes: 174 type = {"bfloat16": ivy.float16}[type] 175 return ivy.as_ivy_dtype(type) if not as_native else ivy.as_native_dtype(type) 176 177 178 backend = "numpy" 179 180 181 # local sub-modules 182 from . import activations 183 from .activations import * 184 from . import creation 185 from .creation import * 186 from . import data_type 187 from .data_type import * 188 from . import device 189 from .device import * 190 from . import elementwise 191 from .elementwise import * 192 from . import general 193 from .general import * 194 from . import gradients 195 from .gradients import * 196 from . import layers 197 from .layers import * 198 from . import linear_algebra as linalg 199 from .linear_algebra import * 200 from . import manipulation 201 from .manipulation import * 202 from . import random 203 from .random import * 204 from . import searching 205 from .searching import * 206 from . import set 207 from .set import * 208 from . import sorting 209 from .sorting import * 210 from . import statistical 211 from .statistical import * 212 from . import utility 213 from .utility import * 214 from . import experimental 215 from .experimental import * 216 from . import control_flow_ops 217 from .control_flow_ops import * 218 219 220 # sub-backends 221 222 from . import sub_backends 223 from .sub_backends import * 224 [end of ivy/functional/backends/numpy/__init__.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/ivy/functional/backends/numpy/__init__.py b/ivy/functional/backends/numpy/__init__.py --- a/ivy/functional/backends/numpy/__init__.py +++ b/ivy/functional/backends/numpy/__init__.py @@ -34,10 +34,8 @@ "bitwise_and": "bitwise_and", "matmul": "matmul", "power": "pow", - "divide": "divide", "subtract": "subtract", "add": "add", - "not_equal": "not_equal", } if ufunc.__name__ in methods.keys(): return eval("ivy." + methods[ufunc.__name__] + "(*inputs, **kwargs)")
{"golden_diff": "diff --git a/ivy/functional/backends/numpy/__init__.py b/ivy/functional/backends/numpy/__init__.py\n--- a/ivy/functional/backends/numpy/__init__.py\n+++ b/ivy/functional/backends/numpy/__init__.py\n@@ -34,10 +34,8 @@\n \"bitwise_and\": \"bitwise_and\",\n \"matmul\": \"matmul\",\n \"power\": \"pow\",\n- \"divide\": \"divide\",\n \"subtract\": \"subtract\",\n \"add\": \"add\",\n- \"not_equal\": \"not_equal\",\n }\n if ufunc.__name__ in methods.keys():\n return eval(\"ivy.\" + methods[ufunc.__name__] + \"(*inputs, **kwargs)\")\n", "issue": "Remove duplicate-keys in `ivy\\functional\\backends\\numpy\\__init__.py`\nIn the file: `ivy\\functional\\backends\\numpy\\__init__.py` duplicate keys are present in the following lines\r\nhttps://github.com/unifyai/ivy/blob/1be0f8c2c8d916b10f06ab538b43fc506a1992a6/ivy/functional/backends/numpy/__init__.py#L25\r\nhttps://github.com/unifyai/ivy/blob/1be0f8c2c8d916b10f06ab538b43fc506a1992a6/ivy/functional/backends/numpy/__init__.py#L40\r\nhttps://github.com/unifyai/ivy/blob/1be0f8c2c8d916b10f06ab538b43fc506a1992a6/ivy/functional/backends/numpy/__init__.py#L31\r\nhttps://github.com/unifyai/ivy/blob/1be0f8c2c8d916b10f06ab538b43fc506a1992a6/ivy/functional/backends/numpy/__init__.py#L37\r\nThey can be removed.\n", "before_files": [{"content": "# global\nimport sys\nimport numpy as np\n\n# local\nimport ivy\nfrom ivy.func_wrapper import _dtype_from_version\n\nbackend_version = {\"version\": np.__version__}\n\n# noinspection PyUnresolvedReferences\nif not ivy.is_local():\n _module_in_memory = sys.modules[__name__]\nelse:\n _module_in_memory = sys.modules[ivy.import_module_path].import_cache[__name__]\n\nuse = ivy.utils.backend.ContextManager(_module_in_memory)\n\n# wrap __array_ufunc__ method of ivy.Array to prioritize Ivy array methods when using numpu backend\n\n\ndef wrap__array_ufunc__(func):\n def rep_method(self, ufunc, method, *inputs, **kwargs):\n methods = {\n \"not_equal\": \"not_equal\",\n \"greater\": \"greater\",\n \"less\": \"less\",\n \"greater_equal\": \"greater_equal\",\n \"less_equal\": \"less_equal\",\n \"multiply\": \"multiply\",\n \"divide\": \"divide\",\n \"remainder\": \"remainder\",\n \"equal\": \"equal\",\n \"bitwise_and\": \"bitwise_and\",\n \"matmul\": \"matmul\",\n \"power\": \"pow\",\n \"divide\": \"divide\",\n \"subtract\": \"subtract\",\n \"add\": \"add\",\n \"not_equal\": \"not_equal\",\n }\n if ufunc.__name__ in methods.keys():\n return eval(\"ivy.\" + methods[ufunc.__name__] + \"(*inputs, **kwargs)\")\n return func(self, ufunc, method, *inputs, **kwargs)\n\n return rep_method\n\n\nivy.Array.__array_ufunc__ = wrap__array_ufunc__(ivy.Array.__array_ufunc__)\n\nNativeArray = np.ndarray\nNativeDevice = str\nNativeDtype = np.dtype\nNativeShape = tuple\n\nNativeSparseArray = None\n\n\n# devices\nvalid_devices = (\"cpu\",)\n\ninvalid_devices = (\"gpu\", \"tpu\")\n\n# native data types\nnative_int8 = np.dtype(\"int8\")\nnative_int16 = np.dtype(\"int16\")\nnative_int32 = np.dtype(\"int32\")\nnative_int64 = np.dtype(\"int64\")\nnative_uint8 = np.dtype(\"uint8\")\nnative_uint16 = np.dtype(\"uint16\")\nnative_uint32 = np.dtype(\"uint32\")\nnative_uint64 = np.dtype(\"uint64\")\nnative_float16 = np.dtype(\"float16\")\nnative_float32 = np.dtype(\"float32\")\nnative_float64 = np.dtype(\"float64\")\nnative_complex64 = np.dtype(\"complex64\")\nnative_complex128 = np.dtype(\"complex128\")\nnative_double = native_float64\nnative_bool = np.dtype(\"bool\")\n\n# valid data types\n# ToDo: Add complex dtypes to valid_dtypes and fix all resulting failures.\n\n# update these to add new dtypes\nvalid_dtypes = {\n \"1.26.0 and below\": (\n ivy.int8,\n ivy.int16,\n ivy.int32,\n ivy.int64,\n ivy.uint8,\n ivy.uint16,\n ivy.uint32,\n ivy.uint64,\n ivy.float16,\n ivy.float32,\n ivy.float64,\n ivy.complex64,\n ivy.complex128,\n ivy.bool,\n )\n}\nvalid_numeric_dtypes = {\n \"1.26.0 and below\": (\n ivy.int8,\n ivy.int16,\n ivy.int32,\n ivy.int64,\n ivy.uint8,\n ivy.uint16,\n ivy.uint32,\n ivy.uint64,\n ivy.float16,\n ivy.float32,\n ivy.float64,\n ivy.complex64,\n ivy.complex128,\n )\n}\nvalid_int_dtypes = {\n \"1.26.0 and below\": (\n ivy.int8,\n ivy.int16,\n ivy.int32,\n ivy.int64,\n ivy.uint8,\n ivy.uint16,\n ivy.uint32,\n ivy.uint64,\n )\n}\nvalid_float_dtypes = {\"1.26.0 and below\": (ivy.float16, ivy.float32, ivy.float64)}\nvalid_uint_dtypes = {\n \"1.26.0 and below\": (ivy.uint8, ivy.uint16, ivy.uint32, ivy.uint64)\n}\nvalid_complex_dtypes = {\"1.26.0 and below\": (ivy.complex64, ivy.complex128)}\n\n# leave these untouched\nvalid_dtypes = _dtype_from_version(valid_dtypes, backend_version)\nvalid_numeric_dtypes = _dtype_from_version(valid_numeric_dtypes, backend_version)\nvalid_int_dtypes = _dtype_from_version(valid_int_dtypes, backend_version)\nvalid_float_dtypes = _dtype_from_version(valid_float_dtypes, backend_version)\nvalid_uint_dtypes = _dtype_from_version(valid_uint_dtypes, backend_version)\nvalid_complex_dtypes = _dtype_from_version(valid_complex_dtypes, backend_version)\n\n# invalid data types\n# update these to add new dtypes\ninvalid_dtypes = {\"1.26.0 and below\": (ivy.bfloat16,)}\ninvalid_numeric_dtypes = {\"1.26.0 and below\": (ivy.bfloat16,)}\ninvalid_int_dtypes = {\"1.26.0 and below\": ()}\ninvalid_float_dtypes = {\"1.26.0 and below\": (ivy.bfloat16,)}\ninvalid_uint_dtypes = {\"1.26.0 and below\": ()}\ninvalid_complex_dtypes = {\"1.26.0 and below\": ()}\n\n\n# leave these untouched\ninvalid_dtypes = _dtype_from_version(invalid_dtypes, backend_version)\ninvalid_numeric_dtypes = _dtype_from_version(invalid_numeric_dtypes, backend_version)\ninvalid_int_dtypes = _dtype_from_version(invalid_int_dtypes, backend_version)\ninvalid_float_dtypes = _dtype_from_version(invalid_float_dtypes, backend_version)\ninvalid_uint_dtypes = _dtype_from_version(invalid_uint_dtypes, backend_version)\ninvalid_complex_dtypes = _dtype_from_version(invalid_complex_dtypes, backend_version)\n\n\nnative_inplace_support = True\n\nsupports_gradients = False\n\n\ndef closest_valid_dtype(type=None, /, as_native=False):\n if type is None:\n type = ivy.default_dtype()\n elif isinstance(type, str) and type in invalid_dtypes:\n type = {\"bfloat16\": ivy.float16}[type]\n return ivy.as_ivy_dtype(type) if not as_native else ivy.as_native_dtype(type)\n\n\nbackend = \"numpy\"\n\n\n# local sub-modules\nfrom . import activations\nfrom .activations import *\nfrom . import creation\nfrom .creation import *\nfrom . import data_type\nfrom .data_type import *\nfrom . import device\nfrom .device import *\nfrom . import elementwise\nfrom .elementwise import *\nfrom . import general\nfrom .general import *\nfrom . import gradients\nfrom .gradients import *\nfrom . import layers\nfrom .layers import *\nfrom . import linear_algebra as linalg\nfrom .linear_algebra import *\nfrom . import manipulation\nfrom .manipulation import *\nfrom . import random\nfrom .random import *\nfrom . import searching\nfrom .searching import *\nfrom . import set\nfrom .set import *\nfrom . import sorting\nfrom .sorting import *\nfrom . import statistical\nfrom .statistical import *\nfrom . import utility\nfrom .utility import *\nfrom . import experimental\nfrom .experimental import *\nfrom . import control_flow_ops\nfrom .control_flow_ops import *\n\n\n# sub-backends\n\nfrom . import sub_backends\nfrom .sub_backends import *\n", "path": "ivy/functional/backends/numpy/__init__.py"}]}
3,160
168
gh_patches_debug_32253
rasdani/github-patches
git_diff
aws__aws-cli-2044
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> [aws deploy push] Destination bundle type does not match extension When specifying a destination with a suffix of ".zip" or ".tar" or ".tar.gz" for use with the `aws deploy push` command the resulting bundle is always only zipped. ``` $ aws deploy push --application-name test --s3-location s3://$AWS_DEPLOY_BUCKET/test/latest.tar --no-ignore-hidden-files --source . To deploy with this revision, run: aws deploy create-deployment --application-name test --s3-location bucket=BUCKET_NAME,key=test/latest.tar,bundleType=zip,eTag=... --deployment-group-name <deployment-group-name> --deployment-config-name <deployment-config-name> --description <description> ``` I downloaded and check the file it come back as a zip file. ``` $ file latest.tar latest.tar: Zip archive data, at least v2.0 to extract ``` </issue> <code> [start of awscli/customizations/codedeploy/push.py] 1 # Copyright 2015 Amazon.com, Inc. or its affiliates. All Rights Reserved. 2 # 3 # Licensed under the Apache License, Version 2.0 (the "License"). You 4 # may not use this file except in compliance with the License. A copy of 5 # the License is located at 6 # 7 # http://aws.amazon.com/apache2.0/ 8 # 9 # or in the "license" file accompanying this file. This file is 10 # distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF 11 # ANY KIND, either express or implied. See the License for the specific 12 # language governing permissions and limitations under the License. 13 14 import os 15 import sys 16 import zipfile 17 import tempfile 18 import contextlib 19 from datetime import datetime 20 21 from botocore.exceptions import ClientError 22 23 from awscli.compat import six 24 from awscli.customizations.codedeploy.utils import validate_s3_location 25 from awscli.customizations.commands import BasicCommand 26 from awscli.compat import ZIP_COMPRESSION_MODE 27 28 29 ONE_MB = 1 << 20 30 MULTIPART_LIMIT = 6 * ONE_MB 31 32 33 class Push(BasicCommand): 34 NAME = 'push' 35 36 DESCRIPTION = ( 37 'Bundles and uploads to Amazon Simple Storage Service (Amazon S3) an ' 38 'application revision, which is an archive file that contains ' 39 'deployable content and an accompanying Application Specification ' 40 'file (AppSpec file). If the upload is successful, a message is ' 41 'returned that describes how to call the create-deployment command to ' 42 'deploy the application revision from Amazon S3 to target Amazon ' 43 'Elastic Compute Cloud (Amazon EC2) instances.' 44 ) 45 46 ARG_TABLE = [ 47 { 48 'name': 'application-name', 49 'synopsis': '--application-name <app-name>', 50 'required': True, 51 'help_text': ( 52 'Required. The name of the AWS CodeDeploy application to be ' 53 'associated with the application revision.' 54 ) 55 }, 56 { 57 'name': 's3-location', 58 'synopsis': '--s3-location s3://<bucket>/<key>', 59 'required': True, 60 'help_text': ( 61 'Required. Information about the location of the application ' 62 'revision to be uploaded to Amazon S3. You must specify both ' 63 'a bucket and a key that represent the Amazon S3 bucket name ' 64 'and the object key name. Use the format ' 65 's3://\<bucket\>/\<key\>' 66 ) 67 }, 68 { 69 'name': 'ignore-hidden-files', 70 'action': 'store_true', 71 'default': False, 72 'group_name': 'ignore-hidden-files', 73 'help_text': ( 74 'Optional. Set the --ignore-hidden-files flag to not bundle ' 75 'and upload hidden files to Amazon S3; otherwise, set the ' 76 '--no-ignore-hidden-files flag (the default) to bundle and ' 77 'upload hidden files to Amazon S3.' 78 ) 79 }, 80 { 81 'name': 'no-ignore-hidden-files', 82 'action': 'store_true', 83 'default': False, 84 'group_name': 'ignore-hidden-files' 85 }, 86 { 87 'name': 'source', 88 'synopsis': '--source <path>', 89 'default': '.', 90 'help_text': ( 91 'Optional. The location of the deployable content and the ' 92 'accompanying AppSpec file on the development machine to be ' 93 'bundled and uploaded to Amazon S3. If not specified, the ' 94 'current directory is used.' 95 ) 96 }, 97 { 98 'name': 'description', 99 'synopsis': '--description <description>', 100 'help_text': ( 101 'Optional. A comment that summarizes the application ' 102 'revision. If not specified, the default string "Uploaded by ' 103 'AWS CLI \'time\' UTC" is used, where \'time\' is the current ' 104 'system time in Coordinated Universal Time (UTC).' 105 ) 106 } 107 ] 108 109 def _run_main(self, parsed_args, parsed_globals): 110 self._validate_args(parsed_args) 111 self.codedeploy = self._session.create_client( 112 'codedeploy', 113 region_name=parsed_globals.region, 114 endpoint_url=parsed_globals.endpoint_url, 115 verify=parsed_globals.verify_ssl 116 ) 117 self.s3 = self._session.create_client( 118 's3', 119 region_name=parsed_globals.region 120 ) 121 self._push(parsed_args) 122 123 def _validate_args(self, parsed_args): 124 validate_s3_location(parsed_args, 's3_location') 125 if parsed_args.ignore_hidden_files \ 126 and parsed_args.no_ignore_hidden_files: 127 raise RuntimeError( 128 'You cannot specify both --ignore-hidden-files and ' 129 '--no-ignore-hidden-files.' 130 ) 131 if not parsed_args.description: 132 parsed_args.description = ( 133 'Uploaded by AWS CLI {0} UTC'.format( 134 datetime.utcnow().isoformat() 135 ) 136 ) 137 138 def _push(self, params): 139 with self._compress( 140 params.source, 141 params.ignore_hidden_files 142 ) as bundle: 143 try: 144 upload_response = self._upload_to_s3(params, bundle) 145 params.eTag = upload_response['ETag'].replace('"', "") 146 if 'VersionId' in upload_response: 147 params.version = upload_response['VersionId'] 148 except Exception as e: 149 raise RuntimeError( 150 'Failed to upload \'%s\' to \'%s\': %s' % 151 (params.source, 152 params.s3_location, 153 str(e)) 154 ) 155 self._register_revision(params) 156 157 if 'version' in params: 158 version_string = ',version={0}'.format(params.version) 159 else: 160 version_string = '' 161 s3location_string = ( 162 '--s3-location bucket={0},key={1},' 163 'bundleType=zip,eTag={2}{3}'.format( 164 params.bucket, 165 params.key, 166 params.eTag, 167 version_string 168 ) 169 ) 170 sys.stdout.write( 171 'To deploy with this revision, run:\n' 172 'aws deploy create-deployment ' 173 '--application-name {0} {1} ' 174 '--deployment-group-name <deployment-group-name> ' 175 '--deployment-config-name <deployment-config-name> ' 176 '--description <description>\n'.format( 177 params.application_name, 178 s3location_string 179 ) 180 ) 181 182 @contextlib.contextmanager 183 def _compress(self, source, ignore_hidden_files=False): 184 source_path = os.path.abspath(source) 185 appspec_path = os.path.sep.join([source_path, 'appspec.yml']) 186 with tempfile.TemporaryFile('w+b') as tf: 187 zf = zipfile.ZipFile(tf, 'w', allowZip64=True) 188 # Using 'try'/'finally' instead of 'with' statement since ZipFile 189 # does not have support context manager in Python 2.6. 190 try: 191 contains_appspec = False 192 for root, dirs, files in os.walk(source, topdown=True): 193 if ignore_hidden_files: 194 files = [fn for fn in files if not fn.startswith('.')] 195 dirs[:] = [dn for dn in dirs if not dn.startswith('.')] 196 for fn in files: 197 filename = os.path.join(root, fn) 198 filename = os.path.abspath(filename) 199 arcname = filename[len(source_path) + 1:] 200 if filename == appspec_path: 201 contains_appspec = True 202 zf.write(filename, arcname, ZIP_COMPRESSION_MODE) 203 if not contains_appspec: 204 raise RuntimeError( 205 '{0} was not found'.format(appspec_path) 206 ) 207 finally: 208 zf.close() 209 yield tf 210 211 def _upload_to_s3(self, params, bundle): 212 size_remaining = self._bundle_size(bundle) 213 if size_remaining < MULTIPART_LIMIT: 214 return self.s3.put_object( 215 Bucket=params.bucket, 216 Key=params.key, 217 Body=bundle 218 ) 219 else: 220 return self._multipart_upload_to_s3( 221 params, 222 bundle, 223 size_remaining 224 ) 225 226 def _bundle_size(self, bundle): 227 bundle.seek(0, 2) 228 size = bundle.tell() 229 bundle.seek(0) 230 return size 231 232 def _multipart_upload_to_s3(self, params, bundle, size_remaining): 233 create_response = self.s3.create_multipart_upload( 234 Bucket=params.bucket, 235 Key=params.key 236 ) 237 upload_id = create_response['UploadId'] 238 try: 239 part_num = 1 240 multipart_list = [] 241 bundle.seek(0) 242 while size_remaining > 0: 243 data = bundle.read(MULTIPART_LIMIT) 244 upload_response = self.s3.upload_part( 245 Bucket=params.bucket, 246 Key=params.key, 247 UploadId=upload_id, 248 PartNumber=part_num, 249 Body=six.BytesIO(data) 250 ) 251 multipart_list.append({ 252 'PartNumber': part_num, 253 'ETag': upload_response['ETag'] 254 }) 255 part_num += 1 256 size_remaining -= len(data) 257 return self.s3.complete_multipart_upload( 258 Bucket=params.bucket, 259 Key=params.key, 260 UploadId=upload_id, 261 MultipartUpload={'Parts': multipart_list} 262 ) 263 except ClientError as e: 264 self.s3.abort_multipart_upload( 265 Bucket=params.bucket, 266 Key=params.key, 267 UploadId=upload_id 268 ) 269 raise e 270 271 def _register_revision(self, params): 272 revision = { 273 'revisionType': 'S3', 274 's3Location': { 275 'bucket': params.bucket, 276 'key': params.key, 277 'bundleType': 'zip', 278 'eTag': params.eTag 279 } 280 } 281 if 'version' in params: 282 revision['s3Location']['version'] = params.version 283 self.codedeploy.register_application_revision( 284 applicationName=params.application_name, 285 revision=revision, 286 description=params.description 287 ) 288 [end of awscli/customizations/codedeploy/push.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/awscli/customizations/codedeploy/push.py b/awscli/customizations/codedeploy/push.py --- a/awscli/customizations/codedeploy/push.py +++ b/awscli/customizations/codedeploy/push.py @@ -35,7 +35,7 @@ DESCRIPTION = ( 'Bundles and uploads to Amazon Simple Storage Service (Amazon S3) an ' - 'application revision, which is an archive file that contains ' + 'application revision, which is a zip archive file that contains ' 'deployable content and an accompanying Application Specification ' 'file (AppSpec file). If the upload is successful, a message is ' 'returned that describes how to call the create-deployment command to ' @@ -61,8 +61,8 @@ 'Required. Information about the location of the application ' 'revision to be uploaded to Amazon S3. You must specify both ' 'a bucket and a key that represent the Amazon S3 bucket name ' - 'and the object key name. Use the format ' - 's3://\<bucket\>/\<key\>' + 'and the object key name. Content will be zipped before ' + 'uploading. Use the format s3://\<bucket\>/\<key\>' ) }, { @@ -90,7 +90,7 @@ 'help_text': ( 'Optional. The location of the deployable content and the ' 'accompanying AppSpec file on the development machine to be ' - 'bundled and uploaded to Amazon S3. If not specified, the ' + 'zipped and uploaded to Amazon S3. If not specified, the ' 'current directory is used.' ) },
{"golden_diff": "diff --git a/awscli/customizations/codedeploy/push.py b/awscli/customizations/codedeploy/push.py\n--- a/awscli/customizations/codedeploy/push.py\n+++ b/awscli/customizations/codedeploy/push.py\n@@ -35,7 +35,7 @@\n \n DESCRIPTION = (\n 'Bundles and uploads to Amazon Simple Storage Service (Amazon S3) an '\n- 'application revision, which is an archive file that contains '\n+ 'application revision, which is a zip archive file that contains '\n 'deployable content and an accompanying Application Specification '\n 'file (AppSpec file). If the upload is successful, a message is '\n 'returned that describes how to call the create-deployment command to '\n@@ -61,8 +61,8 @@\n 'Required. Information about the location of the application '\n 'revision to be uploaded to Amazon S3. You must specify both '\n 'a bucket and a key that represent the Amazon S3 bucket name '\n- 'and the object key name. Use the format '\n- 's3://\\<bucket\\>/\\<key\\>'\n+ 'and the object key name. Content will be zipped before '\n+ 'uploading. Use the format s3://\\<bucket\\>/\\<key\\>'\n )\n },\n {\n@@ -90,7 +90,7 @@\n 'help_text': (\n 'Optional. The location of the deployable content and the '\n 'accompanying AppSpec file on the development machine to be '\n- 'bundled and uploaded to Amazon S3. If not specified, the '\n+ 'zipped and uploaded to Amazon S3. If not specified, the '\n 'current directory is used.'\n )\n },\n", "issue": "[aws deploy push] Destination bundle type does not match extension\nWhen specifying a destination with a suffix of \".zip\" or \".tar\" or \".tar.gz\" for use with the `aws deploy push` command the resulting bundle is always only zipped.\n\n```\n$ aws deploy push --application-name test --s3-location s3://$AWS_DEPLOY_BUCKET/test/latest.tar --no-ignore-hidden-files --source .\nTo deploy with this revision, run:\naws deploy create-deployment --application-name test --s3-location bucket=BUCKET_NAME,key=test/latest.tar,bundleType=zip,eTag=... --deployment-group-name <deployment-group-name> --deployment-config-name <deployment-config-name> --description <description>\n```\n\nI downloaded and check the file it come back as a zip file.\n\n```\n$ file latest.tar \nlatest.tar: Zip archive data, at least v2.0 to extract\n```\n\n", "before_files": [{"content": "# Copyright 2015 Amazon.com, Inc. or its affiliates. All Rights Reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\"). You\n# may not use this file except in compliance with the License. A copy of\n# the License is located at\n#\n# http://aws.amazon.com/apache2.0/\n#\n# or in the \"license\" file accompanying this file. This file is\n# distributed on an \"AS IS\" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF\n# ANY KIND, either express or implied. See the License for the specific\n# language governing permissions and limitations under the License.\n\nimport os\nimport sys\nimport zipfile\nimport tempfile\nimport contextlib\nfrom datetime import datetime\n\nfrom botocore.exceptions import ClientError\n\nfrom awscli.compat import six\nfrom awscli.customizations.codedeploy.utils import validate_s3_location\nfrom awscli.customizations.commands import BasicCommand\nfrom awscli.compat import ZIP_COMPRESSION_MODE\n\n\nONE_MB = 1 << 20\nMULTIPART_LIMIT = 6 * ONE_MB\n\n\nclass Push(BasicCommand):\n NAME = 'push'\n\n DESCRIPTION = (\n 'Bundles and uploads to Amazon Simple Storage Service (Amazon S3) an '\n 'application revision, which is an archive file that contains '\n 'deployable content and an accompanying Application Specification '\n 'file (AppSpec file). If the upload is successful, a message is '\n 'returned that describes how to call the create-deployment command to '\n 'deploy the application revision from Amazon S3 to target Amazon '\n 'Elastic Compute Cloud (Amazon EC2) instances.'\n )\n\n ARG_TABLE = [\n {\n 'name': 'application-name',\n 'synopsis': '--application-name <app-name>',\n 'required': True,\n 'help_text': (\n 'Required. The name of the AWS CodeDeploy application to be '\n 'associated with the application revision.'\n )\n },\n {\n 'name': 's3-location',\n 'synopsis': '--s3-location s3://<bucket>/<key>',\n 'required': True,\n 'help_text': (\n 'Required. Information about the location of the application '\n 'revision to be uploaded to Amazon S3. You must specify both '\n 'a bucket and a key that represent the Amazon S3 bucket name '\n 'and the object key name. Use the format '\n 's3://\\<bucket\\>/\\<key\\>'\n )\n },\n {\n 'name': 'ignore-hidden-files',\n 'action': 'store_true',\n 'default': False,\n 'group_name': 'ignore-hidden-files',\n 'help_text': (\n 'Optional. Set the --ignore-hidden-files flag to not bundle '\n 'and upload hidden files to Amazon S3; otherwise, set the '\n '--no-ignore-hidden-files flag (the default) to bundle and '\n 'upload hidden files to Amazon S3.'\n )\n },\n {\n 'name': 'no-ignore-hidden-files',\n 'action': 'store_true',\n 'default': False,\n 'group_name': 'ignore-hidden-files'\n },\n {\n 'name': 'source',\n 'synopsis': '--source <path>',\n 'default': '.',\n 'help_text': (\n 'Optional. The location of the deployable content and the '\n 'accompanying AppSpec file on the development machine to be '\n 'bundled and uploaded to Amazon S3. If not specified, the '\n 'current directory is used.'\n )\n },\n {\n 'name': 'description',\n 'synopsis': '--description <description>',\n 'help_text': (\n 'Optional. A comment that summarizes the application '\n 'revision. If not specified, the default string \"Uploaded by '\n 'AWS CLI \\'time\\' UTC\" is used, where \\'time\\' is the current '\n 'system time in Coordinated Universal Time (UTC).'\n )\n }\n ]\n\n def _run_main(self, parsed_args, parsed_globals):\n self._validate_args(parsed_args)\n self.codedeploy = self._session.create_client(\n 'codedeploy',\n region_name=parsed_globals.region,\n endpoint_url=parsed_globals.endpoint_url,\n verify=parsed_globals.verify_ssl\n )\n self.s3 = self._session.create_client(\n 's3',\n region_name=parsed_globals.region\n )\n self._push(parsed_args)\n\n def _validate_args(self, parsed_args):\n validate_s3_location(parsed_args, 's3_location')\n if parsed_args.ignore_hidden_files \\\n and parsed_args.no_ignore_hidden_files:\n raise RuntimeError(\n 'You cannot specify both --ignore-hidden-files and '\n '--no-ignore-hidden-files.'\n )\n if not parsed_args.description:\n parsed_args.description = (\n 'Uploaded by AWS CLI {0} UTC'.format(\n datetime.utcnow().isoformat()\n )\n )\n\n def _push(self, params):\n with self._compress(\n params.source,\n params.ignore_hidden_files\n ) as bundle:\n try:\n upload_response = self._upload_to_s3(params, bundle)\n params.eTag = upload_response['ETag'].replace('\"', \"\")\n if 'VersionId' in upload_response:\n params.version = upload_response['VersionId']\n except Exception as e:\n raise RuntimeError(\n 'Failed to upload \\'%s\\' to \\'%s\\': %s' %\n (params.source,\n params.s3_location,\n str(e))\n )\n self._register_revision(params)\n\n if 'version' in params:\n version_string = ',version={0}'.format(params.version)\n else:\n version_string = ''\n s3location_string = (\n '--s3-location bucket={0},key={1},'\n 'bundleType=zip,eTag={2}{3}'.format(\n params.bucket,\n params.key,\n params.eTag,\n version_string\n )\n )\n sys.stdout.write(\n 'To deploy with this revision, run:\\n'\n 'aws deploy create-deployment '\n '--application-name {0} {1} '\n '--deployment-group-name <deployment-group-name> '\n '--deployment-config-name <deployment-config-name> '\n '--description <description>\\n'.format(\n params.application_name,\n s3location_string\n )\n )\n\n @contextlib.contextmanager\n def _compress(self, source, ignore_hidden_files=False):\n source_path = os.path.abspath(source)\n appspec_path = os.path.sep.join([source_path, 'appspec.yml'])\n with tempfile.TemporaryFile('w+b') as tf:\n zf = zipfile.ZipFile(tf, 'w', allowZip64=True)\n # Using 'try'/'finally' instead of 'with' statement since ZipFile\n # does not have support context manager in Python 2.6.\n try:\n contains_appspec = False\n for root, dirs, files in os.walk(source, topdown=True):\n if ignore_hidden_files:\n files = [fn for fn in files if not fn.startswith('.')]\n dirs[:] = [dn for dn in dirs if not dn.startswith('.')]\n for fn in files:\n filename = os.path.join(root, fn)\n filename = os.path.abspath(filename)\n arcname = filename[len(source_path) + 1:]\n if filename == appspec_path:\n contains_appspec = True\n zf.write(filename, arcname, ZIP_COMPRESSION_MODE)\n if not contains_appspec:\n raise RuntimeError(\n '{0} was not found'.format(appspec_path)\n )\n finally:\n zf.close()\n yield tf\n\n def _upload_to_s3(self, params, bundle):\n size_remaining = self._bundle_size(bundle)\n if size_remaining < MULTIPART_LIMIT:\n return self.s3.put_object(\n Bucket=params.bucket,\n Key=params.key,\n Body=bundle\n )\n else:\n return self._multipart_upload_to_s3(\n params,\n bundle,\n size_remaining\n )\n\n def _bundle_size(self, bundle):\n bundle.seek(0, 2)\n size = bundle.tell()\n bundle.seek(0)\n return size\n\n def _multipart_upload_to_s3(self, params, bundle, size_remaining):\n create_response = self.s3.create_multipart_upload(\n Bucket=params.bucket,\n Key=params.key\n )\n upload_id = create_response['UploadId']\n try:\n part_num = 1\n multipart_list = []\n bundle.seek(0)\n while size_remaining > 0:\n data = bundle.read(MULTIPART_LIMIT)\n upload_response = self.s3.upload_part(\n Bucket=params.bucket,\n Key=params.key,\n UploadId=upload_id,\n PartNumber=part_num,\n Body=six.BytesIO(data)\n )\n multipart_list.append({\n 'PartNumber': part_num,\n 'ETag': upload_response['ETag']\n })\n part_num += 1\n size_remaining -= len(data)\n return self.s3.complete_multipart_upload(\n Bucket=params.bucket,\n Key=params.key,\n UploadId=upload_id,\n MultipartUpload={'Parts': multipart_list}\n )\n except ClientError as e:\n self.s3.abort_multipart_upload(\n Bucket=params.bucket,\n Key=params.key,\n UploadId=upload_id\n )\n raise e\n\n def _register_revision(self, params):\n revision = {\n 'revisionType': 'S3',\n 's3Location': {\n 'bucket': params.bucket,\n 'key': params.key,\n 'bundleType': 'zip',\n 'eTag': params.eTag\n }\n }\n if 'version' in params:\n revision['s3Location']['version'] = params.version\n self.codedeploy.register_application_revision(\n applicationName=params.application_name,\n revision=revision,\n description=params.description\n )\n", "path": "awscli/customizations/codedeploy/push.py"}]}
3,620
384
gh_patches_debug_35132
rasdani/github-patches
git_diff
CTFd__CTFd-1352
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> Submission search Search submissions akin to how users are searched </issue> <code> [start of CTFd/admin/submissions.py] 1 from flask import render_template, request 2 3 from CTFd.admin import admin 4 from CTFd.models import Challenges, Submissions 5 from CTFd.utils.decorators import admins_only 6 from CTFd.utils.modes import get_model 7 8 9 @admin.route("/admin/submissions", defaults={"submission_type": None}) 10 @admin.route("/admin/submissions/<submission_type>") 11 @admins_only 12 def submissions_listing(submission_type): 13 filters = {} 14 if submission_type: 15 filters["type"] = submission_type 16 17 curr_page = abs(int(request.args.get("page", 1, type=int))) 18 results_per_page = 50 19 page_start = results_per_page * (curr_page - 1) 20 page_end = results_per_page * (curr_page - 1) + results_per_page 21 sub_count = Submissions.query.filter_by(**filters).count() 22 page_count = int(sub_count / results_per_page) + (sub_count % results_per_page > 0) 23 24 Model = get_model() 25 26 submissions = ( 27 Submissions.query.add_columns( 28 Submissions.id, 29 Submissions.type, 30 Submissions.challenge_id, 31 Submissions.provided, 32 Submissions.account_id, 33 Submissions.date, 34 Challenges.name.label("challenge_name"), 35 Model.name.label("team_name"), 36 ) 37 .filter_by(**filters) 38 .join(Challenges) 39 .join(Model) 40 .order_by(Submissions.date.desc()) 41 .slice(page_start, page_end) 42 .all() 43 ) 44 45 return render_template( 46 "admin/submissions.html", 47 submissions=submissions, 48 page_count=page_count, 49 curr_page=curr_page, 50 type=submission_type, 51 ) 52 [end of CTFd/admin/submissions.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/CTFd/admin/submissions.py b/CTFd/admin/submissions.py --- a/CTFd/admin/submissions.py +++ b/CTFd/admin/submissions.py @@ -1,4 +1,4 @@ -from flask import render_template, request +from flask import render_template, request, url_for from CTFd.admin import admin from CTFd.models import Challenges, Submissions @@ -10,16 +10,21 @@ @admin.route("/admin/submissions/<submission_type>") @admins_only def submissions_listing(submission_type): - filters = {} + filters_by = {} if submission_type: - filters["type"] = submission_type + filters_by["type"] = submission_type + filters = [] - curr_page = abs(int(request.args.get("page", 1, type=int))) - results_per_page = 50 - page_start = results_per_page * (curr_page - 1) - page_end = results_per_page * (curr_page - 1) + results_per_page - sub_count = Submissions.query.filter_by(**filters).count() - page_count = int(sub_count / results_per_page) + (sub_count % results_per_page > 0) + q = request.args.get("q") + field = request.args.get("field") + page = abs(request.args.get("page", 1, type=int)) + + if q: + submissions = [] + if Submissions.__mapper__.has_property( + field + ): # The field exists as an exposed column + filters.append(getattr(Submissions, field).like("%{}%".format(q))) Model = get_model() @@ -34,18 +39,27 @@ Challenges.name.label("challenge_name"), Model.name.label("team_name"), ) - .filter_by(**filters) + .filter_by(**filters_by) + .filter(*filters) .join(Challenges) .join(Model) .order_by(Submissions.date.desc()) - .slice(page_start, page_end) - .all() + .paginate(page=page, per_page=50) ) + args = dict(request.args) + args.pop("page", 1) + return render_template( "admin/submissions.html", submissions=submissions, - page_count=page_count, - curr_page=curr_page, + prev_page=url_for( + request.endpoint, type=submission_type, page=submissions.prev_num, **args + ), + next_page=url_for( + request.endpoint, type=submission_type, page=submissions.next_num, **args + ), type=submission_type, + q=q, + field=field, )
{"golden_diff": "diff --git a/CTFd/admin/submissions.py b/CTFd/admin/submissions.py\n--- a/CTFd/admin/submissions.py\n+++ b/CTFd/admin/submissions.py\n@@ -1,4 +1,4 @@\n-from flask import render_template, request\n+from flask import render_template, request, url_for\n \n from CTFd.admin import admin\n from CTFd.models import Challenges, Submissions\n@@ -10,16 +10,21 @@\n @admin.route(\"/admin/submissions/<submission_type>\")\n @admins_only\n def submissions_listing(submission_type):\n- filters = {}\n+ filters_by = {}\n if submission_type:\n- filters[\"type\"] = submission_type\n+ filters_by[\"type\"] = submission_type\n+ filters = []\n \n- curr_page = abs(int(request.args.get(\"page\", 1, type=int)))\n- results_per_page = 50\n- page_start = results_per_page * (curr_page - 1)\n- page_end = results_per_page * (curr_page - 1) + results_per_page\n- sub_count = Submissions.query.filter_by(**filters).count()\n- page_count = int(sub_count / results_per_page) + (sub_count % results_per_page > 0)\n+ q = request.args.get(\"q\")\n+ field = request.args.get(\"field\")\n+ page = abs(request.args.get(\"page\", 1, type=int))\n+\n+ if q:\n+ submissions = []\n+ if Submissions.__mapper__.has_property(\n+ field\n+ ): # The field exists as an exposed column\n+ filters.append(getattr(Submissions, field).like(\"%{}%\".format(q)))\n \n Model = get_model()\n \n@@ -34,18 +39,27 @@\n Challenges.name.label(\"challenge_name\"),\n Model.name.label(\"team_name\"),\n )\n- .filter_by(**filters)\n+ .filter_by(**filters_by)\n+ .filter(*filters)\n .join(Challenges)\n .join(Model)\n .order_by(Submissions.date.desc())\n- .slice(page_start, page_end)\n- .all()\n+ .paginate(page=page, per_page=50)\n )\n \n+ args = dict(request.args)\n+ args.pop(\"page\", 1)\n+\n return render_template(\n \"admin/submissions.html\",\n submissions=submissions,\n- page_count=page_count,\n- curr_page=curr_page,\n+ prev_page=url_for(\n+ request.endpoint, type=submission_type, page=submissions.prev_num, **args\n+ ),\n+ next_page=url_for(\n+ request.endpoint, type=submission_type, page=submissions.next_num, **args\n+ ),\n type=submission_type,\n+ q=q,\n+ field=field,\n )\n", "issue": "Submission search\nSearch submissions akin to how users are searched\n", "before_files": [{"content": "from flask import render_template, request\n\nfrom CTFd.admin import admin\nfrom CTFd.models import Challenges, Submissions\nfrom CTFd.utils.decorators import admins_only\nfrom CTFd.utils.modes import get_model\n\n\[email protected](\"/admin/submissions\", defaults={\"submission_type\": None})\[email protected](\"/admin/submissions/<submission_type>\")\n@admins_only\ndef submissions_listing(submission_type):\n filters = {}\n if submission_type:\n filters[\"type\"] = submission_type\n\n curr_page = abs(int(request.args.get(\"page\", 1, type=int)))\n results_per_page = 50\n page_start = results_per_page * (curr_page - 1)\n page_end = results_per_page * (curr_page - 1) + results_per_page\n sub_count = Submissions.query.filter_by(**filters).count()\n page_count = int(sub_count / results_per_page) + (sub_count % results_per_page > 0)\n\n Model = get_model()\n\n submissions = (\n Submissions.query.add_columns(\n Submissions.id,\n Submissions.type,\n Submissions.challenge_id,\n Submissions.provided,\n Submissions.account_id,\n Submissions.date,\n Challenges.name.label(\"challenge_name\"),\n Model.name.label(\"team_name\"),\n )\n .filter_by(**filters)\n .join(Challenges)\n .join(Model)\n .order_by(Submissions.date.desc())\n .slice(page_start, page_end)\n .all()\n )\n\n return render_template(\n \"admin/submissions.html\",\n submissions=submissions,\n page_count=page_count,\n curr_page=curr_page,\n type=submission_type,\n )\n", "path": "CTFd/admin/submissions.py"}]}
1,007
615
gh_patches_debug_11056
rasdani/github-patches
git_diff
pantsbuild__pants-14584
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> coursier post processing script choking on null dependency file in report json **Describe the bug** One of our 3rdparty jars (`org.geotools:gt-main:9.2` from the `https://repo.osgeo.org/repository/release` repositiory) transitively depends on a `javax.media:jai_core:1.1.3` jar. For whatever reason, Coursier's resolve report json lists `"file": null` for that jar, which the post processing script doesn't currently handle. **Pants version** 2.9.0 **OS** Encountered on MacOS, though expect we would see this on Linux as well **Additional info** Coursier config: ``` [coursier] repos = [ "https://maven-central.storage-download.googleapis.com/maven2", "https://repo1.maven.org/maven2", "https://repo.osgeo.org/repository/release", ] ``` affected jvm_artifact: ``` jvm_artifact( group="org.geotools", artifact="gt-main", version="9.2", ) ``` [slack discussion](https://pantsbuild.slack.com/archives/C046T6T9U/p1645598534013879) with links to full resolve output and json </issue> <code> [start of src/python/pants/jvm/resolve/coursier_setup.py] 1 # Copyright 2021 Pants project contributors (see CONTRIBUTORS.md). 2 # Licensed under the Apache License, Version 2.0 (see LICENSE). 3 4 from __future__ import annotations 5 6 import os 7 import shlex 8 import textwrap 9 from dataclasses import dataclass 10 from typing import ClassVar, Iterable, Tuple 11 12 from pants.core.util_rules import external_tool 13 from pants.core.util_rules.external_tool import ( 14 DownloadedExternalTool, 15 ExternalToolRequest, 16 TemplatedExternalTool, 17 ) 18 from pants.engine.fs import CreateDigest, Digest, FileContent, MergeDigests 19 from pants.engine.platform import Platform 20 from pants.engine.process import BashBinary, Process 21 from pants.engine.rules import Get, MultiGet, collect_rules, rule 22 from pants.python.binaries import PythonBinary 23 from pants.util.logging import LogLevel 24 25 COURSIER_POST_PROCESSING_SCRIPT = textwrap.dedent( 26 """\ 27 import json 28 import sys 29 import os 30 from pathlib import PurePath 31 from shutil import copyfile 32 33 report = json.load(open(sys.argv[1])) 34 35 # Mapping from dest path to source path. It is ok to capture the same output filename multiple 36 # times if the source is the same as well. 37 classpath = dict() 38 for dep in report['dependencies']: 39 source = PurePath(dep['file']) 40 dest_name = dep['coord'].replace(":", "_") 41 _, ext = os.path.splitext(source) 42 classpath_dest = f"classpath/{dest_name}{ext}" 43 44 existing_source = classpath.get(classpath_dest) 45 if existing_source: 46 if existing_source == source: 47 # We've already captured this file. 48 continue 49 raise Exception( 50 f"Duplicate jar name {classpath_dest} with incompatible source:\\n" 51 f" {source}\\n" 52 f" {existing_source}\\n" 53 ) 54 classpath[classpath_dest] = source 55 copyfile(source, classpath_dest) 56 """ 57 ) 58 59 COURSIER_FETCH_WRAPPER_SCRIPT = textwrap.dedent( 60 """\ 61 set -eux 62 63 coursier_exe="$1" 64 shift 65 json_output_file="$1" 66 shift 67 68 working_dir="$(pwd)" 69 "$coursier_exe" fetch {repos_args} \ 70 --json-output-file="$json_output_file" \ 71 "${{@//{coursier_working_directory}/$working_dir}}" 72 /bin/mkdir -p classpath 73 {python_path} {coursier_bin_dir}/coursier_post_processing_script.py "$json_output_file" 74 """ 75 ) 76 77 78 # TODO: Coursier renders setrlimit error line on macOS. 79 # see https://github.com/pantsbuild/pants/issues/13942. 80 POST_PROCESS_COURSIER_STDERR_SCRIPT = textwrap.dedent( 81 """\ 82 #!{python_path} 83 import sys 84 from subprocess import run, PIPE 85 86 proc = run(sys.argv[1:], stdout=PIPE, stderr=PIPE) 87 88 sys.stdout.buffer.write(proc.stdout) 89 sys.stderr.buffer.write(proc.stderr.replace(b"setrlimit to increase file descriptor limit failed, errno 22\\n", b"")) 90 sys.exit(proc.returncode) 91 """ 92 ) 93 94 95 class CoursierSubsystem(TemplatedExternalTool): 96 options_scope = "coursier" 97 name = "coursier" 98 help = "A dependency resolver for the Maven ecosystem." 99 100 default_version = "v2.0.16-169-g194ebc55c" 101 default_known_versions = [ 102 "v2.0.16-169-g194ebc55c|linux_arm64 |da38c97d55967505b8454c20a90370c518044829398b9bce8b637d194d79abb3|18114472", 103 "v2.0.16-169-g194ebc55c|linux_x86_64|4c61a634c4bd2773b4543fe0fc32210afd343692891121cddb447204b48672e8|18486946", 104 "v2.0.16-169-g194ebc55c|macos_arm64 |15bce235d223ef1d022da30b67b4c64e9228d236b876c834b64e029bbe824c6f|17957182", 105 "v2.0.16-169-g194ebc55c|macos_x86_64|15bce235d223ef1d022da30b67b4c64e9228d236b876c834b64e029bbe824c6f|17957182", 106 ] 107 default_url_template = ( 108 "https://github.com/coursier/coursier/releases/download/{version}/cs-{platform}.gz" 109 ) 110 default_url_platform_mapping = { 111 "macos_arm64": "x86_64-apple-darwin", 112 "macos_x86_64": "x86_64-apple-darwin", 113 "linux_arm64": "aarch64-pc-linux", 114 "linux_x86_64": "x86_64-pc-linux", 115 } 116 117 @classmethod 118 def register_options(cls, register) -> None: 119 super().register_options(register) 120 register( 121 "--repos", 122 type=list, 123 member_type=str, 124 default=[ 125 "https://maven-central.storage-download.googleapis.com/maven2", 126 "https://repo1.maven.org/maven2", 127 ], 128 help=("Maven style repositories to resolve artifacts from."), 129 ) 130 131 def generate_exe(self, plat: Platform) -> str: 132 archive_filename = os.path.basename(self.generate_url(plat)) 133 filename = os.path.splitext(archive_filename)[0] 134 return f"./{filename}" 135 136 137 @dataclass(frozen=True) 138 class Coursier: 139 """The Coursier tool and various utilities, prepared for use via `immutable_input_digests`.""" 140 141 coursier: DownloadedExternalTool 142 _digest: Digest 143 144 bin_dir: ClassVar[str] = "__coursier" 145 fetch_wrapper_script: ClassVar[str] = f"{bin_dir}/coursier_fetch_wrapper_script.sh" 146 post_processing_script: ClassVar[str] = f"{bin_dir}/coursier_post_processing_script.py" 147 post_process_stderr: ClassVar[str] = f"{bin_dir}/coursier_post_process_stderr.py" 148 cache_name: ClassVar[str] = "coursier" 149 cache_dir: ClassVar[str] = ".cache" 150 working_directory_placeholder: ClassVar[str] = "___COURSIER_WORKING_DIRECTORY___" 151 152 def args(self, args: Iterable[str], *, wrapper: Iterable[str] = ()) -> tuple[str, ...]: 153 return ( 154 self.post_process_stderr, 155 *wrapper, 156 os.path.join(self.bin_dir, self.coursier.exe), 157 *args, 158 ) 159 160 @property 161 def env(self) -> dict[str, str]: 162 # NB: These variables have changed a few times, and they change again on `main`. But as of 163 # `v2.0.16+73-gddc6d9cc9` they are accurate. See: 164 # https://github.com/coursier/coursier/blob/v2.0.16+73-gddc6d9cc9/modules/paths/src/main/java/coursier/paths/CoursierPaths.java#L38-L48 165 return { 166 "COURSIER_CACHE": f"{self.cache_dir}/jdk", 167 "COURSIER_ARCHIVE_CACHE": f"{self.cache_dir}/arc", 168 "COURSIER_JVM_CACHE": f"{self.cache_dir}/v1", 169 } 170 171 @property 172 def append_only_caches(self) -> dict[str, str]: 173 return {self.cache_name: self.cache_dir} 174 175 @property 176 def immutable_input_digests(self) -> dict[str, Digest]: 177 return {self.bin_dir: self._digest} 178 179 180 @dataclass(frozen=True) 181 class CoursierWrapperProcess: 182 183 args: Tuple[str, ...] 184 input_digest: Digest 185 output_directories: Tuple[str, ...] 186 output_files: Tuple[str, ...] 187 description: str 188 189 190 @rule 191 async def invoke_coursier_wrapper( 192 bash: BashBinary, 193 coursier: Coursier, 194 request: CoursierWrapperProcess, 195 ) -> Process: 196 197 return Process( 198 argv=coursier.args( 199 request.args, 200 wrapper=[bash.path, coursier.fetch_wrapper_script], 201 ), 202 input_digest=request.input_digest, 203 immutable_input_digests=coursier.immutable_input_digests, 204 output_directories=request.output_directories, 205 output_files=request.output_files, 206 append_only_caches=coursier.append_only_caches, 207 env=coursier.env, 208 description=request.description, 209 level=LogLevel.DEBUG, 210 ) 211 212 213 @rule 214 async def setup_coursier( 215 coursier_subsystem: CoursierSubsystem, 216 python: PythonBinary, 217 ) -> Coursier: 218 repos_args = " ".join(f"-r={shlex.quote(repo)}" for repo in coursier_subsystem.options.repos) 219 coursier_wrapper_script = COURSIER_FETCH_WRAPPER_SCRIPT.format( 220 repos_args=repos_args, 221 coursier_working_directory=Coursier.working_directory_placeholder, 222 python_path=python.path, 223 coursier_bin_dir=Coursier.bin_dir, 224 ) 225 226 post_process_stderr = POST_PROCESS_COURSIER_STDERR_SCRIPT.format(python_path=python.path) 227 228 downloaded_coursier_get = Get( 229 DownloadedExternalTool, 230 ExternalToolRequest, 231 coursier_subsystem.get_request(Platform.current), 232 ) 233 wrapper_scripts_digest_get = Get( 234 Digest, 235 CreateDigest( 236 [ 237 FileContent( 238 os.path.basename(Coursier.fetch_wrapper_script), 239 coursier_wrapper_script.encode("utf-8"), 240 is_executable=True, 241 ), 242 FileContent( 243 os.path.basename(Coursier.post_processing_script), 244 COURSIER_POST_PROCESSING_SCRIPT.encode("utf-8"), 245 is_executable=True, 246 ), 247 FileContent( 248 os.path.basename(Coursier.post_process_stderr), 249 post_process_stderr.encode("utf-8"), 250 is_executable=True, 251 ), 252 ] 253 ), 254 ) 255 256 downloaded_coursier, wrapper_scripts_digest = await MultiGet( 257 downloaded_coursier_get, wrapper_scripts_digest_get 258 ) 259 260 return Coursier( 261 coursier=downloaded_coursier, 262 _digest=await Get( 263 Digest, 264 MergeDigests( 265 [ 266 downloaded_coursier.digest, 267 wrapper_scripts_digest, 268 ] 269 ), 270 ), 271 ) 272 273 274 def rules(): 275 return [ 276 *collect_rules(), 277 *external_tool.rules(), 278 ] 279 [end of src/python/pants/jvm/resolve/coursier_setup.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/src/python/pants/jvm/resolve/coursier_setup.py b/src/python/pants/jvm/resolve/coursier_setup.py --- a/src/python/pants/jvm/resolve/coursier_setup.py +++ b/src/python/pants/jvm/resolve/coursier_setup.py @@ -215,7 +215,10 @@ coursier_subsystem: CoursierSubsystem, python: PythonBinary, ) -> Coursier: - repos_args = " ".join(f"-r={shlex.quote(repo)}" for repo in coursier_subsystem.options.repos) + repos_args = ( + " ".join(f"-r={shlex.quote(repo)}" for repo in coursier_subsystem.options.repos) + + " --no-default" + ) coursier_wrapper_script = COURSIER_FETCH_WRAPPER_SCRIPT.format( repos_args=repos_args, coursier_working_directory=Coursier.working_directory_placeholder,
{"golden_diff": "diff --git a/src/python/pants/jvm/resolve/coursier_setup.py b/src/python/pants/jvm/resolve/coursier_setup.py\n--- a/src/python/pants/jvm/resolve/coursier_setup.py\n+++ b/src/python/pants/jvm/resolve/coursier_setup.py\n@@ -215,7 +215,10 @@\n coursier_subsystem: CoursierSubsystem,\n python: PythonBinary,\n ) -> Coursier:\n- repos_args = \" \".join(f\"-r={shlex.quote(repo)}\" for repo in coursier_subsystem.options.repos)\n+ repos_args = (\n+ \" \".join(f\"-r={shlex.quote(repo)}\" for repo in coursier_subsystem.options.repos)\n+ + \" --no-default\"\n+ )\n coursier_wrapper_script = COURSIER_FETCH_WRAPPER_SCRIPT.format(\n repos_args=repos_args,\n coursier_working_directory=Coursier.working_directory_placeholder,\n", "issue": "coursier post processing script choking on null dependency file in report json\n**Describe the bug**\r\nOne of our 3rdparty jars (`org.geotools:gt-main:9.2` from the `https://repo.osgeo.org/repository/release` repositiory) transitively depends on a `javax.media:jai_core:1.1.3` jar. For whatever reason, Coursier's resolve report json lists `\"file\": null` for that jar, which the post processing script doesn't currently handle.\r\n\r\n**Pants version**\r\n2.9.0\r\n\r\n**OS**\r\nEncountered on MacOS, though expect we would see this on Linux as well\r\n\r\n**Additional info**\r\nCoursier config:\r\n```\r\n[coursier]\r\nrepos = [\r\n \"https://maven-central.storage-download.googleapis.com/maven2\",\r\n \"https://repo1.maven.org/maven2\",\r\n \"https://repo.osgeo.org/repository/release\",\r\n]\r\n```\r\n\r\naffected jvm_artifact:\r\n```\r\njvm_artifact(\r\n group=\"org.geotools\",\r\n artifact=\"gt-main\",\r\n version=\"9.2\",\r\n)\r\n```\r\n\r\n[slack discussion](https://pantsbuild.slack.com/archives/C046T6T9U/p1645598534013879) with links to full resolve output and json\r\n\n", "before_files": [{"content": "# Copyright 2021 Pants project contributors (see CONTRIBUTORS.md).\n# Licensed under the Apache License, Version 2.0 (see LICENSE).\n\nfrom __future__ import annotations\n\nimport os\nimport shlex\nimport textwrap\nfrom dataclasses import dataclass\nfrom typing import ClassVar, Iterable, Tuple\n\nfrom pants.core.util_rules import external_tool\nfrom pants.core.util_rules.external_tool import (\n DownloadedExternalTool,\n ExternalToolRequest,\n TemplatedExternalTool,\n)\nfrom pants.engine.fs import CreateDigest, Digest, FileContent, MergeDigests\nfrom pants.engine.platform import Platform\nfrom pants.engine.process import BashBinary, Process\nfrom pants.engine.rules import Get, MultiGet, collect_rules, rule\nfrom pants.python.binaries import PythonBinary\nfrom pants.util.logging import LogLevel\n\nCOURSIER_POST_PROCESSING_SCRIPT = textwrap.dedent(\n \"\"\"\\\n import json\n import sys\n import os\n from pathlib import PurePath\n from shutil import copyfile\n\n report = json.load(open(sys.argv[1]))\n\n # Mapping from dest path to source path. It is ok to capture the same output filename multiple\n # times if the source is the same as well.\n classpath = dict()\n for dep in report['dependencies']:\n source = PurePath(dep['file'])\n dest_name = dep['coord'].replace(\":\", \"_\")\n _, ext = os.path.splitext(source)\n classpath_dest = f\"classpath/{dest_name}{ext}\"\n\n existing_source = classpath.get(classpath_dest)\n if existing_source:\n if existing_source == source:\n # We've already captured this file.\n continue\n raise Exception(\n f\"Duplicate jar name {classpath_dest} with incompatible source:\\\\n\"\n f\" {source}\\\\n\"\n f\" {existing_source}\\\\n\"\n )\n classpath[classpath_dest] = source\n copyfile(source, classpath_dest)\n \"\"\"\n)\n\nCOURSIER_FETCH_WRAPPER_SCRIPT = textwrap.dedent(\n \"\"\"\\\n set -eux\n\n coursier_exe=\"$1\"\n shift\n json_output_file=\"$1\"\n shift\n\n working_dir=\"$(pwd)\"\n \"$coursier_exe\" fetch {repos_args} \\\n --json-output-file=\"$json_output_file\" \\\n \"${{@//{coursier_working_directory}/$working_dir}}\"\n /bin/mkdir -p classpath\n {python_path} {coursier_bin_dir}/coursier_post_processing_script.py \"$json_output_file\"\n \"\"\"\n)\n\n\n# TODO: Coursier renders setrlimit error line on macOS.\n# see https://github.com/pantsbuild/pants/issues/13942.\nPOST_PROCESS_COURSIER_STDERR_SCRIPT = textwrap.dedent(\n \"\"\"\\\n #!{python_path}\n import sys\n from subprocess import run, PIPE\n\n proc = run(sys.argv[1:], stdout=PIPE, stderr=PIPE)\n\n sys.stdout.buffer.write(proc.stdout)\n sys.stderr.buffer.write(proc.stderr.replace(b\"setrlimit to increase file descriptor limit failed, errno 22\\\\n\", b\"\"))\n sys.exit(proc.returncode)\n \"\"\"\n)\n\n\nclass CoursierSubsystem(TemplatedExternalTool):\n options_scope = \"coursier\"\n name = \"coursier\"\n help = \"A dependency resolver for the Maven ecosystem.\"\n\n default_version = \"v2.0.16-169-g194ebc55c\"\n default_known_versions = [\n \"v2.0.16-169-g194ebc55c|linux_arm64 |da38c97d55967505b8454c20a90370c518044829398b9bce8b637d194d79abb3|18114472\",\n \"v2.0.16-169-g194ebc55c|linux_x86_64|4c61a634c4bd2773b4543fe0fc32210afd343692891121cddb447204b48672e8|18486946\",\n \"v2.0.16-169-g194ebc55c|macos_arm64 |15bce235d223ef1d022da30b67b4c64e9228d236b876c834b64e029bbe824c6f|17957182\",\n \"v2.0.16-169-g194ebc55c|macos_x86_64|15bce235d223ef1d022da30b67b4c64e9228d236b876c834b64e029bbe824c6f|17957182\",\n ]\n default_url_template = (\n \"https://github.com/coursier/coursier/releases/download/{version}/cs-{platform}.gz\"\n )\n default_url_platform_mapping = {\n \"macos_arm64\": \"x86_64-apple-darwin\",\n \"macos_x86_64\": \"x86_64-apple-darwin\",\n \"linux_arm64\": \"aarch64-pc-linux\",\n \"linux_x86_64\": \"x86_64-pc-linux\",\n }\n\n @classmethod\n def register_options(cls, register) -> None:\n super().register_options(register)\n register(\n \"--repos\",\n type=list,\n member_type=str,\n default=[\n \"https://maven-central.storage-download.googleapis.com/maven2\",\n \"https://repo1.maven.org/maven2\",\n ],\n help=(\"Maven style repositories to resolve artifacts from.\"),\n )\n\n def generate_exe(self, plat: Platform) -> str:\n archive_filename = os.path.basename(self.generate_url(plat))\n filename = os.path.splitext(archive_filename)[0]\n return f\"./{filename}\"\n\n\n@dataclass(frozen=True)\nclass Coursier:\n \"\"\"The Coursier tool and various utilities, prepared for use via `immutable_input_digests`.\"\"\"\n\n coursier: DownloadedExternalTool\n _digest: Digest\n\n bin_dir: ClassVar[str] = \"__coursier\"\n fetch_wrapper_script: ClassVar[str] = f\"{bin_dir}/coursier_fetch_wrapper_script.sh\"\n post_processing_script: ClassVar[str] = f\"{bin_dir}/coursier_post_processing_script.py\"\n post_process_stderr: ClassVar[str] = f\"{bin_dir}/coursier_post_process_stderr.py\"\n cache_name: ClassVar[str] = \"coursier\"\n cache_dir: ClassVar[str] = \".cache\"\n working_directory_placeholder: ClassVar[str] = \"___COURSIER_WORKING_DIRECTORY___\"\n\n def args(self, args: Iterable[str], *, wrapper: Iterable[str] = ()) -> tuple[str, ...]:\n return (\n self.post_process_stderr,\n *wrapper,\n os.path.join(self.bin_dir, self.coursier.exe),\n *args,\n )\n\n @property\n def env(self) -> dict[str, str]:\n # NB: These variables have changed a few times, and they change again on `main`. But as of\n # `v2.0.16+73-gddc6d9cc9` they are accurate. See:\n # https://github.com/coursier/coursier/blob/v2.0.16+73-gddc6d9cc9/modules/paths/src/main/java/coursier/paths/CoursierPaths.java#L38-L48\n return {\n \"COURSIER_CACHE\": f\"{self.cache_dir}/jdk\",\n \"COURSIER_ARCHIVE_CACHE\": f\"{self.cache_dir}/arc\",\n \"COURSIER_JVM_CACHE\": f\"{self.cache_dir}/v1\",\n }\n\n @property\n def append_only_caches(self) -> dict[str, str]:\n return {self.cache_name: self.cache_dir}\n\n @property\n def immutable_input_digests(self) -> dict[str, Digest]:\n return {self.bin_dir: self._digest}\n\n\n@dataclass(frozen=True)\nclass CoursierWrapperProcess:\n\n args: Tuple[str, ...]\n input_digest: Digest\n output_directories: Tuple[str, ...]\n output_files: Tuple[str, ...]\n description: str\n\n\n@rule\nasync def invoke_coursier_wrapper(\n bash: BashBinary,\n coursier: Coursier,\n request: CoursierWrapperProcess,\n) -> Process:\n\n return Process(\n argv=coursier.args(\n request.args,\n wrapper=[bash.path, coursier.fetch_wrapper_script],\n ),\n input_digest=request.input_digest,\n immutable_input_digests=coursier.immutable_input_digests,\n output_directories=request.output_directories,\n output_files=request.output_files,\n append_only_caches=coursier.append_only_caches,\n env=coursier.env,\n description=request.description,\n level=LogLevel.DEBUG,\n )\n\n\n@rule\nasync def setup_coursier(\n coursier_subsystem: CoursierSubsystem,\n python: PythonBinary,\n) -> Coursier:\n repos_args = \" \".join(f\"-r={shlex.quote(repo)}\" for repo in coursier_subsystem.options.repos)\n coursier_wrapper_script = COURSIER_FETCH_WRAPPER_SCRIPT.format(\n repos_args=repos_args,\n coursier_working_directory=Coursier.working_directory_placeholder,\n python_path=python.path,\n coursier_bin_dir=Coursier.bin_dir,\n )\n\n post_process_stderr = POST_PROCESS_COURSIER_STDERR_SCRIPT.format(python_path=python.path)\n\n downloaded_coursier_get = Get(\n DownloadedExternalTool,\n ExternalToolRequest,\n coursier_subsystem.get_request(Platform.current),\n )\n wrapper_scripts_digest_get = Get(\n Digest,\n CreateDigest(\n [\n FileContent(\n os.path.basename(Coursier.fetch_wrapper_script),\n coursier_wrapper_script.encode(\"utf-8\"),\n is_executable=True,\n ),\n FileContent(\n os.path.basename(Coursier.post_processing_script),\n COURSIER_POST_PROCESSING_SCRIPT.encode(\"utf-8\"),\n is_executable=True,\n ),\n FileContent(\n os.path.basename(Coursier.post_process_stderr),\n post_process_stderr.encode(\"utf-8\"),\n is_executable=True,\n ),\n ]\n ),\n )\n\n downloaded_coursier, wrapper_scripts_digest = await MultiGet(\n downloaded_coursier_get, wrapper_scripts_digest_get\n )\n\n return Coursier(\n coursier=downloaded_coursier,\n _digest=await Get(\n Digest,\n MergeDigests(\n [\n downloaded_coursier.digest,\n wrapper_scripts_digest,\n ]\n ),\n ),\n )\n\n\ndef rules():\n return [\n *collect_rules(),\n *external_tool.rules(),\n ]\n", "path": "src/python/pants/jvm/resolve/coursier_setup.py"}]}
4,060
208
gh_patches_debug_51710
rasdani/github-patches
git_diff
getsentry__sentry-python-2069
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> Cannot import appengine ### How do you use Sentry? Sentry Saas (sentry.io) ### Version 1.18.0 ### Steps to Reproduce Install the SDK within any project that is not pinning urllib3 < 2.0.0 ### Expected Result ability to import appengine ### Actual Result Cannot import appengine as gaecontrib. As per urllib 2.0.0 release: https://github.com/urllib3/urllib3/tree/2.0.0 Removed urllib3.contrib.appengine.AppEngineManager and support for Google App Engine Standard Environment (https://github.com/urllib3/urllib3/issues/2044). </issue> <code> [start of setup.py] 1 #!/usr/bin/env python 2 3 """ 4 Sentry-Python - Sentry SDK for Python 5 ===================================== 6 7 **Sentry-Python is an SDK for Sentry.** Check out `GitHub 8 <https://github.com/getsentry/sentry-python>`_ to find out more. 9 """ 10 11 import os 12 from setuptools import setup, find_packages 13 14 here = os.path.abspath(os.path.dirname(__file__)) 15 16 17 def get_file_text(file_name): 18 with open(os.path.join(here, file_name)) as in_file: 19 return in_file.read() 20 21 22 setup( 23 name="sentry-sdk", 24 version="1.21.1", 25 author="Sentry Team and Contributors", 26 author_email="[email protected]", 27 url="https://github.com/getsentry/sentry-python", 28 project_urls={ 29 "Documentation": "https://docs.sentry.io/platforms/python/", 30 "Changelog": "https://github.com/getsentry/sentry-python/blob/master/CHANGELOG.md", 31 }, 32 description="Python client for Sentry (https://sentry.io)", 33 long_description=get_file_text("README.md"), 34 long_description_content_type="text/markdown", 35 packages=find_packages(exclude=("tests", "tests.*")), 36 # PEP 561 37 package_data={"sentry_sdk": ["py.typed"]}, 38 zip_safe=False, 39 license="MIT", 40 install_requires=[ 41 'urllib3>=1.25.7; python_version<="3.4"', 42 'urllib3>=1.26.9; python_version=="3.5"', 43 'urllib3>=1.26.11; python_version >="3.6"', 44 "certifi", 45 ], 46 extras_require={ 47 "flask": ["flask>=0.11", "blinker>=1.1"], 48 "quart": ["quart>=0.16.1", "blinker>=1.1"], 49 "bottle": ["bottle>=0.12.13"], 50 "falcon": ["falcon>=1.4"], 51 "django": ["django>=1.8"], 52 "sanic": ["sanic>=0.8"], 53 "celery": ["celery>=3"], 54 "huey": ["huey>=2"], 55 "beam": ["apache-beam>=2.12"], 56 "arq": ["arq>=0.23"], 57 "rq": ["rq>=0.6"], 58 "aiohttp": ["aiohttp>=3.5"], 59 "tornado": ["tornado>=5"], 60 "sqlalchemy": ["sqlalchemy>=1.2"], 61 "pyspark": ["pyspark>=2.4.4"], 62 "pure_eval": ["pure_eval", "executing", "asttokens"], 63 "chalice": ["chalice>=1.16.0"], 64 "httpx": ["httpx>=0.16.0"], 65 "starlette": ["starlette>=0.19.1"], 66 "starlite": ["starlite>=1.48"], 67 "fastapi": ["fastapi>=0.79.0"], 68 "pymongo": ["pymongo>=3.1"], 69 "opentelemetry": ["opentelemetry-distro>=0.35b0"], 70 "grpcio": ["grpcio>=1.21.1"] 71 }, 72 classifiers=[ 73 "Development Status :: 5 - Production/Stable", 74 "Environment :: Web Environment", 75 "Intended Audience :: Developers", 76 "License :: OSI Approved :: BSD License", 77 "Operating System :: OS Independent", 78 "Programming Language :: Python", 79 "Programming Language :: Python :: 2", 80 "Programming Language :: Python :: 2.7", 81 "Programming Language :: Python :: 3", 82 "Programming Language :: Python :: 3.4", 83 "Programming Language :: Python :: 3.5", 84 "Programming Language :: Python :: 3.6", 85 "Programming Language :: Python :: 3.7", 86 "Programming Language :: Python :: 3.8", 87 "Programming Language :: Python :: 3.9", 88 "Programming Language :: Python :: 3.10", 89 "Topic :: Software Development :: Libraries :: Python Modules", 90 ], 91 options={"bdist_wheel": {"universal": "1"}}, 92 ) 93 [end of setup.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/setup.py b/setup.py --- a/setup.py +++ b/setup.py @@ -41,6 +41,7 @@ 'urllib3>=1.25.7; python_version<="3.4"', 'urllib3>=1.26.9; python_version=="3.5"', 'urllib3>=1.26.11; python_version >="3.6"', + 'urllib3<2.0.0', "certifi", ], extras_require={
{"golden_diff": "diff --git a/setup.py b/setup.py\n--- a/setup.py\n+++ b/setup.py\n@@ -41,6 +41,7 @@\n 'urllib3>=1.25.7; python_version<=\"3.4\"',\n 'urllib3>=1.26.9; python_version==\"3.5\"',\n 'urllib3>=1.26.11; python_version >=\"3.6\"',\n+ 'urllib3<2.0.0',\n \"certifi\",\n ],\n extras_require={\n", "issue": "Cannot import appengine\n### How do you use Sentry?\n\nSentry Saas (sentry.io)\n\n### Version\n\n1.18.0\n\n### Steps to Reproduce\n\nInstall the SDK within any project that is not pinning urllib3 < 2.0.0\n\n### Expected Result\n\nability to import appengine\n\n### Actual Result\n\nCannot import appengine as gaecontrib.\r\nAs per urllib 2.0.0 release: https://github.com/urllib3/urllib3/tree/2.0.0\r\n\r\nRemoved urllib3.contrib.appengine.AppEngineManager and support for Google App Engine Standard Environment (https://github.com/urllib3/urllib3/issues/2044).\n", "before_files": [{"content": "#!/usr/bin/env python\n\n\"\"\"\nSentry-Python - Sentry SDK for Python\n=====================================\n\n**Sentry-Python is an SDK for Sentry.** Check out `GitHub\n<https://github.com/getsentry/sentry-python>`_ to find out more.\n\"\"\"\n\nimport os\nfrom setuptools import setup, find_packages\n\nhere = os.path.abspath(os.path.dirname(__file__))\n\n\ndef get_file_text(file_name):\n with open(os.path.join(here, file_name)) as in_file:\n return in_file.read()\n\n\nsetup(\n name=\"sentry-sdk\",\n version=\"1.21.1\",\n author=\"Sentry Team and Contributors\",\n author_email=\"[email protected]\",\n url=\"https://github.com/getsentry/sentry-python\",\n project_urls={\n \"Documentation\": \"https://docs.sentry.io/platforms/python/\",\n \"Changelog\": \"https://github.com/getsentry/sentry-python/blob/master/CHANGELOG.md\",\n },\n description=\"Python client for Sentry (https://sentry.io)\",\n long_description=get_file_text(\"README.md\"),\n long_description_content_type=\"text/markdown\",\n packages=find_packages(exclude=(\"tests\", \"tests.*\")),\n # PEP 561\n package_data={\"sentry_sdk\": [\"py.typed\"]},\n zip_safe=False,\n license=\"MIT\",\n install_requires=[\n 'urllib3>=1.25.7; python_version<=\"3.4\"',\n 'urllib3>=1.26.9; python_version==\"3.5\"',\n 'urllib3>=1.26.11; python_version >=\"3.6\"',\n \"certifi\",\n ],\n extras_require={\n \"flask\": [\"flask>=0.11\", \"blinker>=1.1\"],\n \"quart\": [\"quart>=0.16.1\", \"blinker>=1.1\"],\n \"bottle\": [\"bottle>=0.12.13\"],\n \"falcon\": [\"falcon>=1.4\"],\n \"django\": [\"django>=1.8\"],\n \"sanic\": [\"sanic>=0.8\"],\n \"celery\": [\"celery>=3\"],\n \"huey\": [\"huey>=2\"],\n \"beam\": [\"apache-beam>=2.12\"],\n \"arq\": [\"arq>=0.23\"],\n \"rq\": [\"rq>=0.6\"],\n \"aiohttp\": [\"aiohttp>=3.5\"],\n \"tornado\": [\"tornado>=5\"],\n \"sqlalchemy\": [\"sqlalchemy>=1.2\"],\n \"pyspark\": [\"pyspark>=2.4.4\"],\n \"pure_eval\": [\"pure_eval\", \"executing\", \"asttokens\"],\n \"chalice\": [\"chalice>=1.16.0\"],\n \"httpx\": [\"httpx>=0.16.0\"],\n \"starlette\": [\"starlette>=0.19.1\"],\n \"starlite\": [\"starlite>=1.48\"],\n \"fastapi\": [\"fastapi>=0.79.0\"],\n \"pymongo\": [\"pymongo>=3.1\"],\n \"opentelemetry\": [\"opentelemetry-distro>=0.35b0\"],\n \"grpcio\": [\"grpcio>=1.21.1\"]\n },\n classifiers=[\n \"Development Status :: 5 - Production/Stable\",\n \"Environment :: Web Environment\",\n \"Intended Audience :: Developers\",\n \"License :: OSI Approved :: BSD License\",\n \"Operating System :: OS Independent\",\n \"Programming Language :: Python\",\n \"Programming Language :: Python :: 2\",\n \"Programming Language :: Python :: 2.7\",\n \"Programming Language :: Python :: 3\",\n \"Programming Language :: Python :: 3.4\",\n \"Programming Language :: Python :: 3.5\",\n \"Programming Language :: Python :: 3.6\",\n \"Programming Language :: Python :: 3.7\",\n \"Programming Language :: Python :: 3.8\",\n \"Programming Language :: Python :: 3.9\",\n \"Programming Language :: Python :: 3.10\",\n \"Topic :: Software Development :: Libraries :: Python Modules\",\n ],\n options={\"bdist_wheel\": {\"universal\": \"1\"}},\n)\n", "path": "setup.py"}]}
1,787
119
gh_patches_debug_3474
rasdani/github-patches
git_diff
pymedusa__Medusa-4725
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> YGGTorrent : Unable to bdecode torrent **Describe the bug** The YggTorrent provider seems to have a bug. The downloads are not correctly handled to rtorrent. The "unable to bdecode torrent" occurs. Thank you for your help. **Medusa (please complete the following information):** Branch: master  Commit: e7c4ff3bc9a48b47c932b0e40a539d62e3a62f64 Version: v0.2.6 Database: 44.11 Python Version: | 2.7.12 (default, Dec 4 2017, 14:50:18) [GCC 5.4.0 20160609] SSL Version: | OpenSSL 1.0.2g 1 Mar 2016 OS: | Linux-4.12.14-041214-generic-x86_64-with-Ubuntu-16.04-xenial **Logs:** <details> ``` 2018-07-23 12:45:32 WARNING SEARCHQUEUE-FORCED-325517 :: [e7c4ff3] Unable to bdecode torrent. Invalid torrent: Salvation.S02E01.FASTSUB.VOSTFR.720p.AMZN.WEB-DL.DDP5.1.H.264-GOLD. Deleting cached result if exists 2018-07-23 12:45:32 INFO SEARCHQUEUE-FORCED-325517 :: [e7c4ff3] Downloading Salvation.S02E01.FASTSUB.VOSTFR.720p.AMZN.WEB-DL.DDP5.1.H.264-GOLD with 36 seeders and 0 leechers and size 938.52 MB from Yggtorrent 2018-07-23 12:45:32 INFO SEARCHQUEUE-FORCED-325517 :: [Yggtorrent] :: [e7c4ff3] Picked Salvation.S02E01.FASTSUB.VOSTFR.720p.AMZN.WEB-DL.DDP5.1.H.264-GOLD as the best result. 2018-07-23 12:45:32 INFO SEARCHQUEUE-FORCED-325517 :: [Yggtorrent] :: [e7c4ff3] Quality of Salvation.S02E01.(FASTSUB.VOSTFR.720p.WEBRip.x265.NSP) is 720p WEB-DL 2018-07-23 12:45:32 INFO SEARCHQUEUE-FORCED-325517 :: [Yggtorrent] :: [e7c4ff3] Quality of Salvation.S02E01.FASTSUB.VOSTFR.720p.HDTV.x264.-.ZKH is 720p HDTV 2018-07-23 12:45:32 INFO SEARCHQUEUE-FORCED-325517 :: [Yggtorrent] :: [e7c4ff3] Quality of Salvation.S02E01.FASTSUB.VOSTFR.720p.AMZN.WEB-DL.DDP5.1.H.264-GOLD is 720p WEB-DL 2018-07-23 12:45:31 INFO SEARCHQUEUE-FORCED-325517 :: [Yggtorrent] :: [e7c4ff3] Performing episode search for Salvation``` </details> </issue> <code> [start of medusa/providers/torrent/html/yggtorrent.py] 1 # coding=utf-8 2 3 """Provider code for Yggtorrent.""" 4 5 from __future__ import unicode_literals 6 7 import logging 8 import re 9 10 from medusa import tv 11 from medusa.bs4_parser import BS4Parser 12 from medusa.helper.common import ( 13 convert_size, 14 try_int, 15 ) 16 from medusa.logger.adapters.style import BraceAdapter 17 from medusa.providers.torrent.torrent_provider import TorrentProvider 18 19 from requests.compat import urljoin 20 21 log = BraceAdapter(logging.getLogger(__name__)) 22 log.logger.addHandler(logging.NullHandler()) 23 24 25 class YggtorrentProvider(TorrentProvider): 26 """Yggtorrent Torrent provider.""" 27 28 torrent_id_pattern = re.compile(r'\/(\d+)-') 29 30 def __init__(self): 31 """Initialize the class.""" 32 super(YggtorrentProvider, self).__init__('Yggtorrent') 33 34 # Credentials 35 self.username = None 36 self.password = None 37 38 # URLs 39 self.url = 'https://ww1.yggtorrent.is' 40 self.urls = { 41 'login': urljoin(self.url, 'user/login'), 42 'search': urljoin(self.url, 'engine/search'), 43 'download': urljoin(self.url, 'engine/download_torrent?id={0}') 44 } 45 46 # Proper Strings 47 self.proper_strings = ['PROPER', 'REPACK', 'REAL', 'RERIP'] 48 49 # Torrent Stats 50 self.minseed = None 51 self.minleech = None 52 53 # Cache 54 self.cache = tv.Cache(self, min_time=20) 55 56 def search(self, search_strings, age=0, ep_obj=None, **kwargs): 57 """ 58 Search a provider and parse the results. 59 60 :param search_strings: A dict with mode (key) and the search value (value) 61 :param age: Not used 62 :param ep_obj: Not used 63 :returns: A list of search results (structure) 64 """ 65 results = [] 66 if not self.login(): 67 return results 68 69 # Search Params 70 search_params = { 71 'category': 2145, 72 'do': 'search' 73 } 74 75 for mode in search_strings: 76 log.debug('Search mode: {0}', mode) 77 78 for search_string in search_strings[mode]: 79 80 if mode != 'RSS': 81 log.debug('Search string: {search}', 82 {'search': search_string}) 83 84 search_params['name'] = re.sub(r'[()]', '', search_string) 85 86 response = self.session.get(self.urls['search'], params=search_params) 87 if not response or not response.text: 88 log.debug('No data returned from provider') 89 continue 90 91 results += self.parse(response.text, mode) 92 93 return results 94 95 def parse(self, data, mode): 96 """ 97 Parse search results for items. 98 99 :param data: The raw response from a search 100 :param mode: The current mode used to search, e.g. RSS 101 102 :return: A list of items found 103 """ 104 # Units 105 units = ['O', 'KO', 'MO', 'GO', 'TO', 'PO'] 106 107 items = [] 108 109 with BS4Parser(data, 'html5lib') as html: 110 torrent_table = html.find(class_='table-responsive results') 111 torrent_rows = torrent_table('tr') if torrent_table else [] 112 113 # Continue only if at least one Release is found 114 if len(torrent_rows) < 2: 115 log.debug('Data returned from provider does not contain any torrents') 116 return items 117 118 # Skip column headers 119 for result in torrent_rows[1:]: 120 cells = result('td') 121 if len(cells) < 9: 122 continue 123 124 try: 125 info = cells[1].find('a') 126 title = info.get_text(strip=True) 127 download_url = info.get('href') 128 if not (title and download_url): 129 continue 130 131 torrent_id = self.torrent_id_pattern.search(download_url) 132 download_url = self.urls['download'].format(torrent_id.group(1)) 133 134 seeders = try_int(cells[7].get_text(strip=True), 0) 135 leechers = try_int(cells[8].get_text(strip=True), 0) 136 137 # Filter unseeded torrent 138 if seeders < min(self.minseed, 1): 139 if mode != 'RSS': 140 log.debug("Discarding torrent because it doesn't meet the" 141 " minimum seeders: {0}. Seeders: {1}", 142 title, seeders) 143 continue 144 145 torrent_size = cells[5].get_text() 146 size = convert_size(torrent_size, sep='', units=units, default=-1) 147 148 pubdate_raw = cells[4].find('div', class_='hidden').get_text(strip=True) 149 pubdate = self.parse_pubdate(pubdate_raw, fromtimestamp=True) 150 151 item = { 152 'title': title, 153 'link': download_url, 154 'size': size, 155 'seeders': seeders, 156 'leechers': leechers, 157 'pubdate': pubdate, 158 } 159 if mode != 'RSS': 160 log.debug('Found result: {0} with {1} seeders and {2} leechers', 161 title, seeders, leechers) 162 163 items.append(item) 164 except (AttributeError, TypeError, KeyError, ValueError, IndexError): 165 log.exception('Failed parsing provider.') 166 167 return items 168 169 def login(self): 170 """Login method used for logging in before doing search and torrent downloads.""" 171 login_params = { 172 'id': self.username, 173 'pass': self.password 174 } 175 176 login_resp = self.session.post(self.urls['login'], data=login_params) 177 if not login_resp: 178 log.warning('Invalid username or password. Check your settings') 179 return False 180 181 response = self.session.get(self.url) 182 if not response: 183 log.warning('Unable to connect to provider') 184 return False 185 186 if 'Bienvenue' not in response.text: 187 log.warning('Unable to login to provider') 188 return False 189 190 return True 191 192 193 provider = YggtorrentProvider() 194 [end of medusa/providers/torrent/html/yggtorrent.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/medusa/providers/torrent/html/yggtorrent.py b/medusa/providers/torrent/html/yggtorrent.py --- a/medusa/providers/torrent/html/yggtorrent.py +++ b/medusa/providers/torrent/html/yggtorrent.py @@ -36,7 +36,7 @@ self.password = None # URLs - self.url = 'https://ww1.yggtorrent.is' + self.url = 'https://ww2.yggtorrent.is' self.urls = { 'login': urljoin(self.url, 'user/login'), 'search': urljoin(self.url, 'engine/search'),
{"golden_diff": "diff --git a/medusa/providers/torrent/html/yggtorrent.py b/medusa/providers/torrent/html/yggtorrent.py\n--- a/medusa/providers/torrent/html/yggtorrent.py\n+++ b/medusa/providers/torrent/html/yggtorrent.py\n@@ -36,7 +36,7 @@\n self.password = None\n \n # URLs\n- self.url = 'https://ww1.yggtorrent.is'\n+ self.url = 'https://ww2.yggtorrent.is'\n self.urls = {\n 'login': urljoin(self.url, 'user/login'),\n 'search': urljoin(self.url, 'engine/search'),\n", "issue": "YGGTorrent : Unable to bdecode torrent\n**Describe the bug**\r\nThe YggTorrent provider seems to have a bug. The downloads are not correctly handled to rtorrent. The \"unable to bdecode torrent\" occurs. Thank you for your help. \r\n\r\n\r\n**Medusa (please complete the following information):**\r\n\r\nBranch:\u00a0master\u00a0\r\nCommit:\u00a0e7c4ff3bc9a48b47c932b0e40a539d62e3a62f64\u00a0Version:\u00a0v0.2.6\u00a0Database:\u00a044.11\r\nPython Version: | 2.7.12 (default, Dec 4 2017, 14:50:18) [GCC 5.4.0 20160609]\r\nSSL Version: | OpenSSL 1.0.2g 1 Mar 2016\r\nOS: | Linux-4.12.14-041214-generic-x86_64-with-Ubuntu-16.04-xenial\r\n\r\n**Logs:**\r\n<details>\r\n\r\n```\r\n2018-07-23 12:45:32 WARNING SEARCHQUEUE-FORCED-325517 :: [e7c4ff3] Unable to bdecode torrent. Invalid torrent: Salvation.S02E01.FASTSUB.VOSTFR.720p.AMZN.WEB-DL.DDP5.1.H.264-GOLD. Deleting cached result if exists\r\n2018-07-23 12:45:32 INFO SEARCHQUEUE-FORCED-325517 :: [e7c4ff3] Downloading Salvation.S02E01.FASTSUB.VOSTFR.720p.AMZN.WEB-DL.DDP5.1.H.264-GOLD with 36 seeders and 0 leechers and size 938.52 MB from Yggtorrent\r\n2018-07-23 12:45:32 INFO SEARCHQUEUE-FORCED-325517 :: [Yggtorrent] :: [e7c4ff3] Picked Salvation.S02E01.FASTSUB.VOSTFR.720p.AMZN.WEB-DL.DDP5.1.H.264-GOLD as the best result.\r\n2018-07-23 12:45:32 INFO SEARCHQUEUE-FORCED-325517 :: [Yggtorrent] :: [e7c4ff3] Quality of Salvation.S02E01.(FASTSUB.VOSTFR.720p.WEBRip.x265.NSP) is 720p WEB-DL\r\n2018-07-23 12:45:32 INFO SEARCHQUEUE-FORCED-325517 :: [Yggtorrent] :: [e7c4ff3] Quality of Salvation.S02E01.FASTSUB.VOSTFR.720p.HDTV.x264.-.ZKH is 720p HDTV\r\n2018-07-23 12:45:32 INFO SEARCHQUEUE-FORCED-325517 :: [Yggtorrent] :: [e7c4ff3] Quality of Salvation.S02E01.FASTSUB.VOSTFR.720p.AMZN.WEB-DL.DDP5.1.H.264-GOLD is 720p WEB-DL\r\n2018-07-23 12:45:31 INFO SEARCHQUEUE-FORCED-325517 :: [Yggtorrent] :: [e7c4ff3] Performing episode search for Salvation```\r\n</details>\r\n\r\n\n", "before_files": [{"content": "# coding=utf-8\n\n\"\"\"Provider code for Yggtorrent.\"\"\"\n\nfrom __future__ import unicode_literals\n\nimport logging\nimport re\n\nfrom medusa import tv\nfrom medusa.bs4_parser import BS4Parser\nfrom medusa.helper.common import (\n convert_size,\n try_int,\n)\nfrom medusa.logger.adapters.style import BraceAdapter\nfrom medusa.providers.torrent.torrent_provider import TorrentProvider\n\nfrom requests.compat import urljoin\n\nlog = BraceAdapter(logging.getLogger(__name__))\nlog.logger.addHandler(logging.NullHandler())\n\n\nclass YggtorrentProvider(TorrentProvider):\n \"\"\"Yggtorrent Torrent provider.\"\"\"\n\n torrent_id_pattern = re.compile(r'\\/(\\d+)-')\n\n def __init__(self):\n \"\"\"Initialize the class.\"\"\"\n super(YggtorrentProvider, self).__init__('Yggtorrent')\n\n # Credentials\n self.username = None\n self.password = None\n\n # URLs\n self.url = 'https://ww1.yggtorrent.is'\n self.urls = {\n 'login': urljoin(self.url, 'user/login'),\n 'search': urljoin(self.url, 'engine/search'),\n 'download': urljoin(self.url, 'engine/download_torrent?id={0}')\n }\n\n # Proper Strings\n self.proper_strings = ['PROPER', 'REPACK', 'REAL', 'RERIP']\n\n # Torrent Stats\n self.minseed = None\n self.minleech = None\n\n # Cache\n self.cache = tv.Cache(self, min_time=20)\n\n def search(self, search_strings, age=0, ep_obj=None, **kwargs):\n \"\"\"\n Search a provider and parse the results.\n\n :param search_strings: A dict with mode (key) and the search value (value)\n :param age: Not used\n :param ep_obj: Not used\n :returns: A list of search results (structure)\n \"\"\"\n results = []\n if not self.login():\n return results\n\n # Search Params\n search_params = {\n 'category': 2145,\n 'do': 'search'\n }\n\n for mode in search_strings:\n log.debug('Search mode: {0}', mode)\n\n for search_string in search_strings[mode]:\n\n if mode != 'RSS':\n log.debug('Search string: {search}',\n {'search': search_string})\n\n search_params['name'] = re.sub(r'[()]', '', search_string)\n\n response = self.session.get(self.urls['search'], params=search_params)\n if not response or not response.text:\n log.debug('No data returned from provider')\n continue\n\n results += self.parse(response.text, mode)\n\n return results\n\n def parse(self, data, mode):\n \"\"\"\n Parse search results for items.\n\n :param data: The raw response from a search\n :param mode: The current mode used to search, e.g. RSS\n\n :return: A list of items found\n \"\"\"\n # Units\n units = ['O', 'KO', 'MO', 'GO', 'TO', 'PO']\n\n items = []\n\n with BS4Parser(data, 'html5lib') as html:\n torrent_table = html.find(class_='table-responsive results')\n torrent_rows = torrent_table('tr') if torrent_table else []\n\n # Continue only if at least one Release is found\n if len(torrent_rows) < 2:\n log.debug('Data returned from provider does not contain any torrents')\n return items\n\n # Skip column headers\n for result in torrent_rows[1:]:\n cells = result('td')\n if len(cells) < 9:\n continue\n\n try:\n info = cells[1].find('a')\n title = info.get_text(strip=True)\n download_url = info.get('href')\n if not (title and download_url):\n continue\n\n torrent_id = self.torrent_id_pattern.search(download_url)\n download_url = self.urls['download'].format(torrent_id.group(1))\n\n seeders = try_int(cells[7].get_text(strip=True), 0)\n leechers = try_int(cells[8].get_text(strip=True), 0)\n\n # Filter unseeded torrent\n if seeders < min(self.minseed, 1):\n if mode != 'RSS':\n log.debug(\"Discarding torrent because it doesn't meet the\"\n \" minimum seeders: {0}. Seeders: {1}\",\n title, seeders)\n continue\n\n torrent_size = cells[5].get_text()\n size = convert_size(torrent_size, sep='', units=units, default=-1)\n\n pubdate_raw = cells[4].find('div', class_='hidden').get_text(strip=True)\n pubdate = self.parse_pubdate(pubdate_raw, fromtimestamp=True)\n\n item = {\n 'title': title,\n 'link': download_url,\n 'size': size,\n 'seeders': seeders,\n 'leechers': leechers,\n 'pubdate': pubdate,\n }\n if mode != 'RSS':\n log.debug('Found result: {0} with {1} seeders and {2} leechers',\n title, seeders, leechers)\n\n items.append(item)\n except (AttributeError, TypeError, KeyError, ValueError, IndexError):\n log.exception('Failed parsing provider.')\n\n return items\n\n def login(self):\n \"\"\"Login method used for logging in before doing search and torrent downloads.\"\"\"\n login_params = {\n 'id': self.username,\n 'pass': self.password\n }\n\n login_resp = self.session.post(self.urls['login'], data=login_params)\n if not login_resp:\n log.warning('Invalid username or password. Check your settings')\n return False\n\n response = self.session.get(self.url)\n if not response:\n log.warning('Unable to connect to provider')\n return False\n\n if 'Bienvenue' not in response.text:\n log.warning('Unable to login to provider')\n return False\n\n return True\n\n\nprovider = YggtorrentProvider()\n", "path": "medusa/providers/torrent/html/yggtorrent.py"}]}
3,243
145
gh_patches_debug_51406
rasdani/github-patches
git_diff
pytorch__ignite-1016
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> PyTorch dependency is lacking version constraint ## 🐛 Bug description <!-- A clear and concise description of what the bug is. --> PyTorch is a dependency of Ignite and, thus, is specified in `setup.py` https://github.com/pytorch/ignite/blob/4b311cc82fe45b3082661125cd7ee54007283fb0/setup.py#L24-L26 and `conda.recipe/meta.yaml`: https://github.com/pytorch/ignite/blob/4b311cc82fe45b3082661125cd7ee54007283fb0/conda.recipe/meta.yaml#L15-L23 The PyTorch dependency is lacking a version constraint which may work fine right now, but there is no guarantee that Ignite will be compatible with any future major PyTorch release (e.g. PyTorch v2.x). I suggest to constrain the PyTorch version that Ignite is compatible with, e.g. `>=1.0,<2` or `<2` if any `0.x` and `1.x` version works. If PyTorch has a new major release, even previous Ignite versions can become compatible with the new major PyTorch release (especially if no changes to the code are necessary) by making new bug fix releases with relaxed version constraints to include the new PyTorch version. In my opinion, it is highly preferable to be conservative about dependency version constraints through a [compatible release constraint](https://www.python.org/dev/peps/pep-0440/#compatible-release) in case the dependency conforms with semantic versioning. It is impossible to guarantee compatibility with a future major release of a dependency as its API can change arbitrarily. </issue> <code> [start of setup.py] 1 import os 2 import io 3 import re 4 from setuptools import setup, find_packages 5 6 7 def read(*names, **kwargs): 8 with io.open(os.path.join(os.path.dirname(__file__), *names), encoding=kwargs.get("encoding", "utf8")) as fp: 9 return fp.read() 10 11 12 def find_version(*file_paths): 13 version_file = read(*file_paths) 14 version_match = re.search(r"^__version__ = ['\"]([^'\"]*)['\"]", version_file, re.M) 15 if version_match: 16 return version_match.group(1) 17 raise RuntimeError("Unable to find version string.") 18 19 20 readme = read("README.md") 21 22 VERSION = find_version("ignite", "__init__.py") 23 24 requirements = [ 25 "torch", 26 ] 27 28 setup( 29 # Metadata 30 name="pytorch-ignite", 31 version=VERSION, 32 author="PyTorch Core Team", 33 author_email="[email protected]", 34 url="https://github.com/pytorch/ignite", 35 description="A lightweight library to help with training neural networks in PyTorch.", 36 long_description_content_type="text/markdown", 37 long_description=readme, 38 license="BSD", 39 # Package info 40 packages=find_packages(exclude=("tests", "tests.*",)), 41 zip_safe=True, 42 install_requires=requirements, 43 ) 44 [end of setup.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/setup.py b/setup.py --- a/setup.py +++ b/setup.py @@ -22,7 +22,7 @@ VERSION = find_version("ignite", "__init__.py") requirements = [ - "torch", + "torch>=1.0,<2", ] setup(
{"golden_diff": "diff --git a/setup.py b/setup.py\n--- a/setup.py\n+++ b/setup.py\n@@ -22,7 +22,7 @@\n VERSION = find_version(\"ignite\", \"__init__.py\")\n \n requirements = [\n- \"torch\",\n+ \"torch>=1.0,<2\",\n ]\n \n setup(\n", "issue": "PyTorch dependency is lacking version constraint\n## \ud83d\udc1b Bug description\r\n\r\n<!-- A clear and concise description of what the bug is. -->\r\nPyTorch is a dependency of Ignite and, thus, is specified in `setup.py`\r\n\r\nhttps://github.com/pytorch/ignite/blob/4b311cc82fe45b3082661125cd7ee54007283fb0/setup.py#L24-L26\r\n\r\nand `conda.recipe/meta.yaml`:\r\n\r\nhttps://github.com/pytorch/ignite/blob/4b311cc82fe45b3082661125cd7ee54007283fb0/conda.recipe/meta.yaml#L15-L23\r\n\r\nThe PyTorch dependency is lacking a version constraint which may work fine right now, but there is no guarantee that Ignite will be compatible with any future major PyTorch release (e.g. PyTorch v2.x).\r\n\r\nI suggest to constrain the PyTorch version that Ignite is compatible with, e.g. `>=1.0,<2` or `<2` if any `0.x` and `1.x` version works. If PyTorch has a new major release, even previous Ignite versions can become compatible with the new major PyTorch release (especially if no changes to the code are necessary) by making new bug fix releases with relaxed version constraints to include the new PyTorch version.\r\n\r\nIn my opinion, it is highly preferable to be conservative about dependency version constraints through a [compatible release constraint](https://www.python.org/dev/peps/pep-0440/#compatible-release) in case the dependency conforms with semantic versioning. It is impossible to guarantee compatibility with a future major release of a dependency as its API can change arbitrarily.\n", "before_files": [{"content": "import os\nimport io\nimport re\nfrom setuptools import setup, find_packages\n\n\ndef read(*names, **kwargs):\n with io.open(os.path.join(os.path.dirname(__file__), *names), encoding=kwargs.get(\"encoding\", \"utf8\")) as fp:\n return fp.read()\n\n\ndef find_version(*file_paths):\n version_file = read(*file_paths)\n version_match = re.search(r\"^__version__ = ['\\\"]([^'\\\"]*)['\\\"]\", version_file, re.M)\n if version_match:\n return version_match.group(1)\n raise RuntimeError(\"Unable to find version string.\")\n\n\nreadme = read(\"README.md\")\n\nVERSION = find_version(\"ignite\", \"__init__.py\")\n\nrequirements = [\n \"torch\",\n]\n\nsetup(\n # Metadata\n name=\"pytorch-ignite\",\n version=VERSION,\n author=\"PyTorch Core Team\",\n author_email=\"[email protected]\",\n url=\"https://github.com/pytorch/ignite\",\n description=\"A lightweight library to help with training neural networks in PyTorch.\",\n long_description_content_type=\"text/markdown\",\n long_description=readme,\n license=\"BSD\",\n # Package info\n packages=find_packages(exclude=(\"tests\", \"tests.*\",)),\n zip_safe=True,\n install_requires=requirements,\n)\n", "path": "setup.py"}]}
1,285
68
gh_patches_debug_64103
rasdani/github-patches
git_diff
strawberry-graphql__strawberry-2411
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> App breaking when using Enum as field for Generic <!-- Provide a general summary of the bug in the title above. --> When using an Enum as a field on a Generic, the app breaks, throwing a `NotImplementedError`. <!--- This template is entirely optional and can be removed, but is here to help both you and us. --> <!--- Anything on lines wrapped in comments like these will not show up in the final text. --> ## Describe the Bug The below code is an example of the error. ```python from enum import Enum from typing import Generic, Optional, TypeVar import strawberry T = TypeVar("T") @strawberry.enum class EstimatedValueEnum(Enum): test = "test" testtest = "testtest" @strawberry.type class EstimatedValue(Generic[T]): value: T type: EstimatedValueEnum @strawberry.type class Query: id_translations: Optional[EstimatedValue[int]] schema = strawberry.Schema(query=Query) ``` Are we doing something wrong and this is intended or is this a bug? <!-- A clear and concise description of what the bug is. --> ## System Information - Operating system: Docker - Strawberry version (if applicable): Since [0.149.2](https://github.com/strawberry-graphql/strawberry/blob/main/CHANGELOG.md#01492---2022-12-09) <!-- Add any other relevant information about the problem here. --> </issue> <code> [start of strawberry/enum.py] 1 import dataclasses 2 from enum import EnumMeta 3 from typing import ( 4 Any, 5 Callable, 6 Iterable, 7 List, 8 Mapping, 9 Optional, 10 TypeVar, 11 Union, 12 overload, 13 ) 14 15 from strawberry.type import StrawberryType 16 17 from .exceptions import ObjectIsNotAnEnumError 18 19 20 @dataclasses.dataclass 21 class EnumValue: 22 name: str 23 value: Any 24 deprecation_reason: Optional[str] = None 25 directives: Iterable[object] = () 26 description: Optional[str] = None 27 28 29 @dataclasses.dataclass 30 class EnumDefinition(StrawberryType): 31 wrapped_cls: EnumMeta 32 name: str 33 values: List[EnumValue] 34 description: Optional[str] 35 directives: Iterable[object] = () 36 37 def __hash__(self) -> int: 38 # TODO: Is this enough for unique-ness? 39 return hash(self.name) 40 41 def copy_with( 42 self, type_var_map: Mapping[TypeVar, Union[StrawberryType, type]] 43 ) -> Union[StrawberryType, type]: 44 return super().copy_with(type_var_map) # type: ignore[safe-super] 45 46 @property 47 def is_generic(self) -> bool: 48 return False 49 50 51 # TODO: remove duplication of EnumValueDefinition and EnumValue 52 @dataclasses.dataclass 53 class EnumValueDefinition: 54 value: Any 55 deprecation_reason: Optional[str] = None 56 directives: Iterable[object] = () 57 description: Optional[str] = None 58 59 60 def enum_value( 61 value: Any, 62 deprecation_reason: Optional[str] = None, 63 directives: Iterable[object] = (), 64 description: Optional[str] = None, 65 ) -> EnumValueDefinition: 66 return EnumValueDefinition( 67 value=value, 68 deprecation_reason=deprecation_reason, 69 directives=directives, 70 description=description, 71 ) 72 73 74 EnumType = TypeVar("EnumType", bound=EnumMeta) 75 76 77 def _process_enum( 78 cls: EnumType, 79 name: Optional[str] = None, 80 description: Optional[str] = None, 81 directives: Iterable[object] = (), 82 ) -> EnumType: 83 if not isinstance(cls, EnumMeta): 84 raise ObjectIsNotAnEnumError(cls) 85 86 if not name: 87 name = cls.__name__ 88 89 description = description 90 91 values = [] 92 for item in cls: # type: ignore 93 item_value = item.value 94 item_name = item.name 95 deprecation_reason = None 96 item_directives: Iterable[object] = () 97 enum_value_description = None 98 99 if isinstance(item_value, EnumValueDefinition): 100 item_directives = item_value.directives 101 enum_value_description = item_value.description 102 deprecation_reason = item_value.deprecation_reason 103 item_value = item_value.value 104 105 # update _value2member_map_ so that doing `MyEnum.MY_VALUE` and 106 # `MyEnum['MY_VALUE']` both work 107 cls._value2member_map_[item_value] = item 108 cls._member_map_[item_name]._value_ = item_value 109 110 value = EnumValue( 111 item_name, 112 item_value, 113 deprecation_reason=deprecation_reason, 114 directives=item_directives, 115 description=enum_value_description, 116 ) 117 values.append(value) 118 119 cls._enum_definition = EnumDefinition( # type: ignore 120 wrapped_cls=cls, 121 name=name, 122 values=values, 123 description=description, 124 directives=directives, 125 ) 126 127 return cls 128 129 130 @overload 131 def enum( 132 _cls: EnumType, 133 *, 134 name: Optional[str] = None, 135 description: Optional[str] = None, 136 directives: Iterable[object] = () 137 ) -> EnumType: 138 ... 139 140 141 @overload 142 def enum( 143 _cls: None = None, 144 *, 145 name: Optional[str] = None, 146 description: Optional[str] = None, 147 directives: Iterable[object] = () 148 ) -> Callable[[EnumType], EnumType]: 149 ... 150 151 152 def enum( 153 _cls: Optional[EnumType] = None, 154 *, 155 name: Optional[str] = None, 156 description: Optional[str] = None, 157 directives: Iterable[object] = () 158 ) -> Union[EnumType, Callable[[EnumType], EnumType]]: 159 """Registers the enum in the GraphQL type system. 160 161 If name is passed, the name of the GraphQL type will be 162 the value passed of name instead of the Enum class name. 163 """ 164 165 def wrap(cls: EnumType) -> EnumType: 166 return _process_enum(cls, name, description, directives=directives) 167 168 if not _cls: 169 return wrap 170 171 return wrap(_cls) 172 [end of strawberry/enum.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/strawberry/enum.py b/strawberry/enum.py --- a/strawberry/enum.py +++ b/strawberry/enum.py @@ -41,7 +41,8 @@ def copy_with( self, type_var_map: Mapping[TypeVar, Union[StrawberryType, type]] ) -> Union[StrawberryType, type]: - return super().copy_with(type_var_map) # type: ignore[safe-super] + # enum don't support type parameters, so we can safely return self + return self @property def is_generic(self) -> bool:
{"golden_diff": "diff --git a/strawberry/enum.py b/strawberry/enum.py\n--- a/strawberry/enum.py\n+++ b/strawberry/enum.py\n@@ -41,7 +41,8 @@\n def copy_with(\n self, type_var_map: Mapping[TypeVar, Union[StrawberryType, type]]\n ) -> Union[StrawberryType, type]:\n- return super().copy_with(type_var_map) # type: ignore[safe-super]\n+ # enum don't support type parameters, so we can safely return self\n+ return self\n \n @property\n def is_generic(self) -> bool:\n", "issue": "App breaking when using Enum as field for Generic\n<!-- Provide a general summary of the bug in the title above. -->\r\nWhen using an Enum as a field on a Generic, the app breaks, throwing a `NotImplementedError`.\r\n\r\n<!--- This template is entirely optional and can be removed, but is here to help both you and us. -->\r\n<!--- Anything on lines wrapped in comments like these will not show up in the final text. -->\r\n\r\n## Describe the Bug\r\nThe below code is an example of the error.\r\n```python\r\nfrom enum import Enum\r\nfrom typing import Generic, Optional, TypeVar\r\n\r\nimport strawberry\r\n\r\nT = TypeVar(\"T\")\r\n\r\n\r\[email protected]\r\nclass EstimatedValueEnum(Enum):\r\n test = \"test\"\r\n testtest = \"testtest\"\r\n\r\n\r\[email protected]\r\nclass EstimatedValue(Generic[T]):\r\n value: T\r\n type: EstimatedValueEnum\r\n\r\n\r\[email protected]\r\nclass Query:\r\n id_translations: Optional[EstimatedValue[int]]\r\n\r\n\r\nschema = strawberry.Schema(query=Query)\r\n```\r\nAre we doing something wrong and this is intended or is this a bug?\r\n<!-- A clear and concise description of what the bug is. -->\r\n\r\n## System Information\r\n\r\n - Operating system: Docker\r\n - Strawberry version (if applicable): Since [0.149.2](https://github.com/strawberry-graphql/strawberry/blob/main/CHANGELOG.md#01492---2022-12-09)\r\n\r\n<!-- Add any other relevant information about the problem here. -->\r\n\n", "before_files": [{"content": "import dataclasses\nfrom enum import EnumMeta\nfrom typing import (\n Any,\n Callable,\n Iterable,\n List,\n Mapping,\n Optional,\n TypeVar,\n Union,\n overload,\n)\n\nfrom strawberry.type import StrawberryType\n\nfrom .exceptions import ObjectIsNotAnEnumError\n\n\[email protected]\nclass EnumValue:\n name: str\n value: Any\n deprecation_reason: Optional[str] = None\n directives: Iterable[object] = ()\n description: Optional[str] = None\n\n\[email protected]\nclass EnumDefinition(StrawberryType):\n wrapped_cls: EnumMeta\n name: str\n values: List[EnumValue]\n description: Optional[str]\n directives: Iterable[object] = ()\n\n def __hash__(self) -> int:\n # TODO: Is this enough for unique-ness?\n return hash(self.name)\n\n def copy_with(\n self, type_var_map: Mapping[TypeVar, Union[StrawberryType, type]]\n ) -> Union[StrawberryType, type]:\n return super().copy_with(type_var_map) # type: ignore[safe-super]\n\n @property\n def is_generic(self) -> bool:\n return False\n\n\n# TODO: remove duplication of EnumValueDefinition and EnumValue\[email protected]\nclass EnumValueDefinition:\n value: Any\n deprecation_reason: Optional[str] = None\n directives: Iterable[object] = ()\n description: Optional[str] = None\n\n\ndef enum_value(\n value: Any,\n deprecation_reason: Optional[str] = None,\n directives: Iterable[object] = (),\n description: Optional[str] = None,\n) -> EnumValueDefinition:\n return EnumValueDefinition(\n value=value,\n deprecation_reason=deprecation_reason,\n directives=directives,\n description=description,\n )\n\n\nEnumType = TypeVar(\"EnumType\", bound=EnumMeta)\n\n\ndef _process_enum(\n cls: EnumType,\n name: Optional[str] = None,\n description: Optional[str] = None,\n directives: Iterable[object] = (),\n) -> EnumType:\n if not isinstance(cls, EnumMeta):\n raise ObjectIsNotAnEnumError(cls)\n\n if not name:\n name = cls.__name__\n\n description = description\n\n values = []\n for item in cls: # type: ignore\n item_value = item.value\n item_name = item.name\n deprecation_reason = None\n item_directives: Iterable[object] = ()\n enum_value_description = None\n\n if isinstance(item_value, EnumValueDefinition):\n item_directives = item_value.directives\n enum_value_description = item_value.description\n deprecation_reason = item_value.deprecation_reason\n item_value = item_value.value\n\n # update _value2member_map_ so that doing `MyEnum.MY_VALUE` and\n # `MyEnum['MY_VALUE']` both work\n cls._value2member_map_[item_value] = item\n cls._member_map_[item_name]._value_ = item_value\n\n value = EnumValue(\n item_name,\n item_value,\n deprecation_reason=deprecation_reason,\n directives=item_directives,\n description=enum_value_description,\n )\n values.append(value)\n\n cls._enum_definition = EnumDefinition( # type: ignore\n wrapped_cls=cls,\n name=name,\n values=values,\n description=description,\n directives=directives,\n )\n\n return cls\n\n\n@overload\ndef enum(\n _cls: EnumType,\n *,\n name: Optional[str] = None,\n description: Optional[str] = None,\n directives: Iterable[object] = ()\n) -> EnumType:\n ...\n\n\n@overload\ndef enum(\n _cls: None = None,\n *,\n name: Optional[str] = None,\n description: Optional[str] = None,\n directives: Iterable[object] = ()\n) -> Callable[[EnumType], EnumType]:\n ...\n\n\ndef enum(\n _cls: Optional[EnumType] = None,\n *,\n name: Optional[str] = None,\n description: Optional[str] = None,\n directives: Iterable[object] = ()\n) -> Union[EnumType, Callable[[EnumType], EnumType]]:\n \"\"\"Registers the enum in the GraphQL type system.\n\n If name is passed, the name of the GraphQL type will be\n the value passed of name instead of the Enum class name.\n \"\"\"\n\n def wrap(cls: EnumType) -> EnumType:\n return _process_enum(cls, name, description, directives=directives)\n\n if not _cls:\n return wrap\n\n return wrap(_cls)\n", "path": "strawberry/enum.py"}]}
2,294
146
gh_patches_debug_26348
rasdani/github-patches
git_diff
wagtail__wagtail-4805
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> When changing your password from with admin a user is presented with an edit button. I've just reset my password from within admin and I get this horrible large edit button. <img width="1267" alt="screen shot 2018-01-16 at 12 33 46" src="https://user-images.githubusercontent.com/7290588/34989276-e10b3338-fab9-11e7-9e69-05d1c7846af4.png"> Can we remove the edit button and amend the message to say 'Your details have been updated. You've been logged out for security reasons, please login to continue.' or something like this. When changing your password from with admin a user is presented with an edit button. I've just reset my password from within admin and I get this horrible large edit button. <img width="1267" alt="screen shot 2018-01-16 at 12 33 46" src="https://user-images.githubusercontent.com/7290588/34989276-e10b3338-fab9-11e7-9e69-05d1c7846af4.png"> Can we remove the edit button and amend the message to say 'Your details have been updated. You've been logged out for security reasons, please login to continue.' or something like this. </issue> <code> [start of wagtail/users/views/users.py] 1 from django.conf import settings 2 from django.contrib.auth import get_user_model 3 from django.db.models import Q 4 from django.shortcuts import get_object_or_404, redirect, render 5 from django.urls import reverse 6 from django.utils.translation import ugettext as _ 7 from django.views.decorators.vary import vary_on_headers 8 9 from wagtail.admin import messages 10 from wagtail.admin.forms.search import SearchForm 11 from wagtail.admin.utils import any_permission_required, permission_denied, permission_required 12 from wagtail.core import hooks 13 from wagtail.core.compat import AUTH_USER_APP_LABEL, AUTH_USER_MODEL_NAME 14 from wagtail.users.forms import UserCreationForm, UserEditForm 15 from wagtail.users.utils import user_can_delete_user 16 from wagtail.utils.loading import get_custom_form 17 from wagtail.utils.pagination import paginate 18 19 User = get_user_model() 20 21 # Typically we would check the permission 'auth.change_user' (and 'auth.add_user' / 22 # 'auth.delete_user') for user management actions, but this may vary according to 23 # the AUTH_USER_MODEL setting 24 add_user_perm = "{0}.add_{1}".format(AUTH_USER_APP_LABEL, AUTH_USER_MODEL_NAME.lower()) 25 change_user_perm = "{0}.change_{1}".format(AUTH_USER_APP_LABEL, AUTH_USER_MODEL_NAME.lower()) 26 delete_user_perm = "{0}.delete_{1}".format(AUTH_USER_APP_LABEL, AUTH_USER_MODEL_NAME.lower()) 27 28 29 def get_user_creation_form(): 30 form_setting = 'WAGTAIL_USER_CREATION_FORM' 31 if hasattr(settings, form_setting): 32 return get_custom_form(form_setting) 33 else: 34 return UserCreationForm 35 36 37 def get_user_edit_form(): 38 form_setting = 'WAGTAIL_USER_EDIT_FORM' 39 if hasattr(settings, form_setting): 40 return get_custom_form(form_setting) 41 else: 42 return UserEditForm 43 44 45 @any_permission_required(add_user_perm, change_user_perm, delete_user_perm) 46 @vary_on_headers('X-Requested-With') 47 def index(request): 48 q = None 49 is_searching = False 50 51 model_fields = [f.name for f in User._meta.get_fields()] 52 53 if 'q' in request.GET: 54 form = SearchForm(request.GET, placeholder=_("Search users")) 55 if form.is_valid(): 56 q = form.cleaned_data['q'] 57 is_searching = True 58 conditions = Q() 59 60 for term in q.split(): 61 if 'username' in model_fields: 62 conditions |= Q(username__icontains=term) 63 64 if 'first_name' in model_fields: 65 conditions |= Q(first_name__icontains=term) 66 67 if 'last_name' in model_fields: 68 conditions |= Q(last_name__icontains=term) 69 70 if 'email' in model_fields: 71 conditions |= Q(email__icontains=term) 72 73 users = User.objects.filter(conditions) 74 else: 75 form = SearchForm(placeholder=_("Search users")) 76 77 if not is_searching: 78 users = User.objects.all() 79 80 if 'last_name' in model_fields and 'first_name' in model_fields: 81 users = users.order_by('last_name', 'first_name') 82 83 if 'ordering' in request.GET: 84 ordering = request.GET['ordering'] 85 86 if ordering == 'username': 87 users = users.order_by(User.USERNAME_FIELD) 88 else: 89 ordering = 'name' 90 91 paginator, users = paginate(request, users) 92 93 if request.is_ajax(): 94 return render(request, "wagtailusers/users/results.html", { 95 'users': users, 96 'is_searching': is_searching, 97 'query_string': q, 98 'ordering': ordering, 99 }) 100 else: 101 return render(request, "wagtailusers/users/index.html", { 102 'search_form': form, 103 'users': users, 104 'is_searching': is_searching, 105 'ordering': ordering, 106 'query_string': q, 107 }) 108 109 110 @permission_required(add_user_perm) 111 def create(request): 112 for fn in hooks.get_hooks('before_create_user'): 113 result = fn(request) 114 if hasattr(result, 'status_code'): 115 return result 116 if request.method == 'POST': 117 form = get_user_creation_form()(request.POST, request.FILES) 118 if form.is_valid(): 119 user = form.save() 120 messages.success(request, _("User '{0}' created.").format(user), buttons=[ 121 messages.button(reverse('wagtailusers_users:edit', args=(user.pk,)), _('Edit')) 122 ]) 123 for fn in hooks.get_hooks('after_create_user'): 124 result = fn(request, user) 125 if hasattr(result, 'status_code'): 126 return result 127 return redirect('wagtailusers_users:index') 128 else: 129 messages.error(request, _("The user could not be created due to errors.")) 130 else: 131 form = get_user_creation_form()() 132 133 return render(request, 'wagtailusers/users/create.html', { 134 'form': form, 135 }) 136 137 138 @permission_required(change_user_perm) 139 def edit(request, user_id): 140 user = get_object_or_404(User, pk=user_id) 141 can_delete = user_can_delete_user(request.user, user) 142 editing_self = request.user == user 143 144 for fn in hooks.get_hooks('before_edit_user'): 145 result = fn(request, user) 146 if hasattr(result, 'status_code'): 147 return result 148 if request.method == 'POST': 149 form = get_user_edit_form()(request.POST, request.FILES, instance=user, editing_self=editing_self) 150 if form.is_valid(): 151 user = form.save() 152 messages.success(request, _("Your details have been updated. You've been logged out for security reasons, " 153 "please login to continue.")) 154 for fn in hooks.get_hooks('after_edit_user'): 155 result = fn(request, user) 156 if hasattr(result, 'status_code'): 157 return result 158 return redirect('wagtailusers_users:index') 159 else: 160 messages.error(request, _("The user could not be saved due to errors.")) 161 else: 162 form = get_user_edit_form()(instance=user, editing_self=editing_self) 163 164 return render(request, 'wagtailusers/users/edit.html', { 165 'user': user, 166 'form': form, 167 'can_delete': can_delete, 168 }) 169 170 171 @permission_required(delete_user_perm) 172 def delete(request, user_id): 173 user = get_object_or_404(User, pk=user_id) 174 175 if not user_can_delete_user(request.user, user): 176 return permission_denied(request) 177 178 for fn in hooks.get_hooks('before_delete_user'): 179 result = fn(request, user) 180 if hasattr(result, 'status_code'): 181 return result 182 if request.method == 'POST': 183 user.delete() 184 messages.success(request, _("User '{0}' deleted.").format(user)) 185 for fn in hooks.get_hooks('after_delete_user'): 186 result = fn(request, user) 187 if hasattr(result, 'status_code'): 188 return result 189 return redirect('wagtailusers_users:index') 190 191 return render(request, "wagtailusers/users/confirm_delete.html", { 192 'user': user, 193 }) 194 [end of wagtail/users/views/users.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/wagtail/users/views/users.py b/wagtail/users/views/users.py --- a/wagtail/users/views/users.py +++ b/wagtail/users/views/users.py @@ -1,5 +1,5 @@ from django.conf import settings -from django.contrib.auth import get_user_model +from django.contrib.auth import get_user_model, update_session_auth_hash from django.db.models import Q from django.shortcuts import get_object_or_404, redirect, render from django.urls import reverse @@ -149,8 +149,14 @@ form = get_user_edit_form()(request.POST, request.FILES, instance=user, editing_self=editing_self) if form.is_valid(): user = form.save() - messages.success(request, _("Your details have been updated. You've been logged out for security reasons, " - "please login to continue.")) + + if user == request.user and 'password1' in form.changed_data: + # User is changing their own password; need to update their session hash + update_session_auth_hash(request, user) + + messages.success(request, _("User '{0}' updated.").format(user), buttons=[ + messages.button(reverse('wagtailusers_users:edit', args=(user.pk,)), _('Edit')) + ]) for fn in hooks.get_hooks('after_edit_user'): result = fn(request, user) if hasattr(result, 'status_code'):
{"golden_diff": "diff --git a/wagtail/users/views/users.py b/wagtail/users/views/users.py\n--- a/wagtail/users/views/users.py\n+++ b/wagtail/users/views/users.py\n@@ -1,5 +1,5 @@\n from django.conf import settings\n-from django.contrib.auth import get_user_model\n+from django.contrib.auth import get_user_model, update_session_auth_hash\n from django.db.models import Q\n from django.shortcuts import get_object_or_404, redirect, render\n from django.urls import reverse\n@@ -149,8 +149,14 @@\n form = get_user_edit_form()(request.POST, request.FILES, instance=user, editing_self=editing_self)\n if form.is_valid():\n user = form.save()\n- messages.success(request, _(\"Your details have been updated. You've been logged out for security reasons, \"\n- \"please login to continue.\"))\n+\n+ if user == request.user and 'password1' in form.changed_data:\n+ # User is changing their own password; need to update their session hash\n+ update_session_auth_hash(request, user)\n+\n+ messages.success(request, _(\"User '{0}' updated.\").format(user), buttons=[\n+ messages.button(reverse('wagtailusers_users:edit', args=(user.pk,)), _('Edit'))\n+ ])\n for fn in hooks.get_hooks('after_edit_user'):\n result = fn(request, user)\n if hasattr(result, 'status_code'):\n", "issue": "When changing your password from with admin a user is presented with an edit button. \nI've just reset my password from within admin and I get this horrible large edit button. \r\n<img width=\"1267\" alt=\"screen shot 2018-01-16 at 12 33 46\" src=\"https://user-images.githubusercontent.com/7290588/34989276-e10b3338-fab9-11e7-9e69-05d1c7846af4.png\">\r\n\r\nCan we remove the edit button and amend the message to say 'Your details have been updated. You've been logged out for security reasons, please login to continue.' or something like this. \nWhen changing your password from with admin a user is presented with an edit button. \nI've just reset my password from within admin and I get this horrible large edit button. \r\n<img width=\"1267\" alt=\"screen shot 2018-01-16 at 12 33 46\" src=\"https://user-images.githubusercontent.com/7290588/34989276-e10b3338-fab9-11e7-9e69-05d1c7846af4.png\">\r\n\r\nCan we remove the edit button and amend the message to say 'Your details have been updated. You've been logged out for security reasons, please login to continue.' or something like this. \n", "before_files": [{"content": "from django.conf import settings\nfrom django.contrib.auth import get_user_model\nfrom django.db.models import Q\nfrom django.shortcuts import get_object_or_404, redirect, render\nfrom django.urls import reverse\nfrom django.utils.translation import ugettext as _\nfrom django.views.decorators.vary import vary_on_headers\n\nfrom wagtail.admin import messages\nfrom wagtail.admin.forms.search import SearchForm\nfrom wagtail.admin.utils import any_permission_required, permission_denied, permission_required\nfrom wagtail.core import hooks\nfrom wagtail.core.compat import AUTH_USER_APP_LABEL, AUTH_USER_MODEL_NAME\nfrom wagtail.users.forms import UserCreationForm, UserEditForm\nfrom wagtail.users.utils import user_can_delete_user\nfrom wagtail.utils.loading import get_custom_form\nfrom wagtail.utils.pagination import paginate\n\nUser = get_user_model()\n\n# Typically we would check the permission 'auth.change_user' (and 'auth.add_user' /\n# 'auth.delete_user') for user management actions, but this may vary according to\n# the AUTH_USER_MODEL setting\nadd_user_perm = \"{0}.add_{1}\".format(AUTH_USER_APP_LABEL, AUTH_USER_MODEL_NAME.lower())\nchange_user_perm = \"{0}.change_{1}\".format(AUTH_USER_APP_LABEL, AUTH_USER_MODEL_NAME.lower())\ndelete_user_perm = \"{0}.delete_{1}\".format(AUTH_USER_APP_LABEL, AUTH_USER_MODEL_NAME.lower())\n\n\ndef get_user_creation_form():\n form_setting = 'WAGTAIL_USER_CREATION_FORM'\n if hasattr(settings, form_setting):\n return get_custom_form(form_setting)\n else:\n return UserCreationForm\n\n\ndef get_user_edit_form():\n form_setting = 'WAGTAIL_USER_EDIT_FORM'\n if hasattr(settings, form_setting):\n return get_custom_form(form_setting)\n else:\n return UserEditForm\n\n\n@any_permission_required(add_user_perm, change_user_perm, delete_user_perm)\n@vary_on_headers('X-Requested-With')\ndef index(request):\n q = None\n is_searching = False\n\n model_fields = [f.name for f in User._meta.get_fields()]\n\n if 'q' in request.GET:\n form = SearchForm(request.GET, placeholder=_(\"Search users\"))\n if form.is_valid():\n q = form.cleaned_data['q']\n is_searching = True\n conditions = Q()\n\n for term in q.split():\n if 'username' in model_fields:\n conditions |= Q(username__icontains=term)\n\n if 'first_name' in model_fields:\n conditions |= Q(first_name__icontains=term)\n\n if 'last_name' in model_fields:\n conditions |= Q(last_name__icontains=term)\n\n if 'email' in model_fields:\n conditions |= Q(email__icontains=term)\n\n users = User.objects.filter(conditions)\n else:\n form = SearchForm(placeholder=_(\"Search users\"))\n\n if not is_searching:\n users = User.objects.all()\n\n if 'last_name' in model_fields and 'first_name' in model_fields:\n users = users.order_by('last_name', 'first_name')\n\n if 'ordering' in request.GET:\n ordering = request.GET['ordering']\n\n if ordering == 'username':\n users = users.order_by(User.USERNAME_FIELD)\n else:\n ordering = 'name'\n\n paginator, users = paginate(request, users)\n\n if request.is_ajax():\n return render(request, \"wagtailusers/users/results.html\", {\n 'users': users,\n 'is_searching': is_searching,\n 'query_string': q,\n 'ordering': ordering,\n })\n else:\n return render(request, \"wagtailusers/users/index.html\", {\n 'search_form': form,\n 'users': users,\n 'is_searching': is_searching,\n 'ordering': ordering,\n 'query_string': q,\n })\n\n\n@permission_required(add_user_perm)\ndef create(request):\n for fn in hooks.get_hooks('before_create_user'):\n result = fn(request)\n if hasattr(result, 'status_code'):\n return result\n if request.method == 'POST':\n form = get_user_creation_form()(request.POST, request.FILES)\n if form.is_valid():\n user = form.save()\n messages.success(request, _(\"User '{0}' created.\").format(user), buttons=[\n messages.button(reverse('wagtailusers_users:edit', args=(user.pk,)), _('Edit'))\n ])\n for fn in hooks.get_hooks('after_create_user'):\n result = fn(request, user)\n if hasattr(result, 'status_code'):\n return result\n return redirect('wagtailusers_users:index')\n else:\n messages.error(request, _(\"The user could not be created due to errors.\"))\n else:\n form = get_user_creation_form()()\n\n return render(request, 'wagtailusers/users/create.html', {\n 'form': form,\n })\n\n\n@permission_required(change_user_perm)\ndef edit(request, user_id):\n user = get_object_or_404(User, pk=user_id)\n can_delete = user_can_delete_user(request.user, user)\n editing_self = request.user == user\n\n for fn in hooks.get_hooks('before_edit_user'):\n result = fn(request, user)\n if hasattr(result, 'status_code'):\n return result\n if request.method == 'POST':\n form = get_user_edit_form()(request.POST, request.FILES, instance=user, editing_self=editing_self)\n if form.is_valid():\n user = form.save()\n messages.success(request, _(\"Your details have been updated. You've been logged out for security reasons, \"\n \"please login to continue.\"))\n for fn in hooks.get_hooks('after_edit_user'):\n result = fn(request, user)\n if hasattr(result, 'status_code'):\n return result\n return redirect('wagtailusers_users:index')\n else:\n messages.error(request, _(\"The user could not be saved due to errors.\"))\n else:\n form = get_user_edit_form()(instance=user, editing_self=editing_self)\n\n return render(request, 'wagtailusers/users/edit.html', {\n 'user': user,\n 'form': form,\n 'can_delete': can_delete,\n })\n\n\n@permission_required(delete_user_perm)\ndef delete(request, user_id):\n user = get_object_or_404(User, pk=user_id)\n\n if not user_can_delete_user(request.user, user):\n return permission_denied(request)\n\n for fn in hooks.get_hooks('before_delete_user'):\n result = fn(request, user)\n if hasattr(result, 'status_code'):\n return result\n if request.method == 'POST':\n user.delete()\n messages.success(request, _(\"User '{0}' deleted.\").format(user))\n for fn in hooks.get_hooks('after_delete_user'):\n result = fn(request, user)\n if hasattr(result, 'status_code'):\n return result\n return redirect('wagtailusers_users:index')\n\n return render(request, \"wagtailusers/users/confirm_delete.html\", {\n 'user': user,\n })\n", "path": "wagtail/users/views/users.py"}]}
2,854
310
gh_patches_debug_9098
rasdani/github-patches
git_diff
mindsdb__mindsdb-1590
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> Add option to list tables in PostgreSQL integration :bookmark_tabs: When users create a connection to the database it will be useful to show them tips with a list of tables. To be able to do this we need a new method `get_tables_list` implemented in the PostgreSQL integration class. ## Steps :male_detective: :female_detective: - Frok MindsDB repo - Add new implementation in https://github.com/mindsdb/mindsdb/blob/stable/mindsdb/integrations/postgres/postgres.py#L37 - Make a PR to staging branch ## Additional rewards :1st_place_medal: Each code PR brings :three: point for entry into the draw for a :computer: Deep Learning Laptop powered by the NVIDIA RTX 3080 Max-Q GPU or other swag :shirt: :bear: . For more info check out https://mindsdb.com/hacktoberfest/ </issue> <code> [start of mindsdb/integrations/postgres/postgres.py] 1 from contextlib import closing 2 import pg8000 3 4 from lightwood.api import dtype 5 from mindsdb.integrations.base import Integration 6 from mindsdb.utilities.log import log 7 8 9 class PostgreSQLConnectionChecker: 10 def __init__(self, **kwargs): 11 self.host = kwargs.get('host') 12 self.port = kwargs.get('port') 13 self.user = kwargs.get('user') 14 self.password = kwargs.get('password') 15 self.database = kwargs.get('database', 'postgres') 16 17 def _get_connection(self): 18 return pg8000.connect( 19 database=self.database, 20 user=self.user, 21 password=self.password, 22 host=self.host, 23 port=self.port 24 ) 25 26 def check_connection(self): 27 try: 28 con = self._get_connection() 29 with closing(con) as con: 30 con.run('select 1;') 31 connected = True 32 except Exception: 33 connected = False 34 return connected 35 36 37 class PostgreSQL(Integration, PostgreSQLConnectionChecker): 38 def __init__(self, config, name, db_info): 39 super().__init__(config, name) 40 self.user = db_info.get('user') 41 self.password = db_info.get('password') 42 self.host = db_info.get('host') 43 self.port = db_info.get('port') 44 self.database = db_info.get('database', 'postgres') 45 46 def _to_postgres_table(self, dtype_dict, predicted_cols, columns): 47 subtype_map = { 48 dtype.integer: ' int8', 49 dtype.float: 'float8', 50 dtype.binary: 'bool', 51 dtype.date: 'date', 52 dtype.datetime: 'timestamp', 53 dtype.binary: 'text', 54 dtype.categorical: 'text', 55 dtype.tags: 'text', 56 dtype.image: 'text', 57 dtype.video: 'text', 58 dtype.audio: 'text', 59 dtype.short_text: 'text', 60 dtype.rich_text: 'text', 61 dtype.array: 'text' 62 } 63 64 column_declaration = [] 65 for name in columns: 66 try: 67 col_subtype = dtype_dict[name] 68 new_type = subtype_map[col_subtype] 69 column_declaration.append(f' "{name}" {new_type} ') 70 if name in predicted_cols: 71 column_declaration.append(f' "{name}_original" {new_type} ') 72 except Exception as e: 73 log.error(f'Error: can not determine postgres data type for column {name}: {e}') 74 75 return column_declaration 76 77 def _escape_table_name(self, name): 78 return '"' + name.replace('"', '""') + '"' 79 80 def _query(self, query): 81 con = self._get_connection() 82 with closing(con) as con: 83 84 cur = con.cursor() 85 res = True 86 cur.execute(query) 87 88 try: 89 rows = cur.fetchall() 90 keys = [k[0] if isinstance(k[0], str) else k[0].decode('ascii') for k in cur.description] 91 res = [dict(zip(keys, row)) for row in rows] 92 except Exception: 93 pass 94 95 con.commit() 96 97 return res 98 99 def setup(self): 100 user = f"{self.config['api']['mysql']['user']}_{self.name}" 101 password = self.config['api']['mysql']['password'] 102 host = self.config['api']['mysql']['host'] 103 port = self.config['api']['mysql']['port'] 104 105 try: 106 self._query(''' 107 DO $$ 108 begin 109 if not exists (SELECT 1 FROM pg_extension where extname = 'mysql_fdw') then 110 CREATE EXTENSION mysql_fdw; 111 end if; 112 END 113 $$; 114 ''') 115 except Exception: 116 print('Error: cant find or activate mysql_fdw extension for PostgreSQL.') 117 118 self._query(f'DROP SCHEMA IF EXISTS {self.mindsdb_database} CASCADE') 119 120 self._query(f"DROP USER MAPPING IF EXISTS FOR {self.user} SERVER server_{self.mindsdb_database}") 121 122 self._query(f'DROP SERVER IF EXISTS server_{self.mindsdb_database} CASCADE') 123 124 self._query(f''' 125 CREATE SERVER server_{self.mindsdb_database} 126 FOREIGN DATA WRAPPER mysql_fdw 127 OPTIONS (host '{host}', port '{port}'); 128 ''') 129 130 self._query(f''' 131 CREATE USER MAPPING FOR {self.user} 132 SERVER server_{self.mindsdb_database} 133 OPTIONS (username '{user}', password '{password}'); 134 ''') 135 136 self._query(f'CREATE SCHEMA {self.mindsdb_database}') 137 138 q = f""" 139 CREATE FOREIGN TABLE IF NOT EXISTS {self.mindsdb_database}.predictors ( 140 name text, 141 status text, 142 accuracy text, 143 predict text, 144 select_data_query text, 145 external_datasource text, 146 training_options text 147 ) 148 SERVER server_{self.mindsdb_database} 149 OPTIONS (dbname 'mindsdb', table_name 'predictors'); 150 """ 151 self._query(q) 152 153 q = f""" 154 CREATE FOREIGN TABLE IF NOT EXISTS {self.mindsdb_database}.commands ( 155 command text 156 ) SERVER server_{self.mindsdb_database} 157 OPTIONS (dbname 'mindsdb', table_name 'commands'); 158 """ 159 self._query(q) 160 161 def register_predictors(self, model_data_arr): 162 for model_meta in model_data_arr: 163 name = model_meta['name'] 164 predict = model_meta['predict'] 165 if not isinstance(predict, list): 166 predict = [predict] 167 columns_sql = ','.join(self._to_postgres_table( 168 model_meta['dtype_dict'], 169 predict, 170 list(model_meta['dtype_dict'].keys()) 171 )) 172 columns_sql += ',"select_data_query" text' 173 columns_sql += ',"external_datasource" text' 174 for col in predict: 175 columns_sql += f',"{col}_confidence" float8' 176 if model_meta['dtype_dict'][col] in (dtype.integer, dtype.float): 177 columns_sql += f',"{col}_min" float8' 178 columns_sql += f',"{col}_max" float8' 179 columns_sql += f',"{col}_explain" text' 180 181 self.unregister_predictor(name) 182 q = f""" 183 CREATE FOREIGN TABLE {self.mindsdb_database}.{self._escape_table_name(name)} ( 184 {columns_sql} 185 ) SERVER server_{self.mindsdb_database} 186 OPTIONS (dbname 'mindsdb', table_name '{name}'); 187 """ 188 self._query(q) 189 190 def unregister_predictor(self, name): 191 q = f""" 192 DROP FOREIGN TABLE IF EXISTS {self.mindsdb_database}.{self._escape_table_name(name)}; 193 """ 194 self._query(q) 195 196 def get_row_count(self, query): 197 q = f""" 198 SELECT COUNT(*) as count 199 FROM ({query}) as query;""" 200 result = self._query(q) 201 return result[0]['count'] 202 [end of mindsdb/integrations/postgres/postgres.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/mindsdb/integrations/postgres/postgres.py b/mindsdb/integrations/postgres/postgres.py --- a/mindsdb/integrations/postgres/postgres.py +++ b/mindsdb/integrations/postgres/postgres.py @@ -199,3 +199,13 @@ FROM ({query}) as query;""" result = self._query(q) return result[0]['count'] + + def get_tables_list(self): + q = f""" SELECT table_schema, table_name + FROM information_schema.tables + WHERE table_schema != 'pg_catalog' + AND table_schema != 'information_schema' + ORDER BY table_schema, table_name""" + tables_list = self._query(q) + tables= [f"{table['table_schema']}.{table['table_name']}" for table in tables_list] + return tables \ No newline at end of file
{"golden_diff": "diff --git a/mindsdb/integrations/postgres/postgres.py b/mindsdb/integrations/postgres/postgres.py\n--- a/mindsdb/integrations/postgres/postgres.py\n+++ b/mindsdb/integrations/postgres/postgres.py\n@@ -199,3 +199,13 @@\n FROM ({query}) as query;\"\"\"\n result = self._query(q)\n return result[0]['count']\n+ \n+ def get_tables_list(self):\n+ q = f\"\"\" SELECT table_schema, table_name\n+ FROM information_schema.tables\n+ WHERE table_schema != 'pg_catalog'\n+ AND table_schema != 'information_schema'\n+ ORDER BY table_schema, table_name\"\"\"\n+ tables_list = self._query(q)\n+ tables= [f\"{table['table_schema']}.{table['table_name']}\" for table in tables_list]\n+ return tables\n\\ No newline at end of file\n", "issue": "Add option to list tables in PostgreSQL integration :bookmark_tabs: \nWhen users create a connection to the database it will be useful to show them tips with a list of tables. To be able to do this we need a new method `get_tables_list` implemented in the PostgreSQL integration class.\r\n\r\n## Steps :male_detective: :female_detective: \r\n\r\n- Frok MindsDB repo\r\n- Add new implementation in https://github.com/mindsdb/mindsdb/blob/stable/mindsdb/integrations/postgres/postgres.py#L37\r\n- Make a PR to staging branch\r\n\r\n## Additional rewards :1st_place_medal: \r\n\r\nEach code PR brings :three: point for entry into the draw for a :computer: Deep Learning Laptop powered by the NVIDIA RTX 3080 Max-Q GPU or other swag :shirt: :bear: . For more info check out https://mindsdb.com/hacktoberfest/\n", "before_files": [{"content": "from contextlib import closing\nimport pg8000\n\nfrom lightwood.api import dtype\nfrom mindsdb.integrations.base import Integration\nfrom mindsdb.utilities.log import log\n\n\nclass PostgreSQLConnectionChecker:\n def __init__(self, **kwargs):\n self.host = kwargs.get('host')\n self.port = kwargs.get('port')\n self.user = kwargs.get('user')\n self.password = kwargs.get('password')\n self.database = kwargs.get('database', 'postgres')\n\n def _get_connection(self):\n return pg8000.connect(\n database=self.database,\n user=self.user,\n password=self.password,\n host=self.host,\n port=self.port\n )\n\n def check_connection(self):\n try:\n con = self._get_connection()\n with closing(con) as con:\n con.run('select 1;')\n connected = True\n except Exception:\n connected = False\n return connected\n\n\nclass PostgreSQL(Integration, PostgreSQLConnectionChecker):\n def __init__(self, config, name, db_info):\n super().__init__(config, name)\n self.user = db_info.get('user')\n self.password = db_info.get('password')\n self.host = db_info.get('host')\n self.port = db_info.get('port')\n self.database = db_info.get('database', 'postgres')\n\n def _to_postgres_table(self, dtype_dict, predicted_cols, columns):\n subtype_map = {\n dtype.integer: ' int8',\n dtype.float: 'float8',\n dtype.binary: 'bool',\n dtype.date: 'date',\n dtype.datetime: 'timestamp',\n dtype.binary: 'text',\n dtype.categorical: 'text',\n dtype.tags: 'text',\n dtype.image: 'text',\n dtype.video: 'text',\n dtype.audio: 'text',\n dtype.short_text: 'text',\n dtype.rich_text: 'text',\n dtype.array: 'text'\n }\n\n column_declaration = []\n for name in columns:\n try:\n col_subtype = dtype_dict[name]\n new_type = subtype_map[col_subtype]\n column_declaration.append(f' \"{name}\" {new_type} ')\n if name in predicted_cols:\n column_declaration.append(f' \"{name}_original\" {new_type} ')\n except Exception as e:\n log.error(f'Error: can not determine postgres data type for column {name}: {e}')\n\n return column_declaration\n\n def _escape_table_name(self, name):\n return '\"' + name.replace('\"', '\"\"') + '\"'\n\n def _query(self, query):\n con = self._get_connection()\n with closing(con) as con:\n\n cur = con.cursor()\n res = True\n cur.execute(query)\n\n try:\n rows = cur.fetchall()\n keys = [k[0] if isinstance(k[0], str) else k[0].decode('ascii') for k in cur.description]\n res = [dict(zip(keys, row)) for row in rows]\n except Exception:\n pass\n\n con.commit()\n\n return res\n\n def setup(self):\n user = f\"{self.config['api']['mysql']['user']}_{self.name}\"\n password = self.config['api']['mysql']['password']\n host = self.config['api']['mysql']['host']\n port = self.config['api']['mysql']['port']\n\n try:\n self._query('''\n DO $$\n begin\n if not exists (SELECT 1 FROM pg_extension where extname = 'mysql_fdw') then\n CREATE EXTENSION mysql_fdw;\n end if;\n END\n $$;\n ''')\n except Exception:\n print('Error: cant find or activate mysql_fdw extension for PostgreSQL.')\n\n self._query(f'DROP SCHEMA IF EXISTS {self.mindsdb_database} CASCADE')\n\n self._query(f\"DROP USER MAPPING IF EXISTS FOR {self.user} SERVER server_{self.mindsdb_database}\")\n\n self._query(f'DROP SERVER IF EXISTS server_{self.mindsdb_database} CASCADE')\n\n self._query(f'''\n CREATE SERVER server_{self.mindsdb_database}\n FOREIGN DATA WRAPPER mysql_fdw\n OPTIONS (host '{host}', port '{port}');\n ''')\n\n self._query(f'''\n CREATE USER MAPPING FOR {self.user}\n SERVER server_{self.mindsdb_database}\n OPTIONS (username '{user}', password '{password}');\n ''')\n\n self._query(f'CREATE SCHEMA {self.mindsdb_database}')\n\n q = f\"\"\"\n CREATE FOREIGN TABLE IF NOT EXISTS {self.mindsdb_database}.predictors (\n name text,\n status text,\n accuracy text,\n predict text,\n select_data_query text,\n external_datasource text,\n training_options text\n )\n SERVER server_{self.mindsdb_database}\n OPTIONS (dbname 'mindsdb', table_name 'predictors');\n \"\"\"\n self._query(q)\n\n q = f\"\"\"\n CREATE FOREIGN TABLE IF NOT EXISTS {self.mindsdb_database}.commands (\n command text\n ) SERVER server_{self.mindsdb_database}\n OPTIONS (dbname 'mindsdb', table_name 'commands');\n \"\"\"\n self._query(q)\n\n def register_predictors(self, model_data_arr):\n for model_meta in model_data_arr:\n name = model_meta['name']\n predict = model_meta['predict']\n if not isinstance(predict, list):\n predict = [predict]\n columns_sql = ','.join(self._to_postgres_table(\n model_meta['dtype_dict'],\n predict,\n list(model_meta['dtype_dict'].keys())\n ))\n columns_sql += ',\"select_data_query\" text'\n columns_sql += ',\"external_datasource\" text'\n for col in predict:\n columns_sql += f',\"{col}_confidence\" float8'\n if model_meta['dtype_dict'][col] in (dtype.integer, dtype.float):\n columns_sql += f',\"{col}_min\" float8'\n columns_sql += f',\"{col}_max\" float8'\n columns_sql += f',\"{col}_explain\" text'\n\n self.unregister_predictor(name)\n q = f\"\"\"\n CREATE FOREIGN TABLE {self.mindsdb_database}.{self._escape_table_name(name)} (\n {columns_sql}\n ) SERVER server_{self.mindsdb_database}\n OPTIONS (dbname 'mindsdb', table_name '{name}');\n \"\"\"\n self._query(q)\n\n def unregister_predictor(self, name):\n q = f\"\"\"\n DROP FOREIGN TABLE IF EXISTS {self.mindsdb_database}.{self._escape_table_name(name)};\n \"\"\"\n self._query(q)\n\n def get_row_count(self, query):\n q = f\"\"\" \n SELECT COUNT(*) as count\n FROM ({query}) as query;\"\"\"\n result = self._query(q)\n return result[0]['count']\n", "path": "mindsdb/integrations/postgres/postgres.py"}]}
2,707
204
gh_patches_debug_27460
rasdani/github-patches
git_diff
googleapis__python-bigquery-442
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> Too noise logging about telemetry Hello, In the Apache Airflow project, we use the BigQuery library, but recently we've started to see annoying log message when the library is loaded. It is enough that the library is loaded and there is an message every time. In my opinion, this message should be of a lower level (DEBUG) so that it is not displayed much less often or is displayed only when the client is initialized. ``` import logging logging.basicConfig(level=logging.INFO) from google.cloud import bigquery ``` Output: ``` INFO:google.cloud.bigquery.opentelemetry_tracing:This service is instrumented using OpenTelemetry. OpenTelemetry could not be imported; please add opentelemetry-api and opentelemetry-instrumentation packages in order to get BigQuery Tracing data. ``` Related issue: https://github.com/apache/airflow/issues/13131 CC: @tswast </issue> <code> [start of google/cloud/bigquery/opentelemetry_tracing.py] 1 # Copyright 2020 Google LLC 2 # 3 # Licensed under the Apache License, Version 2.0 (the "License"); 4 # you may not use this file except in compliance with the License. 5 # You may obtain a copy of the License at 6 # 7 # http://www.apache.org/licenses/LICENSE-2.0 8 # 9 # Unless required by applicable law or agreed to in writing, software 10 # distributed under the License is distributed on an "AS IS" BASIS, 11 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 12 # See the License for the specific language governing permissions and 13 # limitations under the License. 14 15 import logging 16 from contextlib import contextmanager 17 from google.api_core.exceptions import GoogleAPICallError 18 19 logger = logging.getLogger(__name__) 20 try: 21 from opentelemetry import trace 22 from opentelemetry.instrumentation.utils import http_status_to_canonical_code 23 from opentelemetry.trace.status import Status 24 25 HAS_OPENTELEMETRY = True 26 27 except ImportError: 28 logger.info( 29 "This service is instrumented using OpenTelemetry. " 30 "OpenTelemetry could not be imported; please " 31 "add opentelemetry-api and opentelemetry-instrumentation " 32 "packages in order to get BigQuery Tracing data." 33 ) 34 35 HAS_OPENTELEMETRY = False 36 37 _default_attributes = { 38 "db.system": "BigQuery" 39 } # static, default values assigned to all spans 40 41 42 @contextmanager 43 def create_span(name, attributes=None, client=None, job_ref=None): 44 """Creates a ContextManager for a Span to be exported to the configured exporter. 45 If no configuration exists yields None. 46 47 Args: 48 name (str): Name that will be set for the span being created 49 attributes (Optional[dict]): 50 Additional attributes that pertain to 51 the specific API call (i.e. not a default attribute) 52 client (Optional[google.cloud.bigquery.client.Client]): 53 Pass in a Client object to extract any attributes that may be 54 relevant to it and add them to the created spans. 55 job_ref (Optional[google.cloud.bigquery.job._AsyncJob]) 56 Pass in a _AsyncJob object to extract any attributes that may be 57 relevant to it and add them to the created spans. 58 59 Yields: 60 opentelemetry.trace.Span: Yields the newly created Span. 61 62 Raises: 63 google.api_core.exceptions.GoogleAPICallError: 64 Raised if a span could not be yielded or issue with call to 65 OpenTelemetry. 66 """ 67 final_attributes = _get_final_span_attributes(attributes, client, job_ref) 68 if not HAS_OPENTELEMETRY: 69 yield None 70 return 71 tracer = trace.get_tracer(__name__) 72 73 # yield new span value 74 with tracer.start_as_current_span(name=name, attributes=final_attributes) as span: 75 try: 76 yield span 77 except GoogleAPICallError as error: 78 if error.code is not None: 79 span.set_status(Status(http_status_to_canonical_code(error.code))) 80 raise 81 82 83 def _get_final_span_attributes(attributes=None, client=None, job_ref=None): 84 final_attributes = {} 85 final_attributes.update(_default_attributes.copy()) 86 if client: 87 client_attributes = _set_client_attributes(client) 88 final_attributes.update(client_attributes) 89 if job_ref: 90 job_attributes = _set_job_attributes(job_ref) 91 final_attributes.update(job_attributes) 92 if attributes: 93 final_attributes.update(attributes) 94 return final_attributes 95 96 97 def _set_client_attributes(client): 98 return {"db.name": client.project, "location": client.location} 99 100 101 def _set_job_attributes(job_ref): 102 job_attributes = { 103 "db.name": job_ref.project, 104 "location": job_ref.location, 105 "num_child_jobs": job_ref.num_child_jobs, 106 "job_id": job_ref.job_id, 107 "parent_job_id": job_ref.parent_job_id, 108 "state": job_ref.state, 109 } 110 111 job_attributes["hasErrors"] = job_ref.error_result is not None 112 113 if job_ref.created is not None: 114 job_attributes["timeCreated"] = job_ref.created.isoformat() 115 116 if job_ref.started is not None: 117 job_attributes["timeStarted"] = job_ref.started.isoformat() 118 119 if job_ref.ended is not None: 120 job_attributes["timeEnded"] = job_ref.ended.isoformat() 121 122 return job_attributes 123 [end of google/cloud/bigquery/opentelemetry_tracing.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/google/cloud/bigquery/opentelemetry_tracing.py b/google/cloud/bigquery/opentelemetry_tracing.py --- a/google/cloud/bigquery/opentelemetry_tracing.py +++ b/google/cloud/bigquery/opentelemetry_tracing.py @@ -23,16 +23,11 @@ from opentelemetry.trace.status import Status HAS_OPENTELEMETRY = True + _warned_telemetry = True except ImportError: - logger.info( - "This service is instrumented using OpenTelemetry. " - "OpenTelemetry could not be imported; please " - "add opentelemetry-api and opentelemetry-instrumentation " - "packages in order to get BigQuery Tracing data." - ) - HAS_OPENTELEMETRY = False + _warned_telemetry = False _default_attributes = { "db.system": "BigQuery" @@ -64,8 +59,18 @@ Raised if a span could not be yielded or issue with call to OpenTelemetry. """ + global _warned_telemetry final_attributes = _get_final_span_attributes(attributes, client, job_ref) if not HAS_OPENTELEMETRY: + if not _warned_telemetry: + logger.debug( + "This service is instrumented using OpenTelemetry. " + "OpenTelemetry could not be imported; please " + "add opentelemetry-api and opentelemetry-instrumentation " + "packages in order to get BigQuery Tracing data." + ) + _warned_telemetry = True + yield None return tracer = trace.get_tracer(__name__)
{"golden_diff": "diff --git a/google/cloud/bigquery/opentelemetry_tracing.py b/google/cloud/bigquery/opentelemetry_tracing.py\n--- a/google/cloud/bigquery/opentelemetry_tracing.py\n+++ b/google/cloud/bigquery/opentelemetry_tracing.py\n@@ -23,16 +23,11 @@\n from opentelemetry.trace.status import Status\n \n HAS_OPENTELEMETRY = True\n+ _warned_telemetry = True\n \n except ImportError:\n- logger.info(\n- \"This service is instrumented using OpenTelemetry. \"\n- \"OpenTelemetry could not be imported; please \"\n- \"add opentelemetry-api and opentelemetry-instrumentation \"\n- \"packages in order to get BigQuery Tracing data.\"\n- )\n-\n HAS_OPENTELEMETRY = False\n+ _warned_telemetry = False\n \n _default_attributes = {\n \"db.system\": \"BigQuery\"\n@@ -64,8 +59,18 @@\n Raised if a span could not be yielded or issue with call to\n OpenTelemetry.\n \"\"\"\n+ global _warned_telemetry\n final_attributes = _get_final_span_attributes(attributes, client, job_ref)\n if not HAS_OPENTELEMETRY:\n+ if not _warned_telemetry:\n+ logger.debug(\n+ \"This service is instrumented using OpenTelemetry. \"\n+ \"OpenTelemetry could not be imported; please \"\n+ \"add opentelemetry-api and opentelemetry-instrumentation \"\n+ \"packages in order to get BigQuery Tracing data.\"\n+ )\n+ _warned_telemetry = True\n+\n yield None\n return\n tracer = trace.get_tracer(__name__)\n", "issue": "Too noise logging about telemetry\nHello,\r\n\r\nIn the Apache Airflow project, we use the BigQuery library, but recently we've started to see annoying log message when the library is loaded. It is enough that the library is loaded and there is an message every time. \r\n\r\nIn my opinion, this message should be of a lower level (DEBUG) so that it is not displayed much less often or is displayed only when the client is initialized. \r\n```\r\nimport logging\r\nlogging.basicConfig(level=logging.INFO)\r\n\r\nfrom google.cloud import bigquery\r\n```\r\nOutput: \r\n```\r\nINFO:google.cloud.bigquery.opentelemetry_tracing:This service is instrumented using OpenTelemetry. OpenTelemetry could not be imported; please add opentelemetry-api and opentelemetry-instrumentation packages in order to get BigQuery Tracing data.\r\n```\r\n\r\nRelated issue: https://github.com/apache/airflow/issues/13131\r\n\r\nCC: @tswast \n", "before_files": [{"content": "# Copyright 2020 Google LLC\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nimport logging\nfrom contextlib import contextmanager\nfrom google.api_core.exceptions import GoogleAPICallError\n\nlogger = logging.getLogger(__name__)\ntry:\n from opentelemetry import trace\n from opentelemetry.instrumentation.utils import http_status_to_canonical_code\n from opentelemetry.trace.status import Status\n\n HAS_OPENTELEMETRY = True\n\nexcept ImportError:\n logger.info(\n \"This service is instrumented using OpenTelemetry. \"\n \"OpenTelemetry could not be imported; please \"\n \"add opentelemetry-api and opentelemetry-instrumentation \"\n \"packages in order to get BigQuery Tracing data.\"\n )\n\n HAS_OPENTELEMETRY = False\n\n_default_attributes = {\n \"db.system\": \"BigQuery\"\n} # static, default values assigned to all spans\n\n\n@contextmanager\ndef create_span(name, attributes=None, client=None, job_ref=None):\n \"\"\"Creates a ContextManager for a Span to be exported to the configured exporter.\n If no configuration exists yields None.\n\n Args:\n name (str): Name that will be set for the span being created\n attributes (Optional[dict]):\n Additional attributes that pertain to\n the specific API call (i.e. not a default attribute)\n client (Optional[google.cloud.bigquery.client.Client]):\n Pass in a Client object to extract any attributes that may be\n relevant to it and add them to the created spans.\n job_ref (Optional[google.cloud.bigquery.job._AsyncJob])\n Pass in a _AsyncJob object to extract any attributes that may be\n relevant to it and add them to the created spans.\n\n Yields:\n opentelemetry.trace.Span: Yields the newly created Span.\n\n Raises:\n google.api_core.exceptions.GoogleAPICallError:\n Raised if a span could not be yielded or issue with call to\n OpenTelemetry.\n \"\"\"\n final_attributes = _get_final_span_attributes(attributes, client, job_ref)\n if not HAS_OPENTELEMETRY:\n yield None\n return\n tracer = trace.get_tracer(__name__)\n\n # yield new span value\n with tracer.start_as_current_span(name=name, attributes=final_attributes) as span:\n try:\n yield span\n except GoogleAPICallError as error:\n if error.code is not None:\n span.set_status(Status(http_status_to_canonical_code(error.code)))\n raise\n\n\ndef _get_final_span_attributes(attributes=None, client=None, job_ref=None):\n final_attributes = {}\n final_attributes.update(_default_attributes.copy())\n if client:\n client_attributes = _set_client_attributes(client)\n final_attributes.update(client_attributes)\n if job_ref:\n job_attributes = _set_job_attributes(job_ref)\n final_attributes.update(job_attributes)\n if attributes:\n final_attributes.update(attributes)\n return final_attributes\n\n\ndef _set_client_attributes(client):\n return {\"db.name\": client.project, \"location\": client.location}\n\n\ndef _set_job_attributes(job_ref):\n job_attributes = {\n \"db.name\": job_ref.project,\n \"location\": job_ref.location,\n \"num_child_jobs\": job_ref.num_child_jobs,\n \"job_id\": job_ref.job_id,\n \"parent_job_id\": job_ref.parent_job_id,\n \"state\": job_ref.state,\n }\n\n job_attributes[\"hasErrors\"] = job_ref.error_result is not None\n\n if job_ref.created is not None:\n job_attributes[\"timeCreated\"] = job_ref.created.isoformat()\n\n if job_ref.started is not None:\n job_attributes[\"timeStarted\"] = job_ref.started.isoformat()\n\n if job_ref.ended is not None:\n job_attributes[\"timeEnded\"] = job_ref.ended.isoformat()\n\n return job_attributes\n", "path": "google/cloud/bigquery/opentelemetry_tracing.py"}]}
1,943
377
gh_patches_debug_9151
rasdani/github-patches
git_diff
python-discord__bot-1199
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> Race condition in wait_for_deletion between a manual delete and reacting Sentry Issue: [BOT-86](https://sentry.io/organizations/python-discord/issues/1861537034/?referrer=github_integration) The message may be deleted before `wait_for_deletion` has a chance to add a reaction. ``` NotFound: 404 Not Found (error code: 10008): Unknown Message File "bot/utils/messages.py", line 38, in wait_for_deletion await message.add_reaction(emoji) File "discord/message.py", line 953, in add_reaction await self._state.http.add_reaction(self.channel.id, self.id, emoji) File "discord/http.py", line 243, in request raise NotFound(r, data) Task exception was never retrieved future: <Task finished name='Task-333258' coro=<wait_for_deletion() done, defined at /bot/bot/utils/messages.py:19> exception=NotFound('404 Not Found (error code: 10008): Unknown Message')> ``` </issue> <code> [start of bot/utils/messages.py] 1 import asyncio 2 import contextlib 3 import logging 4 import random 5 import re 6 from io import BytesIO 7 from typing import List, Optional, Sequence, Union 8 9 import discord 10 from discord.errors import HTTPException 11 from discord.ext.commands import Context 12 13 from bot.constants import Emojis, NEGATIVE_REPLIES 14 15 log = logging.getLogger(__name__) 16 17 18 async def wait_for_deletion( 19 message: discord.Message, 20 user_ids: Sequence[discord.abc.Snowflake], 21 client: discord.Client, 22 deletion_emojis: Sequence[str] = (Emojis.trashcan,), 23 timeout: float = 60 * 5, 24 attach_emojis: bool = True, 25 ) -> None: 26 """ 27 Wait for up to `timeout` seconds for a reaction by any of the specified `user_ids` to delete the message. 28 29 An `attach_emojis` bool may be specified to determine whether to attach the given 30 `deletion_emojis` to the message in the given `context`. 31 """ 32 if message.guild is None: 33 raise ValueError("Message must be sent on a guild") 34 35 if attach_emojis: 36 for emoji in deletion_emojis: 37 await message.add_reaction(emoji) 38 39 def check(reaction: discord.Reaction, user: discord.Member) -> bool: 40 """Check that the deletion emoji is reacted by the appropriate user.""" 41 return ( 42 reaction.message.id == message.id 43 and str(reaction.emoji) in deletion_emojis 44 and user.id in user_ids 45 ) 46 47 with contextlib.suppress(asyncio.TimeoutError): 48 await client.wait_for('reaction_add', check=check, timeout=timeout) 49 await message.delete() 50 51 52 async def send_attachments( 53 message: discord.Message, 54 destination: Union[discord.TextChannel, discord.Webhook], 55 link_large: bool = True 56 ) -> List[str]: 57 """ 58 Re-upload the message's attachments to the destination and return a list of their new URLs. 59 60 Each attachment is sent as a separate message to more easily comply with the request/file size 61 limit. If link_large is True, attachments which are too large are instead grouped into a single 62 embed which links to them. 63 """ 64 large = [] 65 urls = [] 66 for attachment in message.attachments: 67 failure_msg = ( 68 f"Failed to re-upload attachment {attachment.filename} from message {message.id}" 69 ) 70 71 try: 72 # Allow 512 bytes of leeway for the rest of the request. 73 # This should avoid most files that are too large, 74 # but some may get through hence the try-catch. 75 if attachment.size <= destination.guild.filesize_limit - 512: 76 with BytesIO() as file: 77 await attachment.save(file, use_cached=True) 78 attachment_file = discord.File(file, filename=attachment.filename) 79 80 if isinstance(destination, discord.TextChannel): 81 msg = await destination.send(file=attachment_file) 82 urls.append(msg.attachments[0].url) 83 else: 84 await destination.send( 85 file=attachment_file, 86 username=sub_clyde(message.author.display_name), 87 avatar_url=message.author.avatar_url 88 ) 89 elif link_large: 90 large.append(attachment) 91 else: 92 log.info(f"{failure_msg} because it's too large.") 93 except HTTPException as e: 94 if link_large and e.status == 413: 95 large.append(attachment) 96 else: 97 log.warning(f"{failure_msg} with status {e.status}.", exc_info=e) 98 99 if link_large and large: 100 desc = "\n".join(f"[{attachment.filename}]({attachment.url})" for attachment in large) 101 embed = discord.Embed(description=desc) 102 embed.set_footer(text="Attachments exceed upload size limit.") 103 104 if isinstance(destination, discord.TextChannel): 105 await destination.send(embed=embed) 106 else: 107 await destination.send( 108 embed=embed, 109 username=sub_clyde(message.author.display_name), 110 avatar_url=message.author.avatar_url 111 ) 112 113 return urls 114 115 116 def sub_clyde(username: Optional[str]) -> Optional[str]: 117 """ 118 Replace "e"/"E" in any "clyde" in `username` with a Cyrillic "е"/"E" and return the new string. 119 120 Discord disallows "clyde" anywhere in the username for webhooks. It will return a 400. 121 Return None only if `username` is None. 122 """ 123 def replace_e(match: re.Match) -> str: 124 char = "е" if match[2] == "e" else "Е" 125 return match[1] + char 126 127 if username: 128 return re.sub(r"(clyd)(e)", replace_e, username, flags=re.I) 129 else: 130 return username # Empty string or None 131 132 133 async def send_denial(ctx: Context, reason: str) -> None: 134 """Send an embed denying the user with the given reason.""" 135 embed = discord.Embed() 136 embed.colour = discord.Colour.red() 137 embed.title = random.choice(NEGATIVE_REPLIES) 138 embed.description = reason 139 140 await ctx.send(embed=embed) 141 142 143 def format_user(user: discord.abc.User) -> str: 144 """Return a string for `user` which has their mention and ID.""" 145 return f"{user.mention} (`{user.id}`)" 146 [end of bot/utils/messages.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/bot/utils/messages.py b/bot/utils/messages.py --- a/bot/utils/messages.py +++ b/bot/utils/messages.py @@ -34,7 +34,11 @@ if attach_emojis: for emoji in deletion_emojis: - await message.add_reaction(emoji) + try: + await message.add_reaction(emoji) + except discord.NotFound: + log.trace(f"Aborting wait_for_deletion: message {message.id} deleted prematurely.") + return def check(reaction: discord.Reaction, user: discord.Member) -> bool: """Check that the deletion emoji is reacted by the appropriate user."""
{"golden_diff": "diff --git a/bot/utils/messages.py b/bot/utils/messages.py\n--- a/bot/utils/messages.py\n+++ b/bot/utils/messages.py\n@@ -34,7 +34,11 @@\n \n if attach_emojis:\n for emoji in deletion_emojis:\n- await message.add_reaction(emoji)\n+ try:\n+ await message.add_reaction(emoji)\n+ except discord.NotFound:\n+ log.trace(f\"Aborting wait_for_deletion: message {message.id} deleted prematurely.\")\n+ return\n \n def check(reaction: discord.Reaction, user: discord.Member) -> bool:\n \"\"\"Check that the deletion emoji is reacted by the appropriate user.\"\"\"\n", "issue": "Race condition in wait_for_deletion between a manual delete and reacting\nSentry Issue: [BOT-86](https://sentry.io/organizations/python-discord/issues/1861537034/?referrer=github_integration)\r\n\r\nThe message may be deleted before `wait_for_deletion` has a chance to add a reaction.\r\n\r\n```\r\nNotFound: 404 Not Found (error code: 10008): Unknown Message\r\n File \"bot/utils/messages.py\", line 38, in wait_for_deletion\r\n await message.add_reaction(emoji)\r\n File \"discord/message.py\", line 953, in add_reaction\r\n await self._state.http.add_reaction(self.channel.id, self.id, emoji)\r\n File \"discord/http.py\", line 243, in request\r\n raise NotFound(r, data)\r\n\r\nTask exception was never retrieved\r\nfuture: <Task finished name='Task-333258' coro=<wait_for_deletion() done, defined at /bot/bot/utils/messages.py:19> exception=NotFound('404 Not Found (error code: 10008): Unknown Message')>\r\n```\n", "before_files": [{"content": "import asyncio\nimport contextlib\nimport logging\nimport random\nimport re\nfrom io import BytesIO\nfrom typing import List, Optional, Sequence, Union\n\nimport discord\nfrom discord.errors import HTTPException\nfrom discord.ext.commands import Context\n\nfrom bot.constants import Emojis, NEGATIVE_REPLIES\n\nlog = logging.getLogger(__name__)\n\n\nasync def wait_for_deletion(\n message: discord.Message,\n user_ids: Sequence[discord.abc.Snowflake],\n client: discord.Client,\n deletion_emojis: Sequence[str] = (Emojis.trashcan,),\n timeout: float = 60 * 5,\n attach_emojis: bool = True,\n) -> None:\n \"\"\"\n Wait for up to `timeout` seconds for a reaction by any of the specified `user_ids` to delete the message.\n\n An `attach_emojis` bool may be specified to determine whether to attach the given\n `deletion_emojis` to the message in the given `context`.\n \"\"\"\n if message.guild is None:\n raise ValueError(\"Message must be sent on a guild\")\n\n if attach_emojis:\n for emoji in deletion_emojis:\n await message.add_reaction(emoji)\n\n def check(reaction: discord.Reaction, user: discord.Member) -> bool:\n \"\"\"Check that the deletion emoji is reacted by the appropriate user.\"\"\"\n return (\n reaction.message.id == message.id\n and str(reaction.emoji) in deletion_emojis\n and user.id in user_ids\n )\n\n with contextlib.suppress(asyncio.TimeoutError):\n await client.wait_for('reaction_add', check=check, timeout=timeout)\n await message.delete()\n\n\nasync def send_attachments(\n message: discord.Message,\n destination: Union[discord.TextChannel, discord.Webhook],\n link_large: bool = True\n) -> List[str]:\n \"\"\"\n Re-upload the message's attachments to the destination and return a list of their new URLs.\n\n Each attachment is sent as a separate message to more easily comply with the request/file size\n limit. If link_large is True, attachments which are too large are instead grouped into a single\n embed which links to them.\n \"\"\"\n large = []\n urls = []\n for attachment in message.attachments:\n failure_msg = (\n f\"Failed to re-upload attachment {attachment.filename} from message {message.id}\"\n )\n\n try:\n # Allow 512 bytes of leeway for the rest of the request.\n # This should avoid most files that are too large,\n # but some may get through hence the try-catch.\n if attachment.size <= destination.guild.filesize_limit - 512:\n with BytesIO() as file:\n await attachment.save(file, use_cached=True)\n attachment_file = discord.File(file, filename=attachment.filename)\n\n if isinstance(destination, discord.TextChannel):\n msg = await destination.send(file=attachment_file)\n urls.append(msg.attachments[0].url)\n else:\n await destination.send(\n file=attachment_file,\n username=sub_clyde(message.author.display_name),\n avatar_url=message.author.avatar_url\n )\n elif link_large:\n large.append(attachment)\n else:\n log.info(f\"{failure_msg} because it's too large.\")\n except HTTPException as e:\n if link_large and e.status == 413:\n large.append(attachment)\n else:\n log.warning(f\"{failure_msg} with status {e.status}.\", exc_info=e)\n\n if link_large and large:\n desc = \"\\n\".join(f\"[{attachment.filename}]({attachment.url})\" for attachment in large)\n embed = discord.Embed(description=desc)\n embed.set_footer(text=\"Attachments exceed upload size limit.\")\n\n if isinstance(destination, discord.TextChannel):\n await destination.send(embed=embed)\n else:\n await destination.send(\n embed=embed,\n username=sub_clyde(message.author.display_name),\n avatar_url=message.author.avatar_url\n )\n\n return urls\n\n\ndef sub_clyde(username: Optional[str]) -> Optional[str]:\n \"\"\"\n Replace \"e\"/\"E\" in any \"clyde\" in `username` with a Cyrillic \"\u0435\"/\"E\" and return the new string.\n\n Discord disallows \"clyde\" anywhere in the username for webhooks. It will return a 400.\n Return None only if `username` is None.\n \"\"\"\n def replace_e(match: re.Match) -> str:\n char = \"\u0435\" if match[2] == \"e\" else \"\u0415\"\n return match[1] + char\n\n if username:\n return re.sub(r\"(clyd)(e)\", replace_e, username, flags=re.I)\n else:\n return username # Empty string or None\n\n\nasync def send_denial(ctx: Context, reason: str) -> None:\n \"\"\"Send an embed denying the user with the given reason.\"\"\"\n embed = discord.Embed()\n embed.colour = discord.Colour.red()\n embed.title = random.choice(NEGATIVE_REPLIES)\n embed.description = reason\n\n await ctx.send(embed=embed)\n\n\ndef format_user(user: discord.abc.User) -> str:\n \"\"\"Return a string for `user` which has their mention and ID.\"\"\"\n return f\"{user.mention} (`{user.id}`)\"\n", "path": "bot/utils/messages.py"}]}
2,264
147
gh_patches_debug_7832
rasdani/github-patches
git_diff
Mailu__Mailu-1974
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> DEFER_ON_TLS_ERROR in mailu.env doesn't appear to be applied. Thank you for opening an issue with Mailu. Please understand that issues are meant for bugs and enhancement-requests. For **user-support questions**, reach out to us on [matrix](https://matrix.to/#/#mailu:tedomum.net). To be able to help you best, we need some more information. ## Before you open your issue - [X] Check if no issue or pull-request for this already exists. - [X] Check [documentation](https://mailu.io/master/) and [FAQ](https://mailu.io/master/faq.html). (Tip, use the search function on the documentation page) - [X] You understand `Mailu` is made by volunteers in their **free time** — be conscise, civil and accept that delays can occur. - [X] The title of the issue should be short and simple. It should contain specific terms related to the actual issue. Be specific while writing the title. ## Environment & Versions ### Environment - [X] docker-compose - [ ] kubernetes - [ ] docker swarm ### Versions master (d464187) ## Description DEFER_ON_TLS_ERROR in mailu.env doesn't appear to be applied. I've set this environment variable to "false" and my emails still appear to be deferred according to the SMTP container. `(status=deferred (client TLS configuration problem))` I've set this to "false" because MTA-STS isn't as widely deployed as one would hope, and I don't want my emails blackholed and deferred until my destination's policy is published (which may be never)... ## Replication Steps In mailu.env - set `DEFER_ON_TLS_ERROR=false`, reload containers, then try sending an email to a domain without an MTA-STS policy... Then output logs for the SMTP container and note the message deferral. ## Expected behaviour I expect the email to continue being sent despite the MTA-STS error. ## Logs ``` smtp_1 | Sep 09 13:42:08 mail postfix/cleanup[416]: 85B6147DDA: message-id=<[email protected]> smtp_1 | Sep 09 13:42:09 mail postfix/qmgr[391]: 85B6147DDA: from=<[email protected]>, size=6074, nrcpt=1 (queue active) smtp_1 | Sep 09 13:42:09 mail postfix/smtp[417]: warning: connect to /tmp/mta-sts.socket: No such file or directory smtp_1 | Sep 09 13:42:09 mail postfix/smtp[417]: warning: table socketmap:unix:/tmp/mta-sts.socket:postfix lookup error: No such file or directory smtp_1 | Sep 09 13:42:09 mail postfix/smtp[417]: warning: socketmap:unix:/tmp/mta-sts.socket:postfix lookup error for "ssl-tools.net" smtp_1 | Sep 09 13:42:09 mail postfix/smtp[417]: warning: smtp_tls_policy_maps, next-hop destination "ssl-tools.net": policy table lookup error smtp_1 | Sep 09 13:42:09 mail postfix/smtp[417]: warning: TLS policy lookup for ssl-tools.net/ssl-tools.net: client TLS configuration problem smtp_1 | Sep 09 13:42:09 mail postfix/smtp[417]: warning: TLS policy lookup for ssl-tools.net/ssl-tools.net: client TLS configuration problem smtp_1 | Sep 09 13:42:09 mail postfix/smtp[417]: 85B6147DDA: to=<[email protected]>, relay=none, delay=1.1, delays=1/0.03/0.01/0, dsn=4.7.5, status=deferred (client TLS configuration problem) ``` </issue> <code> [start of core/postfix/start.py] 1 #!/usr/bin/python3 2 3 import os 4 import glob 5 import shutil 6 import multiprocessing 7 import logging as log 8 import sys 9 10 from podop import run_server 11 from pwd import getpwnam 12 from socrate import system, conf 13 14 log.basicConfig(stream=sys.stderr, level=os.environ.get("LOG_LEVEL", "WARNING")) 15 16 def start_podop(): 17 os.setuid(getpwnam('postfix').pw_uid) 18 os.mkdir('/dev/shm/postfix',mode=0o700) 19 url = "http://" + os.environ["ADMIN_ADDRESS"] + "/internal/postfix/" 20 # TODO: Remove verbosity setting from Podop? 21 run_server(0, "postfix", "/tmp/podop.socket", [ 22 ("transport", "url", url + "transport/§"), 23 ("alias", "url", url + "alias/§"), 24 ("dane", "url", url + "dane/§"), 25 ("domain", "url", url + "domain/§"), 26 ("mailbox", "url", url + "mailbox/§"), 27 ("recipientmap", "url", url + "recipient/map/§"), 28 ("sendermap", "url", url + "sender/map/§"), 29 ("senderaccess", "url", url + "sender/access/§"), 30 ("senderlogin", "url", url + "sender/login/§"), 31 ("senderrate", "url", url + "sender/rate/§") 32 ]) 33 34 def start_mta_sts_daemon(): 35 os.chmod("/root/", 0o755) # read access to /root/.netrc required 36 os.setuid(getpwnam('postfix').pw_uid) 37 from postfix_mta_sts_resolver import daemon 38 daemon.main() 39 40 def is_valid_postconf_line(line): 41 return not line.startswith("#") \ 42 and not line == '' 43 44 # Actual startup script 45 os.environ["FRONT_ADDRESS"] = system.get_host_address_from_environment("FRONT", "front") 46 os.environ["ADMIN_ADDRESS"] = system.get_host_address_from_environment("ADMIN", "admin") 47 os.environ["ANTISPAM_MILTER_ADDRESS"] = system.get_host_address_from_environment("ANTISPAM_MILTER", "antispam:11332") 48 os.environ["LMTP_ADDRESS"] = system.get_host_address_from_environment("LMTP", "imap:2525") 49 os.environ["OUTCLEAN"] = os.environ["HOSTNAMES"].split(",")[0] 50 try: 51 _to_lookup = os.environ["OUTCLEAN"] 52 # Ensure we lookup a FQDN: @see #1884 53 if not _to_lookup.endswith('.'): 54 _to_lookup += '.' 55 os.environ["OUTCLEAN_ADDRESS"] = system.resolve_hostname(_to_lookup) 56 except: 57 os.environ["OUTCLEAN_ADDRESS"] = "10.10.10.10" 58 59 for postfix_file in glob.glob("/conf/*.cf"): 60 conf.jinja(postfix_file, os.environ, os.path.join("/etc/postfix", os.path.basename(postfix_file))) 61 62 if os.path.exists("/overrides/postfix.cf"): 63 for line in open("/overrides/postfix.cf").read().strip().split("\n"): 64 if is_valid_postconf_line(line): 65 os.system('postconf -e "{}"'.format(line)) 66 67 if os.path.exists("/overrides/postfix.master"): 68 for line in open("/overrides/postfix.master").read().strip().split("\n"): 69 if is_valid_postconf_line(line): 70 os.system('postconf -Me "{}"'.format(line)) 71 72 for map_file in glob.glob("/overrides/*.map"): 73 destination = os.path.join("/etc/postfix", os.path.basename(map_file)) 74 shutil.copyfile(map_file, destination) 75 os.system("postmap {}".format(destination)) 76 os.remove(destination) 77 78 if os.path.exists("/overrides/mta-sts-daemon.yml"): 79 shutil.copyfile("/overrides/mta-sts-daemon.yml", "/etc/mta-sts-daemon.yml") 80 conf.jinja("/etc/mta-sts-daemon.yml", os.environ, "/etc/mta-sts-daemon.yml") 81 82 if not os.path.exists("/etc/postfix/tls_policy.map.lmdb"): 83 open("/etc/postfix/tls_policy.map", "a").close() 84 os.system("postmap /etc/postfix/tls_policy.map") 85 86 if "RELAYUSER" in os.environ: 87 path = "/etc/postfix/sasl_passwd" 88 conf.jinja("/conf/sasl_passwd", os.environ, path) 89 os.system("postmap {}".format(path)) 90 91 # Run Podop and Postfix 92 multiprocessing.Process(target=start_podop).start() 93 multiprocessing.Process(target=start_mta_sts_daemon).start() 94 os.system("/usr/libexec/postfix/post-install meta_directory=/etc/postfix create-missing") 95 # Before starting postfix, we need to check permissions on /queue 96 # in the event that postfix,postdrop id have changed 97 os.system("postfix set-permissions") 98 os.system("postfix start-fg") 99 [end of core/postfix/start.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/core/postfix/start.py b/core/postfix/start.py --- a/core/postfix/start.py +++ b/core/postfix/start.py @@ -77,7 +77,8 @@ if os.path.exists("/overrides/mta-sts-daemon.yml"): shutil.copyfile("/overrides/mta-sts-daemon.yml", "/etc/mta-sts-daemon.yml") -conf.jinja("/etc/mta-sts-daemon.yml", os.environ, "/etc/mta-sts-daemon.yml") +else: + conf.jinja("/conf/mta-sts-daemon.yml", os.environ, "/etc/mta-sts-daemon.yml") if not os.path.exists("/etc/postfix/tls_policy.map.lmdb"): open("/etc/postfix/tls_policy.map", "a").close()
{"golden_diff": "diff --git a/core/postfix/start.py b/core/postfix/start.py\n--- a/core/postfix/start.py\n+++ b/core/postfix/start.py\n@@ -77,7 +77,8 @@\n \n if os.path.exists(\"/overrides/mta-sts-daemon.yml\"):\n shutil.copyfile(\"/overrides/mta-sts-daemon.yml\", \"/etc/mta-sts-daemon.yml\")\n-conf.jinja(\"/etc/mta-sts-daemon.yml\", os.environ, \"/etc/mta-sts-daemon.yml\")\n+else:\n+ conf.jinja(\"/conf/mta-sts-daemon.yml\", os.environ, \"/etc/mta-sts-daemon.yml\")\n \n if not os.path.exists(\"/etc/postfix/tls_policy.map.lmdb\"):\n open(\"/etc/postfix/tls_policy.map\", \"a\").close()\n", "issue": "DEFER_ON_TLS_ERROR in mailu.env doesn't appear to be applied.\nThank you for opening an issue with Mailu. Please understand that issues are meant for bugs and enhancement-requests.\r\nFor **user-support questions**, reach out to us on [matrix](https://matrix.to/#/#mailu:tedomum.net).\r\n\r\nTo be able to help you best, we need some more information.\r\n\r\n## Before you open your issue\r\n- [X] Check if no issue or pull-request for this already exists.\r\n- [X] Check [documentation](https://mailu.io/master/) and [FAQ](https://mailu.io/master/faq.html). (Tip, use the search function on the documentation page)\r\n- [X] You understand `Mailu` is made by volunteers in their **free time** \u2014 be conscise, civil and accept that delays can occur.\r\n- [X] The title of the issue should be short and simple. It should contain specific terms related to the actual issue. Be specific while writing the title.\r\n\r\n## Environment & Versions\r\n### Environment\r\n - [X] docker-compose\r\n - [ ] kubernetes\r\n - [ ] docker swarm\r\n\r\n### Versions\r\nmaster (d464187)\r\n\r\n## Description\r\nDEFER_ON_TLS_ERROR in mailu.env doesn't appear to be applied. I've set this environment variable to \"false\" and my emails still appear to be deferred according to the SMTP container. \r\n`(status=deferred (client TLS configuration problem))`\r\n\r\nI've set this to \"false\" because MTA-STS isn't as widely deployed as one would hope, and I don't want my emails blackholed and deferred until my destination's policy is published (which may be never)... \r\n\r\n## Replication Steps\r\nIn mailu.env - set `DEFER_ON_TLS_ERROR=false`, reload containers, then try sending an email to a domain without an MTA-STS policy...\r\n\r\nThen output logs for the SMTP container and note the message deferral.\r\n\r\n## Expected behaviour\r\nI expect the email to continue being sent despite the MTA-STS error.\r\n\r\n## Logs\r\n```\r\nsmtp_1 | Sep 09 13:42:08 mail postfix/cleanup[416]: 85B6147DDA: message-id=<[email protected]>\r\nsmtp_1 | Sep 09 13:42:09 mail postfix/qmgr[391]: 85B6147DDA: from=<[email protected]>, size=6074, nrcpt=1 (queue active)\r\nsmtp_1 | Sep 09 13:42:09 mail postfix/smtp[417]: warning: connect to /tmp/mta-sts.socket: No such file or directory\r\nsmtp_1 | Sep 09 13:42:09 mail postfix/smtp[417]: warning: table socketmap:unix:/tmp/mta-sts.socket:postfix lookup error: No such file or directory\r\nsmtp_1 | Sep 09 13:42:09 mail postfix/smtp[417]: warning: socketmap:unix:/tmp/mta-sts.socket:postfix lookup error for \"ssl-tools.net\"\r\nsmtp_1 | Sep 09 13:42:09 mail postfix/smtp[417]: warning: smtp_tls_policy_maps, next-hop destination \"ssl-tools.net\": policy table lookup error\r\nsmtp_1 | Sep 09 13:42:09 mail postfix/smtp[417]: warning: TLS policy lookup for ssl-tools.net/ssl-tools.net: client TLS configuration problem\r\nsmtp_1 | Sep 09 13:42:09 mail postfix/smtp[417]: warning: TLS policy lookup for ssl-tools.net/ssl-tools.net: client TLS configuration problem\r\nsmtp_1 | Sep 09 13:42:09 mail postfix/smtp[417]: 85B6147DDA: to=<[email protected]>, relay=none, delay=1.1, delays=1/0.03/0.01/0, dsn=4.7.5, status=deferred (client TLS configuration problem)\r\n```\r\n\n", "before_files": [{"content": "#!/usr/bin/python3\n\nimport os\nimport glob\nimport shutil\nimport multiprocessing\nimport logging as log\nimport sys\n\nfrom podop import run_server\nfrom pwd import getpwnam\nfrom socrate import system, conf\n\nlog.basicConfig(stream=sys.stderr, level=os.environ.get(\"LOG_LEVEL\", \"WARNING\"))\n\ndef start_podop():\n os.setuid(getpwnam('postfix').pw_uid)\n os.mkdir('/dev/shm/postfix',mode=0o700)\n url = \"http://\" + os.environ[\"ADMIN_ADDRESS\"] + \"/internal/postfix/\"\n # TODO: Remove verbosity setting from Podop?\n run_server(0, \"postfix\", \"/tmp/podop.socket\", [\n (\"transport\", \"url\", url + \"transport/\u00a7\"),\n (\"alias\", \"url\", url + \"alias/\u00a7\"),\n (\"dane\", \"url\", url + \"dane/\u00a7\"),\n (\"domain\", \"url\", url + \"domain/\u00a7\"),\n (\"mailbox\", \"url\", url + \"mailbox/\u00a7\"),\n (\"recipientmap\", \"url\", url + \"recipient/map/\u00a7\"),\n (\"sendermap\", \"url\", url + \"sender/map/\u00a7\"),\n (\"senderaccess\", \"url\", url + \"sender/access/\u00a7\"),\n (\"senderlogin\", \"url\", url + \"sender/login/\u00a7\"),\n (\"senderrate\", \"url\", url + \"sender/rate/\u00a7\")\n ])\n\ndef start_mta_sts_daemon():\n os.chmod(\"/root/\", 0o755) # read access to /root/.netrc required\n os.setuid(getpwnam('postfix').pw_uid)\n from postfix_mta_sts_resolver import daemon\n daemon.main()\n\ndef is_valid_postconf_line(line):\n return not line.startswith(\"#\") \\\n and not line == ''\n\n# Actual startup script\nos.environ[\"FRONT_ADDRESS\"] = system.get_host_address_from_environment(\"FRONT\", \"front\")\nos.environ[\"ADMIN_ADDRESS\"] = system.get_host_address_from_environment(\"ADMIN\", \"admin\")\nos.environ[\"ANTISPAM_MILTER_ADDRESS\"] = system.get_host_address_from_environment(\"ANTISPAM_MILTER\", \"antispam:11332\")\nos.environ[\"LMTP_ADDRESS\"] = system.get_host_address_from_environment(\"LMTP\", \"imap:2525\")\nos.environ[\"OUTCLEAN\"] = os.environ[\"HOSTNAMES\"].split(\",\")[0]\ntry:\n _to_lookup = os.environ[\"OUTCLEAN\"]\n # Ensure we lookup a FQDN: @see #1884\n if not _to_lookup.endswith('.'):\n _to_lookup += '.'\n os.environ[\"OUTCLEAN_ADDRESS\"] = system.resolve_hostname(_to_lookup)\nexcept:\n os.environ[\"OUTCLEAN_ADDRESS\"] = \"10.10.10.10\"\n\nfor postfix_file in glob.glob(\"/conf/*.cf\"):\n conf.jinja(postfix_file, os.environ, os.path.join(\"/etc/postfix\", os.path.basename(postfix_file)))\n\nif os.path.exists(\"/overrides/postfix.cf\"):\n for line in open(\"/overrides/postfix.cf\").read().strip().split(\"\\n\"):\n if is_valid_postconf_line(line):\n os.system('postconf -e \"{}\"'.format(line))\n\nif os.path.exists(\"/overrides/postfix.master\"):\n for line in open(\"/overrides/postfix.master\").read().strip().split(\"\\n\"):\n if is_valid_postconf_line(line):\n os.system('postconf -Me \"{}\"'.format(line))\n\nfor map_file in glob.glob(\"/overrides/*.map\"):\n destination = os.path.join(\"/etc/postfix\", os.path.basename(map_file))\n shutil.copyfile(map_file, destination)\n os.system(\"postmap {}\".format(destination))\n os.remove(destination)\n\nif os.path.exists(\"/overrides/mta-sts-daemon.yml\"):\n shutil.copyfile(\"/overrides/mta-sts-daemon.yml\", \"/etc/mta-sts-daemon.yml\")\nconf.jinja(\"/etc/mta-sts-daemon.yml\", os.environ, \"/etc/mta-sts-daemon.yml\")\n\nif not os.path.exists(\"/etc/postfix/tls_policy.map.lmdb\"):\n open(\"/etc/postfix/tls_policy.map\", \"a\").close()\n os.system(\"postmap /etc/postfix/tls_policy.map\")\n\nif \"RELAYUSER\" in os.environ:\n path = \"/etc/postfix/sasl_passwd\"\n conf.jinja(\"/conf/sasl_passwd\", os.environ, path)\n os.system(\"postmap {}\".format(path))\n\n# Run Podop and Postfix\nmultiprocessing.Process(target=start_podop).start()\nmultiprocessing.Process(target=start_mta_sts_daemon).start()\nos.system(\"/usr/libexec/postfix/post-install meta_directory=/etc/postfix create-missing\")\n# Before starting postfix, we need to check permissions on /queue\n# in the event that postfix,postdrop id have changed\nos.system(\"postfix set-permissions\")\nos.system(\"postfix start-fg\")\n", "path": "core/postfix/start.py"}]}
2,770
180
gh_patches_debug_23666
rasdani/github-patches
git_diff
pypi__warehouse-7927
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> Support for properly displaying the site translated to RTL languages **What's the problem this feature will solve?** Display PyPI.org propely when translated to RTL languages such as Arabic, Farsi and Hebrew. **Describe the solution you'd like** I can't suggest a specific technical solution, since I'm not familiar with the website's design and implementation. **Additional context** The context is the recent work on translation into additional languages. </issue> <code> [start of warehouse/i18n/__init__.py] 1 # Licensed under the Apache License, Version 2.0 (the "License"); 2 # you may not use this file except in compliance with the License. 3 # You may obtain a copy of the License at 4 # 5 # http://www.apache.org/licenses/LICENSE-2.0 6 # 7 # Unless required by applicable law or agreed to in writing, software 8 # distributed under the License is distributed on an "AS IS" BASIS, 9 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 10 # See the License for the specific language governing permissions and 11 # limitations under the License. 12 13 import functools 14 15 from babel.core import Locale 16 from pyramid import viewderivers 17 from pyramid.i18n import TranslationStringFactory, default_locale_negotiator 18 from pyramid.threadlocal import get_current_request 19 20 from warehouse.cache.http import add_vary 21 22 # Taken from: 23 # https://github.com/django/django/blob/master/django/conf/locale/__init__.py 24 KNOWN_LOCALES = { 25 "en": "English", # English 26 "es": "español", # Spanish 27 "fr": "français", # French 28 "ja": "日本語", # Japanese 29 "pt_BR": "Português Brasileiro", # Brazilian Portugeuse 30 "uk": "Українська", # Ukrainian 31 "el": "Ελληνικά", # Greek 32 "de": "Deutsch", # German 33 "zh_Hans": "简体中文", # Simplified Chinese 34 "ru": "Русский", # Russian 35 } 36 37 LOCALE_ATTR = "_LOCALE_" 38 39 _translation_factory = TranslationStringFactory("messages") 40 41 42 class LazyString: 43 def __init__(self, fn, *args, **kwargs): 44 self.fn = fn 45 self.args = args 46 self.mapping = kwargs.get("mapping", {}) 47 self.kwargs = kwargs 48 49 def __json__(self, request): 50 return str(self) 51 52 def __mod__(self, new_mapping): 53 mapping = self.mapping.copy() 54 mapping.update(new_mapping) 55 return LazyString(self.fn, *self.args, mapping=new_mapping, **self.kwargs) 56 57 def __str__(self): 58 return self.fn(*self.args, **self.kwargs) 59 60 61 def _locale(request): 62 """ 63 Computes a babel.core:Locale() object for this request. 64 """ 65 return Locale.parse(request.locale_name, sep="_") 66 67 68 def _negotiate_locale(request): 69 locale_name = getattr(request, LOCALE_ATTR, None) 70 if locale_name is not None: 71 return locale_name 72 73 locale_name = request.params.get(LOCALE_ATTR) 74 if locale_name is not None: 75 return locale_name 76 77 locale_name = request.cookies.get(LOCALE_ATTR) 78 if locale_name is not None: 79 return locale_name 80 81 if not request.accept_language: 82 return default_locale_negotiator(request) 83 84 return request.accept_language.best_match( 85 tuple(KNOWN_LOCALES.keys()), default_match=default_locale_negotiator(request) 86 ) 87 88 89 def _localize(request, message, **kwargs): 90 """ 91 To be used on the request directly, e.g. `request._(message)` 92 """ 93 return request.localizer.translate(_translation_factory(message, **kwargs)) 94 95 96 def localize(message, **kwargs): 97 """ 98 To be used when we don't have the request context, e.g. 99 `from warehouse.i18n import localize as _` 100 """ 101 102 def _lazy_localize(message, **kwargs): 103 request = get_current_request() 104 return _localize(request, message, **kwargs) 105 106 return LazyString(_lazy_localize, message, **kwargs) 107 108 109 class InvalidLocalizer: 110 def _fail(self): 111 raise RuntimeError("Cannot use localizer without has_translations=True") 112 113 @property 114 def locale_name(self): 115 self._fail() 116 117 def pluralize(self, *args, **kwargs): 118 self._fail() 119 120 def translate(self, *args, **kwargs): 121 self._fail() 122 123 124 def translated_view(view, info): 125 if info.options.get("has_translations"): 126 # If this page can be translated, then we'll add a Vary: PyPI-Locale 127 # Vary header. 128 # Note: This will give weird results if hitting PyPI directly instead of through 129 # the Fastly VCL which sets PyPI-Locale. 130 return add_vary("PyPI-Locale")(view) 131 elif info.exception_only: 132 return view 133 else: 134 # If we're not using translations on this view, then we'll wrap the view 135 # with a wrapper that just ensures that the localizer cannot be used. 136 @functools.wraps(view) 137 def wrapped(context, request): 138 # This whole method is a little bit of an odd duck, we want to make 139 # sure that we don't actually *access* request.localizer, because 140 # doing so triggers the machinery to create a new localizer. So 141 # instead we will dig into the request object __dict__ to 142 # effectively do the same thing, just without triggering an access 143 # on request.localizer. 144 145 # Save the original session so that we can restore it once the 146 # inner views have been called. 147 nothing = object() 148 original_localizer = request.__dict__.get("localizer", nothing) 149 150 # This particular view hasn't been set to allow access to the 151 # translations, so we'll just assign an InvalidLocalizer to 152 # request.localizer 153 request.__dict__["localizer"] = InvalidLocalizer() 154 155 try: 156 # Invoke the real view 157 return view(context, request) 158 finally: 159 # Restore the original session so that things like 160 # pyramid_debugtoolbar can access it. 161 if original_localizer is nothing: 162 del request.__dict__["localizer"] 163 else: 164 request.__dict__["localizer"] = original_localizer 165 166 return wrapped 167 168 169 translated_view.options = {"has_translations"} 170 171 172 def includeme(config): 173 # Add the request attributes 174 config.add_request_method(_locale, name="locale", reify=True) 175 config.add_request_method(_localize, name="_") 176 177 # Register our translation directory. 178 config.add_translation_dirs("warehouse:locale/") 179 180 config.set_locale_negotiator(_negotiate_locale) 181 182 # Register our i18n/l10n filters for Jinja2 183 filters = config.get_settings().setdefault("jinja2.filters", {}) 184 filters.setdefault("format_date", "warehouse.i18n.filters:format_date") 185 filters.setdefault("format_datetime", "warehouse.i18n.filters:format_datetime") 186 filters.setdefault( 187 "format_rfc822_datetime", "warehouse.i18n.filters:format_rfc822_datetime" 188 ) 189 filters.setdefault("format_number", "warehouse.i18n.filters:format_number") 190 191 jglobals = config.get_settings().setdefault("jinja2.globals", {}) 192 jglobals.setdefault("KNOWN_LOCALES", "warehouse.i18n:KNOWN_LOCALES") 193 194 config.add_view_deriver( 195 translated_view, over="rendered_view", under=viewderivers.INGRESS 196 ) 197 [end of warehouse/i18n/__init__.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/warehouse/i18n/__init__.py b/warehouse/i18n/__init__.py --- a/warehouse/i18n/__init__.py +++ b/warehouse/i18n/__init__.py @@ -19,19 +19,21 @@ from warehouse.cache.http import add_vary -# Taken from: -# https://github.com/django/django/blob/master/django/conf/locale/__init__.py KNOWN_LOCALES = { - "en": "English", # English - "es": "español", # Spanish - "fr": "français", # French - "ja": "日本語", # Japanese - "pt_BR": "Português Brasileiro", # Brazilian Portugeuse - "uk": "Українська", # Ukrainian - "el": "Ελληνικά", # Greek - "de": "Deutsch", # German - "zh_Hans": "简体中文", # Simplified Chinese - "ru": "Русский", # Russian + identifier: Locale.parse(identifier, sep="_") + for identifier in [ + "en", # English + "es", # Spanish + "fr", # French + "ja", # Japanese + "pt_BR", # Brazilian Portugeuse + "uk", # Ukranian + "el", # Greek + "de", # German + "zh_Hans", # Simplified Chinese + "ru", # Russian + "he", # Hebrew + ] } LOCALE_ATTR = "_LOCALE_" @@ -60,9 +62,9 @@ def _locale(request): """ - Computes a babel.core:Locale() object for this request. + Gets a babel.core:Locale() object for this request. """ - return Locale.parse(request.locale_name, sep="_") + return KNOWN_LOCALES.get(request.locale_name, "en") def _negotiate_locale(request):
{"golden_diff": "diff --git a/warehouse/i18n/__init__.py b/warehouse/i18n/__init__.py\n--- a/warehouse/i18n/__init__.py\n+++ b/warehouse/i18n/__init__.py\n@@ -19,19 +19,21 @@\n \n from warehouse.cache.http import add_vary\n \n-# Taken from:\n-# https://github.com/django/django/blob/master/django/conf/locale/__init__.py\n KNOWN_LOCALES = {\n- \"en\": \"English\", # English\n- \"es\": \"espa\u00f1ol\", # Spanish\n- \"fr\": \"fran\u00e7ais\", # French\n- \"ja\": \"\u65e5\u672c\u8a9e\", # Japanese\n- \"pt_BR\": \"Portugu\u00eas Brasileiro\", # Brazilian Portugeuse\n- \"uk\": \"\u0423\u043a\u0440\u0430\u0457\u043d\u0441\u044c\u043a\u0430\", # Ukrainian\n- \"el\": \"\u0395\u03bb\u03bb\u03b7\u03bd\u03b9\u03ba\u03ac\", # Greek\n- \"de\": \"Deutsch\", # German\n- \"zh_Hans\": \"\u7b80\u4f53\u4e2d\u6587\", # Simplified Chinese\n- \"ru\": \"\u0420\u0443\u0441\u0441\u043a\u0438\u0439\", # Russian\n+ identifier: Locale.parse(identifier, sep=\"_\")\n+ for identifier in [\n+ \"en\", # English\n+ \"es\", # Spanish\n+ \"fr\", # French\n+ \"ja\", # Japanese\n+ \"pt_BR\", # Brazilian Portugeuse\n+ \"uk\", # Ukranian\n+ \"el\", # Greek\n+ \"de\", # German\n+ \"zh_Hans\", # Simplified Chinese\n+ \"ru\", # Russian\n+ \"he\", # Hebrew\n+ ]\n }\n \n LOCALE_ATTR = \"_LOCALE_\"\n@@ -60,9 +62,9 @@\n \n def _locale(request):\n \"\"\"\n- Computes a babel.core:Locale() object for this request.\n+ Gets a babel.core:Locale() object for this request.\n \"\"\"\n- return Locale.parse(request.locale_name, sep=\"_\")\n+ return KNOWN_LOCALES.get(request.locale_name, \"en\")\n \n \n def _negotiate_locale(request):\n", "issue": "Support for properly displaying the site translated to RTL languages\n**What's the problem this feature will solve?**\r\nDisplay PyPI.org propely when translated to RTL languages such as Arabic, Farsi and Hebrew.\r\n\r\n**Describe the solution you'd like**\r\nI can't suggest a specific technical solution, since I'm not familiar with the website's design and implementation.\r\n\r\n**Additional context**\r\nThe context is the recent work on translation into additional languages.\n", "before_files": [{"content": "# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nimport functools\n\nfrom babel.core import Locale\nfrom pyramid import viewderivers\nfrom pyramid.i18n import TranslationStringFactory, default_locale_negotiator\nfrom pyramid.threadlocal import get_current_request\n\nfrom warehouse.cache.http import add_vary\n\n# Taken from:\n# https://github.com/django/django/blob/master/django/conf/locale/__init__.py\nKNOWN_LOCALES = {\n \"en\": \"English\", # English\n \"es\": \"espa\u00f1ol\", # Spanish\n \"fr\": \"fran\u00e7ais\", # French\n \"ja\": \"\u65e5\u672c\u8a9e\", # Japanese\n \"pt_BR\": \"Portugu\u00eas Brasileiro\", # Brazilian Portugeuse\n \"uk\": \"\u0423\u043a\u0440\u0430\u0457\u043d\u0441\u044c\u043a\u0430\", # Ukrainian\n \"el\": \"\u0395\u03bb\u03bb\u03b7\u03bd\u03b9\u03ba\u03ac\", # Greek\n \"de\": \"Deutsch\", # German\n \"zh_Hans\": \"\u7b80\u4f53\u4e2d\u6587\", # Simplified Chinese\n \"ru\": \"\u0420\u0443\u0441\u0441\u043a\u0438\u0439\", # Russian\n}\n\nLOCALE_ATTR = \"_LOCALE_\"\n\n_translation_factory = TranslationStringFactory(\"messages\")\n\n\nclass LazyString:\n def __init__(self, fn, *args, **kwargs):\n self.fn = fn\n self.args = args\n self.mapping = kwargs.get(\"mapping\", {})\n self.kwargs = kwargs\n\n def __json__(self, request):\n return str(self)\n\n def __mod__(self, new_mapping):\n mapping = self.mapping.copy()\n mapping.update(new_mapping)\n return LazyString(self.fn, *self.args, mapping=new_mapping, **self.kwargs)\n\n def __str__(self):\n return self.fn(*self.args, **self.kwargs)\n\n\ndef _locale(request):\n \"\"\"\n Computes a babel.core:Locale() object for this request.\n \"\"\"\n return Locale.parse(request.locale_name, sep=\"_\")\n\n\ndef _negotiate_locale(request):\n locale_name = getattr(request, LOCALE_ATTR, None)\n if locale_name is not None:\n return locale_name\n\n locale_name = request.params.get(LOCALE_ATTR)\n if locale_name is not None:\n return locale_name\n\n locale_name = request.cookies.get(LOCALE_ATTR)\n if locale_name is not None:\n return locale_name\n\n if not request.accept_language:\n return default_locale_negotiator(request)\n\n return request.accept_language.best_match(\n tuple(KNOWN_LOCALES.keys()), default_match=default_locale_negotiator(request)\n )\n\n\ndef _localize(request, message, **kwargs):\n \"\"\"\n To be used on the request directly, e.g. `request._(message)`\n \"\"\"\n return request.localizer.translate(_translation_factory(message, **kwargs))\n\n\ndef localize(message, **kwargs):\n \"\"\"\n To be used when we don't have the request context, e.g.\n `from warehouse.i18n import localize as _`\n \"\"\"\n\n def _lazy_localize(message, **kwargs):\n request = get_current_request()\n return _localize(request, message, **kwargs)\n\n return LazyString(_lazy_localize, message, **kwargs)\n\n\nclass InvalidLocalizer:\n def _fail(self):\n raise RuntimeError(\"Cannot use localizer without has_translations=True\")\n\n @property\n def locale_name(self):\n self._fail()\n\n def pluralize(self, *args, **kwargs):\n self._fail()\n\n def translate(self, *args, **kwargs):\n self._fail()\n\n\ndef translated_view(view, info):\n if info.options.get(\"has_translations\"):\n # If this page can be translated, then we'll add a Vary: PyPI-Locale\n # Vary header.\n # Note: This will give weird results if hitting PyPI directly instead of through\n # the Fastly VCL which sets PyPI-Locale.\n return add_vary(\"PyPI-Locale\")(view)\n elif info.exception_only:\n return view\n else:\n # If we're not using translations on this view, then we'll wrap the view\n # with a wrapper that just ensures that the localizer cannot be used.\n @functools.wraps(view)\n def wrapped(context, request):\n # This whole method is a little bit of an odd duck, we want to make\n # sure that we don't actually *access* request.localizer, because\n # doing so triggers the machinery to create a new localizer. So\n # instead we will dig into the request object __dict__ to\n # effectively do the same thing, just without triggering an access\n # on request.localizer.\n\n # Save the original session so that we can restore it once the\n # inner views have been called.\n nothing = object()\n original_localizer = request.__dict__.get(\"localizer\", nothing)\n\n # This particular view hasn't been set to allow access to the\n # translations, so we'll just assign an InvalidLocalizer to\n # request.localizer\n request.__dict__[\"localizer\"] = InvalidLocalizer()\n\n try:\n # Invoke the real view\n return view(context, request)\n finally:\n # Restore the original session so that things like\n # pyramid_debugtoolbar can access it.\n if original_localizer is nothing:\n del request.__dict__[\"localizer\"]\n else:\n request.__dict__[\"localizer\"] = original_localizer\n\n return wrapped\n\n\ntranslated_view.options = {\"has_translations\"}\n\n\ndef includeme(config):\n # Add the request attributes\n config.add_request_method(_locale, name=\"locale\", reify=True)\n config.add_request_method(_localize, name=\"_\")\n\n # Register our translation directory.\n config.add_translation_dirs(\"warehouse:locale/\")\n\n config.set_locale_negotiator(_negotiate_locale)\n\n # Register our i18n/l10n filters for Jinja2\n filters = config.get_settings().setdefault(\"jinja2.filters\", {})\n filters.setdefault(\"format_date\", \"warehouse.i18n.filters:format_date\")\n filters.setdefault(\"format_datetime\", \"warehouse.i18n.filters:format_datetime\")\n filters.setdefault(\n \"format_rfc822_datetime\", \"warehouse.i18n.filters:format_rfc822_datetime\"\n )\n filters.setdefault(\"format_number\", \"warehouse.i18n.filters:format_number\")\n\n jglobals = config.get_settings().setdefault(\"jinja2.globals\", {})\n jglobals.setdefault(\"KNOWN_LOCALES\", \"warehouse.i18n:KNOWN_LOCALES\")\n\n config.add_view_deriver(\n translated_view, over=\"rendered_view\", under=viewderivers.INGRESS\n )\n", "path": "warehouse/i18n/__init__.py"}]}
2,693
492
gh_patches_debug_13887
rasdani/github-patches
git_diff
apluslms__a-plus-474
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> The exercise results page does not show that a module is open for reading while the exercises are still closed. The UI leaves the impression that the module is completely closed (between the reading opening time and the opening time). The "opens" badge in the module box also only shows the normal open time. The UI should show clearly that a module is open for reading. The collapsing box should be opened by default for read-open modules. </issue> <code> [start of exercise/templatetags/exercise.py] 1 import json 2 from django import template 3 from django.db.models import Max, Min 4 from django.template.loader import render_to_string 5 from django.utils import timezone 6 from django.utils.translation import ugettext_lazy as _ 7 8 from course.models import CourseModule 9 from lib.errors import TagUsageError 10 from ..cache.content import CachedContent 11 from ..cache.points import CachedPoints 12 from ..exercise_summary import UserExerciseSummary 13 from ..models import LearningObjectDisplay, LearningObject, Submission, BaseExercise 14 15 16 register = template.Library() 17 18 19 def _prepare_now(context): 20 if not 'now' in context: 21 context['now'] = timezone.now() 22 return context['now'] 23 24 25 def _prepare_context(context, student=None): 26 if not 'instance' in context: 27 raise TagUsageError() 28 instance = context['instance'] 29 _prepare_now(context) 30 if not 'content' in context: 31 context['content'] = CachedContent(instance) 32 def points(user, key): 33 if not key in context: 34 context[key] = CachedPoints(instance, user, context['content']) 35 return context[key] 36 if student: 37 return points(student, 'studentpoints') 38 return points(context['request'].user, 'points') 39 40 41 def _get_toc(context, student=None): 42 points = _prepare_context(context, student) 43 context = context.flatten() 44 context.update({ 45 'modules': points.modules_flatted(), 46 'categories': points.categories(), 47 'total': points.total(), 48 'is_course_staff': context.get('is_course_staff', False), 49 }) 50 return context 51 52 53 def _is_accessible(context, entry, t): 54 if t and t > _prepare_now(context): 55 return False 56 if entry.get('requirements'): 57 points = _prepare_context(context) 58 module = CourseModule.objects.get(id=entry['id']) 59 return module.are_requirements_passed(points) 60 return True 61 62 63 @register.inclusion_tag("exercise/_user_results.html", takes_context=True) 64 def user_results(context, student=None): 65 values = _get_toc(context, student) 66 values['total_json'] = json.dumps(values['total']) 67 if student: 68 values['is_course_staff'] = False 69 return values 70 71 72 @register.inclusion_tag("exercise/_user_toc.html", takes_context=True) 73 def user_toc(context, student=None): 74 return _get_toc(context, student) 75 76 77 @register.inclusion_tag("exercise/_user_last.html", takes_context=True) 78 def user_last(context): 79 user = context['request'].user 80 points = _prepare_context(context) 81 if user.is_authenticated: 82 last = LearningObjectDisplay.objects.filter( 83 profile=user.userprofile, 84 learning_object__status=LearningObject.STATUS.READY, 85 learning_object__course_module__course_instance=context['instance'], 86 ).select_related('learning_object').order_by('-timestamp').first() 87 if last: 88 entry,_,_,_ = points.find(last.learning_object) 89 return { 90 'last': entry, 91 'last_time': last.timestamp, 92 } 93 return { 94 'begin': points.begin(), 95 'instance': context['instance'], 96 } 97 98 99 @register.inclusion_tag("exercise/_category_points.html", takes_context=True) 100 def category_points(context, student=None): 101 return _get_toc(context, student) 102 103 104 @register.inclusion_tag("exercise/_submission_list.html", takes_context=True) 105 def latest_submissions(context): 106 submissions = context["profile"].submissions \ 107 .filter(exercise__course_module__course_instance=context["instance"]) \ 108 .order_by("-id")[:10] 109 return { 110 "submissions": submissions, 111 "title": _("Latest submissions"), 112 "empty": _("No submissions for this course."), 113 } 114 115 116 @register.filter 117 def max_submissions(exercise, user_profile): 118 return exercise.max_submissions_for_student(user_profile) 119 120 121 @register.filter 122 def percent(decimal): 123 return int(decimal * 100) 124 125 126 @register.filter 127 def submission_status(status): 128 return Submission.STATUS[status] 129 130 131 def _points_data(obj, classes=None): 132 if isinstance(obj, UserExerciseSummary): 133 exercise = obj.exercise 134 data = { 135 'points': obj.get_points(), 136 'max': exercise.max_points, 137 'difficulty': exercise.difficulty, 138 'required': exercise.points_to_pass, 139 'confirm_the_level': exercise.category.confirm_the_level, 140 'missing_points': obj.is_missing_points(), 141 'passed': obj.is_passed(), 142 'full_score': obj.is_full_points(), 143 'submitted': obj.is_submitted(), 144 'graded': obj.is_graded(), 145 'official': not obj.is_unofficial(), 146 'exercise_page': True, 147 } 148 elif isinstance(obj, Submission): 149 exercise = obj.exercise 150 data = { 151 'points': obj.grade, 152 'max': exercise.max_points, 153 'difficulty': exercise.difficulty, 154 'required': exercise.points_to_pass, 155 'confirm_the_level': exercise.category.confirm_the_level, 156 'missing_points': obj.grade < exercise.points_to_pass, 157 'passed': obj.grade >= exercise.points_to_pass, 158 'full_score': obj.grade >= exercise.max_points, 159 'submitted': True, 160 'graded': obj.is_graded, 161 'official': obj.status != Submission.STATUS.UNOFFICIAL, 162 } 163 if not obj.is_graded and ( 164 not exercise.category.confirm_the_level 165 or obj.status != Submission.STATUS.WAITING 166 ): 167 data['status'] = obj.status 168 else: 169 points = obj.get('points', 0) 170 max_points = obj.get('max_points', 0) 171 required = obj.get('points_to_pass', 0) 172 data = { 173 'points': points, 174 'max': max_points, 175 'difficulty': obj.get('difficulty', ''), 176 'required': required, 177 'confirm_the_level': obj.get('confirm_the_level', False), 178 'missing_points': points < required, 179 'passed': obj.get('passed', True), 180 'full_score': points >= max_points, 181 'submitted': obj.get('submission_count', 0) > 0, 182 'graded': obj.get('graded', True), 183 'status': obj.get('submission_status', False), 184 'unconfirmed': obj.get('unconfirmed', False), 185 'official': not obj.get('unofficial', False), 186 'confirmable_points': obj.get('confirmable_points', False), 187 } 188 percentage = 0 189 required_percentage = None 190 if data['max'] > 0: 191 percentage = int(round(100.0 * data['points'] / data['max'])) 192 if data['required']: 193 required_percentage = int(round(100.0 * data['required'] / data['max'])) 194 data.update({ 195 'classes': classes, 196 'percentage': percentage, 197 'required_percentage': required_percentage, 198 }) 199 return data 200 201 202 @register.inclusion_tag("exercise/_points_progress.html") 203 def points_progress(obj): 204 return _points_data(obj) 205 206 207 @register.inclusion_tag("exercise/_points_badge.html") 208 def points_badge(obj, classes=None): 209 return _points_data(obj, classes) 210 211 212 @register.simple_tag(takes_context=True) 213 def max_group_size(context): 214 points = _prepare_context(context) 215 return points.total()['max_group_size'] 216 217 218 @register.simple_tag(takes_context=True) 219 def min_group_size(context): 220 points = _prepare_context(context) 221 return points.total()['min_group_size'] 222 223 224 @register.simple_tag(takes_context=True) 225 def module_accessible(context, entry): 226 t = entry.get('reading_opening_time', entry.get('opening_time')) 227 return _is_accessible(context, entry, t) 228 229 230 @register.simple_tag(takes_context=True) 231 def exercise_accessible(context, entry): 232 t = entry.get('opening_time') 233 return _is_accessible(context, entry, t) 234 235 236 @register.simple_tag 237 def get_grading_errors(submission): 238 if not isinstance(submission.grading_data, dict): 239 return "" 240 grading_data = submission.grading_data.get('grading_data') 241 if not isinstance(grading_data, str): 242 return "" 243 if grading_data.startswith('<pre>'): 244 return grading_data[5:-6] 245 try: 246 return json.loads(grading_data).get('errors', "") 247 except (AttributeError, TypeError, ValueError): 248 return "" 249 250 251 @register.inclusion_tag("exercise/_text_stats.html", takes_context=True) 252 def exercise_text_stats(context, exercise): 253 if not 'instance' in context: 254 raise TagUsageError() 255 instance = context['instance'] 256 257 if not 'student_count' in context: 258 context['student_count'] = instance.students.count() 259 total = context['student_count'] 260 261 if isinstance(exercise, int): 262 num = instance.students.filter(submissions__exercise_id=exercise).distinct().count() 263 else: 264 num = exercise.number_of_submitters() if exercise else 0 265 return { 266 "number": num, 267 "percentage": int(100 * num / total) if total else 0, 268 } 269 270 271 @register.simple_tag 272 def get_format_info(format): 273 format_infos = { 274 'json' : { 275 'name': 'json', 276 'verbose_name': 'JSON', 277 }, 278 'csv': { 279 'name': 'csv', 280 'verbose_name': 'CSV', 281 }, 282 'excel.csv': { 283 'name': 'excel.csv', 284 'verbose_name': _('Excel compatible CSV'), 285 }, 286 } 287 try: 288 return format_infos[format] 289 except KeyError as e: 290 raise RuntimeError('Invalid format: \'{}\''.format(format)) from e 291 292 293 @register.simple_tag 294 def get_format_info_list(formats): 295 return [get_format_info(format) for format in formats.split()] 296 297 [end of exercise/templatetags/exercise.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/exercise/templatetags/exercise.py b/exercise/templatetags/exercise.py --- a/exercise/templatetags/exercise.py +++ b/exercise/templatetags/exercise.py @@ -223,8 +223,11 @@ @register.simple_tag(takes_context=True) def module_accessible(context, entry): - t = entry.get('reading_opening_time', entry.get('opening_time')) - return _is_accessible(context, entry, t) + t = entry.get('reading_opening_time') + if t: + return _is_accessible(context, entry, t) + else: + return exercise_accessible(context, entry) @register.simple_tag(takes_context=True) @@ -293,4 +296,3 @@ @register.simple_tag def get_format_info_list(formats): return [get_format_info(format) for format in formats.split()] -
{"golden_diff": "diff --git a/exercise/templatetags/exercise.py b/exercise/templatetags/exercise.py\n--- a/exercise/templatetags/exercise.py\n+++ b/exercise/templatetags/exercise.py\n@@ -223,8 +223,11 @@\n \n @register.simple_tag(takes_context=True)\n def module_accessible(context, entry):\n- t = entry.get('reading_opening_time', entry.get('opening_time'))\n- return _is_accessible(context, entry, t)\n+ t = entry.get('reading_opening_time')\n+ if t:\n+ return _is_accessible(context, entry, t)\n+ else:\n+ return exercise_accessible(context, entry)\n \n \n @register.simple_tag(takes_context=True)\n@@ -293,4 +296,3 @@\n @register.simple_tag\n def get_format_info_list(formats):\n return [get_format_info(format) for format in formats.split()]\n-\n", "issue": "The exercise results page does not show that a module is open for reading while the exercises are still closed.\nThe UI leaves the impression that the module is completely closed (between the reading opening time and the opening time). The \"opens\" badge in the module box also only shows the normal open time. The UI should show clearly that a module is open for reading. The collapsing box should be opened by default for read-open modules.\n", "before_files": [{"content": "import json\nfrom django import template\nfrom django.db.models import Max, Min\nfrom django.template.loader import render_to_string\nfrom django.utils import timezone\nfrom django.utils.translation import ugettext_lazy as _\n\nfrom course.models import CourseModule\nfrom lib.errors import TagUsageError\nfrom ..cache.content import CachedContent\nfrom ..cache.points import CachedPoints\nfrom ..exercise_summary import UserExerciseSummary\nfrom ..models import LearningObjectDisplay, LearningObject, Submission, BaseExercise\n\n\nregister = template.Library()\n\n\ndef _prepare_now(context):\n if not 'now' in context:\n context['now'] = timezone.now()\n return context['now']\n\n\ndef _prepare_context(context, student=None):\n if not 'instance' in context:\n raise TagUsageError()\n instance = context['instance']\n _prepare_now(context)\n if not 'content' in context:\n context['content'] = CachedContent(instance)\n def points(user, key):\n if not key in context:\n context[key] = CachedPoints(instance, user, context['content'])\n return context[key]\n if student:\n return points(student, 'studentpoints')\n return points(context['request'].user, 'points')\n\n\ndef _get_toc(context, student=None):\n points = _prepare_context(context, student)\n context = context.flatten()\n context.update({\n 'modules': points.modules_flatted(),\n 'categories': points.categories(),\n 'total': points.total(),\n 'is_course_staff': context.get('is_course_staff', False),\n })\n return context\n\n\ndef _is_accessible(context, entry, t):\n if t and t > _prepare_now(context):\n return False\n if entry.get('requirements'):\n points = _prepare_context(context)\n module = CourseModule.objects.get(id=entry['id'])\n return module.are_requirements_passed(points)\n return True\n\n\[email protected]_tag(\"exercise/_user_results.html\", takes_context=True)\ndef user_results(context, student=None):\n values = _get_toc(context, student)\n values['total_json'] = json.dumps(values['total'])\n if student:\n values['is_course_staff'] = False\n return values\n\n\[email protected]_tag(\"exercise/_user_toc.html\", takes_context=True)\ndef user_toc(context, student=None):\n return _get_toc(context, student)\n\n\[email protected]_tag(\"exercise/_user_last.html\", takes_context=True)\ndef user_last(context):\n user = context['request'].user\n points = _prepare_context(context)\n if user.is_authenticated:\n last = LearningObjectDisplay.objects.filter(\n profile=user.userprofile,\n learning_object__status=LearningObject.STATUS.READY,\n learning_object__course_module__course_instance=context['instance'],\n ).select_related('learning_object').order_by('-timestamp').first()\n if last:\n entry,_,_,_ = points.find(last.learning_object)\n return {\n 'last': entry,\n 'last_time': last.timestamp,\n }\n return {\n 'begin': points.begin(),\n 'instance': context['instance'],\n }\n\n\[email protected]_tag(\"exercise/_category_points.html\", takes_context=True)\ndef category_points(context, student=None):\n return _get_toc(context, student)\n\n\[email protected]_tag(\"exercise/_submission_list.html\", takes_context=True)\ndef latest_submissions(context):\n submissions = context[\"profile\"].submissions \\\n .filter(exercise__course_module__course_instance=context[\"instance\"]) \\\n .order_by(\"-id\")[:10]\n return {\n \"submissions\": submissions,\n \"title\": _(\"Latest submissions\"),\n \"empty\": _(\"No submissions for this course.\"),\n }\n\n\[email protected]\ndef max_submissions(exercise, user_profile):\n return exercise.max_submissions_for_student(user_profile)\n\n\[email protected]\ndef percent(decimal):\n return int(decimal * 100)\n\n\[email protected]\ndef submission_status(status):\n return Submission.STATUS[status]\n\n\ndef _points_data(obj, classes=None):\n if isinstance(obj, UserExerciseSummary):\n exercise = obj.exercise\n data = {\n 'points': obj.get_points(),\n 'max': exercise.max_points,\n 'difficulty': exercise.difficulty,\n 'required': exercise.points_to_pass,\n 'confirm_the_level': exercise.category.confirm_the_level,\n 'missing_points': obj.is_missing_points(),\n 'passed': obj.is_passed(),\n 'full_score': obj.is_full_points(),\n 'submitted': obj.is_submitted(),\n 'graded': obj.is_graded(),\n 'official': not obj.is_unofficial(),\n 'exercise_page': True,\n }\n elif isinstance(obj, Submission):\n exercise = obj.exercise\n data = {\n 'points': obj.grade,\n 'max': exercise.max_points,\n 'difficulty': exercise.difficulty,\n 'required': exercise.points_to_pass,\n 'confirm_the_level': exercise.category.confirm_the_level,\n 'missing_points': obj.grade < exercise.points_to_pass,\n 'passed': obj.grade >= exercise.points_to_pass,\n 'full_score': obj.grade >= exercise.max_points,\n 'submitted': True,\n 'graded': obj.is_graded,\n 'official': obj.status != Submission.STATUS.UNOFFICIAL,\n }\n if not obj.is_graded and (\n not exercise.category.confirm_the_level\n or obj.status != Submission.STATUS.WAITING\n ):\n data['status'] = obj.status\n else:\n points = obj.get('points', 0)\n max_points = obj.get('max_points', 0)\n required = obj.get('points_to_pass', 0)\n data = {\n 'points': points,\n 'max': max_points,\n 'difficulty': obj.get('difficulty', ''),\n 'required': required,\n 'confirm_the_level': obj.get('confirm_the_level', False),\n 'missing_points': points < required,\n 'passed': obj.get('passed', True),\n 'full_score': points >= max_points,\n 'submitted': obj.get('submission_count', 0) > 0,\n 'graded': obj.get('graded', True),\n 'status': obj.get('submission_status', False),\n 'unconfirmed': obj.get('unconfirmed', False),\n 'official': not obj.get('unofficial', False),\n 'confirmable_points': obj.get('confirmable_points', False),\n }\n percentage = 0\n required_percentage = None\n if data['max'] > 0:\n percentage = int(round(100.0 * data['points'] / data['max']))\n if data['required']:\n required_percentage = int(round(100.0 * data['required'] / data['max']))\n data.update({\n 'classes': classes,\n 'percentage': percentage,\n 'required_percentage': required_percentage,\n })\n return data\n\n\[email protected]_tag(\"exercise/_points_progress.html\")\ndef points_progress(obj):\n return _points_data(obj)\n\n\[email protected]_tag(\"exercise/_points_badge.html\")\ndef points_badge(obj, classes=None):\n return _points_data(obj, classes)\n\n\[email protected]_tag(takes_context=True)\ndef max_group_size(context):\n points = _prepare_context(context)\n return points.total()['max_group_size']\n\n\[email protected]_tag(takes_context=True)\ndef min_group_size(context):\n points = _prepare_context(context)\n return points.total()['min_group_size']\n\n\[email protected]_tag(takes_context=True)\ndef module_accessible(context, entry):\n t = entry.get('reading_opening_time', entry.get('opening_time'))\n return _is_accessible(context, entry, t)\n\n\[email protected]_tag(takes_context=True)\ndef exercise_accessible(context, entry):\n t = entry.get('opening_time')\n return _is_accessible(context, entry, t)\n\n\[email protected]_tag\ndef get_grading_errors(submission):\n if not isinstance(submission.grading_data, dict):\n return \"\"\n grading_data = submission.grading_data.get('grading_data')\n if not isinstance(grading_data, str):\n return \"\"\n if grading_data.startswith('<pre>'):\n return grading_data[5:-6]\n try:\n return json.loads(grading_data).get('errors', \"\")\n except (AttributeError, TypeError, ValueError):\n return \"\"\n\n\[email protected]_tag(\"exercise/_text_stats.html\", takes_context=True)\ndef exercise_text_stats(context, exercise):\n if not 'instance' in context:\n raise TagUsageError()\n instance = context['instance']\n\n if not 'student_count' in context:\n context['student_count'] = instance.students.count()\n total = context['student_count']\n\n if isinstance(exercise, int):\n num = instance.students.filter(submissions__exercise_id=exercise).distinct().count()\n else:\n num = exercise.number_of_submitters() if exercise else 0\n return {\n \"number\": num,\n \"percentage\": int(100 * num / total) if total else 0,\n }\n\n\[email protected]_tag\ndef get_format_info(format):\n format_infos = {\n 'json' : {\n 'name': 'json',\n 'verbose_name': 'JSON',\n },\n 'csv': {\n 'name': 'csv',\n 'verbose_name': 'CSV',\n },\n 'excel.csv': {\n 'name': 'excel.csv',\n 'verbose_name': _('Excel compatible CSV'),\n },\n }\n try:\n return format_infos[format]\n except KeyError as e:\n raise RuntimeError('Invalid format: \\'{}\\''.format(format)) from e\n\n\[email protected]_tag\ndef get_format_info_list(formats):\n return [get_format_info(format) for format in formats.split()]\n\n", "path": "exercise/templatetags/exercise.py"}]}
3,499
208
gh_patches_debug_35
rasdani/github-patches
git_diff
StackStorm__st2-5104
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> Add version string to st2tests to make it installable Prior to this change, this will fail: cd st2tests/st2tests pip install . After this change that command successfully installs the `st2tests` package. This will also work for installing via GitHub as in: pip install -e git+https://github.com/StackStorm/[email protected]#egg=st2tests&subdirectory=st2tests The original request in #2574 is to get st2tests onto PyPI, and I'm not sure if this will accomplish that request, but this is a good first step. </issue> <code> [start of st2tests/st2tests/__init__.py] 1 # Copyright 2020 The StackStorm Authors. 2 # Copyright 2019 Extreme Networks, Inc. 3 # 4 # Licensed under the Apache License, Version 2.0 (the "License"); 5 # you may not use this file except in compliance with the License. 6 # You may obtain a copy of the License at 7 # 8 # http://www.apache.org/licenses/LICENSE-2.0 9 # 10 # Unless required by applicable law or agreed to in writing, software 11 # distributed under the License is distributed on an "AS IS" BASIS, 12 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 13 # See the License for the specific language governing permissions and 14 # limitations under the License. 15 16 from __future__ import absolute_import 17 18 from st2tests.base import EventletTestCase 19 from st2tests.base import DbTestCase 20 from st2tests.base import ExecutionDbTestCase 21 from st2tests.base import DbModelTestCase 22 from st2tests.base import WorkflowTestCase 23 24 25 __all__ = [ 26 'EventletTestCase', 27 'DbTestCase', 28 'ExecutionDbTestCase', 29 'DbModelTestCase', 30 'WorkflowTestCase' 31 ] 32 [end of st2tests/st2tests/__init__.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/st2tests/st2tests/__init__.py b/st2tests/st2tests/__init__.py --- a/st2tests/st2tests/__init__.py +++ b/st2tests/st2tests/__init__.py @@ -29,3 +29,5 @@ 'DbModelTestCase', 'WorkflowTestCase' ] + +__version__ = '3.3dev'
{"golden_diff": "diff --git a/st2tests/st2tests/__init__.py b/st2tests/st2tests/__init__.py\n--- a/st2tests/st2tests/__init__.py\n+++ b/st2tests/st2tests/__init__.py\n@@ -29,3 +29,5 @@\n 'DbModelTestCase',\n 'WorkflowTestCase'\n ]\n+\n+__version__ = '3.3dev'\n", "issue": "Add version string to st2tests to make it installable\nPrior to this change, this will fail:\r\n\r\n cd st2tests/st2tests\r\n pip install .\r\n\r\nAfter this change that command successfully installs the `st2tests` package. This will also work for installing via GitHub as in:\r\n\r\n pip install -e git+https://github.com/StackStorm/[email protected]#egg=st2tests&subdirectory=st2tests\r\n\r\nThe original request in #2574 is to get st2tests onto PyPI, and I'm not sure if this will accomplish that request, but this is a good first step.\n", "before_files": [{"content": "# Copyright 2020 The StackStorm Authors.\n# Copyright 2019 Extreme Networks, Inc.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom __future__ import absolute_import\n\nfrom st2tests.base import EventletTestCase\nfrom st2tests.base import DbTestCase\nfrom st2tests.base import ExecutionDbTestCase\nfrom st2tests.base import DbModelTestCase\nfrom st2tests.base import WorkflowTestCase\n\n\n__all__ = [\n 'EventletTestCase',\n 'DbTestCase',\n 'ExecutionDbTestCase',\n 'DbModelTestCase',\n 'WorkflowTestCase'\n]\n", "path": "st2tests/st2tests/__init__.py"}]}
977
88
gh_patches_debug_24126
rasdani/github-patches
git_diff
kartoza__prj.app-279
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> Sponsors for a release need to be filtered for the project and the date There is a serious bug with the sponsors whereby if you set the release date for a release in one project, it will show on the release page sponsors from any other project. <img width="1387" alt="screen shot 2016-03-31 at 13 36 43" src="https://cloud.githubusercontent.com/assets/178003/14167536/1eff7c90-f746-11e5-9a2e-ed648c9d22d6.png"> Note that you need to have a date for the version otherwise it does not replicate: <img width="1210" alt="screen shot 2016-03-31 at 13 37 21" src="https://cloud.githubusercontent.com/assets/178003/14167565/50e68c58-f746-11e5-96b6-e59970bf6c14.png"> </issue> <code> [start of django_project/changes/models/version.py] 1 # coding=utf-8 2 from django.core.urlresolvers import reverse 3 # from django.utils.text import slugify 4 from common.utilities import version_slugify 5 import os 6 import logging 7 from core.settings.contrib import STOP_WORDS 8 from django.conf.global_settings import MEDIA_ROOT 9 from django.db import models 10 from .entry import Entry 11 from .sponsorship_period import SponsorshipPeriod 12 from django.contrib.auth.models import User 13 from django.utils.translation import ugettext_lazy as _ 14 15 logger = logging.getLogger(__name__) 16 17 18 class ApprovedVersionManager(models.Manager): 19 """Custom version manager that shows only approved records.""" 20 21 def get_queryset(self): 22 """Query set generator""" 23 return super( 24 ApprovedVersionManager, self).get_queryset().filter( 25 approved=True) 26 27 28 class UnapprovedVersionManager(models.Manager): 29 """Custom version manager that shows only unapproved records.""" 30 31 def get_queryset(self): 32 """Query set generator""" 33 return super( 34 UnapprovedVersionManager, self).get_queryset().filter( 35 approved=False) 36 37 38 # noinspection PyUnresolvedReferences 39 class Version(models.Model): 40 """A version model that the changelog is associated with..""" 41 42 name = models.CharField( 43 help_text='Name of this release e.g. 1.0.1.', 44 max_length=255, 45 null=False, 46 blank=False, 47 unique=False) 48 49 padded_version = models.CharField( 50 help_text=( 51 'Numeric version for this release e.g. 001000001 for 1.0.1 ' 52 'calculated by zero padding each component of maj/minor/bugfix ' 53 'elements from name.'), 54 max_length=9, 55 null=False, 56 blank=True, 57 unique=False) 58 59 approved = models.BooleanField( 60 help_text=( 61 'Whether this version has been approved for use by the ' 62 'project owner.'), 63 default=False) 64 65 image_file = models.ImageField( 66 help_text=( 67 'An optional image for this version e.g. a splashscreen. ' 68 'Most browsers support dragging the image directly on to the ' 69 '"Choose File" button above.'), 70 upload_to=os.path.join(MEDIA_ROOT, 'images/projects'), 71 blank=True) 72 73 description = models.TextField( 74 null=True, 75 blank=True, 76 help_text='Describe the new version. Markdown is supported.') 77 78 release_date = models.DateField( 79 _('Release date (yyyy-mm-dd)'), 80 help_text='Date of official release', 81 null=True, 82 blank=True) 83 84 author = models.ForeignKey(User) 85 slug = models.SlugField() 86 project = models.ForeignKey('base.Project') 87 objects = models.Manager() 88 approved_objects = ApprovedVersionManager() 89 unapproved_objects = UnapprovedVersionManager() 90 91 # noinspection PyClassicStyleClass 92 class Meta: 93 """Meta options for the version class.""" 94 unique_together = ( 95 ('name', 'project'), 96 ('slug', 'project'), 97 ) 98 app_label = 'changes' 99 # ordering = ['-datetime_created'] 100 101 def save(self, *args, **kwargs): 102 if not self.pk: 103 words = self.name.split() 104 filtered_words = [t for t in words if t.lower() not in STOP_WORDS] 105 new_list = ' '.join(filtered_words) 106 self.slug = version_slugify(new_list)[:50] 107 self.padded_version = self.pad_name(self.name) 108 super(Version, self).save(*args, **kwargs) 109 110 def pad_name(self, version): 111 """Create a 0 padded version of the version name. 112 113 e.g. input: 2.10.1 114 e.g. output: 002010100 115 116 This will ensure we have sortable version names. 117 118 :param version: A text version in the form 0.0.0 - if the version is 119 not in this form, we return the version unaltered. 120 :type version: str 121 122 :returns: Zero padded representation of the version e.g. 001010100 123 :rtype: str 124 125 """ 126 tokens = version.split('.') 127 if len(tokens) != 3: 128 return version 129 result = '' 130 for token in tokens: 131 result += token.zfill(3) 132 return result 133 134 def __unicode__(self): 135 return u'%s : %s' % (self.project.name, self.name) 136 137 def get_absolute_url(self): 138 return reverse('version-detail', kwargs={ 139 'slug': self.slug, 140 'project_slug': self.project.slug 141 }) 142 143 def entries(self): 144 """Get the entries for this version.""" 145 qs = Entry.objects.filter(version=self).order_by('category') 146 return qs 147 148 def _entries_for_category(self, category): 149 """All entries for this version and filtered by the given category. 150 151 :param category: Category to filter by. 152 :type category: Category 153 154 .. note:: only approved entries returned. 155 """ 156 qs = Entry.objects.filter(version=self, 157 category=category, 158 approved=True) 159 return qs 160 161 def categories(self): 162 """Get a list of categories where there are one or more entries. 163 164 Example use in template:: 165 {% for row in version.categories %} 166 <h2 class="text-muted">{{ row.category.name }}</h2> 167 <ul> 168 {% for entry in row.entries %} 169 <li>{{ entry.name }}</li> 170 {% endfor %} 171 </ul> 172 {% endfor %} 173 """ 174 qs = self.entries() 175 used = [] 176 categories = [] 177 for entry in qs: 178 category = entry.category 179 if category not in used: 180 row = { 181 'category': category, 182 'entries': self._entries_for_category(category) 183 } 184 categories.append(row) 185 used.append(category) 186 return categories 187 188 def sponsors(self): 189 """Return a list of sponsors current at time of this version release. 190 191 :returns: A list of SponsorPeriod objects whose release date coincides 192 with the version release date. Only approved sponsors are returned. 193 Returns None if the release date (which is optional) is not set. 194 :rtype: Queryset, None 195 """ 196 if self.release_date is None: 197 return None 198 sponsors = SponsorshipPeriod.approved_objects.filter( 199 end_date__gte=self.release_date).filter( 200 start_date__lte=self.release_date).order_by( 201 'start_date').order_by( 202 '-sponsorship_level__value') 203 return sponsors 204 [end of django_project/changes/models/version.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/django_project/changes/models/version.py b/django_project/changes/models/version.py --- a/django_project/changes/models/version.py +++ b/django_project/changes/models/version.py @@ -188,8 +188,9 @@ def sponsors(self): """Return a list of sponsors current at time of this version release. - :returns: A list of SponsorPeriod objects whose release date coincides - with the version release date. Only approved sponsors are returned. + :returns: A list of SponsorPeriod objects for current project + whose release date coincides with the version release date. + Only approved sponsors are returned. Returns None if the release date (which is optional) is not set. :rtype: Queryset, None """ @@ -197,7 +198,8 @@ return None sponsors = SponsorshipPeriod.approved_objects.filter( end_date__gte=self.release_date).filter( - start_date__lte=self.release_date).order_by( + start_date__lte=self.release_date).filter( + project=self.project).order_by( 'start_date').order_by( '-sponsorship_level__value') return sponsors
{"golden_diff": "diff --git a/django_project/changes/models/version.py b/django_project/changes/models/version.py\n--- a/django_project/changes/models/version.py\n+++ b/django_project/changes/models/version.py\n@@ -188,8 +188,9 @@\n def sponsors(self):\n \"\"\"Return a list of sponsors current at time of this version release.\n \n- :returns: A list of SponsorPeriod objects whose release date coincides\n- with the version release date. Only approved sponsors are returned.\n+ :returns: A list of SponsorPeriod objects for current project\n+ whose release date coincides with the version release date.\n+ Only approved sponsors are returned.\n Returns None if the release date (which is optional) is not set.\n :rtype: Queryset, None\n \"\"\"\n@@ -197,7 +198,8 @@\n return None\n sponsors = SponsorshipPeriod.approved_objects.filter(\n end_date__gte=self.release_date).filter(\n- start_date__lte=self.release_date).order_by(\n+ start_date__lte=self.release_date).filter(\n+ project=self.project).order_by(\n 'start_date').order_by(\n '-sponsorship_level__value')\n return sponsors\n", "issue": "Sponsors for a release need to be filtered for the project and the date\nThere is a serious bug with the sponsors whereby if you set the release date for a release in one project, it will show on the release page sponsors from any other project.\n\n<img width=\"1387\" alt=\"screen shot 2016-03-31 at 13 36 43\" src=\"https://cloud.githubusercontent.com/assets/178003/14167536/1eff7c90-f746-11e5-9a2e-ed648c9d22d6.png\">\n\nNote that you need to have a date for the version otherwise it does not replicate:\n\n<img width=\"1210\" alt=\"screen shot 2016-03-31 at 13 37 21\" src=\"https://cloud.githubusercontent.com/assets/178003/14167565/50e68c58-f746-11e5-96b6-e59970bf6c14.png\">\n\n", "before_files": [{"content": "# coding=utf-8\nfrom django.core.urlresolvers import reverse\n# from django.utils.text import slugify\nfrom common.utilities import version_slugify\nimport os\nimport logging\nfrom core.settings.contrib import STOP_WORDS\nfrom django.conf.global_settings import MEDIA_ROOT\nfrom django.db import models\nfrom .entry import Entry\nfrom .sponsorship_period import SponsorshipPeriod\nfrom django.contrib.auth.models import User\nfrom django.utils.translation import ugettext_lazy as _\n\nlogger = logging.getLogger(__name__)\n\n\nclass ApprovedVersionManager(models.Manager):\n \"\"\"Custom version manager that shows only approved records.\"\"\"\n\n def get_queryset(self):\n \"\"\"Query set generator\"\"\"\n return super(\n ApprovedVersionManager, self).get_queryset().filter(\n approved=True)\n\n\nclass UnapprovedVersionManager(models.Manager):\n \"\"\"Custom version manager that shows only unapproved records.\"\"\"\n\n def get_queryset(self):\n \"\"\"Query set generator\"\"\"\n return super(\n UnapprovedVersionManager, self).get_queryset().filter(\n approved=False)\n\n\n# noinspection PyUnresolvedReferences\nclass Version(models.Model):\n \"\"\"A version model that the changelog is associated with..\"\"\"\n\n name = models.CharField(\n help_text='Name of this release e.g. 1.0.1.',\n max_length=255,\n null=False,\n blank=False,\n unique=False)\n\n padded_version = models.CharField(\n help_text=(\n 'Numeric version for this release e.g. 001000001 for 1.0.1 '\n 'calculated by zero padding each component of maj/minor/bugfix '\n 'elements from name.'),\n max_length=9,\n null=False,\n blank=True,\n unique=False)\n\n approved = models.BooleanField(\n help_text=(\n 'Whether this version has been approved for use by the '\n 'project owner.'),\n default=False)\n\n image_file = models.ImageField(\n help_text=(\n 'An optional image for this version e.g. a splashscreen. '\n 'Most browsers support dragging the image directly on to the '\n '\"Choose File\" button above.'),\n upload_to=os.path.join(MEDIA_ROOT, 'images/projects'),\n blank=True)\n\n description = models.TextField(\n null=True,\n blank=True,\n help_text='Describe the new version. Markdown is supported.')\n\n release_date = models.DateField(\n _('Release date (yyyy-mm-dd)'),\n help_text='Date of official release',\n null=True,\n blank=True)\n\n author = models.ForeignKey(User)\n slug = models.SlugField()\n project = models.ForeignKey('base.Project')\n objects = models.Manager()\n approved_objects = ApprovedVersionManager()\n unapproved_objects = UnapprovedVersionManager()\n\n # noinspection PyClassicStyleClass\n class Meta:\n \"\"\"Meta options for the version class.\"\"\"\n unique_together = (\n ('name', 'project'),\n ('slug', 'project'),\n )\n app_label = 'changes'\n # ordering = ['-datetime_created']\n\n def save(self, *args, **kwargs):\n if not self.pk:\n words = self.name.split()\n filtered_words = [t for t in words if t.lower() not in STOP_WORDS]\n new_list = ' '.join(filtered_words)\n self.slug = version_slugify(new_list)[:50]\n self.padded_version = self.pad_name(self.name)\n super(Version, self).save(*args, **kwargs)\n\n def pad_name(self, version):\n \"\"\"Create a 0 padded version of the version name.\n\n e.g. input: 2.10.1\n e.g. output: 002010100\n\n This will ensure we have sortable version names.\n\n :param version: A text version in the form 0.0.0 - if the version is\n not in this form, we return the version unaltered.\n :type version: str\n\n :returns: Zero padded representation of the version e.g. 001010100\n :rtype: str\n\n \"\"\"\n tokens = version.split('.')\n if len(tokens) != 3:\n return version\n result = ''\n for token in tokens:\n result += token.zfill(3)\n return result\n\n def __unicode__(self):\n return u'%s : %s' % (self.project.name, self.name)\n\n def get_absolute_url(self):\n return reverse('version-detail', kwargs={\n 'slug': self.slug,\n 'project_slug': self.project.slug\n })\n\n def entries(self):\n \"\"\"Get the entries for this version.\"\"\"\n qs = Entry.objects.filter(version=self).order_by('category')\n return qs\n\n def _entries_for_category(self, category):\n \"\"\"All entries for this version and filtered by the given category.\n\n :param category: Category to filter by.\n :type category: Category\n\n .. note:: only approved entries returned.\n \"\"\"\n qs = Entry.objects.filter(version=self,\n category=category,\n approved=True)\n return qs\n\n def categories(self):\n \"\"\"Get a list of categories where there are one or more entries.\n\n Example use in template::\n {% for row in version.categories %}\n <h2 class=\"text-muted\">{{ row.category.name }}</h2>\n <ul>\n {% for entry in row.entries %}\n <li>{{ entry.name }}</li>\n {% endfor %}\n </ul>\n {% endfor %}\n \"\"\"\n qs = self.entries()\n used = []\n categories = []\n for entry in qs:\n category = entry.category\n if category not in used:\n row = {\n 'category': category,\n 'entries': self._entries_for_category(category)\n }\n categories.append(row)\n used.append(category)\n return categories\n\n def sponsors(self):\n \"\"\"Return a list of sponsors current at time of this version release.\n\n :returns: A list of SponsorPeriod objects whose release date coincides\n with the version release date. Only approved sponsors are returned.\n Returns None if the release date (which is optional) is not set.\n :rtype: Queryset, None\n \"\"\"\n if self.release_date is None:\n return None\n sponsors = SponsorshipPeriod.approved_objects.filter(\n end_date__gte=self.release_date).filter(\n start_date__lte=self.release_date).order_by(\n 'start_date').order_by(\n '-sponsorship_level__value')\n return sponsors\n", "path": "django_project/changes/models/version.py"}]}
2,690
268
gh_patches_debug_6928
rasdani/github-patches
git_diff
sanic-org__sanic-2001
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> Hard error on invalid max-age cookie **Describe the bug** Currently when setting the `max-age` cookie value, it's possible for a valid value to not be set as expected, as well as an invalid value from raising a hard error. In both cases the values are replaced by a `0` `max-age`. **Code snippet** ```python response.cookie["my-cookie"]["max-age"] = 10.0 # max-age is set to 0 response.cookie["my-cookie"]["max-age"] = 10.5 # max-age is set to 0 response.cookie["my-cookie"]["max-age"] = "ten" # max-age is set to 0 response.cookie["my-cookie"]["max-age"] = "10" # max-age is set to 10 response.cookie["my-cookie"]["max-age"] = 10 # max-age is set to 10 ``` **Expected behavior** Here's what I think the expected behaviour should be (akin to how the `expires` cookie attribute is handled; raising an error if not a `datetime.datetime`). ```python response.cookie["my-cookie"]["max-age"] = 10.0 # max-age is set to 10 response.cookie["my-cookie"]["max-age"] = 10.5 # raise ValueError response.cookie["my-cookie"]["max-age"] = "ten" # raise ValueError response.cookie["my-cookie"]["max-age"] = "10" # max-age is set to 10 response.cookie["my-cookie"]["max-age"] = 10 # max-age is set to 10 ``` **Environment (please complete the following information):** - OS: macOS - Version 19.12.2 **Additional context** I've created a pull request for this here #1809. Here's the issue relating to the original implementation #1452. Creating this issue so I can have an issue number for the changelog. </issue> <code> [start of sanic/cookies.py] 1 import re 2 import string 3 4 from datetime import datetime 5 6 7 DEFAULT_MAX_AGE = 0 8 9 # ------------------------------------------------------------ # 10 # SimpleCookie 11 # ------------------------------------------------------------ # 12 13 # Straight up copied this section of dark magic from SimpleCookie 14 15 _LegalChars = string.ascii_letters + string.digits + "!#$%&'*+-.^_`|~:" 16 _UnescapedChars = _LegalChars + " ()/<=>?@[]{}" 17 18 _Translator = { 19 n: "\\%03o" % n for n in set(range(256)) - set(map(ord, _UnescapedChars)) 20 } 21 _Translator.update({ord('"'): '\\"', ord("\\"): "\\\\"}) 22 23 24 def _quote(str): 25 r"""Quote a string for use in a cookie header. 26 If the string does not need to be double-quoted, then just return the 27 string. Otherwise, surround the string in doublequotes and quote 28 (with a \) special characters. 29 """ 30 if str is None or _is_legal_key(str): 31 return str 32 else: 33 return '"' + str.translate(_Translator) + '"' 34 35 36 _is_legal_key = re.compile("[%s]+" % re.escape(_LegalChars)).fullmatch 37 38 # ------------------------------------------------------------ # 39 # Custom SimpleCookie 40 # ------------------------------------------------------------ # 41 42 43 class CookieJar(dict): 44 """CookieJar dynamically writes headers as cookies are added and removed 45 It gets around the limitation of one header per name by using the 46 MultiHeader class to provide a unique key that encodes to Set-Cookie. 47 """ 48 49 def __init__(self, headers): 50 super().__init__() 51 self.headers = headers 52 self.cookie_headers = {} 53 self.header_key = "Set-Cookie" 54 55 def __setitem__(self, key, value): 56 # If this cookie doesn't exist, add it to the header keys 57 if not self.cookie_headers.get(key): 58 cookie = Cookie(key, value) 59 cookie["path"] = "/" 60 self.cookie_headers[key] = self.header_key 61 self.headers.add(self.header_key, cookie) 62 return super().__setitem__(key, cookie) 63 else: 64 self[key].value = value 65 66 def __delitem__(self, key): 67 if key not in self.cookie_headers: 68 self[key] = "" 69 self[key]["max-age"] = 0 70 else: 71 cookie_header = self.cookie_headers[key] 72 # remove it from header 73 cookies = self.headers.popall(cookie_header) 74 for cookie in cookies: 75 if cookie.key != key: 76 self.headers.add(cookie_header, cookie) 77 del self.cookie_headers[key] 78 return super().__delitem__(key) 79 80 81 class Cookie(dict): 82 """A stripped down version of Morsel from SimpleCookie #gottagofast""" 83 84 _keys = { 85 "expires": "expires", 86 "path": "Path", 87 "comment": "Comment", 88 "domain": "Domain", 89 "max-age": "Max-Age", 90 "secure": "Secure", 91 "httponly": "HttpOnly", 92 "version": "Version", 93 "samesite": "SameSite", 94 } 95 _flags = {"secure", "httponly"} 96 97 def __init__(self, key, value): 98 if key in self._keys: 99 raise KeyError("Cookie name is a reserved word") 100 if not _is_legal_key(key): 101 raise KeyError("Cookie key contains illegal characters") 102 self.key = key 103 self.value = value 104 super().__init__() 105 106 def __setitem__(self, key, value): 107 if key not in self._keys: 108 raise KeyError("Unknown cookie property") 109 if value is not False: 110 if key.lower() == "max-age": 111 if not str(value).isdigit(): 112 value = DEFAULT_MAX_AGE 113 elif key.lower() == "expires": 114 if not isinstance(value, datetime): 115 raise TypeError( 116 "Cookie 'expires' property must be a datetime" 117 ) 118 return super().__setitem__(key, value) 119 120 def encode(self, encoding): 121 """ 122 Encode the cookie content in a specific type of encoding instructed 123 by the developer. Leverages the :func:`str.encode` method provided 124 by python. 125 126 This method can be used to encode and embed ``utf-8`` content into 127 the cookies. 128 129 :param encoding: Encoding to be used with the cookie 130 :return: Cookie encoded in a codec of choosing. 131 :except: UnicodeEncodeError 132 """ 133 return str(self).encode(encoding) 134 135 def __str__(self): 136 """Format as a Set-Cookie header value.""" 137 output = ["%s=%s" % (self.key, _quote(self.value))] 138 for key, value in self.items(): 139 if key == "max-age": 140 try: 141 output.append("%s=%d" % (self._keys[key], value)) 142 except TypeError: 143 output.append("%s=%s" % (self._keys[key], value)) 144 elif key == "expires": 145 output.append( 146 "%s=%s" 147 % (self._keys[key], value.strftime("%a, %d-%b-%Y %T GMT")) 148 ) 149 elif key in self._flags and self[key]: 150 output.append(self._keys[key]) 151 else: 152 output.append("%s=%s" % (self._keys[key], value)) 153 154 return "; ".join(output) 155 [end of sanic/cookies.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/sanic/cookies.py b/sanic/cookies.py --- a/sanic/cookies.py +++ b/sanic/cookies.py @@ -109,7 +109,7 @@ if value is not False: if key.lower() == "max-age": if not str(value).isdigit(): - value = DEFAULT_MAX_AGE + raise ValueError("Cookie max-age must be an integer") elif key.lower() == "expires": if not isinstance(value, datetime): raise TypeError(
{"golden_diff": "diff --git a/sanic/cookies.py b/sanic/cookies.py\n--- a/sanic/cookies.py\n+++ b/sanic/cookies.py\n@@ -109,7 +109,7 @@\n if value is not False:\n if key.lower() == \"max-age\":\n if not str(value).isdigit():\n- value = DEFAULT_MAX_AGE\n+ raise ValueError(\"Cookie max-age must be an integer\")\n elif key.lower() == \"expires\":\n if not isinstance(value, datetime):\n raise TypeError(\n", "issue": "Hard error on invalid max-age cookie\n**Describe the bug**\r\nCurrently when setting the `max-age` cookie value, it's possible for a valid value to not be set as expected, as well as an invalid value from raising a hard error. In both cases the values are replaced by a `0` `max-age`.\r\n\r\n\r\n**Code snippet**\r\n```python\r\nresponse.cookie[\"my-cookie\"][\"max-age\"] = 10.0 # max-age is set to 0\r\nresponse.cookie[\"my-cookie\"][\"max-age\"] = 10.5 # max-age is set to 0\r\nresponse.cookie[\"my-cookie\"][\"max-age\"] = \"ten\" # max-age is set to 0\r\nresponse.cookie[\"my-cookie\"][\"max-age\"] = \"10\" # max-age is set to 10\r\nresponse.cookie[\"my-cookie\"][\"max-age\"] = 10 # max-age is set to 10\r\n```\r\n\r\n**Expected behavior**\r\nHere's what I think the expected behaviour should be (akin to how the `expires` cookie attribute is handled; raising an error if not a `datetime.datetime`).\r\n\r\n```python\r\nresponse.cookie[\"my-cookie\"][\"max-age\"] = 10.0 # max-age is set to 10\r\nresponse.cookie[\"my-cookie\"][\"max-age\"] = 10.5 # raise ValueError\r\nresponse.cookie[\"my-cookie\"][\"max-age\"] = \"ten\" # raise ValueError\r\nresponse.cookie[\"my-cookie\"][\"max-age\"] = \"10\" # max-age is set to 10\r\nresponse.cookie[\"my-cookie\"][\"max-age\"] = 10 # max-age is set to 10\r\n```\r\n\r\n**Environment (please complete the following information):**\r\n - OS: macOS\r\n - Version 19.12.2\r\n\r\n\r\n**Additional context**\r\nI've created a pull request for this here #1809. Here's the issue relating to the original implementation #1452.\r\n\r\nCreating this issue so I can have an issue number for the changelog.\r\n\r\n\n", "before_files": [{"content": "import re\nimport string\n\nfrom datetime import datetime\n\n\nDEFAULT_MAX_AGE = 0\n\n# ------------------------------------------------------------ #\n# SimpleCookie\n# ------------------------------------------------------------ #\n\n# Straight up copied this section of dark magic from SimpleCookie\n\n_LegalChars = string.ascii_letters + string.digits + \"!#$%&'*+-.^_`|~:\"\n_UnescapedChars = _LegalChars + \" ()/<=>?@[]{}\"\n\n_Translator = {\n n: \"\\\\%03o\" % n for n in set(range(256)) - set(map(ord, _UnescapedChars))\n}\n_Translator.update({ord('\"'): '\\\\\"', ord(\"\\\\\"): \"\\\\\\\\\"})\n\n\ndef _quote(str):\n r\"\"\"Quote a string for use in a cookie header.\n If the string does not need to be double-quoted, then just return the\n string. Otherwise, surround the string in doublequotes and quote\n (with a \\) special characters.\n \"\"\"\n if str is None or _is_legal_key(str):\n return str\n else:\n return '\"' + str.translate(_Translator) + '\"'\n\n\n_is_legal_key = re.compile(\"[%s]+\" % re.escape(_LegalChars)).fullmatch\n\n# ------------------------------------------------------------ #\n# Custom SimpleCookie\n# ------------------------------------------------------------ #\n\n\nclass CookieJar(dict):\n \"\"\"CookieJar dynamically writes headers as cookies are added and removed\n It gets around the limitation of one header per name by using the\n MultiHeader class to provide a unique key that encodes to Set-Cookie.\n \"\"\"\n\n def __init__(self, headers):\n super().__init__()\n self.headers = headers\n self.cookie_headers = {}\n self.header_key = \"Set-Cookie\"\n\n def __setitem__(self, key, value):\n # If this cookie doesn't exist, add it to the header keys\n if not self.cookie_headers.get(key):\n cookie = Cookie(key, value)\n cookie[\"path\"] = \"/\"\n self.cookie_headers[key] = self.header_key\n self.headers.add(self.header_key, cookie)\n return super().__setitem__(key, cookie)\n else:\n self[key].value = value\n\n def __delitem__(self, key):\n if key not in self.cookie_headers:\n self[key] = \"\"\n self[key][\"max-age\"] = 0\n else:\n cookie_header = self.cookie_headers[key]\n # remove it from header\n cookies = self.headers.popall(cookie_header)\n for cookie in cookies:\n if cookie.key != key:\n self.headers.add(cookie_header, cookie)\n del self.cookie_headers[key]\n return super().__delitem__(key)\n\n\nclass Cookie(dict):\n \"\"\"A stripped down version of Morsel from SimpleCookie #gottagofast\"\"\"\n\n _keys = {\n \"expires\": \"expires\",\n \"path\": \"Path\",\n \"comment\": \"Comment\",\n \"domain\": \"Domain\",\n \"max-age\": \"Max-Age\",\n \"secure\": \"Secure\",\n \"httponly\": \"HttpOnly\",\n \"version\": \"Version\",\n \"samesite\": \"SameSite\",\n }\n _flags = {\"secure\", \"httponly\"}\n\n def __init__(self, key, value):\n if key in self._keys:\n raise KeyError(\"Cookie name is a reserved word\")\n if not _is_legal_key(key):\n raise KeyError(\"Cookie key contains illegal characters\")\n self.key = key\n self.value = value\n super().__init__()\n\n def __setitem__(self, key, value):\n if key not in self._keys:\n raise KeyError(\"Unknown cookie property\")\n if value is not False:\n if key.lower() == \"max-age\":\n if not str(value).isdigit():\n value = DEFAULT_MAX_AGE\n elif key.lower() == \"expires\":\n if not isinstance(value, datetime):\n raise TypeError(\n \"Cookie 'expires' property must be a datetime\"\n )\n return super().__setitem__(key, value)\n\n def encode(self, encoding):\n \"\"\"\n Encode the cookie content in a specific type of encoding instructed\n by the developer. Leverages the :func:`str.encode` method provided\n by python.\n\n This method can be used to encode and embed ``utf-8`` content into\n the cookies.\n\n :param encoding: Encoding to be used with the cookie\n :return: Cookie encoded in a codec of choosing.\n :except: UnicodeEncodeError\n \"\"\"\n return str(self).encode(encoding)\n\n def __str__(self):\n \"\"\"Format as a Set-Cookie header value.\"\"\"\n output = [\"%s=%s\" % (self.key, _quote(self.value))]\n for key, value in self.items():\n if key == \"max-age\":\n try:\n output.append(\"%s=%d\" % (self._keys[key], value))\n except TypeError:\n output.append(\"%s=%s\" % (self._keys[key], value))\n elif key == \"expires\":\n output.append(\n \"%s=%s\"\n % (self._keys[key], value.strftime(\"%a, %d-%b-%Y %T GMT\"))\n )\n elif key in self._flags and self[key]:\n output.append(self._keys[key])\n else:\n output.append(\"%s=%s\" % (self._keys[key], value))\n\n return \"; \".join(output)\n", "path": "sanic/cookies.py"}]}
2,481
114
gh_patches_debug_29289
rasdani/github-patches
git_diff
google__openhtf-186
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> Attaching binary file using test.attach raises UnicodeDecodeError If I attach a png or avi I see the following in OutputTestRecord Python2.7/site-packages/openhtf/**init**.py", line 185, in OutputTestRecord output_cb(test_record) File "virtualenv/local/lib/python2.7/site-packages/openhtf/**init**.py", line 83, in **call** f.write(self.encode(as_dict)) File "/usr/lib/python2.7/json/encoder.py", line 209, in encode chunks = list(chunks) File "/usr/lib/python2.7/json/encoder.py", line 434, in _iterencode for chunk in _iterencode_dict(o, _current_indent_level): File "/usr/lib/python2.7/json/encoder.py", line 408, in _iterencode_dict for chunk in chunks: File "/usr/lib/python2.7/json/encoder.py", line 332, in _iterencode_list for chunk in chunks: File "/usr/lib/python2.7/json/encoder.py", line 408, in _iterencode_dict for chunk in chunks: File "/usr/lib/python2.7/json/encoder.py", line 408, in _iterencode_dict for chunk in chunks: File "/usr/lib/python2.7/json/encoder.py", line 390, in _iterencode_dict yield _encoder(value) UnicodeDecodeError: 'utf8' codec can't decode byte 0x89 in position 0: invalid start byte </issue> <code> [start of openhtf/io/output/json_factory.py] 1 """Module for outputting test record to JSON-formatted files.""" 2 3 from json import JSONEncoder 4 5 from openhtf import util 6 from openhtf.exe import test_state 7 8 9 class OutputToJSON(JSONEncoder): 10 """Return an output callback that writes JSON Test Records. 11 12 An example filename_pattern might be: 13 '/data/test_records/%(dut_id)s.%(start_time_millis)s' 14 15 To use this output mechanism: 16 test = openhtf.Test(PhaseOne, PhaseTwo) 17 test.AddOutputCallback(openhtf.OutputToJson( 18 '/data/test_records/%(dut_id)s.%(start_time_millis)s')) 19 20 Args: 21 filename_pattern: A format string specifying the filename to write to, 22 will be formatted with the Test Record as a dictionary. 23 inline_attachments: Whether attachments should be included inline in the 24 output. Set to False if you expect to have large binary attachments. 25 """ 26 27 def __init__(self, filename_pattern=None, inline_attachments=True, **kwargs): 28 super(OutputToJSON, self).__init__(**kwargs) 29 self.filename_pattern = filename_pattern 30 self.inline_attachments = inline_attachments 31 32 def default(self, obj): 33 if isinstance(obj, BaseException): 34 # Just repr exceptions. 35 return repr(obj) 36 return super(OutputToJSON, self).default(obj) 37 38 # pylint: disable=invalid-name 39 def __call__(self, test_record): 40 assert self.filename_pattern, 'filename_pattern required' 41 if self.inline_attachments: 42 as_dict = util.ConvertToBaseTypes(test_record) 43 else: 44 as_dict = util.ConvertToBaseTypes(test_record, ignore_keys='attachments') 45 with open(self.filename_pattern % as_dict, 'w') as f: 46 f.write(self.encode(as_dict)) 47 # pylint: enable=invalid-name 48 [end of openhtf/io/output/json_factory.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/openhtf/io/output/json_factory.py b/openhtf/io/output/json_factory.py --- a/openhtf/io/output/json_factory.py +++ b/openhtf/io/output/json_factory.py @@ -1,5 +1,6 @@ """Module for outputting test record to JSON-formatted files.""" +import base64 from json import JSONEncoder from openhtf import util @@ -21,7 +22,9 @@ filename_pattern: A format string specifying the filename to write to, will be formatted with the Test Record as a dictionary. inline_attachments: Whether attachments should be included inline in the - output. Set to False if you expect to have large binary attachments. + output. Set to False if you expect to have large binary attachments. If + True (the default), then attachments are base64 encoded to allow for + binary data that's not supported by JSON directly. """ def __init__(self, filename_pattern=None, inline_attachments=True, **kwargs): @@ -40,6 +43,9 @@ assert self.filename_pattern, 'filename_pattern required' if self.inline_attachments: as_dict = util.ConvertToBaseTypes(test_record) + for phase in as_dict['phases']: + for value in phase['attachments'].itervalues(): + value['data'] = base64.standard_b64encode(value['data']) else: as_dict = util.ConvertToBaseTypes(test_record, ignore_keys='attachments') with open(self.filename_pattern % as_dict, 'w') as f:
{"golden_diff": "diff --git a/openhtf/io/output/json_factory.py b/openhtf/io/output/json_factory.py\n--- a/openhtf/io/output/json_factory.py\n+++ b/openhtf/io/output/json_factory.py\n@@ -1,5 +1,6 @@\n \"\"\"Module for outputting test record to JSON-formatted files.\"\"\"\n \n+import base64\n from json import JSONEncoder\n \n from openhtf import util\n@@ -21,7 +22,9 @@\n filename_pattern: A format string specifying the filename to write to,\n will be formatted with the Test Record as a dictionary.\n inline_attachments: Whether attachments should be included inline in the\n- output. Set to False if you expect to have large binary attachments.\n+ output. Set to False if you expect to have large binary attachments. If\n+ True (the default), then attachments are base64 encoded to allow for\n+ binary data that's not supported by JSON directly.\n \"\"\"\n \n def __init__(self, filename_pattern=None, inline_attachments=True, **kwargs):\n@@ -40,6 +43,9 @@\n assert self.filename_pattern, 'filename_pattern required'\n if self.inline_attachments:\n as_dict = util.ConvertToBaseTypes(test_record)\n+ for phase in as_dict['phases']:\n+ for value in phase['attachments'].itervalues():\n+ value['data'] = base64.standard_b64encode(value['data'])\n else:\n as_dict = util.ConvertToBaseTypes(test_record, ignore_keys='attachments')\n with open(self.filename_pattern % as_dict, 'w') as f:\n", "issue": "Attaching binary file using test.attach raises UnicodeDecodeError\nIf I attach a png or avi I see the following in OutputTestRecord\n\nPython2.7/site-packages/openhtf/**init**.py\", line 185, in OutputTestRecord\n output_cb(test_record)\n File \"virtualenv/local/lib/python2.7/site-packages/openhtf/**init**.py\", line 83, in **call**\n f.write(self.encode(as_dict))\n File \"/usr/lib/python2.7/json/encoder.py\", line 209, in encode\n chunks = list(chunks)\n File \"/usr/lib/python2.7/json/encoder.py\", line 434, in _iterencode\n for chunk in _iterencode_dict(o, _current_indent_level):\n File \"/usr/lib/python2.7/json/encoder.py\", line 408, in _iterencode_dict\n for chunk in chunks:\n File \"/usr/lib/python2.7/json/encoder.py\", line 332, in _iterencode_list\n for chunk in chunks:\n File \"/usr/lib/python2.7/json/encoder.py\", line 408, in _iterencode_dict\n for chunk in chunks:\n File \"/usr/lib/python2.7/json/encoder.py\", line 408, in _iterencode_dict\n for chunk in chunks:\n File \"/usr/lib/python2.7/json/encoder.py\", line 390, in _iterencode_dict\n yield _encoder(value)\nUnicodeDecodeError: 'utf8' codec can't decode byte 0x89 in position 0: invalid start byte\n\n", "before_files": [{"content": "\"\"\"Module for outputting test record to JSON-formatted files.\"\"\"\n\nfrom json import JSONEncoder\n\nfrom openhtf import util\nfrom openhtf.exe import test_state\n\n\nclass OutputToJSON(JSONEncoder):\n \"\"\"Return an output callback that writes JSON Test Records.\n\n An example filename_pattern might be:\n '/data/test_records/%(dut_id)s.%(start_time_millis)s'\n\n To use this output mechanism:\n test = openhtf.Test(PhaseOne, PhaseTwo)\n test.AddOutputCallback(openhtf.OutputToJson(\n '/data/test_records/%(dut_id)s.%(start_time_millis)s'))\n\n Args:\n filename_pattern: A format string specifying the filename to write to,\n will be formatted with the Test Record as a dictionary.\n inline_attachments: Whether attachments should be included inline in the\n output. Set to False if you expect to have large binary attachments.\n \"\"\"\n\n def __init__(self, filename_pattern=None, inline_attachments=True, **kwargs):\n super(OutputToJSON, self).__init__(**kwargs)\n self.filename_pattern = filename_pattern\n self.inline_attachments = inline_attachments\n\n def default(self, obj):\n if isinstance(obj, BaseException):\n # Just repr exceptions.\n return repr(obj)\n return super(OutputToJSON, self).default(obj)\n\n # pylint: disable=invalid-name\n def __call__(self, test_record):\n assert self.filename_pattern, 'filename_pattern required'\n if self.inline_attachments:\n as_dict = util.ConvertToBaseTypes(test_record)\n else:\n as_dict = util.ConvertToBaseTypes(test_record, ignore_keys='attachments')\n with open(self.filename_pattern % as_dict, 'w') as f:\n f.write(self.encode(as_dict))\n # pylint: enable=invalid-name\n", "path": "openhtf/io/output/json_factory.py"}]}
1,372
344
gh_patches_debug_24992
rasdani/github-patches
git_diff
fedora-infra__bodhi-2733
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> Staging is currently returning HTML to bodhi CLI requests I am not sure why this is happening, but it seems that staging Bodhi is currently returning HTML to CLI requests. This also happens to requests with ```http``` or ```curl```. I recall a problem with the unit tests where they would sometimes receive HTML when they didn't explicitly use a request header to ask for a JSON response once we started testing under Python 3. We ended up adjusting the tests to pass that header since this did not seem to happen when serving Bodhi with ```pserve-3```. It turns out that there really is some problem that seems related to Python 3 since staging Bodhi started doing this same thing. </issue> <code> [start of bodhi/server/webapp.py] 1 # -*- coding: utf-8 -*- 2 # Copyright © 2018 Red Hat, Inc. 3 # 4 # This file is part of Bodhi. 5 # 6 # This program is free software; you can redistribute it and/or 7 # modify it under the terms of the GNU General Public License 8 # as published by the Free Software Foundation; either version 2 9 # of the License, or (at your option) any later version. 10 # 11 # This program is distributed in the hope that it will be useful, 12 # but WITHOUT ANY WARRANTY; without even the implied warranty of 13 # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the 14 # GNU General Public License for more details. 15 # 16 # You should have received a copy of the GNU General Public License 17 # along with this program; if not, write to the Free Software 18 # Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. 19 """ 20 Define Bodhi's WSGI application. 21 22 As of the writing of this docblock, this module is a bit misnamed since the webapp is actually 23 defined in bodhi.server.__init__. However, that is an anti-pattern with lots of nasty in-line 24 imports due to circular dependencies, and this module is intended to solve that problem. 25 Unfortunately, it is a backwards-incompatible change to move main() here, so it will remain in 26 __init__ until we make a major Bodhi release. See https://github.com/fedora-infra/bodhi/issues/2294 27 """ 28 29 from pyramid.events import NewRequest, subscriber 30 31 from bodhi import server 32 33 34 def _complete_database_session(request): 35 """ 36 Commit the database changes if no exceptions occurred. 37 38 This is a post-request hook. It handles rolling back or committing the session based on whether 39 an exception occurred or not. To get a database session that's not tied to the request/response 40 cycle, just use the :data:`Session` scoped session. 41 42 Args: 43 request (pyramid.request.Request): The current web request. 44 """ 45 _rollback_or_commit(request) 46 server.Session().close() 47 server.Session.remove() 48 49 50 @subscriber(NewRequest) 51 def _prepare_request(event): 52 """ 53 Add callbacks onto every new request. 54 55 This function adds a callback to clean up the database session when the request is finished. 56 57 Args: 58 event (pyramid.events.NewRequest): The new request event. 59 """ 60 event.request.add_finished_callback(_complete_database_session) 61 62 63 def _rollback_or_commit(request): 64 """ 65 Commit the transaction if there are no exceptions, otherwise rollback. 66 67 Args: 68 request (pyramid.request.Request): The current web request. 69 """ 70 if request.exception is not None: 71 server.Session().rollback() 72 else: 73 server.Session().commit() 74 [end of bodhi/server/webapp.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/bodhi/server/webapp.py b/bodhi/server/webapp.py --- a/bodhi/server/webapp.py +++ b/bodhi/server/webapp.py @@ -50,13 +50,25 @@ @subscriber(NewRequest) def _prepare_request(event): """ - Add callbacks onto every new request. + Prepare each incoming request to Bodhi. - This function adds a callback to clean up the database session when the request is finished. + This function does two things: + * If requests do not have an Accept header, or if their Accept header is "*/*", it sets the + header to application/json. Pyramid has undefined behavior when an ambiguous or missing + Accept header is received, and multiple views are defined that handle specific Accept + headers. For example, we have a view that returns html or JSON for /composes/, depending + on the Accept header, but if a request has no Accept header or has */*, Pyramid will + consider both views to be a match for the request and so it is undefined which view will + handle the request. Let's force ambibuous requests to receive a JSON response so we have a + defined behavior. See https://github.com/fedora-infra/bodhi/issues/2731. + * It adds a callback to clean up the database session when the request is finished. Args: event (pyramid.events.NewRequest): The new request event. """ + if 'Accept' not in event.request.headers or event.request.headers['Accept'] == '*/*': + event.request.headers['Accept'] = 'application/json' + event.request.add_finished_callback(_complete_database_session)
{"golden_diff": "diff --git a/bodhi/server/webapp.py b/bodhi/server/webapp.py\n--- a/bodhi/server/webapp.py\n+++ b/bodhi/server/webapp.py\n@@ -50,13 +50,25 @@\n @subscriber(NewRequest)\n def _prepare_request(event):\n \"\"\"\n- Add callbacks onto every new request.\n+ Prepare each incoming request to Bodhi.\n \n- This function adds a callback to clean up the database session when the request is finished.\n+ This function does two things:\n+ * If requests do not have an Accept header, or if their Accept header is \"*/*\", it sets the\n+ header to application/json. Pyramid has undefined behavior when an ambiguous or missing\n+ Accept header is received, and multiple views are defined that handle specific Accept\n+ headers. For example, we have a view that returns html or JSON for /composes/, depending\n+ on the Accept header, but if a request has no Accept header or has */*, Pyramid will\n+ consider both views to be a match for the request and so it is undefined which view will\n+ handle the request. Let's force ambibuous requests to receive a JSON response so we have a\n+ defined behavior. See https://github.com/fedora-infra/bodhi/issues/2731.\n+ * It adds a callback to clean up the database session when the request is finished.\n \n Args:\n event (pyramid.events.NewRequest): The new request event.\n \"\"\"\n+ if 'Accept' not in event.request.headers or event.request.headers['Accept'] == '*/*':\n+ event.request.headers['Accept'] = 'application/json'\n+\n event.request.add_finished_callback(_complete_database_session)\n", "issue": "Staging is currently returning HTML to bodhi CLI requests\nI am not sure why this is happening, but it seems that staging Bodhi is currently returning HTML to CLI requests. This also happens to requests with ```http``` or ```curl```.\r\n\r\nI recall a problem with the unit tests where they would sometimes receive HTML when they didn't explicitly use a request header to ask for a JSON response once we started testing under Python 3. We ended up adjusting the tests to pass that header since this did not seem to happen when serving Bodhi with ```pserve-3```.\r\n\r\nIt turns out that there really is some problem that seems related to Python 3 since staging Bodhi started doing this same thing.\n", "before_files": [{"content": "# -*- coding: utf-8 -*-\n# Copyright \u00a9 2018 Red Hat, Inc.\n#\n# This file is part of Bodhi.\n#\n# This program is free software; you can redistribute it and/or\n# modify it under the terms of the GNU General Public License\n# as published by the Free Software Foundation; either version 2\n# of the License, or (at your option) any later version.\n#\n# This program is distributed in the hope that it will be useful,\n# but WITHOUT ANY WARRANTY; without even the implied warranty of\n# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the\n# GNU General Public License for more details.\n#\n# You should have received a copy of the GNU General Public License\n# along with this program; if not, write to the Free Software\n# Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA.\n\"\"\"\nDefine Bodhi's WSGI application.\n\nAs of the writing of this docblock, this module is a bit misnamed since the webapp is actually\ndefined in bodhi.server.__init__. However, that is an anti-pattern with lots of nasty in-line\nimports due to circular dependencies, and this module is intended to solve that problem.\nUnfortunately, it is a backwards-incompatible change to move main() here, so it will remain in\n__init__ until we make a major Bodhi release. See https://github.com/fedora-infra/bodhi/issues/2294\n\"\"\"\n\nfrom pyramid.events import NewRequest, subscriber\n\nfrom bodhi import server\n\n\ndef _complete_database_session(request):\n \"\"\"\n Commit the database changes if no exceptions occurred.\n\n This is a post-request hook. It handles rolling back or committing the session based on whether\n an exception occurred or not. To get a database session that's not tied to the request/response\n cycle, just use the :data:`Session` scoped session.\n\n Args:\n request (pyramid.request.Request): The current web request.\n \"\"\"\n _rollback_or_commit(request)\n server.Session().close()\n server.Session.remove()\n\n\n@subscriber(NewRequest)\ndef _prepare_request(event):\n \"\"\"\n Add callbacks onto every new request.\n\n This function adds a callback to clean up the database session when the request is finished.\n\n Args:\n event (pyramid.events.NewRequest): The new request event.\n \"\"\"\n event.request.add_finished_callback(_complete_database_session)\n\n\ndef _rollback_or_commit(request):\n \"\"\"\n Commit the transaction if there are no exceptions, otherwise rollback.\n\n Args:\n request (pyramid.request.Request): The current web request.\n \"\"\"\n if request.exception is not None:\n server.Session().rollback()\n else:\n server.Session().commit()\n", "path": "bodhi/server/webapp.py"}]}
1,414
372
gh_patches_debug_53973
rasdani/github-patches
git_diff
Mailu__Mailu-2563
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> postfix throttling afer a hard shutdown After a hard shutdown, I noticed that postfix would not restart because of master.pid is found in /queue/pid/master.pid We should check and remove this file during container start up (start.py) postfix throttling afer a hard shutdown After a hard shutdown, I noticed that postfix would not restart because of master.pid is found in /queue/pid/master.pid We should check and remove this file during container start up (start.py) </issue> <code> [start of core/postfix/start.py] 1 #!/usr/bin/env python3 2 3 import os 4 import glob 5 import shutil 6 import multiprocessing 7 import logging as log 8 import sys 9 import re 10 11 from podop import run_server 12 from pwd import getpwnam 13 from socrate import system, conf 14 15 log.basicConfig(stream=sys.stderr, level=os.environ.get("LOG_LEVEL", "WARNING")) 16 17 def start_podop(): 18 os.setuid(getpwnam('postfix').pw_uid) 19 os.makedirs('/dev/shm/postfix',mode=0o700, exist_ok=True) 20 url = "http://" + os.environ["ADMIN_ADDRESS"] + "/internal/postfix/" 21 # TODO: Remove verbosity setting from Podop? 22 run_server(0, "postfix", "/tmp/podop.socket", [ 23 ("transport", "url", url + "transport/§"), 24 ("alias", "url", url + "alias/§"), 25 ("dane", "url", url + "dane/§"), 26 ("domain", "url", url + "domain/§"), 27 ("mailbox", "url", url + "mailbox/§"), 28 ("recipientmap", "url", url + "recipient/map/§"), 29 ("sendermap", "url", url + "sender/map/§"), 30 ("senderlogin", "url", url + "sender/login/§"), 31 ("senderrate", "url", url + "sender/rate/§") 32 ]) 33 34 def start_mta_sts_daemon(): 35 os.chmod("/root/", 0o755) # read access to /root/.netrc required 36 os.setuid(getpwnam('postfix').pw_uid) 37 from postfix_mta_sts_resolver import daemon 38 daemon.main() 39 40 def is_valid_postconf_line(line): 41 return not line.startswith("#") \ 42 and not line == '' 43 44 # Actual startup script 45 os.environ['DEFER_ON_TLS_ERROR'] = os.environ['DEFER_ON_TLS_ERROR'] if 'DEFER_ON_TLS_ERROR' in os.environ else 'True' 46 os.environ["FRONT_ADDRESS"] = system.get_host_address_from_environment("FRONT", "front") 47 os.environ["ADMIN_ADDRESS"] = system.get_host_address_from_environment("ADMIN", "admin") 48 os.environ["ANTISPAM_MILTER_ADDRESS"] = system.get_host_address_from_environment("ANTISPAM_MILTER", "antispam:11332") 49 os.environ["LMTP_ADDRESS"] = system.get_host_address_from_environment("LMTP", "imap:2525") 50 os.environ["POSTFIX_LOG_SYSLOG"] = os.environ.get("POSTFIX_LOG_SYSLOG","local") 51 os.environ["POSTFIX_LOG_FILE"] = os.environ.get("POSTFIX_LOG_FILE", "") 52 53 # Postfix requires IPv6 addresses to be wrapped in square brackets 54 if 'RELAYNETS' in os.environ: 55 os.environ["RELAYNETS"] = re.sub(r'([0-9a-fA-F]+:[0-9a-fA-F:]+)/', '[\\1]/', os.environ["RELAYNETS"]) 56 57 for postfix_file in glob.glob("/conf/*.cf"): 58 conf.jinja(postfix_file, os.environ, os.path.join("/etc/postfix", os.path.basename(postfix_file))) 59 60 if os.path.exists("/overrides/postfix.cf"): 61 for line in open("/overrides/postfix.cf").read().strip().split("\n"): 62 if is_valid_postconf_line(line): 63 os.system('postconf -e "{}"'.format(line)) 64 65 if os.path.exists("/overrides/postfix.master"): 66 for line in open("/overrides/postfix.master").read().strip().split("\n"): 67 if is_valid_postconf_line(line): 68 os.system('postconf -Me "{}"'.format(line)) 69 70 for map_file in glob.glob("/overrides/*.map"): 71 destination = os.path.join("/etc/postfix", os.path.basename(map_file)) 72 shutil.copyfile(map_file, destination) 73 os.system("postmap {}".format(destination)) 74 os.remove(destination) 75 76 if os.path.exists("/overrides/mta-sts-daemon.yml"): 77 shutil.copyfile("/overrides/mta-sts-daemon.yml", "/etc/mta-sts-daemon.yml") 78 else: 79 conf.jinja("/conf/mta-sts-daemon.yml", os.environ, "/etc/mta-sts-daemon.yml") 80 81 for policy in ['tls_policy', 'transport']: 82 if not os.path.exists(f'/etc/postfix/{policy}.map.lmdb'): 83 open(f'/etc/postfix/{policy}.map', 'a').close() 84 os.system(f'postmap /etc/postfix/{policy}.map') 85 86 if "RELAYUSER" in os.environ: 87 path = "/etc/postfix/sasl_passwd" 88 conf.jinja("/conf/sasl_passwd", os.environ, path) 89 os.system("postmap {}".format(path)) 90 91 # Configure and start local rsyslog server 92 conf.jinja("/conf/rsyslog.conf", os.environ, "/etc/rsyslog.conf") 93 os.system("/usr/sbin/rsyslogd -niNONE &") 94 # Configure logrotate and start crond 95 if os.environ["POSTFIX_LOG_FILE"] != "": 96 conf.jinja("/conf/logrotate.conf", os.environ, "/etc/logrotate.d/postfix.conf") 97 os.system("/usr/sbin/crond") 98 if os.path.exists("/overrides/logrotate.conf"): 99 shutil.copyfile("/overrides/logrotate.conf", "/etc/logrotate.d/postfix.conf") 100 101 # Run Podop and Postfix 102 multiprocessing.Process(target=start_podop).start() 103 multiprocessing.Process(target=start_mta_sts_daemon).start() 104 os.system("/usr/libexec/postfix/post-install meta_directory=/etc/postfix create-missing") 105 # Before starting postfix, we need to check permissions on /queue 106 # in the event that postfix,postdrop id have changed 107 os.system("postfix set-permissions") 108 os.system("postfix start-fg") 109 [end of core/postfix/start.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/core/postfix/start.py b/core/postfix/start.py --- a/core/postfix/start.py +++ b/core/postfix/start.py @@ -14,6 +14,8 @@ log.basicConfig(stream=sys.stderr, level=os.environ.get("LOG_LEVEL", "WARNING")) +os.system("flock -n /queue/pid/master.pid rm /queue/pid/master.pid") + def start_podop(): os.setuid(getpwnam('postfix').pw_uid) os.makedirs('/dev/shm/postfix',mode=0o700, exist_ok=True)
{"golden_diff": "diff --git a/core/postfix/start.py b/core/postfix/start.py\n--- a/core/postfix/start.py\n+++ b/core/postfix/start.py\n@@ -14,6 +14,8 @@\n \n log.basicConfig(stream=sys.stderr, level=os.environ.get(\"LOG_LEVEL\", \"WARNING\"))\n \n+os.system(\"flock -n /queue/pid/master.pid rm /queue/pid/master.pid\")\n+\n def start_podop():\n os.setuid(getpwnam('postfix').pw_uid)\n os.makedirs('/dev/shm/postfix',mode=0o700, exist_ok=True)\n", "issue": "postfix throttling afer a hard shutdown\nAfter a hard shutdown, I noticed that postfix would not restart because of master.pid is found in /queue/pid/master.pid\r\nWe should check and remove this file during container start up (start.py)\npostfix throttling afer a hard shutdown\nAfter a hard shutdown, I noticed that postfix would not restart because of master.pid is found in /queue/pid/master.pid\r\nWe should check and remove this file during container start up (start.py)\n", "before_files": [{"content": "#!/usr/bin/env python3\n\nimport os\nimport glob\nimport shutil\nimport multiprocessing\nimport logging as log\nimport sys\nimport re\n\nfrom podop import run_server\nfrom pwd import getpwnam\nfrom socrate import system, conf\n\nlog.basicConfig(stream=sys.stderr, level=os.environ.get(\"LOG_LEVEL\", \"WARNING\"))\n\ndef start_podop():\n os.setuid(getpwnam('postfix').pw_uid)\n os.makedirs('/dev/shm/postfix',mode=0o700, exist_ok=True)\n url = \"http://\" + os.environ[\"ADMIN_ADDRESS\"] + \"/internal/postfix/\"\n # TODO: Remove verbosity setting from Podop?\n run_server(0, \"postfix\", \"/tmp/podop.socket\", [\n (\"transport\", \"url\", url + \"transport/\u00a7\"),\n (\"alias\", \"url\", url + \"alias/\u00a7\"),\n (\"dane\", \"url\", url + \"dane/\u00a7\"),\n (\"domain\", \"url\", url + \"domain/\u00a7\"),\n (\"mailbox\", \"url\", url + \"mailbox/\u00a7\"),\n (\"recipientmap\", \"url\", url + \"recipient/map/\u00a7\"),\n (\"sendermap\", \"url\", url + \"sender/map/\u00a7\"),\n (\"senderlogin\", \"url\", url + \"sender/login/\u00a7\"),\n (\"senderrate\", \"url\", url + \"sender/rate/\u00a7\")\n ])\n\ndef start_mta_sts_daemon():\n os.chmod(\"/root/\", 0o755) # read access to /root/.netrc required\n os.setuid(getpwnam('postfix').pw_uid)\n from postfix_mta_sts_resolver import daemon\n daemon.main()\n\ndef is_valid_postconf_line(line):\n return not line.startswith(\"#\") \\\n and not line == ''\n\n# Actual startup script\nos.environ['DEFER_ON_TLS_ERROR'] = os.environ['DEFER_ON_TLS_ERROR'] if 'DEFER_ON_TLS_ERROR' in os.environ else 'True'\nos.environ[\"FRONT_ADDRESS\"] = system.get_host_address_from_environment(\"FRONT\", \"front\")\nos.environ[\"ADMIN_ADDRESS\"] = system.get_host_address_from_environment(\"ADMIN\", \"admin\")\nos.environ[\"ANTISPAM_MILTER_ADDRESS\"] = system.get_host_address_from_environment(\"ANTISPAM_MILTER\", \"antispam:11332\")\nos.environ[\"LMTP_ADDRESS\"] = system.get_host_address_from_environment(\"LMTP\", \"imap:2525\")\nos.environ[\"POSTFIX_LOG_SYSLOG\"] = os.environ.get(\"POSTFIX_LOG_SYSLOG\",\"local\")\nos.environ[\"POSTFIX_LOG_FILE\"] = os.environ.get(\"POSTFIX_LOG_FILE\", \"\")\n\n# Postfix requires IPv6 addresses to be wrapped in square brackets\nif 'RELAYNETS' in os.environ:\n os.environ[\"RELAYNETS\"] = re.sub(r'([0-9a-fA-F]+:[0-9a-fA-F:]+)/', '[\\\\1]/', os.environ[\"RELAYNETS\"])\n\nfor postfix_file in glob.glob(\"/conf/*.cf\"):\n conf.jinja(postfix_file, os.environ, os.path.join(\"/etc/postfix\", os.path.basename(postfix_file)))\n\nif os.path.exists(\"/overrides/postfix.cf\"):\n for line in open(\"/overrides/postfix.cf\").read().strip().split(\"\\n\"):\n if is_valid_postconf_line(line):\n os.system('postconf -e \"{}\"'.format(line))\n\nif os.path.exists(\"/overrides/postfix.master\"):\n for line in open(\"/overrides/postfix.master\").read().strip().split(\"\\n\"):\n if is_valid_postconf_line(line):\n os.system('postconf -Me \"{}\"'.format(line))\n\nfor map_file in glob.glob(\"/overrides/*.map\"):\n destination = os.path.join(\"/etc/postfix\", os.path.basename(map_file))\n shutil.copyfile(map_file, destination)\n os.system(\"postmap {}\".format(destination))\n os.remove(destination)\n\nif os.path.exists(\"/overrides/mta-sts-daemon.yml\"):\n shutil.copyfile(\"/overrides/mta-sts-daemon.yml\", \"/etc/mta-sts-daemon.yml\")\nelse:\n conf.jinja(\"/conf/mta-sts-daemon.yml\", os.environ, \"/etc/mta-sts-daemon.yml\")\n\nfor policy in ['tls_policy', 'transport']:\n if not os.path.exists(f'/etc/postfix/{policy}.map.lmdb'):\n open(f'/etc/postfix/{policy}.map', 'a').close()\n os.system(f'postmap /etc/postfix/{policy}.map')\n\nif \"RELAYUSER\" in os.environ:\n path = \"/etc/postfix/sasl_passwd\"\n conf.jinja(\"/conf/sasl_passwd\", os.environ, path)\n os.system(\"postmap {}\".format(path))\n\n# Configure and start local rsyslog server\nconf.jinja(\"/conf/rsyslog.conf\", os.environ, \"/etc/rsyslog.conf\")\nos.system(\"/usr/sbin/rsyslogd -niNONE &\")\n# Configure logrotate and start crond\nif os.environ[\"POSTFIX_LOG_FILE\"] != \"\":\n conf.jinja(\"/conf/logrotate.conf\", os.environ, \"/etc/logrotate.d/postfix.conf\")\n os.system(\"/usr/sbin/crond\")\n if os.path.exists(\"/overrides/logrotate.conf\"):\n shutil.copyfile(\"/overrides/logrotate.conf\", \"/etc/logrotate.d/postfix.conf\")\n\n# Run Podop and Postfix\nmultiprocessing.Process(target=start_podop).start()\nmultiprocessing.Process(target=start_mta_sts_daemon).start()\nos.system(\"/usr/libexec/postfix/post-install meta_directory=/etc/postfix create-missing\")\n# Before starting postfix, we need to check permissions on /queue\n# in the event that postfix,postdrop id have changed\nos.system(\"postfix set-permissions\")\nos.system(\"postfix start-fg\")\n", "path": "core/postfix/start.py"}]}
2,106
128
gh_patches_debug_10941
rasdani/github-patches
git_diff
mesonbuild__meson-8978
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> get_variable with a file object as default value: Argument of type File is not held by an ObjectHolder **Describe the bug** After updating Meson, I see this error in a previously working build: ``` build/analysis/vale/meson.build:24:0: ERROR: Argument build/analysis/vale/vale-styleguide/config/documentation.vale.ini of type File is not held by an ObjectHolder. This is a Meson bug and should be reported! ``` The file is being specified in this manner: ``` # Supply a style file, which will use this file instead of the default .vale.ini vale_config_file = get_variable('vale_config_file', files('vale-styleguide/config/documentation.vale.ini')) ``` The default variable option is being used - I'm not overriding it. The same is happening in a Doxygen module I use: ``` doxyfile_input = get_variable('doxyfile_input', files('Doxyfile.in')) ``` I tried moving the file object into another variable: ``` vale_default_config_file = files('vale-styleguide/config/documentation.vale.ini') vale_config_file = get_variable('vale_config_file', vale_default_config_file) ``` With teh same result - the error is reported on the `get_variable` line. **system parameters** * Is this a [cross build](https://mesonbuild.com/Cross-compilation.html) or just a plain native build (for the same computer)? **native** * what operating system (e.g. MacOS Catalina, Windows 10, CentOS 8.0, Ubuntu 18.04, etc.) **MacOS 10.15.7** * what Python version are you using e.g. 3.8.0 **Python 3.9.6** * what `meson --version` **0.59.0.rc1** </issue> <code> [start of mesonbuild/interpreterbase/_unholder.py] 1 # Copyright 2013-2021 The Meson development team 2 3 # Licensed under the Apache License, Version 2.0 (the "License"); 4 # you may not use this file except in compliance with the License. 5 # You may obtain a copy of the License at 6 7 # http://www.apache.org/licenses/LICENSE-2.0 8 9 # Unless required by applicable law or agreed to in writing, software 10 # distributed under the License is distributed on an "AS IS" BASIS, 11 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 12 # See the License for the specific language governing permissions and 13 # limitations under the License. 14 15 from .baseobjects import InterpreterObject, MesonInterpreterObject, ObjectHolder, TYPE_var 16 from .exceptions import InvalidArguments 17 from ..mesonlib import HoldableObject, MesonBugException 18 19 import typing as T 20 21 def _unholder(obj: T.Union[TYPE_var, InterpreterObject], *, permissive: bool = False) -> TYPE_var: 22 if isinstance(obj, (int, bool, str)): 23 return obj 24 elif isinstance(obj, list): 25 return [_unholder(x) for x in obj] 26 elif isinstance(obj, dict): 27 return {k: _unholder(v) for k, v in obj.items()} 28 elif isinstance(obj, ObjectHolder): 29 assert isinstance(obj.held_object, HoldableObject) 30 return obj.held_object 31 elif isinstance(obj, MesonInterpreterObject): 32 return obj 33 elif isinstance(obj, HoldableObject) and permissive: 34 return obj 35 elif isinstance(obj, HoldableObject): 36 raise MesonBugException(f'Argument {obj} of type {type(obj).__name__} is not held by an ObjectHolder.') 37 elif isinstance(obj, InterpreterObject): 38 raise InvalidArguments(f'Argument {obj} of type {type(obj).__name__} cannot be passed to a method or function') 39 raise MesonBugException(f'Unknown object {obj} of type {type(obj).__name__} in the parameters.') 40 [end of mesonbuild/interpreterbase/_unholder.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/mesonbuild/interpreterbase/_unholder.py b/mesonbuild/interpreterbase/_unholder.py --- a/mesonbuild/interpreterbase/_unholder.py +++ b/mesonbuild/interpreterbase/_unholder.py @@ -22,9 +22,9 @@ if isinstance(obj, (int, bool, str)): return obj elif isinstance(obj, list): - return [_unholder(x) for x in obj] + return [_unholder(x, permissive=permissive) for x in obj] elif isinstance(obj, dict): - return {k: _unholder(v) for k, v in obj.items()} + return {k: _unholder(v, permissive=permissive) for k, v in obj.items()} elif isinstance(obj, ObjectHolder): assert isinstance(obj.held_object, HoldableObject) return obj.held_object
{"golden_diff": "diff --git a/mesonbuild/interpreterbase/_unholder.py b/mesonbuild/interpreterbase/_unholder.py\n--- a/mesonbuild/interpreterbase/_unholder.py\n+++ b/mesonbuild/interpreterbase/_unholder.py\n@@ -22,9 +22,9 @@\n if isinstance(obj, (int, bool, str)):\n return obj\n elif isinstance(obj, list):\n- return [_unholder(x) for x in obj]\n+ return [_unholder(x, permissive=permissive) for x in obj]\n elif isinstance(obj, dict):\n- return {k: _unholder(v) for k, v in obj.items()}\n+ return {k: _unholder(v, permissive=permissive) for k, v in obj.items()}\n elif isinstance(obj, ObjectHolder):\n assert isinstance(obj.held_object, HoldableObject)\n return obj.held_object\n", "issue": "get_variable with a file object as default value: Argument of type File is not held by an ObjectHolder\n**Describe the bug**\r\nAfter updating Meson, I see this error in a previously working build:\r\n\r\n```\r\nbuild/analysis/vale/meson.build:24:0: ERROR: Argument build/analysis/vale/vale-styleguide/config/documentation.vale.ini of type File is not held by an ObjectHolder.\r\n\r\n This is a Meson bug and should be reported!\r\n```\r\n\r\nThe file is being specified in this manner:\r\n\r\n```\r\n# Supply a style file, which will use this file instead of the default .vale.ini\r\nvale_config_file = get_variable('vale_config_file',\r\n\tfiles('vale-styleguide/config/documentation.vale.ini'))\r\n```\r\n\r\nThe default variable option is being used - I'm not overriding it.\r\n\r\nThe same is happening in a Doxygen module I use:\r\n\r\n```\r\ndoxyfile_input = get_variable('doxyfile_input', files('Doxyfile.in'))\r\n```\r\n\r\nI tried moving the file object into another variable:\r\n\r\n```\r\nvale_default_config_file = files('vale-styleguide/config/documentation.vale.ini')\r\nvale_config_file = get_variable('vale_config_file', vale_default_config_file)\r\n```\r\n\r\nWith teh same result - the error is reported on the `get_variable` line.\r\n\r\n**system parameters**\r\n* Is this a [cross build](https://mesonbuild.com/Cross-compilation.html) or just a plain native build (for the same computer)? **native**\r\n* what operating system (e.g. MacOS Catalina, Windows 10, CentOS 8.0, Ubuntu 18.04, etc.) **MacOS 10.15.7**\r\n* what Python version are you using e.g. 3.8.0 **Python 3.9.6**\r\n* what `meson --version` **0.59.0.rc1**\r\n\n", "before_files": [{"content": "# Copyright 2013-2021 The Meson development team\n\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n\n# http://www.apache.org/licenses/LICENSE-2.0\n\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom .baseobjects import InterpreterObject, MesonInterpreterObject, ObjectHolder, TYPE_var\nfrom .exceptions import InvalidArguments\nfrom ..mesonlib import HoldableObject, MesonBugException\n\nimport typing as T\n\ndef _unholder(obj: T.Union[TYPE_var, InterpreterObject], *, permissive: bool = False) -> TYPE_var:\n if isinstance(obj, (int, bool, str)):\n return obj\n elif isinstance(obj, list):\n return [_unholder(x) for x in obj]\n elif isinstance(obj, dict):\n return {k: _unholder(v) for k, v in obj.items()}\n elif isinstance(obj, ObjectHolder):\n assert isinstance(obj.held_object, HoldableObject)\n return obj.held_object\n elif isinstance(obj, MesonInterpreterObject):\n return obj\n elif isinstance(obj, HoldableObject) and permissive:\n return obj\n elif isinstance(obj, HoldableObject):\n raise MesonBugException(f'Argument {obj} of type {type(obj).__name__} is not held by an ObjectHolder.')\n elif isinstance(obj, InterpreterObject):\n raise InvalidArguments(f'Argument {obj} of type {type(obj).__name__} cannot be passed to a method or function')\n raise MesonBugException(f'Unknown object {obj} of type {type(obj).__name__} in the parameters.')\n", "path": "mesonbuild/interpreterbase/_unholder.py"}]}
1,441
196
gh_patches_debug_40786
rasdani/github-patches
git_diff
kornia__kornia-2017
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> No automatic `nearest` interpolation for masks in `RandomElasticTransform` ### Describe the bug Usually, the data type `DataKey.MASK` uses the `nearest` interpolation method to preserve label index values (e.g. when using `RandomAffine`). However, this does not work for the `RandomElasticTransform` augmentation since the corresponding parameter is named `mode` and not `resample`. Hence the default argument to `AugmentationSequential` ```python extra_args={DataKey.MASK: dict(resample=Resample.NEAREST, align_corners=True)} ``` does not trigger the `nearest` interpolation mode. ### Reproduction steps Here is an example reproducing the issue of the current default interpolation for mask input types: ```python import torch import kornia.augmentation as K features = torch.rand(1, 1, 4, 4) labels = torch.ones(1, 1, 4, 4, dtype=torch.float32) * 10 labels[:, :, :, :2] = 0 labels[:, :, :2, :] = 0 # tensor([[[[ 0., 0., 0., 0.], # [ 0., 0., 0., 0.], # [ 0., 0., 10., 10.], # [ 0., 0., 10., 10.]]]]) compose = K.AugmentationSequential(K.RandomElasticTransform(alpha=(10, 10))) torch.manual_seed(0) compose(features, labels, data_keys=["input", "mask"])[1] # tensor([[[[0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00], # [0.0000e+00, 0.0000e+00, 4.4417e-03, 4.3297e-03], # [0.0000e+00, 0.0000e+00, 9.8832e+00, 1.0000e+01], # [0.0000e+00, 0.0000e+00, 9.8833e+00, 1.0000e+01]]]]) ``` ### Expected behavior The default behavior for the elastic transformation should be similar to the other transformations, i.e. `nearest` interpolation for mask types. ### Environment - PyTorch Version (e.g., 1.0): 1.13 - OS (e.g., Linux): Linux - How you installed PyTorch (`conda`, `pip`, source): pip - Build command you used (if compiling from source): - Python version: 3.10.4 - CUDA/cuDNN version: 11.7 - GPU models and configuration: 3090 - Any other relevant information: ### Additional context Current workaround for this issue is to manually set the interpolation mode: ```python compose = K.AugmentationSequential(*transforms, extra_args={DataKey.MASK: dict(resample=Resample.NEAREST, align_corners=True, mode="nearest")}) ``` One option would be to rename the `mode` parameter to `resample`, than the default mask handling triggers as well. This would be a breaking change but I see that elastic transformation might get some exciting new functionality in the future anyway (issue #994) so this could be a combined change. </issue> <code> [start of kornia/augmentation/_2d/geometric/elastic_transform.py] 1 from typing import Any, Dict, Optional, Tuple 2 3 import torch 4 from torch import Tensor 5 6 from kornia.augmentation._2d.geometric.base import GeometricAugmentationBase2D 7 from kornia.geometry.transform import elastic_transform2d 8 9 10 class RandomElasticTransform(GeometricAugmentationBase2D): 11 r"""Add random elastic transformation to a tensor image. 12 13 .. image:: _static/img/RandomElasticTransform.png 14 15 Args: 16 kernel_size: the size of the Gaussian kernel. 17 sigma: The standard deviation of the Gaussian in the y and x directions, 18 respectively. Larger sigma results in smaller pixel displacements. 19 alpha: The scaling factor that controls the intensity of the deformation 20 in the y and x directions, respectively. 21 align_corners: Interpolation flag used by `grid_sample`. 22 mode: Interpolation mode used by `grid_sample`. Either 'bilinear' or 'nearest'. 23 padding_mode: The padding used by ```grid_sample```. Either 'zeros', 'border' or 'refection'. 24 same_on_batch: apply the same transformation across the batch. 25 p: probability of applying the transformation. 26 keepdim: whether to keep the output shape the same as input (True) or broadcast it 27 to the batch form (False). 28 29 .. note:: 30 This function internally uses :func:`kornia.geometry.transform.elastic_transform2d`. 31 32 Examples: 33 >>> import torch 34 >>> img = torch.ones(1, 1, 2, 2) 35 >>> out = RandomElasticTransform()(img) 36 >>> out.shape 37 torch.Size([1, 1, 2, 2]) 38 39 To apply the exact augmenation again, you may take the advantage of the previous parameter state: 40 >>> input = torch.randn(1, 3, 32, 32) 41 >>> aug = RandomElasticTransform(p=1.) 42 >>> (aug(input) == aug(input, params=aug._params)).all() 43 tensor(True) 44 """ 45 46 def __init__( 47 self, 48 kernel_size: Tuple[int, int] = (63, 63), 49 sigma: Tuple[float, float] = (32.0, 32.0), 50 alpha: Tuple[float, float] = (1.0, 1.0), 51 align_corners: bool = False, 52 mode: str = "bilinear", 53 padding_mode: str = "zeros", 54 same_on_batch: bool = False, 55 p: float = 0.5, 56 keepdim: bool = False, 57 return_transform: Optional[bool] = None, 58 ) -> None: 59 super().__init__( 60 p=p, return_transform=return_transform, same_on_batch=same_on_batch, p_batch=1.0, keepdim=keepdim 61 ) 62 self.flags = dict( 63 kernel_size=kernel_size, 64 sigma=sigma, 65 alpha=alpha, 66 align_corners=align_corners, 67 mode=mode, 68 padding_mode=padding_mode, 69 ) 70 71 def generate_parameters(self, shape: torch.Size) -> Dict[str, Tensor]: 72 B, _, H, W = shape 73 if self.same_on_batch: 74 noise = torch.rand(1, 2, H, W, device=self.device, dtype=self.dtype).repeat(B, 1, 1, 1) 75 else: 76 noise = torch.rand(B, 2, H, W, device=self.device, dtype=self.dtype) 77 return dict(noise=noise * 2 - 1) 78 79 # TODO: It is incorrect to return identity 80 def compute_transformation(self, input: Tensor, params: Dict[str, Tensor], flags: Dict[str, Any]) -> Tensor: 81 return self.identity_matrix(input) 82 83 def apply_transform( 84 self, input: Tensor, params: Dict[str, Tensor], flags: Dict[str, Any], transform: Optional[Tensor] = None 85 ) -> Tensor: 86 return elastic_transform2d( 87 input, 88 params["noise"].to(input), 89 flags["kernel_size"], 90 flags["sigma"], 91 flags["alpha"], 92 flags["align_corners"], 93 flags["mode"], 94 flags["padding_mode"], 95 ) 96 [end of kornia/augmentation/_2d/geometric/elastic_transform.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/kornia/augmentation/_2d/geometric/elastic_transform.py b/kornia/augmentation/_2d/geometric/elastic_transform.py --- a/kornia/augmentation/_2d/geometric/elastic_transform.py +++ b/kornia/augmentation/_2d/geometric/elastic_transform.py @@ -1,9 +1,11 @@ -from typing import Any, Dict, Optional, Tuple +import warnings +from typing import Any, Dict, Optional, Tuple, Union import torch from torch import Tensor from kornia.augmentation._2d.geometric.base import GeometricAugmentationBase2D +from kornia.constants import Resample from kornia.geometry.transform import elastic_transform2d @@ -19,7 +21,9 @@ alpha: The scaling factor that controls the intensity of the deformation in the y and x directions, respectively. align_corners: Interpolation flag used by `grid_sample`. - mode: Interpolation mode used by `grid_sample`. Either 'bilinear' or 'nearest'. + resample: Interpolation mode used by `grid_sample`. Either 'nearest' (0) or 'bilinear' (1). + mode: Deprecated: Interpolation mode used by `grid_sample`. Either 'bilinear' or 'nearest'. + Please use the `resample` argument instead. padding_mode: The padding used by ```grid_sample```. Either 'zeros', 'border' or 'refection'. same_on_batch: apply the same transformation across the batch. p: probability of applying the transformation. @@ -49,7 +53,8 @@ sigma: Tuple[float, float] = (32.0, 32.0), alpha: Tuple[float, float] = (1.0, 1.0), align_corners: bool = False, - mode: str = "bilinear", + resample: Union[str, int, Resample] = Resample.BILINEAR.name, + mode: Optional[str] = None, padding_mode: str = "zeros", same_on_batch: bool = False, p: float = 0.5, @@ -59,12 +64,21 @@ super().__init__( p=p, return_transform=return_transform, same_on_batch=same_on_batch, p_batch=1.0, keepdim=keepdim ) + if mode is not None: + resample = mode + if not torch.jit.is_scripting(): + warnings.warn( + "`mode` argument has been deprecated. Please use the `resample` argument instead", + DeprecationWarning, + stacklevel=2, + ) + self.flags = dict( kernel_size=kernel_size, sigma=sigma, alpha=alpha, align_corners=align_corners, - mode=mode, + resample=Resample.get(resample), padding_mode=padding_mode, ) @@ -90,6 +104,6 @@ flags["sigma"], flags["alpha"], flags["align_corners"], - flags["mode"], + flags["resample"].name.lower(), flags["padding_mode"], )
{"golden_diff": "diff --git a/kornia/augmentation/_2d/geometric/elastic_transform.py b/kornia/augmentation/_2d/geometric/elastic_transform.py\n--- a/kornia/augmentation/_2d/geometric/elastic_transform.py\n+++ b/kornia/augmentation/_2d/geometric/elastic_transform.py\n@@ -1,9 +1,11 @@\n-from typing import Any, Dict, Optional, Tuple\n+import warnings\n+from typing import Any, Dict, Optional, Tuple, Union\n \n import torch\n from torch import Tensor\n \n from kornia.augmentation._2d.geometric.base import GeometricAugmentationBase2D\n+from kornia.constants import Resample\n from kornia.geometry.transform import elastic_transform2d\n \n \n@@ -19,7 +21,9 @@\n alpha: The scaling factor that controls the intensity of the deformation\n in the y and x directions, respectively.\n align_corners: Interpolation flag used by `grid_sample`.\n- mode: Interpolation mode used by `grid_sample`. Either 'bilinear' or 'nearest'.\n+ resample: Interpolation mode used by `grid_sample`. Either 'nearest' (0) or 'bilinear' (1).\n+ mode: Deprecated: Interpolation mode used by `grid_sample`. Either 'bilinear' or 'nearest'.\n+ Please use the `resample` argument instead.\n padding_mode: The padding used by ```grid_sample```. Either 'zeros', 'border' or 'refection'.\n same_on_batch: apply the same transformation across the batch.\n p: probability of applying the transformation.\n@@ -49,7 +53,8 @@\n sigma: Tuple[float, float] = (32.0, 32.0),\n alpha: Tuple[float, float] = (1.0, 1.0),\n align_corners: bool = False,\n- mode: str = \"bilinear\",\n+ resample: Union[str, int, Resample] = Resample.BILINEAR.name,\n+ mode: Optional[str] = None,\n padding_mode: str = \"zeros\",\n same_on_batch: bool = False,\n p: float = 0.5,\n@@ -59,12 +64,21 @@\n super().__init__(\n p=p, return_transform=return_transform, same_on_batch=same_on_batch, p_batch=1.0, keepdim=keepdim\n )\n+ if mode is not None:\n+ resample = mode\n+ if not torch.jit.is_scripting():\n+ warnings.warn(\n+ \"`mode` argument has been deprecated. Please use the `resample` argument instead\",\n+ DeprecationWarning,\n+ stacklevel=2,\n+ )\n+\n self.flags = dict(\n kernel_size=kernel_size,\n sigma=sigma,\n alpha=alpha,\n align_corners=align_corners,\n- mode=mode,\n+ resample=Resample.get(resample),\n padding_mode=padding_mode,\n )\n \n@@ -90,6 +104,6 @@\n flags[\"sigma\"],\n flags[\"alpha\"],\n flags[\"align_corners\"],\n- flags[\"mode\"],\n+ flags[\"resample\"].name.lower(),\n flags[\"padding_mode\"],\n )\n", "issue": "No automatic `nearest` interpolation for masks in `RandomElasticTransform`\n### Describe the bug\r\n\r\nUsually, the data type `DataKey.MASK` uses the `nearest` interpolation method to preserve label index values (e.g. when using `RandomAffine`). However, this does not work for the `RandomElasticTransform` augmentation since the corresponding parameter is named `mode` and not `resample`. Hence the default argument to `AugmentationSequential`\r\n```python\r\nextra_args={DataKey.MASK: dict(resample=Resample.NEAREST, align_corners=True)}\r\n```\r\ndoes not trigger the `nearest` interpolation mode.\r\n\r\n### Reproduction steps\r\n\r\nHere is an example reproducing the issue of the current default interpolation for mask input types:\r\n```python\r\nimport torch\r\nimport kornia.augmentation as K\r\nfeatures = torch.rand(1, 1, 4, 4)\r\nlabels = torch.ones(1, 1, 4, 4, dtype=torch.float32) * 10\r\nlabels[:, :, :, :2] = 0\r\nlabels[:, :, :2, :] = 0\r\n# tensor([[[[ 0., 0., 0., 0.],\r\n# [ 0., 0., 0., 0.],\r\n# [ 0., 0., 10., 10.],\r\n# [ 0., 0., 10., 10.]]]])\r\n\r\ncompose = K.AugmentationSequential(K.RandomElasticTransform(alpha=(10, 10)))\r\ntorch.manual_seed(0)\r\ncompose(features, labels, data_keys=[\"input\", \"mask\"])[1]\r\n# tensor([[[[0.0000e+00, 0.0000e+00, 0.0000e+00, 0.0000e+00],\r\n# [0.0000e+00, 0.0000e+00, 4.4417e-03, 4.3297e-03],\r\n# [0.0000e+00, 0.0000e+00, 9.8832e+00, 1.0000e+01],\r\n# [0.0000e+00, 0.0000e+00, 9.8833e+00, 1.0000e+01]]]])\r\n```\r\n\r\n\r\n### Expected behavior\r\n\r\nThe default behavior for the elastic transformation should be similar to the other transformations, i.e. `nearest` interpolation for mask types.\r\n\r\n### Environment\r\n- PyTorch Version (e.g., 1.0): 1.13\r\n- OS (e.g., Linux): Linux\r\n- How you installed PyTorch (`conda`, `pip`, source): pip\r\n- Build command you used (if compiling from source):\r\n- Python version: 3.10.4\r\n- CUDA/cuDNN version: 11.7\r\n- GPU models and configuration: 3090\r\n- Any other relevant information:\r\n\r\n\r\n### Additional context\r\n\r\nCurrent workaround for this issue is to manually set the interpolation mode:\r\n```python\r\ncompose = K.AugmentationSequential(*transforms, extra_args={DataKey.MASK: dict(resample=Resample.NEAREST, align_corners=True, mode=\"nearest\")})\r\n```\r\n\r\nOne option would be to rename the `mode` parameter to `resample`, than the default mask handling triggers as well. This would be a breaking change but I see that elastic transformation might get some exciting new functionality in the future anyway (issue #994) so this could be a combined change.\n", "before_files": [{"content": "from typing import Any, Dict, Optional, Tuple\n\nimport torch\nfrom torch import Tensor\n\nfrom kornia.augmentation._2d.geometric.base import GeometricAugmentationBase2D\nfrom kornia.geometry.transform import elastic_transform2d\n\n\nclass RandomElasticTransform(GeometricAugmentationBase2D):\n r\"\"\"Add random elastic transformation to a tensor image.\n\n .. image:: _static/img/RandomElasticTransform.png\n\n Args:\n kernel_size: the size of the Gaussian kernel.\n sigma: The standard deviation of the Gaussian in the y and x directions,\n respectively. Larger sigma results in smaller pixel displacements.\n alpha: The scaling factor that controls the intensity of the deformation\n in the y and x directions, respectively.\n align_corners: Interpolation flag used by `grid_sample`.\n mode: Interpolation mode used by `grid_sample`. Either 'bilinear' or 'nearest'.\n padding_mode: The padding used by ```grid_sample```. Either 'zeros', 'border' or 'refection'.\n same_on_batch: apply the same transformation across the batch.\n p: probability of applying the transformation.\n keepdim: whether to keep the output shape the same as input (True) or broadcast it\n to the batch form (False).\n\n .. note::\n This function internally uses :func:`kornia.geometry.transform.elastic_transform2d`.\n\n Examples:\n >>> import torch\n >>> img = torch.ones(1, 1, 2, 2)\n >>> out = RandomElasticTransform()(img)\n >>> out.shape\n torch.Size([1, 1, 2, 2])\n\n To apply the exact augmenation again, you may take the advantage of the previous parameter state:\n >>> input = torch.randn(1, 3, 32, 32)\n >>> aug = RandomElasticTransform(p=1.)\n >>> (aug(input) == aug(input, params=aug._params)).all()\n tensor(True)\n \"\"\"\n\n def __init__(\n self,\n kernel_size: Tuple[int, int] = (63, 63),\n sigma: Tuple[float, float] = (32.0, 32.0),\n alpha: Tuple[float, float] = (1.0, 1.0),\n align_corners: bool = False,\n mode: str = \"bilinear\",\n padding_mode: str = \"zeros\",\n same_on_batch: bool = False,\n p: float = 0.5,\n keepdim: bool = False,\n return_transform: Optional[bool] = None,\n ) -> None:\n super().__init__(\n p=p, return_transform=return_transform, same_on_batch=same_on_batch, p_batch=1.0, keepdim=keepdim\n )\n self.flags = dict(\n kernel_size=kernel_size,\n sigma=sigma,\n alpha=alpha,\n align_corners=align_corners,\n mode=mode,\n padding_mode=padding_mode,\n )\n\n def generate_parameters(self, shape: torch.Size) -> Dict[str, Tensor]:\n B, _, H, W = shape\n if self.same_on_batch:\n noise = torch.rand(1, 2, H, W, device=self.device, dtype=self.dtype).repeat(B, 1, 1, 1)\n else:\n noise = torch.rand(B, 2, H, W, device=self.device, dtype=self.dtype)\n return dict(noise=noise * 2 - 1)\n\n # TODO: It is incorrect to return identity\n def compute_transformation(self, input: Tensor, params: Dict[str, Tensor], flags: Dict[str, Any]) -> Tensor:\n return self.identity_matrix(input)\n\n def apply_transform(\n self, input: Tensor, params: Dict[str, Tensor], flags: Dict[str, Any], transform: Optional[Tensor] = None\n ) -> Tensor:\n return elastic_transform2d(\n input,\n params[\"noise\"].to(input),\n flags[\"kernel_size\"],\n flags[\"sigma\"],\n flags[\"alpha\"],\n flags[\"align_corners\"],\n flags[\"mode\"],\n flags[\"padding_mode\"],\n )\n", "path": "kornia/augmentation/_2d/geometric/elastic_transform.py"}]}
2,461
708
gh_patches_debug_1316
rasdani/github-patches
git_diff
mozilla__bugbug-3334
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> Use information on how a bug is filed as a feature This could be especially useful for the Spam model. https://bugzilla.mozilla.org/show_bug.cgi?id=1565403 </issue> <code> [start of bugbug/models/spambug.py] 1 # -*- coding: utf-8 -*- 2 # This Source Code Form is subject to the terms of the Mozilla Public 3 # License, v. 2.0. If a copy of the MPL was not distributed with this file, 4 # You can obtain one at http://mozilla.org/MPL/2.0/. 5 6 import xgboost 7 from imblearn.over_sampling import BorderlineSMOTE 8 from sklearn.compose import ColumnTransformer 9 from sklearn.feature_extraction import DictVectorizer 10 from sklearn.pipeline import Pipeline 11 12 from bugbug import bug_features, bugzilla, feature_cleanup, utils 13 from bugbug.model import BugModel 14 15 16 class SpamBugModel(BugModel): 17 def __init__(self, lemmatization=False): 18 BugModel.__init__(self, lemmatization) 19 20 self.sampler = BorderlineSMOTE(random_state=0) 21 self.calculate_importance = False 22 23 feature_extractors = [ 24 bug_features.has_str(), 25 bug_features.has_regression_range(), 26 bug_features.severity(), 27 bug_features.has_crash_signature(), 28 bug_features.has_url(), 29 bug_features.whiteboard(), 30 bug_features.product(), 31 # TODO: We would like to use the component at the time of filing too, 32 # but we can't because the rollback script doesn't support changes to 33 # components yet. 34 # bug_features.component(), 35 bug_features.num_words_title(), 36 bug_features.num_words_comments(), 37 bug_features.keywords(), 38 bug_features.priority(), 39 bug_features.version(), 40 bug_features.target_milestone(), 41 bug_features.has_attachment(), 42 bug_features.platform(), 43 bug_features.op_sys(), 44 ] 45 46 cleanup_functions = [ 47 feature_cleanup.fileref(), 48 feature_cleanup.url(), 49 feature_cleanup.synonyms(), 50 ] 51 52 self.extraction_pipeline = Pipeline( 53 [ 54 ( 55 "bug_extractor", 56 bug_features.BugExtractor( 57 feature_extractors, cleanup_functions, rollback=True 58 ), 59 ), 60 ( 61 "union", 62 ColumnTransformer( 63 [ 64 ("data", DictVectorizer(), "data"), 65 ("title", self.text_vectorizer(min_df=0.0001), "title"), 66 ( 67 "comments", 68 self.text_vectorizer(min_df=0.0001), 69 "comments", 70 ), 71 ] 72 ), 73 ), 74 ] 75 ) 76 77 self.clf = xgboost.XGBClassifier(n_jobs=utils.get_physical_cpu_count()) 78 self.clf.set_params(predictor="cpu_predictor") 79 80 def get_labels(self): 81 classes = {} 82 83 for bug_data in bugzilla.get_bugs(include_invalid=True): 84 bug_id = bug_data["id"] 85 86 # Skip bugs filed by Mozillians, since we are sure they are not spam. 87 if "@mozilla" in bug_data["creator"]: 88 continue 89 90 # A bug that was moved out of 'Invalid Bugs' is definitely a legitimate bug. 91 for history in bug_data["history"]: 92 for change in history["changes"]: 93 if ( 94 change["field_name"] == "product" 95 and change["removed"] == "Invalid Bugs" 96 ): 97 classes[bug_id] = 0 98 99 # A fixed bug is definitely a legitimate bug. 100 if bug_data["resolution"] == "FIXED": 101 classes[bug_id] = 0 102 103 # A bug in the 'Invalid Bugs' product is definitely a spam bug. 104 elif bug_data["product"] == "Invalid Bugs": 105 classes[bug_id] = 1 106 107 print( 108 "{} bugs are classified as non-spam".format( 109 sum(1 for label in classes.values() if label == 0) 110 ) 111 ) 112 print( 113 "{} bugs are classified as spam".format( 114 sum(1 for label in classes.values() if label == 1) 115 ) 116 ) 117 118 return classes, [0, 1] 119 120 def items_gen(self, classes): 121 # Overwriting this method to add include_invalid=True to get_bugs to 122 # include spam bugs. 123 return ( 124 (bug, classes[bug["id"]]) 125 for bug in bugzilla.get_bugs(include_invalid=True) 126 if bug["id"] in classes 127 ) 128 129 def get_feature_names(self): 130 return self.extraction_pipeline.named_steps["union"].get_feature_names_out() 131 132 def overwrite_classes(self, bugs, classes, probabilities): 133 for i, bug in enumerate(bugs): 134 if "@mozilla" in bug["creator"]: 135 if probabilities: 136 classes[i] = [1.0, 0.0] 137 else: 138 classes[i] = 0 139 140 return classes 141 [end of bugbug/models/spambug.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/bugbug/models/spambug.py b/bugbug/models/spambug.py --- a/bugbug/models/spambug.py +++ b/bugbug/models/spambug.py @@ -41,6 +41,7 @@ bug_features.has_attachment(), bug_features.platform(), bug_features.op_sys(), + bug_features.filed_via(), ] cleanup_functions = [
{"golden_diff": "diff --git a/bugbug/models/spambug.py b/bugbug/models/spambug.py\n--- a/bugbug/models/spambug.py\n+++ b/bugbug/models/spambug.py\n@@ -41,6 +41,7 @@\n bug_features.has_attachment(),\n bug_features.platform(),\n bug_features.op_sys(),\n+ bug_features.filed_via(),\n ]\n \n cleanup_functions = [\n", "issue": "Use information on how a bug is filed as a feature\nThis could be especially useful for the Spam model.\r\n\r\nhttps://bugzilla.mozilla.org/show_bug.cgi?id=1565403\n", "before_files": [{"content": "# -*- coding: utf-8 -*-\n# This Source Code Form is subject to the terms of the Mozilla Public\n# License, v. 2.0. If a copy of the MPL was not distributed with this file,\n# You can obtain one at http://mozilla.org/MPL/2.0/.\n\nimport xgboost\nfrom imblearn.over_sampling import BorderlineSMOTE\nfrom sklearn.compose import ColumnTransformer\nfrom sklearn.feature_extraction import DictVectorizer\nfrom sklearn.pipeline import Pipeline\n\nfrom bugbug import bug_features, bugzilla, feature_cleanup, utils\nfrom bugbug.model import BugModel\n\n\nclass SpamBugModel(BugModel):\n def __init__(self, lemmatization=False):\n BugModel.__init__(self, lemmatization)\n\n self.sampler = BorderlineSMOTE(random_state=0)\n self.calculate_importance = False\n\n feature_extractors = [\n bug_features.has_str(),\n bug_features.has_regression_range(),\n bug_features.severity(),\n bug_features.has_crash_signature(),\n bug_features.has_url(),\n bug_features.whiteboard(),\n bug_features.product(),\n # TODO: We would like to use the component at the time of filing too,\n # but we can't because the rollback script doesn't support changes to\n # components yet.\n # bug_features.component(),\n bug_features.num_words_title(),\n bug_features.num_words_comments(),\n bug_features.keywords(),\n bug_features.priority(),\n bug_features.version(),\n bug_features.target_milestone(),\n bug_features.has_attachment(),\n bug_features.platform(),\n bug_features.op_sys(),\n ]\n\n cleanup_functions = [\n feature_cleanup.fileref(),\n feature_cleanup.url(),\n feature_cleanup.synonyms(),\n ]\n\n self.extraction_pipeline = Pipeline(\n [\n (\n \"bug_extractor\",\n bug_features.BugExtractor(\n feature_extractors, cleanup_functions, rollback=True\n ),\n ),\n (\n \"union\",\n ColumnTransformer(\n [\n (\"data\", DictVectorizer(), \"data\"),\n (\"title\", self.text_vectorizer(min_df=0.0001), \"title\"),\n (\n \"comments\",\n self.text_vectorizer(min_df=0.0001),\n \"comments\",\n ),\n ]\n ),\n ),\n ]\n )\n\n self.clf = xgboost.XGBClassifier(n_jobs=utils.get_physical_cpu_count())\n self.clf.set_params(predictor=\"cpu_predictor\")\n\n def get_labels(self):\n classes = {}\n\n for bug_data in bugzilla.get_bugs(include_invalid=True):\n bug_id = bug_data[\"id\"]\n\n # Skip bugs filed by Mozillians, since we are sure they are not spam.\n if \"@mozilla\" in bug_data[\"creator\"]:\n continue\n\n # A bug that was moved out of 'Invalid Bugs' is definitely a legitimate bug.\n for history in bug_data[\"history\"]:\n for change in history[\"changes\"]:\n if (\n change[\"field_name\"] == \"product\"\n and change[\"removed\"] == \"Invalid Bugs\"\n ):\n classes[bug_id] = 0\n\n # A fixed bug is definitely a legitimate bug.\n if bug_data[\"resolution\"] == \"FIXED\":\n classes[bug_id] = 0\n\n # A bug in the 'Invalid Bugs' product is definitely a spam bug.\n elif bug_data[\"product\"] == \"Invalid Bugs\":\n classes[bug_id] = 1\n\n print(\n \"{} bugs are classified as non-spam\".format(\n sum(1 for label in classes.values() if label == 0)\n )\n )\n print(\n \"{} bugs are classified as spam\".format(\n sum(1 for label in classes.values() if label == 1)\n )\n )\n\n return classes, [0, 1]\n\n def items_gen(self, classes):\n # Overwriting this method to add include_invalid=True to get_bugs to\n # include spam bugs.\n return (\n (bug, classes[bug[\"id\"]])\n for bug in bugzilla.get_bugs(include_invalid=True)\n if bug[\"id\"] in classes\n )\n\n def get_feature_names(self):\n return self.extraction_pipeline.named_steps[\"union\"].get_feature_names_out()\n\n def overwrite_classes(self, bugs, classes, probabilities):\n for i, bug in enumerate(bugs):\n if \"@mozilla\" in bug[\"creator\"]:\n if probabilities:\n classes[i] = [1.0, 0.0]\n else:\n classes[i] = 0\n\n return classes\n", "path": "bugbug/models/spambug.py"}]}
1,865
89
gh_patches_debug_22536
rasdani/github-patches
git_diff
opsdroid__opsdroid-1860
You will be provided with a partial code base and an issue statement explaining a problem to resolve. <issue> User configurable connection for mongo-based databases So the pymongo client has a multitude of ways for connecting to different mongo services So for MongoDB Atlas users the connection string is given as such for python connections to the mongo db atlas `mongodb+srv://<username>:<password>@<cluster-name>.mongodb.net/myFirstDatabase` In making the mongo connection to be user configurable we can specify different types of mongo services versus just asking for the basic connection arguments like port, user name, pass, and also we can give users an easier way to connect versus making assumptions about the type of mongodb the kinds of credentials they might have. As long as the pymongo client accepts the connection and connects the user to the database and the collection they want I think this would be great! Thanks again guys! </issue> <code> [start of opsdroid/database/mongo/__init__.py] 1 # -*- coding: utf-8 -*- 2 """A module for opsdroid to allow persist in mongo database.""" 3 import logging 4 from contextlib import asynccontextmanager 5 from motor.motor_asyncio import AsyncIOMotorClient 6 from voluptuous import Any 7 8 from opsdroid.database import Database 9 10 _LOGGER = logging.getLogger(__name__) 11 CONFIG_SCHEMA = { 12 "host": str, 13 "port": Any(int, str), 14 "database": str, 15 "user": str, 16 "password": str, 17 "collection": str, 18 } 19 20 21 class DatabaseMongo(Database): 22 """A module for opsdroid to allow memory to persist in a mongo database.""" 23 24 def __init__(self, config, opsdroid=None): 25 """Create the connection. 26 27 Set some basic properties from the database config such as the name 28 of this database. 29 30 Args: 31 config (dict): The config for this database specified in the 32 `configuration.yaml` file. 33 opsdroid (OpsDroid): An instance of opsdroid.core. 34 35 """ 36 super().__init__(config, opsdroid=opsdroid) 37 _LOGGER.debug("Loaded mongo database connector.") 38 self.name = "mongo" 39 self.config = config 40 self.client = None 41 self.database = None 42 self.collection = config.get("collection", "opsdroid") 43 44 async def connect(self): 45 """Connect to the database.""" 46 host = self.config.get("host", "localhost") 47 port = self.config.get("port", "27017") 48 database = self.config.get("database", "opsdroid") 49 user = self.config.get("user") 50 pwd = self.config.get("password") 51 if user and pwd: 52 path = "mongodb://{user}:{pwd}@{host}:{port}".format( 53 user=user, pwd=pwd, host=host, port=port 54 ) 55 else: 56 path = "mongodb://{host}:{port}".format(host=host, port=port) 57 self.client = AsyncIOMotorClient(path) 58 self.database = self.client[database] 59 _LOGGER.info("Connected to MongoDB.") 60 61 async def put(self, key, data): 62 """Insert or replace an object into the database for a given key. 63 64 Args: 65 key (str): the key is the document lookup key. 66 data (object): the data to be inserted or replaced 67 68 """ 69 _LOGGER.debug("Putting %s into MongoDB collection %s", key, self.collection) 70 71 if isinstance(data, str): 72 data = {"value": data} 73 if "key" not in data: 74 data["key"] = key 75 76 return await self.database[self.collection].update_one( 77 {"key": data["key"]}, {"$set": data}, upsert=True 78 ) 79 80 async def get(self, key): 81 """Get a document from the database (key). 82 83 Args: 84 key (str): the key is the document lookup key. 85 86 """ 87 _LOGGER.debug("Getting %s from MongoDB collection %s", key, self.collection) 88 89 response = await self.database[self.collection].find_one( 90 {"$query": {"key": key}, "$orderby": {"$natural": -1}} 91 ) 92 if response.keys() == {"_id", "key", "value"}: 93 response = response["value"] 94 return response 95 96 async def delete(self, key): 97 """Delete a document from the database (key). 98 99 Args: 100 key (str): the key is the document lookup key. 101 102 """ 103 _LOGGER.debug("Deleting %s from MongoDB collection %s.", key, self.collection) 104 105 return await self.database[self.collection].delete_one({"key": key}) 106 107 @asynccontextmanager 108 async def memory_in_collection(self, collection): 109 """Use the specified collection rather than the default.""" 110 db_copy = DatabaseMongo(self.config, self.opsdroid) 111 try: 112 await db_copy.connect() 113 db_copy.collection = collection 114 yield db_copy 115 finally: 116 if db_copy.client: 117 db_copy.client.close() 118 [end of opsdroid/database/mongo/__init__.py] </code> I need you to solve this issue by generating a single patch file that I can apply directly to this repository using git apply. Please respond with a single patch file in the following format. <patch> diff --git a/file.py b/file.py --- a/file.py +++ b/file.py @@ -1,27 +1,35 @@ def euclidean(a, b): - while b: - a, b = b, a % b - return a + if b == 0: + return a + return euclidean(b, a % b) def bresenham(x0, y0, x1, y1): points = [] dx = abs(x1 - x0) dy = abs(y1 - y0) - sx = 1 if x0 < x1 else -1 - sy = 1 if y0 < y1 else -1 - err = dx - dy + x, y = x0, y0 + sx = -1 if x0 > x1 else 1 + sy = -1 if y0 > y1 else 1 - while True: - points.append((x0, y0)) - if x0 == x1 and y0 == y1: - break - e2 = 2 * err - if e2 > -dy: - err -= dy - x0 += sx - if e2 < dx: - err += dx - y0 += sy + if dx > dy: + err = dx / 2.0 + while x != x1: + points.append((x, y)) + err -= dy + if err < 0: + y += sy + err += dx + x += sx + else: + err = dy / 2.0 + while y != y1: + points.append((x, y)) + err -= dx + if err < 0: + x += sx + err += dy + y += sy + + points.append((x, y)) return points </patch>
diff --git a/opsdroid/database/mongo/__init__.py b/opsdroid/database/mongo/__init__.py --- a/opsdroid/database/mongo/__init__.py +++ b/opsdroid/database/mongo/__init__.py @@ -44,17 +44,18 @@ async def connect(self): """Connect to the database.""" host = self.config.get("host", "localhost") + protocol = self.config.get("protocol", "mongodb").replace("://", "") port = self.config.get("port", "27017") + if port != "27017": + host = f"{host}:{port}" database = self.config.get("database", "opsdroid") user = self.config.get("user") pwd = self.config.get("password") if user and pwd: - path = "mongodb://{user}:{pwd}@{host}:{port}".format( - user=user, pwd=pwd, host=host, port=port - ) + self.db_url = f"{protocol}://{user}:{pwd}@{host}" else: - path = "mongodb://{host}:{port}".format(host=host, port=port) - self.client = AsyncIOMotorClient(path) + self.db_url = f"{protocol}://{host}" + self.client = AsyncIOMotorClient(self.db_url) self.database = self.client[database] _LOGGER.info("Connected to MongoDB.")
{"golden_diff": "diff --git a/opsdroid/database/mongo/__init__.py b/opsdroid/database/mongo/__init__.py\n--- a/opsdroid/database/mongo/__init__.py\n+++ b/opsdroid/database/mongo/__init__.py\n@@ -44,17 +44,18 @@\n async def connect(self):\n \"\"\"Connect to the database.\"\"\"\n host = self.config.get(\"host\", \"localhost\")\n+ protocol = self.config.get(\"protocol\", \"mongodb\").replace(\"://\", \"\")\n port = self.config.get(\"port\", \"27017\")\n+ if port != \"27017\":\n+ host = f\"{host}:{port}\"\n database = self.config.get(\"database\", \"opsdroid\")\n user = self.config.get(\"user\")\n pwd = self.config.get(\"password\")\n if user and pwd:\n- path = \"mongodb://{user}:{pwd}@{host}:{port}\".format(\n- user=user, pwd=pwd, host=host, port=port\n- )\n+ self.db_url = f\"{protocol}://{user}:{pwd}@{host}\"\n else:\n- path = \"mongodb://{host}:{port}\".format(host=host, port=port)\n- self.client = AsyncIOMotorClient(path)\n+ self.db_url = f\"{protocol}://{host}\"\n+ self.client = AsyncIOMotorClient(self.db_url)\n self.database = self.client[database]\n _LOGGER.info(\"Connected to MongoDB.\")\n", "issue": "User configurable connection for mongo-based databases\nSo the pymongo client has a multitude of ways for connecting to different mongo services\r\n\r\nSo for MongoDB Atlas users the connection string is given as such \r\nfor python connections to the mongo db atlas \r\n\r\n`mongodb+srv://<username>:<password>@<cluster-name>.mongodb.net/myFirstDatabase`\r\n\r\nIn making the mongo connection to be user configurable we can specify different types of mongo services versus\r\njust asking for the basic connection arguments like port, user name, pass, and also we can give users an easier way to connect versus making assumptions about the type of mongodb the kinds of credentials they might have. \r\n\r\nAs long as the pymongo client accepts the connection and connects the user to the database and the collection they want I think this would be great!\r\n\r\nThanks again guys!\r\n\n", "before_files": [{"content": "# -*- coding: utf-8 -*-\n\"\"\"A module for opsdroid to allow persist in mongo database.\"\"\"\nimport logging\nfrom contextlib import asynccontextmanager\nfrom motor.motor_asyncio import AsyncIOMotorClient\nfrom voluptuous import Any\n\nfrom opsdroid.database import Database\n\n_LOGGER = logging.getLogger(__name__)\nCONFIG_SCHEMA = {\n \"host\": str,\n \"port\": Any(int, str),\n \"database\": str,\n \"user\": str,\n \"password\": str,\n \"collection\": str,\n}\n\n\nclass DatabaseMongo(Database):\n \"\"\"A module for opsdroid to allow memory to persist in a mongo database.\"\"\"\n\n def __init__(self, config, opsdroid=None):\n \"\"\"Create the connection.\n\n Set some basic properties from the database config such as the name\n of this database.\n\n Args:\n config (dict): The config for this database specified in the\n `configuration.yaml` file.\n opsdroid (OpsDroid): An instance of opsdroid.core.\n\n \"\"\"\n super().__init__(config, opsdroid=opsdroid)\n _LOGGER.debug(\"Loaded mongo database connector.\")\n self.name = \"mongo\"\n self.config = config\n self.client = None\n self.database = None\n self.collection = config.get(\"collection\", \"opsdroid\")\n\n async def connect(self):\n \"\"\"Connect to the database.\"\"\"\n host = self.config.get(\"host\", \"localhost\")\n port = self.config.get(\"port\", \"27017\")\n database = self.config.get(\"database\", \"opsdroid\")\n user = self.config.get(\"user\")\n pwd = self.config.get(\"password\")\n if user and pwd:\n path = \"mongodb://{user}:{pwd}@{host}:{port}\".format(\n user=user, pwd=pwd, host=host, port=port\n )\n else:\n path = \"mongodb://{host}:{port}\".format(host=host, port=port)\n self.client = AsyncIOMotorClient(path)\n self.database = self.client[database]\n _LOGGER.info(\"Connected to MongoDB.\")\n\n async def put(self, key, data):\n \"\"\"Insert or replace an object into the database for a given key.\n\n Args:\n key (str): the key is the document lookup key.\n data (object): the data to be inserted or replaced\n\n \"\"\"\n _LOGGER.debug(\"Putting %s into MongoDB collection %s\", key, self.collection)\n\n if isinstance(data, str):\n data = {\"value\": data}\n if \"key\" not in data:\n data[\"key\"] = key\n\n return await self.database[self.collection].update_one(\n {\"key\": data[\"key\"]}, {\"$set\": data}, upsert=True\n )\n\n async def get(self, key):\n \"\"\"Get a document from the database (key).\n\n Args:\n key (str): the key is the document lookup key.\n\n \"\"\"\n _LOGGER.debug(\"Getting %s from MongoDB collection %s\", key, self.collection)\n\n response = await self.database[self.collection].find_one(\n {\"$query\": {\"key\": key}, \"$orderby\": {\"$natural\": -1}}\n )\n if response.keys() == {\"_id\", \"key\", \"value\"}:\n response = response[\"value\"]\n return response\n\n async def delete(self, key):\n \"\"\"Delete a document from the database (key).\n\n Args:\n key (str): the key is the document lookup key.\n\n \"\"\"\n _LOGGER.debug(\"Deleting %s from MongoDB collection %s.\", key, self.collection)\n\n return await self.database[self.collection].delete_one({\"key\": key})\n\n @asynccontextmanager\n async def memory_in_collection(self, collection):\n \"\"\"Use the specified collection rather than the default.\"\"\"\n db_copy = DatabaseMongo(self.config, self.opsdroid)\n try:\n await db_copy.connect()\n db_copy.collection = collection\n yield db_copy\n finally:\n if db_copy.client:\n db_copy.client.close()\n", "path": "opsdroid/database/mongo/__init__.py"}]}
1,831
325