problem_id
stringlengths 18
22
| source
stringclasses 1
value | task_type
stringclasses 1
value | in_source_id
stringlengths 13
58
| prompt
stringlengths 1.1k
25.4k
| golden_diff
stringlengths 145
5.13k
| verification_info
stringlengths 582
39.1k
| num_tokens
int64 271
4.1k
| num_tokens_diff
int64 47
1.02k
|
---|---|---|---|---|---|---|---|---|
gh_patches_debug_28066 | rasdani/github-patches | git_diff | Parsl__parsl-203 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
User tracking to report parsl version
User tracking as described in #34 tracks a range of script attributes and anonymous user information.
It would help to track the version of Parsl and python. This would help us understand:
1. Rate of upgrade to newer releases
2. Python version in use, and support requirements
3. Support requirements for older parsl releases.
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `parsl/dataflow/usage_tracking/usage.py`
Content:
```
1 import uuid
2 import time
3 import hashlib
4 import os
5 import getpass
6 import json
7 import logging
8 import socket
9
10 from parsl.dataflow.states import States
11
12 logger = logging.getLogger(__name__)
13
14
15 class UsageTracker (object):
16 """Anonymized Usage Tracking for Parsl.
17
18 Client for this is here : https://github.com/Parsl/parsl_tracking
19 This issue captures the discussion that went into functionality
20 implemented here : https://github.com/Parsl/parsl/issues/34
21
22 """
23
24 def __init__(self, dfk, ip='52.3.111.203', port=50077,
25 domain_name='tracking.parsl-project.org'):
26 """Initialize usage tracking unless the user has opted-out.
27
28 Tracks usage stats by inspecting the internal state of the dfk.
29
30 Args:
31 - dfk (DFK object) : Data Flow Kernel object
32
33 KWargs:
34 - ip (string) : IP address
35 - port (int) : Port number, Default:50077
36 - domain_name (string) : Domain name, will override IP
37 Default: tracking.parsl-project.org
38 """
39 if domain_name:
40 try:
41 self.UDP_IP = socket.gethostbyname(domain_name)
42 except Exception:
43 logging.debug("Could not lookup domain_name, defaulting to 52.3.111.203")
44 self.UDP_IP = ip
45 else:
46 self.UDP_IP = ip
47 self.UDP_PORT = port
48 self.dfk = dfk
49 self.config = self.dfk.config
50 self.uuid = str(uuid.uuid4())
51 self.test_mode, self.tracking_enabled = self.check_tracking_enabled()
52 logger.debug("Tracking status: {}".format(self.tracking_enabled))
53 logger.debug("Testing mode : {}".format(self.test_mode))
54 self.initialized = False # Once first message is sent this will be True
55
56 def check_tracking_enabled(self):
57 """By default tracking is enabled.
58
59 If Test mode is set via env variable PARSL_TESTING, a test flag is set
60
61 Tracking is disabled if :
62 1. config["globals"]["usageTracking"] is set to False (Bool)
63 2. Environment variable PARSL_TRACKING is set to false (case insensitive)
64
65 """
66 track = True # By default we track usage
67 test = False # By default we are not in testing mode
68
69 testvar = str(os.environ.get("PARSL_TESTING", 'None')).lower()
70 if testvar == 'true':
71 test = True
72
73 if self.config and self.config["globals"]["usageTracking"] is False:
74 track = False
75
76 envvar = str(os.environ.get("PARSL_TRACKING", True)).lower()
77 if envvar == "false":
78 track = False
79
80 return test, track
81
82 def construct_start_message(self):
83 """Collect preliminary run info at the start of the DFK.
84
85 Returns :
86 - Message dict dumped as json string, ready for UDP
87 """
88 uname = getpass.getuser().encode('latin1')
89 hashed_username = hashlib.sha256(uname).hexdigest()[0:10]
90 hname = socket.gethostname().encode('latin1')
91 hashed_hostname = hashlib.sha256(hname).hexdigest()[0:10]
92 message = {'uuid': self.uuid,
93 'uname': hashed_username,
94 'hname': hashed_hostname,
95 'test': self.test_mode,
96 'start': time.time()}
97
98 return json.dumps(message)
99
100 def construct_end_message(self):
101 """Collect the final run information at the time of DFK cleanup.
102
103 Returns:
104 - Message dict dumped as json string, ready for UDP
105 """
106 app_count = self.dfk.task_count
107
108 site_count = 0
109 if self.dfk._executors_managed:
110 site_count = len(self.dfk.config['sites'])
111
112 failed_states = (States.failed, States.dep_fail)
113 app_fails = len([t for t in self.dfk.tasks if
114 self.dfk.tasks[t]['status'] in failed_states])
115
116 message = {'uuid': self.uuid,
117 'end': time.time(),
118 't_apps': app_count,
119 'sites': site_count,
120 'c_time': None,
121 'failed': app_fails,
122 'test': self.test_mode,
123 }
124
125 return json.dumps(message)
126
127 def send_UDP_message(self, message):
128 """Send UDP message."""
129 if self.tracking_enabled:
130 try:
131 sock = socket.socket(socket.AF_INET, socket.SOCK_DGRAM) # UDP
132 x = sock.sendto(bytes(message, "utf-8"), (self.UDP_IP, self.UDP_PORT))
133 sock.close()
134 except OSError:
135 logger.debug("Unable to reach the network to send usage data")
136 x = 0
137 else:
138 x = -1
139
140 return x
141
142 def send_message(self):
143 """Send message over UDP.
144
145 If tracking is disables, the bytes_sent will always be set to -1
146
147 Returns:
148 (bytes_sent, time_taken)
149 """
150 start = time.time()
151 message = None
152 if not self.initialized:
153 message = self.construct_start_message()
154 self.initialized = True
155 else:
156 message = self.construct_end_message()
157
158 x = self.send_UDP_message(message)
159 end = time.time()
160
161 return x, end - start
162
163
164 if __name__ == '__main__':
165
166 from parsl import *
167
168 workers = ThreadPoolExecutor(max_workers=4)
169 dfk = DataFlowKernel(executors=[workers])
170
171 # ut = UsageTracker(dfk, ip='52.3.111.203')
172 ut = UsageTracker(dfk, domain_name='tracking.parsl-project.org')
173
174 for i in range(0, 2):
175 x = ut.send_message()
176 print(x)
177
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/parsl/dataflow/usage_tracking/usage.py b/parsl/dataflow/usage_tracking/usage.py
--- a/parsl/dataflow/usage_tracking/usage.py
+++ b/parsl/dataflow/usage_tracking/usage.py
@@ -6,8 +6,10 @@
import json
import logging
import socket
+import sys
from parsl.dataflow.states import States
+from parsl.version import VERSION as PARSL_VERSION
logger = logging.getLogger(__name__)
@@ -48,6 +50,10 @@
self.dfk = dfk
self.config = self.dfk.config
self.uuid = str(uuid.uuid4())
+ self.parsl_version = PARSL_VERSION
+ self.python_version = "{}.{}.{}".format(sys.version_info.major,
+ sys.version_info.minor,
+ sys.version_info.micro)
self.test_mode, self.tracking_enabled = self.check_tracking_enabled()
logger.debug("Tracking status: {}".format(self.tracking_enabled))
logger.debug("Testing mode : {}".format(self.test_mode))
@@ -93,6 +99,8 @@
'uname': hashed_username,
'hname': hashed_hostname,
'test': self.test_mode,
+ 'parsl_v': self.parsl_version,
+ 'python_v': self.python_version,
'start': time.time()}
return json.dumps(message)
| {"golden_diff": "diff --git a/parsl/dataflow/usage_tracking/usage.py b/parsl/dataflow/usage_tracking/usage.py\n--- a/parsl/dataflow/usage_tracking/usage.py\n+++ b/parsl/dataflow/usage_tracking/usage.py\n@@ -6,8 +6,10 @@\n import json\n import logging\n import socket\n+import sys\n \n from parsl.dataflow.states import States\n+from parsl.version import VERSION as PARSL_VERSION\n \n logger = logging.getLogger(__name__)\n \n@@ -48,6 +50,10 @@\n self.dfk = dfk\n self.config = self.dfk.config\n self.uuid = str(uuid.uuid4())\n+ self.parsl_version = PARSL_VERSION\n+ self.python_version = \"{}.{}.{}\".format(sys.version_info.major,\n+ sys.version_info.minor,\n+ sys.version_info.micro)\n self.test_mode, self.tracking_enabled = self.check_tracking_enabled()\n logger.debug(\"Tracking status: {}\".format(self.tracking_enabled))\n logger.debug(\"Testing mode : {}\".format(self.test_mode))\n@@ -93,6 +99,8 @@\n 'uname': hashed_username,\n 'hname': hashed_hostname,\n 'test': self.test_mode,\n+ 'parsl_v': self.parsl_version,\n+ 'python_v': self.python_version,\n 'start': time.time()}\n \n return json.dumps(message)\n", "issue": "User tracking to report parsl version\nUser tracking as described in #34 tracks a range of script attributes and anonymous user information.\r\nIt would help to track the version of Parsl and python. This would help us understand:\r\n1. Rate of upgrade to newer releases\r\n2. Python version in use, and support requirements\r\n3. Support requirements for older parsl releases.\r\n\n", "before_files": [{"content": "import uuid\nimport time\nimport hashlib\nimport os\nimport getpass\nimport json\nimport logging\nimport socket\n\nfrom parsl.dataflow.states import States\n\nlogger = logging.getLogger(__name__)\n\n\nclass UsageTracker (object):\n \"\"\"Anonymized Usage Tracking for Parsl.\n\n Client for this is here : https://github.com/Parsl/parsl_tracking\n This issue captures the discussion that went into functionality\n implemented here : https://github.com/Parsl/parsl/issues/34\n\n \"\"\"\n\n def __init__(self, dfk, ip='52.3.111.203', port=50077,\n domain_name='tracking.parsl-project.org'):\n \"\"\"Initialize usage tracking unless the user has opted-out.\n\n Tracks usage stats by inspecting the internal state of the dfk.\n\n Args:\n - dfk (DFK object) : Data Flow Kernel object\n\n KWargs:\n - ip (string) : IP address\n - port (int) : Port number, Default:50077\n - domain_name (string) : Domain name, will override IP\n Default: tracking.parsl-project.org\n \"\"\"\n if domain_name:\n try:\n self.UDP_IP = socket.gethostbyname(domain_name)\n except Exception:\n logging.debug(\"Could not lookup domain_name, defaulting to 52.3.111.203\")\n self.UDP_IP = ip\n else:\n self.UDP_IP = ip\n self.UDP_PORT = port\n self.dfk = dfk\n self.config = self.dfk.config\n self.uuid = str(uuid.uuid4())\n self.test_mode, self.tracking_enabled = self.check_tracking_enabled()\n logger.debug(\"Tracking status: {}\".format(self.tracking_enabled))\n logger.debug(\"Testing mode : {}\".format(self.test_mode))\n self.initialized = False # Once first message is sent this will be True\n\n def check_tracking_enabled(self):\n \"\"\"By default tracking is enabled.\n\n If Test mode is set via env variable PARSL_TESTING, a test flag is set\n\n Tracking is disabled if :\n 1. config[\"globals\"][\"usageTracking\"] is set to False (Bool)\n 2. Environment variable PARSL_TRACKING is set to false (case insensitive)\n\n \"\"\"\n track = True # By default we track usage\n test = False # By default we are not in testing mode\n\n testvar = str(os.environ.get(\"PARSL_TESTING\", 'None')).lower()\n if testvar == 'true':\n test = True\n\n if self.config and self.config[\"globals\"][\"usageTracking\"] is False:\n track = False\n\n envvar = str(os.environ.get(\"PARSL_TRACKING\", True)).lower()\n if envvar == \"false\":\n track = False\n\n return test, track\n\n def construct_start_message(self):\n \"\"\"Collect preliminary run info at the start of the DFK.\n\n Returns :\n - Message dict dumped as json string, ready for UDP\n \"\"\"\n uname = getpass.getuser().encode('latin1')\n hashed_username = hashlib.sha256(uname).hexdigest()[0:10]\n hname = socket.gethostname().encode('latin1')\n hashed_hostname = hashlib.sha256(hname).hexdigest()[0:10]\n message = {'uuid': self.uuid,\n 'uname': hashed_username,\n 'hname': hashed_hostname,\n 'test': self.test_mode,\n 'start': time.time()}\n\n return json.dumps(message)\n\n def construct_end_message(self):\n \"\"\"Collect the final run information at the time of DFK cleanup.\n\n Returns:\n - Message dict dumped as json string, ready for UDP\n \"\"\"\n app_count = self.dfk.task_count\n\n site_count = 0\n if self.dfk._executors_managed:\n site_count = len(self.dfk.config['sites'])\n\n failed_states = (States.failed, States.dep_fail)\n app_fails = len([t for t in self.dfk.tasks if\n self.dfk.tasks[t]['status'] in failed_states])\n\n message = {'uuid': self.uuid,\n 'end': time.time(),\n 't_apps': app_count,\n 'sites': site_count,\n 'c_time': None,\n 'failed': app_fails,\n 'test': self.test_mode,\n }\n\n return json.dumps(message)\n\n def send_UDP_message(self, message):\n \"\"\"Send UDP message.\"\"\"\n if self.tracking_enabled:\n try:\n sock = socket.socket(socket.AF_INET, socket.SOCK_DGRAM) # UDP\n x = sock.sendto(bytes(message, \"utf-8\"), (self.UDP_IP, self.UDP_PORT))\n sock.close()\n except OSError:\n logger.debug(\"Unable to reach the network to send usage data\")\n x = 0\n else:\n x = -1\n\n return x\n\n def send_message(self):\n \"\"\"Send message over UDP.\n\n If tracking is disables, the bytes_sent will always be set to -1\n\n Returns:\n (bytes_sent, time_taken)\n \"\"\"\n start = time.time()\n message = None\n if not self.initialized:\n message = self.construct_start_message()\n self.initialized = True\n else:\n message = self.construct_end_message()\n\n x = self.send_UDP_message(message)\n end = time.time()\n\n return x, end - start\n\n\nif __name__ == '__main__':\n\n from parsl import *\n\n workers = ThreadPoolExecutor(max_workers=4)\n dfk = DataFlowKernel(executors=[workers])\n\n # ut = UsageTracker(dfk, ip='52.3.111.203')\n ut = UsageTracker(dfk, domain_name='tracking.parsl-project.org')\n\n for i in range(0, 2):\n x = ut.send_message()\n print(x)\n", "path": "parsl/dataflow/usage_tracking/usage.py"}], "after_files": [{"content": "import uuid\nimport time\nimport hashlib\nimport os\nimport getpass\nimport json\nimport logging\nimport socket\nimport sys\n\nfrom parsl.dataflow.states import States\nfrom parsl.version import VERSION as PARSL_VERSION\n\nlogger = logging.getLogger(__name__)\n\n\nclass UsageTracker (object):\n \"\"\"Anonymized Usage Tracking for Parsl.\n\n Client for this is here : https://github.com/Parsl/parsl_tracking\n This issue captures the discussion that went into functionality\n implemented here : https://github.com/Parsl/parsl/issues/34\n\n \"\"\"\n\n def __init__(self, dfk, ip='52.3.111.203', port=50077,\n domain_name='tracking.parsl-project.org'):\n \"\"\"Initialize usage tracking unless the user has opted-out.\n\n Tracks usage stats by inspecting the internal state of the dfk.\n\n Args:\n - dfk (DFK object) : Data Flow Kernel object\n\n KWargs:\n - ip (string) : IP address\n - port (int) : Port number, Default:50077\n - domain_name (string) : Domain name, will override IP\n Default: tracking.parsl-project.org\n \"\"\"\n if domain_name:\n try:\n self.UDP_IP = socket.gethostbyname(domain_name)\n except Exception:\n logging.debug(\"Could not lookup domain_name, defaulting to 52.3.111.203\")\n self.UDP_IP = ip\n else:\n self.UDP_IP = ip\n self.UDP_PORT = port\n self.dfk = dfk\n self.config = self.dfk.config\n self.uuid = str(uuid.uuid4())\n self.parsl_version = PARSL_VERSION\n self.python_version = \"{}.{}.{}\".format(sys.version_info.major,\n sys.version_info.minor,\n sys.version_info.micro)\n self.test_mode, self.tracking_enabled = self.check_tracking_enabled()\n logger.debug(\"Tracking status: {}\".format(self.tracking_enabled))\n logger.debug(\"Testing mode : {}\".format(self.test_mode))\n self.initialized = False # Once first message is sent this will be True\n\n def check_tracking_enabled(self):\n \"\"\"By default tracking is enabled.\n\n If Test mode is set via env variable PARSL_TESTING, a test flag is set\n\n Tracking is disabled if :\n 1. config[\"globals\"][\"usageTracking\"] is set to False (Bool)\n 2. Environment variable PARSL_TRACKING is set to false (case insensitive)\n\n \"\"\"\n track = True # By default we track usage\n test = False # By default we are not in testing mode\n\n testvar = str(os.environ.get(\"PARSL_TESTING\", 'None')).lower()\n if testvar == 'true':\n test = True\n\n if self.config and self.config[\"globals\"][\"usageTracking\"] is False:\n track = False\n\n envvar = str(os.environ.get(\"PARSL_TRACKING\", True)).lower()\n if envvar == \"false\":\n track = False\n\n return test, track\n\n def construct_start_message(self):\n \"\"\"Collect preliminary run info at the start of the DFK.\n\n Returns :\n - Message dict dumped as json string, ready for UDP\n \"\"\"\n uname = getpass.getuser().encode('latin1')\n hashed_username = hashlib.sha256(uname).hexdigest()[0:10]\n hname = socket.gethostname().encode('latin1')\n hashed_hostname = hashlib.sha256(hname).hexdigest()[0:10]\n message = {'uuid': self.uuid,\n 'uname': hashed_username,\n 'hname': hashed_hostname,\n 'test': self.test_mode,\n 'parsl_v': self.parsl_version,\n 'python_v': self.python_version,\n 'start': time.time()}\n\n return json.dumps(message)\n\n def construct_end_message(self):\n \"\"\"Collect the final run information at the time of DFK cleanup.\n\n Returns:\n - Message dict dumped as json string, ready for UDP\n \"\"\"\n app_count = self.dfk.task_count\n\n site_count = 0\n if self.dfk._executors_managed:\n site_count = len(self.dfk.config['sites'])\n\n failed_states = (States.failed, States.dep_fail)\n app_fails = len([t for t in self.dfk.tasks if\n self.dfk.tasks[t]['status'] in failed_states])\n\n message = {'uuid': self.uuid,\n 'end': time.time(),\n 't_apps': app_count,\n 'sites': site_count,\n 'c_time': None,\n 'failed': app_fails,\n 'test': self.test_mode,\n }\n\n return json.dumps(message)\n\n def send_UDP_message(self, message):\n \"\"\"Send UDP message.\"\"\"\n if self.tracking_enabled:\n try:\n sock = socket.socket(socket.AF_INET, socket.SOCK_DGRAM) # UDP\n x = sock.sendto(bytes(message, \"utf-8\"), (self.UDP_IP, self.UDP_PORT))\n sock.close()\n except OSError:\n logger.debug(\"Unable to reach the network to send usage data\")\n x = 0\n else:\n x = -1\n\n return x\n\n def send_message(self):\n \"\"\"Send message over UDP.\n\n If tracking is disables, the bytes_sent will always be set to -1\n\n Returns:\n (bytes_sent, time_taken)\n \"\"\"\n start = time.time()\n message = None\n if not self.initialized:\n message = self.construct_start_message()\n self.initialized = True\n else:\n message = self.construct_end_message()\n\n x = self.send_UDP_message(message)\n end = time.time()\n\n return x, end - start\n\n\nif __name__ == '__main__':\n\n from parsl import *\n\n workers = ThreadPoolExecutor(max_workers=4)\n dfk = DataFlowKernel(executors=[workers])\n\n # ut = UsageTracker(dfk, ip='52.3.111.203')\n ut = UsageTracker(dfk, domain_name='tracking.parsl-project.org')\n\n for i in range(0, 2):\n x = ut.send_message()\n print(x)\n", "path": "parsl/dataflow/usage_tracking/usage.py"}]} | 2,073 | 306 |
gh_patches_debug_30335 | rasdani/github-patches | git_diff | encode__uvicorn-109 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
uvicorn 0.2 not compatible with django channels
daphe server for a project(avilpage) can be started with
```
daphne avilpage.asgi:application
```
uvicorn fails with
```
->uvicorn avilpage.asgi:application
Usage: uvicorn [OPTIONS] APP
Error: Error loading ASGI app. Could not import module "avilpage.asgi".
```
eventhough
```
python -c 'import avilpage.asgi'
```
works fine without any errors.
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `uvicorn/main.py`
Content:
```
1 from uvicorn.protocols.http import H11Protocol, HttpToolsProtocol
2
3 import asyncio
4 import click
5 import importlib
6 import signal
7 import os
8 import logging
9 import sys
10
11
12 LOOP_CHOICES = click.Choice(["uvloop", "asyncio"])
13 LEVEL_CHOICES = click.Choice(["debug", "info", "warning", "error", "critical"])
14 HTTP_CHOICES = click.Choice(["httptools", "h11"])
15 LOG_LEVELS = {
16 "critical": logging.CRITICAL,
17 "error": logging.ERROR,
18 "warning": logging.WARNING,
19 "info": logging.INFO,
20 "debug": logging.DEBUG,
21 }
22 HTTP_PROTOCOLS = {"h11": H11Protocol, "httptools": HttpToolsProtocol}
23
24
25 @click.command()
26 @click.argument("app")
27 @click.option("--host", type=str, default="127.0.0.1", help="Host")
28 @click.option("--port", type=int, default=8000, help="Port")
29 @click.option("--loop", type=LOOP_CHOICES, default="uvloop", help="Event loop")
30 @click.option("--http", type=HTTP_CHOICES, default="httptools", help="HTTP Handler")
31 @click.option("--workers", type=int, default=1, help="Number of worker processes")
32 @click.option("--log-level", type=LEVEL_CHOICES, default="info", help="Log level")
33 def main(app, host: str, port: int, loop: str, http: str, workers: int, log_level: str):
34 log_level = LOG_LEVELS[log_level]
35 logging.basicConfig(format="%(levelname)s: %(message)s", level=log_level)
36
37 app = load_app(app)
38 loop = get_event_loop(loop)
39 logger = logging.getLogger()
40 protocol_class = HTTP_PROTOCOLS[http]
41
42 if workers != 1:
43 raise click.UsageError(
44 'Not yet available. For multiple worker processes, use gunicorn. '
45 'eg. "gunicorn -w 4 -k uvicorn.workers.UvicornWorker".'
46 )
47
48 server = Server(app, host, port, loop, logger, protocol_class)
49 server.run()
50
51
52 def run(app, host="127.0.0.1", port=8000, log_level="info"):
53 log_level = LOG_LEVELS[log_level]
54 logging.basicConfig(format="%(levelname)s: %(message)s", level=log_level)
55
56 loop = get_event_loop("uvloop")
57 logger = logging.getLogger()
58 protocol_class = HttpToolsProtocol
59
60 server = Server(app, host, port, loop, logger, protocol_class)
61 server.run()
62
63
64 def load_app(app):
65 if not isinstance(app, str):
66 return app
67
68 if ":" not in app:
69 message = 'Invalid app string "{app}". Must be in format "<module>:<app>".'
70 raise click.UsageError(message.format(app=app))
71
72 module_str, _, attr = app.partition(":")
73 try:
74 module = importlib.import_module(module_str)
75 except ModuleNotFoundError:
76 message = 'Error loading ASGI app. Could not import module "{module_str}".'
77 raise click.UsageError(message.format(module_str=module_str))
78
79 try:
80 return getattr(module, attr)
81 except AttributeError:
82 message = 'Error loading ASGI app. No attribute "{attr}" found in module "{module_str}".'
83 raise click.UsageError(message.format(attr=attr, module_str=module_str))
84
85
86 def get_event_loop(loop):
87 if loop == "uvloop":
88 import uvloop
89
90 asyncio.get_event_loop().close()
91 asyncio.set_event_loop_policy(uvloop.EventLoopPolicy())
92 return asyncio.get_event_loop()
93
94
95 class Server:
96 def __init__(
97 self,
98 app,
99 host="127.0.0.1",
100 port=8000,
101 loop=None,
102 logger=None,
103 protocol_class=None,
104 ):
105 self.app = app
106 self.host = host
107 self.port = port
108 self.loop = loop or asyncio.get_event_loop()
109 self.logger = logger or logging.getLogger()
110 self.server = None
111 self.should_exit = False
112 self.pid = os.getpid()
113 self.protocol_class = protocol_class
114
115 def set_signal_handlers(self):
116 handled = (signal.SIGQUIT, signal.SIGTERM, signal.SIGINT, signal.SIGABRT)
117 for sig in handled:
118 self.loop.add_signal_handler(sig, self.handle_exit, sig, None)
119
120 def run(self):
121 self.set_signal_handlers()
122 self.loop.run_until_complete(self.create_server())
123 if self.server is not None:
124 message = "* Uvicorn running on http://%s:%d 🦄 (Press CTRL+C to quit)"
125 click.echo(message % (self.host, self.port))
126 self.logger.info("Started worker [{}]".format(self.pid))
127 self.loop.create_task(self.tick())
128 self.loop.run_forever()
129
130 def handle_exit(self, sig, frame):
131 self.logger.warning("Received signal {}. Shutting down.".format(sig.name))
132 self.should_exit = True
133
134 def create_protocol(self):
135 try:
136 return self.protocol_class(app=self.app, loop=self.loop, logger=self.logger)
137 except Exception as exc:
138 self.logger.error(exc)
139 self.should_exit = True
140
141 async def create_server(self):
142 try:
143 self.server = await self.loop.create_server(
144 self.create_protocol, host=self.host, port=self.port
145 )
146 except Exception as exc:
147 self.logger.error(exc)
148
149 async def tick(self):
150 while not self.should_exit:
151 self.protocol_class.tick()
152 await asyncio.sleep(1)
153
154 self.logger.info("Stopping worker [{}]".format(self.pid))
155 self.server.close()
156 await self.server.wait_closed()
157 self.loop.stop()
158
159
160 if __name__ == "__main__":
161 main()
162
```
Path: `uvicorn/__init__.py`
Content:
```
1 from uvicorn.main import main, run
2
3 __version__ = "0.2.3"
4 __all__ = ["main", "run"]
5
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/uvicorn/__init__.py b/uvicorn/__init__.py
--- a/uvicorn/__init__.py
+++ b/uvicorn/__init__.py
@@ -1,4 +1,4 @@
from uvicorn.main import main, run
-__version__ = "0.2.3"
+__version__ = "0.2.4"
__all__ = ["main", "run"]
diff --git a/uvicorn/main.py b/uvicorn/main.py
--- a/uvicorn/main.py
+++ b/uvicorn/main.py
@@ -34,6 +34,7 @@
log_level = LOG_LEVELS[log_level]
logging.basicConfig(format="%(levelname)s: %(message)s", level=log_level)
+ sys.path.insert(0, ".")
app = load_app(app)
loop = get_event_loop(loop)
logger = logging.getLogger()
@@ -69,7 +70,7 @@
message = 'Invalid app string "{app}". Must be in format "<module>:<app>".'
raise click.UsageError(message.format(app=app))
- module_str, _, attr = app.partition(":")
+ module_str, attrs = app.split(":", 1)
try:
module = importlib.import_module(module_str)
except ModuleNotFoundError:
@@ -77,10 +78,13 @@
raise click.UsageError(message.format(module_str=module_str))
try:
- return getattr(module, attr)
+ for attr in attrs.split('.'):
+ asgi_app = getattr(module, attr)
except AttributeError:
- message = 'Error loading ASGI app. No attribute "{attr}" found in module "{module_str}".'
- raise click.UsageError(message.format(attr=attr, module_str=module_str))
+ message = 'Error loading ASGI app. No app "{attrs}" found in module "{module_str}".'
+ raise click.UsageError(message.format(attrs=attrs, module_str=module_str))
+
+ return asgi_app
def get_event_loop(loop):
| {"golden_diff": "diff --git a/uvicorn/__init__.py b/uvicorn/__init__.py\n--- a/uvicorn/__init__.py\n+++ b/uvicorn/__init__.py\n@@ -1,4 +1,4 @@\n from uvicorn.main import main, run\n \n-__version__ = \"0.2.3\"\n+__version__ = \"0.2.4\"\n __all__ = [\"main\", \"run\"]\ndiff --git a/uvicorn/main.py b/uvicorn/main.py\n--- a/uvicorn/main.py\n+++ b/uvicorn/main.py\n@@ -34,6 +34,7 @@\n log_level = LOG_LEVELS[log_level]\n logging.basicConfig(format=\"%(levelname)s: %(message)s\", level=log_level)\n \n+ sys.path.insert(0, \".\")\n app = load_app(app)\n loop = get_event_loop(loop)\n logger = logging.getLogger()\n@@ -69,7 +70,7 @@\n message = 'Invalid app string \"{app}\". Must be in format \"<module>:<app>\".'\n raise click.UsageError(message.format(app=app))\n \n- module_str, _, attr = app.partition(\":\")\n+ module_str, attrs = app.split(\":\", 1)\n try:\n module = importlib.import_module(module_str)\n except ModuleNotFoundError:\n@@ -77,10 +78,13 @@\n raise click.UsageError(message.format(module_str=module_str))\n \n try:\n- return getattr(module, attr)\n+ for attr in attrs.split('.'):\n+ asgi_app = getattr(module, attr)\n except AttributeError:\n- message = 'Error loading ASGI app. No attribute \"{attr}\" found in module \"{module_str}\".'\n- raise click.UsageError(message.format(attr=attr, module_str=module_str))\n+ message = 'Error loading ASGI app. No app \"{attrs}\" found in module \"{module_str}\".'\n+ raise click.UsageError(message.format(attrs=attrs, module_str=module_str))\n+\n+ return asgi_app\n \n \n def get_event_loop(loop):\n", "issue": "uvicorn 0.2 not compatible with django channels\ndaphe server for a project(avilpage) can be started with\r\n\r\n```\r\ndaphne avilpage.asgi:application \r\n```\r\n\r\nuvicorn fails with \r\n\r\n```\r\n->uvicorn avilpage.asgi:application\r\nUsage: uvicorn [OPTIONS] APP\r\n\r\nError: Error loading ASGI app. Could not import module \"avilpage.asgi\".\r\n```\r\n\r\neventhough \r\n\r\n```\r\npython -c 'import avilpage.asgi' \r\n```\r\nworks fine without any errors.\r\n\n", "before_files": [{"content": "from uvicorn.protocols.http import H11Protocol, HttpToolsProtocol\n\nimport asyncio\nimport click\nimport importlib\nimport signal\nimport os\nimport logging\nimport sys\n\n\nLOOP_CHOICES = click.Choice([\"uvloop\", \"asyncio\"])\nLEVEL_CHOICES = click.Choice([\"debug\", \"info\", \"warning\", \"error\", \"critical\"])\nHTTP_CHOICES = click.Choice([\"httptools\", \"h11\"])\nLOG_LEVELS = {\n \"critical\": logging.CRITICAL,\n \"error\": logging.ERROR,\n \"warning\": logging.WARNING,\n \"info\": logging.INFO,\n \"debug\": logging.DEBUG,\n}\nHTTP_PROTOCOLS = {\"h11\": H11Protocol, \"httptools\": HttpToolsProtocol}\n\n\[email protected]()\[email protected](\"app\")\[email protected](\"--host\", type=str, default=\"127.0.0.1\", help=\"Host\")\[email protected](\"--port\", type=int, default=8000, help=\"Port\")\[email protected](\"--loop\", type=LOOP_CHOICES, default=\"uvloop\", help=\"Event loop\")\[email protected](\"--http\", type=HTTP_CHOICES, default=\"httptools\", help=\"HTTP Handler\")\[email protected](\"--workers\", type=int, default=1, help=\"Number of worker processes\")\[email protected](\"--log-level\", type=LEVEL_CHOICES, default=\"info\", help=\"Log level\")\ndef main(app, host: str, port: int, loop: str, http: str, workers: int, log_level: str):\n log_level = LOG_LEVELS[log_level]\n logging.basicConfig(format=\"%(levelname)s: %(message)s\", level=log_level)\n\n app = load_app(app)\n loop = get_event_loop(loop)\n logger = logging.getLogger()\n protocol_class = HTTP_PROTOCOLS[http]\n\n if workers != 1:\n raise click.UsageError(\n 'Not yet available. For multiple worker processes, use gunicorn. '\n 'eg. \"gunicorn -w 4 -k uvicorn.workers.UvicornWorker\".'\n )\n\n server = Server(app, host, port, loop, logger, protocol_class)\n server.run()\n\n\ndef run(app, host=\"127.0.0.1\", port=8000, log_level=\"info\"):\n log_level = LOG_LEVELS[log_level]\n logging.basicConfig(format=\"%(levelname)s: %(message)s\", level=log_level)\n\n loop = get_event_loop(\"uvloop\")\n logger = logging.getLogger()\n protocol_class = HttpToolsProtocol\n\n server = Server(app, host, port, loop, logger, protocol_class)\n server.run()\n\n\ndef load_app(app):\n if not isinstance(app, str):\n return app\n\n if \":\" not in app:\n message = 'Invalid app string \"{app}\". Must be in format \"<module>:<app>\".'\n raise click.UsageError(message.format(app=app))\n\n module_str, _, attr = app.partition(\":\")\n try:\n module = importlib.import_module(module_str)\n except ModuleNotFoundError:\n message = 'Error loading ASGI app. Could not import module \"{module_str}\".'\n raise click.UsageError(message.format(module_str=module_str))\n\n try:\n return getattr(module, attr)\n except AttributeError:\n message = 'Error loading ASGI app. No attribute \"{attr}\" found in module \"{module_str}\".'\n raise click.UsageError(message.format(attr=attr, module_str=module_str))\n\n\ndef get_event_loop(loop):\n if loop == \"uvloop\":\n import uvloop\n\n asyncio.get_event_loop().close()\n asyncio.set_event_loop_policy(uvloop.EventLoopPolicy())\n return asyncio.get_event_loop()\n\n\nclass Server:\n def __init__(\n self,\n app,\n host=\"127.0.0.1\",\n port=8000,\n loop=None,\n logger=None,\n protocol_class=None,\n ):\n self.app = app\n self.host = host\n self.port = port\n self.loop = loop or asyncio.get_event_loop()\n self.logger = logger or logging.getLogger()\n self.server = None\n self.should_exit = False\n self.pid = os.getpid()\n self.protocol_class = protocol_class\n\n def set_signal_handlers(self):\n handled = (signal.SIGQUIT, signal.SIGTERM, signal.SIGINT, signal.SIGABRT)\n for sig in handled:\n self.loop.add_signal_handler(sig, self.handle_exit, sig, None)\n\n def run(self):\n self.set_signal_handlers()\n self.loop.run_until_complete(self.create_server())\n if self.server is not None:\n message = \"* Uvicorn running on http://%s:%d \ud83e\udd84 (Press CTRL+C to quit)\"\n click.echo(message % (self.host, self.port))\n self.logger.info(\"Started worker [{}]\".format(self.pid))\n self.loop.create_task(self.tick())\n self.loop.run_forever()\n\n def handle_exit(self, sig, frame):\n self.logger.warning(\"Received signal {}. Shutting down.\".format(sig.name))\n self.should_exit = True\n\n def create_protocol(self):\n try:\n return self.protocol_class(app=self.app, loop=self.loop, logger=self.logger)\n except Exception as exc:\n self.logger.error(exc)\n self.should_exit = True\n\n async def create_server(self):\n try:\n self.server = await self.loop.create_server(\n self.create_protocol, host=self.host, port=self.port\n )\n except Exception as exc:\n self.logger.error(exc)\n\n async def tick(self):\n while not self.should_exit:\n self.protocol_class.tick()\n await asyncio.sleep(1)\n\n self.logger.info(\"Stopping worker [{}]\".format(self.pid))\n self.server.close()\n await self.server.wait_closed()\n self.loop.stop()\n\n\nif __name__ == \"__main__\":\n main()\n", "path": "uvicorn/main.py"}, {"content": "from uvicorn.main import main, run\n\n__version__ = \"0.2.3\"\n__all__ = [\"main\", \"run\"]\n", "path": "uvicorn/__init__.py"}], "after_files": [{"content": "from uvicorn.protocols.http import H11Protocol, HttpToolsProtocol\n\nimport asyncio\nimport click\nimport importlib\nimport signal\nimport os\nimport logging\nimport sys\n\n\nLOOP_CHOICES = click.Choice([\"uvloop\", \"asyncio\"])\nLEVEL_CHOICES = click.Choice([\"debug\", \"info\", \"warning\", \"error\", \"critical\"])\nHTTP_CHOICES = click.Choice([\"httptools\", \"h11\"])\nLOG_LEVELS = {\n \"critical\": logging.CRITICAL,\n \"error\": logging.ERROR,\n \"warning\": logging.WARNING,\n \"info\": logging.INFO,\n \"debug\": logging.DEBUG,\n}\nHTTP_PROTOCOLS = {\"h11\": H11Protocol, \"httptools\": HttpToolsProtocol}\n\n\[email protected]()\[email protected](\"app\")\[email protected](\"--host\", type=str, default=\"127.0.0.1\", help=\"Host\")\[email protected](\"--port\", type=int, default=8000, help=\"Port\")\[email protected](\"--loop\", type=LOOP_CHOICES, default=\"uvloop\", help=\"Event loop\")\[email protected](\"--http\", type=HTTP_CHOICES, default=\"httptools\", help=\"HTTP Handler\")\[email protected](\"--workers\", type=int, default=1, help=\"Number of worker processes\")\[email protected](\"--log-level\", type=LEVEL_CHOICES, default=\"info\", help=\"Log level\")\ndef main(app, host: str, port: int, loop: str, http: str, workers: int, log_level: str):\n log_level = LOG_LEVELS[log_level]\n logging.basicConfig(format=\"%(levelname)s: %(message)s\", level=log_level)\n\n sys.path.insert(0, \".\")\n app = load_app(app)\n loop = get_event_loop(loop)\n logger = logging.getLogger()\n protocol_class = HTTP_PROTOCOLS[http]\n\n if workers != 1:\n raise click.UsageError(\n 'Not yet available. For multiple worker processes, use gunicorn. '\n 'eg. \"gunicorn -w 4 -k uvicorn.workers.UvicornWorker\".'\n )\n\n server = Server(app, host, port, loop, logger, protocol_class)\n server.run()\n\n\ndef run(app, host=\"127.0.0.1\", port=8000, log_level=\"info\"):\n log_level = LOG_LEVELS[log_level]\n logging.basicConfig(format=\"%(levelname)s: %(message)s\", level=log_level)\n\n loop = get_event_loop(\"uvloop\")\n logger = logging.getLogger()\n protocol_class = HttpToolsProtocol\n\n server = Server(app, host, port, loop, logger, protocol_class)\n server.run()\n\n\ndef load_app(app):\n if not isinstance(app, str):\n return app\n\n if \":\" not in app:\n message = 'Invalid app string \"{app}\". Must be in format \"<module>:<app>\".'\n raise click.UsageError(message.format(app=app))\n\n module_str, attrs = app.split(\":\", 1)\n try:\n module = importlib.import_module(module_str)\n except ModuleNotFoundError:\n message = 'Error loading ASGI app. Could not import module \"{module_str}\".'\n raise click.UsageError(message.format(module_str=module_str))\n\n try:\n for attr in attrs.split('.'):\n asgi_app = getattr(module, attr)\n except AttributeError:\n message = 'Error loading ASGI app. No app \"{attrs}\" found in module \"{module_str}\".'\n raise click.UsageError(message.format(attrs=attrs, module_str=module_str))\n\n return asgi_app\n\n\ndef get_event_loop(loop):\n if loop == \"uvloop\":\n import uvloop\n\n asyncio.get_event_loop().close()\n asyncio.set_event_loop_policy(uvloop.EventLoopPolicy())\n return asyncio.get_event_loop()\n\n\nclass Server:\n def __init__(\n self,\n app,\n host=\"127.0.0.1\",\n port=8000,\n loop=None,\n logger=None,\n protocol_class=None,\n ):\n self.app = app\n self.host = host\n self.port = port\n self.loop = loop or asyncio.get_event_loop()\n self.logger = logger or logging.getLogger()\n self.server = None\n self.should_exit = False\n self.pid = os.getpid()\n self.protocol_class = protocol_class\n\n def set_signal_handlers(self):\n handled = (signal.SIGQUIT, signal.SIGTERM, signal.SIGINT, signal.SIGABRT)\n for sig in handled:\n self.loop.add_signal_handler(sig, self.handle_exit, sig, None)\n\n def run(self):\n self.set_signal_handlers()\n self.loop.run_until_complete(self.create_server())\n if self.server is not None:\n message = \"* Uvicorn running on http://%s:%d \ud83e\udd84 (Press CTRL+C to quit)\"\n click.echo(message % (self.host, self.port))\n self.logger.info(\"Started worker [{}]\".format(self.pid))\n self.loop.create_task(self.tick())\n self.loop.run_forever()\n\n def handle_exit(self, sig, frame):\n self.logger.warning(\"Received signal {}. Shutting down.\".format(sig.name))\n self.should_exit = True\n\n def create_protocol(self):\n try:\n return self.protocol_class(app=self.app, loop=self.loop, logger=self.logger)\n except Exception as exc:\n self.logger.error(exc)\n self.should_exit = True\n\n async def create_server(self):\n try:\n self.server = await self.loop.create_server(\n self.create_protocol, host=self.host, port=self.port\n )\n except Exception as exc:\n self.logger.error(exc)\n\n async def tick(self):\n while not self.should_exit:\n self.protocol_class.tick()\n await asyncio.sleep(1)\n\n self.logger.info(\"Stopping worker [{}]\".format(self.pid))\n self.server.close()\n await self.server.wait_closed()\n self.loop.stop()\n\n\nif __name__ == \"__main__\":\n main()\n", "path": "uvicorn/main.py"}, {"content": "from uvicorn.main import main, run\n\n__version__ = \"0.2.4\"\n__all__ = [\"main\", \"run\"]\n", "path": "uvicorn/__init__.py"}]} | 2,070 | 448 |
gh_patches_debug_31456 | rasdani/github-patches | git_diff | googleapis__python-bigquery-1721 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Make compiled packages such as "grpc" and "proto-plus" optional extras
I believe proto-plus / protobuf / grpc packages were only needed for `google.cloud.bigquery_v2`. These modules have been deprecated and disconnected from the `google.cloud.bigquery` modules in version 3.x. I believe this means we can move those dependencies to "extras" (a.k.a. optional dependencies) and make our package lighter-weight by default.
Follow-up to https://github.com/googleapis/python-bigquery/pull/1282/files#discussion_r928022130 and https://github.com/googleapis/python-bigquery/pull/1282/files#discussion_r928022317
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `setup.py`
Content:
```
1 # Copyright 2018 Google LLC
2 #
3 # Licensed under the Apache License, Version 2.0 (the "License");
4 # you may not use this file except in compliance with the License.
5 # You may obtain a copy of the License at
6 #
7 # http://www.apache.org/licenses/LICENSE-2.0
8 #
9 # Unless required by applicable law or agreed to in writing, software
10 # distributed under the License is distributed on an "AS IS" BASIS,
11 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12 # See the License for the specific language governing permissions and
13 # limitations under the License.
14
15 import io
16 import os
17
18 import setuptools
19
20
21 # Package metadata.
22
23 name = "google-cloud-bigquery"
24 description = "Google BigQuery API client library"
25
26 # Should be one of:
27 # 'Development Status :: 3 - Alpha'
28 # 'Development Status :: 4 - Beta'
29 # 'Development Status :: 5 - Production/Stable'
30 release_status = "Development Status :: 5 - Production/Stable"
31 dependencies = [
32 "grpcio >= 1.47.0, < 2.0dev", # https://github.com/googleapis/python-bigquery/issues/1262
33 "grpcio >= 1.49.1, < 2.0dev; python_version>='3.11'",
34 # NOTE: Maintainers, please do not require google-api-core>=2.x.x
35 # Until this issue is closed
36 # https://github.com/googleapis/google-cloud-python/issues/10566
37 "google-api-core[grpc] >= 1.31.5, <3.0.0dev,!=2.0.*,!=2.1.*,!=2.2.*,!=2.3.0",
38 "proto-plus >= 1.15.0, <2.0.0dev",
39 # NOTE: Maintainers, please do not require google-cloud-core>=2.x.x
40 # Until this issue is closed
41 # https://github.com/googleapis/google-cloud-python/issues/10566
42 "google-cloud-core >= 1.6.0, <3.0.0dev",
43 "google-resumable-media >= 0.6.0, < 3.0dev",
44 "packaging >= 20.0.0",
45 "protobuf>=3.19.5,<5.0.0dev,!=3.20.0,!=3.20.1,!=4.21.0,!=4.21.1,!=4.21.2,!=4.21.3,!=4.21.4,!=4.21.5", # For the legacy proto-based types.
46 "python-dateutil >= 2.7.2, <3.0dev",
47 "requests >= 2.21.0, < 3.0.0dev",
48 ]
49 pyarrow_dependency = "pyarrow >= 3.0.0"
50 extras = {
51 # Keep the no-op bqstorage extra for backward compatibility.
52 # See: https://github.com/googleapis/python-bigquery/issues/757
53 "bqstorage": [
54 "google-cloud-bigquery-storage >= 2.6.0, <3.0.0dev",
55 # Due to an issue in pip's dependency resolver, the `grpc` extra is not
56 # installed, even though `google-cloud-bigquery-storage` specifies it
57 # as `google-api-core[grpc]`. We thus need to explicitly specify it here.
58 # See: https://github.com/googleapis/python-bigquery/issues/83 The
59 # grpc.Channel.close() method isn't added until 1.32.0.
60 # https://github.com/grpc/grpc/pull/15254
61 "grpcio >= 1.47.0, < 2.0dev",
62 "grpcio >= 1.49.1, < 2.0dev; python_version>='3.11'",
63 pyarrow_dependency,
64 ],
65 "pandas": [
66 "pandas>=1.1.0",
67 pyarrow_dependency,
68 "db-dtypes>=0.3.0,<2.0.0dev",
69 ],
70 "ipywidgets": [
71 "ipywidgets>=7.7.0",
72 "ipykernel>=6.0.0",
73 ],
74 "geopandas": ["geopandas>=0.9.0, <1.0dev", "Shapely>=1.8.4, <3.0.0dev"],
75 "ipython": [
76 "ipython>=7.23.1,!=8.1.0",
77 "ipykernel>=6.0.0",
78 ],
79 "tqdm": ["tqdm >= 4.7.4, <5.0.0dev"],
80 "opentelemetry": [
81 "opentelemetry-api >= 1.1.0",
82 "opentelemetry-sdk >= 1.1.0",
83 "opentelemetry-instrumentation >= 0.20b0",
84 ],
85 }
86
87 all_extras = []
88
89 for extra in extras:
90 all_extras.extend(extras[extra])
91
92 extras["all"] = all_extras
93
94 # Setup boilerplate below this line.
95
96 package_root = os.path.abspath(os.path.dirname(__file__))
97
98 readme_filename = os.path.join(package_root, "README.rst")
99 with io.open(readme_filename, encoding="utf-8") as readme_file:
100 readme = readme_file.read()
101
102 version = {}
103 with open(os.path.join(package_root, "google/cloud/bigquery/version.py")) as fp:
104 exec(fp.read(), version)
105 version = version["__version__"]
106
107 # Only include packages under the 'google' namespace. Do not include tests,
108 # benchmarks, etc.
109 packages = [
110 package
111 for package in setuptools.PEP420PackageFinder.find()
112 if package.startswith("google")
113 ]
114
115 # Determine which namespaces are needed.
116 namespaces = ["google"]
117 if "google.cloud" in packages:
118 namespaces.append("google.cloud")
119
120
121 setuptools.setup(
122 name=name,
123 version=version,
124 description=description,
125 long_description=readme,
126 author="Google LLC",
127 author_email="[email protected]",
128 license="Apache 2.0",
129 url="https://github.com/googleapis/python-bigquery",
130 classifiers=[
131 release_status,
132 "Intended Audience :: Developers",
133 "License :: OSI Approved :: Apache Software License",
134 "Programming Language :: Python",
135 "Programming Language :: Python :: 3",
136 "Programming Language :: Python :: 3.7",
137 "Programming Language :: Python :: 3.8",
138 "Programming Language :: Python :: 3.9",
139 "Programming Language :: Python :: 3.10",
140 "Programming Language :: Python :: 3.11",
141 "Operating System :: OS Independent",
142 "Topic :: Internet",
143 ],
144 platforms="Posix; MacOS X; Windows",
145 packages=packages,
146 namespace_packages=namespaces,
147 install_requires=dependencies,
148 extras_require=extras,
149 python_requires=">=3.7",
150 include_package_data=True,
151 zip_safe=False,
152 )
153
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/setup.py b/setup.py
--- a/setup.py
+++ b/setup.py
@@ -29,20 +29,16 @@
# 'Development Status :: 5 - Production/Stable'
release_status = "Development Status :: 5 - Production/Stable"
dependencies = [
- "grpcio >= 1.47.0, < 2.0dev", # https://github.com/googleapis/python-bigquery/issues/1262
- "grpcio >= 1.49.1, < 2.0dev; python_version>='3.11'",
# NOTE: Maintainers, please do not require google-api-core>=2.x.x
# Until this issue is closed
# https://github.com/googleapis/google-cloud-python/issues/10566
- "google-api-core[grpc] >= 1.31.5, <3.0.0dev,!=2.0.*,!=2.1.*,!=2.2.*,!=2.3.0",
- "proto-plus >= 1.15.0, <2.0.0dev",
+ "google-api-core >= 1.31.5, <3.0.0dev,!=2.0.*,!=2.1.*,!=2.2.*,!=2.3.0",
# NOTE: Maintainers, please do not require google-cloud-core>=2.x.x
# Until this issue is closed
# https://github.com/googleapis/google-cloud-python/issues/10566
"google-cloud-core >= 1.6.0, <3.0.0dev",
"google-resumable-media >= 0.6.0, < 3.0dev",
"packaging >= 20.0.0",
- "protobuf>=3.19.5,<5.0.0dev,!=3.20.0,!=3.20.1,!=4.21.0,!=4.21.1,!=4.21.2,!=4.21.3,!=4.21.4,!=4.21.5", # For the legacy proto-based types.
"python-dateutil >= 2.7.2, <3.0dev",
"requests >= 2.21.0, < 3.0.0dev",
]
@@ -82,6 +78,10 @@
"opentelemetry-sdk >= 1.1.0",
"opentelemetry-instrumentation >= 0.20b0",
],
+ "bigquery_v2": [
+ "proto-plus >= 1.15.0, <2.0.0dev",
+ "protobuf>=3.19.5,<5.0.0dev,!=3.20.0,!=3.20.1,!=4.21.0,!=4.21.1,!=4.21.2,!=4.21.3,!=4.21.4,!=4.21.5", # For the legacy proto-based types.
+ ],
}
all_extras = []
| {"golden_diff": "diff --git a/setup.py b/setup.py\n--- a/setup.py\n+++ b/setup.py\n@@ -29,20 +29,16 @@\n # 'Development Status :: 5 - Production/Stable'\n release_status = \"Development Status :: 5 - Production/Stable\"\n dependencies = [\n- \"grpcio >= 1.47.0, < 2.0dev\", # https://github.com/googleapis/python-bigquery/issues/1262\n- \"grpcio >= 1.49.1, < 2.0dev; python_version>='3.11'\",\n # NOTE: Maintainers, please do not require google-api-core>=2.x.x\n # Until this issue is closed\n # https://github.com/googleapis/google-cloud-python/issues/10566\n- \"google-api-core[grpc] >= 1.31.5, <3.0.0dev,!=2.0.*,!=2.1.*,!=2.2.*,!=2.3.0\",\n- \"proto-plus >= 1.15.0, <2.0.0dev\",\n+ \"google-api-core >= 1.31.5, <3.0.0dev,!=2.0.*,!=2.1.*,!=2.2.*,!=2.3.0\",\n # NOTE: Maintainers, please do not require google-cloud-core>=2.x.x\n # Until this issue is closed\n # https://github.com/googleapis/google-cloud-python/issues/10566\n \"google-cloud-core >= 1.6.0, <3.0.0dev\",\n \"google-resumable-media >= 0.6.0, < 3.0dev\",\n \"packaging >= 20.0.0\",\n- \"protobuf>=3.19.5,<5.0.0dev,!=3.20.0,!=3.20.1,!=4.21.0,!=4.21.1,!=4.21.2,!=4.21.3,!=4.21.4,!=4.21.5\", # For the legacy proto-based types.\n \"python-dateutil >= 2.7.2, <3.0dev\",\n \"requests >= 2.21.0, < 3.0.0dev\",\n ]\n@@ -82,6 +78,10 @@\n \"opentelemetry-sdk >= 1.1.0\",\n \"opentelemetry-instrumentation >= 0.20b0\",\n ],\n+ \"bigquery_v2\": [\n+ \"proto-plus >= 1.15.0, <2.0.0dev\",\n+ \"protobuf>=3.19.5,<5.0.0dev,!=3.20.0,!=3.20.1,!=4.21.0,!=4.21.1,!=4.21.2,!=4.21.3,!=4.21.4,!=4.21.5\", # For the legacy proto-based types.\n+ ],\n }\n \n all_extras = []\n", "issue": "Make compiled packages such as \"grpc\" and \"proto-plus\" optional extras\nI believe proto-plus / protobuf / grpc packages were only needed for `google.cloud.bigquery_v2`. These modules have been deprecated and disconnected from the `google.cloud.bigquery` modules in version 3.x. I believe this means we can move those dependencies to \"extras\" (a.k.a. optional dependencies) and make our package lighter-weight by default.\r\n\r\nFollow-up to https://github.com/googleapis/python-bigquery/pull/1282/files#discussion_r928022130 and https://github.com/googleapis/python-bigquery/pull/1282/files#discussion_r928022317\n", "before_files": [{"content": "# Copyright 2018 Google LLC\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nimport io\nimport os\n\nimport setuptools\n\n\n# Package metadata.\n\nname = \"google-cloud-bigquery\"\ndescription = \"Google BigQuery API client library\"\n\n# Should be one of:\n# 'Development Status :: 3 - Alpha'\n# 'Development Status :: 4 - Beta'\n# 'Development Status :: 5 - Production/Stable'\nrelease_status = \"Development Status :: 5 - Production/Stable\"\ndependencies = [\n \"grpcio >= 1.47.0, < 2.0dev\", # https://github.com/googleapis/python-bigquery/issues/1262\n \"grpcio >= 1.49.1, < 2.0dev; python_version>='3.11'\",\n # NOTE: Maintainers, please do not require google-api-core>=2.x.x\n # Until this issue is closed\n # https://github.com/googleapis/google-cloud-python/issues/10566\n \"google-api-core[grpc] >= 1.31.5, <3.0.0dev,!=2.0.*,!=2.1.*,!=2.2.*,!=2.3.0\",\n \"proto-plus >= 1.15.0, <2.0.0dev\",\n # NOTE: Maintainers, please do not require google-cloud-core>=2.x.x\n # Until this issue is closed\n # https://github.com/googleapis/google-cloud-python/issues/10566\n \"google-cloud-core >= 1.6.0, <3.0.0dev\",\n \"google-resumable-media >= 0.6.0, < 3.0dev\",\n \"packaging >= 20.0.0\",\n \"protobuf>=3.19.5,<5.0.0dev,!=3.20.0,!=3.20.1,!=4.21.0,!=4.21.1,!=4.21.2,!=4.21.3,!=4.21.4,!=4.21.5\", # For the legacy proto-based types.\n \"python-dateutil >= 2.7.2, <3.0dev\",\n \"requests >= 2.21.0, < 3.0.0dev\",\n]\npyarrow_dependency = \"pyarrow >= 3.0.0\"\nextras = {\n # Keep the no-op bqstorage extra for backward compatibility.\n # See: https://github.com/googleapis/python-bigquery/issues/757\n \"bqstorage\": [\n \"google-cloud-bigquery-storage >= 2.6.0, <3.0.0dev\",\n # Due to an issue in pip's dependency resolver, the `grpc` extra is not\n # installed, even though `google-cloud-bigquery-storage` specifies it\n # as `google-api-core[grpc]`. We thus need to explicitly specify it here.\n # See: https://github.com/googleapis/python-bigquery/issues/83 The\n # grpc.Channel.close() method isn't added until 1.32.0.\n # https://github.com/grpc/grpc/pull/15254\n \"grpcio >= 1.47.0, < 2.0dev\",\n \"grpcio >= 1.49.1, < 2.0dev; python_version>='3.11'\",\n pyarrow_dependency,\n ],\n \"pandas\": [\n \"pandas>=1.1.0\",\n pyarrow_dependency,\n \"db-dtypes>=0.3.0,<2.0.0dev\",\n ],\n \"ipywidgets\": [\n \"ipywidgets>=7.7.0\",\n \"ipykernel>=6.0.0\",\n ],\n \"geopandas\": [\"geopandas>=0.9.0, <1.0dev\", \"Shapely>=1.8.4, <3.0.0dev\"],\n \"ipython\": [\n \"ipython>=7.23.1,!=8.1.0\",\n \"ipykernel>=6.0.0\",\n ],\n \"tqdm\": [\"tqdm >= 4.7.4, <5.0.0dev\"],\n \"opentelemetry\": [\n \"opentelemetry-api >= 1.1.0\",\n \"opentelemetry-sdk >= 1.1.0\",\n \"opentelemetry-instrumentation >= 0.20b0\",\n ],\n}\n\nall_extras = []\n\nfor extra in extras:\n all_extras.extend(extras[extra])\n\nextras[\"all\"] = all_extras\n\n# Setup boilerplate below this line.\n\npackage_root = os.path.abspath(os.path.dirname(__file__))\n\nreadme_filename = os.path.join(package_root, \"README.rst\")\nwith io.open(readme_filename, encoding=\"utf-8\") as readme_file:\n readme = readme_file.read()\n\nversion = {}\nwith open(os.path.join(package_root, \"google/cloud/bigquery/version.py\")) as fp:\n exec(fp.read(), version)\nversion = version[\"__version__\"]\n\n# Only include packages under the 'google' namespace. Do not include tests,\n# benchmarks, etc.\npackages = [\n package\n for package in setuptools.PEP420PackageFinder.find()\n if package.startswith(\"google\")\n]\n\n# Determine which namespaces are needed.\nnamespaces = [\"google\"]\nif \"google.cloud\" in packages:\n namespaces.append(\"google.cloud\")\n\n\nsetuptools.setup(\n name=name,\n version=version,\n description=description,\n long_description=readme,\n author=\"Google LLC\",\n author_email=\"[email protected]\",\n license=\"Apache 2.0\",\n url=\"https://github.com/googleapis/python-bigquery\",\n classifiers=[\n release_status,\n \"Intended Audience :: Developers\",\n \"License :: OSI Approved :: Apache Software License\",\n \"Programming Language :: Python\",\n \"Programming Language :: Python :: 3\",\n \"Programming Language :: Python :: 3.7\",\n \"Programming Language :: Python :: 3.8\",\n \"Programming Language :: Python :: 3.9\",\n \"Programming Language :: Python :: 3.10\",\n \"Programming Language :: Python :: 3.11\",\n \"Operating System :: OS Independent\",\n \"Topic :: Internet\",\n ],\n platforms=\"Posix; MacOS X; Windows\",\n packages=packages,\n namespace_packages=namespaces,\n install_requires=dependencies,\n extras_require=extras,\n python_requires=\">=3.7\",\n include_package_data=True,\n zip_safe=False,\n)\n", "path": "setup.py"}], "after_files": [{"content": "# Copyright 2018 Google LLC\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nimport io\nimport os\n\nimport setuptools\n\n\n# Package metadata.\n\nname = \"google-cloud-bigquery\"\ndescription = \"Google BigQuery API client library\"\n\n# Should be one of:\n# 'Development Status :: 3 - Alpha'\n# 'Development Status :: 4 - Beta'\n# 'Development Status :: 5 - Production/Stable'\nrelease_status = \"Development Status :: 5 - Production/Stable\"\ndependencies = [\n # NOTE: Maintainers, please do not require google-api-core>=2.x.x\n # Until this issue is closed\n # https://github.com/googleapis/google-cloud-python/issues/10566\n \"google-api-core >= 1.31.5, <3.0.0dev,!=2.0.*,!=2.1.*,!=2.2.*,!=2.3.0\",\n # NOTE: Maintainers, please do not require google-cloud-core>=2.x.x\n # Until this issue is closed\n # https://github.com/googleapis/google-cloud-python/issues/10566\n \"google-cloud-core >= 1.6.0, <3.0.0dev\",\n \"google-resumable-media >= 0.6.0, < 3.0dev\",\n \"packaging >= 20.0.0\",\n \"python-dateutil >= 2.7.2, <3.0dev\",\n \"requests >= 2.21.0, < 3.0.0dev\",\n]\npyarrow_dependency = \"pyarrow >= 3.0.0\"\nextras = {\n # Keep the no-op bqstorage extra for backward compatibility.\n # See: https://github.com/googleapis/python-bigquery/issues/757\n \"bqstorage\": [\n \"google-cloud-bigquery-storage >= 2.6.0, <3.0.0dev\",\n # Due to an issue in pip's dependency resolver, the `grpc` extra is not\n # installed, even though `google-cloud-bigquery-storage` specifies it\n # as `google-api-core[grpc]`. We thus need to explicitly specify it here.\n # See: https://github.com/googleapis/python-bigquery/issues/83 The\n # grpc.Channel.close() method isn't added until 1.32.0.\n # https://github.com/grpc/grpc/pull/15254\n \"grpcio >= 1.47.0, < 2.0dev\",\n \"grpcio >= 1.49.1, < 2.0dev; python_version>='3.11'\",\n pyarrow_dependency,\n ],\n \"pandas\": [\n \"pandas>=1.1.0\",\n pyarrow_dependency,\n \"db-dtypes>=0.3.0,<2.0.0dev\",\n ],\n \"ipywidgets\": [\n \"ipywidgets>=7.7.0\",\n \"ipykernel>=6.0.0\",\n ],\n \"geopandas\": [\"geopandas>=0.9.0, <1.0dev\", \"Shapely>=1.8.4, <3.0.0dev\"],\n \"ipython\": [\n \"ipython>=7.23.1,!=8.1.0\",\n \"ipykernel>=6.0.0\",\n ],\n \"tqdm\": [\"tqdm >= 4.7.4, <5.0.0dev\"],\n \"opentelemetry\": [\n \"opentelemetry-api >= 1.1.0\",\n \"opentelemetry-sdk >= 1.1.0\",\n \"opentelemetry-instrumentation >= 0.20b0\",\n ],\n \"bigquery_v2\": [\n \"proto-plus >= 1.15.0, <2.0.0dev\",\n \"protobuf>=3.19.5,<5.0.0dev,!=3.20.0,!=3.20.1,!=4.21.0,!=4.21.1,!=4.21.2,!=4.21.3,!=4.21.4,!=4.21.5\", # For the legacy proto-based types.\n ],\n}\n\nall_extras = []\n\nfor extra in extras:\n all_extras.extend(extras[extra])\n\nextras[\"all\"] = all_extras\n\n# Setup boilerplate below this line.\n\npackage_root = os.path.abspath(os.path.dirname(__file__))\n\nreadme_filename = os.path.join(package_root, \"README.rst\")\nwith io.open(readme_filename, encoding=\"utf-8\") as readme_file:\n readme = readme_file.read()\n\nversion = {}\nwith open(os.path.join(package_root, \"google/cloud/bigquery/version.py\")) as fp:\n exec(fp.read(), version)\nversion = version[\"__version__\"]\n\n# Only include packages under the 'google' namespace. Do not include tests,\n# benchmarks, etc.\npackages = [\n package\n for package in setuptools.PEP420PackageFinder.find()\n if package.startswith(\"google\")\n]\n\n# Determine which namespaces are needed.\nnamespaces = [\"google\"]\nif \"google.cloud\" in packages:\n namespaces.append(\"google.cloud\")\n\n\nsetuptools.setup(\n name=name,\n version=version,\n description=description,\n long_description=readme,\n author=\"Google LLC\",\n author_email=\"[email protected]\",\n license=\"Apache 2.0\",\n url=\"https://github.com/googleapis/python-bigquery\",\n classifiers=[\n release_status,\n \"Intended Audience :: Developers\",\n \"License :: OSI Approved :: Apache Software License\",\n \"Programming Language :: Python\",\n \"Programming Language :: Python :: 3\",\n \"Programming Language :: Python :: 3.7\",\n \"Programming Language :: Python :: 3.8\",\n \"Programming Language :: Python :: 3.9\",\n \"Programming Language :: Python :: 3.10\",\n \"Programming Language :: Python :: 3.11\",\n \"Operating System :: OS Independent\",\n \"Topic :: Internet\",\n ],\n platforms=\"Posix; MacOS X; Windows\",\n packages=packages,\n namespace_packages=namespaces,\n install_requires=dependencies,\n extras_require=extras,\n python_requires=\">=3.7\",\n include_package_data=True,\n zip_safe=False,\n)\n", "path": "setup.py"}]} | 2,327 | 707 |
gh_patches_debug_54074 | rasdani/github-patches | git_diff | e-valuation__EvaP-290 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Update installation instructions
https://evap.readthedocs.org/en/latest/installation.html
Someone should follow these instructions and see if they are correct and complete.
The short version at https://github.com/fsr-itse/EvaP should also be checked again.
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `evap/settings.py`
Content:
```
1 # Django settings for evap project.
2
3 # automatically determine SITE_ROOT, used for absolute paths below
4 import os.path
5 SITE_ROOT = os.path.dirname(os.path.realpath(__file__))
6
7 DEBUG = True
8 TEMPLATE_DEBUG = DEBUG
9
10 ADMINS = (
11 # ('Your Name', '[email protected]'),
12 )
13
14 MANAGERS = ADMINS
15
16 DATABASES = {
17 'default': {
18 'ENGINE': 'django.db.backends.sqlite3', # 'postgresql_psycopg2', 'postgresql', 'mysql', 'sqlite3' or 'oracle'.
19 'NAME': os.path.join(SITE_ROOT, 'database.sqlite3'), # Or path to database file if using sqlite3.
20 'USER': '', # Not used with sqlite3.
21 'PASSWORD': '', # Not used with sqlite3.
22 'HOST': '', # Set to empty string for localhost. Not used with sqlite3.
23 'PORT': '', # Set to empty string for default. Not used with sqlite3.
24 }
25 }
26
27 CACHES = {
28 'default': {
29 # 'BACKEND': 'django.core.cache.backends.dummy.DummyCache',
30 'BACKEND': 'django.core.cache.backends.locmem.LocMemCache',
31 }
32 }
33
34 # config for feedback links
35 FEEDBACK_EMAIL = "webmaster@localhost"
36 TRACKER_URL = "https://github.com/fsr-itse/EvaP"
37
38 # config for mail system
39 DEFAULT_FROM_EMAIL = "webmaster@localhost"
40 REPLY_TO_EMAIL = DEFAULT_FROM_EMAIL
41 if DEBUG:
42 EMAIL_BACKEND = 'django.core.mail.backends.console.EmailBackend'
43
44 # key authentication settings
45 LOGIN_KEY_VALIDITY = 210 # days, so roughly 7 months
46
47 # minimum answers needed for publishing
48 MIN_ANSWER_COUNT = 2
49 MIN_ANSWER_PERCENTAGE = 0.2
50
51 # days before end date to send reminder
52 REMIND_X_DAYS_AHEAD_OF_END_DATE = 2
53
54 # email domains for the internal users of the hosting institution used to
55 # figure out who can login with username and password and who needs a login key
56 INSTITUTION_EMAIL_DOMAINS = ["hpi.uni-potsdam.de", "student.hpi.uni-potsdam.de"]
57
58 # Local time zone for this installation. Choices can be found here:
59 # http://en.wikipedia.org/wiki/List_of_tz_zones_by_name
60 # although not all choices may be available on all operating systems.
61 # On Unix systems, a value of None will cause Django to use the same
62 # timezone as the operating system.
63 # If running in a Windows environment this must be set to the same as your
64 # system time zone.
65 TIME_ZONE = 'Europe/Berlin'
66
67 # Language code for this installation. All choices can be found here:
68 # http://www.i18nguy.com/unicode/language-identifiers.html
69 LANGUAGE_CODE = 'en'
70
71 LANGUAGES = (
72 ('en', "English"),
73 ('de', "Deutsch"),
74 )
75
76 SITE_ID = 1
77
78 # If you set this to False, Django will make some optimizations so as not
79 # to load the internationalization machinery.
80 USE_I18N = True
81
82 # If you set this to False, Django will not format dates, numbers and
83 # calendars according to the current locale
84 USE_L10N = True
85
86 # Locale paths
87 LOCALE_PATHS = (
88 os.path.join(SITE_ROOT, "locale"),
89 )
90
91 # Absolute filesystem path to the directory that will hold user-uploaded files.
92 # Example: "/home/media/media.lawrence.com/media/"
93 MEDIA_ROOT = os.path.join(SITE_ROOT, "upload")
94
95 # URL that handles the media served from MEDIA_ROOT. Make sure to use a
96 # trailing slash.
97 # Examples: "http://media.lawrence.com/media/", "http://example.com/media/"
98 MEDIA_URL = '/media/'
99
100 # Absolute path to the directory static files should be collected to.
101 # Don't put anything in this directory yourself; store your static files
102 # in apps' "static/" subdirectories and in STATICFILES_DIRS.
103 # Example: "/home/media/media.lawrence.com/static/"
104 STATIC_ROOT = os.path.join(SITE_ROOT, "staticfiles")
105
106 # URL prefix for static files.
107 # Example: "http://media.lawrence.com/static/"
108 STATIC_URL = '/static/'
109
110 # URL prefix for admin static files -- CSS, JavaScript and images.
111 # Make sure to use a trailing slash.
112 # Examples: "http://foo.com/static/admin/", "/static/admin/".
113 ADMIN_MEDIA_PREFIX = '/static/admin/'
114
115 # Additional locations of static files
116 STATICFILES_DIRS = (
117 # Put strings here, like "/home/html/static" or "C:/www/django/static".
118 # Always use forward slashes, even on Windows.
119 # Don't forget to use absolute paths, not relative paths.
120 os.path.join(SITE_ROOT, "static"),
121 )
122
123 # List of finder classes that know how to find static files in
124 # various locations.
125 STATICFILES_FINDERS = (
126 'django.contrib.staticfiles.finders.FileSystemFinder',
127 'django.contrib.staticfiles.finders.AppDirectoriesFinder',
128 # 'django.contrib.staticfiles.finders.DefaultStorageFinder',
129 )
130
131 # Make this unique, and don't share it with anybody.
132 SECRET_KEY = 'k9-)vh3c_dtm6bpi7j(!*s_^91v0!ekjt_#o&0i$e22tnn^-vb'
133
134 # List of callables that know how to import templates from various sources.
135 TEMPLATE_LOADERS = (
136 'django.template.loaders.filesystem.Loader',
137 'django.template.loaders.app_directories.Loader',
138 # 'django.template.loaders.eggs.Loader',
139 )
140
141 TEMPLATE_CONTEXT_PROCESSORS = (
142 "django.contrib.auth.context_processors.auth",
143 "django.core.context_processors.debug",
144 "django.core.context_processors.i18n",
145 "django.core.context_processors.media",
146 "django.core.context_processors.static",
147 "django.core.context_processors.request",
148 "django.contrib.messages.context_processors.messages",
149 )
150
151 MIDDLEWARE_CLASSES = (
152 'django.middleware.common.CommonMiddleware',
153 'django.contrib.sessions.middleware.SessionMiddleware',
154 'django.middleware.locale.LocaleMiddleware',
155 'django.middleware.csrf.CsrfViewMiddleware',
156 'django.contrib.auth.middleware.AuthenticationMiddleware',
157 'django.contrib.messages.middleware.MessageMiddleware',
158 'evap.evaluation.auth.RequestAuthMiddleware',
159 'evap.evaluation.403.Django403Middleware',
160 )
161
162 AUTHENTICATION_BACKENDS = (
163 'evap.evaluation.auth.RequestAuthUserBackend',
164 'django.contrib.auth.backends.ModelBackend',
165 )
166
167 LOGIN_URL = "/"
168
169 ROOT_URLCONF = 'evap.urls'
170
171 TEMPLATE_DIRS = (
172 # Put strings here, like "/home/html/django_templates" or "C:/www/django/templates".
173 # Always use forward slashes, even on Windows.
174 # Don't forget to use absolute paths, not relative paths.
175 os.path.join(SITE_ROOT, "templates"),
176 )
177
178 INSTALLED_APPS = (
179 'django.contrib.auth',
180 'django.contrib.contenttypes',
181 'django.contrib.sessions',
182 'django.contrib.sites',
183 'django.contrib.messages',
184 'django.contrib.staticfiles',
185 'django.contrib.admin',
186 'south',
187 'widget_tweaks',
188 'evap.evaluation',
189 'evap.fsr',
190 'evap.results',
191 'evap.student',
192 'evap.contributor',
193 )
194 if not DEBUG:
195 INSTALLED_APPS += (
196 'raven.contrib.django.raven_compat',
197 )
198
199 RAVEN_CONFIG = {
200 'dsn': 'http://public:[email protected]/1',
201 }
202
203 # A sample logging configuration. The only tangible logging
204 # performed by this configuration is to send an email to
205 # the site admins on every HTTP 500 error.
206 # See http://docs.djangoproject.com/en/dev/topics/logging for
207 # more details on how to customize your logging configuration.
208 LOGGING = {
209 'version': 1,
210 'disable_existing_loggers': False,
211 'filters': {
212 'require_debug_false': {
213 '()': 'django.utils.log.RequireDebugFalse'
214 }
215 },
216 'handlers': {
217 'sentry': {
218 'level': 'ERROR',
219 'class': 'raven.contrib.django.handlers.SentryHandler',
220 },
221 'mail_admins': {
222 'level': 'ERROR',
223 'filters': ['require_debug_false'],
224 'class': 'django.utils.log.AdminEmailHandler'
225 },
226 'console': {
227 'level': 'INFO',
228 'class': 'logging.StreamHandler'
229 }
230 },
231 'loggers': {
232 'django.request': {
233 'handlers': ['mail_admins'],
234 'level': 'ERROR',
235 'propagate': True,
236 },
237 'evap.evaluation.management.commands.import_evaj': {
238 'handlers': ['console'],
239 'level': 'INFO'
240 },
241 'raven': {
242 'level': 'DEBUG',
243 'handlers': ['console'],
244 'propagate': False,
245 },
246 'sentry.errors': {
247 'level': 'DEBUG',
248 'handlers': ['console'],
249 'propagate': False,
250 }
251 }
252 }
253
254 # Create a local_settings.py if you want to override settings per machine
255 # or user, e.g. for development or different settings in deployments using
256 # multiple servers.
257 _LOCAL_SETTINGS_FILENAME = os.path.join(SITE_ROOT, "localsettings.py")
258 if os.path.exists(_LOCAL_SETTINGS_FILENAME):
259 execfile(_LOCAL_SETTINGS_FILENAME)
260 del _LOCAL_SETTINGS_FILENAME
261
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/evap/settings.py b/evap/settings.py
--- a/evap/settings.py
+++ b/evap/settings.py
@@ -251,7 +251,7 @@
}
}
-# Create a local_settings.py if you want to override settings per machine
+# Create a localsettings.py if you want to override settings per machine
# or user, e.g. for development or different settings in deployments using
# multiple servers.
_LOCAL_SETTINGS_FILENAME = os.path.join(SITE_ROOT, "localsettings.py")
| {"golden_diff": "diff --git a/evap/settings.py b/evap/settings.py\n--- a/evap/settings.py\n+++ b/evap/settings.py\n@@ -251,7 +251,7 @@\n }\n }\n \n-# Create a local_settings.py if you want to override settings per machine\n+# Create a localsettings.py if you want to override settings per machine\n # or user, e.g. for development or different settings in deployments using\n # multiple servers.\n _LOCAL_SETTINGS_FILENAME = os.path.join(SITE_ROOT, \"localsettings.py\")\n", "issue": "Update installation instructions\nhttps://evap.readthedocs.org/en/latest/installation.html\n\nSomeone should follow these instructions and see if they are correct and complete.\n\nThe short version at https://github.com/fsr-itse/EvaP should also be checked again.\n\n", "before_files": [{"content": "# Django settings for evap project.\n\n# automatically determine SITE_ROOT, used for absolute paths below\nimport os.path\nSITE_ROOT = os.path.dirname(os.path.realpath(__file__))\n\nDEBUG = True\nTEMPLATE_DEBUG = DEBUG\n\nADMINS = (\n # ('Your Name', '[email protected]'),\n)\n\nMANAGERS = ADMINS\n\nDATABASES = {\n 'default': {\n 'ENGINE': 'django.db.backends.sqlite3', # 'postgresql_psycopg2', 'postgresql', 'mysql', 'sqlite3' or 'oracle'.\n 'NAME': os.path.join(SITE_ROOT, 'database.sqlite3'), # Or path to database file if using sqlite3.\n 'USER': '', # Not used with sqlite3.\n 'PASSWORD': '', # Not used with sqlite3.\n 'HOST': '', # Set to empty string for localhost. Not used with sqlite3.\n 'PORT': '', # Set to empty string for default. Not used with sqlite3.\n }\n}\n\nCACHES = {\n 'default': {\n # 'BACKEND': 'django.core.cache.backends.dummy.DummyCache',\n 'BACKEND': 'django.core.cache.backends.locmem.LocMemCache',\n }\n}\n\n# config for feedback links\nFEEDBACK_EMAIL = \"webmaster@localhost\"\nTRACKER_URL = \"https://github.com/fsr-itse/EvaP\"\n\n# config for mail system\nDEFAULT_FROM_EMAIL = \"webmaster@localhost\"\nREPLY_TO_EMAIL = DEFAULT_FROM_EMAIL\nif DEBUG:\n EMAIL_BACKEND = 'django.core.mail.backends.console.EmailBackend'\n\n# key authentication settings\nLOGIN_KEY_VALIDITY = 210 # days, so roughly 7 months\n\n# minimum answers needed for publishing\nMIN_ANSWER_COUNT = 2\nMIN_ANSWER_PERCENTAGE = 0.2\n\n# days before end date to send reminder\nREMIND_X_DAYS_AHEAD_OF_END_DATE = 2\n\n# email domains for the internal users of the hosting institution used to\n# figure out who can login with username and password and who needs a login key\nINSTITUTION_EMAIL_DOMAINS = [\"hpi.uni-potsdam.de\", \"student.hpi.uni-potsdam.de\"]\n\n# Local time zone for this installation. Choices can be found here:\n# http://en.wikipedia.org/wiki/List_of_tz_zones_by_name\n# although not all choices may be available on all operating systems.\n# On Unix systems, a value of None will cause Django to use the same\n# timezone as the operating system.\n# If running in a Windows environment this must be set to the same as your\n# system time zone.\nTIME_ZONE = 'Europe/Berlin'\n\n# Language code for this installation. All choices can be found here:\n# http://www.i18nguy.com/unicode/language-identifiers.html\nLANGUAGE_CODE = 'en'\n\nLANGUAGES = (\n ('en', \"English\"),\n ('de', \"Deutsch\"),\n)\n\nSITE_ID = 1\n\n# If you set this to False, Django will make some optimizations so as not\n# to load the internationalization machinery.\nUSE_I18N = True\n\n# If you set this to False, Django will not format dates, numbers and\n# calendars according to the current locale\nUSE_L10N = True\n\n# Locale paths\nLOCALE_PATHS = (\n os.path.join(SITE_ROOT, \"locale\"),\n)\n\n# Absolute filesystem path to the directory that will hold user-uploaded files.\n# Example: \"/home/media/media.lawrence.com/media/\"\nMEDIA_ROOT = os.path.join(SITE_ROOT, \"upload\")\n\n# URL that handles the media served from MEDIA_ROOT. Make sure to use a\n# trailing slash.\n# Examples: \"http://media.lawrence.com/media/\", \"http://example.com/media/\"\nMEDIA_URL = '/media/'\n\n# Absolute path to the directory static files should be collected to.\n# Don't put anything in this directory yourself; store your static files\n# in apps' \"static/\" subdirectories and in STATICFILES_DIRS.\n# Example: \"/home/media/media.lawrence.com/static/\"\nSTATIC_ROOT = os.path.join(SITE_ROOT, \"staticfiles\")\n\n# URL prefix for static files.\n# Example: \"http://media.lawrence.com/static/\"\nSTATIC_URL = '/static/'\n\n# URL prefix for admin static files -- CSS, JavaScript and images.\n# Make sure to use a trailing slash.\n# Examples: \"http://foo.com/static/admin/\", \"/static/admin/\".\nADMIN_MEDIA_PREFIX = '/static/admin/'\n\n# Additional locations of static files\nSTATICFILES_DIRS = (\n # Put strings here, like \"/home/html/static\" or \"C:/www/django/static\".\n # Always use forward slashes, even on Windows.\n # Don't forget to use absolute paths, not relative paths.\n os.path.join(SITE_ROOT, \"static\"),\n)\n\n# List of finder classes that know how to find static files in\n# various locations.\nSTATICFILES_FINDERS = (\n 'django.contrib.staticfiles.finders.FileSystemFinder',\n 'django.contrib.staticfiles.finders.AppDirectoriesFinder',\n# 'django.contrib.staticfiles.finders.DefaultStorageFinder',\n)\n\n# Make this unique, and don't share it with anybody.\nSECRET_KEY = 'k9-)vh3c_dtm6bpi7j(!*s_^91v0!ekjt_#o&0i$e22tnn^-vb'\n\n# List of callables that know how to import templates from various sources.\nTEMPLATE_LOADERS = (\n 'django.template.loaders.filesystem.Loader',\n 'django.template.loaders.app_directories.Loader',\n# 'django.template.loaders.eggs.Loader',\n)\n\nTEMPLATE_CONTEXT_PROCESSORS = (\n \"django.contrib.auth.context_processors.auth\",\n \"django.core.context_processors.debug\",\n \"django.core.context_processors.i18n\",\n \"django.core.context_processors.media\",\n \"django.core.context_processors.static\",\n \"django.core.context_processors.request\",\n \"django.contrib.messages.context_processors.messages\",\n)\n\nMIDDLEWARE_CLASSES = (\n 'django.middleware.common.CommonMiddleware',\n 'django.contrib.sessions.middleware.SessionMiddleware',\n 'django.middleware.locale.LocaleMiddleware',\n 'django.middleware.csrf.CsrfViewMiddleware',\n 'django.contrib.auth.middleware.AuthenticationMiddleware',\n 'django.contrib.messages.middleware.MessageMiddleware',\n 'evap.evaluation.auth.RequestAuthMiddleware',\n 'evap.evaluation.403.Django403Middleware',\n)\n\nAUTHENTICATION_BACKENDS = (\n 'evap.evaluation.auth.RequestAuthUserBackend',\n 'django.contrib.auth.backends.ModelBackend',\n)\n\nLOGIN_URL = \"/\"\n\nROOT_URLCONF = 'evap.urls'\n\nTEMPLATE_DIRS = (\n # Put strings here, like \"/home/html/django_templates\" or \"C:/www/django/templates\".\n # Always use forward slashes, even on Windows.\n # Don't forget to use absolute paths, not relative paths.\n os.path.join(SITE_ROOT, \"templates\"),\n)\n\nINSTALLED_APPS = (\n 'django.contrib.auth',\n 'django.contrib.contenttypes',\n 'django.contrib.sessions',\n 'django.contrib.sites',\n 'django.contrib.messages',\n 'django.contrib.staticfiles',\n 'django.contrib.admin',\n 'south',\n 'widget_tweaks',\n 'evap.evaluation',\n 'evap.fsr',\n 'evap.results',\n 'evap.student',\n 'evap.contributor',\n)\nif not DEBUG:\n INSTALLED_APPS += (\n 'raven.contrib.django.raven_compat',\n )\n\nRAVEN_CONFIG = {\n 'dsn': 'http://public:[email protected]/1',\n}\n\n# A sample logging configuration. The only tangible logging\n# performed by this configuration is to send an email to\n# the site admins on every HTTP 500 error.\n# See http://docs.djangoproject.com/en/dev/topics/logging for\n# more details on how to customize your logging configuration.\nLOGGING = {\n 'version': 1,\n 'disable_existing_loggers': False,\n 'filters': {\n 'require_debug_false': {\n '()': 'django.utils.log.RequireDebugFalse'\n }\n },\n 'handlers': {\n 'sentry': {\n 'level': 'ERROR',\n 'class': 'raven.contrib.django.handlers.SentryHandler',\n },\n 'mail_admins': {\n 'level': 'ERROR',\n 'filters': ['require_debug_false'],\n 'class': 'django.utils.log.AdminEmailHandler'\n },\n 'console': {\n 'level': 'INFO',\n 'class': 'logging.StreamHandler'\n }\n },\n 'loggers': {\n 'django.request': {\n 'handlers': ['mail_admins'],\n 'level': 'ERROR',\n 'propagate': True,\n },\n 'evap.evaluation.management.commands.import_evaj': {\n 'handlers': ['console'],\n 'level': 'INFO'\n },\n 'raven': {\n 'level': 'DEBUG',\n 'handlers': ['console'],\n 'propagate': False,\n },\n 'sentry.errors': {\n 'level': 'DEBUG',\n 'handlers': ['console'],\n 'propagate': False,\n }\n }\n}\n\n# Create a local_settings.py if you want to override settings per machine\n# or user, e.g. for development or different settings in deployments using\n# multiple servers.\n_LOCAL_SETTINGS_FILENAME = os.path.join(SITE_ROOT, \"localsettings.py\")\nif os.path.exists(_LOCAL_SETTINGS_FILENAME):\n execfile(_LOCAL_SETTINGS_FILENAME)\ndel _LOCAL_SETTINGS_FILENAME\n", "path": "evap/settings.py"}], "after_files": [{"content": "# Django settings for evap project.\n\n# automatically determine SITE_ROOT, used for absolute paths below\nimport os.path\nSITE_ROOT = os.path.dirname(os.path.realpath(__file__))\n\nDEBUG = True\nTEMPLATE_DEBUG = DEBUG\n\nADMINS = (\n # ('Your Name', '[email protected]'),\n)\n\nMANAGERS = ADMINS\n\nDATABASES = {\n 'default': {\n 'ENGINE': 'django.db.backends.sqlite3', # 'postgresql_psycopg2', 'postgresql', 'mysql', 'sqlite3' or 'oracle'.\n 'NAME': os.path.join(SITE_ROOT, 'database.sqlite3'), # Or path to database file if using sqlite3.\n 'USER': '', # Not used with sqlite3.\n 'PASSWORD': '', # Not used with sqlite3.\n 'HOST': '', # Set to empty string for localhost. Not used with sqlite3.\n 'PORT': '', # Set to empty string for default. Not used with sqlite3.\n }\n}\n\nCACHES = {\n 'default': {\n # 'BACKEND': 'django.core.cache.backends.dummy.DummyCache',\n 'BACKEND': 'django.core.cache.backends.locmem.LocMemCache',\n }\n}\n\n# config for feedback links\nFEEDBACK_EMAIL = \"webmaster@localhost\"\nTRACKER_URL = \"https://github.com/fsr-itse/EvaP\"\n\n# config for mail system\nDEFAULT_FROM_EMAIL = \"webmaster@localhost\"\nREPLY_TO_EMAIL = DEFAULT_FROM_EMAIL\nif DEBUG:\n EMAIL_BACKEND = 'django.core.mail.backends.console.EmailBackend'\n\n# key authentication settings\nLOGIN_KEY_VALIDITY = 210 # days, so roughly 7 months\n\n# minimum answers needed for publishing\nMIN_ANSWER_COUNT = 2\nMIN_ANSWER_PERCENTAGE = 0.2\n\n# days before end date to send reminder\nREMIND_X_DAYS_AHEAD_OF_END_DATE = 2\n\n# email domains for the internal users of the hosting institution used to\n# figure out who can login with username and password and who needs a login key\nINSTITUTION_EMAIL_DOMAINS = [\"hpi.uni-potsdam.de\", \"student.hpi.uni-potsdam.de\"]\n\n# Local time zone for this installation. Choices can be found here:\n# http://en.wikipedia.org/wiki/List_of_tz_zones_by_name\n# although not all choices may be available on all operating systems.\n# On Unix systems, a value of None will cause Django to use the same\n# timezone as the operating system.\n# If running in a Windows environment this must be set to the same as your\n# system time zone.\nTIME_ZONE = 'Europe/Berlin'\n\n# Language code for this installation. All choices can be found here:\n# http://www.i18nguy.com/unicode/language-identifiers.html\nLANGUAGE_CODE = 'en'\n\nLANGUAGES = (\n ('en', \"English\"),\n ('de', \"Deutsch\"),\n)\n\nSITE_ID = 1\n\n# If you set this to False, Django will make some optimizations so as not\n# to load the internationalization machinery.\nUSE_I18N = True\n\n# If you set this to False, Django will not format dates, numbers and\n# calendars according to the current locale\nUSE_L10N = True\n\n# Locale paths\nLOCALE_PATHS = (\n os.path.join(SITE_ROOT, \"locale\"),\n)\n\n# Absolute filesystem path to the directory that will hold user-uploaded files.\n# Example: \"/home/media/media.lawrence.com/media/\"\nMEDIA_ROOT = os.path.join(SITE_ROOT, \"upload\")\n\n# URL that handles the media served from MEDIA_ROOT. Make sure to use a\n# trailing slash.\n# Examples: \"http://media.lawrence.com/media/\", \"http://example.com/media/\"\nMEDIA_URL = '/media/'\n\n# Absolute path to the directory static files should be collected to.\n# Don't put anything in this directory yourself; store your static files\n# in apps' \"static/\" subdirectories and in STATICFILES_DIRS.\n# Example: \"/home/media/media.lawrence.com/static/\"\nSTATIC_ROOT = os.path.join(SITE_ROOT, \"staticfiles\")\n\n# URL prefix for static files.\n# Example: \"http://media.lawrence.com/static/\"\nSTATIC_URL = '/static/'\n\n# URL prefix for admin static files -- CSS, JavaScript and images.\n# Make sure to use a trailing slash.\n# Examples: \"http://foo.com/static/admin/\", \"/static/admin/\".\nADMIN_MEDIA_PREFIX = '/static/admin/'\n\n# Additional locations of static files\nSTATICFILES_DIRS = (\n # Put strings here, like \"/home/html/static\" or \"C:/www/django/static\".\n # Always use forward slashes, even on Windows.\n # Don't forget to use absolute paths, not relative paths.\n os.path.join(SITE_ROOT, \"static\"),\n)\n\n# List of finder classes that know how to find static files in\n# various locations.\nSTATICFILES_FINDERS = (\n 'django.contrib.staticfiles.finders.FileSystemFinder',\n 'django.contrib.staticfiles.finders.AppDirectoriesFinder',\n# 'django.contrib.staticfiles.finders.DefaultStorageFinder',\n)\n\n# Make this unique, and don't share it with anybody.\nSECRET_KEY = 'k9-)vh3c_dtm6bpi7j(!*s_^91v0!ekjt_#o&0i$e22tnn^-vb'\n\n# List of callables that know how to import templates from various sources.\nTEMPLATE_LOADERS = (\n 'django.template.loaders.filesystem.Loader',\n 'django.template.loaders.app_directories.Loader',\n# 'django.template.loaders.eggs.Loader',\n)\n\nTEMPLATE_CONTEXT_PROCESSORS = (\n \"django.contrib.auth.context_processors.auth\",\n \"django.core.context_processors.debug\",\n \"django.core.context_processors.i18n\",\n \"django.core.context_processors.media\",\n \"django.core.context_processors.static\",\n \"django.core.context_processors.request\",\n \"django.contrib.messages.context_processors.messages\",\n)\n\nMIDDLEWARE_CLASSES = (\n 'django.middleware.common.CommonMiddleware',\n 'django.contrib.sessions.middleware.SessionMiddleware',\n 'django.middleware.locale.LocaleMiddleware',\n 'django.middleware.csrf.CsrfViewMiddleware',\n 'django.contrib.auth.middleware.AuthenticationMiddleware',\n 'django.contrib.messages.middleware.MessageMiddleware',\n 'evap.evaluation.auth.RequestAuthMiddleware',\n 'evap.evaluation.403.Django403Middleware',\n)\n\nAUTHENTICATION_BACKENDS = (\n 'evap.evaluation.auth.RequestAuthUserBackend',\n 'django.contrib.auth.backends.ModelBackend',\n)\n\nLOGIN_URL = \"/\"\n\nROOT_URLCONF = 'evap.urls'\n\nTEMPLATE_DIRS = (\n # Put strings here, like \"/home/html/django_templates\" or \"C:/www/django/templates\".\n # Always use forward slashes, even on Windows.\n # Don't forget to use absolute paths, not relative paths.\n os.path.join(SITE_ROOT, \"templates\"),\n)\n\nINSTALLED_APPS = (\n 'django.contrib.auth',\n 'django.contrib.contenttypes',\n 'django.contrib.sessions',\n 'django.contrib.sites',\n 'django.contrib.messages',\n 'django.contrib.staticfiles',\n 'django.contrib.admin',\n 'south',\n 'widget_tweaks',\n 'evap.evaluation',\n 'evap.fsr',\n 'evap.results',\n 'evap.student',\n 'evap.contributor',\n)\nif not DEBUG:\n INSTALLED_APPS += (\n 'raven.contrib.django.raven_compat',\n )\n\nRAVEN_CONFIG = {\n 'dsn': 'http://public:[email protected]/1',\n}\n\n# A sample logging configuration. The only tangible logging\n# performed by this configuration is to send an email to\n# the site admins on every HTTP 500 error.\n# See http://docs.djangoproject.com/en/dev/topics/logging for\n# more details on how to customize your logging configuration.\nLOGGING = {\n 'version': 1,\n 'disable_existing_loggers': False,\n 'filters': {\n 'require_debug_false': {\n '()': 'django.utils.log.RequireDebugFalse'\n }\n },\n 'handlers': {\n 'sentry': {\n 'level': 'ERROR',\n 'class': 'raven.contrib.django.handlers.SentryHandler',\n },\n 'mail_admins': {\n 'level': 'ERROR',\n 'filters': ['require_debug_false'],\n 'class': 'django.utils.log.AdminEmailHandler'\n },\n 'console': {\n 'level': 'INFO',\n 'class': 'logging.StreamHandler'\n }\n },\n 'loggers': {\n 'django.request': {\n 'handlers': ['mail_admins'],\n 'level': 'ERROR',\n 'propagate': True,\n },\n 'evap.evaluation.management.commands.import_evaj': {\n 'handlers': ['console'],\n 'level': 'INFO'\n },\n 'raven': {\n 'level': 'DEBUG',\n 'handlers': ['console'],\n 'propagate': False,\n },\n 'sentry.errors': {\n 'level': 'DEBUG',\n 'handlers': ['console'],\n 'propagate': False,\n }\n }\n}\n\n# Create a localsettings.py if you want to override settings per machine\n# or user, e.g. for development or different settings in deployments using\n# multiple servers.\n_LOCAL_SETTINGS_FILENAME = os.path.join(SITE_ROOT, \"localsettings.py\")\nif os.path.exists(_LOCAL_SETTINGS_FILENAME):\n execfile(_LOCAL_SETTINGS_FILENAME)\ndel _LOCAL_SETTINGS_FILENAME\n", "path": "evap/settings.py"}]} | 3,013 | 118 |
gh_patches_debug_16603 | rasdani/github-patches | git_diff | paperless-ngx__paperless-ngx-1220 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
[BUG] OS reports file as busy – wrong error message?!
### Description
Hello everyone,
I have setup paperless-ngx in Docker in my Synology NAS. I have created a bind mount for the `data/consume` folder and threw in some pdf files. However I always got an error `OS reports file as busy still`. Now, I've made sure that no other client is accessing the file. And the file can be opened from any other network device and the NAS web interface. Turns out, this was actually a permissions issue with wrong User IDs. So, the error message is factually not correct.
So I am under the impression that the error message is not very precise. From my little Python knowledge, I've looked[ at the code](https://github.com/paperless-ngx/paperless-ngx/blob/4df065d8d524870ec18e8fbf2fc488449939a044/src/documents/management/commands/document_consumer.py#L74) and it seems to be a very [generic Exception](https://docs.python.org/3/library/exceptions.html#OSError). It could also be thrown, when the disk is full, etc. There is a specific error number in the Exception. Would it make sense, to add that to the error output in the log and/or to change the message?
### Steps to reproduce
There's already a test for it, if you want to trigger the error message on purpose.
### Webserver logs
```bash
[2022-06-20 10:22:37,016] [WARNING] [paperless.management.consumer] Not consuming file /data/consume/E79221B2X115146_06202022_122344_000002.pdf: OS reports file as busy still
[2022-06-20 10:26:20,714] [WARNING] [paperless.management.consumer] Not consuming file /data/consume/E79221B2X115146_06202022_122344_000003.pdf: OS reports file as busy still
[2022-06-20 10:26:47,094] [WARNING] [paperless.management.consumer] Not consuming file /data/consume/.DS_Store: Unknown file extension.
[2022-06-20 10:26:48,738] [WARNING] [paperless.management.consumer] Not consuming file /data/consume/Fragebogen-Mehrfachbeschäftigung-Minijob.pdf: OS reports file as busy still
[2022-06-20 10:26:53,120] [WARNING] [paperless.management.consumer] Not consuming file /data/consume/.DS_Store: Unknown file extension.
```
### Paperless-ngx version
1.7.1
### Host OS
Synology DSM 7
### Installation method
Docker
### Browser
_No response_
### Configuration changes
_No response_
### Other
_No response_
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `src/documents/management/commands/document_consumer.py`
Content:
```
1 import logging
2 import os
3 from pathlib import Path
4 from pathlib import PurePath
5 from threading import Thread
6 from time import monotonic
7 from time import sleep
8 from typing import Final
9
10 from django.conf import settings
11 from django.core.management.base import BaseCommand
12 from django.core.management.base import CommandError
13 from django_q.tasks import async_task
14 from documents.models import Tag
15 from documents.parsers import is_file_ext_supported
16 from watchdog.events import FileSystemEventHandler
17 from watchdog.observers.polling import PollingObserver
18
19 try:
20 from inotifyrecursive import INotify, flags
21 except ImportError:
22 INotify = flags = None
23
24 logger = logging.getLogger("paperless.management.consumer")
25
26
27 def _tags_from_path(filepath):
28 """Walk up the directory tree from filepath to CONSUMPTION_DIR
29 and get or create Tag IDs for every directory.
30 """
31 normalized_consumption_dir = os.path.abspath(
32 os.path.normpath(settings.CONSUMPTION_DIR),
33 )
34 tag_ids = set()
35 path_parts = Path(filepath).relative_to(normalized_consumption_dir).parent.parts
36 for part in path_parts:
37 tag_ids.add(
38 Tag.objects.get_or_create(name__iexact=part, defaults={"name": part})[0].pk,
39 )
40
41 return tag_ids
42
43
44 def _is_ignored(filepath: str) -> bool:
45 normalized_consumption_dir = os.path.abspath(
46 os.path.normpath(settings.CONSUMPTION_DIR),
47 )
48 filepath_relative = PurePath(filepath).relative_to(normalized_consumption_dir)
49 return any(filepath_relative.match(p) for p in settings.CONSUMER_IGNORE_PATTERNS)
50
51
52 def _consume(filepath):
53 if os.path.isdir(filepath) or _is_ignored(filepath):
54 return
55
56 if not os.path.isfile(filepath):
57 logger.debug(f"Not consuming file {filepath}: File has moved.")
58 return
59
60 if not is_file_ext_supported(os.path.splitext(filepath)[1]):
61 logger.warning(f"Not consuming file {filepath}: Unknown file extension.")
62 return
63
64 # Total wait time: up to 500ms
65 os_error_retry_count: Final[int] = 50
66 os_error_retry_wait: Final[float] = 0.01
67
68 read_try_count = 0
69 file_open_ok = False
70
71 while (read_try_count < os_error_retry_count) and not file_open_ok:
72 try:
73 with open(filepath, "rb"):
74 file_open_ok = True
75 except OSError:
76 read_try_count += 1
77 sleep(os_error_retry_wait)
78
79 if read_try_count >= os_error_retry_count:
80 logger.warning(f"Not consuming file {filepath}: OS reports file as busy still")
81 return
82
83 tag_ids = None
84 try:
85 if settings.CONSUMER_SUBDIRS_AS_TAGS:
86 tag_ids = _tags_from_path(filepath)
87 except Exception:
88 logger.exception("Error creating tags from path")
89
90 try:
91 logger.info(f"Adding {filepath} to the task queue.")
92 async_task(
93 "documents.tasks.consume_file",
94 filepath,
95 override_tag_ids=tag_ids if tag_ids else None,
96 task_name=os.path.basename(filepath)[:100],
97 )
98 except Exception:
99 # Catch all so that the consumer won't crash.
100 # This is also what the test case is listening for to check for
101 # errors.
102 logger.exception("Error while consuming document")
103
104
105 def _consume_wait_unmodified(file):
106 if _is_ignored(file):
107 return
108
109 logger.debug(f"Waiting for file {file} to remain unmodified")
110 mtime = -1
111 size = -1
112 current_try = 0
113 while current_try < settings.CONSUMER_POLLING_RETRY_COUNT:
114 try:
115 stat_data = os.stat(file)
116 new_mtime = stat_data.st_mtime
117 new_size = stat_data.st_size
118 except FileNotFoundError:
119 logger.debug(
120 f"File {file} moved while waiting for it to remain " f"unmodified.",
121 )
122 return
123 if new_mtime == mtime and new_size == size:
124 _consume(file)
125 return
126 mtime = new_mtime
127 size = new_size
128 sleep(settings.CONSUMER_POLLING_DELAY)
129 current_try += 1
130
131 logger.error(f"Timeout while waiting on file {file} to remain unmodified.")
132
133
134 class Handler(FileSystemEventHandler):
135 def on_created(self, event):
136 Thread(target=_consume_wait_unmodified, args=(event.src_path,)).start()
137
138 def on_moved(self, event):
139 Thread(target=_consume_wait_unmodified, args=(event.dest_path,)).start()
140
141
142 class Command(BaseCommand):
143 """
144 On every iteration of an infinite loop, consume what we can from the
145 consumption directory.
146 """
147
148 # This is here primarily for the tests and is irrelevant in production.
149 stop_flag = False
150
151 observer = None
152
153 def add_arguments(self, parser):
154 parser.add_argument(
155 "directory",
156 default=settings.CONSUMPTION_DIR,
157 nargs="?",
158 help="The consumption directory.",
159 )
160 parser.add_argument("--oneshot", action="store_true", help="Run only once.")
161
162 def handle(self, *args, **options):
163 directory = options["directory"]
164 recursive = settings.CONSUMER_RECURSIVE
165
166 if not directory:
167 raise CommandError("CONSUMPTION_DIR does not appear to be set.")
168
169 directory = os.path.abspath(directory)
170
171 if not os.path.isdir(directory):
172 raise CommandError(f"Consumption directory {directory} does not exist")
173
174 if recursive:
175 for dirpath, _, filenames in os.walk(directory):
176 for filename in filenames:
177 filepath = os.path.join(dirpath, filename)
178 _consume(filepath)
179 else:
180 for entry in os.scandir(directory):
181 _consume(entry.path)
182
183 if options["oneshot"]:
184 return
185
186 if settings.CONSUMER_POLLING == 0 and INotify:
187 self.handle_inotify(directory, recursive)
188 else:
189 self.handle_polling(directory, recursive)
190
191 logger.debug("Consumer exiting.")
192
193 def handle_polling(self, directory, recursive):
194 logger.info(f"Polling directory for changes: {directory}")
195 self.observer = PollingObserver(timeout=settings.CONSUMER_POLLING)
196 self.observer.schedule(Handler(), directory, recursive=recursive)
197 self.observer.start()
198 try:
199 while self.observer.is_alive():
200 self.observer.join(1)
201 if self.stop_flag:
202 self.observer.stop()
203 except KeyboardInterrupt:
204 self.observer.stop()
205 self.observer.join()
206
207 def handle_inotify(self, directory, recursive):
208 logger.info(f"Using inotify to watch directory for changes: {directory}")
209
210 inotify = INotify()
211 inotify_flags = flags.CLOSE_WRITE | flags.MOVED_TO
212 if recursive:
213 descriptor = inotify.add_watch_recursive(directory, inotify_flags)
214 else:
215 descriptor = inotify.add_watch(directory, inotify_flags)
216
217 try:
218
219 inotify_debounce: Final[float] = settings.CONSUMER_INOTIFY_DELAY
220 notified_files = {}
221
222 while not self.stop_flag:
223
224 for event in inotify.read(timeout=1000):
225 if recursive:
226 path = inotify.get_path(event.wd)
227 else:
228 path = directory
229 filepath = os.path.join(path, event.name)
230 notified_files[filepath] = monotonic()
231
232 # Check the files against the timeout
233 still_waiting = {}
234 for filepath in notified_files:
235 # Time of the last inotify event for this file
236 last_event_time = notified_files[filepath]
237
238 # Current time - last time over the configured timeout
239 waited_long_enough = (
240 monotonic() - last_event_time
241 ) > inotify_debounce
242
243 # Also make sure the file exists still, some scanners might write a
244 # temporary file first
245 file_still_exists = os.path.exists(filepath) and os.path.isfile(
246 filepath,
247 )
248
249 if waited_long_enough and file_still_exists:
250 _consume(filepath)
251 elif file_still_exists:
252 still_waiting[filepath] = last_event_time
253
254 # These files are still waiting to hit the timeout
255 notified_files = still_waiting
256
257 except KeyboardInterrupt:
258 pass
259
260 inotify.rm_watch(descriptor)
261 inotify.close()
262
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/src/documents/management/commands/document_consumer.py b/src/documents/management/commands/document_consumer.py
--- a/src/documents/management/commands/document_consumer.py
+++ b/src/documents/management/commands/document_consumer.py
@@ -67,17 +67,19 @@
read_try_count = 0
file_open_ok = False
+ os_error_str = None
while (read_try_count < os_error_retry_count) and not file_open_ok:
try:
with open(filepath, "rb"):
file_open_ok = True
- except OSError:
+ except OSError as e:
read_try_count += 1
+ os_error_str = str(e)
sleep(os_error_retry_wait)
if read_try_count >= os_error_retry_count:
- logger.warning(f"Not consuming file {filepath}: OS reports file as busy still")
+ logger.warning(f"Not consuming file {filepath}: OS reports {os_error_str}")
return
tag_ids = None
| {"golden_diff": "diff --git a/src/documents/management/commands/document_consumer.py b/src/documents/management/commands/document_consumer.py\n--- a/src/documents/management/commands/document_consumer.py\n+++ b/src/documents/management/commands/document_consumer.py\n@@ -67,17 +67,19 @@\n \n read_try_count = 0\n file_open_ok = False\n+ os_error_str = None\n \n while (read_try_count < os_error_retry_count) and not file_open_ok:\n try:\n with open(filepath, \"rb\"):\n file_open_ok = True\n- except OSError:\n+ except OSError as e:\n read_try_count += 1\n+ os_error_str = str(e)\n sleep(os_error_retry_wait)\n \n if read_try_count >= os_error_retry_count:\n- logger.warning(f\"Not consuming file {filepath}: OS reports file as busy still\")\n+ logger.warning(f\"Not consuming file {filepath}: OS reports {os_error_str}\")\n return\n \n tag_ids = None\n", "issue": "[BUG] OS reports file as busy \u2013 wrong error message?!\n### Description\n\nHello everyone,\r\n\r\nI have setup paperless-ngx in Docker in my Synology NAS. I have created a bind mount for the `data/consume` folder and threw in some pdf files. However I always got an error `OS reports file as busy still`. Now, I've made sure that no other client is accessing the file. And the file can be opened from any other network device and the NAS web interface. Turns out, this was actually a permissions issue with wrong User IDs. So, the error message is factually not correct.\r\n\r\nSo I am under the impression that the error message is not very precise. From my little Python knowledge, I've looked[ at the code](https://github.com/paperless-ngx/paperless-ngx/blob/4df065d8d524870ec18e8fbf2fc488449939a044/src/documents/management/commands/document_consumer.py#L74) and it seems to be a very [generic Exception](https://docs.python.org/3/library/exceptions.html#OSError). It could also be thrown, when the disk is full, etc. There is a specific error number in the Exception. Would it make sense, to add that to the error output in the log and/or to change the message?\n\n### Steps to reproduce\n\nThere's already a test for it, if you want to trigger the error message on purpose.\n\n### Webserver logs\n\n```bash\n[2022-06-20 10:22:37,016] [WARNING] [paperless.management.consumer] Not consuming file /data/consume/E79221B2X115146_06202022_122344_000002.pdf: OS reports file as busy still\r\n\r\n[2022-06-20 10:26:20,714] [WARNING] [paperless.management.consumer] Not consuming file /data/consume/E79221B2X115146_06202022_122344_000003.pdf: OS reports file as busy still\r\n\r\n[2022-06-20 10:26:47,094] [WARNING] [paperless.management.consumer] Not consuming file /data/consume/.DS_Store: Unknown file extension.\r\n\r\n[2022-06-20 10:26:48,738] [WARNING] [paperless.management.consumer] Not consuming file /data/consume/Fragebogen-Mehrfachbesch\u00e4ftigung-Minijob.pdf: OS reports file as busy still\r\n\r\n[2022-06-20 10:26:53,120] [WARNING] [paperless.management.consumer] Not consuming file /data/consume/.DS_Store: Unknown file extension.\n```\n\n\n### Paperless-ngx version\n\n1.7.1\n\n### Host OS\n\nSynology DSM 7\n\n### Installation method\n\nDocker\n\n### Browser\n\n_No response_\n\n### Configuration changes\n\n_No response_\n\n### Other\n\n_No response_\n", "before_files": [{"content": "import logging\nimport os\nfrom pathlib import Path\nfrom pathlib import PurePath\nfrom threading import Thread\nfrom time import monotonic\nfrom time import sleep\nfrom typing import Final\n\nfrom django.conf import settings\nfrom django.core.management.base import BaseCommand\nfrom django.core.management.base import CommandError\nfrom django_q.tasks import async_task\nfrom documents.models import Tag\nfrom documents.parsers import is_file_ext_supported\nfrom watchdog.events import FileSystemEventHandler\nfrom watchdog.observers.polling import PollingObserver\n\ntry:\n from inotifyrecursive import INotify, flags\nexcept ImportError:\n INotify = flags = None\n\nlogger = logging.getLogger(\"paperless.management.consumer\")\n\n\ndef _tags_from_path(filepath):\n \"\"\"Walk up the directory tree from filepath to CONSUMPTION_DIR\n and get or create Tag IDs for every directory.\n \"\"\"\n normalized_consumption_dir = os.path.abspath(\n os.path.normpath(settings.CONSUMPTION_DIR),\n )\n tag_ids = set()\n path_parts = Path(filepath).relative_to(normalized_consumption_dir).parent.parts\n for part in path_parts:\n tag_ids.add(\n Tag.objects.get_or_create(name__iexact=part, defaults={\"name\": part})[0].pk,\n )\n\n return tag_ids\n\n\ndef _is_ignored(filepath: str) -> bool:\n normalized_consumption_dir = os.path.abspath(\n os.path.normpath(settings.CONSUMPTION_DIR),\n )\n filepath_relative = PurePath(filepath).relative_to(normalized_consumption_dir)\n return any(filepath_relative.match(p) for p in settings.CONSUMER_IGNORE_PATTERNS)\n\n\ndef _consume(filepath):\n if os.path.isdir(filepath) or _is_ignored(filepath):\n return\n\n if not os.path.isfile(filepath):\n logger.debug(f\"Not consuming file {filepath}: File has moved.\")\n return\n\n if not is_file_ext_supported(os.path.splitext(filepath)[1]):\n logger.warning(f\"Not consuming file {filepath}: Unknown file extension.\")\n return\n\n # Total wait time: up to 500ms\n os_error_retry_count: Final[int] = 50\n os_error_retry_wait: Final[float] = 0.01\n\n read_try_count = 0\n file_open_ok = False\n\n while (read_try_count < os_error_retry_count) and not file_open_ok:\n try:\n with open(filepath, \"rb\"):\n file_open_ok = True\n except OSError:\n read_try_count += 1\n sleep(os_error_retry_wait)\n\n if read_try_count >= os_error_retry_count:\n logger.warning(f\"Not consuming file {filepath}: OS reports file as busy still\")\n return\n\n tag_ids = None\n try:\n if settings.CONSUMER_SUBDIRS_AS_TAGS:\n tag_ids = _tags_from_path(filepath)\n except Exception:\n logger.exception(\"Error creating tags from path\")\n\n try:\n logger.info(f\"Adding {filepath} to the task queue.\")\n async_task(\n \"documents.tasks.consume_file\",\n filepath,\n override_tag_ids=tag_ids if tag_ids else None,\n task_name=os.path.basename(filepath)[:100],\n )\n except Exception:\n # Catch all so that the consumer won't crash.\n # This is also what the test case is listening for to check for\n # errors.\n logger.exception(\"Error while consuming document\")\n\n\ndef _consume_wait_unmodified(file):\n if _is_ignored(file):\n return\n\n logger.debug(f\"Waiting for file {file} to remain unmodified\")\n mtime = -1\n size = -1\n current_try = 0\n while current_try < settings.CONSUMER_POLLING_RETRY_COUNT:\n try:\n stat_data = os.stat(file)\n new_mtime = stat_data.st_mtime\n new_size = stat_data.st_size\n except FileNotFoundError:\n logger.debug(\n f\"File {file} moved while waiting for it to remain \" f\"unmodified.\",\n )\n return\n if new_mtime == mtime and new_size == size:\n _consume(file)\n return\n mtime = new_mtime\n size = new_size\n sleep(settings.CONSUMER_POLLING_DELAY)\n current_try += 1\n\n logger.error(f\"Timeout while waiting on file {file} to remain unmodified.\")\n\n\nclass Handler(FileSystemEventHandler):\n def on_created(self, event):\n Thread(target=_consume_wait_unmodified, args=(event.src_path,)).start()\n\n def on_moved(self, event):\n Thread(target=_consume_wait_unmodified, args=(event.dest_path,)).start()\n\n\nclass Command(BaseCommand):\n \"\"\"\n On every iteration of an infinite loop, consume what we can from the\n consumption directory.\n \"\"\"\n\n # This is here primarily for the tests and is irrelevant in production.\n stop_flag = False\n\n observer = None\n\n def add_arguments(self, parser):\n parser.add_argument(\n \"directory\",\n default=settings.CONSUMPTION_DIR,\n nargs=\"?\",\n help=\"The consumption directory.\",\n )\n parser.add_argument(\"--oneshot\", action=\"store_true\", help=\"Run only once.\")\n\n def handle(self, *args, **options):\n directory = options[\"directory\"]\n recursive = settings.CONSUMER_RECURSIVE\n\n if not directory:\n raise CommandError(\"CONSUMPTION_DIR does not appear to be set.\")\n\n directory = os.path.abspath(directory)\n\n if not os.path.isdir(directory):\n raise CommandError(f\"Consumption directory {directory} does not exist\")\n\n if recursive:\n for dirpath, _, filenames in os.walk(directory):\n for filename in filenames:\n filepath = os.path.join(dirpath, filename)\n _consume(filepath)\n else:\n for entry in os.scandir(directory):\n _consume(entry.path)\n\n if options[\"oneshot\"]:\n return\n\n if settings.CONSUMER_POLLING == 0 and INotify:\n self.handle_inotify(directory, recursive)\n else:\n self.handle_polling(directory, recursive)\n\n logger.debug(\"Consumer exiting.\")\n\n def handle_polling(self, directory, recursive):\n logger.info(f\"Polling directory for changes: {directory}\")\n self.observer = PollingObserver(timeout=settings.CONSUMER_POLLING)\n self.observer.schedule(Handler(), directory, recursive=recursive)\n self.observer.start()\n try:\n while self.observer.is_alive():\n self.observer.join(1)\n if self.stop_flag:\n self.observer.stop()\n except KeyboardInterrupt:\n self.observer.stop()\n self.observer.join()\n\n def handle_inotify(self, directory, recursive):\n logger.info(f\"Using inotify to watch directory for changes: {directory}\")\n\n inotify = INotify()\n inotify_flags = flags.CLOSE_WRITE | flags.MOVED_TO\n if recursive:\n descriptor = inotify.add_watch_recursive(directory, inotify_flags)\n else:\n descriptor = inotify.add_watch(directory, inotify_flags)\n\n try:\n\n inotify_debounce: Final[float] = settings.CONSUMER_INOTIFY_DELAY\n notified_files = {}\n\n while not self.stop_flag:\n\n for event in inotify.read(timeout=1000):\n if recursive:\n path = inotify.get_path(event.wd)\n else:\n path = directory\n filepath = os.path.join(path, event.name)\n notified_files[filepath] = monotonic()\n\n # Check the files against the timeout\n still_waiting = {}\n for filepath in notified_files:\n # Time of the last inotify event for this file\n last_event_time = notified_files[filepath]\n\n # Current time - last time over the configured timeout\n waited_long_enough = (\n monotonic() - last_event_time\n ) > inotify_debounce\n\n # Also make sure the file exists still, some scanners might write a\n # temporary file first\n file_still_exists = os.path.exists(filepath) and os.path.isfile(\n filepath,\n )\n\n if waited_long_enough and file_still_exists:\n _consume(filepath)\n elif file_still_exists:\n still_waiting[filepath] = last_event_time\n\n # These files are still waiting to hit the timeout\n notified_files = still_waiting\n\n except KeyboardInterrupt:\n pass\n\n inotify.rm_watch(descriptor)\n inotify.close()\n", "path": "src/documents/management/commands/document_consumer.py"}], "after_files": [{"content": "import logging\nimport os\nfrom pathlib import Path\nfrom pathlib import PurePath\nfrom threading import Thread\nfrom time import monotonic\nfrom time import sleep\nfrom typing import Final\n\nfrom django.conf import settings\nfrom django.core.management.base import BaseCommand\nfrom django.core.management.base import CommandError\nfrom django_q.tasks import async_task\nfrom documents.models import Tag\nfrom documents.parsers import is_file_ext_supported\nfrom watchdog.events import FileSystemEventHandler\nfrom watchdog.observers.polling import PollingObserver\n\ntry:\n from inotifyrecursive import INotify, flags\nexcept ImportError:\n INotify = flags = None\n\nlogger = logging.getLogger(\"paperless.management.consumer\")\n\n\ndef _tags_from_path(filepath):\n \"\"\"Walk up the directory tree from filepath to CONSUMPTION_DIR\n and get or create Tag IDs for every directory.\n \"\"\"\n normalized_consumption_dir = os.path.abspath(\n os.path.normpath(settings.CONSUMPTION_DIR),\n )\n tag_ids = set()\n path_parts = Path(filepath).relative_to(normalized_consumption_dir).parent.parts\n for part in path_parts:\n tag_ids.add(\n Tag.objects.get_or_create(name__iexact=part, defaults={\"name\": part})[0].pk,\n )\n\n return tag_ids\n\n\ndef _is_ignored(filepath: str) -> bool:\n normalized_consumption_dir = os.path.abspath(\n os.path.normpath(settings.CONSUMPTION_DIR),\n )\n filepath_relative = PurePath(filepath).relative_to(normalized_consumption_dir)\n return any(filepath_relative.match(p) for p in settings.CONSUMER_IGNORE_PATTERNS)\n\n\ndef _consume(filepath):\n if os.path.isdir(filepath) or _is_ignored(filepath):\n return\n\n if not os.path.isfile(filepath):\n logger.debug(f\"Not consuming file {filepath}: File has moved.\")\n return\n\n if not is_file_ext_supported(os.path.splitext(filepath)[1]):\n logger.warning(f\"Not consuming file {filepath}: Unknown file extension.\")\n return\n\n # Total wait time: up to 500ms\n os_error_retry_count: Final[int] = 50\n os_error_retry_wait: Final[float] = 0.01\n\n read_try_count = 0\n file_open_ok = False\n os_error_str = None\n\n while (read_try_count < os_error_retry_count) and not file_open_ok:\n try:\n with open(filepath, \"rb\"):\n file_open_ok = True\n except OSError as e:\n read_try_count += 1\n os_error_str = str(e)\n sleep(os_error_retry_wait)\n\n if read_try_count >= os_error_retry_count:\n logger.warning(f\"Not consuming file {filepath}: OS reports {os_error_str}\")\n return\n\n tag_ids = None\n try:\n if settings.CONSUMER_SUBDIRS_AS_TAGS:\n tag_ids = _tags_from_path(filepath)\n except Exception:\n logger.exception(\"Error creating tags from path\")\n\n try:\n logger.info(f\"Adding {filepath} to the task queue.\")\n async_task(\n \"documents.tasks.consume_file\",\n filepath,\n override_tag_ids=tag_ids if tag_ids else None,\n task_name=os.path.basename(filepath)[:100],\n )\n except Exception:\n # Catch all so that the consumer won't crash.\n # This is also what the test case is listening for to check for\n # errors.\n logger.exception(\"Error while consuming document\")\n\n\ndef _consume_wait_unmodified(file):\n if _is_ignored(file):\n return\n\n logger.debug(f\"Waiting for file {file} to remain unmodified\")\n mtime = -1\n size = -1\n current_try = 0\n while current_try < settings.CONSUMER_POLLING_RETRY_COUNT:\n try:\n stat_data = os.stat(file)\n new_mtime = stat_data.st_mtime\n new_size = stat_data.st_size\n except FileNotFoundError:\n logger.debug(\n f\"File {file} moved while waiting for it to remain \" f\"unmodified.\",\n )\n return\n if new_mtime == mtime and new_size == size:\n _consume(file)\n return\n mtime = new_mtime\n size = new_size\n sleep(settings.CONSUMER_POLLING_DELAY)\n current_try += 1\n\n logger.error(f\"Timeout while waiting on file {file} to remain unmodified.\")\n\n\nclass Handler(FileSystemEventHandler):\n def on_created(self, event):\n Thread(target=_consume_wait_unmodified, args=(event.src_path,)).start()\n\n def on_moved(self, event):\n Thread(target=_consume_wait_unmodified, args=(event.dest_path,)).start()\n\n\nclass Command(BaseCommand):\n \"\"\"\n On every iteration of an infinite loop, consume what we can from the\n consumption directory.\n \"\"\"\n\n # This is here primarily for the tests and is irrelevant in production.\n stop_flag = False\n\n observer = None\n\n def add_arguments(self, parser):\n parser.add_argument(\n \"directory\",\n default=settings.CONSUMPTION_DIR,\n nargs=\"?\",\n help=\"The consumption directory.\",\n )\n parser.add_argument(\"--oneshot\", action=\"store_true\", help=\"Run only once.\")\n\n def handle(self, *args, **options):\n directory = options[\"directory\"]\n recursive = settings.CONSUMER_RECURSIVE\n\n if not directory:\n raise CommandError(\"CONSUMPTION_DIR does not appear to be set.\")\n\n directory = os.path.abspath(directory)\n\n if not os.path.isdir(directory):\n raise CommandError(f\"Consumption directory {directory} does not exist\")\n\n if recursive:\n for dirpath, _, filenames in os.walk(directory):\n for filename in filenames:\n filepath = os.path.join(dirpath, filename)\n _consume(filepath)\n else:\n for entry in os.scandir(directory):\n _consume(entry.path)\n\n if options[\"oneshot\"]:\n return\n\n if settings.CONSUMER_POLLING == 0 and INotify:\n self.handle_inotify(directory, recursive)\n else:\n self.handle_polling(directory, recursive)\n\n logger.debug(\"Consumer exiting.\")\n\n def handle_polling(self, directory, recursive):\n logger.info(f\"Polling directory for changes: {directory}\")\n self.observer = PollingObserver(timeout=settings.CONSUMER_POLLING)\n self.observer.schedule(Handler(), directory, recursive=recursive)\n self.observer.start()\n try:\n while self.observer.is_alive():\n self.observer.join(1)\n if self.stop_flag:\n self.observer.stop()\n except KeyboardInterrupt:\n self.observer.stop()\n self.observer.join()\n\n def handle_inotify(self, directory, recursive):\n logger.info(f\"Using inotify to watch directory for changes: {directory}\")\n\n inotify = INotify()\n inotify_flags = flags.CLOSE_WRITE | flags.MOVED_TO\n if recursive:\n descriptor = inotify.add_watch_recursive(directory, inotify_flags)\n else:\n descriptor = inotify.add_watch(directory, inotify_flags)\n\n try:\n\n inotify_debounce: Final[float] = settings.CONSUMER_INOTIFY_DELAY\n notified_files = {}\n\n while not self.stop_flag:\n\n for event in inotify.read(timeout=1000):\n if recursive:\n path = inotify.get_path(event.wd)\n else:\n path = directory\n filepath = os.path.join(path, event.name)\n notified_files[filepath] = monotonic()\n\n # Check the files against the timeout\n still_waiting = {}\n for filepath in notified_files:\n # Time of the last inotify event for this file\n last_event_time = notified_files[filepath]\n\n # Current time - last time over the configured timeout\n waited_long_enough = (\n monotonic() - last_event_time\n ) > inotify_debounce\n\n # Also make sure the file exists still, some scanners might write a\n # temporary file first\n file_still_exists = os.path.exists(filepath) and os.path.isfile(\n filepath,\n )\n\n if waited_long_enough and file_still_exists:\n _consume(filepath)\n elif file_still_exists:\n still_waiting[filepath] = last_event_time\n\n # These files are still waiting to hit the timeout\n notified_files = still_waiting\n\n except KeyboardInterrupt:\n pass\n\n inotify.rm_watch(descriptor)\n inotify.close()\n", "path": "src/documents/management/commands/document_consumer.py"}]} | 3,449 | 215 |
gh_patches_debug_2014 | rasdani/github-patches | git_diff | pandas-dev__pandas-7007 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Matplotlib cursor position wrong after using asfreq method to change freq of DateTimeIndex from None to something
After using the `asfreq` method to change the frequency of a time-series DataFrame from `None` to something e.g. `15Min` the cursor position in matplotlib graphs of that DataFrame is no longer correct (usually shows a datetime just after the unix epoch). The following demonstrates this (NB dt in df1 is not a constant):
```
df1 = pandas.read_csv('tseries1.csv', names=['tstamp', 'Q'], parse_dates=True,
index_col='tstamp').clip_lower(0).fillna(0)
df1['T'] = pandas.read_csv('tseries2.csv', names=['tstamp', 'T'], parse_dates=True,
index_col='tstamp', squeeze=True).clip_lower(0).fillna(0)
df2 = df1.asfreq(freq='15Min', method='ffill')
# NB df1.index.freq is None
# NB df2.index.freq is <15 * Minutes>
df1.plot()
df2.plot()
plt.show()
```
I find the Matplotlib cursor position to be invaluable when looking for features in very long time-series.
Versions:
- pandas master (commit ID 764b444)
- numpy 1.8
- matplotlib 1.3.0
Matplotlib cursor position wrong after using asfreq method to change freq of DateTimeIndex from None to something
After using the `asfreq` method to change the frequency of a time-series DataFrame from `None` to something e.g. `15Min` the cursor position in matplotlib graphs of that DataFrame is no longer correct (usually shows a datetime just after the unix epoch). The following demonstrates this (NB dt in df1 is not a constant):
```
df1 = pandas.read_csv('tseries1.csv', names=['tstamp', 'Q'], parse_dates=True,
index_col='tstamp').clip_lower(0).fillna(0)
df1['T'] = pandas.read_csv('tseries2.csv', names=['tstamp', 'T'], parse_dates=True,
index_col='tstamp', squeeze=True).clip_lower(0).fillna(0)
df2 = df1.asfreq(freq='15Min', method='ffill')
# NB df1.index.freq is None
# NB df2.index.freq is <15 * Minutes>
df1.plot()
df2.plot()
plt.show()
```
I find the Matplotlib cursor position to be invaluable when looking for features in very long time-series.
Versions:
- pandas master (commit ID 764b444)
- numpy 1.8
- matplotlib 1.3.0
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `pandas/tseries/plotting.py`
Content:
```
1 """
2 Period formatters and locators adapted from scikits.timeseries by
3 Pierre GF Gerard-Marchant & Matt Knox
4 """
5
6 #!!! TODO: Use the fact that axis can have units to simplify the process
7 import datetime as pydt
8 from datetime import datetime
9
10 from matplotlib import pylab
11 import matplotlib.units as units
12
13 import numpy as np
14
15 from pandas import isnull
16 from pandas.tseries.period import Period
17 from pandas.tseries.offsets import DateOffset
18 import pandas.tseries.frequencies as frequencies
19 from pandas.tseries.index import DatetimeIndex
20 import pandas.core.common as com
21
22 from pandas.tseries.converter import (PeriodConverter, TimeSeries_DateLocator,
23 TimeSeries_DateFormatter)
24
25 #----------------------------------------------------------------------
26 # Plotting functions and monkey patches
27
28
29 def tsplot(series, plotf, **kwargs):
30 """
31 Plots a Series on the given Matplotlib axes or the current axes
32
33 Parameters
34 ----------
35 axes : Axes
36 series : Series
37
38 Notes
39 _____
40 Supports same kwargs as Axes.plot
41
42 """
43 # Used inferred freq is possible, need a test case for inferred
44 if 'ax' in kwargs:
45 ax = kwargs.pop('ax')
46 else:
47 import matplotlib.pyplot as plt
48 ax = plt.gca()
49
50 freq = _get_freq(ax, series)
51 # resample against axes freq if necessary
52 if freq is None: # pragma: no cover
53 raise ValueError('Cannot use dynamic axis without frequency info')
54 else:
55 # Convert DatetimeIndex to PeriodIndex
56 if isinstance(series.index, DatetimeIndex):
57 series = series.to_period(freq=freq)
58 freq, ax_freq, series = _maybe_resample(series, ax, freq, plotf,
59 kwargs)
60
61 # Set ax with freq info
62 _decorate_axes(ax, freq, kwargs)
63
64 # mask missing values
65 args = _maybe_mask(series)
66
67 # how to make sure ax.clear() flows through?
68 if not hasattr(ax, '_plot_data'):
69 ax._plot_data = []
70 ax._plot_data.append((series, kwargs))
71
72 # styles
73 style = kwargs.pop('style', None)
74 if style is not None:
75 args.append(style)
76
77 lines = plotf(ax, *args, **kwargs)
78 label = kwargs.get('label', None)
79
80 # set date formatter, locators and rescale limits
81 format_dateaxis(ax, ax.freq)
82 left, right = _get_xlim(ax.get_lines())
83 ax.set_xlim(left, right)
84
85 # x and y coord info
86 tz = series.index.to_datetime().tz
87 ax.format_coord = lambda t, y : "t = {} y = {:8f}".format(datetime.fromtimestamp(t, tz), y)
88
89 return lines
90
91
92 def _maybe_resample(series, ax, freq, plotf, kwargs):
93 ax_freq = _get_ax_freq(ax)
94 if ax_freq is not None and freq != ax_freq:
95 if frequencies.is_superperiod(freq, ax_freq): # upsample input
96 series = series.copy()
97 series.index = series.index.asfreq(ax_freq, how='s')
98 freq = ax_freq
99 elif _is_sup(freq, ax_freq): # one is weekly
100 how = kwargs.pop('how', 'last')
101 series = series.resample('D', how=how).dropna()
102 series = series.resample(ax_freq, how=how).dropna()
103 freq = ax_freq
104 elif frequencies.is_subperiod(freq, ax_freq) or _is_sub(freq, ax_freq):
105 _upsample_others(ax, freq, plotf, kwargs)
106 ax_freq = freq
107 else: # pragma: no cover
108 raise ValueError('Incompatible frequency conversion')
109 return freq, ax_freq, series
110
111
112 def _get_ax_freq(ax):
113 ax_freq = getattr(ax, 'freq', None)
114 if ax_freq is None:
115 if hasattr(ax, 'left_ax'):
116 ax_freq = getattr(ax.left_ax, 'freq', None)
117 if hasattr(ax, 'right_ax'):
118 ax_freq = getattr(ax.right_ax, 'freq', None)
119 return ax_freq
120
121
122 def _is_sub(f1, f2):
123 return ((f1.startswith('W') and frequencies.is_subperiod('D', f2)) or
124 (f2.startswith('W') and frequencies.is_subperiod(f1, 'D')))
125
126
127 def _is_sup(f1, f2):
128 return ((f1.startswith('W') and frequencies.is_superperiod('D', f2)) or
129 (f2.startswith('W') and frequencies.is_superperiod(f1, 'D')))
130
131
132 def _upsample_others(ax, freq, plotf, kwargs):
133 legend = ax.get_legend()
134 lines, labels = _replot_ax(ax, freq, plotf, kwargs)
135
136 other_ax = None
137 if hasattr(ax, 'left_ax'):
138 other_ax = ax.left_ax
139 if hasattr(ax, 'right_ax'):
140 other_ax = ax.right_ax
141
142 if other_ax is not None:
143 rlines, rlabels = _replot_ax(other_ax, freq, plotf, kwargs)
144 lines.extend(rlines)
145 labels.extend(rlabels)
146
147 if (legend is not None and kwargs.get('legend', True) and
148 len(lines) > 0):
149 title = legend.get_title().get_text()
150 if title == 'None':
151 title = None
152 ax.legend(lines, labels, loc='best', title=title)
153
154
155 def _replot_ax(ax, freq, plotf, kwargs):
156 data = getattr(ax, '_plot_data', None)
157 ax._plot_data = []
158 ax.clear()
159 _decorate_axes(ax, freq, kwargs)
160
161 lines = []
162 labels = []
163 if data is not None:
164 for series, kwds in data:
165 series = series.copy()
166 idx = series.index.asfreq(freq, how='S')
167 series.index = idx
168 ax._plot_data.append(series)
169 args = _maybe_mask(series)
170 lines.append(plotf(ax, *args, **kwds)[0])
171 labels.append(com.pprint_thing(series.name))
172
173 return lines, labels
174
175
176 def _decorate_axes(ax, freq, kwargs):
177 ax.freq = freq
178 xaxis = ax.get_xaxis()
179 xaxis.freq = freq
180 if not hasattr(ax, 'legendlabels'):
181 ax.legendlabels = [kwargs.get('label', None)]
182 else:
183 ax.legendlabels.append(kwargs.get('label', None))
184 ax.view_interval = None
185 ax.date_axis_info = None
186
187
188 def _maybe_mask(series):
189 mask = isnull(series)
190 if mask.any():
191 masked_array = np.ma.array(series.values)
192 masked_array = np.ma.masked_where(mask, masked_array)
193 args = [series.index, masked_array]
194 else:
195 args = [series.index, series.values]
196 return args
197
198
199 def _get_freq(ax, series):
200 # get frequency from data
201 freq = getattr(series.index, 'freq', None)
202 if freq is None:
203 freq = getattr(series.index, 'inferred_freq', None)
204
205 ax_freq = getattr(ax, 'freq', None)
206
207 # use axes freq if no data freq
208 if freq is None:
209 freq = ax_freq
210
211 # get the period frequency
212 if isinstance(freq, DateOffset):
213 freq = freq.rule_code
214 else:
215 freq = frequencies.get_base_alias(freq)
216
217 freq = frequencies.get_period_alias(freq)
218
219 return freq
220
221
222 def _get_xlim(lines):
223 left, right = np.inf, -np.inf
224 for l in lines:
225 x = l.get_xdata()
226 left = min(x[0].ordinal, left)
227 right = max(x[-1].ordinal, right)
228 return left, right
229
230 # Patch methods for subplot. Only format_dateaxis is currently used.
231 # Do we need the rest for convenience?
232
233
234 def format_dateaxis(subplot, freq):
235 """
236 Pretty-formats the date axis (x-axis).
237
238 Major and minor ticks are automatically set for the frequency of the
239 current underlying series. As the dynamic mode is activated by
240 default, changing the limits of the x axis will intelligently change
241 the positions of the ticks.
242 """
243 majlocator = TimeSeries_DateLocator(freq, dynamic_mode=True,
244 minor_locator=False,
245 plot_obj=subplot)
246 minlocator = TimeSeries_DateLocator(freq, dynamic_mode=True,
247 minor_locator=True,
248 plot_obj=subplot)
249 subplot.xaxis.set_major_locator(majlocator)
250 subplot.xaxis.set_minor_locator(minlocator)
251
252 majformatter = TimeSeries_DateFormatter(freq, dynamic_mode=True,
253 minor_locator=False,
254 plot_obj=subplot)
255 minformatter = TimeSeries_DateFormatter(freq, dynamic_mode=True,
256 minor_locator=True,
257 plot_obj=subplot)
258 subplot.xaxis.set_major_formatter(majformatter)
259 subplot.xaxis.set_minor_formatter(minformatter)
260 pylab.draw_if_interactive()
261
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/pandas/tseries/plotting.py b/pandas/tseries/plotting.py
--- a/pandas/tseries/plotting.py
+++ b/pandas/tseries/plotting.py
@@ -83,8 +83,7 @@
ax.set_xlim(left, right)
# x and y coord info
- tz = series.index.to_datetime().tz
- ax.format_coord = lambda t, y : "t = {} y = {:8f}".format(datetime.fromtimestamp(t, tz), y)
+ ax.format_coord = lambda t, y: "t = {} y = {:8f}".format(Period(ordinal=int(t), freq=ax.freq), y)
return lines
| {"golden_diff": "diff --git a/pandas/tseries/plotting.py b/pandas/tseries/plotting.py\n--- a/pandas/tseries/plotting.py\n+++ b/pandas/tseries/plotting.py\n@@ -83,8 +83,7 @@\n ax.set_xlim(left, right)\n \n # x and y coord info\n- tz = series.index.to_datetime().tz\n- ax.format_coord = lambda t, y : \"t = {} y = {:8f}\".format(datetime.fromtimestamp(t, tz), y)\n+ ax.format_coord = lambda t, y: \"t = {} y = {:8f}\".format(Period(ordinal=int(t), freq=ax.freq), y)\n \n return lines\n", "issue": "Matplotlib cursor position wrong after using asfreq method to change freq of DateTimeIndex from None to something\nAfter using the `asfreq` method to change the frequency of a time-series DataFrame from `None` to something e.g. `15Min` the cursor position in matplotlib graphs of that DataFrame is no longer correct (usually shows a datetime just after the unix epoch). The following demonstrates this (NB dt in df1 is not a constant):\n\n```\ndf1 = pandas.read_csv('tseries1.csv', names=['tstamp', 'Q'], parse_dates=True, \n index_col='tstamp').clip_lower(0).fillna(0)\ndf1['T'] = pandas.read_csv('tseries2.csv', names=['tstamp', 'T'], parse_dates=True, \n index_col='tstamp', squeeze=True).clip_lower(0).fillna(0)\n\ndf2 = df1.asfreq(freq='15Min', method='ffill')\n# NB df1.index.freq is None\n# NB df2.index.freq is <15 * Minutes>\ndf1.plot()\ndf2.plot()\nplt.show()\n```\n\nI find the Matplotlib cursor position to be invaluable when looking for features in very long time-series.\n\nVersions:\n- pandas master (commit ID 764b444)\n- numpy 1.8\n- matplotlib 1.3.0\n\nMatplotlib cursor position wrong after using asfreq method to change freq of DateTimeIndex from None to something\nAfter using the `asfreq` method to change the frequency of a time-series DataFrame from `None` to something e.g. `15Min` the cursor position in matplotlib graphs of that DataFrame is no longer correct (usually shows a datetime just after the unix epoch). The following demonstrates this (NB dt in df1 is not a constant):\n\n```\ndf1 = pandas.read_csv('tseries1.csv', names=['tstamp', 'Q'], parse_dates=True, \n index_col='tstamp').clip_lower(0).fillna(0)\ndf1['T'] = pandas.read_csv('tseries2.csv', names=['tstamp', 'T'], parse_dates=True, \n index_col='tstamp', squeeze=True).clip_lower(0).fillna(0)\n\ndf2 = df1.asfreq(freq='15Min', method='ffill')\n# NB df1.index.freq is None\n# NB df2.index.freq is <15 * Minutes>\ndf1.plot()\ndf2.plot()\nplt.show()\n```\n\nI find the Matplotlib cursor position to be invaluable when looking for features in very long time-series.\n\nVersions:\n- pandas master (commit ID 764b444)\n- numpy 1.8\n- matplotlib 1.3.0\n\n", "before_files": [{"content": "\"\"\"\nPeriod formatters and locators adapted from scikits.timeseries by\nPierre GF Gerard-Marchant & Matt Knox\n\"\"\"\n\n#!!! TODO: Use the fact that axis can have units to simplify the process\nimport datetime as pydt\nfrom datetime import datetime\n\nfrom matplotlib import pylab\nimport matplotlib.units as units\n\nimport numpy as np\n\nfrom pandas import isnull\nfrom pandas.tseries.period import Period\nfrom pandas.tseries.offsets import DateOffset\nimport pandas.tseries.frequencies as frequencies\nfrom pandas.tseries.index import DatetimeIndex\nimport pandas.core.common as com\n\nfrom pandas.tseries.converter import (PeriodConverter, TimeSeries_DateLocator,\n TimeSeries_DateFormatter)\n\n#----------------------------------------------------------------------\n# Plotting functions and monkey patches\n\n\ndef tsplot(series, plotf, **kwargs):\n \"\"\"\n Plots a Series on the given Matplotlib axes or the current axes\n\n Parameters\n ----------\n axes : Axes\n series : Series\n\n Notes\n _____\n Supports same kwargs as Axes.plot\n\n \"\"\"\n # Used inferred freq is possible, need a test case for inferred\n if 'ax' in kwargs:\n ax = kwargs.pop('ax')\n else:\n import matplotlib.pyplot as plt\n ax = plt.gca()\n\n freq = _get_freq(ax, series)\n # resample against axes freq if necessary\n if freq is None: # pragma: no cover\n raise ValueError('Cannot use dynamic axis without frequency info')\n else:\n # Convert DatetimeIndex to PeriodIndex\n if isinstance(series.index, DatetimeIndex):\n series = series.to_period(freq=freq)\n freq, ax_freq, series = _maybe_resample(series, ax, freq, plotf,\n kwargs)\n\n # Set ax with freq info\n _decorate_axes(ax, freq, kwargs)\n\n # mask missing values\n args = _maybe_mask(series)\n\n # how to make sure ax.clear() flows through?\n if not hasattr(ax, '_plot_data'):\n ax._plot_data = []\n ax._plot_data.append((series, kwargs))\n\n # styles\n style = kwargs.pop('style', None)\n if style is not None:\n args.append(style)\n\n lines = plotf(ax, *args, **kwargs)\n label = kwargs.get('label', None)\n\n # set date formatter, locators and rescale limits\n format_dateaxis(ax, ax.freq)\n left, right = _get_xlim(ax.get_lines())\n ax.set_xlim(left, right)\n\n # x and y coord info\n tz = series.index.to_datetime().tz\n ax.format_coord = lambda t, y : \"t = {} y = {:8f}\".format(datetime.fromtimestamp(t, tz), y)\n\n return lines\n\n\ndef _maybe_resample(series, ax, freq, plotf, kwargs):\n ax_freq = _get_ax_freq(ax)\n if ax_freq is not None and freq != ax_freq:\n if frequencies.is_superperiod(freq, ax_freq): # upsample input\n series = series.copy()\n series.index = series.index.asfreq(ax_freq, how='s')\n freq = ax_freq\n elif _is_sup(freq, ax_freq): # one is weekly\n how = kwargs.pop('how', 'last')\n series = series.resample('D', how=how).dropna()\n series = series.resample(ax_freq, how=how).dropna()\n freq = ax_freq\n elif frequencies.is_subperiod(freq, ax_freq) or _is_sub(freq, ax_freq):\n _upsample_others(ax, freq, plotf, kwargs)\n ax_freq = freq\n else: # pragma: no cover\n raise ValueError('Incompatible frequency conversion')\n return freq, ax_freq, series\n\n\ndef _get_ax_freq(ax):\n ax_freq = getattr(ax, 'freq', None)\n if ax_freq is None:\n if hasattr(ax, 'left_ax'):\n ax_freq = getattr(ax.left_ax, 'freq', None)\n if hasattr(ax, 'right_ax'):\n ax_freq = getattr(ax.right_ax, 'freq', None)\n return ax_freq\n\n\ndef _is_sub(f1, f2):\n return ((f1.startswith('W') and frequencies.is_subperiod('D', f2)) or\n (f2.startswith('W') and frequencies.is_subperiod(f1, 'D')))\n\n\ndef _is_sup(f1, f2):\n return ((f1.startswith('W') and frequencies.is_superperiod('D', f2)) or\n (f2.startswith('W') and frequencies.is_superperiod(f1, 'D')))\n\n\ndef _upsample_others(ax, freq, plotf, kwargs):\n legend = ax.get_legend()\n lines, labels = _replot_ax(ax, freq, plotf, kwargs)\n\n other_ax = None\n if hasattr(ax, 'left_ax'):\n other_ax = ax.left_ax\n if hasattr(ax, 'right_ax'):\n other_ax = ax.right_ax\n\n if other_ax is not None:\n rlines, rlabels = _replot_ax(other_ax, freq, plotf, kwargs)\n lines.extend(rlines)\n labels.extend(rlabels)\n\n if (legend is not None and kwargs.get('legend', True) and\n len(lines) > 0):\n title = legend.get_title().get_text()\n if title == 'None':\n title = None\n ax.legend(lines, labels, loc='best', title=title)\n\n\ndef _replot_ax(ax, freq, plotf, kwargs):\n data = getattr(ax, '_plot_data', None)\n ax._plot_data = []\n ax.clear()\n _decorate_axes(ax, freq, kwargs)\n\n lines = []\n labels = []\n if data is not None:\n for series, kwds in data:\n series = series.copy()\n idx = series.index.asfreq(freq, how='S')\n series.index = idx\n ax._plot_data.append(series)\n args = _maybe_mask(series)\n lines.append(plotf(ax, *args, **kwds)[0])\n labels.append(com.pprint_thing(series.name))\n\n return lines, labels\n\n\ndef _decorate_axes(ax, freq, kwargs):\n ax.freq = freq\n xaxis = ax.get_xaxis()\n xaxis.freq = freq\n if not hasattr(ax, 'legendlabels'):\n ax.legendlabels = [kwargs.get('label', None)]\n else:\n ax.legendlabels.append(kwargs.get('label', None))\n ax.view_interval = None\n ax.date_axis_info = None\n\n\ndef _maybe_mask(series):\n mask = isnull(series)\n if mask.any():\n masked_array = np.ma.array(series.values)\n masked_array = np.ma.masked_where(mask, masked_array)\n args = [series.index, masked_array]\n else:\n args = [series.index, series.values]\n return args\n\n\ndef _get_freq(ax, series):\n # get frequency from data\n freq = getattr(series.index, 'freq', None)\n if freq is None:\n freq = getattr(series.index, 'inferred_freq', None)\n\n ax_freq = getattr(ax, 'freq', None)\n\n # use axes freq if no data freq\n if freq is None:\n freq = ax_freq\n\n # get the period frequency\n if isinstance(freq, DateOffset):\n freq = freq.rule_code\n else:\n freq = frequencies.get_base_alias(freq)\n\n freq = frequencies.get_period_alias(freq)\n\n return freq\n\n\ndef _get_xlim(lines):\n left, right = np.inf, -np.inf\n for l in lines:\n x = l.get_xdata()\n left = min(x[0].ordinal, left)\n right = max(x[-1].ordinal, right)\n return left, right\n\n# Patch methods for subplot. Only format_dateaxis is currently used.\n# Do we need the rest for convenience?\n\n\ndef format_dateaxis(subplot, freq):\n \"\"\"\n Pretty-formats the date axis (x-axis).\n\n Major and minor ticks are automatically set for the frequency of the\n current underlying series. As the dynamic mode is activated by\n default, changing the limits of the x axis will intelligently change\n the positions of the ticks.\n \"\"\"\n majlocator = TimeSeries_DateLocator(freq, dynamic_mode=True,\n minor_locator=False,\n plot_obj=subplot)\n minlocator = TimeSeries_DateLocator(freq, dynamic_mode=True,\n minor_locator=True,\n plot_obj=subplot)\n subplot.xaxis.set_major_locator(majlocator)\n subplot.xaxis.set_minor_locator(minlocator)\n\n majformatter = TimeSeries_DateFormatter(freq, dynamic_mode=True,\n minor_locator=False,\n plot_obj=subplot)\n minformatter = TimeSeries_DateFormatter(freq, dynamic_mode=True,\n minor_locator=True,\n plot_obj=subplot)\n subplot.xaxis.set_major_formatter(majformatter)\n subplot.xaxis.set_minor_formatter(minformatter)\n pylab.draw_if_interactive()\n", "path": "pandas/tseries/plotting.py"}], "after_files": [{"content": "\"\"\"\nPeriod formatters and locators adapted from scikits.timeseries by\nPierre GF Gerard-Marchant & Matt Knox\n\"\"\"\n\n#!!! TODO: Use the fact that axis can have units to simplify the process\nimport datetime as pydt\nfrom datetime import datetime\n\nfrom matplotlib import pylab\nimport matplotlib.units as units\n\nimport numpy as np\n\nfrom pandas import isnull\nfrom pandas.tseries.period import Period\nfrom pandas.tseries.offsets import DateOffset\nimport pandas.tseries.frequencies as frequencies\nfrom pandas.tseries.index import DatetimeIndex\nimport pandas.core.common as com\n\nfrom pandas.tseries.converter import (PeriodConverter, TimeSeries_DateLocator,\n TimeSeries_DateFormatter)\n\n#----------------------------------------------------------------------\n# Plotting functions and monkey patches\n\n\ndef tsplot(series, plotf, **kwargs):\n \"\"\"\n Plots a Series on the given Matplotlib axes or the current axes\n\n Parameters\n ----------\n axes : Axes\n series : Series\n\n Notes\n _____\n Supports same kwargs as Axes.plot\n\n \"\"\"\n # Used inferred freq is possible, need a test case for inferred\n if 'ax' in kwargs:\n ax = kwargs.pop('ax')\n else:\n import matplotlib.pyplot as plt\n ax = plt.gca()\n\n freq = _get_freq(ax, series)\n # resample against axes freq if necessary\n if freq is None: # pragma: no cover\n raise ValueError('Cannot use dynamic axis without frequency info')\n else:\n # Convert DatetimeIndex to PeriodIndex\n if isinstance(series.index, DatetimeIndex):\n series = series.to_period(freq=freq)\n freq, ax_freq, series = _maybe_resample(series, ax, freq, plotf,\n kwargs)\n\n # Set ax with freq info\n _decorate_axes(ax, freq, kwargs)\n\n # mask missing values\n args = _maybe_mask(series)\n\n # how to make sure ax.clear() flows through?\n if not hasattr(ax, '_plot_data'):\n ax._plot_data = []\n ax._plot_data.append((series, kwargs))\n\n # styles\n style = kwargs.pop('style', None)\n if style is not None:\n args.append(style)\n\n lines = plotf(ax, *args, **kwargs)\n label = kwargs.get('label', None)\n\n # set date formatter, locators and rescale limits\n format_dateaxis(ax, ax.freq)\n left, right = _get_xlim(ax.get_lines())\n ax.set_xlim(left, right)\n\n # x and y coord info\n ax.format_coord = lambda t, y: \"t = {} y = {:8f}\".format(Period(ordinal=int(t), freq=ax.freq), y)\n\n return lines\n\n\ndef _maybe_resample(series, ax, freq, plotf, kwargs):\n ax_freq = _get_ax_freq(ax)\n if ax_freq is not None and freq != ax_freq:\n if frequencies.is_superperiod(freq, ax_freq): # upsample input\n series = series.copy()\n series.index = series.index.asfreq(ax_freq, how='s')\n freq = ax_freq\n elif _is_sup(freq, ax_freq): # one is weekly\n how = kwargs.pop('how', 'last')\n series = series.resample('D', how=how).dropna()\n series = series.resample(ax_freq, how=how).dropna()\n freq = ax_freq\n elif frequencies.is_subperiod(freq, ax_freq) or _is_sub(freq, ax_freq):\n _upsample_others(ax, freq, plotf, kwargs)\n ax_freq = freq\n else: # pragma: no cover\n raise ValueError('Incompatible frequency conversion')\n return freq, ax_freq, series\n\n\ndef _get_ax_freq(ax):\n ax_freq = getattr(ax, 'freq', None)\n if ax_freq is None:\n if hasattr(ax, 'left_ax'):\n ax_freq = getattr(ax.left_ax, 'freq', None)\n if hasattr(ax, 'right_ax'):\n ax_freq = getattr(ax.right_ax, 'freq', None)\n return ax_freq\n\n\ndef _is_sub(f1, f2):\n return ((f1.startswith('W') and frequencies.is_subperiod('D', f2)) or\n (f2.startswith('W') and frequencies.is_subperiod(f1, 'D')))\n\n\ndef _is_sup(f1, f2):\n return ((f1.startswith('W') and frequencies.is_superperiod('D', f2)) or\n (f2.startswith('W') and frequencies.is_superperiod(f1, 'D')))\n\n\ndef _upsample_others(ax, freq, plotf, kwargs):\n legend = ax.get_legend()\n lines, labels = _replot_ax(ax, freq, plotf, kwargs)\n\n other_ax = None\n if hasattr(ax, 'left_ax'):\n other_ax = ax.left_ax\n if hasattr(ax, 'right_ax'):\n other_ax = ax.right_ax\n\n if other_ax is not None:\n rlines, rlabels = _replot_ax(other_ax, freq, plotf, kwargs)\n lines.extend(rlines)\n labels.extend(rlabels)\n\n if (legend is not None and kwargs.get('legend', True) and\n len(lines) > 0):\n title = legend.get_title().get_text()\n if title == 'None':\n title = None\n ax.legend(lines, labels, loc='best', title=title)\n\n\ndef _replot_ax(ax, freq, plotf, kwargs):\n data = getattr(ax, '_plot_data', None)\n ax._plot_data = []\n ax.clear()\n _decorate_axes(ax, freq, kwargs)\n\n lines = []\n labels = []\n if data is not None:\n for series, kwds in data:\n series = series.copy()\n idx = series.index.asfreq(freq, how='S')\n series.index = idx\n ax._plot_data.append(series)\n args = _maybe_mask(series)\n lines.append(plotf(ax, *args, **kwds)[0])\n labels.append(com.pprint_thing(series.name))\n\n return lines, labels\n\n\ndef _decorate_axes(ax, freq, kwargs):\n ax.freq = freq\n xaxis = ax.get_xaxis()\n xaxis.freq = freq\n if not hasattr(ax, 'legendlabels'):\n ax.legendlabels = [kwargs.get('label', None)]\n else:\n ax.legendlabels.append(kwargs.get('label', None))\n ax.view_interval = None\n ax.date_axis_info = None\n\n\ndef _maybe_mask(series):\n mask = isnull(series)\n if mask.any():\n masked_array = np.ma.array(series.values)\n masked_array = np.ma.masked_where(mask, masked_array)\n args = [series.index, masked_array]\n else:\n args = [series.index, series.values]\n return args\n\n\ndef _get_freq(ax, series):\n # get frequency from data\n freq = getattr(series.index, 'freq', None)\n if freq is None:\n freq = getattr(series.index, 'inferred_freq', None)\n\n ax_freq = getattr(ax, 'freq', None)\n\n # use axes freq if no data freq\n if freq is None:\n freq = ax_freq\n\n # get the period frequency\n if isinstance(freq, DateOffset):\n freq = freq.rule_code\n else:\n freq = frequencies.get_base_alias(freq)\n\n freq = frequencies.get_period_alias(freq)\n\n return freq\n\n\ndef _get_xlim(lines):\n left, right = np.inf, -np.inf\n for l in lines:\n x = l.get_xdata()\n left = min(x[0].ordinal, left)\n right = max(x[-1].ordinal, right)\n return left, right\n\n# Patch methods for subplot. Only format_dateaxis is currently used.\n# Do we need the rest for convenience?\n\n\ndef format_dateaxis(subplot, freq):\n \"\"\"\n Pretty-formats the date axis (x-axis).\n\n Major and minor ticks are automatically set for the frequency of the\n current underlying series. As the dynamic mode is activated by\n default, changing the limits of the x axis will intelligently change\n the positions of the ticks.\n \"\"\"\n majlocator = TimeSeries_DateLocator(freq, dynamic_mode=True,\n minor_locator=False,\n plot_obj=subplot)\n minlocator = TimeSeries_DateLocator(freq, dynamic_mode=True,\n minor_locator=True,\n plot_obj=subplot)\n subplot.xaxis.set_major_locator(majlocator)\n subplot.xaxis.set_minor_locator(minlocator)\n\n majformatter = TimeSeries_DateFormatter(freq, dynamic_mode=True,\n minor_locator=False,\n plot_obj=subplot)\n minformatter = TimeSeries_DateFormatter(freq, dynamic_mode=True,\n minor_locator=True,\n plot_obj=subplot)\n subplot.xaxis.set_major_formatter(majformatter)\n subplot.xaxis.set_minor_formatter(minformatter)\n pylab.draw_if_interactive()\n", "path": "pandas/tseries/plotting.py"}]} | 3,455 | 157 |
gh_patches_debug_10861 | rasdani/github-patches | git_diff | horovod__horovod-2039 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Running horovod.spark.run with env=os.environ fails
Example:
horovod.spark.run(fn, num_proc=2, env=os.environ)
That `env` is an object, not a dictionary. It cannot be pickled:
```
Traceback (most recent call last):
File "horovod/run/common/util/tiny_shell_exec.py", line 32, in execute
exit_code = safe_shell_exec.execute(command, env=env, stdout=output, stderr=output)
File "horovod/run/common/util/safe_shell_exec.py", line 183, in execute
middleman.start()
File "multiprocessing/process.py", line 105, in start
self._popen = self._Popen(self)
File "multiprocessing/context.py", line 284, in _Popen
return Popen(process_obj)
File "multiprocessing/popen_spawn_posix.py", line 32, in __init__
super().__init__(process_obj)
File "multiprocessing/popen_fork.py", line 19, in __init__
self._launch(process_obj)
File "multiprocessing/popen_spawn_posix.py", line 47, in _launch
reduction.dump(process_obj, fp)
File "multiprocessing/reduction.py", line 60, in dump
ForkingPickler(file, protocol).dump(obj)
AttributeError: Can't pickle local object '_createenviron.<locals>.encode'
```
It works with
horovod.spark.run(fn, num_proc=2, env=os.environ.copy())
The `run` function needs to copy `env` itself first.
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `horovod/run/mpi_run.py`
Content:
```
1 # Copyright 2019 Uber Technologies, Inc. All Rights Reserved.
2 #
3 # Licensed under the Apache License, Version 2.0 (the "License");
4 # you may not use this file except in compliance with the License.
5 # You may obtain a copy of the License at
6 #
7 # http://www.apache.org/licenses/LICENSE-2.0
8 #
9 # Unless required by applicable law or agreed to in writing, software
10 # distributed under the License is distributed on an "AS IS" BASIS,
11 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12 # See the License for the specific language governing permissions and
13 # limitations under the License.
14 # ==============================================================================
15
16 import copy
17 import os
18 import sys
19
20 from shlex import quote
21
22 from horovod.run.common.util import env as env_util, safe_shell_exec, tiny_shell_exec
23
24 # MPI implementations
25 _OMPI_IMPL = 'OpenMPI'
26 _SMPI_IMPL = 'SpectrumMPI'
27 _MPICH_IMPL = 'MPICH'
28 _UNKNOWN_IMPL = 'Unknown'
29 _MISSING_IMPL = 'Missing'
30
31 # Open MPI Flags
32 _OMPI_FLAGS = ['-mca pml ob1', '-mca btl ^openib']
33 # Spectrum MPI Flags
34 _SMPI_FLAGS = []
35 _SMPI_FLAGS_TCP = ['-tcp']
36 # MPICH Flags
37 _MPICH_FLAGS = []
38
39 # Threshold for large cluster MPI issues:
40 _LARGE_CLUSTER_THRESHOLD = 64
41 # No process binding args
42 _NO_BINDING_ARGS = ['-bind-to none', '-map-by slot']
43 # Process socket binding args
44 _SOCKET_BINDING_ARGS = ['-bind-to socket', '-map-by socket', '-rank-by core']
45
46 # MPI not found error message
47 _MPI_NOT_FOUND_ERROR_MSG= ('horovod does not find an installed MPI.\n\n'
48 'Choose one of:\n'
49 '1. Install Open MPI 4.0.0+ or IBM Spectrum MPI or MPICH and re-install Horovod '
50 '(use --no-cache-dir pip option).\n'
51 '2. Run distributed '
52 'training script using the standard way provided by your'
53 ' MPI distribution (usually mpirun, srun, or jsrun).\n'
54 '3. Use built-in gloo option (horovodrun --gloo ...).')
55
56
57 def mpi_available(env=None):
58 return _get_mpi_implementation(env) not in {_UNKNOWN_IMPL, _MISSING_IMPL}
59
60
61 def is_open_mpi(env=None):
62 return _get_mpi_implementation(env) == _OMPI_IMPL
63
64
65 def is_spectrum_mpi(env=None):
66 return _get_mpi_implementation(env) == _SMPI_IMPL
67
68
69 def is_mpich(env=None):
70 return _get_mpi_implementation(env) == _MPICH_IMPL
71
72
73 def _get_mpi_implementation(env=None):
74 """
75 Detects the available MPI implementation by invoking `mpirun --version`.
76 This command is executed by the given execute function, which takes the
77 command as the only argument and returns (output, exit code). Output
78 represents the stdout and stderr as a string.
79
80 Returns one of:
81 - _OMPI_IMPL, _SMPI_IMPL or _MPICH_IMPL for known implementations
82 - _UNKNOWN_IMPL for any unknown implementation
83 - _MISSING_IMPL if `mpirun --version` could not be executed.
84
85 :param env: environment variable to use to run mpirun
86 :return: string representing identified implementation
87 """
88 command = 'mpirun --version'
89 res = tiny_shell_exec.execute(command, env)
90 if res is None:
91 return _MISSING_IMPL
92 (output, exit_code) = res
93
94 if exit_code == 0:
95 if 'Open MPI' in output or 'OpenRTE' in output:
96 return _OMPI_IMPL
97 elif 'IBM Spectrum MPI' in output:
98 return _SMPI_IMPL
99 elif 'MPICH' in output:
100 return _MPICH_IMPL
101
102 print('Unknown MPI implementation given in output of mpirun --version:', file=sys.stderr)
103 print(output, file=sys.stderr)
104 return _UNKNOWN_IMPL
105 else:
106 print('Was unable to run {command}:'.format(command=command), file=sys.stderr)
107 print(output, file=sys.stderr)
108 return _MISSING_IMPL
109
110
111 def _get_mpi_implementation_flags(tcp_flag, env=None):
112 if is_open_mpi(env):
113 return list(_OMPI_FLAGS), list(_NO_BINDING_ARGS)
114 elif is_spectrum_mpi(env):
115 return list(_SMPI_FLAGS) if not tcp_flag else list(_SMPI_FLAGS_TCP), list(_SOCKET_BINDING_ARGS)
116 elif is_mpich(env):
117 return list(_MPICH_FLAGS), list(_NO_BINDING_ARGS)
118 else:
119 return None, None
120
121
122 def mpi_run(settings, nics, env, command, stdout=None, stderr=None):
123 """
124 Runs mpi_run.
125
126 Args:
127 settings: Settings for running MPI.
128 Note: settings.num_proc and settings.hosts must not be None.
129 nics: Interfaces to include by MPI.
130 env: Environment dictionary to use for running command.
131 command: Command and arguments to run as a list of string.
132 stdout: Stdout of the mpi process.
133 Only used when settings.run_func_mode is True.
134 stderr: Stderr of the mpi process.
135 Only used when settings.run_func_mode is True.
136 """
137 mpi_impl_flags, impl_binding_args = _get_mpi_implementation_flags(settings.tcp_flag, env=env)
138 if mpi_impl_flags is None:
139 raise Exception(_MPI_NOT_FOUND_ERROR_MSG)
140
141 ssh_port_arg = '-mca plm_rsh_args \"-p {ssh_port}\"'.format(
142 ssh_port=settings.ssh_port) if settings.ssh_port else ''
143
144 # if user does not specify any hosts, mpirun by default uses local host.
145 # There is no need to specify localhost.
146 hosts_arg = '-H {hosts}'.format(hosts=settings.hosts)
147
148 tcp_intf_arg = '-mca btl_tcp_if_include {nics}'.format(
149 nics=','.join(nics)) if nics else ''
150 nccl_socket_intf_arg = '-x NCCL_SOCKET_IFNAME={nics}'.format(
151 nics=','.join(nics)) if nics else ''
152
153 # On large cluster runs (e.g. Summit), we need extra settings to work around OpenMPI issues
154 if settings.num_hosts and settings.num_hosts >= _LARGE_CLUSTER_THRESHOLD:
155 mpi_impl_flags.append('-mca plm_rsh_no_tree_spawn true')
156 mpi_impl_flags.append('-mca plm_rsh_num_concurrent {}'.format(settings.num_hosts))
157
158 binding_args = settings.binding_args if settings.binding_args else ' '.join(impl_binding_args)
159
160 # Pass all the env variables to the mpirun command.
161 mpirun_command = (
162 'mpirun --allow-run-as-root --tag-output '
163 '-np {num_proc} {hosts_arg} '
164 '{binding_args} '
165 '{mpi_args} '
166 '{ssh_port_arg} '
167 '{tcp_intf_arg} '
168 '{nccl_socket_intf_arg} '
169 '{output_filename_arg} '
170 '{env} {extra_mpi_args} {command}' # expect a lot of environment variables
171 .format(num_proc=settings.num_proc,
172 hosts_arg=hosts_arg,
173 binding_args=binding_args,
174 mpi_args=' '.join(mpi_impl_flags),
175 tcp_intf_arg=tcp_intf_arg,
176 nccl_socket_intf_arg=nccl_socket_intf_arg,
177 ssh_port_arg=ssh_port_arg,
178 output_filename_arg='--output-filename ' + settings.output_filename
179 if settings.output_filename else '',
180 env=' '.join('-x %s' % key for key in sorted(env.keys())
181 if env_util.is_exportable(key)),
182
183 extra_mpi_args=settings.extra_mpi_args if settings.extra_mpi_args else '',
184 command=' '.join(quote(par) for par in command))
185 )
186
187 if settings.verbose >= 2:
188 print(mpirun_command)
189
190 # we need the driver's PATH in env to run mpirun,
191 # env for mpirun is different to env encoded in mpirun_command
192 if 'PATH' not in env and 'PATH' in os.environ:
193 env = copy.copy(env) # copy env so we do not leak env modifications
194 env['PATH'] = os.environ['PATH']
195
196 # Execute the mpirun command.
197 if settings.run_func_mode:
198 exit_code = safe_shell_exec.execute(mpirun_command, env=env, stdout=stdout, stderr=stderr)
199 if exit_code != 0:
200 raise RuntimeError("mpirun failed with exit code {exit_code}".format(exit_code=exit_code))
201 else:
202 os.execve('/bin/sh', ['/bin/sh', '-c', mpirun_command], env)
203
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/horovod/run/mpi_run.py b/horovod/run/mpi_run.py
--- a/horovod/run/mpi_run.py
+++ b/horovod/run/mpi_run.py
@@ -134,6 +134,10 @@
stderr: Stderr of the mpi process.
Only used when settings.run_func_mode is True.
"""
+ if env is not None and not isinstance(env, dict):
+ raise Exception('env argument must be a dict, not {type}: {env}'
+ .format(type=type(env), env=env))
+
mpi_impl_flags, impl_binding_args = _get_mpi_implementation_flags(settings.tcp_flag, env=env)
if mpi_impl_flags is None:
raise Exception(_MPI_NOT_FOUND_ERROR_MSG)
| {"golden_diff": "diff --git a/horovod/run/mpi_run.py b/horovod/run/mpi_run.py\n--- a/horovod/run/mpi_run.py\n+++ b/horovod/run/mpi_run.py\n@@ -134,6 +134,10 @@\n stderr: Stderr of the mpi process.\n Only used when settings.run_func_mode is True.\n \"\"\"\n+ if env is not None and not isinstance(env, dict):\n+ raise Exception('env argument must be a dict, not {type}: {env}'\n+ .format(type=type(env), env=env))\n+\n mpi_impl_flags, impl_binding_args = _get_mpi_implementation_flags(settings.tcp_flag, env=env)\n if mpi_impl_flags is None:\n raise Exception(_MPI_NOT_FOUND_ERROR_MSG)\n", "issue": "Running horovod.spark.run with env=os.environ fails\nExample:\r\n\r\n horovod.spark.run(fn, num_proc=2, env=os.environ)\r\n\r\nThat `env` is an object, not a dictionary. It cannot be pickled:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"horovod/run/common/util/tiny_shell_exec.py\", line 32, in execute\r\n exit_code = safe_shell_exec.execute(command, env=env, stdout=output, stderr=output)\r\n File \"horovod/run/common/util/safe_shell_exec.py\", line 183, in execute\r\n middleman.start()\r\n File \"multiprocessing/process.py\", line 105, in start\r\n self._popen = self._Popen(self)\r\n File \"multiprocessing/context.py\", line 284, in _Popen\r\n return Popen(process_obj)\r\n File \"multiprocessing/popen_spawn_posix.py\", line 32, in __init__\r\n super().__init__(process_obj)\r\n File \"multiprocessing/popen_fork.py\", line 19, in __init__\r\n self._launch(process_obj)\r\n File \"multiprocessing/popen_spawn_posix.py\", line 47, in _launch\r\n reduction.dump(process_obj, fp)\r\n File \"multiprocessing/reduction.py\", line 60, in dump\r\n ForkingPickler(file, protocol).dump(obj)\r\nAttributeError: Can't pickle local object '_createenviron.<locals>.encode'\r\n```\r\n\r\nIt works with\r\n\r\n horovod.spark.run(fn, num_proc=2, env=os.environ.copy())\r\n\r\nThe `run` function needs to copy `env` itself first.\n", "before_files": [{"content": "# Copyright 2019 Uber Technologies, Inc. All Rights Reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n# ==============================================================================\n\nimport copy\nimport os\nimport sys\n\nfrom shlex import quote\n\nfrom horovod.run.common.util import env as env_util, safe_shell_exec, tiny_shell_exec\n\n# MPI implementations\n_OMPI_IMPL = 'OpenMPI'\n_SMPI_IMPL = 'SpectrumMPI'\n_MPICH_IMPL = 'MPICH'\n_UNKNOWN_IMPL = 'Unknown'\n_MISSING_IMPL = 'Missing'\n\n# Open MPI Flags\n_OMPI_FLAGS = ['-mca pml ob1', '-mca btl ^openib']\n# Spectrum MPI Flags\n_SMPI_FLAGS = []\n_SMPI_FLAGS_TCP = ['-tcp']\n# MPICH Flags\n_MPICH_FLAGS = []\n\n# Threshold for large cluster MPI issues:\n_LARGE_CLUSTER_THRESHOLD = 64\n# No process binding args\n_NO_BINDING_ARGS = ['-bind-to none', '-map-by slot']\n# Process socket binding args\n_SOCKET_BINDING_ARGS = ['-bind-to socket', '-map-by socket', '-rank-by core']\n\n# MPI not found error message\n_MPI_NOT_FOUND_ERROR_MSG= ('horovod does not find an installed MPI.\\n\\n'\n 'Choose one of:\\n'\n '1. Install Open MPI 4.0.0+ or IBM Spectrum MPI or MPICH and re-install Horovod '\n '(use --no-cache-dir pip option).\\n'\n '2. Run distributed '\n 'training script using the standard way provided by your'\n ' MPI distribution (usually mpirun, srun, or jsrun).\\n'\n '3. Use built-in gloo option (horovodrun --gloo ...).')\n\n\ndef mpi_available(env=None):\n return _get_mpi_implementation(env) not in {_UNKNOWN_IMPL, _MISSING_IMPL}\n\n\ndef is_open_mpi(env=None):\n return _get_mpi_implementation(env) == _OMPI_IMPL\n\n\ndef is_spectrum_mpi(env=None):\n return _get_mpi_implementation(env) == _SMPI_IMPL\n\n\ndef is_mpich(env=None):\n return _get_mpi_implementation(env) == _MPICH_IMPL\n\n\ndef _get_mpi_implementation(env=None):\n \"\"\"\n Detects the available MPI implementation by invoking `mpirun --version`.\n This command is executed by the given execute function, which takes the\n command as the only argument and returns (output, exit code). Output\n represents the stdout and stderr as a string.\n\n Returns one of:\n - _OMPI_IMPL, _SMPI_IMPL or _MPICH_IMPL for known implementations\n - _UNKNOWN_IMPL for any unknown implementation\n - _MISSING_IMPL if `mpirun --version` could not be executed.\n\n :param env: environment variable to use to run mpirun\n :return: string representing identified implementation\n \"\"\"\n command = 'mpirun --version'\n res = tiny_shell_exec.execute(command, env)\n if res is None:\n return _MISSING_IMPL\n (output, exit_code) = res\n\n if exit_code == 0:\n if 'Open MPI' in output or 'OpenRTE' in output:\n return _OMPI_IMPL\n elif 'IBM Spectrum MPI' in output:\n return _SMPI_IMPL\n elif 'MPICH' in output:\n return _MPICH_IMPL\n\n print('Unknown MPI implementation given in output of mpirun --version:', file=sys.stderr)\n print(output, file=sys.stderr)\n return _UNKNOWN_IMPL\n else:\n print('Was unable to run {command}:'.format(command=command), file=sys.stderr)\n print(output, file=sys.stderr)\n return _MISSING_IMPL\n\n\ndef _get_mpi_implementation_flags(tcp_flag, env=None):\n if is_open_mpi(env):\n return list(_OMPI_FLAGS), list(_NO_BINDING_ARGS)\n elif is_spectrum_mpi(env):\n return list(_SMPI_FLAGS) if not tcp_flag else list(_SMPI_FLAGS_TCP), list(_SOCKET_BINDING_ARGS)\n elif is_mpich(env):\n return list(_MPICH_FLAGS), list(_NO_BINDING_ARGS)\n else:\n return None, None\n\n\ndef mpi_run(settings, nics, env, command, stdout=None, stderr=None):\n \"\"\"\n Runs mpi_run.\n\n Args:\n settings: Settings for running MPI.\n Note: settings.num_proc and settings.hosts must not be None.\n nics: Interfaces to include by MPI.\n env: Environment dictionary to use for running command.\n command: Command and arguments to run as a list of string.\n stdout: Stdout of the mpi process.\n Only used when settings.run_func_mode is True.\n stderr: Stderr of the mpi process.\n Only used when settings.run_func_mode is True.\n \"\"\"\n mpi_impl_flags, impl_binding_args = _get_mpi_implementation_flags(settings.tcp_flag, env=env)\n if mpi_impl_flags is None:\n raise Exception(_MPI_NOT_FOUND_ERROR_MSG)\n\n ssh_port_arg = '-mca plm_rsh_args \\\"-p {ssh_port}\\\"'.format(\n ssh_port=settings.ssh_port) if settings.ssh_port else ''\n\n # if user does not specify any hosts, mpirun by default uses local host.\n # There is no need to specify localhost.\n hosts_arg = '-H {hosts}'.format(hosts=settings.hosts)\n\n tcp_intf_arg = '-mca btl_tcp_if_include {nics}'.format(\n nics=','.join(nics)) if nics else ''\n nccl_socket_intf_arg = '-x NCCL_SOCKET_IFNAME={nics}'.format(\n nics=','.join(nics)) if nics else ''\n\n # On large cluster runs (e.g. Summit), we need extra settings to work around OpenMPI issues\n if settings.num_hosts and settings.num_hosts >= _LARGE_CLUSTER_THRESHOLD:\n mpi_impl_flags.append('-mca plm_rsh_no_tree_spawn true')\n mpi_impl_flags.append('-mca plm_rsh_num_concurrent {}'.format(settings.num_hosts))\n\n binding_args = settings.binding_args if settings.binding_args else ' '.join(impl_binding_args)\n\n # Pass all the env variables to the mpirun command.\n mpirun_command = (\n 'mpirun --allow-run-as-root --tag-output '\n '-np {num_proc} {hosts_arg} '\n '{binding_args} '\n '{mpi_args} '\n '{ssh_port_arg} '\n '{tcp_intf_arg} '\n '{nccl_socket_intf_arg} '\n '{output_filename_arg} '\n '{env} {extra_mpi_args} {command}' # expect a lot of environment variables\n .format(num_proc=settings.num_proc,\n hosts_arg=hosts_arg,\n binding_args=binding_args,\n mpi_args=' '.join(mpi_impl_flags),\n tcp_intf_arg=tcp_intf_arg,\n nccl_socket_intf_arg=nccl_socket_intf_arg,\n ssh_port_arg=ssh_port_arg,\n output_filename_arg='--output-filename ' + settings.output_filename\n if settings.output_filename else '',\n env=' '.join('-x %s' % key for key in sorted(env.keys())\n if env_util.is_exportable(key)),\n\n extra_mpi_args=settings.extra_mpi_args if settings.extra_mpi_args else '',\n command=' '.join(quote(par) for par in command))\n )\n\n if settings.verbose >= 2:\n print(mpirun_command)\n\n # we need the driver's PATH in env to run mpirun,\n # env for mpirun is different to env encoded in mpirun_command\n if 'PATH' not in env and 'PATH' in os.environ:\n env = copy.copy(env) # copy env so we do not leak env modifications\n env['PATH'] = os.environ['PATH']\n\n # Execute the mpirun command.\n if settings.run_func_mode:\n exit_code = safe_shell_exec.execute(mpirun_command, env=env, stdout=stdout, stderr=stderr)\n if exit_code != 0:\n raise RuntimeError(\"mpirun failed with exit code {exit_code}\".format(exit_code=exit_code))\n else:\n os.execve('/bin/sh', ['/bin/sh', '-c', mpirun_command], env)\n", "path": "horovod/run/mpi_run.py"}], "after_files": [{"content": "# Copyright 2019 Uber Technologies, Inc. All Rights Reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n# ==============================================================================\n\nimport copy\nimport os\nimport sys\n\nfrom shlex import quote\n\nfrom horovod.run.common.util import env as env_util, safe_shell_exec, tiny_shell_exec\n\n# MPI implementations\n_OMPI_IMPL = 'OpenMPI'\n_SMPI_IMPL = 'SpectrumMPI'\n_MPICH_IMPL = 'MPICH'\n_UNKNOWN_IMPL = 'Unknown'\n_MISSING_IMPL = 'Missing'\n\n# Open MPI Flags\n_OMPI_FLAGS = ['-mca pml ob1', '-mca btl ^openib']\n# Spectrum MPI Flags\n_SMPI_FLAGS = []\n_SMPI_FLAGS_TCP = ['-tcp']\n# MPICH Flags\n_MPICH_FLAGS = []\n\n# Threshold for large cluster MPI issues:\n_LARGE_CLUSTER_THRESHOLD = 64\n# No process binding args\n_NO_BINDING_ARGS = ['-bind-to none', '-map-by slot']\n# Process socket binding args\n_SOCKET_BINDING_ARGS = ['-bind-to socket', '-map-by socket', '-rank-by core']\n\n# MPI not found error message\n_MPI_NOT_FOUND_ERROR_MSG= ('horovod does not find an installed MPI.\\n\\n'\n 'Choose one of:\\n'\n '1. Install Open MPI 4.0.0+ or IBM Spectrum MPI or MPICH and re-install Horovod '\n '(use --no-cache-dir pip option).\\n'\n '2. Run distributed '\n 'training script using the standard way provided by your'\n ' MPI distribution (usually mpirun, srun, or jsrun).\\n'\n '3. Use built-in gloo option (horovodrun --gloo ...).')\n\n\ndef mpi_available(env=None):\n return _get_mpi_implementation(env) not in {_UNKNOWN_IMPL, _MISSING_IMPL}\n\n\ndef is_open_mpi(env=None):\n return _get_mpi_implementation(env) == _OMPI_IMPL\n\n\ndef is_spectrum_mpi(env=None):\n return _get_mpi_implementation(env) == _SMPI_IMPL\n\n\ndef is_mpich(env=None):\n return _get_mpi_implementation(env) == _MPICH_IMPL\n\n\ndef _get_mpi_implementation(env=None):\n \"\"\"\n Detects the available MPI implementation by invoking `mpirun --version`.\n This command is executed by the given execute function, which takes the\n command as the only argument and returns (output, exit code). Output\n represents the stdout and stderr as a string.\n\n Returns one of:\n - _OMPI_IMPL, _SMPI_IMPL or _MPICH_IMPL for known implementations\n - _UNKNOWN_IMPL for any unknown implementation\n - _MISSING_IMPL if `mpirun --version` could not be executed.\n\n :param env: environment variable to use to run mpirun\n :return: string representing identified implementation\n \"\"\"\n command = 'mpirun --version'\n res = tiny_shell_exec.execute(command, env)\n if res is None:\n return _MISSING_IMPL\n (output, exit_code) = res\n\n if exit_code == 0:\n if 'Open MPI' in output or 'OpenRTE' in output:\n return _OMPI_IMPL\n elif 'IBM Spectrum MPI' in output:\n return _SMPI_IMPL\n elif 'MPICH' in output:\n return _MPICH_IMPL\n\n print('Unknown MPI implementation given in output of mpirun --version:', file=sys.stderr)\n print(output, file=sys.stderr)\n return _UNKNOWN_IMPL\n else:\n print('Was unable to run {command}:'.format(command=command), file=sys.stderr)\n print(output, file=sys.stderr)\n return _MISSING_IMPL\n\n\ndef _get_mpi_implementation_flags(tcp_flag, env=None):\n if is_open_mpi(env):\n return list(_OMPI_FLAGS), list(_NO_BINDING_ARGS)\n elif is_spectrum_mpi(env):\n return list(_SMPI_FLAGS) if not tcp_flag else list(_SMPI_FLAGS_TCP), list(_SOCKET_BINDING_ARGS)\n elif is_mpich(env):\n return list(_MPICH_FLAGS), list(_NO_BINDING_ARGS)\n else:\n return None, None\n\n\ndef mpi_run(settings, nics, env, command, stdout=None, stderr=None):\n \"\"\"\n Runs mpi_run.\n\n Args:\n settings: Settings for running MPI.\n Note: settings.num_proc and settings.hosts must not be None.\n nics: Interfaces to include by MPI.\n env: Environment dictionary to use for running command.\n command: Command and arguments to run as a list of string.\n stdout: Stdout of the mpi process.\n Only used when settings.run_func_mode is True.\n stderr: Stderr of the mpi process.\n Only used when settings.run_func_mode is True.\n \"\"\"\n if env is not None and not isinstance(env, dict):\n raise Exception('env argument must be a dict, not {type}: {env}'\n .format(type=type(env), env=env))\n\n mpi_impl_flags, impl_binding_args = _get_mpi_implementation_flags(settings.tcp_flag, env=env)\n if mpi_impl_flags is None:\n raise Exception(_MPI_NOT_FOUND_ERROR_MSG)\n\n ssh_port_arg = '-mca plm_rsh_args \\\"-p {ssh_port}\\\"'.format(\n ssh_port=settings.ssh_port) if settings.ssh_port else ''\n\n # if user does not specify any hosts, mpirun by default uses local host.\n # There is no need to specify localhost.\n hosts_arg = '-H {hosts}'.format(hosts=settings.hosts)\n\n tcp_intf_arg = '-mca btl_tcp_if_include {nics}'.format(\n nics=','.join(nics)) if nics else ''\n nccl_socket_intf_arg = '-x NCCL_SOCKET_IFNAME={nics}'.format(\n nics=','.join(nics)) if nics else ''\n\n # On large cluster runs (e.g. Summit), we need extra settings to work around OpenMPI issues\n if settings.num_hosts and settings.num_hosts >= _LARGE_CLUSTER_THRESHOLD:\n mpi_impl_flags.append('-mca plm_rsh_no_tree_spawn true')\n mpi_impl_flags.append('-mca plm_rsh_num_concurrent {}'.format(settings.num_hosts))\n\n binding_args = settings.binding_args if settings.binding_args else ' '.join(impl_binding_args)\n\n # Pass all the env variables to the mpirun command.\n mpirun_command = (\n 'mpirun --allow-run-as-root --tag-output '\n '-np {num_proc} {hosts_arg} '\n '{binding_args} '\n '{mpi_args} '\n '{ssh_port_arg} '\n '{tcp_intf_arg} '\n '{nccl_socket_intf_arg} '\n '{output_filename_arg} '\n '{env} {extra_mpi_args} {command}' # expect a lot of environment variables\n .format(num_proc=settings.num_proc,\n hosts_arg=hosts_arg,\n binding_args=binding_args,\n mpi_args=' '.join(mpi_impl_flags),\n tcp_intf_arg=tcp_intf_arg,\n nccl_socket_intf_arg=nccl_socket_intf_arg,\n ssh_port_arg=ssh_port_arg,\n output_filename_arg='--output-filename ' + settings.output_filename\n if settings.output_filename else '',\n env=' '.join('-x %s' % key for key in sorted(env.keys())\n if env_util.is_exportable(key)),\n\n extra_mpi_args=settings.extra_mpi_args if settings.extra_mpi_args else '',\n command=' '.join(quote(par) for par in command))\n )\n\n if settings.verbose >= 2:\n print(mpirun_command)\n\n # we need the driver's PATH in env to run mpirun,\n # env for mpirun is different to env encoded in mpirun_command\n if 'PATH' not in env and 'PATH' in os.environ:\n env = copy.copy(env) # copy env so we do not leak env modifications\n env['PATH'] = os.environ['PATH']\n\n # Execute the mpirun command.\n if settings.run_func_mode:\n exit_code = safe_shell_exec.execute(mpirun_command, env=env, stdout=stdout, stderr=stderr)\n if exit_code != 0:\n raise RuntimeError(\"mpirun failed with exit code {exit_code}\".format(exit_code=exit_code))\n else:\n os.execve('/bin/sh', ['/bin/sh', '-c', mpirun_command], env)\n", "path": "horovod/run/mpi_run.py"}]} | 3,012 | 174 |
gh_patches_debug_23834 | rasdani/github-patches | git_diff | NVIDIA__NVFlare-380 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Yaml loader should be replaced by safe_loader or other more secure loader
To load yaml files from unknown source, we should avoid using yaml's loader. A better way is to use either safe_loader or other mechanism.
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `nvflare/lighter/provision.py`
Content:
```
1 # Copyright (c) 2021-2022, NVIDIA CORPORATION. All rights reserved.
2 #
3 # Licensed under the Apache License, Version 2.0 (the "License");
4 # you may not use this file except in compliance with the License.
5 # You may obtain a copy of the License at
6 #
7 # http://www.apache.org/licenses/LICENSE-2.0
8 #
9 # Unless required by applicable law or agreed to in writing, software
10 # distributed under the License is distributed on an "AS IS" BASIS,
11 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12 # See the License for the specific language governing permissions and
13 # limitations under the License.
14
15 from __future__ import absolute_import
16
17 import argparse
18 import os
19 import pathlib
20 import shutil
21 import sys
22 import webbrowser
23
24 import yaml
25
26 from nvflare.fuel.utils.class_utils import instantiate_class
27 from nvflare.lighter.spec import Participant, Project, Provisioner
28
29
30 def main():
31 parser = argparse.ArgumentParser()
32 parser.add_argument("-p", "--project_file", type=str, default="project.yml", help="file to describe FL project")
33 parser.add_argument("-w", "--workspace", type=str, default="workspace", help="directory used by provision")
34 parser.add_argument("-c", "--custom_folder", type=str, default=".", help="additional folder to load python codes")
35 parser.add_argument(
36 "-u",
37 "--ui_tool",
38 action="store_true",
39 help="Run provisioning UI tool to generate project.yml file",
40 )
41
42 args = parser.parse_args()
43
44 file_path = pathlib.Path(__file__).parent.absolute()
45 current_path = os.getcwd()
46 custom_folder_path = os.path.join(current_path, args.custom_folder)
47 sys.path.append(custom_folder_path)
48 print("Path list (sys.path) for python codes loading: {}".format(sys.path))
49
50 # main project file
51 project_file = args.project_file
52 current_project_yml = os.path.join(current_path, "project.yml")
53 if len(sys.argv) == 1 and not os.path.exists(current_project_yml):
54 answer = input(
55 f"No project.yml found in current folder. Is it OK to generate one at {current_project_yml} for you? (y/N) "
56 )
57 if answer.strip().upper() == "Y":
58 shutil.copyfile(os.path.join(file_path, "project.yml"), current_project_yml)
59 print(f"{current_project_yml} was created. Please edit it to fit your FL configuration.")
60 exit(0)
61
62 if args.ui_tool:
63 ui_helper_path = os.path.join(file_path, "provision_helper.html")
64 ui_helper_url = f"file://{ui_helper_path}"
65 webbrowser.open_new_tab(ui_helper_url)
66 print(
67 "\n******\n"
68 "Now launching provisioning UI tool.\n"
69 "After generating project.yml in the browser and saving it to your local folder,\n"
70 "please re-run provision with -p option, pointing to the generated project.yml, to generate all packages.\n******\n"
71 )
72 exit(0)
73
74 workspace = args.workspace
75 workspace_full_path = os.path.join(current_path, workspace)
76
77 project_full_path = os.path.join(current_path, project_file)
78 print(f"Project yaml file: {project_full_path}.")
79
80 project_dict = yaml.load(open(project_full_path, "r"), Loader=yaml.Loader)
81 api_version = project_dict.get("api_version")
82 if api_version not in [3]:
83 raise ValueError(f"API version expected 3 but found {api_version}")
84
85 project_name = project_dict.get("name")
86 project_description = project_dict.get("description", "")
87 participants = list()
88 for p in project_dict.get("participants"):
89 participants.append(Participant(**p))
90 project = Project(name=project_name, description=project_description, participants=participants)
91
92 builders = list()
93 for b in project_dict.get("builders"):
94 path = b.get("path")
95 args = b.get("args")
96 builders.append(instantiate_class(path, args))
97
98 provisioner = Provisioner(workspace_full_path, builders)
99
100 provisioner.provision(project)
101
102
103 if __name__ == "__main__":
104 main()
105
```
Path: `nvflare/lighter/utils.py`
Content:
```
1 # Copyright (c) 2021-2022, NVIDIA CORPORATION. All rights reserved.
2 #
3 # Licensed under the Apache License, Version 2.0 (the "License");
4 # you may not use this file except in compliance with the License.
5 # You may obtain a copy of the License at
6 #
7 # http://www.apache.org/licenses/LICENSE-2.0
8 #
9 # Unless required by applicable law or agreed to in writing, software
10 # distributed under the License is distributed on an "AS IS" BASIS,
11 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12 # See the License for the specific language governing permissions and
13 # limitations under the License.
14
15 import os
16 import random
17 from base64 import b64encode
18
19 import yaml
20 from cryptography.hazmat.primitives import hashes
21 from cryptography.hazmat.primitives.asymmetric import padding
22
23
24 def generate_password():
25 s = "abcdefghijklmnopqrstuvwxyz01234567890ABCDEFGHIJKLMNOPQRSTUVWXYZ"
26 passlen = 16
27 p = "".join(random.sample(s, passlen))
28 return p
29
30
31 def sign_all(content_folder, signing_pri_key):
32 signatures = dict()
33 for f in os.listdir(content_folder):
34 path = os.path.join(content_folder, f)
35 if os.path.isfile(path):
36 signature = signing_pri_key.sign(
37 data=open(path, "rb").read(),
38 padding=padding.PSS(
39 mgf=padding.MGF1(hashes.SHA256()),
40 salt_length=padding.PSS.MAX_LENGTH,
41 ),
42 algorithm=hashes.SHA256(),
43 )
44 signatures[f] = b64encode(signature).decode("utf-8")
45 return signatures
46
47
48 def load_yaml(file_name):
49 return yaml.load(open(file_name, "r"), Loader=yaml.Loader)
50
51
52 def sh_replace(src, mapping_dict):
53 result = src
54 for k, v in mapping_dict.items():
55 result = result.replace("{~~" + k + "~~}", str(v))
56 return result
57
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/nvflare/lighter/provision.py b/nvflare/lighter/provision.py
--- a/nvflare/lighter/provision.py
+++ b/nvflare/lighter/provision.py
@@ -21,10 +21,9 @@
import sys
import webbrowser
-import yaml
-
from nvflare.fuel.utils.class_utils import instantiate_class
from nvflare.lighter.spec import Participant, Project, Provisioner
+from nvflare.lighter.utils import load_yaml
def main():
@@ -77,7 +76,7 @@
project_full_path = os.path.join(current_path, project_file)
print(f"Project yaml file: {project_full_path}.")
- project_dict = yaml.load(open(project_full_path, "r"), Loader=yaml.Loader)
+ project_dict = load_yaml(project_full_path)
api_version = project_dict.get("api_version")
if api_version not in [3]:
raise ValueError(f"API version expected 3 but found {api_version}")
diff --git a/nvflare/lighter/utils.py b/nvflare/lighter/utils.py
--- a/nvflare/lighter/utils.py
+++ b/nvflare/lighter/utils.py
@@ -46,7 +46,7 @@
def load_yaml(file_name):
- return yaml.load(open(file_name, "r"), Loader=yaml.Loader)
+ return yaml.safe_load(open(file_name, "r"))
def sh_replace(src, mapping_dict):
| {"golden_diff": "diff --git a/nvflare/lighter/provision.py b/nvflare/lighter/provision.py\n--- a/nvflare/lighter/provision.py\n+++ b/nvflare/lighter/provision.py\n@@ -21,10 +21,9 @@\n import sys\n import webbrowser\n \n-import yaml\n-\n from nvflare.fuel.utils.class_utils import instantiate_class\n from nvflare.lighter.spec import Participant, Project, Provisioner\n+from nvflare.lighter.utils import load_yaml\n \n \n def main():\n@@ -77,7 +76,7 @@\n project_full_path = os.path.join(current_path, project_file)\n print(f\"Project yaml file: {project_full_path}.\")\n \n- project_dict = yaml.load(open(project_full_path, \"r\"), Loader=yaml.Loader)\n+ project_dict = load_yaml(project_full_path)\n api_version = project_dict.get(\"api_version\")\n if api_version not in [3]:\n raise ValueError(f\"API version expected 3 but found {api_version}\")\ndiff --git a/nvflare/lighter/utils.py b/nvflare/lighter/utils.py\n--- a/nvflare/lighter/utils.py\n+++ b/nvflare/lighter/utils.py\n@@ -46,7 +46,7 @@\n \n \n def load_yaml(file_name):\n- return yaml.load(open(file_name, \"r\"), Loader=yaml.Loader)\n+ return yaml.safe_load(open(file_name, \"r\"))\n \n \n def sh_replace(src, mapping_dict):\n", "issue": "Yaml loader should be replaced by safe_loader or other more secure loader\nTo load yaml files from unknown source, we should avoid using yaml's loader. A better way is to use either safe_loader or other mechanism.\n", "before_files": [{"content": "# Copyright (c) 2021-2022, NVIDIA CORPORATION. All rights reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom __future__ import absolute_import\n\nimport argparse\nimport os\nimport pathlib\nimport shutil\nimport sys\nimport webbrowser\n\nimport yaml\n\nfrom nvflare.fuel.utils.class_utils import instantiate_class\nfrom nvflare.lighter.spec import Participant, Project, Provisioner\n\n\ndef main():\n parser = argparse.ArgumentParser()\n parser.add_argument(\"-p\", \"--project_file\", type=str, default=\"project.yml\", help=\"file to describe FL project\")\n parser.add_argument(\"-w\", \"--workspace\", type=str, default=\"workspace\", help=\"directory used by provision\")\n parser.add_argument(\"-c\", \"--custom_folder\", type=str, default=\".\", help=\"additional folder to load python codes\")\n parser.add_argument(\n \"-u\",\n \"--ui_tool\",\n action=\"store_true\",\n help=\"Run provisioning UI tool to generate project.yml file\",\n )\n\n args = parser.parse_args()\n\n file_path = pathlib.Path(__file__).parent.absolute()\n current_path = os.getcwd()\n custom_folder_path = os.path.join(current_path, args.custom_folder)\n sys.path.append(custom_folder_path)\n print(\"Path list (sys.path) for python codes loading: {}\".format(sys.path))\n\n # main project file\n project_file = args.project_file\n current_project_yml = os.path.join(current_path, \"project.yml\")\n if len(sys.argv) == 1 and not os.path.exists(current_project_yml):\n answer = input(\n f\"No project.yml found in current folder. Is it OK to generate one at {current_project_yml} for you? (y/N) \"\n )\n if answer.strip().upper() == \"Y\":\n shutil.copyfile(os.path.join(file_path, \"project.yml\"), current_project_yml)\n print(f\"{current_project_yml} was created. Please edit it to fit your FL configuration.\")\n exit(0)\n\n if args.ui_tool:\n ui_helper_path = os.path.join(file_path, \"provision_helper.html\")\n ui_helper_url = f\"file://{ui_helper_path}\"\n webbrowser.open_new_tab(ui_helper_url)\n print(\n \"\\n******\\n\"\n \"Now launching provisioning UI tool.\\n\"\n \"After generating project.yml in the browser and saving it to your local folder,\\n\"\n \"please re-run provision with -p option, pointing to the generated project.yml, to generate all packages.\\n******\\n\"\n )\n exit(0)\n\n workspace = args.workspace\n workspace_full_path = os.path.join(current_path, workspace)\n\n project_full_path = os.path.join(current_path, project_file)\n print(f\"Project yaml file: {project_full_path}.\")\n\n project_dict = yaml.load(open(project_full_path, \"r\"), Loader=yaml.Loader)\n api_version = project_dict.get(\"api_version\")\n if api_version not in [3]:\n raise ValueError(f\"API version expected 3 but found {api_version}\")\n\n project_name = project_dict.get(\"name\")\n project_description = project_dict.get(\"description\", \"\")\n participants = list()\n for p in project_dict.get(\"participants\"):\n participants.append(Participant(**p))\n project = Project(name=project_name, description=project_description, participants=participants)\n\n builders = list()\n for b in project_dict.get(\"builders\"):\n path = b.get(\"path\")\n args = b.get(\"args\")\n builders.append(instantiate_class(path, args))\n\n provisioner = Provisioner(workspace_full_path, builders)\n\n provisioner.provision(project)\n\n\nif __name__ == \"__main__\":\n main()\n", "path": "nvflare/lighter/provision.py"}, {"content": "# Copyright (c) 2021-2022, NVIDIA CORPORATION. All rights reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nimport os\nimport random\nfrom base64 import b64encode\n\nimport yaml\nfrom cryptography.hazmat.primitives import hashes\nfrom cryptography.hazmat.primitives.asymmetric import padding\n\n\ndef generate_password():\n s = \"abcdefghijklmnopqrstuvwxyz01234567890ABCDEFGHIJKLMNOPQRSTUVWXYZ\"\n passlen = 16\n p = \"\".join(random.sample(s, passlen))\n return p\n\n\ndef sign_all(content_folder, signing_pri_key):\n signatures = dict()\n for f in os.listdir(content_folder):\n path = os.path.join(content_folder, f)\n if os.path.isfile(path):\n signature = signing_pri_key.sign(\n data=open(path, \"rb\").read(),\n padding=padding.PSS(\n mgf=padding.MGF1(hashes.SHA256()),\n salt_length=padding.PSS.MAX_LENGTH,\n ),\n algorithm=hashes.SHA256(),\n )\n signatures[f] = b64encode(signature).decode(\"utf-8\")\n return signatures\n\n\ndef load_yaml(file_name):\n return yaml.load(open(file_name, \"r\"), Loader=yaml.Loader)\n\n\ndef sh_replace(src, mapping_dict):\n result = src\n for k, v in mapping_dict.items():\n result = result.replace(\"{~~\" + k + \"~~}\", str(v))\n return result\n", "path": "nvflare/lighter/utils.py"}], "after_files": [{"content": "# Copyright (c) 2021-2022, NVIDIA CORPORATION. All rights reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom __future__ import absolute_import\n\nimport argparse\nimport os\nimport pathlib\nimport shutil\nimport sys\nimport webbrowser\n\nfrom nvflare.fuel.utils.class_utils import instantiate_class\nfrom nvflare.lighter.spec import Participant, Project, Provisioner\nfrom nvflare.lighter.utils import load_yaml\n\n\ndef main():\n parser = argparse.ArgumentParser()\n parser.add_argument(\"-p\", \"--project_file\", type=str, default=\"project.yml\", help=\"file to describe FL project\")\n parser.add_argument(\"-w\", \"--workspace\", type=str, default=\"workspace\", help=\"directory used by provision\")\n parser.add_argument(\"-c\", \"--custom_folder\", type=str, default=\".\", help=\"additional folder to load python codes\")\n parser.add_argument(\n \"-u\",\n \"--ui_tool\",\n action=\"store_true\",\n help=\"Run provisioning UI tool to generate project.yml file\",\n )\n\n args = parser.parse_args()\n\n file_path = pathlib.Path(__file__).parent.absolute()\n current_path = os.getcwd()\n custom_folder_path = os.path.join(current_path, args.custom_folder)\n sys.path.append(custom_folder_path)\n print(\"Path list (sys.path) for python codes loading: {}\".format(sys.path))\n\n # main project file\n project_file = args.project_file\n current_project_yml = os.path.join(current_path, \"project.yml\")\n if len(sys.argv) == 1 and not os.path.exists(current_project_yml):\n answer = input(\n f\"No project.yml found in current folder. Is it OK to generate one at {current_project_yml} for you? (y/N) \"\n )\n if answer.strip().upper() == \"Y\":\n shutil.copyfile(os.path.join(file_path, \"project.yml\"), current_project_yml)\n print(f\"{current_project_yml} was created. Please edit it to fit your FL configuration.\")\n exit(0)\n\n if args.ui_tool:\n ui_helper_path = os.path.join(file_path, \"provision_helper.html\")\n ui_helper_url = f\"file://{ui_helper_path}\"\n webbrowser.open_new_tab(ui_helper_url)\n print(\n \"\\n******\\n\"\n \"Now launching provisioning UI tool.\\n\"\n \"After generating project.yml in the browser and saving it to your local folder,\\n\"\n \"please re-run provision with -p option, pointing to the generated project.yml, to generate all packages.\\n******\\n\"\n )\n exit(0)\n\n workspace = args.workspace\n workspace_full_path = os.path.join(current_path, workspace)\n\n project_full_path = os.path.join(current_path, project_file)\n print(f\"Project yaml file: {project_full_path}.\")\n\n project_dict = load_yaml(project_full_path)\n api_version = project_dict.get(\"api_version\")\n if api_version not in [3]:\n raise ValueError(f\"API version expected 3 but found {api_version}\")\n\n project_name = project_dict.get(\"name\")\n project_description = project_dict.get(\"description\", \"\")\n participants = list()\n for p in project_dict.get(\"participants\"):\n participants.append(Participant(**p))\n project = Project(name=project_name, description=project_description, participants=participants)\n\n builders = list()\n for b in project_dict.get(\"builders\"):\n path = b.get(\"path\")\n args = b.get(\"args\")\n builders.append(instantiate_class(path, args))\n\n provisioner = Provisioner(workspace_full_path, builders)\n\n provisioner.provision(project)\n\n\nif __name__ == \"__main__\":\n main()\n", "path": "nvflare/lighter/provision.py"}, {"content": "# Copyright (c) 2021-2022, NVIDIA CORPORATION. All rights reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nimport os\nimport random\nfrom base64 import b64encode\n\nimport yaml\nfrom cryptography.hazmat.primitives import hashes\nfrom cryptography.hazmat.primitives.asymmetric import padding\n\n\ndef generate_password():\n s = \"abcdefghijklmnopqrstuvwxyz01234567890ABCDEFGHIJKLMNOPQRSTUVWXYZ\"\n passlen = 16\n p = \"\".join(random.sample(s, passlen))\n return p\n\n\ndef sign_all(content_folder, signing_pri_key):\n signatures = dict()\n for f in os.listdir(content_folder):\n path = os.path.join(content_folder, f)\n if os.path.isfile(path):\n signature = signing_pri_key.sign(\n data=open(path, \"rb\").read(),\n padding=padding.PSS(\n mgf=padding.MGF1(hashes.SHA256()),\n salt_length=padding.PSS.MAX_LENGTH,\n ),\n algorithm=hashes.SHA256(),\n )\n signatures[f] = b64encode(signature).decode(\"utf-8\")\n return signatures\n\n\ndef load_yaml(file_name):\n return yaml.safe_load(open(file_name, \"r\"))\n\n\ndef sh_replace(src, mapping_dict):\n result = src\n for k, v in mapping_dict.items():\n result = result.replace(\"{~~\" + k + \"~~}\", str(v))\n return result\n", "path": "nvflare/lighter/utils.py"}]} | 1,980 | 315 |
gh_patches_debug_16051 | rasdani/github-patches | git_diff | docker__docker-py-812 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
requests 2.8.0 re-introduces bug #647
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `docker/unixconn/unixconn.py`
Content:
```
1 # Copyright 2013 dotCloud inc.
2
3 # Licensed under the Apache License, Version 2.0 (the "License");
4 # you may not use this file except in compliance with the License.
5 # You may obtain a copy of the License at
6
7 # http://www.apache.org/licenses/LICENSE-2.0
8
9 # Unless required by applicable law or agreed to in writing, software
10 # distributed under the License is distributed on an "AS IS" BASIS,
11 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12 # See the License for the specific language governing permissions and
13 # limitations under the License.
14 import six
15 import requests.adapters
16 import socket
17
18 if six.PY3:
19 import http.client as httplib
20 else:
21 import httplib
22
23 try:
24 import requests.packages.urllib3 as urllib3
25 except ImportError:
26 import urllib3
27
28 RecentlyUsedContainer = urllib3._collections.RecentlyUsedContainer
29
30
31 class UnixHTTPConnection(httplib.HTTPConnection, object):
32 def __init__(self, base_url, unix_socket, timeout=60):
33 httplib.HTTPConnection.__init__(self, 'localhost', timeout=timeout)
34 self.base_url = base_url
35 self.unix_socket = unix_socket
36 self.timeout = timeout
37
38 def connect(self):
39 sock = socket.socket(socket.AF_UNIX, socket.SOCK_STREAM)
40 sock.settimeout(self.timeout)
41 sock.connect(self.unix_socket)
42 self.sock = sock
43
44
45 class UnixHTTPConnectionPool(urllib3.connectionpool.HTTPConnectionPool):
46 def __init__(self, base_url, socket_path, timeout=60):
47 urllib3.connectionpool.HTTPConnectionPool.__init__(
48 self, 'localhost', timeout=timeout
49 )
50 self.base_url = base_url
51 self.socket_path = socket_path
52 self.timeout = timeout
53
54 def _new_conn(self):
55 return UnixHTTPConnection(self.base_url, self.socket_path,
56 self.timeout)
57
58
59 class UnixAdapter(requests.adapters.HTTPAdapter):
60 def __init__(self, socket_url, timeout=60):
61 socket_path = socket_url.replace('http+unix://', '')
62 if not socket_path.startswith('/'):
63 socket_path = '/' + socket_path
64 self.socket_path = socket_path
65 self.timeout = timeout
66 self.pools = RecentlyUsedContainer(10,
67 dispose_func=lambda p: p.close())
68 super(UnixAdapter, self).__init__()
69
70 def get_connection(self, url, proxies=None):
71 with self.pools.lock:
72 pool = self.pools.get(url)
73 if pool:
74 return pool
75
76 pool = UnixHTTPConnectionPool(url,
77 self.socket_path,
78 self.timeout)
79 self.pools[url] = pool
80
81 return pool
82
83 def close(self):
84 self.pools.clear()
85
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/docker/unixconn/unixconn.py b/docker/unixconn/unixconn.py
--- a/docker/unixconn/unixconn.py
+++ b/docker/unixconn/unixconn.py
@@ -73,12 +73,20 @@
if pool:
return pool
- pool = UnixHTTPConnectionPool(url,
- self.socket_path,
- self.timeout)
+ pool = UnixHTTPConnectionPool(
+ url, self.socket_path, self.timeout
+ )
self.pools[url] = pool
return pool
+ def request_url(self, request, proxies):
+ # The select_proxy utility in requests errors out when the provided URL
+ # doesn't have a hostname, like is the case when using a UNIX socket.
+ # Since proxies are an irrelevant notion in the case of UNIX sockets
+ # anyway, we simply return the path URL directly.
+ # See also: https://github.com/docker/docker-py/issues/811
+ return request.path_url
+
def close(self):
self.pools.clear()
| {"golden_diff": "diff --git a/docker/unixconn/unixconn.py b/docker/unixconn/unixconn.py\n--- a/docker/unixconn/unixconn.py\n+++ b/docker/unixconn/unixconn.py\n@@ -73,12 +73,20 @@\n if pool:\n return pool\n \n- pool = UnixHTTPConnectionPool(url,\n- self.socket_path,\n- self.timeout)\n+ pool = UnixHTTPConnectionPool(\n+ url, self.socket_path, self.timeout\n+ )\n self.pools[url] = pool\n \n return pool\n \n+ def request_url(self, request, proxies):\n+ # The select_proxy utility in requests errors out when the provided URL\n+ # doesn't have a hostname, like is the case when using a UNIX socket.\n+ # Since proxies are an irrelevant notion in the case of UNIX sockets\n+ # anyway, we simply return the path URL directly.\n+ # See also: https://github.com/docker/docker-py/issues/811\n+ return request.path_url\n+\n def close(self):\n self.pools.clear()\n", "issue": "requests 2.8.0 re-introduces bug #647\n\n", "before_files": [{"content": "# Copyright 2013 dotCloud inc.\n\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n\n# http://www.apache.org/licenses/LICENSE-2.0\n\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\nimport six\nimport requests.adapters\nimport socket\n\nif six.PY3:\n import http.client as httplib\nelse:\n import httplib\n\ntry:\n import requests.packages.urllib3 as urllib3\nexcept ImportError:\n import urllib3\n\nRecentlyUsedContainer = urllib3._collections.RecentlyUsedContainer\n\n\nclass UnixHTTPConnection(httplib.HTTPConnection, object):\n def __init__(self, base_url, unix_socket, timeout=60):\n httplib.HTTPConnection.__init__(self, 'localhost', timeout=timeout)\n self.base_url = base_url\n self.unix_socket = unix_socket\n self.timeout = timeout\n\n def connect(self):\n sock = socket.socket(socket.AF_UNIX, socket.SOCK_STREAM)\n sock.settimeout(self.timeout)\n sock.connect(self.unix_socket)\n self.sock = sock\n\n\nclass UnixHTTPConnectionPool(urllib3.connectionpool.HTTPConnectionPool):\n def __init__(self, base_url, socket_path, timeout=60):\n urllib3.connectionpool.HTTPConnectionPool.__init__(\n self, 'localhost', timeout=timeout\n )\n self.base_url = base_url\n self.socket_path = socket_path\n self.timeout = timeout\n\n def _new_conn(self):\n return UnixHTTPConnection(self.base_url, self.socket_path,\n self.timeout)\n\n\nclass UnixAdapter(requests.adapters.HTTPAdapter):\n def __init__(self, socket_url, timeout=60):\n socket_path = socket_url.replace('http+unix://', '')\n if not socket_path.startswith('/'):\n socket_path = '/' + socket_path\n self.socket_path = socket_path\n self.timeout = timeout\n self.pools = RecentlyUsedContainer(10,\n dispose_func=lambda p: p.close())\n super(UnixAdapter, self).__init__()\n\n def get_connection(self, url, proxies=None):\n with self.pools.lock:\n pool = self.pools.get(url)\n if pool:\n return pool\n\n pool = UnixHTTPConnectionPool(url,\n self.socket_path,\n self.timeout)\n self.pools[url] = pool\n\n return pool\n\n def close(self):\n self.pools.clear()\n", "path": "docker/unixconn/unixconn.py"}], "after_files": [{"content": "# Copyright 2013 dotCloud inc.\n\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n\n# http://www.apache.org/licenses/LICENSE-2.0\n\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\nimport six\nimport requests.adapters\nimport socket\n\nif six.PY3:\n import http.client as httplib\nelse:\n import httplib\n\ntry:\n import requests.packages.urllib3 as urllib3\nexcept ImportError:\n import urllib3\n\nRecentlyUsedContainer = urllib3._collections.RecentlyUsedContainer\n\n\nclass UnixHTTPConnection(httplib.HTTPConnection, object):\n def __init__(self, base_url, unix_socket, timeout=60):\n httplib.HTTPConnection.__init__(self, 'localhost', timeout=timeout)\n self.base_url = base_url\n self.unix_socket = unix_socket\n self.timeout = timeout\n\n def connect(self):\n sock = socket.socket(socket.AF_UNIX, socket.SOCK_STREAM)\n sock.settimeout(self.timeout)\n sock.connect(self.unix_socket)\n self.sock = sock\n\n\nclass UnixHTTPConnectionPool(urllib3.connectionpool.HTTPConnectionPool):\n def __init__(self, base_url, socket_path, timeout=60):\n urllib3.connectionpool.HTTPConnectionPool.__init__(\n self, 'localhost', timeout=timeout\n )\n self.base_url = base_url\n self.socket_path = socket_path\n self.timeout = timeout\n\n def _new_conn(self):\n return UnixHTTPConnection(self.base_url, self.socket_path,\n self.timeout)\n\n\nclass UnixAdapter(requests.adapters.HTTPAdapter):\n def __init__(self, socket_url, timeout=60):\n socket_path = socket_url.replace('http+unix://', '')\n if not socket_path.startswith('/'):\n socket_path = '/' + socket_path\n self.socket_path = socket_path\n self.timeout = timeout\n self.pools = RecentlyUsedContainer(10,\n dispose_func=lambda p: p.close())\n super(UnixAdapter, self).__init__()\n\n def get_connection(self, url, proxies=None):\n with self.pools.lock:\n pool = self.pools.get(url)\n if pool:\n return pool\n\n pool = UnixHTTPConnectionPool(\n url, self.socket_path, self.timeout\n )\n self.pools[url] = pool\n\n return pool\n\n def request_url(self, request, proxies):\n # The select_proxy utility in requests errors out when the provided URL\n # doesn't have a hostname, like is the case when using a UNIX socket.\n # Since proxies are an irrelevant notion in the case of UNIX sockets\n # anyway, we simply return the path URL directly.\n # See also: https://github.com/docker/docker-py/issues/811\n return request.path_url\n\n def close(self):\n self.pools.clear()\n", "path": "docker/unixconn/unixconn.py"}]} | 1,047 | 237 |
gh_patches_debug_2925 | rasdani/github-patches | git_diff | spack__spack-20572 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
improve installation of Zoltran: imposing +int64 constrains on parmetis
<!--*Please add a concise summary of your suggestion here.*-->
### Rationale
zoltan spec has a variant called `int64` which imposes the corresponding constrain on metis.
https://github.com/spack/spack/blob/6947951aaf9954b1dfd12ca7a9266d7335f07105/var/spack/repos/builtin/packages/zoltan/package.py#L37-L44
The same constrain must be applied to parmetis.
<!--*Is your feature request related to a problem? Please describe it!*-->
### Description
I guess a solution can be something like
```
depends_on('parmetis@4:', when='+parmetis')
depends_on('parmetis@4: +int64', when='+parmetis+int64')
```
<!--*Describe the solution you'd like and the alternatives you have considered.*-->
### Additional information
<!--*Add any other context about the feature request here.*-->
I guess this happens because parmetis package has been recently updated and `int64` has been added. Because there was no such an option in parmetis for a long time people came up with a workaround by specifying `metis+int64` explicitly in their script. The parametis update brings an inconsistency because `int64` is off by default in parmetis, however, and the ''legacy'' workaround imposes `int64` on metis.
My spack version is 0.16.0
### General information
- [x] I have run `spack --version` and reported the version of Spack
- [x] I have searched the issues of this repo and believe this is not a duplicate
<!--If you want to ask a question about the tool (how to use it, what it can currently do, etc.), try the `#general` channel on our Slack first. We have a welcoming community and chances are you'll get your reply faster and without opening an issue.
Other than that, thanks for taking the time to contribute to Spack!
-->
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `var/spack/repos/builtin/packages/zoltan/package.py`
Content:
```
1 # Copyright 2013-2020 Lawrence Livermore National Security, LLC and other
2 # Spack Project Developers. See the top-level COPYRIGHT file for details.
3 #
4 # SPDX-License-Identifier: (Apache-2.0 OR MIT)
5
6
7 from spack import *
8 import re
9
10
11 class Zoltan(AutotoolsPackage):
12 """The Zoltan library is a toolkit of parallel combinatorial algorithms
13 for parallel, unstructured, and/or adaptive scientific
14 applications. Zoltan's largest component is a suite of dynamic
15 load-balancing and partitioning algorithms that increase
16 applications' parallel performance by reducing idle time. Zoltan
17 also has graph coloring and graph ordering algorithms, which are
18 useful in task schedulers and parallel preconditioners.
19
20 """
21
22 homepage = "http://www.cs.sandia.gov/zoltan"
23 url = "http://www.cs.sandia.gov/~kddevin/Zoltan_Distributions/zoltan_distrib_v3.83.tar.gz"
24
25 version('3.83', sha256='d0d78fdeab7a385c87d3666b8a8dc748994ff04d3fd846872a4845e12d79c1bb')
26 version('3.8', sha256='5bdd46548fb9c73b225bbcf3d206c558c318cb292f0b19645e536315d14aafb7')
27 version('3.6', sha256='d2cb41e5fb72ca564b24bc5f21d82d9f7992f2c977bc82b243a01a8a8ee4eb9c')
28 version('3.3', sha256='8a90585674ab1bbd011dab29f778b9816519712c78d0aab4cdde9c68f02b30dc')
29
30 patch('notparallel.patch', when='@3.8')
31
32 variant('debug', default=False, description='Builds a debug version of the library.')
33 variant('shared', default=True, description='Builds a shared version of the library.')
34
35 variant('fortran', default=True, description='Enable Fortran support.')
36 variant('mpi', default=True, description='Enable MPI support.')
37 variant('parmetis', default=False, description='Enable ParMETIS support.')
38 variant('int64', default=False, description='Enable 64bit indices.')
39
40 depends_on('mpi', when='+mpi')
41
42 depends_on('parmetis@4:', when='+parmetis')
43 depends_on('metis+int64', when='+parmetis+int64')
44 depends_on('metis', when='+parmetis')
45
46 depends_on('perl@:5.21', type='build', when='@:3.6')
47 depends_on('autoconf', type='build')
48 depends_on('automake', type='build')
49 depends_on('m4', type='build')
50
51 conflicts('+parmetis', when='~mpi')
52
53 build_directory = 'spack-build'
54
55 @property
56 def configure_directory(self):
57 spec = self.spec
58
59 # FIXME: The older Zoltan versions fail to compile the F90 MPI wrappers
60 # because of some complicated generic type problem.
61 if spec.satisfies('@:3.6+fortran+mpi'):
62 raise RuntimeError(('Cannot build Zoltan v{0} with +fortran and '
63 '+mpi; please disable one of these features '
64 'or upgrade versions.').format(self.version))
65 if spec.satisfies('@:3.6'):
66 zoltan_path = 'Zoltan_v{0}'.format(self.version)
67 return zoltan_path
68 return '.'
69
70 @property
71 def parallel(self):
72 # NOTE: Earlier versions of Zoltan cannot be built in parallel
73 # because they contain nested Makefile dependency bugs.
74 return not self.spec.satisfies('@:3.6+fortran')
75
76 def autoreconf(self, spec, prefix):
77 autoreconf = which('autoreconf')
78 with working_dir(self.configure_directory):
79 autoreconf('-ivf')
80
81 def configure_args(self):
82 spec = self.spec
83
84 config_args = [
85 self.get_config_flag('f90interface', 'fortran'),
86 self.get_config_flag('mpi', 'mpi'),
87 ]
88 config_cflags = [
89 '-O0' if '+debug' in spec else '-O3',
90 '-g' if '+debug' in spec else '',
91 ]
92
93 config_ldflags = []
94 # PGI runtime libraries
95 if '%pgi' in spec:
96 config_ldflags.append('-pgf90libs')
97 if '+shared' in spec:
98 config_args.extend([
99 'RANLIB=echo',
100 '--with-ar=$(CXX) -shared $(LDFLAGS) -o'
101 ])
102 config_cflags.append(self.compiler.cc_pic_flag)
103 if spec.satisfies('%gcc'):
104 config_args.append('--with-libs=-lgfortran')
105 if spec.satisfies('%intel'):
106 config_args.append('--with-libs=-lifcore')
107
108 if '+int64' in spec:
109 config_args.append('--with-id-type=ulong')
110
111 if '+parmetis' in spec:
112 parmetis_prefix = spec['parmetis'].prefix
113 config_args.extend([
114 '--with-parmetis',
115 '--with-parmetis-libdir={0}'.format(parmetis_prefix.lib),
116 '--with-parmetis-incdir={0}'.format(parmetis_prefix.include),
117 '--with-incdirs=-I{0}'.format(spec['metis'].prefix.include),
118 '--with-ldflags=-L{0}'.format(spec['metis'].prefix.lib)
119 ])
120 if '+int64' in spec['metis']:
121 config_args.append('--with-id-type=ulong')
122 else:
123 config_args.append('--with-id-type=uint')
124
125 if '+mpi' in spec:
126 config_args.extend([
127 'CC={0}'.format(spec['mpi'].mpicc),
128 'CXX={0}'.format(spec['mpi'].mpicxx),
129 'FC={0}'.format(spec['mpi'].mpifc),
130 '--with-mpi={0}'.format(spec['mpi'].prefix),
131
132 # NOTE: Zoltan assumes that it's linking against an MPI library
133 # that can be found with '-lmpi' which isn't the case for many
134 # MPI packages. We rely on the MPI-wrappers to automatically
135 # add what is required for linking and thus pass an empty
136 # list of libs
137 '--with-mpi-libs= '
138 ])
139
140 config_fcflags = config_cflags[:]
141 if spec.satisfies('%gcc@10:+fortran'):
142 config_fcflags.append('-fallow-argument-mismatch')
143 # NOTE: Early versions of Zoltan come packaged with a few embedded
144 # library packages (e.g. ParMETIS, Scotch), which messes with Spack's
145 # ability to descend directly into the package's source directory.
146 config_args.extend([
147 '--with-cflags={0}'.format(' '.join(config_cflags)),
148 '--with-cxxflags={0}'.format(' '.join(config_cflags)),
149 '--with-fcflags={0}'.format(' '.join(config_fcflags)),
150 '--with-ldflags={0}'.format(' '.join(config_ldflags))
151 ])
152 return config_args
153
154 # NOTE: Unfortunately, Zoltan doesn't provide any configuration
155 # options for the extension of the output library files, so this
156 # script must change these extensions as a post-processing step.
157 @run_after('install')
158 def solib_install(self):
159 if '+shared' in self.spec:
160 for lib_path in find(self.spec.prefix.lib, 'lib*.a'):
161 lib_shared_name = re.sub(r'\.a$', '.{0}'.format(dso_suffix),
162 lib_path)
163 move(lib_path, lib_shared_name)
164
165 def get_config_flag(self, flag_name, flag_variant):
166 flag_pre = 'en' if '+{0}'.format(flag_variant) in self.spec else 'dis'
167 return '--{0}able-{1}'.format(flag_pre, flag_name)
168
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/var/spack/repos/builtin/packages/zoltan/package.py b/var/spack/repos/builtin/packages/zoltan/package.py
--- a/var/spack/repos/builtin/packages/zoltan/package.py
+++ b/var/spack/repos/builtin/packages/zoltan/package.py
@@ -39,6 +39,7 @@
depends_on('mpi', when='+mpi')
+ depends_on('parmetis@4: +int64', when='+parmetis+int64')
depends_on('parmetis@4:', when='+parmetis')
depends_on('metis+int64', when='+parmetis+int64')
depends_on('metis', when='+parmetis')
| {"golden_diff": "diff --git a/var/spack/repos/builtin/packages/zoltan/package.py b/var/spack/repos/builtin/packages/zoltan/package.py\n--- a/var/spack/repos/builtin/packages/zoltan/package.py\n+++ b/var/spack/repos/builtin/packages/zoltan/package.py\n@@ -39,6 +39,7 @@\n \n depends_on('mpi', when='+mpi')\n \n+ depends_on('parmetis@4: +int64', when='+parmetis+int64')\n depends_on('parmetis@4:', when='+parmetis')\n depends_on('metis+int64', when='+parmetis+int64')\n depends_on('metis', when='+parmetis')\n", "issue": "improve installation of Zoltran: imposing +int64 constrains on parmetis\n<!--*Please add a concise summary of your suggestion here.*-->\r\n\r\n### Rationale\r\nzoltan spec has a variant called `int64` which imposes the corresponding constrain on metis. \r\nhttps://github.com/spack/spack/blob/6947951aaf9954b1dfd12ca7a9266d7335f07105/var/spack/repos/builtin/packages/zoltan/package.py#L37-L44\r\n\r\nThe same constrain must be applied to parmetis. \r\n\r\n\r\n<!--*Is your feature request related to a problem? Please describe it!*-->\r\n\r\n### Description\r\nI guess a solution can be something like\r\n```\r\ndepends_on('parmetis@4:', when='+parmetis') \r\ndepends_on('parmetis@4: +int64', when='+parmetis+int64')\r\n```\r\n\r\n<!--*Describe the solution you'd like and the alternatives you have considered.*-->\r\n\r\n\r\n### Additional information\r\n<!--*Add any other context about the feature request here.*-->\r\nI guess this happens because parmetis package has been recently updated and `int64` has been added. Because there was no such an option in parmetis for a long time people came up with a workaround by specifying `metis+int64` explicitly in their script. The parametis update brings an inconsistency because `int64` is off by default in parmetis, however, and the ''legacy'' workaround imposes `int64` on metis.\r\nMy spack version is 0.16.0\r\n\r\n### General information\r\n\r\n- [x] I have run `spack --version` and reported the version of Spack\r\n- [x] I have searched the issues of this repo and believe this is not a duplicate\r\n\r\n\r\n\r\n<!--If you want to ask a question about the tool (how to use it, what it can currently do, etc.), try the `#general` channel on our Slack first. We have a welcoming community and chances are you'll get your reply faster and without opening an issue.\r\n\r\nOther than that, thanks for taking the time to contribute to Spack!\r\n-->\n", "before_files": [{"content": "# Copyright 2013-2020 Lawrence Livermore National Security, LLC and other\n# Spack Project Developers. See the top-level COPYRIGHT file for details.\n#\n# SPDX-License-Identifier: (Apache-2.0 OR MIT)\n\n\nfrom spack import *\nimport re\n\n\nclass Zoltan(AutotoolsPackage):\n \"\"\"The Zoltan library is a toolkit of parallel combinatorial algorithms\n for parallel, unstructured, and/or adaptive scientific\n applications. Zoltan's largest component is a suite of dynamic\n load-balancing and partitioning algorithms that increase\n applications' parallel performance by reducing idle time. Zoltan\n also has graph coloring and graph ordering algorithms, which are\n useful in task schedulers and parallel preconditioners.\n\n \"\"\"\n\n homepage = \"http://www.cs.sandia.gov/zoltan\"\n url = \"http://www.cs.sandia.gov/~kddevin/Zoltan_Distributions/zoltan_distrib_v3.83.tar.gz\"\n\n version('3.83', sha256='d0d78fdeab7a385c87d3666b8a8dc748994ff04d3fd846872a4845e12d79c1bb')\n version('3.8', sha256='5bdd46548fb9c73b225bbcf3d206c558c318cb292f0b19645e536315d14aafb7')\n version('3.6', sha256='d2cb41e5fb72ca564b24bc5f21d82d9f7992f2c977bc82b243a01a8a8ee4eb9c')\n version('3.3', sha256='8a90585674ab1bbd011dab29f778b9816519712c78d0aab4cdde9c68f02b30dc')\n\n patch('notparallel.patch', when='@3.8')\n\n variant('debug', default=False, description='Builds a debug version of the library.')\n variant('shared', default=True, description='Builds a shared version of the library.')\n\n variant('fortran', default=True, description='Enable Fortran support.')\n variant('mpi', default=True, description='Enable MPI support.')\n variant('parmetis', default=False, description='Enable ParMETIS support.')\n variant('int64', default=False, description='Enable 64bit indices.')\n\n depends_on('mpi', when='+mpi')\n\n depends_on('parmetis@4:', when='+parmetis')\n depends_on('metis+int64', when='+parmetis+int64')\n depends_on('metis', when='+parmetis')\n\n depends_on('perl@:5.21', type='build', when='@:3.6')\n depends_on('autoconf', type='build')\n depends_on('automake', type='build')\n depends_on('m4', type='build')\n\n conflicts('+parmetis', when='~mpi')\n\n build_directory = 'spack-build'\n\n @property\n def configure_directory(self):\n spec = self.spec\n\n # FIXME: The older Zoltan versions fail to compile the F90 MPI wrappers\n # because of some complicated generic type problem.\n if spec.satisfies('@:3.6+fortran+mpi'):\n raise RuntimeError(('Cannot build Zoltan v{0} with +fortran and '\n '+mpi; please disable one of these features '\n 'or upgrade versions.').format(self.version))\n if spec.satisfies('@:3.6'):\n zoltan_path = 'Zoltan_v{0}'.format(self.version)\n return zoltan_path\n return '.'\n\n @property\n def parallel(self):\n # NOTE: Earlier versions of Zoltan cannot be built in parallel\n # because they contain nested Makefile dependency bugs.\n return not self.spec.satisfies('@:3.6+fortran')\n\n def autoreconf(self, spec, prefix):\n autoreconf = which('autoreconf')\n with working_dir(self.configure_directory):\n autoreconf('-ivf')\n\n def configure_args(self):\n spec = self.spec\n\n config_args = [\n self.get_config_flag('f90interface', 'fortran'),\n self.get_config_flag('mpi', 'mpi'),\n ]\n config_cflags = [\n '-O0' if '+debug' in spec else '-O3',\n '-g' if '+debug' in spec else '',\n ]\n\n config_ldflags = []\n # PGI runtime libraries\n if '%pgi' in spec:\n config_ldflags.append('-pgf90libs')\n if '+shared' in spec:\n config_args.extend([\n 'RANLIB=echo',\n '--with-ar=$(CXX) -shared $(LDFLAGS) -o'\n ])\n config_cflags.append(self.compiler.cc_pic_flag)\n if spec.satisfies('%gcc'):\n config_args.append('--with-libs=-lgfortran')\n if spec.satisfies('%intel'):\n config_args.append('--with-libs=-lifcore')\n\n if '+int64' in spec:\n config_args.append('--with-id-type=ulong')\n\n if '+parmetis' in spec:\n parmetis_prefix = spec['parmetis'].prefix\n config_args.extend([\n '--with-parmetis',\n '--with-parmetis-libdir={0}'.format(parmetis_prefix.lib),\n '--with-parmetis-incdir={0}'.format(parmetis_prefix.include),\n '--with-incdirs=-I{0}'.format(spec['metis'].prefix.include),\n '--with-ldflags=-L{0}'.format(spec['metis'].prefix.lib)\n ])\n if '+int64' in spec['metis']:\n config_args.append('--with-id-type=ulong')\n else:\n config_args.append('--with-id-type=uint')\n\n if '+mpi' in spec:\n config_args.extend([\n 'CC={0}'.format(spec['mpi'].mpicc),\n 'CXX={0}'.format(spec['mpi'].mpicxx),\n 'FC={0}'.format(spec['mpi'].mpifc),\n '--with-mpi={0}'.format(spec['mpi'].prefix),\n\n # NOTE: Zoltan assumes that it's linking against an MPI library\n # that can be found with '-lmpi' which isn't the case for many\n # MPI packages. We rely on the MPI-wrappers to automatically\n # add what is required for linking and thus pass an empty\n # list of libs\n '--with-mpi-libs= '\n ])\n\n config_fcflags = config_cflags[:]\n if spec.satisfies('%gcc@10:+fortran'):\n config_fcflags.append('-fallow-argument-mismatch')\n # NOTE: Early versions of Zoltan come packaged with a few embedded\n # library packages (e.g. ParMETIS, Scotch), which messes with Spack's\n # ability to descend directly into the package's source directory.\n config_args.extend([\n '--with-cflags={0}'.format(' '.join(config_cflags)),\n '--with-cxxflags={0}'.format(' '.join(config_cflags)),\n '--with-fcflags={0}'.format(' '.join(config_fcflags)),\n '--with-ldflags={0}'.format(' '.join(config_ldflags))\n ])\n return config_args\n\n # NOTE: Unfortunately, Zoltan doesn't provide any configuration\n # options for the extension of the output library files, so this\n # script must change these extensions as a post-processing step.\n @run_after('install')\n def solib_install(self):\n if '+shared' in self.spec:\n for lib_path in find(self.spec.prefix.lib, 'lib*.a'):\n lib_shared_name = re.sub(r'\\.a$', '.{0}'.format(dso_suffix),\n lib_path)\n move(lib_path, lib_shared_name)\n\n def get_config_flag(self, flag_name, flag_variant):\n flag_pre = 'en' if '+{0}'.format(flag_variant) in self.spec else 'dis'\n return '--{0}able-{1}'.format(flag_pre, flag_name)\n", "path": "var/spack/repos/builtin/packages/zoltan/package.py"}], "after_files": [{"content": "# Copyright 2013-2020 Lawrence Livermore National Security, LLC and other\n# Spack Project Developers. See the top-level COPYRIGHT file for details.\n#\n# SPDX-License-Identifier: (Apache-2.0 OR MIT)\n\n\nfrom spack import *\nimport re\n\n\nclass Zoltan(AutotoolsPackage):\n \"\"\"The Zoltan library is a toolkit of parallel combinatorial algorithms\n for parallel, unstructured, and/or adaptive scientific\n applications. Zoltan's largest component is a suite of dynamic\n load-balancing and partitioning algorithms that increase\n applications' parallel performance by reducing idle time. Zoltan\n also has graph coloring and graph ordering algorithms, which are\n useful in task schedulers and parallel preconditioners.\n\n \"\"\"\n\n homepage = \"http://www.cs.sandia.gov/zoltan\"\n url = \"http://www.cs.sandia.gov/~kddevin/Zoltan_Distributions/zoltan_distrib_v3.83.tar.gz\"\n\n version('3.83', sha256='d0d78fdeab7a385c87d3666b8a8dc748994ff04d3fd846872a4845e12d79c1bb')\n version('3.8', sha256='5bdd46548fb9c73b225bbcf3d206c558c318cb292f0b19645e536315d14aafb7')\n version('3.6', sha256='d2cb41e5fb72ca564b24bc5f21d82d9f7992f2c977bc82b243a01a8a8ee4eb9c')\n version('3.3', sha256='8a90585674ab1bbd011dab29f778b9816519712c78d0aab4cdde9c68f02b30dc')\n\n patch('notparallel.patch', when='@3.8')\n\n variant('debug', default=False, description='Builds a debug version of the library.')\n variant('shared', default=True, description='Builds a shared version of the library.')\n\n variant('fortran', default=True, description='Enable Fortran support.')\n variant('mpi', default=True, description='Enable MPI support.')\n variant('parmetis', default=False, description='Enable ParMETIS support.')\n variant('int64', default=False, description='Enable 64bit indices.')\n\n depends_on('mpi', when='+mpi')\n\n depends_on('parmetis@4: +int64', when='+parmetis+int64')\n depends_on('parmetis@4:', when='+parmetis')\n depends_on('metis+int64', when='+parmetis+int64')\n depends_on('metis', when='+parmetis')\n\n depends_on('perl@:5.21', type='build', when='@:3.6')\n depends_on('autoconf', type='build')\n depends_on('automake', type='build')\n depends_on('m4', type='build')\n\n conflicts('+parmetis', when='~mpi')\n\n build_directory = 'spack-build'\n\n @property\n def configure_directory(self):\n spec = self.spec\n\n # FIXME: The older Zoltan versions fail to compile the F90 MPI wrappers\n # because of some complicated generic type problem.\n if spec.satisfies('@:3.6+fortran+mpi'):\n raise RuntimeError(('Cannot build Zoltan v{0} with +fortran and '\n '+mpi; please disable one of these features '\n 'or upgrade versions.').format(self.version))\n if spec.satisfies('@:3.6'):\n zoltan_path = 'Zoltan_v{0}'.format(self.version)\n return zoltan_path\n return '.'\n\n @property\n def parallel(self):\n # NOTE: Earlier versions of Zoltan cannot be built in parallel\n # because they contain nested Makefile dependency bugs.\n return not self.spec.satisfies('@:3.6+fortran')\n\n def autoreconf(self, spec, prefix):\n autoreconf = which('autoreconf')\n with working_dir(self.configure_directory):\n autoreconf('-ivf')\n\n def configure_args(self):\n spec = self.spec\n\n config_args = [\n self.get_config_flag('f90interface', 'fortran'),\n self.get_config_flag('mpi', 'mpi'),\n ]\n config_cflags = [\n '-O0' if '+debug' in spec else '-O3',\n '-g' if '+debug' in spec else '',\n ]\n\n config_ldflags = []\n # PGI runtime libraries\n if '%pgi' in spec:\n config_ldflags.append('-pgf90libs')\n if '+shared' in spec:\n config_args.extend([\n 'RANLIB=echo',\n '--with-ar=$(CXX) -shared $(LDFLAGS) -o'\n ])\n config_cflags.append(self.compiler.cc_pic_flag)\n if spec.satisfies('%gcc'):\n config_args.append('--with-libs=-lgfortran')\n if spec.satisfies('%intel'):\n config_args.append('--with-libs=-lifcore')\n\n if '+int64' in spec:\n config_args.append('--with-id-type=ulong')\n\n if '+parmetis' in spec:\n parmetis_prefix = spec['parmetis'].prefix\n config_args.extend([\n '--with-parmetis',\n '--with-parmetis-libdir={0}'.format(parmetis_prefix.lib),\n '--with-parmetis-incdir={0}'.format(parmetis_prefix.include),\n '--with-incdirs=-I{0}'.format(spec['metis'].prefix.include),\n '--with-ldflags=-L{0}'.format(spec['metis'].prefix.lib)\n ])\n if '+int64' in spec['metis']:\n config_args.append('--with-id-type=ulong')\n else:\n config_args.append('--with-id-type=uint')\n\n if '+mpi' in spec:\n config_args.extend([\n 'CC={0}'.format(spec['mpi'].mpicc),\n 'CXX={0}'.format(spec['mpi'].mpicxx),\n 'FC={0}'.format(spec['mpi'].mpifc),\n '--with-mpi={0}'.format(spec['mpi'].prefix),\n\n # NOTE: Zoltan assumes that it's linking against an MPI library\n # that can be found with '-lmpi' which isn't the case for many\n # MPI packages. We rely on the MPI-wrappers to automatically\n # add what is required for linking and thus pass an empty\n # list of libs\n '--with-mpi-libs= '\n ])\n\n config_fcflags = config_cflags[:]\n if spec.satisfies('%gcc@10:+fortran'):\n config_fcflags.append('-fallow-argument-mismatch')\n # NOTE: Early versions of Zoltan come packaged with a few embedded\n # library packages (e.g. ParMETIS, Scotch), which messes with Spack's\n # ability to descend directly into the package's source directory.\n config_args.extend([\n '--with-cflags={0}'.format(' '.join(config_cflags)),\n '--with-cxxflags={0}'.format(' '.join(config_cflags)),\n '--with-fcflags={0}'.format(' '.join(config_fcflags)),\n '--with-ldflags={0}'.format(' '.join(config_ldflags))\n ])\n return config_args\n\n # NOTE: Unfortunately, Zoltan doesn't provide any configuration\n # options for the extension of the output library files, so this\n # script must change these extensions as a post-processing step.\n @run_after('install')\n def solib_install(self):\n if '+shared' in self.spec:\n for lib_path in find(self.spec.prefix.lib, 'lib*.a'):\n lib_shared_name = re.sub(r'\\.a$', '.{0}'.format(dso_suffix),\n lib_path)\n move(lib_path, lib_shared_name)\n\n def get_config_flag(self, flag_name, flag_variant):\n flag_pre = 'en' if '+{0}'.format(flag_variant) in self.spec else 'dis'\n return '--{0}able-{1}'.format(flag_pre, flag_name)\n", "path": "var/spack/repos/builtin/packages/zoltan/package.py"}]} | 3,014 | 152 |
gh_patches_debug_16696 | rasdani/github-patches | git_diff | mlcommons__GaNDLF-857 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
[FEATURE] Upgrade PyTorch to 2.2.1
### Is your feature request related to a problem? Please describe.
PyTorch 2.2 has been released since 30th Jan, and it would be good to update the dependency to reflect this. Full release notes are [here](https://github.com/pytorch/pytorch/releases/tag/v2.2.0).
### Describe the solution you'd like
Update the requirements.
### Describe alternatives you've considered
N.A.
### Additional context
N.A.
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `setup.py`
Content:
```
1 #!/usr/bin/env python
2
3 """The setup script."""
4
5
6 import sys, re, os
7 from setuptools import setup, find_packages
8 from setuptools.command.install import install
9 from setuptools.command.develop import develop
10 from setuptools.command.egg_info import egg_info
11
12 try:
13 with open("README.md") as readme_file:
14 readme = readme_file.read()
15 except Exception as error:
16 readme = "No README information found."
17 sys.stderr.write("Warning: Could not open '%s' due %s\n" % ("README.md", error))
18
19
20 class CustomInstallCommand(install):
21 def run(self):
22 install.run(self)
23
24
25 class CustomDevelopCommand(develop):
26 def run(self):
27 develop.run(self)
28
29
30 class CustomEggInfoCommand(egg_info):
31 def run(self):
32 egg_info.run(self)
33
34
35 try:
36 filepath = "GANDLF/version.py"
37 version_file = open(filepath)
38 (__version__,) = re.findall('__version__ = "(.*)"', version_file.read())
39
40 except Exception as error:
41 __version__ = "0.0.1"
42 sys.stderr.write("Warning: Could not open '%s' due %s\n" % (filepath, error))
43
44 # Handle cases where specific files need to be bundled into the final package as installed via PyPI
45 dockerfiles = [
46 item
47 for item in os.listdir(os.path.dirname(os.path.abspath(__file__)))
48 if (os.path.isfile(item) and item.startswith("Dockerfile-"))
49 ]
50 setup_files = ["setup.py", ".dockerignore", "pyproject.toml", "MANIFEST.in"]
51 all_extra_files = dockerfiles + setup_files
52 all_extra_files_pathcorrected = [os.path.join("../", item) for item in all_extra_files]
53 # find_packages should only ever find these as subpackages of gandlf, not as top-level packages
54 # generate this dynamically?
55 # GANDLF.GANDLF is needed to prevent recursion madness in deployments
56 toplevel_package_excludes = [
57 "GANDLF.GANDLF",
58 "anonymize",
59 "cli",
60 "compute",
61 "data",
62 "grad_clipping",
63 "losses",
64 "metrics",
65 "models",
66 "optimizers",
67 "schedulers",
68 "utils",
69 ]
70
71 # specifying version for `black` separately because it is also used to [check for lint](https://github.com/mlcommons/GaNDLF/blob/master/.github/workflows/black.yml)
72 black_version = "23.11.0"
73 requirements = [
74 "torch==2.1.2",
75 f"black=={black_version}",
76 "numpy==1.25.0",
77 "scipy",
78 "SimpleITK!=2.0.*",
79 "SimpleITK!=2.2.1", # https://github.com/mlcommons/GaNDLF/issues/536
80 "torchvision",
81 "tqdm",
82 "torchio==0.19.5",
83 "pandas>=2.0.0",
84 "scikit-learn>=0.23.2",
85 "scikit-image>=0.19.1",
86 "setuptools",
87 "seaborn",
88 "pyyaml",
89 "tiffslide",
90 "matplotlib",
91 "gdown==5.1.0",
92 "pytest",
93 "coverage",
94 "pytest-cov",
95 "psutil",
96 "medcam",
97 "opencv-python",
98 "torchmetrics==1.1.2",
99 "zarr==2.10.3",
100 "pydicom",
101 "onnx",
102 "torchinfo==1.7.0",
103 "segmentation-models-pytorch==0.3.3",
104 "ACSConv==0.1.1",
105 "docker",
106 "dicom-anonymizer==1.0.12",
107 "twine",
108 "zarr",
109 "keyring",
110 "monai==1.3.0",
111 "click>=8.0.0",
112 "deprecated",
113 "packaging==24.0",
114 "typer==0.9.0",
115 ]
116
117 if __name__ == "__main__":
118 setup(
119 name="GANDLF",
120 version=__version__,
121 author="MLCommons",
122 author_email="[email protected]",
123 python_requires=">3.8, <3.12",
124 packages=find_packages(
125 where=os.path.dirname(os.path.abspath(__file__)),
126 exclude=toplevel_package_excludes,
127 ),
128 cmdclass={
129 "install": CustomInstallCommand,
130 "develop": CustomDevelopCommand,
131 "egg_info": CustomEggInfoCommand,
132 },
133 entry_points={
134 "console_scripts": [
135 "gandlf=GANDLF.entrypoints.cli_tool:gandlf",
136 # old entrypoints
137 "gandlf_run=GANDLF.entrypoints.run:old_way",
138 "gandlf_constructCSV=GANDLF.entrypoints.construct_csv:old_way",
139 "gandlf_collectStats=GANDLF.entrypoints.collect_stats:old_way",
140 "gandlf_patchMiner=GANDLF.entrypoints.patch_miner:old_way",
141 "gandlf_preprocess=GANDLF.entrypoints.preprocess:old_way",
142 "gandlf_anonymizer=GANDLF.entrypoints.anonymizer:old_way",
143 "gandlf_configGenerator=GANDLF.entrypoints.config_generator:old_way",
144 "gandlf_verifyInstall=GANDLF.entrypoints.verify_install:old_way",
145 "gandlf_recoverConfig=GANDLF.entrypoints.recover_config:old_way",
146 "gandlf_deploy=GANDLF.entrypoints.deploy:old_way",
147 "gandlf_optimizeModel=GANDLF.entrypoints.optimize_model:old_way",
148 "gandlf_generateMetrics=GANDLF.entrypoints.generate_metrics:old_way",
149 "gandlf_debugInfo=GANDLF.entrypoints.debug_info:old_way",
150 "gandlf_splitCSV=GANDLF.entrypoints.split_csv:old_way",
151 ]
152 },
153 classifiers=[
154 "Development Status :: 3 - Alpha",
155 "Intended Audience :: Science/Research",
156 "License :: OSI Approved :: Apache Software License",
157 "Natural Language :: English",
158 "Operating System :: OS Independent",
159 "Programming Language :: Python :: 3.9",
160 "Programming Language :: Python :: 3.10",
161 "Programming Language :: Python :: 3.11",
162 "Topic :: Scientific/Engineering :: Medical Science Apps.",
163 ],
164 description=(
165 "PyTorch-based framework that handles segmentation/regression/classification using various DL architectures for medical imaging."
166 ),
167 install_requires=requirements,
168 license="Apache-2.0",
169 long_description=readme,
170 long_description_content_type="text/markdown",
171 include_package_data=True,
172 package_data={"GANDLF": all_extra_files_pathcorrected},
173 keywords="semantic, segmentation, regression, classification, data-augmentation, medical-imaging, clinical-workflows, deep-learning, pytorch",
174 zip_safe=False,
175 )
176
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/setup.py b/setup.py
--- a/setup.py
+++ b/setup.py
@@ -71,7 +71,7 @@
# specifying version for `black` separately because it is also used to [check for lint](https://github.com/mlcommons/GaNDLF/blob/master/.github/workflows/black.yml)
black_version = "23.11.0"
requirements = [
- "torch==2.1.2",
+ "torch==2.2.1",
f"black=={black_version}",
"numpy==1.25.0",
"scipy",
@@ -79,7 +79,7 @@
"SimpleITK!=2.2.1", # https://github.com/mlcommons/GaNDLF/issues/536
"torchvision",
"tqdm",
- "torchio==0.19.5",
+ "torchio==0.19.6",
"pandas>=2.0.0",
"scikit-learn>=0.23.2",
"scikit-image>=0.19.1",
| {"golden_diff": "diff --git a/setup.py b/setup.py\n--- a/setup.py\n+++ b/setup.py\n@@ -71,7 +71,7 @@\n # specifying version for `black` separately because it is also used to [check for lint](https://github.com/mlcommons/GaNDLF/blob/master/.github/workflows/black.yml)\n black_version = \"23.11.0\"\n requirements = [\n- \"torch==2.1.2\",\n+ \"torch==2.2.1\",\n f\"black=={black_version}\",\n \"numpy==1.25.0\",\n \"scipy\",\n@@ -79,7 +79,7 @@\n \"SimpleITK!=2.2.1\", # https://github.com/mlcommons/GaNDLF/issues/536\n \"torchvision\",\n \"tqdm\",\n- \"torchio==0.19.5\",\n+ \"torchio==0.19.6\",\n \"pandas>=2.0.0\",\n \"scikit-learn>=0.23.2\",\n \"scikit-image>=0.19.1\",\n", "issue": "[FEATURE] Upgrade PyTorch to 2.2.1\n### Is your feature request related to a problem? Please describe.\r\nPyTorch 2.2 has been released since 30th Jan, and it would be good to update the dependency to reflect this. Full release notes are [here](https://github.com/pytorch/pytorch/releases/tag/v2.2.0).\r\n\r\n### Describe the solution you'd like\r\nUpdate the requirements.\r\n\r\n### Describe alternatives you've considered\r\nN.A.\r\n\r\n### Additional context\r\nN.A.\r\n\n", "before_files": [{"content": "#!/usr/bin/env python\n\n\"\"\"The setup script.\"\"\"\n\n\nimport sys, re, os\nfrom setuptools import setup, find_packages\nfrom setuptools.command.install import install\nfrom setuptools.command.develop import develop\nfrom setuptools.command.egg_info import egg_info\n\ntry:\n with open(\"README.md\") as readme_file:\n readme = readme_file.read()\nexcept Exception as error:\n readme = \"No README information found.\"\n sys.stderr.write(\"Warning: Could not open '%s' due %s\\n\" % (\"README.md\", error))\n\n\nclass CustomInstallCommand(install):\n def run(self):\n install.run(self)\n\n\nclass CustomDevelopCommand(develop):\n def run(self):\n develop.run(self)\n\n\nclass CustomEggInfoCommand(egg_info):\n def run(self):\n egg_info.run(self)\n\n\ntry:\n filepath = \"GANDLF/version.py\"\n version_file = open(filepath)\n (__version__,) = re.findall('__version__ = \"(.*)\"', version_file.read())\n\nexcept Exception as error:\n __version__ = \"0.0.1\"\n sys.stderr.write(\"Warning: Could not open '%s' due %s\\n\" % (filepath, error))\n\n# Handle cases where specific files need to be bundled into the final package as installed via PyPI\ndockerfiles = [\n item\n for item in os.listdir(os.path.dirname(os.path.abspath(__file__)))\n if (os.path.isfile(item) and item.startswith(\"Dockerfile-\"))\n]\nsetup_files = [\"setup.py\", \".dockerignore\", \"pyproject.toml\", \"MANIFEST.in\"]\nall_extra_files = dockerfiles + setup_files\nall_extra_files_pathcorrected = [os.path.join(\"../\", item) for item in all_extra_files]\n# find_packages should only ever find these as subpackages of gandlf, not as top-level packages\n# generate this dynamically?\n# GANDLF.GANDLF is needed to prevent recursion madness in deployments\ntoplevel_package_excludes = [\n \"GANDLF.GANDLF\",\n \"anonymize\",\n \"cli\",\n \"compute\",\n \"data\",\n \"grad_clipping\",\n \"losses\",\n \"metrics\",\n \"models\",\n \"optimizers\",\n \"schedulers\",\n \"utils\",\n]\n\n# specifying version for `black` separately because it is also used to [check for lint](https://github.com/mlcommons/GaNDLF/blob/master/.github/workflows/black.yml)\nblack_version = \"23.11.0\"\nrequirements = [\n \"torch==2.1.2\",\n f\"black=={black_version}\",\n \"numpy==1.25.0\",\n \"scipy\",\n \"SimpleITK!=2.0.*\",\n \"SimpleITK!=2.2.1\", # https://github.com/mlcommons/GaNDLF/issues/536\n \"torchvision\",\n \"tqdm\",\n \"torchio==0.19.5\",\n \"pandas>=2.0.0\",\n \"scikit-learn>=0.23.2\",\n \"scikit-image>=0.19.1\",\n \"setuptools\",\n \"seaborn\",\n \"pyyaml\",\n \"tiffslide\",\n \"matplotlib\",\n \"gdown==5.1.0\",\n \"pytest\",\n \"coverage\",\n \"pytest-cov\",\n \"psutil\",\n \"medcam\",\n \"opencv-python\",\n \"torchmetrics==1.1.2\",\n \"zarr==2.10.3\",\n \"pydicom\",\n \"onnx\",\n \"torchinfo==1.7.0\",\n \"segmentation-models-pytorch==0.3.3\",\n \"ACSConv==0.1.1\",\n \"docker\",\n \"dicom-anonymizer==1.0.12\",\n \"twine\",\n \"zarr\",\n \"keyring\",\n \"monai==1.3.0\",\n \"click>=8.0.0\",\n \"deprecated\",\n \"packaging==24.0\",\n \"typer==0.9.0\",\n]\n\nif __name__ == \"__main__\":\n setup(\n name=\"GANDLF\",\n version=__version__,\n author=\"MLCommons\",\n author_email=\"[email protected]\",\n python_requires=\">3.8, <3.12\",\n packages=find_packages(\n where=os.path.dirname(os.path.abspath(__file__)),\n exclude=toplevel_package_excludes,\n ),\n cmdclass={\n \"install\": CustomInstallCommand,\n \"develop\": CustomDevelopCommand,\n \"egg_info\": CustomEggInfoCommand,\n },\n entry_points={\n \"console_scripts\": [\n \"gandlf=GANDLF.entrypoints.cli_tool:gandlf\",\n # old entrypoints\n \"gandlf_run=GANDLF.entrypoints.run:old_way\",\n \"gandlf_constructCSV=GANDLF.entrypoints.construct_csv:old_way\",\n \"gandlf_collectStats=GANDLF.entrypoints.collect_stats:old_way\",\n \"gandlf_patchMiner=GANDLF.entrypoints.patch_miner:old_way\",\n \"gandlf_preprocess=GANDLF.entrypoints.preprocess:old_way\",\n \"gandlf_anonymizer=GANDLF.entrypoints.anonymizer:old_way\",\n \"gandlf_configGenerator=GANDLF.entrypoints.config_generator:old_way\",\n \"gandlf_verifyInstall=GANDLF.entrypoints.verify_install:old_way\",\n \"gandlf_recoverConfig=GANDLF.entrypoints.recover_config:old_way\",\n \"gandlf_deploy=GANDLF.entrypoints.deploy:old_way\",\n \"gandlf_optimizeModel=GANDLF.entrypoints.optimize_model:old_way\",\n \"gandlf_generateMetrics=GANDLF.entrypoints.generate_metrics:old_way\",\n \"gandlf_debugInfo=GANDLF.entrypoints.debug_info:old_way\",\n \"gandlf_splitCSV=GANDLF.entrypoints.split_csv:old_way\",\n ]\n },\n classifiers=[\n \"Development Status :: 3 - Alpha\",\n \"Intended Audience :: Science/Research\",\n \"License :: OSI Approved :: Apache Software License\",\n \"Natural Language :: English\",\n \"Operating System :: OS Independent\",\n \"Programming Language :: Python :: 3.9\",\n \"Programming Language :: Python :: 3.10\",\n \"Programming Language :: Python :: 3.11\",\n \"Topic :: Scientific/Engineering :: Medical Science Apps.\",\n ],\n description=(\n \"PyTorch-based framework that handles segmentation/regression/classification using various DL architectures for medical imaging.\"\n ),\n install_requires=requirements,\n license=\"Apache-2.0\",\n long_description=readme,\n long_description_content_type=\"text/markdown\",\n include_package_data=True,\n package_data={\"GANDLF\": all_extra_files_pathcorrected},\n keywords=\"semantic, segmentation, regression, classification, data-augmentation, medical-imaging, clinical-workflows, deep-learning, pytorch\",\n zip_safe=False,\n )\n", "path": "setup.py"}], "after_files": [{"content": "#!/usr/bin/env python\n\n\"\"\"The setup script.\"\"\"\n\n\nimport sys, re, os\nfrom setuptools import setup, find_packages\nfrom setuptools.command.install import install\nfrom setuptools.command.develop import develop\nfrom setuptools.command.egg_info import egg_info\n\ntry:\n with open(\"README.md\") as readme_file:\n readme = readme_file.read()\nexcept Exception as error:\n readme = \"No README information found.\"\n sys.stderr.write(\"Warning: Could not open '%s' due %s\\n\" % (\"README.md\", error))\n\n\nclass CustomInstallCommand(install):\n def run(self):\n install.run(self)\n\n\nclass CustomDevelopCommand(develop):\n def run(self):\n develop.run(self)\n\n\nclass CustomEggInfoCommand(egg_info):\n def run(self):\n egg_info.run(self)\n\n\ntry:\n filepath = \"GANDLF/version.py\"\n version_file = open(filepath)\n (__version__,) = re.findall('__version__ = \"(.*)\"', version_file.read())\n\nexcept Exception as error:\n __version__ = \"0.0.1\"\n sys.stderr.write(\"Warning: Could not open '%s' due %s\\n\" % (filepath, error))\n\n# Handle cases where specific files need to be bundled into the final package as installed via PyPI\ndockerfiles = [\n item\n for item in os.listdir(os.path.dirname(os.path.abspath(__file__)))\n if (os.path.isfile(item) and item.startswith(\"Dockerfile-\"))\n]\nsetup_files = [\"setup.py\", \".dockerignore\", \"pyproject.toml\", \"MANIFEST.in\"]\nall_extra_files = dockerfiles + setup_files\nall_extra_files_pathcorrected = [os.path.join(\"../\", item) for item in all_extra_files]\n# find_packages should only ever find these as subpackages of gandlf, not as top-level packages\n# generate this dynamically?\n# GANDLF.GANDLF is needed to prevent recursion madness in deployments\ntoplevel_package_excludes = [\n \"GANDLF.GANDLF\",\n \"anonymize\",\n \"cli\",\n \"compute\",\n \"data\",\n \"grad_clipping\",\n \"losses\",\n \"metrics\",\n \"models\",\n \"optimizers\",\n \"schedulers\",\n \"utils\",\n]\n\n# specifying version for `black` separately because it is also used to [check for lint](https://github.com/mlcommons/GaNDLF/blob/master/.github/workflows/black.yml)\nblack_version = \"23.11.0\"\nrequirements = [\n \"torch==2.2.1\",\n f\"black=={black_version}\",\n \"numpy==1.25.0\",\n \"scipy\",\n \"SimpleITK!=2.0.*\",\n \"SimpleITK!=2.2.1\", # https://github.com/mlcommons/GaNDLF/issues/536\n \"torchvision\",\n \"tqdm\",\n \"torchio==0.19.6\",\n \"pandas>=2.0.0\",\n \"scikit-learn>=0.23.2\",\n \"scikit-image>=0.19.1\",\n \"setuptools\",\n \"seaborn\",\n \"pyyaml\",\n \"tiffslide\",\n \"matplotlib\",\n \"gdown==5.1.0\",\n \"pytest\",\n \"coverage\",\n \"pytest-cov\",\n \"psutil\",\n \"medcam\",\n \"opencv-python\",\n \"torchmetrics==1.1.2\",\n \"zarr==2.10.3\",\n \"pydicom\",\n \"onnx\",\n \"torchinfo==1.7.0\",\n \"segmentation-models-pytorch==0.3.3\",\n \"ACSConv==0.1.1\",\n \"docker\",\n \"dicom-anonymizer==1.0.12\",\n \"twine\",\n \"zarr\",\n \"keyring\",\n \"monai==1.3.0\",\n \"click>=8.0.0\",\n \"deprecated\",\n \"packaging==24.0\",\n \"typer==0.9.0\",\n]\n\nif __name__ == \"__main__\":\n setup(\n name=\"GANDLF\",\n version=__version__,\n author=\"MLCommons\",\n author_email=\"[email protected]\",\n python_requires=\">3.8, <3.12\",\n packages=find_packages(\n where=os.path.dirname(os.path.abspath(__file__)),\n exclude=toplevel_package_excludes,\n ),\n cmdclass={\n \"install\": CustomInstallCommand,\n \"develop\": CustomDevelopCommand,\n \"egg_info\": CustomEggInfoCommand,\n },\n entry_points={\n \"console_scripts\": [\n \"gandlf=GANDLF.entrypoints.cli_tool:gandlf\",\n # old entrypoints\n \"gandlf_run=GANDLF.entrypoints.run:old_way\",\n \"gandlf_constructCSV=GANDLF.entrypoints.construct_csv:old_way\",\n \"gandlf_collectStats=GANDLF.entrypoints.collect_stats:old_way\",\n \"gandlf_patchMiner=GANDLF.entrypoints.patch_miner:old_way\",\n \"gandlf_preprocess=GANDLF.entrypoints.preprocess:old_way\",\n \"gandlf_anonymizer=GANDLF.entrypoints.anonymizer:old_way\",\n \"gandlf_configGenerator=GANDLF.entrypoints.config_generator:old_way\",\n \"gandlf_verifyInstall=GANDLF.entrypoints.verify_install:old_way\",\n \"gandlf_recoverConfig=GANDLF.entrypoints.recover_config:old_way\",\n \"gandlf_deploy=GANDLF.entrypoints.deploy:old_way\",\n \"gandlf_optimizeModel=GANDLF.entrypoints.optimize_model:old_way\",\n \"gandlf_generateMetrics=GANDLF.entrypoints.generate_metrics:old_way\",\n \"gandlf_debugInfo=GANDLF.entrypoints.debug_info:old_way\",\n \"gandlf_splitCSV=GANDLF.entrypoints.split_csv:old_way\",\n ]\n },\n classifiers=[\n \"Development Status :: 3 - Alpha\",\n \"Intended Audience :: Science/Research\",\n \"License :: OSI Approved :: Apache Software License\",\n \"Natural Language :: English\",\n \"Operating System :: OS Independent\",\n \"Programming Language :: Python :: 3.9\",\n \"Programming Language :: Python :: 3.10\",\n \"Programming Language :: Python :: 3.11\",\n \"Topic :: Scientific/Engineering :: Medical Science Apps.\",\n ],\n description=(\n \"PyTorch-based framework that handles segmentation/regression/classification using various DL architectures for medical imaging.\"\n ),\n install_requires=requirements,\n license=\"Apache-2.0\",\n long_description=readme,\n long_description_content_type=\"text/markdown\",\n include_package_data=True,\n package_data={\"GANDLF\": all_extra_files_pathcorrected},\n keywords=\"semantic, segmentation, regression, classification, data-augmentation, medical-imaging, clinical-workflows, deep-learning, pytorch\",\n zip_safe=False,\n )\n", "path": "setup.py"}]} | 2,311 | 249 |
gh_patches_debug_35797 | rasdani/github-patches | git_diff | mathesar-foundation__mathesar-691 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Table, Database, & Schema APIs should support sorting by ID & name
## Problem
<!-- Please provide a clear and concise description of the problem that this feature request is designed to solve.-->
The table, database, and schema APIs don't currently support sorting results.
## Proposed solution
<!-- A clear and concise description of your proposed solution or feature. -->
All three APIs should support sorting by ID and name.
## Additional context
<!-- Add any other context or screenshots about the feature request here.-->
We can use `django-filter` / `django-property-filter` for this.
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `mathesar/api/filters.py`
Content:
```
1 from django_filters import BooleanFilter, DateTimeFromToRangeFilter
2 from django_property_filter import PropertyFilterSet, PropertyBaseInFilter, PropertyCharFilter
3
4 from mathesar.database.types import MathesarTypeIdentifier
5 from mathesar.models import Schema, Table, Database
6
7 FILTER_OPTIONS_BY_TYPE_IDENTIFIER = {
8 MathesarTypeIdentifier.BOOLEAN.value:
9 {
10 "db_type": "BOOLEAN",
11 "options": [{
12 "op": "eq",
13 "value": {
14 "allowed_types": ["BOOLEAN"],
15 }
16 }, {
17 "op": "is_null",
18 "value": "null",
19 }]
20 }
21 }
22
23
24 class CharInFilter(PropertyBaseInFilter, PropertyCharFilter):
25 pass
26
27
28 class SchemaFilter(PropertyFilterSet):
29 database = CharInFilter(field_name='database__name', lookup_expr='in')
30 name = CharInFilter(field_name='name', lookup_expr='in')
31
32 class Meta:
33 model = Schema
34 fields = ['name']
35
36
37 class TableFilter(PropertyFilterSet):
38 name = CharInFilter(field_name='name', lookup_expr='in')
39 created = DateTimeFromToRangeFilter(field_name='created_at')
40 updated = DateTimeFromToRangeFilter(field_name='updated_at')
41 not_imported = BooleanFilter(lookup_expr="isnull", field_name='import_verified')
42
43 class Meta:
44 model = Table
45 fields = ['name', 'schema', 'created_at', 'updated_at', 'import_verified']
46
47
48 class DatabaseFilter(PropertyFilterSet):
49 class Meta:
50 model = Database
51 fields = ['deleted']
52
```
Path: `config/settings.py`
Content:
```
1 """
2 Django settings for config project.
3
4 Generated by 'django-admin startproject' using Django 3.1.7.
5
6 For more information on this file, see
7 https://docs.djangoproject.com/en/3.1/topics/settings/
8
9 For the full list of settings and their values, see
10 https://docs.djangoproject.com/en/3.1/ref/settings/
11 """
12
13 import os
14 from pathlib import Path
15
16 from decouple import Csv, config as decouple_config
17 from dj_database_url import parse as db_url
18
19
20 # We use a 'tuple' with pipes as delimiters as decople naively splits the global
21 # variables on commas when casting to Csv()
22 def pipe_delim(pipe_string):
23 # Remove opening and closing brackets
24 pipe_string = pipe_string[1:-1]
25 # Split on pipe delim
26 return pipe_string.split("|")
27
28
29 # Build paths inside the project like this: BASE_DIR / 'subdir'.
30 BASE_DIR = Path(__file__).resolve().parent.parent
31
32 # Application definition
33
34 INSTALLED_APPS = [
35 "django.contrib.admin",
36 "django.contrib.auth",
37 "django.contrib.contenttypes",
38 "django.contrib.sessions",
39 "django.contrib.messages",
40 "django.contrib.staticfiles",
41 "rest_framework",
42 "django_filters",
43 "django_property_filter",
44 "mathesar",
45 ]
46
47 MIDDLEWARE = [
48 "django.middleware.security.SecurityMiddleware",
49 "django.contrib.sessions.middleware.SessionMiddleware",
50 "django.middleware.common.CommonMiddleware",
51 "django.middleware.csrf.CsrfViewMiddleware",
52 "django.contrib.auth.middleware.AuthenticationMiddleware",
53 "django.contrib.messages.middleware.MessageMiddleware",
54 "django.middleware.clickjacking.XFrameOptionsMiddleware",
55 ]
56
57 ROOT_URLCONF = "config.urls"
58
59 TEMPLATES = [
60 {
61 "BACKEND": "django.template.backends.django.DjangoTemplates",
62 "DIRS": [],
63 "APP_DIRS": True,
64 "OPTIONS": {
65 "context_processors": [
66 "config.context_processors.frontend_settings",
67 "django.template.context_processors.debug",
68 "django.template.context_processors.request",
69 "django.contrib.auth.context_processors.auth",
70 "django.contrib.messages.context_processors.messages",
71 ],
72 },
73 },
74 ]
75
76 WSGI_APPLICATION = "config.wsgi.application"
77
78 # Database
79 # https://docs.djangoproject.com/en/3.1/ref/settings/#databases
80
81 # TODO: Add to documentation that database keys should not be than 128 characters.
82
83 # MATHESAR_DATABASES should be of the form '({db_name}|{db_url}), ({db_name}|{db_url})'
84 # See pipe_delim above for why we use pipes as delimiters
85 DATABASES = {
86 db_key: db_url(url_string)
87 for db_key, url_string in decouple_config('MATHESAR_DATABASES', cast=Csv(pipe_delim))
88 }
89 DATABASES[decouple_config('DJANGO_DATABASE_KEY')] = decouple_config('DJANGO_DATABASE_URL', cast=db_url)
90
91 for db_key, db_dict in DATABASES.items():
92 # Engine can be '.postgresql' or '.postgresql_psycopg2'
93 if not db_dict['ENGINE'].startswith('django.db.backends.postgresql'):
94 raise ValueError(
95 f"{db_key} is not a PostgreSQL database. "
96 f"{db_dict['ENGINE']} found for {db_key}'s engine."
97 )
98
99
100 # pytest-django will create a new database named 'test_{DATABASES[table_db]['NAME']}'
101 # and use it for our API tests if we don't specify DATABASES[table_db]['TEST']['NAME']
102 if decouple_config('TEST', default=False, cast=bool):
103 for db_key, _ in decouple_config('MATHESAR_DATABASES', cast=Csv(pipe_delim)):
104 DATABASES[db_key]['TEST'] = {'NAME': DATABASES[db_key]['NAME']}
105
106
107 # Quick-start development settings - unsuitable for production
108 # See https://docs.djangoproject.com/en/3.1/howto/deployment/checklist/
109
110 # SECURITY WARNING: keep the secret key used in production secret!
111 SECRET_KEY = decouple_config('SECRET_KEY')
112
113 # SECURITY WARNING: don't run with debug turned on in production!
114 DEBUG = decouple_config('DEBUG', default=False, cast=bool)
115
116 ALLOWED_HOSTS = decouple_config('ALLOWED_HOSTS', cast=Csv())
117
118 # Password validation
119 # https://docs.djangoproject.com/en/3.1/ref/settings/#auth-password-validators
120
121 AUTH_PASSWORD_VALIDATORS = [
122 {
123 "NAME": "django.contrib.auth.password_validation.UserAttributeSimilarityValidator",
124 },
125 {
126 "NAME": "django.contrib.auth.password_validation.MinimumLengthValidator",
127 },
128 {
129 "NAME": "django.contrib.auth.password_validation.CommonPasswordValidator",
130 },
131 {
132 "NAME": "django.contrib.auth.password_validation.NumericPasswordValidator",
133 },
134 ]
135
136
137 # Internationalization
138 # https://docs.djangoproject.com/en/3.1/topics/i18n/
139
140 LANGUAGE_CODE = "en-us"
141
142 TIME_ZONE = "UTC"
143
144 USE_I18N = True
145
146 USE_L10N = True
147
148 USE_TZ = True
149
150
151 # Static files (CSS, JavaScript, Images)
152 # https://docs.djangoproject.com/en/3.1/howto/static-files/
153 # https://docs.djangoproject.com/en/3.1/ref/contrib/staticfiles/
154
155 STATIC_URL = "/static/"
156
157 # When running with DEBUG=False, the webserver needs to serve files from this location
158 # python manage.py collectstatic has to be run to collect all static files into this location
159 # The files need to served in brotli or gzip compressed format
160 STATIC_ROOT = os.path.join(BASE_DIR, 'static/')
161
162 # Media files (uploaded by the user)
163
164 MEDIA_ROOT = os.path.join(BASE_DIR, '.media/')
165
166 MEDIA_URL = "/media/"
167
168 # Update Authentication classes, removed BasicAuthentication
169 # Defaults: https://www.django-rest-framework.org/api-guide/settings/
170 REST_FRAMEWORK = {
171 'DEFAULT_AUTHENTICATION_CLASSES': [
172 'rest_framework.authentication.SessionAuthentication'
173 ]
174 }
175
176 # Mathesar settings
177 MATHESAR_MODE = decouple_config('MODE', default='PRODUCTION')
178 MATHESAR_UI_BUILD_LOCATION = os.path.join(BASE_DIR, 'mathesar/static/mathesar/')
179 MATHESAR_MANIFEST_LOCATION = os.path.join(MATHESAR_UI_BUILD_LOCATION, 'manifest.json')
180 MATHESAR_CLIENT_DEV_URL = 'http://localhost:3000'
181
182
183 STATICFILES_DIRS = [MATHESAR_UI_BUILD_LOCATION]
184
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/config/settings.py b/config/settings.py
--- a/config/settings.py
+++ b/config/settings.py
@@ -170,7 +170,11 @@
REST_FRAMEWORK = {
'DEFAULT_AUTHENTICATION_CLASSES': [
'rest_framework.authentication.SessionAuthentication'
- ]
+ ],
+ 'DEFAULT_FILTER_BACKENDS': (
+ 'django_filters.rest_framework.DjangoFilterBackend',
+ 'rest_framework.filters.OrderingFilter',
+ ),
}
# Mathesar settings
diff --git a/mathesar/api/filters.py b/mathesar/api/filters.py
--- a/mathesar/api/filters.py
+++ b/mathesar/api/filters.py
@@ -1,5 +1,5 @@
-from django_filters import BooleanFilter, DateTimeFromToRangeFilter
-from django_property_filter import PropertyFilterSet, PropertyBaseInFilter, PropertyCharFilter
+from django_filters import BooleanFilter, DateTimeFromToRangeFilter, OrderingFilter
+from django_property_filter import PropertyFilterSet, PropertyBaseInFilter, PropertyCharFilter, PropertyOrderingFilter
from mathesar.database.types import MathesarTypeIdentifier
from mathesar.models import Schema, Table, Database
@@ -29,6 +29,14 @@
database = CharInFilter(field_name='database__name', lookup_expr='in')
name = CharInFilter(field_name='name', lookup_expr='in')
+ sort_by = PropertyOrderingFilter(
+ fields=(
+ ('id', 'id'),
+ ('name', 'name'),
+ ),
+ label="Sort By",
+ )
+
class Meta:
model = Schema
fields = ['name']
@@ -40,12 +48,28 @@
updated = DateTimeFromToRangeFilter(field_name='updated_at')
not_imported = BooleanFilter(lookup_expr="isnull", field_name='import_verified')
+ sort_by = PropertyOrderingFilter(
+ fields=(
+ ('id', 'id'),
+ ('name', 'name'),
+ ),
+ label="Sort By",
+ )
+
class Meta:
model = Table
fields = ['name', 'schema', 'created_at', 'updated_at', 'import_verified']
class DatabaseFilter(PropertyFilterSet):
+ sort_by = OrderingFilter(
+ fields=(
+ ('id', 'id'),
+ ('name', 'name'),
+ ),
+ label="Sort By",
+ )
+
class Meta:
model = Database
fields = ['deleted']
| {"golden_diff": "diff --git a/config/settings.py b/config/settings.py\n--- a/config/settings.py\n+++ b/config/settings.py\n@@ -170,7 +170,11 @@\n REST_FRAMEWORK = {\n 'DEFAULT_AUTHENTICATION_CLASSES': [\n 'rest_framework.authentication.SessionAuthentication'\n- ]\n+ ],\n+ 'DEFAULT_FILTER_BACKENDS': (\n+ 'django_filters.rest_framework.DjangoFilterBackend',\n+ 'rest_framework.filters.OrderingFilter',\n+ ),\n }\n \n # Mathesar settings\ndiff --git a/mathesar/api/filters.py b/mathesar/api/filters.py\n--- a/mathesar/api/filters.py\n+++ b/mathesar/api/filters.py\n@@ -1,5 +1,5 @@\n-from django_filters import BooleanFilter, DateTimeFromToRangeFilter\n-from django_property_filter import PropertyFilterSet, PropertyBaseInFilter, PropertyCharFilter\n+from django_filters import BooleanFilter, DateTimeFromToRangeFilter, OrderingFilter\n+from django_property_filter import PropertyFilterSet, PropertyBaseInFilter, PropertyCharFilter, PropertyOrderingFilter\n \n from mathesar.database.types import MathesarTypeIdentifier\n from mathesar.models import Schema, Table, Database\n@@ -29,6 +29,14 @@\n database = CharInFilter(field_name='database__name', lookup_expr='in')\n name = CharInFilter(field_name='name', lookup_expr='in')\n \n+ sort_by = PropertyOrderingFilter(\n+ fields=(\n+ ('id', 'id'),\n+ ('name', 'name'),\n+ ),\n+ label=\"Sort By\",\n+ )\n+\n class Meta:\n model = Schema\n fields = ['name']\n@@ -40,12 +48,28 @@\n updated = DateTimeFromToRangeFilter(field_name='updated_at')\n not_imported = BooleanFilter(lookup_expr=\"isnull\", field_name='import_verified')\n \n+ sort_by = PropertyOrderingFilter(\n+ fields=(\n+ ('id', 'id'),\n+ ('name', 'name'),\n+ ),\n+ label=\"Sort By\",\n+ )\n+\n class Meta:\n model = Table\n fields = ['name', 'schema', 'created_at', 'updated_at', 'import_verified']\n \n \n class DatabaseFilter(PropertyFilterSet):\n+ sort_by = OrderingFilter(\n+ fields=(\n+ ('id', 'id'),\n+ ('name', 'name'),\n+ ),\n+ label=\"Sort By\",\n+ )\n+\n class Meta:\n model = Database\n fields = ['deleted']\n", "issue": "Table, Database, & Schema APIs should support sorting by ID & name\n## Problem\r\n<!-- Please provide a clear and concise description of the problem that this feature request is designed to solve.-->\r\nThe table, database, and schema APIs don't currently support sorting results.\r\n\r\n## Proposed solution\r\n<!-- A clear and concise description of your proposed solution or feature. -->\r\nAll three APIs should support sorting by ID and name.\r\n\r\n## Additional context\r\n<!-- Add any other context or screenshots about the feature request here.-->\r\nWe can use `django-filter` / `django-property-filter` for this.\n", "before_files": [{"content": "from django_filters import BooleanFilter, DateTimeFromToRangeFilter\nfrom django_property_filter import PropertyFilterSet, PropertyBaseInFilter, PropertyCharFilter\n\nfrom mathesar.database.types import MathesarTypeIdentifier\nfrom mathesar.models import Schema, Table, Database\n\nFILTER_OPTIONS_BY_TYPE_IDENTIFIER = {\n MathesarTypeIdentifier.BOOLEAN.value:\n {\n \"db_type\": \"BOOLEAN\",\n \"options\": [{\n \"op\": \"eq\",\n \"value\": {\n \"allowed_types\": [\"BOOLEAN\"],\n }\n }, {\n \"op\": \"is_null\",\n \"value\": \"null\",\n }]\n }\n}\n\n\nclass CharInFilter(PropertyBaseInFilter, PropertyCharFilter):\n pass\n\n\nclass SchemaFilter(PropertyFilterSet):\n database = CharInFilter(field_name='database__name', lookup_expr='in')\n name = CharInFilter(field_name='name', lookup_expr='in')\n\n class Meta:\n model = Schema\n fields = ['name']\n\n\nclass TableFilter(PropertyFilterSet):\n name = CharInFilter(field_name='name', lookup_expr='in')\n created = DateTimeFromToRangeFilter(field_name='created_at')\n updated = DateTimeFromToRangeFilter(field_name='updated_at')\n not_imported = BooleanFilter(lookup_expr=\"isnull\", field_name='import_verified')\n\n class Meta:\n model = Table\n fields = ['name', 'schema', 'created_at', 'updated_at', 'import_verified']\n\n\nclass DatabaseFilter(PropertyFilterSet):\n class Meta:\n model = Database\n fields = ['deleted']\n", "path": "mathesar/api/filters.py"}, {"content": "\"\"\"\nDjango settings for config project.\n\nGenerated by 'django-admin startproject' using Django 3.1.7.\n\nFor more information on this file, see\nhttps://docs.djangoproject.com/en/3.1/topics/settings/\n\nFor the full list of settings and their values, see\nhttps://docs.djangoproject.com/en/3.1/ref/settings/\n\"\"\"\n\nimport os\nfrom pathlib import Path\n\nfrom decouple import Csv, config as decouple_config\nfrom dj_database_url import parse as db_url\n\n\n# We use a 'tuple' with pipes as delimiters as decople naively splits the global\n# variables on commas when casting to Csv()\ndef pipe_delim(pipe_string):\n # Remove opening and closing brackets\n pipe_string = pipe_string[1:-1]\n # Split on pipe delim\n return pipe_string.split(\"|\")\n\n\n# Build paths inside the project like this: BASE_DIR / 'subdir'.\nBASE_DIR = Path(__file__).resolve().parent.parent\n\n# Application definition\n\nINSTALLED_APPS = [\n \"django.contrib.admin\",\n \"django.contrib.auth\",\n \"django.contrib.contenttypes\",\n \"django.contrib.sessions\",\n \"django.contrib.messages\",\n \"django.contrib.staticfiles\",\n \"rest_framework\",\n \"django_filters\",\n \"django_property_filter\",\n \"mathesar\",\n]\n\nMIDDLEWARE = [\n \"django.middleware.security.SecurityMiddleware\",\n \"django.contrib.sessions.middleware.SessionMiddleware\",\n \"django.middleware.common.CommonMiddleware\",\n \"django.middleware.csrf.CsrfViewMiddleware\",\n \"django.contrib.auth.middleware.AuthenticationMiddleware\",\n \"django.contrib.messages.middleware.MessageMiddleware\",\n \"django.middleware.clickjacking.XFrameOptionsMiddleware\",\n]\n\nROOT_URLCONF = \"config.urls\"\n\nTEMPLATES = [\n {\n \"BACKEND\": \"django.template.backends.django.DjangoTemplates\",\n \"DIRS\": [],\n \"APP_DIRS\": True,\n \"OPTIONS\": {\n \"context_processors\": [\n \"config.context_processors.frontend_settings\",\n \"django.template.context_processors.debug\",\n \"django.template.context_processors.request\",\n \"django.contrib.auth.context_processors.auth\",\n \"django.contrib.messages.context_processors.messages\",\n ],\n },\n },\n]\n\nWSGI_APPLICATION = \"config.wsgi.application\"\n\n# Database\n# https://docs.djangoproject.com/en/3.1/ref/settings/#databases\n\n# TODO: Add to documentation that database keys should not be than 128 characters.\n\n# MATHESAR_DATABASES should be of the form '({db_name}|{db_url}), ({db_name}|{db_url})'\n# See pipe_delim above for why we use pipes as delimiters\nDATABASES = {\n db_key: db_url(url_string)\n for db_key, url_string in decouple_config('MATHESAR_DATABASES', cast=Csv(pipe_delim))\n}\nDATABASES[decouple_config('DJANGO_DATABASE_KEY')] = decouple_config('DJANGO_DATABASE_URL', cast=db_url)\n\nfor db_key, db_dict in DATABASES.items():\n # Engine can be '.postgresql' or '.postgresql_psycopg2'\n if not db_dict['ENGINE'].startswith('django.db.backends.postgresql'):\n raise ValueError(\n f\"{db_key} is not a PostgreSQL database. \"\n f\"{db_dict['ENGINE']} found for {db_key}'s engine.\"\n )\n\n\n# pytest-django will create a new database named 'test_{DATABASES[table_db]['NAME']}'\n# and use it for our API tests if we don't specify DATABASES[table_db]['TEST']['NAME']\nif decouple_config('TEST', default=False, cast=bool):\n for db_key, _ in decouple_config('MATHESAR_DATABASES', cast=Csv(pipe_delim)):\n DATABASES[db_key]['TEST'] = {'NAME': DATABASES[db_key]['NAME']}\n\n\n# Quick-start development settings - unsuitable for production\n# See https://docs.djangoproject.com/en/3.1/howto/deployment/checklist/\n\n# SECURITY WARNING: keep the secret key used in production secret!\nSECRET_KEY = decouple_config('SECRET_KEY')\n\n# SECURITY WARNING: don't run with debug turned on in production!\nDEBUG = decouple_config('DEBUG', default=False, cast=bool)\n\nALLOWED_HOSTS = decouple_config('ALLOWED_HOSTS', cast=Csv())\n\n# Password validation\n# https://docs.djangoproject.com/en/3.1/ref/settings/#auth-password-validators\n\nAUTH_PASSWORD_VALIDATORS = [\n {\n \"NAME\": \"django.contrib.auth.password_validation.UserAttributeSimilarityValidator\",\n },\n {\n \"NAME\": \"django.contrib.auth.password_validation.MinimumLengthValidator\",\n },\n {\n \"NAME\": \"django.contrib.auth.password_validation.CommonPasswordValidator\",\n },\n {\n \"NAME\": \"django.contrib.auth.password_validation.NumericPasswordValidator\",\n },\n]\n\n\n# Internationalization\n# https://docs.djangoproject.com/en/3.1/topics/i18n/\n\nLANGUAGE_CODE = \"en-us\"\n\nTIME_ZONE = \"UTC\"\n\nUSE_I18N = True\n\nUSE_L10N = True\n\nUSE_TZ = True\n\n\n# Static files (CSS, JavaScript, Images)\n# https://docs.djangoproject.com/en/3.1/howto/static-files/\n# https://docs.djangoproject.com/en/3.1/ref/contrib/staticfiles/\n\nSTATIC_URL = \"/static/\"\n\n# When running with DEBUG=False, the webserver needs to serve files from this location\n# python manage.py collectstatic has to be run to collect all static files into this location\n# The files need to served in brotli or gzip compressed format\nSTATIC_ROOT = os.path.join(BASE_DIR, 'static/')\n\n# Media files (uploaded by the user)\n\nMEDIA_ROOT = os.path.join(BASE_DIR, '.media/')\n\nMEDIA_URL = \"/media/\"\n\n# Update Authentication classes, removed BasicAuthentication\n# Defaults: https://www.django-rest-framework.org/api-guide/settings/\nREST_FRAMEWORK = {\n 'DEFAULT_AUTHENTICATION_CLASSES': [\n 'rest_framework.authentication.SessionAuthentication'\n ]\n}\n\n# Mathesar settings\nMATHESAR_MODE = decouple_config('MODE', default='PRODUCTION')\nMATHESAR_UI_BUILD_LOCATION = os.path.join(BASE_DIR, 'mathesar/static/mathesar/')\nMATHESAR_MANIFEST_LOCATION = os.path.join(MATHESAR_UI_BUILD_LOCATION, 'manifest.json')\nMATHESAR_CLIENT_DEV_URL = 'http://localhost:3000'\n\n\nSTATICFILES_DIRS = [MATHESAR_UI_BUILD_LOCATION]\n", "path": "config/settings.py"}], "after_files": [{"content": "from django_filters import BooleanFilter, DateTimeFromToRangeFilter, OrderingFilter\nfrom django_property_filter import PropertyFilterSet, PropertyBaseInFilter, PropertyCharFilter, PropertyOrderingFilter\n\nfrom mathesar.database.types import MathesarTypeIdentifier\nfrom mathesar.models import Schema, Table, Database\n\nFILTER_OPTIONS_BY_TYPE_IDENTIFIER = {\n MathesarTypeIdentifier.BOOLEAN.value:\n {\n \"db_type\": \"BOOLEAN\",\n \"options\": [{\n \"op\": \"eq\",\n \"value\": {\n \"allowed_types\": [\"BOOLEAN\"],\n }\n }, {\n \"op\": \"is_null\",\n \"value\": \"null\",\n }]\n }\n}\n\n\nclass CharInFilter(PropertyBaseInFilter, PropertyCharFilter):\n pass\n\n\nclass SchemaFilter(PropertyFilterSet):\n database = CharInFilter(field_name='database__name', lookup_expr='in')\n name = CharInFilter(field_name='name', lookup_expr='in')\n\n sort_by = PropertyOrderingFilter(\n fields=(\n ('id', 'id'),\n ('name', 'name'),\n ),\n label=\"Sort By\",\n )\n\n class Meta:\n model = Schema\n fields = ['name']\n\n\nclass TableFilter(PropertyFilterSet):\n name = CharInFilter(field_name='name', lookup_expr='in')\n created = DateTimeFromToRangeFilter(field_name='created_at')\n updated = DateTimeFromToRangeFilter(field_name='updated_at')\n not_imported = BooleanFilter(lookup_expr=\"isnull\", field_name='import_verified')\n\n sort_by = PropertyOrderingFilter(\n fields=(\n ('id', 'id'),\n ('name', 'name'),\n ),\n label=\"Sort By\",\n )\n\n class Meta:\n model = Table\n fields = ['name', 'schema', 'created_at', 'updated_at', 'import_verified']\n\n\nclass DatabaseFilter(PropertyFilterSet):\n sort_by = OrderingFilter(\n fields=(\n ('id', 'id'),\n ('name', 'name'),\n ),\n label=\"Sort By\",\n )\n\n class Meta:\n model = Database\n fields = ['deleted']\n", "path": "mathesar/api/filters.py"}, {"content": "\"\"\"\nDjango settings for config project.\n\nGenerated by 'django-admin startproject' using Django 3.1.7.\n\nFor more information on this file, see\nhttps://docs.djangoproject.com/en/3.1/topics/settings/\n\nFor the full list of settings and their values, see\nhttps://docs.djangoproject.com/en/3.1/ref/settings/\n\"\"\"\n\nimport os\nfrom pathlib import Path\n\nfrom decouple import Csv, config as decouple_config\nfrom dj_database_url import parse as db_url\n\n\n# We use a 'tuple' with pipes as delimiters as decople naively splits the global\n# variables on commas when casting to Csv()\ndef pipe_delim(pipe_string):\n # Remove opening and closing brackets\n pipe_string = pipe_string[1:-1]\n # Split on pipe delim\n return pipe_string.split(\"|\")\n\n\n# Build paths inside the project like this: BASE_DIR / 'subdir'.\nBASE_DIR = Path(__file__).resolve().parent.parent\n\n# Application definition\n\nINSTALLED_APPS = [\n \"django.contrib.admin\",\n \"django.contrib.auth\",\n \"django.contrib.contenttypes\",\n \"django.contrib.sessions\",\n \"django.contrib.messages\",\n \"django.contrib.staticfiles\",\n \"rest_framework\",\n \"django_filters\",\n \"django_property_filter\",\n \"mathesar\",\n]\n\nMIDDLEWARE = [\n \"django.middleware.security.SecurityMiddleware\",\n \"django.contrib.sessions.middleware.SessionMiddleware\",\n \"django.middleware.common.CommonMiddleware\",\n \"django.middleware.csrf.CsrfViewMiddleware\",\n \"django.contrib.auth.middleware.AuthenticationMiddleware\",\n \"django.contrib.messages.middleware.MessageMiddleware\",\n \"django.middleware.clickjacking.XFrameOptionsMiddleware\",\n]\n\nROOT_URLCONF = \"config.urls\"\n\nTEMPLATES = [\n {\n \"BACKEND\": \"django.template.backends.django.DjangoTemplates\",\n \"DIRS\": [],\n \"APP_DIRS\": True,\n \"OPTIONS\": {\n \"context_processors\": [\n \"config.context_processors.frontend_settings\",\n \"django.template.context_processors.debug\",\n \"django.template.context_processors.request\",\n \"django.contrib.auth.context_processors.auth\",\n \"django.contrib.messages.context_processors.messages\",\n ],\n },\n },\n]\n\nWSGI_APPLICATION = \"config.wsgi.application\"\n\n# Database\n# https://docs.djangoproject.com/en/3.1/ref/settings/#databases\n\n# TODO: Add to documentation that database keys should not be than 128 characters.\n\n# MATHESAR_DATABASES should be of the form '({db_name}|{db_url}), ({db_name}|{db_url})'\n# See pipe_delim above for why we use pipes as delimiters\nDATABASES = {\n db_key: db_url(url_string)\n for db_key, url_string in decouple_config('MATHESAR_DATABASES', cast=Csv(pipe_delim))\n}\nDATABASES[decouple_config('DJANGO_DATABASE_KEY')] = decouple_config('DJANGO_DATABASE_URL', cast=db_url)\n\nfor db_key, db_dict in DATABASES.items():\n # Engine can be '.postgresql' or '.postgresql_psycopg2'\n if not db_dict['ENGINE'].startswith('django.db.backends.postgresql'):\n raise ValueError(\n f\"{db_key} is not a PostgreSQL database. \"\n f\"{db_dict['ENGINE']} found for {db_key}'s engine.\"\n )\n\n\n# pytest-django will create a new database named 'test_{DATABASES[table_db]['NAME']}'\n# and use it for our API tests if we don't specify DATABASES[table_db]['TEST']['NAME']\nif decouple_config('TEST', default=False, cast=bool):\n for db_key, _ in decouple_config('MATHESAR_DATABASES', cast=Csv(pipe_delim)):\n DATABASES[db_key]['TEST'] = {'NAME': DATABASES[db_key]['NAME']}\n\n\n# Quick-start development settings - unsuitable for production\n# See https://docs.djangoproject.com/en/3.1/howto/deployment/checklist/\n\n# SECURITY WARNING: keep the secret key used in production secret!\nSECRET_KEY = decouple_config('SECRET_KEY')\n\n# SECURITY WARNING: don't run with debug turned on in production!\nDEBUG = decouple_config('DEBUG', default=False, cast=bool)\n\nALLOWED_HOSTS = decouple_config('ALLOWED_HOSTS', cast=Csv())\n\n# Password validation\n# https://docs.djangoproject.com/en/3.1/ref/settings/#auth-password-validators\n\nAUTH_PASSWORD_VALIDATORS = [\n {\n \"NAME\": \"django.contrib.auth.password_validation.UserAttributeSimilarityValidator\",\n },\n {\n \"NAME\": \"django.contrib.auth.password_validation.MinimumLengthValidator\",\n },\n {\n \"NAME\": \"django.contrib.auth.password_validation.CommonPasswordValidator\",\n },\n {\n \"NAME\": \"django.contrib.auth.password_validation.NumericPasswordValidator\",\n },\n]\n\n\n# Internationalization\n# https://docs.djangoproject.com/en/3.1/topics/i18n/\n\nLANGUAGE_CODE = \"en-us\"\n\nTIME_ZONE = \"UTC\"\n\nUSE_I18N = True\n\nUSE_L10N = True\n\nUSE_TZ = True\n\n\n# Static files (CSS, JavaScript, Images)\n# https://docs.djangoproject.com/en/3.1/howto/static-files/\n# https://docs.djangoproject.com/en/3.1/ref/contrib/staticfiles/\n\nSTATIC_URL = \"/static/\"\n\n# When running with DEBUG=False, the webserver needs to serve files from this location\n# python manage.py collectstatic has to be run to collect all static files into this location\n# The files need to served in brotli or gzip compressed format\nSTATIC_ROOT = os.path.join(BASE_DIR, 'static/')\n\n# Media files (uploaded by the user)\n\nMEDIA_ROOT = os.path.join(BASE_DIR, '.media/')\n\nMEDIA_URL = \"/media/\"\n\n# Update Authentication classes, removed BasicAuthentication\n# Defaults: https://www.django-rest-framework.org/api-guide/settings/\nREST_FRAMEWORK = {\n 'DEFAULT_AUTHENTICATION_CLASSES': [\n 'rest_framework.authentication.SessionAuthentication'\n ],\n 'DEFAULT_FILTER_BACKENDS': (\n 'django_filters.rest_framework.DjangoFilterBackend',\n 'rest_framework.filters.OrderingFilter',\n ),\n}\n\n# Mathesar settings\nMATHESAR_MODE = decouple_config('MODE', default='PRODUCTION')\nMATHESAR_UI_BUILD_LOCATION = os.path.join(BASE_DIR, 'mathesar/static/mathesar/')\nMATHESAR_MANIFEST_LOCATION = os.path.join(MATHESAR_UI_BUILD_LOCATION, 'manifest.json')\nMATHESAR_CLIENT_DEV_URL = 'http://localhost:3000'\n\n\nSTATICFILES_DIRS = [MATHESAR_UI_BUILD_LOCATION]\n", "path": "config/settings.py"}]} | 2,643 | 542 |
gh_patches_debug_13689 | rasdani/github-patches | git_diff | conan-io__conan-2592 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
conan search gives AttributeError: 'UserIO' object has no attribute 'warn'
Version: 1.1.1
OS: Linux Ubuntu 14.04
conda: v4.2.7
Repro steps:
* `conda create -n conan python=2.7`
* `source activate conan`
* `pip install conan`
* `conan search zlib/1.2.11@conan/stable -r=conan-center`
Gives the following python stack:
```
(conan) ~ $ conan search zlib/1.2.11@conan/stable -r=conan-center
Traceback (most recent call last):
File "/home/mgodbolt/apps/miniconda/envs/conan/lib/python2.7/site-packages/conans/client/command.py", line 1131, in run
method(args[0][1:])
File "/home/mgodbolt/apps/miniconda/envs/conan/lib/python2.7/site-packages/conans/client/command.py", line 814, in search
outdated=args.outdated)
File "/home/mgodbolt/apps/miniconda/envs/conan/lib/python2.7/site-packages/conans/client/conan_api.py", line 64, in wrapper
return f(*args, **kwargs)
File "/home/mgodbolt/apps/miniconda/envs/conan/lib/python2.7/site-packages/conans/client/conan_api.py", line 595, in search_packages
outdated=outdated)
File "/home/mgodbolt/apps/miniconda/envs/conan/lib/python2.7/site-packages/conans/client/cmd/search.py", line 44, in search_packages
remote = RemoteRegistry(self._client_cache.registry, self._user_io).remote(remote)
File "/home/mgodbolt/apps/miniconda/envs/conan/lib/python2.7/site-packages/conans/client/remote_registry.py", line 95, in remote
remotes, _ = self._load()
File "/home/mgodbolt/apps/miniconda/envs/conan/lib/python2.7/site-packages/conans/client/remote_registry.py", line 65, in _load
self._output.warn("Remotes registry file missing, creating default one in %s"
AttributeError: 'UserIO' object has no attribute 'warn'
ERROR: 'UserIO' object has no attribute 'warn'
```
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `conans/client/cmd/search.py`
Content:
```
1 from conans.search.search import DiskSearchManager, filter_outdated
2 from collections import OrderedDict
3 from conans.client.remote_registry import RemoteRegistry
4
5
6 class Search(object):
7 def __init__(self, client_cache, remote_manager, user_io):
8 self._client_cache = client_cache
9 self._remote_manager = remote_manager
10 self._user_io = user_io
11
12 def search_recipes(self, pattern, remote=None, case_sensitive=False):
13 ignorecase = not case_sensitive
14 if not remote:
15 return DiskSearchManager(self._client_cache).search_recipes(pattern, ignorecase)
16
17 registry = RemoteRegistry(self._client_cache.registry, self._user_io.out)
18 if remote == 'all':
19 remotes = registry.remotes
20 # We have to check if there is a remote called "all"
21 # Deprecate: 2.0 can remove this check
22 if 'all' not in (r.name for r in remotes):
23 references = {}
24 for remote in remotes:
25 result = self._remote_manager.search_recipes(remote, pattern, ignorecase)
26 if result:
27 references[remote.name] = result
28 return references
29 # single remote
30 remote = registry.remote(remote)
31 return self._remote_manager.search_recipes(remote, pattern, ignorecase)
32
33 def search_packages(self, reference=None, remote=None, query=None, outdated=False):
34 """ Return the single information saved in conan.vars about all the packages
35 or the packages which match with a pattern
36
37 Attributes:
38 pattern = string to match packages
39 remote = search on another origin to get packages info
40 packages_pattern = String query with binary
41 packages properties: "arch=x86 AND os=Windows"
42 """
43 if remote:
44 remote = RemoteRegistry(self._client_cache.registry, self._user_io).remote(remote)
45 packages_props = self._remote_manager.search_packages(remote, reference, query)
46 ordered_packages = OrderedDict(sorted(packages_props.items()))
47 manifest = self._remote_manager.get_conan_digest(reference, remote)
48 recipe_hash = manifest.summary_hash
49 else:
50 searcher = DiskSearchManager(self._client_cache)
51 packages_props = searcher.search_packages(reference, query)
52 ordered_packages = OrderedDict(sorted(packages_props.items()))
53 try:
54 recipe_hash = self._client_cache.load_manifest(reference).summary_hash
55 except IOError: # It could not exist in local
56 recipe_hash = None
57 if outdated and recipe_hash:
58 ordered_packages = filter_outdated(ordered_packages, recipe_hash)
59 return ordered_packages, reference, recipe_hash, query
60
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/conans/client/cmd/search.py b/conans/client/cmd/search.py
--- a/conans/client/cmd/search.py
+++ b/conans/client/cmd/search.py
@@ -41,7 +41,7 @@
packages properties: "arch=x86 AND os=Windows"
"""
if remote:
- remote = RemoteRegistry(self._client_cache.registry, self._user_io).remote(remote)
+ remote = RemoteRegistry(self._client_cache.registry, self._user_io.out).remote(remote)
packages_props = self._remote_manager.search_packages(remote, reference, query)
ordered_packages = OrderedDict(sorted(packages_props.items()))
manifest = self._remote_manager.get_conan_digest(reference, remote)
| {"golden_diff": "diff --git a/conans/client/cmd/search.py b/conans/client/cmd/search.py\n--- a/conans/client/cmd/search.py\n+++ b/conans/client/cmd/search.py\n@@ -41,7 +41,7 @@\n packages properties: \"arch=x86 AND os=Windows\"\n \"\"\"\n if remote:\n- remote = RemoteRegistry(self._client_cache.registry, self._user_io).remote(remote)\n+ remote = RemoteRegistry(self._client_cache.registry, self._user_io.out).remote(remote)\n packages_props = self._remote_manager.search_packages(remote, reference, query)\n ordered_packages = OrderedDict(sorted(packages_props.items()))\n manifest = self._remote_manager.get_conan_digest(reference, remote)\n", "issue": "conan search gives AttributeError: 'UserIO' object has no attribute 'warn'\nVersion: 1.1.1\r\nOS: Linux Ubuntu 14.04 \r\nconda: v4.2.7\r\n\r\nRepro steps:\r\n* `conda create -n conan python=2.7`\r\n* `source activate conan`\r\n* `pip install conan`\r\n* `conan search zlib/1.2.11@conan/stable -r=conan-center`\r\n\r\nGives the following python stack:\r\n\r\n```\r\n(conan) ~ $ conan search zlib/1.2.11@conan/stable -r=conan-center\r\nTraceback (most recent call last):\r\n File \"/home/mgodbolt/apps/miniconda/envs/conan/lib/python2.7/site-packages/conans/client/command.py\", line 1131, in run\r\n method(args[0][1:])\r\n File \"/home/mgodbolt/apps/miniconda/envs/conan/lib/python2.7/site-packages/conans/client/command.py\", line 814, in search\r\n outdated=args.outdated)\r\n File \"/home/mgodbolt/apps/miniconda/envs/conan/lib/python2.7/site-packages/conans/client/conan_api.py\", line 64, in wrapper\r\n return f(*args, **kwargs)\r\n File \"/home/mgodbolt/apps/miniconda/envs/conan/lib/python2.7/site-packages/conans/client/conan_api.py\", line 595, in search_packages\r\n outdated=outdated)\r\n File \"/home/mgodbolt/apps/miniconda/envs/conan/lib/python2.7/site-packages/conans/client/cmd/search.py\", line 44, in search_packages\r\n remote = RemoteRegistry(self._client_cache.registry, self._user_io).remote(remote)\r\n File \"/home/mgodbolt/apps/miniconda/envs/conan/lib/python2.7/site-packages/conans/client/remote_registry.py\", line 95, in remote\r\n remotes, _ = self._load()\r\n File \"/home/mgodbolt/apps/miniconda/envs/conan/lib/python2.7/site-packages/conans/client/remote_registry.py\", line 65, in _load\r\n self._output.warn(\"Remotes registry file missing, creating default one in %s\"\r\nAttributeError: 'UserIO' object has no attribute 'warn'\r\n\r\nERROR: 'UserIO' object has no attribute 'warn'\r\n```\r\n\n", "before_files": [{"content": "from conans.search.search import DiskSearchManager, filter_outdated\nfrom collections import OrderedDict\nfrom conans.client.remote_registry import RemoteRegistry\n\n\nclass Search(object):\n def __init__(self, client_cache, remote_manager, user_io):\n self._client_cache = client_cache\n self._remote_manager = remote_manager\n self._user_io = user_io\n\n def search_recipes(self, pattern, remote=None, case_sensitive=False):\n ignorecase = not case_sensitive\n if not remote:\n return DiskSearchManager(self._client_cache).search_recipes(pattern, ignorecase)\n\n registry = RemoteRegistry(self._client_cache.registry, self._user_io.out)\n if remote == 'all':\n remotes = registry.remotes\n # We have to check if there is a remote called \"all\"\n # Deprecate: 2.0 can remove this check\n if 'all' not in (r.name for r in remotes):\n references = {}\n for remote in remotes:\n result = self._remote_manager.search_recipes(remote, pattern, ignorecase)\n if result:\n references[remote.name] = result\n return references\n # single remote\n remote = registry.remote(remote)\n return self._remote_manager.search_recipes(remote, pattern, ignorecase)\n\n def search_packages(self, reference=None, remote=None, query=None, outdated=False):\n \"\"\" Return the single information saved in conan.vars about all the packages\n or the packages which match with a pattern\n\n Attributes:\n pattern = string to match packages\n remote = search on another origin to get packages info\n packages_pattern = String query with binary\n packages properties: \"arch=x86 AND os=Windows\"\n \"\"\"\n if remote:\n remote = RemoteRegistry(self._client_cache.registry, self._user_io).remote(remote)\n packages_props = self._remote_manager.search_packages(remote, reference, query)\n ordered_packages = OrderedDict(sorted(packages_props.items()))\n manifest = self._remote_manager.get_conan_digest(reference, remote)\n recipe_hash = manifest.summary_hash\n else:\n searcher = DiskSearchManager(self._client_cache)\n packages_props = searcher.search_packages(reference, query)\n ordered_packages = OrderedDict(sorted(packages_props.items()))\n try:\n recipe_hash = self._client_cache.load_manifest(reference).summary_hash\n except IOError: # It could not exist in local\n recipe_hash = None\n if outdated and recipe_hash:\n ordered_packages = filter_outdated(ordered_packages, recipe_hash)\n return ordered_packages, reference, recipe_hash, query\n", "path": "conans/client/cmd/search.py"}], "after_files": [{"content": "from conans.search.search import DiskSearchManager, filter_outdated\nfrom collections import OrderedDict\nfrom conans.client.remote_registry import RemoteRegistry\n\n\nclass Search(object):\n def __init__(self, client_cache, remote_manager, user_io):\n self._client_cache = client_cache\n self._remote_manager = remote_manager\n self._user_io = user_io\n\n def search_recipes(self, pattern, remote=None, case_sensitive=False):\n ignorecase = not case_sensitive\n if not remote:\n return DiskSearchManager(self._client_cache).search_recipes(pattern, ignorecase)\n\n registry = RemoteRegistry(self._client_cache.registry, self._user_io.out)\n if remote == 'all':\n remotes = registry.remotes\n # We have to check if there is a remote called \"all\"\n # Deprecate: 2.0 can remove this check\n if 'all' not in (r.name for r in remotes):\n references = {}\n for remote in remotes:\n result = self._remote_manager.search_recipes(remote, pattern, ignorecase)\n if result:\n references[remote.name] = result\n return references\n # single remote\n remote = registry.remote(remote)\n return self._remote_manager.search_recipes(remote, pattern, ignorecase)\n\n def search_packages(self, reference=None, remote=None, query=None, outdated=False):\n \"\"\" Return the single information saved in conan.vars about all the packages\n or the packages which match with a pattern\n\n Attributes:\n pattern = string to match packages\n remote = search on another origin to get packages info\n packages_pattern = String query with binary\n packages properties: \"arch=x86 AND os=Windows\"\n \"\"\"\n if remote:\n remote = RemoteRegistry(self._client_cache.registry, self._user_io.out).remote(remote)\n packages_props = self._remote_manager.search_packages(remote, reference, query)\n ordered_packages = OrderedDict(sorted(packages_props.items()))\n manifest = self._remote_manager.get_conan_digest(reference, remote)\n recipe_hash = manifest.summary_hash\n else:\n searcher = DiskSearchManager(self._client_cache)\n packages_props = searcher.search_packages(reference, query)\n ordered_packages = OrderedDict(sorted(packages_props.items()))\n try:\n recipe_hash = self._client_cache.load_manifest(reference).summary_hash\n except IOError: # It could not exist in local\n recipe_hash = None\n if outdated and recipe_hash:\n ordered_packages = filter_outdated(ordered_packages, recipe_hash)\n return ordered_packages, reference, recipe_hash, query\n", "path": "conans/client/cmd/search.py"}]} | 1,453 | 151 |
gh_patches_debug_1659 | rasdani/github-patches | git_diff | huggingface__diffusers-6012 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
logging.remove_handler() has a faulty assertion, doesn't allow registered handlers to be removed
### Describe the bug
in `utils/logging.py` there the function remove_handler seems to have a faulty assertion in it.
```py
def add_handler(handler: logging.Handler) -> None:
"""adds a handler to the HuggingFace Diffusers' root logger."""
_configure_library_root_logger()
assert handler is not None
_get_library_root_logger().addHandler(handler)
def remove_handler(handler: logging.Handler) -> None:
"""removes given handler from the HuggingFace Diffusers' root logger."""
_configure_library_root_logger()
assert handler is not None and handler not in _get_library_root_logger().handlers # <<< THIS ONE
_get_library_root_logger().removeHandler(handler)
```
That line seems to have a `not` that shouldn't be there. (the `not in` to be precise)
Normally I'd just do a PR to remove it, but as I'm not that familiar with the codebase, nor Python in general, and don't yet have a solid grasp of what these things actually do (I'm just playing around trying to familiarize myself), I decided to make an issue instead so people who actually know the codebase can do it.
### Reproduction
```py
from diffusers import logging
from logging import Handler
class TestHandler(Handler):
def __init__(self):
super().__init__()
def emit(self):
pass
handler = TestHandler()
logging.add_handler(handler)
logging.remove_handler(handler)
```
### Logs
```shell
Traceback (most recent call last):
File ".\test.py", line 14, in <module>
logging.remove_handler(handler)
File "C:\Users\XXX\XXX\venv\lib\site-packages\diffusers\utils\logging.py", line 221, in remove_handler
assert handler is not None and handler not in _get_library_root_logger().handlers
AssertionError
```
### System Info
- `diffusers` version: 0.3.0
- Platform: Windows-10-10.0.19041-SP0
- Python version: 3.8.9
- PyTorch version (GPU?): 1.12.1+cpu (False)
- Huggingface_hub version: 0.9.1
- Transformers version: 4.21.3
- Using GPU in script?: no
- Using distributed or parallel set-up in script?: Don't understand the question, but seems irrelevant.
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `src/diffusers/utils/logging.py`
Content:
```
1 # coding=utf-8
2 # Copyright 2023 Optuna, Hugging Face
3 #
4 # Licensed under the Apache License, Version 2.0 (the "License");
5 # you may not use this file except in compliance with the License.
6 # You may obtain a copy of the License at
7 #
8 # http://www.apache.org/licenses/LICENSE-2.0
9 #
10 # Unless required by applicable law or agreed to in writing, software
11 # distributed under the License is distributed on an "AS IS" BASIS,
12 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13 # See the License for the specific language governing permissions and
14 # limitations under the License.
15 """ Logging utilities."""
16
17 import logging
18 import os
19 import sys
20 import threading
21 from logging import (
22 CRITICAL, # NOQA
23 DEBUG, # NOQA
24 ERROR, # NOQA
25 FATAL, # NOQA
26 INFO, # NOQA
27 NOTSET, # NOQA
28 WARN, # NOQA
29 WARNING, # NOQA
30 )
31 from typing import Dict, Optional
32
33 from tqdm import auto as tqdm_lib
34
35
36 _lock = threading.Lock()
37 _default_handler: Optional[logging.Handler] = None
38
39 log_levels = {
40 "debug": logging.DEBUG,
41 "info": logging.INFO,
42 "warning": logging.WARNING,
43 "error": logging.ERROR,
44 "critical": logging.CRITICAL,
45 }
46
47 _default_log_level = logging.WARNING
48
49 _tqdm_active = True
50
51
52 def _get_default_logging_level() -> int:
53 """
54 If DIFFUSERS_VERBOSITY env var is set to one of the valid choices return that as the new default level. If it is
55 not - fall back to `_default_log_level`
56 """
57 env_level_str = os.getenv("DIFFUSERS_VERBOSITY", None)
58 if env_level_str:
59 if env_level_str in log_levels:
60 return log_levels[env_level_str]
61 else:
62 logging.getLogger().warning(
63 f"Unknown option DIFFUSERS_VERBOSITY={env_level_str}, "
64 f"has to be one of: { ', '.join(log_levels.keys()) }"
65 )
66 return _default_log_level
67
68
69 def _get_library_name() -> str:
70 return __name__.split(".")[0]
71
72
73 def _get_library_root_logger() -> logging.Logger:
74 return logging.getLogger(_get_library_name())
75
76
77 def _configure_library_root_logger() -> None:
78 global _default_handler
79
80 with _lock:
81 if _default_handler:
82 # This library has already configured the library root logger.
83 return
84 _default_handler = logging.StreamHandler() # Set sys.stderr as stream.
85 _default_handler.flush = sys.stderr.flush
86
87 # Apply our default configuration to the library root logger.
88 library_root_logger = _get_library_root_logger()
89 library_root_logger.addHandler(_default_handler)
90 library_root_logger.setLevel(_get_default_logging_level())
91 library_root_logger.propagate = False
92
93
94 def _reset_library_root_logger() -> None:
95 global _default_handler
96
97 with _lock:
98 if not _default_handler:
99 return
100
101 library_root_logger = _get_library_root_logger()
102 library_root_logger.removeHandler(_default_handler)
103 library_root_logger.setLevel(logging.NOTSET)
104 _default_handler = None
105
106
107 def get_log_levels_dict() -> Dict[str, int]:
108 return log_levels
109
110
111 def get_logger(name: Optional[str] = None) -> logging.Logger:
112 """
113 Return a logger with the specified name.
114
115 This function is not supposed to be directly accessed unless you are writing a custom diffusers module.
116 """
117
118 if name is None:
119 name = _get_library_name()
120
121 _configure_library_root_logger()
122 return logging.getLogger(name)
123
124
125 def get_verbosity() -> int:
126 """
127 Return the current level for the 🤗 Diffusers' root logger as an `int`.
128
129 Returns:
130 `int`:
131 Logging level integers which can be one of:
132
133 - `50`: `diffusers.logging.CRITICAL` or `diffusers.logging.FATAL`
134 - `40`: `diffusers.logging.ERROR`
135 - `30`: `diffusers.logging.WARNING` or `diffusers.logging.WARN`
136 - `20`: `diffusers.logging.INFO`
137 - `10`: `diffusers.logging.DEBUG`
138
139 """
140
141 _configure_library_root_logger()
142 return _get_library_root_logger().getEffectiveLevel()
143
144
145 def set_verbosity(verbosity: int) -> None:
146 """
147 Set the verbosity level for the 🤗 Diffusers' root logger.
148
149 Args:
150 verbosity (`int`):
151 Logging level which can be one of:
152
153 - `diffusers.logging.CRITICAL` or `diffusers.logging.FATAL`
154 - `diffusers.logging.ERROR`
155 - `diffusers.logging.WARNING` or `diffusers.logging.WARN`
156 - `diffusers.logging.INFO`
157 - `diffusers.logging.DEBUG`
158 """
159
160 _configure_library_root_logger()
161 _get_library_root_logger().setLevel(verbosity)
162
163
164 def set_verbosity_info() -> None:
165 """Set the verbosity to the `INFO` level."""
166 return set_verbosity(INFO)
167
168
169 def set_verbosity_warning() -> None:
170 """Set the verbosity to the `WARNING` level."""
171 return set_verbosity(WARNING)
172
173
174 def set_verbosity_debug() -> None:
175 """Set the verbosity to the `DEBUG` level."""
176 return set_verbosity(DEBUG)
177
178
179 def set_verbosity_error() -> None:
180 """Set the verbosity to the `ERROR` level."""
181 return set_verbosity(ERROR)
182
183
184 def disable_default_handler() -> None:
185 """Disable the default handler of the 🤗 Diffusers' root logger."""
186
187 _configure_library_root_logger()
188
189 assert _default_handler is not None
190 _get_library_root_logger().removeHandler(_default_handler)
191
192
193 def enable_default_handler() -> None:
194 """Enable the default handler of the 🤗 Diffusers' root logger."""
195
196 _configure_library_root_logger()
197
198 assert _default_handler is not None
199 _get_library_root_logger().addHandler(_default_handler)
200
201
202 def add_handler(handler: logging.Handler) -> None:
203 """adds a handler to the HuggingFace Diffusers' root logger."""
204
205 _configure_library_root_logger()
206
207 assert handler is not None
208 _get_library_root_logger().addHandler(handler)
209
210
211 def remove_handler(handler: logging.Handler) -> None:
212 """removes given handler from the HuggingFace Diffusers' root logger."""
213
214 _configure_library_root_logger()
215
216 assert handler is not None and handler not in _get_library_root_logger().handlers
217 _get_library_root_logger().removeHandler(handler)
218
219
220 def disable_propagation() -> None:
221 """
222 Disable propagation of the library log outputs. Note that log propagation is disabled by default.
223 """
224
225 _configure_library_root_logger()
226 _get_library_root_logger().propagate = False
227
228
229 def enable_propagation() -> None:
230 """
231 Enable propagation of the library log outputs. Please disable the HuggingFace Diffusers' default handler to prevent
232 double logging if the root logger has been configured.
233 """
234
235 _configure_library_root_logger()
236 _get_library_root_logger().propagate = True
237
238
239 def enable_explicit_format() -> None:
240 """
241 Enable explicit formatting for every 🤗 Diffusers' logger. The explicit formatter is as follows:
242 ```
243 [LEVELNAME|FILENAME|LINE NUMBER] TIME >> MESSAGE
244 ```
245 All handlers currently bound to the root logger are affected by this method.
246 """
247 handlers = _get_library_root_logger().handlers
248
249 for handler in handlers:
250 formatter = logging.Formatter("[%(levelname)s|%(filename)s:%(lineno)s] %(asctime)s >> %(message)s")
251 handler.setFormatter(formatter)
252
253
254 def reset_format() -> None:
255 """
256 Resets the formatting for 🤗 Diffusers' loggers.
257
258 All handlers currently bound to the root logger are affected by this method.
259 """
260 handlers = _get_library_root_logger().handlers
261
262 for handler in handlers:
263 handler.setFormatter(None)
264
265
266 def warning_advice(self, *args, **kwargs) -> None:
267 """
268 This method is identical to `logger.warning()`, but if env var DIFFUSERS_NO_ADVISORY_WARNINGS=1 is set, this
269 warning will not be printed
270 """
271 no_advisory_warnings = os.getenv("DIFFUSERS_NO_ADVISORY_WARNINGS", False)
272 if no_advisory_warnings:
273 return
274 self.warning(*args, **kwargs)
275
276
277 logging.Logger.warning_advice = warning_advice
278
279
280 class EmptyTqdm:
281 """Dummy tqdm which doesn't do anything."""
282
283 def __init__(self, *args, **kwargs): # pylint: disable=unused-argument
284 self._iterator = args[0] if args else None
285
286 def __iter__(self):
287 return iter(self._iterator)
288
289 def __getattr__(self, _):
290 """Return empty function."""
291
292 def empty_fn(*args, **kwargs): # pylint: disable=unused-argument
293 return
294
295 return empty_fn
296
297 def __enter__(self):
298 return self
299
300 def __exit__(self, type_, value, traceback):
301 return
302
303
304 class _tqdm_cls:
305 def __call__(self, *args, **kwargs):
306 if _tqdm_active:
307 return tqdm_lib.tqdm(*args, **kwargs)
308 else:
309 return EmptyTqdm(*args, **kwargs)
310
311 def set_lock(self, *args, **kwargs):
312 self._lock = None
313 if _tqdm_active:
314 return tqdm_lib.tqdm.set_lock(*args, **kwargs)
315
316 def get_lock(self):
317 if _tqdm_active:
318 return tqdm_lib.tqdm.get_lock()
319
320
321 tqdm = _tqdm_cls()
322
323
324 def is_progress_bar_enabled() -> bool:
325 """Return a boolean indicating whether tqdm progress bars are enabled."""
326 global _tqdm_active
327 return bool(_tqdm_active)
328
329
330 def enable_progress_bar() -> None:
331 """Enable tqdm progress bar."""
332 global _tqdm_active
333 _tqdm_active = True
334
335
336 def disable_progress_bar() -> None:
337 """Disable tqdm progress bar."""
338 global _tqdm_active
339 _tqdm_active = False
340
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/src/diffusers/utils/logging.py b/src/diffusers/utils/logging.py
--- a/src/diffusers/utils/logging.py
+++ b/src/diffusers/utils/logging.py
@@ -213,7 +213,7 @@
_configure_library_root_logger()
- assert handler is not None and handler not in _get_library_root_logger().handlers
+ assert handler is not None and handler in _get_library_root_logger().handlers
_get_library_root_logger().removeHandler(handler)
| {"golden_diff": "diff --git a/src/diffusers/utils/logging.py b/src/diffusers/utils/logging.py\n--- a/src/diffusers/utils/logging.py\n+++ b/src/diffusers/utils/logging.py\n@@ -213,7 +213,7 @@\n \n _configure_library_root_logger()\n \n- assert handler is not None and handler not in _get_library_root_logger().handlers\n+ assert handler is not None and handler in _get_library_root_logger().handlers\n _get_library_root_logger().removeHandler(handler)\n", "issue": "logging.remove_handler() has a faulty assertion, doesn't allow registered handlers to be removed\n### Describe the bug\r\n\r\nin `utils/logging.py` there the function remove_handler seems to have a faulty assertion in it.\r\n\r\n```py\r\ndef add_handler(handler: logging.Handler) -> None:\r\n \"\"\"adds a handler to the HuggingFace Diffusers' root logger.\"\"\"\r\n\r\n _configure_library_root_logger()\r\n\r\n assert handler is not None\r\n _get_library_root_logger().addHandler(handler)\r\n\r\n\r\ndef remove_handler(handler: logging.Handler) -> None:\r\n \"\"\"removes given handler from the HuggingFace Diffusers' root logger.\"\"\"\r\n\r\n _configure_library_root_logger()\r\n\r\n assert handler is not None and handler not in _get_library_root_logger().handlers # <<< THIS ONE\r\n _get_library_root_logger().removeHandler(handler)\r\n```\r\n\r\nThat line seems to have a `not` that shouldn't be there. (the `not in` to be precise)\r\n\r\nNormally I'd just do a PR to remove it, but as I'm not that familiar with the codebase, nor Python in general, and don't yet have a solid grasp of what these things actually do (I'm just playing around trying to familiarize myself), I decided to make an issue instead so people who actually know the codebase can do it.\r\n\r\n### Reproduction\r\n\r\n```py\r\nfrom diffusers import logging\r\nfrom logging import Handler\r\n\r\nclass TestHandler(Handler):\r\n def __init__(self):\r\n super().__init__()\r\n\r\n def emit(self):\r\n pass\r\n\r\nhandler = TestHandler()\r\nlogging.add_handler(handler)\r\nlogging.remove_handler(handler)\r\n```\r\n\r\n### Logs\r\n\r\n```shell\r\nTraceback (most recent call last):\r\n File \".\\test.py\", line 14, in <module>\r\n logging.remove_handler(handler)\r\n File \"C:\\Users\\XXX\\XXX\\venv\\lib\\site-packages\\diffusers\\utils\\logging.py\", line 221, in remove_handler\r\n assert handler is not None and handler not in _get_library_root_logger().handlers\r\nAssertionError\r\n```\r\n\r\n\r\n### System Info\r\n\r\n- `diffusers` version: 0.3.0\r\n- Platform: Windows-10-10.0.19041-SP0\r\n- Python version: 3.8.9\r\n- PyTorch version (GPU?): 1.12.1+cpu (False)\r\n- Huggingface_hub version: 0.9.1\r\n- Transformers version: 4.21.3\r\n- Using GPU in script?: no\r\n- Using distributed or parallel set-up in script?: Don't understand the question, but seems irrelevant.\n", "before_files": [{"content": "# coding=utf-8\n# Copyright 2023 Optuna, Hugging Face\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\"\"\" Logging utilities.\"\"\"\n\nimport logging\nimport os\nimport sys\nimport threading\nfrom logging import (\n CRITICAL, # NOQA\n DEBUG, # NOQA\n ERROR, # NOQA\n FATAL, # NOQA\n INFO, # NOQA\n NOTSET, # NOQA\n WARN, # NOQA\n WARNING, # NOQA\n)\nfrom typing import Dict, Optional\n\nfrom tqdm import auto as tqdm_lib\n\n\n_lock = threading.Lock()\n_default_handler: Optional[logging.Handler] = None\n\nlog_levels = {\n \"debug\": logging.DEBUG,\n \"info\": logging.INFO,\n \"warning\": logging.WARNING,\n \"error\": logging.ERROR,\n \"critical\": logging.CRITICAL,\n}\n\n_default_log_level = logging.WARNING\n\n_tqdm_active = True\n\n\ndef _get_default_logging_level() -> int:\n \"\"\"\n If DIFFUSERS_VERBOSITY env var is set to one of the valid choices return that as the new default level. If it is\n not - fall back to `_default_log_level`\n \"\"\"\n env_level_str = os.getenv(\"DIFFUSERS_VERBOSITY\", None)\n if env_level_str:\n if env_level_str in log_levels:\n return log_levels[env_level_str]\n else:\n logging.getLogger().warning(\n f\"Unknown option DIFFUSERS_VERBOSITY={env_level_str}, \"\n f\"has to be one of: { ', '.join(log_levels.keys()) }\"\n )\n return _default_log_level\n\n\ndef _get_library_name() -> str:\n return __name__.split(\".\")[0]\n\n\ndef _get_library_root_logger() -> logging.Logger:\n return logging.getLogger(_get_library_name())\n\n\ndef _configure_library_root_logger() -> None:\n global _default_handler\n\n with _lock:\n if _default_handler:\n # This library has already configured the library root logger.\n return\n _default_handler = logging.StreamHandler() # Set sys.stderr as stream.\n _default_handler.flush = sys.stderr.flush\n\n # Apply our default configuration to the library root logger.\n library_root_logger = _get_library_root_logger()\n library_root_logger.addHandler(_default_handler)\n library_root_logger.setLevel(_get_default_logging_level())\n library_root_logger.propagate = False\n\n\ndef _reset_library_root_logger() -> None:\n global _default_handler\n\n with _lock:\n if not _default_handler:\n return\n\n library_root_logger = _get_library_root_logger()\n library_root_logger.removeHandler(_default_handler)\n library_root_logger.setLevel(logging.NOTSET)\n _default_handler = None\n\n\ndef get_log_levels_dict() -> Dict[str, int]:\n return log_levels\n\n\ndef get_logger(name: Optional[str] = None) -> logging.Logger:\n \"\"\"\n Return a logger with the specified name.\n\n This function is not supposed to be directly accessed unless you are writing a custom diffusers module.\n \"\"\"\n\n if name is None:\n name = _get_library_name()\n\n _configure_library_root_logger()\n return logging.getLogger(name)\n\n\ndef get_verbosity() -> int:\n \"\"\"\n Return the current level for the \ud83e\udd17 Diffusers' root logger as an `int`.\n\n Returns:\n `int`:\n Logging level integers which can be one of:\n\n - `50`: `diffusers.logging.CRITICAL` or `diffusers.logging.FATAL`\n - `40`: `diffusers.logging.ERROR`\n - `30`: `diffusers.logging.WARNING` or `diffusers.logging.WARN`\n - `20`: `diffusers.logging.INFO`\n - `10`: `diffusers.logging.DEBUG`\n\n \"\"\"\n\n _configure_library_root_logger()\n return _get_library_root_logger().getEffectiveLevel()\n\n\ndef set_verbosity(verbosity: int) -> None:\n \"\"\"\n Set the verbosity level for the \ud83e\udd17 Diffusers' root logger.\n\n Args:\n verbosity (`int`):\n Logging level which can be one of:\n\n - `diffusers.logging.CRITICAL` or `diffusers.logging.FATAL`\n - `diffusers.logging.ERROR`\n - `diffusers.logging.WARNING` or `diffusers.logging.WARN`\n - `diffusers.logging.INFO`\n - `diffusers.logging.DEBUG`\n \"\"\"\n\n _configure_library_root_logger()\n _get_library_root_logger().setLevel(verbosity)\n\n\ndef set_verbosity_info() -> None:\n \"\"\"Set the verbosity to the `INFO` level.\"\"\"\n return set_verbosity(INFO)\n\n\ndef set_verbosity_warning() -> None:\n \"\"\"Set the verbosity to the `WARNING` level.\"\"\"\n return set_verbosity(WARNING)\n\n\ndef set_verbosity_debug() -> None:\n \"\"\"Set the verbosity to the `DEBUG` level.\"\"\"\n return set_verbosity(DEBUG)\n\n\ndef set_verbosity_error() -> None:\n \"\"\"Set the verbosity to the `ERROR` level.\"\"\"\n return set_verbosity(ERROR)\n\n\ndef disable_default_handler() -> None:\n \"\"\"Disable the default handler of the \ud83e\udd17 Diffusers' root logger.\"\"\"\n\n _configure_library_root_logger()\n\n assert _default_handler is not None\n _get_library_root_logger().removeHandler(_default_handler)\n\n\ndef enable_default_handler() -> None:\n \"\"\"Enable the default handler of the \ud83e\udd17 Diffusers' root logger.\"\"\"\n\n _configure_library_root_logger()\n\n assert _default_handler is not None\n _get_library_root_logger().addHandler(_default_handler)\n\n\ndef add_handler(handler: logging.Handler) -> None:\n \"\"\"adds a handler to the HuggingFace Diffusers' root logger.\"\"\"\n\n _configure_library_root_logger()\n\n assert handler is not None\n _get_library_root_logger().addHandler(handler)\n\n\ndef remove_handler(handler: logging.Handler) -> None:\n \"\"\"removes given handler from the HuggingFace Diffusers' root logger.\"\"\"\n\n _configure_library_root_logger()\n\n assert handler is not None and handler not in _get_library_root_logger().handlers\n _get_library_root_logger().removeHandler(handler)\n\n\ndef disable_propagation() -> None:\n \"\"\"\n Disable propagation of the library log outputs. Note that log propagation is disabled by default.\n \"\"\"\n\n _configure_library_root_logger()\n _get_library_root_logger().propagate = False\n\n\ndef enable_propagation() -> None:\n \"\"\"\n Enable propagation of the library log outputs. Please disable the HuggingFace Diffusers' default handler to prevent\n double logging if the root logger has been configured.\n \"\"\"\n\n _configure_library_root_logger()\n _get_library_root_logger().propagate = True\n\n\ndef enable_explicit_format() -> None:\n \"\"\"\n Enable explicit formatting for every \ud83e\udd17 Diffusers' logger. The explicit formatter is as follows:\n ```\n [LEVELNAME|FILENAME|LINE NUMBER] TIME >> MESSAGE\n ```\n All handlers currently bound to the root logger are affected by this method.\n \"\"\"\n handlers = _get_library_root_logger().handlers\n\n for handler in handlers:\n formatter = logging.Formatter(\"[%(levelname)s|%(filename)s:%(lineno)s] %(asctime)s >> %(message)s\")\n handler.setFormatter(formatter)\n\n\ndef reset_format() -> None:\n \"\"\"\n Resets the formatting for \ud83e\udd17 Diffusers' loggers.\n\n All handlers currently bound to the root logger are affected by this method.\n \"\"\"\n handlers = _get_library_root_logger().handlers\n\n for handler in handlers:\n handler.setFormatter(None)\n\n\ndef warning_advice(self, *args, **kwargs) -> None:\n \"\"\"\n This method is identical to `logger.warning()`, but if env var DIFFUSERS_NO_ADVISORY_WARNINGS=1 is set, this\n warning will not be printed\n \"\"\"\n no_advisory_warnings = os.getenv(\"DIFFUSERS_NO_ADVISORY_WARNINGS\", False)\n if no_advisory_warnings:\n return\n self.warning(*args, **kwargs)\n\n\nlogging.Logger.warning_advice = warning_advice\n\n\nclass EmptyTqdm:\n \"\"\"Dummy tqdm which doesn't do anything.\"\"\"\n\n def __init__(self, *args, **kwargs): # pylint: disable=unused-argument\n self._iterator = args[0] if args else None\n\n def __iter__(self):\n return iter(self._iterator)\n\n def __getattr__(self, _):\n \"\"\"Return empty function.\"\"\"\n\n def empty_fn(*args, **kwargs): # pylint: disable=unused-argument\n return\n\n return empty_fn\n\n def __enter__(self):\n return self\n\n def __exit__(self, type_, value, traceback):\n return\n\n\nclass _tqdm_cls:\n def __call__(self, *args, **kwargs):\n if _tqdm_active:\n return tqdm_lib.tqdm(*args, **kwargs)\n else:\n return EmptyTqdm(*args, **kwargs)\n\n def set_lock(self, *args, **kwargs):\n self._lock = None\n if _tqdm_active:\n return tqdm_lib.tqdm.set_lock(*args, **kwargs)\n\n def get_lock(self):\n if _tqdm_active:\n return tqdm_lib.tqdm.get_lock()\n\n\ntqdm = _tqdm_cls()\n\n\ndef is_progress_bar_enabled() -> bool:\n \"\"\"Return a boolean indicating whether tqdm progress bars are enabled.\"\"\"\n global _tqdm_active\n return bool(_tqdm_active)\n\n\ndef enable_progress_bar() -> None:\n \"\"\"Enable tqdm progress bar.\"\"\"\n global _tqdm_active\n _tqdm_active = True\n\n\ndef disable_progress_bar() -> None:\n \"\"\"Disable tqdm progress bar.\"\"\"\n global _tqdm_active\n _tqdm_active = False\n", "path": "src/diffusers/utils/logging.py"}], "after_files": [{"content": "# coding=utf-8\n# Copyright 2023 Optuna, Hugging Face\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\"\"\" Logging utilities.\"\"\"\n\nimport logging\nimport os\nimport sys\nimport threading\nfrom logging import (\n CRITICAL, # NOQA\n DEBUG, # NOQA\n ERROR, # NOQA\n FATAL, # NOQA\n INFO, # NOQA\n NOTSET, # NOQA\n WARN, # NOQA\n WARNING, # NOQA\n)\nfrom typing import Dict, Optional\n\nfrom tqdm import auto as tqdm_lib\n\n\n_lock = threading.Lock()\n_default_handler: Optional[logging.Handler] = None\n\nlog_levels = {\n \"debug\": logging.DEBUG,\n \"info\": logging.INFO,\n \"warning\": logging.WARNING,\n \"error\": logging.ERROR,\n \"critical\": logging.CRITICAL,\n}\n\n_default_log_level = logging.WARNING\n\n_tqdm_active = True\n\n\ndef _get_default_logging_level() -> int:\n \"\"\"\n If DIFFUSERS_VERBOSITY env var is set to one of the valid choices return that as the new default level. If it is\n not - fall back to `_default_log_level`\n \"\"\"\n env_level_str = os.getenv(\"DIFFUSERS_VERBOSITY\", None)\n if env_level_str:\n if env_level_str in log_levels:\n return log_levels[env_level_str]\n else:\n logging.getLogger().warning(\n f\"Unknown option DIFFUSERS_VERBOSITY={env_level_str}, \"\n f\"has to be one of: { ', '.join(log_levels.keys()) }\"\n )\n return _default_log_level\n\n\ndef _get_library_name() -> str:\n return __name__.split(\".\")[0]\n\n\ndef _get_library_root_logger() -> logging.Logger:\n return logging.getLogger(_get_library_name())\n\n\ndef _configure_library_root_logger() -> None:\n global _default_handler\n\n with _lock:\n if _default_handler:\n # This library has already configured the library root logger.\n return\n _default_handler = logging.StreamHandler() # Set sys.stderr as stream.\n _default_handler.flush = sys.stderr.flush\n\n # Apply our default configuration to the library root logger.\n library_root_logger = _get_library_root_logger()\n library_root_logger.addHandler(_default_handler)\n library_root_logger.setLevel(_get_default_logging_level())\n library_root_logger.propagate = False\n\n\ndef _reset_library_root_logger() -> None:\n global _default_handler\n\n with _lock:\n if not _default_handler:\n return\n\n library_root_logger = _get_library_root_logger()\n library_root_logger.removeHandler(_default_handler)\n library_root_logger.setLevel(logging.NOTSET)\n _default_handler = None\n\n\ndef get_log_levels_dict() -> Dict[str, int]:\n return log_levels\n\n\ndef get_logger(name: Optional[str] = None) -> logging.Logger:\n \"\"\"\n Return a logger with the specified name.\n\n This function is not supposed to be directly accessed unless you are writing a custom diffusers module.\n \"\"\"\n\n if name is None:\n name = _get_library_name()\n\n _configure_library_root_logger()\n return logging.getLogger(name)\n\n\ndef get_verbosity() -> int:\n \"\"\"\n Return the current level for the \ud83e\udd17 Diffusers' root logger as an `int`.\n\n Returns:\n `int`:\n Logging level integers which can be one of:\n\n - `50`: `diffusers.logging.CRITICAL` or `diffusers.logging.FATAL`\n - `40`: `diffusers.logging.ERROR`\n - `30`: `diffusers.logging.WARNING` or `diffusers.logging.WARN`\n - `20`: `diffusers.logging.INFO`\n - `10`: `diffusers.logging.DEBUG`\n\n \"\"\"\n\n _configure_library_root_logger()\n return _get_library_root_logger().getEffectiveLevel()\n\n\ndef set_verbosity(verbosity: int) -> None:\n \"\"\"\n Set the verbosity level for the \ud83e\udd17 Diffusers' root logger.\n\n Args:\n verbosity (`int`):\n Logging level which can be one of:\n\n - `diffusers.logging.CRITICAL` or `diffusers.logging.FATAL`\n - `diffusers.logging.ERROR`\n - `diffusers.logging.WARNING` or `diffusers.logging.WARN`\n - `diffusers.logging.INFO`\n - `diffusers.logging.DEBUG`\n \"\"\"\n\n _configure_library_root_logger()\n _get_library_root_logger().setLevel(verbosity)\n\n\ndef set_verbosity_info() -> None:\n \"\"\"Set the verbosity to the `INFO` level.\"\"\"\n return set_verbosity(INFO)\n\n\ndef set_verbosity_warning() -> None:\n \"\"\"Set the verbosity to the `WARNING` level.\"\"\"\n return set_verbosity(WARNING)\n\n\ndef set_verbosity_debug() -> None:\n \"\"\"Set the verbosity to the `DEBUG` level.\"\"\"\n return set_verbosity(DEBUG)\n\n\ndef set_verbosity_error() -> None:\n \"\"\"Set the verbosity to the `ERROR` level.\"\"\"\n return set_verbosity(ERROR)\n\n\ndef disable_default_handler() -> None:\n \"\"\"Disable the default handler of the \ud83e\udd17 Diffusers' root logger.\"\"\"\n\n _configure_library_root_logger()\n\n assert _default_handler is not None\n _get_library_root_logger().removeHandler(_default_handler)\n\n\ndef enable_default_handler() -> None:\n \"\"\"Enable the default handler of the \ud83e\udd17 Diffusers' root logger.\"\"\"\n\n _configure_library_root_logger()\n\n assert _default_handler is not None\n _get_library_root_logger().addHandler(_default_handler)\n\n\ndef add_handler(handler: logging.Handler) -> None:\n \"\"\"adds a handler to the HuggingFace Diffusers' root logger.\"\"\"\n\n _configure_library_root_logger()\n\n assert handler is not None\n _get_library_root_logger().addHandler(handler)\n\n\ndef remove_handler(handler: logging.Handler) -> None:\n \"\"\"removes given handler from the HuggingFace Diffusers' root logger.\"\"\"\n\n _configure_library_root_logger()\n\n assert handler is not None and handler in _get_library_root_logger().handlers\n _get_library_root_logger().removeHandler(handler)\n\n\ndef disable_propagation() -> None:\n \"\"\"\n Disable propagation of the library log outputs. Note that log propagation is disabled by default.\n \"\"\"\n\n _configure_library_root_logger()\n _get_library_root_logger().propagate = False\n\n\ndef enable_propagation() -> None:\n \"\"\"\n Enable propagation of the library log outputs. Please disable the HuggingFace Diffusers' default handler to prevent\n double logging if the root logger has been configured.\n \"\"\"\n\n _configure_library_root_logger()\n _get_library_root_logger().propagate = True\n\n\ndef enable_explicit_format() -> None:\n \"\"\"\n Enable explicit formatting for every \ud83e\udd17 Diffusers' logger. The explicit formatter is as follows:\n ```\n [LEVELNAME|FILENAME|LINE NUMBER] TIME >> MESSAGE\n ```\n All handlers currently bound to the root logger are affected by this method.\n \"\"\"\n handlers = _get_library_root_logger().handlers\n\n for handler in handlers:\n formatter = logging.Formatter(\"[%(levelname)s|%(filename)s:%(lineno)s] %(asctime)s >> %(message)s\")\n handler.setFormatter(formatter)\n\n\ndef reset_format() -> None:\n \"\"\"\n Resets the formatting for \ud83e\udd17 Diffusers' loggers.\n\n All handlers currently bound to the root logger are affected by this method.\n \"\"\"\n handlers = _get_library_root_logger().handlers\n\n for handler in handlers:\n handler.setFormatter(None)\n\n\ndef warning_advice(self, *args, **kwargs) -> None:\n \"\"\"\n This method is identical to `logger.warning()`, but if env var DIFFUSERS_NO_ADVISORY_WARNINGS=1 is set, this\n warning will not be printed\n \"\"\"\n no_advisory_warnings = os.getenv(\"DIFFUSERS_NO_ADVISORY_WARNINGS\", False)\n if no_advisory_warnings:\n return\n self.warning(*args, **kwargs)\n\n\nlogging.Logger.warning_advice = warning_advice\n\n\nclass EmptyTqdm:\n \"\"\"Dummy tqdm which doesn't do anything.\"\"\"\n\n def __init__(self, *args, **kwargs): # pylint: disable=unused-argument\n self._iterator = args[0] if args else None\n\n def __iter__(self):\n return iter(self._iterator)\n\n def __getattr__(self, _):\n \"\"\"Return empty function.\"\"\"\n\n def empty_fn(*args, **kwargs): # pylint: disable=unused-argument\n return\n\n return empty_fn\n\n def __enter__(self):\n return self\n\n def __exit__(self, type_, value, traceback):\n return\n\n\nclass _tqdm_cls:\n def __call__(self, *args, **kwargs):\n if _tqdm_active:\n return tqdm_lib.tqdm(*args, **kwargs)\n else:\n return EmptyTqdm(*args, **kwargs)\n\n def set_lock(self, *args, **kwargs):\n self._lock = None\n if _tqdm_active:\n return tqdm_lib.tqdm.set_lock(*args, **kwargs)\n\n def get_lock(self):\n if _tqdm_active:\n return tqdm_lib.tqdm.get_lock()\n\n\ntqdm = _tqdm_cls()\n\n\ndef is_progress_bar_enabled() -> bool:\n \"\"\"Return a boolean indicating whether tqdm progress bars are enabled.\"\"\"\n global _tqdm_active\n return bool(_tqdm_active)\n\n\ndef enable_progress_bar() -> None:\n \"\"\"Enable tqdm progress bar.\"\"\"\n global _tqdm_active\n _tqdm_active = True\n\n\ndef disable_progress_bar() -> None:\n \"\"\"Disable tqdm progress bar.\"\"\"\n global _tqdm_active\n _tqdm_active = False\n", "path": "src/diffusers/utils/logging.py"}]} | 3,995 | 111 |
gh_patches_debug_10399 | rasdani/github-patches | git_diff | horovod__horovod-3505 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
hvd.DistributedOptimizer gradient accumulation doesn't clean up infinite gradient correctly
**Environment:**
1. Framework: (TensorFlow, Keras, PyTorch, MXNet) Keras
2. Framework version: 2.4
3. Horovod version: 2.3
4. MPI version:
5. CUDA version:
6. NCCL version:
7. Python version:
8. Spark / PySpark version:
9. Ray version:
10. OS and version:
11. GCC version:
12. CMake version:
**Checklist:**
1. Did you search issues to find if somebody asked this question before?
2. If your question is about hang, did you read [this doc](https://github.com/horovod/horovod/blob/master/docs/running.rst)?
3. If your question is about docker, did you read [this doc](https://github.com/horovod/horovod/blob/master/docs/docker.rst)?
4. Did you check if you question is answered in the [troubleshooting guide](https://github.com/horovod/horovod/blob/master/docs/troubleshooting.rst)?
**Bug report:**
We were training in TensorFlow [FP16 mixed precision](https://www.tensorflow.org/guide/mixed_precision) with keras `model.fit()` and with gradient accumulation/aggregation (`backward_pass_per_step` in `hvd.DistributedOptimizer`) and noticed that the [GradientAggregationHelperEager](https://github.com/horovod/horovod/blob/master/horovod/tensorflow/gradient_aggregation_eager.py#L8) doesn't work correctly with FP16 when the loss goes infinite. Details:
It is kind of expected that at the very first 2-15 steps of the training, the gradient out of TF [LossScaleOptimizer](https://github.com/keras-team/keras/blob/v2.8.0/keras/mixed_precision/loss_scale_optimizer.py#L258-L844) is infinite (because the default initial loss scale factor is as large as `2**15`). Dynamic LSO can handle this gracefully, it just skips applying gradient of that step and divides the scale factor by half. However horovod GradientAggregationHelper will anyway add the infinite gradient up locally, and the infinite gradient will never be correctly cleaned up in [this way](https://github.com/horovod/horovod/blob/133ef0725253db83cfb82a4ed4003df76d189829/horovod/tensorflow/gradient_aggregation_eager.py#L119-L123):
```
def _clear_vars(self):
self.counter.assign(0)
for idx in self.locally_aggregated_grads.keys():
self.locally_aggregated_grads[idx].assign_add(
-1 * self.locally_aggregated_grads[idx])
```
as the result of adding inf value by its negative val will be NaN.
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `horovod/tensorflow/gradient_aggregation_eager.py`
Content:
```
1 from distutils.version import LooseVersion
2
3 import tensorflow as tf
4
5 _POST_TF_2_4_0 = LooseVersion(tf.__version__) >= LooseVersion('2.4.0')
6
7
8 class LocalGradientAggregationHelperEager:
9 """
10 LocalGradientAggregationHelperEager aggregates gradient updates
11 locally, and communicates the updates across machines only once
12 every backward_passes_per_step. Only supports eager execution.
13 """
14
15 def __init__(
16 self,
17 backward_passes_per_step,
18 allreduce_func,
19 sparse_as_dense,
20 average_aggregated_gradients,
21 ):
22 self.allreduce_grads = allreduce_func
23 self.sparse_as_dense = sparse_as_dense
24
25 # backward_passes_per_step controls how often gradient updates are
26 # synchronized.
27 self.backward_passes_per_step = backward_passes_per_step
28 if self.backward_passes_per_step <= 0:
29 raise ValueError("backward_passes_per_step must be > 0")
30
31 # average_aggregated_gradients controls whether gradient updates that are
32 # aggregated, should be divided by `backward_passes_per_step`.
33 self.average_aggregated_gradients = average_aggregated_gradients
34
35 # This is going to be N data structure holding the aggregated gradient updates
36 # for parameter updates. N is the number of parameters.
37 self.locally_aggregated_grads = {}
38
39 # Used to know when to allreduce and apply gradients. We allreduce when `self.counter`
40 # is equal to `self.backward_passes_per_step`. We apply gradients when `self.counter`
41 # is equal to 0.
42 self.counter = tf.Variable(initial_value=0)
43
44 def compute_gradients(self, grads, vars):
45 # On steps where allreduce happens, resulting_grads returns the allreduced
46 # gradients, on other steps it returns the locally aggregated
47 # gradients.
48 resulting_grads = []
49
50 for idx, grad in enumerate(grads):
51 # Handle IndexedSlices.
52 if self.sparse_as_dense and isinstance(grad, tf.IndexedSlices):
53 grad = tf.convert_to_tensor(grad)
54 elif isinstance(grad, tf.IndexedSlices):
55 raise ValueError(
56 "IndexedSlices are not supported when "
57 "`backward_passes_per_step` > 1 and "
58 "`sparse_as_dense` is False."
59 )
60
61 # Create variables to store to aggregate gradients if they don't
62 # already exist. Skip variables that are None.
63 if idx not in self.locally_aggregated_grads.keys():
64 if grad is not None:
65 self.locally_aggregated_grads[idx] = tf.Variable(
66 initial_value=tf.zeros_like(grad),
67 trainable=False,
68 dtype=grad.dtype,
69 )
70
71 if grad is None:
72 resulting_grads.append(None)
73 else:
74 self.locally_aggregated_grads[idx].assign_add(grad)
75 resulting_grads.append(
76 self.locally_aggregated_grads[idx].read_value())
77 assert len(self.locally_aggregated_grads) == len(grads)
78
79 # Increment counter.
80 self.counter.assign_add(1)
81
82 def _all_reduce_and_clear_aggregated_variables(aggregated_gradients, vars):
83 # Performs allreduce. If `average_aggregated_gradients` is
84 # set to True divides result by `backward_passes_per_step`.
85 reduced_gradients = self._allreduce_helper(aggregated_gradients, vars)
86 assert len(reduced_gradients) == len(grads)
87
88 self._clear_vars()
89 return reduced_gradients
90
91 def _do_nothing(aggregated_gradients):
92 return aggregated_gradients
93
94 resulting_grads = tf.cond(
95 pred=tf.equal(self.counter, self.backward_passes_per_step),
96 true_fn=lambda: _all_reduce_and_clear_aggregated_variables(resulting_grads, vars),
97 false_fn=lambda: _do_nothing(resulting_grads),
98 )
99
100 return resulting_grads
101
102 def _allreduce_helper(self, grads, vars):
103 allreduced_grads = self.allreduce_grads(grads, vars)
104
105 if not self.average_aggregated_gradients:
106 return allreduced_grads
107
108 scaled_allreduced_grads = []
109 for grad in allreduced_grads:
110 if grad is None:
111 scaled_allreduced_grads.append(grad)
112 continue
113
114 scaled_allreduced_grads.append(
115 grad / self.backward_passes_per_step)
116
117 return scaled_allreduced_grads
118
119 def _clear_vars(self):
120 self.counter.assign(0)
121 for idx in self.locally_aggregated_grads.keys():
122 self.locally_aggregated_grads[idx].assign_add(
123 -1 * self.locally_aggregated_grads[idx])
124
125 def apply_gradients(self, apply_grads_closure, optimizer, *args, **kwargs):
126 def increment_optimizer_iteration():
127 if hasattr(optimizer, "_iterations") and optimizer._iterations is not None:
128 return optimizer._iterations.assign_add(1).op
129 return tf.no_op()
130
131 def non_aggregation_step():
132 if _POST_TF_2_4_0:
133 # In TF 2.4+ `_aggregate_gradients()` is called from inside of `apply_gradients()`.
134 # We account for this by calling `_aggregate_gradients()` for steps where we do
135 # not call `apply_gradients()`.
136 transformed_grads_and_vars = optimizer._transform_unaggregated_gradients(
137 args[0])
138 _ = optimizer._aggregate_gradients(transformed_grads_and_vars)
139
140 return increment_optimizer_iteration()
141
142 def is_aggregation_step():
143 return tf.equal(self.counter, 0)
144
145 return tf.cond(
146 pred=is_aggregation_step(),
147 true_fn=apply_grads_closure,
148 false_fn=non_aggregation_step,
149 )
150
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/horovod/tensorflow/gradient_aggregation_eager.py b/horovod/tensorflow/gradient_aggregation_eager.py
--- a/horovod/tensorflow/gradient_aggregation_eager.py
+++ b/horovod/tensorflow/gradient_aggregation_eager.py
@@ -119,8 +119,8 @@
def _clear_vars(self):
self.counter.assign(0)
for idx in self.locally_aggregated_grads.keys():
- self.locally_aggregated_grads[idx].assign_add(
- -1 * self.locally_aggregated_grads[idx])
+ self.locally_aggregated_grads[idx].assign(
+ tf.zeros_like(self.locally_aggregated_grads[idx]))
def apply_gradients(self, apply_grads_closure, optimizer, *args, **kwargs):
def increment_optimizer_iteration():
| {"golden_diff": "diff --git a/horovod/tensorflow/gradient_aggregation_eager.py b/horovod/tensorflow/gradient_aggregation_eager.py\n--- a/horovod/tensorflow/gradient_aggregation_eager.py\n+++ b/horovod/tensorflow/gradient_aggregation_eager.py\n@@ -119,8 +119,8 @@\n def _clear_vars(self):\n self.counter.assign(0)\n for idx in self.locally_aggregated_grads.keys():\n- self.locally_aggregated_grads[idx].assign_add(\n- -1 * self.locally_aggregated_grads[idx])\n+ self.locally_aggregated_grads[idx].assign(\n+ tf.zeros_like(self.locally_aggregated_grads[idx]))\n \n def apply_gradients(self, apply_grads_closure, optimizer, *args, **kwargs):\n def increment_optimizer_iteration():\n", "issue": "hvd.DistributedOptimizer gradient accumulation doesn't clean up infinite gradient correctly\n**Environment:**\r\n1. Framework: (TensorFlow, Keras, PyTorch, MXNet) Keras\r\n2. Framework version: 2.4\r\n3. Horovod version: 2.3\r\n4. MPI version: \r\n5. CUDA version: \r\n6. NCCL version:\r\n7. Python version:\r\n8. Spark / PySpark version:\r\n9. Ray version:\r\n10. OS and version:\r\n11. GCC version:\r\n12. CMake version:\r\n\r\n**Checklist:**\r\n1. Did you search issues to find if somebody asked this question before?\r\n2. If your question is about hang, did you read [this doc](https://github.com/horovod/horovod/blob/master/docs/running.rst)?\r\n3. If your question is about docker, did you read [this doc](https://github.com/horovod/horovod/blob/master/docs/docker.rst)?\r\n4. Did you check if you question is answered in the [troubleshooting guide](https://github.com/horovod/horovod/blob/master/docs/troubleshooting.rst)?\r\n\r\n**Bug report:**\r\n\r\nWe were training in TensorFlow [FP16 mixed precision](https://www.tensorflow.org/guide/mixed_precision) with keras `model.fit()` and with gradient accumulation/aggregation (`backward_pass_per_step` in `hvd.DistributedOptimizer`) and noticed that the [GradientAggregationHelperEager](https://github.com/horovod/horovod/blob/master/horovod/tensorflow/gradient_aggregation_eager.py#L8) doesn't work correctly with FP16 when the loss goes infinite. Details:\r\n\r\nIt is kind of expected that at the very first 2-15 steps of the training, the gradient out of TF [LossScaleOptimizer](https://github.com/keras-team/keras/blob/v2.8.0/keras/mixed_precision/loss_scale_optimizer.py#L258-L844) is infinite (because the default initial loss scale factor is as large as `2**15`). Dynamic LSO can handle this gracefully, it just skips applying gradient of that step and divides the scale factor by half. However horovod GradientAggregationHelper will anyway add the infinite gradient up locally, and the infinite gradient will never be correctly cleaned up in [this way](https://github.com/horovod/horovod/blob/133ef0725253db83cfb82a4ed4003df76d189829/horovod/tensorflow/gradient_aggregation_eager.py#L119-L123):\r\n```\r\n def _clear_vars(self):\r\n self.counter.assign(0)\r\n for idx in self.locally_aggregated_grads.keys():\r\n self.locally_aggregated_grads[idx].assign_add(\r\n -1 * self.locally_aggregated_grads[idx])\r\n```\r\n\r\nas the result of adding inf value by its negative val will be NaN. \r\n\n", "before_files": [{"content": "from distutils.version import LooseVersion\n\nimport tensorflow as tf\n\n_POST_TF_2_4_0 = LooseVersion(tf.__version__) >= LooseVersion('2.4.0')\n\n\nclass LocalGradientAggregationHelperEager:\n \"\"\"\n LocalGradientAggregationHelperEager aggregates gradient updates\n locally, and communicates the updates across machines only once\n every backward_passes_per_step. Only supports eager execution.\n \"\"\"\n\n def __init__(\n self,\n backward_passes_per_step,\n allreduce_func,\n sparse_as_dense,\n average_aggregated_gradients,\n ):\n self.allreduce_grads = allreduce_func\n self.sparse_as_dense = sparse_as_dense\n\n # backward_passes_per_step controls how often gradient updates are\n # synchronized.\n self.backward_passes_per_step = backward_passes_per_step\n if self.backward_passes_per_step <= 0:\n raise ValueError(\"backward_passes_per_step must be > 0\")\n\n # average_aggregated_gradients controls whether gradient updates that are\n # aggregated, should be divided by `backward_passes_per_step`.\n self.average_aggregated_gradients = average_aggregated_gradients\n\n # This is going to be N data structure holding the aggregated gradient updates\n # for parameter updates. N is the number of parameters.\n self.locally_aggregated_grads = {}\n\n # Used to know when to allreduce and apply gradients. We allreduce when `self.counter`\n # is equal to `self.backward_passes_per_step`. We apply gradients when `self.counter`\n # is equal to 0.\n self.counter = tf.Variable(initial_value=0)\n\n def compute_gradients(self, grads, vars):\n # On steps where allreduce happens, resulting_grads returns the allreduced\n # gradients, on other steps it returns the locally aggregated\n # gradients.\n resulting_grads = []\n\n for idx, grad in enumerate(grads):\n # Handle IndexedSlices.\n if self.sparse_as_dense and isinstance(grad, tf.IndexedSlices):\n grad = tf.convert_to_tensor(grad)\n elif isinstance(grad, tf.IndexedSlices):\n raise ValueError(\n \"IndexedSlices are not supported when \"\n \"`backward_passes_per_step` > 1 and \"\n \"`sparse_as_dense` is False.\"\n )\n\n # Create variables to store to aggregate gradients if they don't\n # already exist. Skip variables that are None.\n if idx not in self.locally_aggregated_grads.keys():\n if grad is not None:\n self.locally_aggregated_grads[idx] = tf.Variable(\n initial_value=tf.zeros_like(grad),\n trainable=False,\n dtype=grad.dtype,\n )\n\n if grad is None:\n resulting_grads.append(None)\n else:\n self.locally_aggregated_grads[idx].assign_add(grad)\n resulting_grads.append(\n self.locally_aggregated_grads[idx].read_value())\n assert len(self.locally_aggregated_grads) == len(grads)\n\n # Increment counter.\n self.counter.assign_add(1)\n\n def _all_reduce_and_clear_aggregated_variables(aggregated_gradients, vars):\n # Performs allreduce. If `average_aggregated_gradients` is\n # set to True divides result by `backward_passes_per_step`.\n reduced_gradients = self._allreduce_helper(aggregated_gradients, vars)\n assert len(reduced_gradients) == len(grads)\n\n self._clear_vars()\n return reduced_gradients\n\n def _do_nothing(aggregated_gradients):\n return aggregated_gradients\n\n resulting_grads = tf.cond(\n pred=tf.equal(self.counter, self.backward_passes_per_step),\n true_fn=lambda: _all_reduce_and_clear_aggregated_variables(resulting_grads, vars),\n false_fn=lambda: _do_nothing(resulting_grads),\n )\n\n return resulting_grads\n\n def _allreduce_helper(self, grads, vars):\n allreduced_grads = self.allreduce_grads(grads, vars)\n\n if not self.average_aggregated_gradients:\n return allreduced_grads\n\n scaled_allreduced_grads = []\n for grad in allreduced_grads:\n if grad is None:\n scaled_allreduced_grads.append(grad)\n continue\n\n scaled_allreduced_grads.append(\n grad / self.backward_passes_per_step)\n\n return scaled_allreduced_grads\n\n def _clear_vars(self):\n self.counter.assign(0)\n for idx in self.locally_aggregated_grads.keys():\n self.locally_aggregated_grads[idx].assign_add(\n -1 * self.locally_aggregated_grads[idx])\n\n def apply_gradients(self, apply_grads_closure, optimizer, *args, **kwargs):\n def increment_optimizer_iteration():\n if hasattr(optimizer, \"_iterations\") and optimizer._iterations is not None:\n return optimizer._iterations.assign_add(1).op\n return tf.no_op()\n\n def non_aggregation_step():\n if _POST_TF_2_4_0:\n # In TF 2.4+ `_aggregate_gradients()` is called from inside of `apply_gradients()`.\n # We account for this by calling `_aggregate_gradients()` for steps where we do\n # not call `apply_gradients()`.\n transformed_grads_and_vars = optimizer._transform_unaggregated_gradients(\n args[0])\n _ = optimizer._aggregate_gradients(transformed_grads_and_vars)\n\n return increment_optimizer_iteration()\n\n def is_aggregation_step():\n return tf.equal(self.counter, 0)\n\n return tf.cond(\n pred=is_aggregation_step(),\n true_fn=apply_grads_closure,\n false_fn=non_aggregation_step,\n )\n", "path": "horovod/tensorflow/gradient_aggregation_eager.py"}], "after_files": [{"content": "from distutils.version import LooseVersion\n\nimport tensorflow as tf\n\n_POST_TF_2_4_0 = LooseVersion(tf.__version__) >= LooseVersion('2.4.0')\n\n\nclass LocalGradientAggregationHelperEager:\n \"\"\"\n LocalGradientAggregationHelperEager aggregates gradient updates\n locally, and communicates the updates across machines only once\n every backward_passes_per_step. Only supports eager execution.\n \"\"\"\n\n def __init__(\n self,\n backward_passes_per_step,\n allreduce_func,\n sparse_as_dense,\n average_aggregated_gradients,\n ):\n self.allreduce_grads = allreduce_func\n self.sparse_as_dense = sparse_as_dense\n\n # backward_passes_per_step controls how often gradient updates are\n # synchronized.\n self.backward_passes_per_step = backward_passes_per_step\n if self.backward_passes_per_step <= 0:\n raise ValueError(\"backward_passes_per_step must be > 0\")\n\n # average_aggregated_gradients controls whether gradient updates that are\n # aggregated, should be divided by `backward_passes_per_step`.\n self.average_aggregated_gradients = average_aggregated_gradients\n\n # This is going to be N data structure holding the aggregated gradient updates\n # for parameter updates. N is the number of parameters.\n self.locally_aggregated_grads = {}\n\n # Used to know when to allreduce and apply gradients. We allreduce when `self.counter`\n # is equal to `self.backward_passes_per_step`. We apply gradients when `self.counter`\n # is equal to 0.\n self.counter = tf.Variable(initial_value=0)\n\n def compute_gradients(self, grads, vars):\n # On steps where allreduce happens, resulting_grads returns the allreduced\n # gradients, on other steps it returns the locally aggregated\n # gradients.\n resulting_grads = []\n\n for idx, grad in enumerate(grads):\n # Handle IndexedSlices.\n if self.sparse_as_dense and isinstance(grad, tf.IndexedSlices):\n grad = tf.convert_to_tensor(grad)\n elif isinstance(grad, tf.IndexedSlices):\n raise ValueError(\n \"IndexedSlices are not supported when \"\n \"`backward_passes_per_step` > 1 and \"\n \"`sparse_as_dense` is False.\"\n )\n\n # Create variables to store to aggregate gradients if they don't\n # already exist. Skip variables that are None.\n if idx not in self.locally_aggregated_grads.keys():\n if grad is not None:\n self.locally_aggregated_grads[idx] = tf.Variable(\n initial_value=tf.zeros_like(grad),\n trainable=False,\n dtype=grad.dtype,\n )\n\n if grad is None:\n resulting_grads.append(None)\n else:\n self.locally_aggregated_grads[idx].assign_add(grad)\n resulting_grads.append(\n self.locally_aggregated_grads[idx].read_value())\n assert len(self.locally_aggregated_grads) == len(grads)\n\n # Increment counter.\n self.counter.assign_add(1)\n\n def _all_reduce_and_clear_aggregated_variables(aggregated_gradients, vars):\n # Performs allreduce. If `average_aggregated_gradients` is\n # set to True divides result by `backward_passes_per_step`.\n reduced_gradients = self._allreduce_helper(aggregated_gradients, vars)\n assert len(reduced_gradients) == len(grads)\n\n self._clear_vars()\n return reduced_gradients\n\n def _do_nothing(aggregated_gradients):\n return aggregated_gradients\n\n resulting_grads = tf.cond(\n pred=tf.equal(self.counter, self.backward_passes_per_step),\n true_fn=lambda: _all_reduce_and_clear_aggregated_variables(resulting_grads, vars),\n false_fn=lambda: _do_nothing(resulting_grads),\n )\n\n return resulting_grads\n\n def _allreduce_helper(self, grads, vars):\n allreduced_grads = self.allreduce_grads(grads, vars)\n\n if not self.average_aggregated_gradients:\n return allreduced_grads\n\n scaled_allreduced_grads = []\n for grad in allreduced_grads:\n if grad is None:\n scaled_allreduced_grads.append(grad)\n continue\n\n scaled_allreduced_grads.append(\n grad / self.backward_passes_per_step)\n\n return scaled_allreduced_grads\n\n def _clear_vars(self):\n self.counter.assign(0)\n for idx in self.locally_aggregated_grads.keys():\n self.locally_aggregated_grads[idx].assign(\n tf.zeros_like(self.locally_aggregated_grads[idx]))\n\n def apply_gradients(self, apply_grads_closure, optimizer, *args, **kwargs):\n def increment_optimizer_iteration():\n if hasattr(optimizer, \"_iterations\") and optimizer._iterations is not None:\n return optimizer._iterations.assign_add(1).op\n return tf.no_op()\n\n def non_aggregation_step():\n if _POST_TF_2_4_0:\n # In TF 2.4+ `_aggregate_gradients()` is called from inside of `apply_gradients()`.\n # We account for this by calling `_aggregate_gradients()` for steps where we do\n # not call `apply_gradients()`.\n transformed_grads_and_vars = optimizer._transform_unaggregated_gradients(\n args[0])\n _ = optimizer._aggregate_gradients(transformed_grads_and_vars)\n\n return increment_optimizer_iteration()\n\n def is_aggregation_step():\n return tf.equal(self.counter, 0)\n\n return tf.cond(\n pred=is_aggregation_step(),\n true_fn=apply_grads_closure,\n false_fn=non_aggregation_step,\n )\n", "path": "horovod/tensorflow/gradient_aggregation_eager.py"}]} | 2,487 | 190 |
gh_patches_debug_35497 | rasdani/github-patches | git_diff | iterative__dvc-2765 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
lock: improve error message when .dvc/lock is taken
https://github.com/iterative/dvc/pull/2519#discussion_r326844119
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `dvc/lock.py`
Content:
```
1 """Manages dvc lock file."""
2 from __future__ import unicode_literals
3
4 import hashlib
5 import os
6 import time
7 from datetime import timedelta
8
9 from funcy.py3 import lkeep
10
11 from dvc.exceptions import DvcException
12 from dvc.utils import makedirs
13 from dvc.utils.compat import is_py3
14
15
16 DEFAULT_TIMEOUT = 5
17
18
19 class LockError(DvcException):
20 """Thrown when unable to acquire the lock for dvc repo."""
21
22
23 if is_py3:
24 import flufl.lock
25
26 class Lock(flufl.lock.Lock):
27 """Class for dvc repo lock.
28
29 Args:
30 lockfile (str): the lock filename
31 in.
32 tmp_dir (str): a directory to store claim files.
33 """
34
35 def __init__(self, lockfile, tmp_dir=None):
36 import socket
37
38 self._tmp_dir = tmp_dir
39 if self._tmp_dir is not None:
40 makedirs(self._tmp_dir, exist_ok=True)
41
42 # NOTE: this is basically Lock.__init__ copy-paste, except that
43 # instead of using `socket.getfqdn()` we use `socket.gethostname()`
44 # to speed this up. We've seen [1] `getfqdn()` take ~5sec to return
45 # anything, which is way too slow. `gethostname()` is actually a
46 # fallback for `getfqdn()` when it is not able to resolve a
47 # canonical hostname through network. The claimfile that uses
48 # `self._hostname` is still usable, as it uses `pid` and random
49 # number to generate the resulting lock file name, which is unique
50 # enough for our application.
51 #
52 # [1] https://github.com/iterative/dvc/issues/2582
53 self._hostname = socket.gethostname()
54
55 self._lockfile = lockfile
56 self._lifetime = timedelta(days=365) # Lock for good by default
57 self._separator = flufl.lock.SEP
58 self._set_claimfile()
59 self._owned = True
60 self._retry_errnos = []
61
62 @property
63 def lockfile(self):
64 return self._lockfile
65
66 @property
67 def files(self):
68 return lkeep([self._lockfile, self._tmp_dir])
69
70 def lock(self):
71 try:
72 super(Lock, self).lock(timedelta(seconds=DEFAULT_TIMEOUT))
73 except flufl.lock.TimeOutError:
74 raise LockError(
75 "cannot perform the cmd since DVC is busy and "
76 "locked. Please retry the cmd later."
77 )
78
79 def _set_claimfile(self, pid=None):
80 super(Lock, self)._set_claimfile(pid)
81
82 if self._tmp_dir is not None:
83 # Under Windows file path length is limited so we hash it
84 filename = hashlib.md5(self._claimfile.encode()).hexdigest()
85 self._claimfile = os.path.join(
86 self._tmp_dir, filename + ".lock"
87 )
88
89 # Fix for __del__ bug in flufl.lock [1] which is causing errors on
90 # Python shutdown [2].
91 # [1] https://gitlab.com/warsaw/flufl.lock/issues/7
92 # [2] https://github.com/iterative/dvc/issues/2573
93 def __del__(self):
94 try:
95 if self._owned:
96 self.finalize()
97 except ImportError:
98 pass
99
100
101 else:
102 import zc.lockfile
103
104 class Lock(object):
105 """Class for dvc repo lock.
106
107 Uses zc.lockfile as backend.
108 """
109
110 def __init__(self, lockfile, tmp_dir=None):
111 self.lockfile = lockfile
112 self._lock = None
113
114 @property
115 def files(self):
116 return [self.lockfile]
117
118 def _do_lock(self):
119 try:
120 self._lock = zc.lockfile.LockFile(self.lockfile)
121 except zc.lockfile.LockError:
122 raise LockError(
123 "cannot perform the cmd since DVC is busy and "
124 "locked. Please retry the cmd later."
125 )
126
127 def lock(self):
128 try:
129 self._do_lock()
130 return
131 except LockError:
132 time.sleep(DEFAULT_TIMEOUT)
133
134 self._do_lock()
135
136 def unlock(self):
137 self._lock.close()
138 self._lock = None
139
140 def __enter__(self):
141 self.lock()
142
143 def __exit__(self, typ, value, tbck):
144 self.unlock()
145
```
Path: `dvc/main.py`
Content:
```
1 """Main entry point for dvc CLI."""
2 from __future__ import unicode_literals
3
4 import logging
5
6 from dvc.analytics import Analytics
7 from dvc.cli import parse_args
8 from dvc.config import ConfigError
9 from dvc.exceptions import DvcParserError
10 from dvc.exceptions import NotDvcRepoError
11 from dvc.external_repo import clean_repos
12 from dvc.lock import LockError
13 from dvc.logger import FOOTER
14 from dvc.remote.pool import close_pools
15 from dvc.utils.compat import is_py2
16
17
18 # Workaround for CPython bug. See [1] and [2] for more info.
19 # [1] https://github.com/aws/aws-cli/blob/1.16.277/awscli/clidriver.py#L55
20 # [2] https://bugs.python.org/issue29288
21 "".encode("idna")
22
23
24 logger = logging.getLogger("dvc")
25
26
27 def main(argv=None):
28 """Run dvc CLI command.
29
30 Args:
31 argv: optional list of arguments to parse. sys.argv is used by default.
32
33 Returns:
34 int: command's return code.
35 """
36 args = None
37 cmd = None
38
39 outerLogLevel = logger.level
40 try:
41 args = parse_args(argv)
42
43 if args.quiet:
44 logger.setLevel(logging.CRITICAL)
45
46 elif args.verbose:
47 logger.setLevel(logging.DEBUG)
48
49 cmd = args.func(args)
50 ret = cmd.run()
51 except LockError:
52 logger.exception("failed to lock before running a command")
53 ret = 250
54 except ConfigError:
55 logger.exception("configuration error")
56 ret = 251
57 except KeyboardInterrupt:
58 logger.exception("interrupted by the user")
59 ret = 252
60 except NotDvcRepoError:
61 logger.exception("")
62 ret = 253
63 except DvcParserError:
64 ret = 254
65 except Exception as exc: # pylint: disable=broad-except
66 if isinstance(exc, UnicodeError) and is_py2:
67 logger.exception(
68 "unicode is not supported in DVC for Python 2 "
69 "(end-of-life January 1, 2020), please upgrade to Python 3"
70 )
71 else:
72 logger.exception("unexpected error")
73 ret = 255
74 finally:
75 logger.setLevel(outerLogLevel)
76
77 # Python 2 fails to close these clean occasionally and users see
78 # weird error messages, so we do it manually
79 close_pools()
80
81 # Remove cached repos in the end of the call, these are anonymous
82 # so won't be reused by any other subsequent run anyway.
83 clean_repos()
84
85 if ret != 0:
86 logger.info(FOOTER)
87
88 Analytics().send_cmd(cmd, args, ret)
89
90 return ret
91
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/dvc/lock.py b/dvc/lock.py
--- a/dvc/lock.py
+++ b/dvc/lock.py
@@ -15,6 +15,12 @@
DEFAULT_TIMEOUT = 5
+FAILED_TO_LOCK_MESSAGE = (
+ "cannot perform the command because another DVC process seems to be "
+ "running on this project. If that is not the case, manually remove "
+ "`.dvc/lock` and try again."
+)
+
class LockError(DvcException):
"""Thrown when unable to acquire the lock for dvc repo."""
@@ -71,10 +77,7 @@
try:
super(Lock, self).lock(timedelta(seconds=DEFAULT_TIMEOUT))
except flufl.lock.TimeOutError:
- raise LockError(
- "cannot perform the cmd since DVC is busy and "
- "locked. Please retry the cmd later."
- )
+ raise LockError(FAILED_TO_LOCK_MESSAGE)
def _set_claimfile(self, pid=None):
super(Lock, self)._set_claimfile(pid)
@@ -119,10 +122,7 @@
try:
self._lock = zc.lockfile.LockFile(self.lockfile)
except zc.lockfile.LockError:
- raise LockError(
- "cannot perform the cmd since DVC is busy and "
- "locked. Please retry the cmd later."
- )
+ raise LockError(FAILED_TO_LOCK_MESSAGE)
def lock(self):
try:
diff --git a/dvc/main.py b/dvc/main.py
--- a/dvc/main.py
+++ b/dvc/main.py
@@ -9,7 +9,6 @@
from dvc.exceptions import DvcParserError
from dvc.exceptions import NotDvcRepoError
from dvc.external_repo import clean_repos
-from dvc.lock import LockError
from dvc.logger import FOOTER
from dvc.remote.pool import close_pools
from dvc.utils.compat import is_py2
@@ -48,9 +47,6 @@
cmd = args.func(args)
ret = cmd.run()
- except LockError:
- logger.exception("failed to lock before running a command")
- ret = 250
except ConfigError:
logger.exception("configuration error")
ret = 251
| {"golden_diff": "diff --git a/dvc/lock.py b/dvc/lock.py\n--- a/dvc/lock.py\n+++ b/dvc/lock.py\n@@ -15,6 +15,12 @@\n \n DEFAULT_TIMEOUT = 5\n \n+FAILED_TO_LOCK_MESSAGE = (\n+ \"cannot perform the command because another DVC process seems to be \"\n+ \"running on this project. If that is not the case, manually remove \"\n+ \"`.dvc/lock` and try again.\"\n+)\n+\n \n class LockError(DvcException):\n \"\"\"Thrown when unable to acquire the lock for dvc repo.\"\"\"\n@@ -71,10 +77,7 @@\n try:\n super(Lock, self).lock(timedelta(seconds=DEFAULT_TIMEOUT))\n except flufl.lock.TimeOutError:\n- raise LockError(\n- \"cannot perform the cmd since DVC is busy and \"\n- \"locked. Please retry the cmd later.\"\n- )\n+ raise LockError(FAILED_TO_LOCK_MESSAGE)\n \n def _set_claimfile(self, pid=None):\n super(Lock, self)._set_claimfile(pid)\n@@ -119,10 +122,7 @@\n try:\n self._lock = zc.lockfile.LockFile(self.lockfile)\n except zc.lockfile.LockError:\n- raise LockError(\n- \"cannot perform the cmd since DVC is busy and \"\n- \"locked. Please retry the cmd later.\"\n- )\n+ raise LockError(FAILED_TO_LOCK_MESSAGE)\n \n def lock(self):\n try:\ndiff --git a/dvc/main.py b/dvc/main.py\n--- a/dvc/main.py\n+++ b/dvc/main.py\n@@ -9,7 +9,6 @@\n from dvc.exceptions import DvcParserError\n from dvc.exceptions import NotDvcRepoError\n from dvc.external_repo import clean_repos\n-from dvc.lock import LockError\n from dvc.logger import FOOTER\n from dvc.remote.pool import close_pools\n from dvc.utils.compat import is_py2\n@@ -48,9 +47,6 @@\n \n cmd = args.func(args)\n ret = cmd.run()\n- except LockError:\n- logger.exception(\"failed to lock before running a command\")\n- ret = 250\n except ConfigError:\n logger.exception(\"configuration error\")\n ret = 251\n", "issue": "lock: improve error message when .dvc/lock is taken\nhttps://github.com/iterative/dvc/pull/2519#discussion_r326844119\n", "before_files": [{"content": "\"\"\"Manages dvc lock file.\"\"\"\nfrom __future__ import unicode_literals\n\nimport hashlib\nimport os\nimport time\nfrom datetime import timedelta\n\nfrom funcy.py3 import lkeep\n\nfrom dvc.exceptions import DvcException\nfrom dvc.utils import makedirs\nfrom dvc.utils.compat import is_py3\n\n\nDEFAULT_TIMEOUT = 5\n\n\nclass LockError(DvcException):\n \"\"\"Thrown when unable to acquire the lock for dvc repo.\"\"\"\n\n\nif is_py3:\n import flufl.lock\n\n class Lock(flufl.lock.Lock):\n \"\"\"Class for dvc repo lock.\n\n Args:\n lockfile (str): the lock filename\n in.\n tmp_dir (str): a directory to store claim files.\n \"\"\"\n\n def __init__(self, lockfile, tmp_dir=None):\n import socket\n\n self._tmp_dir = tmp_dir\n if self._tmp_dir is not None:\n makedirs(self._tmp_dir, exist_ok=True)\n\n # NOTE: this is basically Lock.__init__ copy-paste, except that\n # instead of using `socket.getfqdn()` we use `socket.gethostname()`\n # to speed this up. We've seen [1] `getfqdn()` take ~5sec to return\n # anything, which is way too slow. `gethostname()` is actually a\n # fallback for `getfqdn()` when it is not able to resolve a\n # canonical hostname through network. The claimfile that uses\n # `self._hostname` is still usable, as it uses `pid` and random\n # number to generate the resulting lock file name, which is unique\n # enough for our application.\n #\n # [1] https://github.com/iterative/dvc/issues/2582\n self._hostname = socket.gethostname()\n\n self._lockfile = lockfile\n self._lifetime = timedelta(days=365) # Lock for good by default\n self._separator = flufl.lock.SEP\n self._set_claimfile()\n self._owned = True\n self._retry_errnos = []\n\n @property\n def lockfile(self):\n return self._lockfile\n\n @property\n def files(self):\n return lkeep([self._lockfile, self._tmp_dir])\n\n def lock(self):\n try:\n super(Lock, self).lock(timedelta(seconds=DEFAULT_TIMEOUT))\n except flufl.lock.TimeOutError:\n raise LockError(\n \"cannot perform the cmd since DVC is busy and \"\n \"locked. Please retry the cmd later.\"\n )\n\n def _set_claimfile(self, pid=None):\n super(Lock, self)._set_claimfile(pid)\n\n if self._tmp_dir is not None:\n # Under Windows file path length is limited so we hash it\n filename = hashlib.md5(self._claimfile.encode()).hexdigest()\n self._claimfile = os.path.join(\n self._tmp_dir, filename + \".lock\"\n )\n\n # Fix for __del__ bug in flufl.lock [1] which is causing errors on\n # Python shutdown [2].\n # [1] https://gitlab.com/warsaw/flufl.lock/issues/7\n # [2] https://github.com/iterative/dvc/issues/2573\n def __del__(self):\n try:\n if self._owned:\n self.finalize()\n except ImportError:\n pass\n\n\nelse:\n import zc.lockfile\n\n class Lock(object):\n \"\"\"Class for dvc repo lock.\n\n Uses zc.lockfile as backend.\n \"\"\"\n\n def __init__(self, lockfile, tmp_dir=None):\n self.lockfile = lockfile\n self._lock = None\n\n @property\n def files(self):\n return [self.lockfile]\n\n def _do_lock(self):\n try:\n self._lock = zc.lockfile.LockFile(self.lockfile)\n except zc.lockfile.LockError:\n raise LockError(\n \"cannot perform the cmd since DVC is busy and \"\n \"locked. Please retry the cmd later.\"\n )\n\n def lock(self):\n try:\n self._do_lock()\n return\n except LockError:\n time.sleep(DEFAULT_TIMEOUT)\n\n self._do_lock()\n\n def unlock(self):\n self._lock.close()\n self._lock = None\n\n def __enter__(self):\n self.lock()\n\n def __exit__(self, typ, value, tbck):\n self.unlock()\n", "path": "dvc/lock.py"}, {"content": "\"\"\"Main entry point for dvc CLI.\"\"\"\nfrom __future__ import unicode_literals\n\nimport logging\n\nfrom dvc.analytics import Analytics\nfrom dvc.cli import parse_args\nfrom dvc.config import ConfigError\nfrom dvc.exceptions import DvcParserError\nfrom dvc.exceptions import NotDvcRepoError\nfrom dvc.external_repo import clean_repos\nfrom dvc.lock import LockError\nfrom dvc.logger import FOOTER\nfrom dvc.remote.pool import close_pools\nfrom dvc.utils.compat import is_py2\n\n\n# Workaround for CPython bug. See [1] and [2] for more info.\n# [1] https://github.com/aws/aws-cli/blob/1.16.277/awscli/clidriver.py#L55\n# [2] https://bugs.python.org/issue29288\n\"\".encode(\"idna\")\n\n\nlogger = logging.getLogger(\"dvc\")\n\n\ndef main(argv=None):\n \"\"\"Run dvc CLI command.\n\n Args:\n argv: optional list of arguments to parse. sys.argv is used by default.\n\n Returns:\n int: command's return code.\n \"\"\"\n args = None\n cmd = None\n\n outerLogLevel = logger.level\n try:\n args = parse_args(argv)\n\n if args.quiet:\n logger.setLevel(logging.CRITICAL)\n\n elif args.verbose:\n logger.setLevel(logging.DEBUG)\n\n cmd = args.func(args)\n ret = cmd.run()\n except LockError:\n logger.exception(\"failed to lock before running a command\")\n ret = 250\n except ConfigError:\n logger.exception(\"configuration error\")\n ret = 251\n except KeyboardInterrupt:\n logger.exception(\"interrupted by the user\")\n ret = 252\n except NotDvcRepoError:\n logger.exception(\"\")\n ret = 253\n except DvcParserError:\n ret = 254\n except Exception as exc: # pylint: disable=broad-except\n if isinstance(exc, UnicodeError) and is_py2:\n logger.exception(\n \"unicode is not supported in DVC for Python 2 \"\n \"(end-of-life January 1, 2020), please upgrade to Python 3\"\n )\n else:\n logger.exception(\"unexpected error\")\n ret = 255\n finally:\n logger.setLevel(outerLogLevel)\n\n # Python 2 fails to close these clean occasionally and users see\n # weird error messages, so we do it manually\n close_pools()\n\n # Remove cached repos in the end of the call, these are anonymous\n # so won't be reused by any other subsequent run anyway.\n clean_repos()\n\n if ret != 0:\n logger.info(FOOTER)\n\n Analytics().send_cmd(cmd, args, ret)\n\n return ret\n", "path": "dvc/main.py"}], "after_files": [{"content": "\"\"\"Manages dvc lock file.\"\"\"\nfrom __future__ import unicode_literals\n\nimport hashlib\nimport os\nimport time\nfrom datetime import timedelta\n\nfrom funcy.py3 import lkeep\n\nfrom dvc.exceptions import DvcException\nfrom dvc.utils import makedirs\nfrom dvc.utils.compat import is_py3\n\n\nDEFAULT_TIMEOUT = 5\n\nFAILED_TO_LOCK_MESSAGE = (\n \"cannot perform the command because another DVC process seems to be \"\n \"running on this project. If that is not the case, manually remove \"\n \"`.dvc/lock` and try again.\"\n)\n\n\nclass LockError(DvcException):\n \"\"\"Thrown when unable to acquire the lock for dvc repo.\"\"\"\n\n\nif is_py3:\n import flufl.lock\n\n class Lock(flufl.lock.Lock):\n \"\"\"Class for dvc repo lock.\n\n Args:\n lockfile (str): the lock filename\n in.\n tmp_dir (str): a directory to store claim files.\n \"\"\"\n\n def __init__(self, lockfile, tmp_dir=None):\n import socket\n\n self._tmp_dir = tmp_dir\n if self._tmp_dir is not None:\n makedirs(self._tmp_dir, exist_ok=True)\n\n # NOTE: this is basically Lock.__init__ copy-paste, except that\n # instead of using `socket.getfqdn()` we use `socket.gethostname()`\n # to speed this up. We've seen [1] `getfqdn()` take ~5sec to return\n # anything, which is way too slow. `gethostname()` is actually a\n # fallback for `getfqdn()` when it is not able to resolve a\n # canonical hostname through network. The claimfile that uses\n # `self._hostname` is still usable, as it uses `pid` and random\n # number to generate the resulting lock file name, which is unique\n # enough for our application.\n #\n # [1] https://github.com/iterative/dvc/issues/2582\n self._hostname = socket.gethostname()\n\n self._lockfile = lockfile\n self._lifetime = timedelta(days=365) # Lock for good by default\n self._separator = flufl.lock.SEP\n self._set_claimfile()\n self._owned = True\n self._retry_errnos = []\n\n @property\n def lockfile(self):\n return self._lockfile\n\n @property\n def files(self):\n return lkeep([self._lockfile, self._tmp_dir])\n\n def lock(self):\n try:\n super(Lock, self).lock(timedelta(seconds=DEFAULT_TIMEOUT))\n except flufl.lock.TimeOutError:\n raise LockError(FAILED_TO_LOCK_MESSAGE)\n\n def _set_claimfile(self, pid=None):\n super(Lock, self)._set_claimfile(pid)\n\n if self._tmp_dir is not None:\n # Under Windows file path length is limited so we hash it\n filename = hashlib.md5(self._claimfile.encode()).hexdigest()\n self._claimfile = os.path.join(\n self._tmp_dir, filename + \".lock\"\n )\n\n # Fix for __del__ bug in flufl.lock [1] which is causing errors on\n # Python shutdown [2].\n # [1] https://gitlab.com/warsaw/flufl.lock/issues/7\n # [2] https://github.com/iterative/dvc/issues/2573\n def __del__(self):\n try:\n if self._owned:\n self.finalize()\n except ImportError:\n pass\n\n\nelse:\n import zc.lockfile\n\n class Lock(object):\n \"\"\"Class for dvc repo lock.\n\n Uses zc.lockfile as backend.\n \"\"\"\n\n def __init__(self, lockfile, tmp_dir=None):\n self.lockfile = lockfile\n self._lock = None\n\n @property\n def files(self):\n return [self.lockfile]\n\n def _do_lock(self):\n try:\n self._lock = zc.lockfile.LockFile(self.lockfile)\n except zc.lockfile.LockError:\n raise LockError(FAILED_TO_LOCK_MESSAGE)\n\n def lock(self):\n try:\n self._do_lock()\n return\n except LockError:\n time.sleep(DEFAULT_TIMEOUT)\n\n self._do_lock()\n\n def unlock(self):\n self._lock.close()\n self._lock = None\n\n def __enter__(self):\n self.lock()\n\n def __exit__(self, typ, value, tbck):\n self.unlock()\n", "path": "dvc/lock.py"}, {"content": "\"\"\"Main entry point for dvc CLI.\"\"\"\nfrom __future__ import unicode_literals\n\nimport logging\n\nfrom dvc.analytics import Analytics\nfrom dvc.cli import parse_args\nfrom dvc.config import ConfigError\nfrom dvc.exceptions import DvcParserError\nfrom dvc.exceptions import NotDvcRepoError\nfrom dvc.external_repo import clean_repos\nfrom dvc.logger import FOOTER\nfrom dvc.remote.pool import close_pools\nfrom dvc.utils.compat import is_py2\n\n\n# Workaround for CPython bug. See [1] and [2] for more info.\n# [1] https://github.com/aws/aws-cli/blob/1.16.277/awscli/clidriver.py#L55\n# [2] https://bugs.python.org/issue29288\n\"\".encode(\"idna\")\n\n\nlogger = logging.getLogger(\"dvc\")\n\n\ndef main(argv=None):\n \"\"\"Run dvc CLI command.\n\n Args:\n argv: optional list of arguments to parse. sys.argv is used by default.\n\n Returns:\n int: command's return code.\n \"\"\"\n args = None\n cmd = None\n\n outerLogLevel = logger.level\n try:\n args = parse_args(argv)\n\n if args.quiet:\n logger.setLevel(logging.CRITICAL)\n\n elif args.verbose:\n logger.setLevel(logging.DEBUG)\n\n cmd = args.func(args)\n ret = cmd.run()\n except ConfigError:\n logger.exception(\"configuration error\")\n ret = 251\n except KeyboardInterrupt:\n logger.exception(\"interrupted by the user\")\n ret = 252\n except NotDvcRepoError:\n logger.exception(\"\")\n ret = 253\n except DvcParserError:\n ret = 254\n except Exception as exc: # pylint: disable=broad-except\n if isinstance(exc, UnicodeError) and is_py2:\n logger.exception(\n \"unicode is not supported in DVC for Python 2 \"\n \"(end-of-life January 1, 2020), please upgrade to Python 3\"\n )\n else:\n logger.exception(\"unexpected error\")\n ret = 255\n finally:\n logger.setLevel(outerLogLevel)\n\n # Python 2 fails to close these clean occasionally and users see\n # weird error messages, so we do it manually\n close_pools()\n\n # Remove cached repos in the end of the call, these are anonymous\n # so won't be reused by any other subsequent run anyway.\n clean_repos()\n\n if ret != 0:\n logger.info(FOOTER)\n\n Analytics().send_cmd(cmd, args, ret)\n\n return ret\n", "path": "dvc/main.py"}]} | 2,429 | 516 |
gh_patches_debug_38054 | rasdani/github-patches | git_diff | translate__pootle-4613 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
TP creation email should use BCC
Let's BCC by default when Pootle is communicating to lists of people. When you create a TP all relevant people are listed in the To field.
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `pootle/core/mail.py`
Content:
```
1 # -*- coding: utf-8 -*-
2 #
3 # Copyright (C) Pootle contributors.
4 #
5 # This file is a part of the Pootle project. It is distributed under the GPL3
6 # or later license. See the LICENSE file for a copy of the license and the
7 # AUTHORS file for copyright and authorship information.
8
9 from django.core.mail import EmailMultiAlternatives, get_connection
10
11
12 def send_mail(subject, message, from_email, recipient_list,
13 fail_silently=False, auth_user=None, auth_password=None,
14 connection=None, html_message=None, headers=None):
15 """Override django send_mail function to allow use of custom email headers.
16 """
17
18 connection = connection or get_connection(username=auth_user,
19 password=auth_password,
20 fail_silently=fail_silently)
21
22 mail = EmailMultiAlternatives(subject, message,
23 from_email, recipient_list,
24 connection=connection, headers=headers)
25
26 if html_message:
27 mail.attach_alternative(html_message, 'text/html')
28
29 return mail.send()
30
```
Path: `pootle/apps/pootle_translationproject/receivers.py`
Content:
```
1 #!/usr/bin/env python
2 # -*- coding: utf-8 -*-
3 #
4 # Copyright (C) Pootle contributors.
5 #
6 # This file is a part of the Pootle project. It is distributed under the GPL3
7 # or later license. See the LICENSE file for a copy of the license and the
8 # AUTHORS file for copyright and authorship information.
9
10 from django.contrib.auth import get_user_model
11 from django.core.mail import send_mail
12 from django.db.models import Q
13 from django.dispatch import receiver
14 from django.template.loader import render_to_string
15 from django.utils.translation import ugettext_lazy as _
16
17 from pootle.core.url_helpers import urljoin
18
19 from .models import TranslationProject
20 from .signals import tp_init_failed_async, tp_inited_async
21
22
23 def get_recipients(project):
24 User = get_user_model()
25 return list(set(User.objects.filter(
26 Q(permissionset__positive_permissions__codename="administrate",
27 permissionset__directory__pootle_path=project.pootle_path) |
28 Q(is_superuser=True)).values_list("email", flat=True)))
29
30
31 @receiver(tp_inited_async, sender=TranslationProject)
32 def tp_inited_async(instance, response_url, **kwargs):
33 ctx = {"tp": instance,
34 "url": urljoin(response_url, instance.get_absolute_url())}
35 message = render_to_string(
36 'projects/admin/email/translation_project_created.txt', ctx)
37 subject = _(u"Translation project (%s) created" % instance)
38 recipients = get_recipients(instance.project)
39 send_mail(subject, message, from_email=None,
40 recipient_list=recipients, fail_silently=True)
41
42
43 @receiver(tp_init_failed_async, sender=TranslationProject)
44 def tp_init_failed_async(instance, **kwargs):
45 ctx = {"tp": instance}
46 message = render_to_string(
47 'projects/admin/email/translation_project_creation_failed.txt', ctx)
48 subject = _(u"Translation project (%s) creation failed" % instance)
49 recipients = get_recipients(instance.project)
50 send_mail(subject, message, from_email=None,
51 recipient_list=recipients, fail_silently=True)
52
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/pootle/apps/pootle_translationproject/receivers.py b/pootle/apps/pootle_translationproject/receivers.py
--- a/pootle/apps/pootle_translationproject/receivers.py
+++ b/pootle/apps/pootle_translationproject/receivers.py
@@ -8,12 +8,12 @@
# AUTHORS file for copyright and authorship information.
from django.contrib.auth import get_user_model
-from django.core.mail import send_mail
from django.db.models import Q
from django.dispatch import receiver
from django.template.loader import render_to_string
from django.utils.translation import ugettext_lazy as _
+from pootle.core.mail import send_mail
from pootle.core.url_helpers import urljoin
from .models import TranslationProject
@@ -37,7 +37,7 @@
subject = _(u"Translation project (%s) created" % instance)
recipients = get_recipients(instance.project)
send_mail(subject, message, from_email=None,
- recipient_list=recipients, fail_silently=True)
+ recipient_list=[], fail_silently=True, bcc=recipients)
@receiver(tp_init_failed_async, sender=TranslationProject)
@@ -48,4 +48,4 @@
subject = _(u"Translation project (%s) creation failed" % instance)
recipients = get_recipients(instance.project)
send_mail(subject, message, from_email=None,
- recipient_list=recipients, fail_silently=True)
+ recipient_list=[], fail_silently=True, bcc=recipients)
diff --git a/pootle/core/mail.py b/pootle/core/mail.py
--- a/pootle/core/mail.py
+++ b/pootle/core/mail.py
@@ -11,7 +11,8 @@
def send_mail(subject, message, from_email, recipient_list,
fail_silently=False, auth_user=None, auth_password=None,
- connection=None, html_message=None, headers=None):
+ connection=None, html_message=None, headers=None,
+ cc=None, bcc=None):
"""Override django send_mail function to allow use of custom email headers.
"""
@@ -21,7 +22,8 @@
mail = EmailMultiAlternatives(subject, message,
from_email, recipient_list,
- connection=connection, headers=headers)
+ connection=connection, headers=headers,
+ cc=cc, bcc=bcc)
if html_message:
mail.attach_alternative(html_message, 'text/html')
| {"golden_diff": "diff --git a/pootle/apps/pootle_translationproject/receivers.py b/pootle/apps/pootle_translationproject/receivers.py\n--- a/pootle/apps/pootle_translationproject/receivers.py\n+++ b/pootle/apps/pootle_translationproject/receivers.py\n@@ -8,12 +8,12 @@\n # AUTHORS file for copyright and authorship information.\n \n from django.contrib.auth import get_user_model\n-from django.core.mail import send_mail\n from django.db.models import Q\n from django.dispatch import receiver\n from django.template.loader import render_to_string\n from django.utils.translation import ugettext_lazy as _\n \n+from pootle.core.mail import send_mail\n from pootle.core.url_helpers import urljoin\n \n from .models import TranslationProject\n@@ -37,7 +37,7 @@\n subject = _(u\"Translation project (%s) created\" % instance)\n recipients = get_recipients(instance.project)\n send_mail(subject, message, from_email=None,\n- recipient_list=recipients, fail_silently=True)\n+ recipient_list=[], fail_silently=True, bcc=recipients)\n \n \n @receiver(tp_init_failed_async, sender=TranslationProject)\n@@ -48,4 +48,4 @@\n subject = _(u\"Translation project (%s) creation failed\" % instance)\n recipients = get_recipients(instance.project)\n send_mail(subject, message, from_email=None,\n- recipient_list=recipients, fail_silently=True)\n+ recipient_list=[], fail_silently=True, bcc=recipients)\ndiff --git a/pootle/core/mail.py b/pootle/core/mail.py\n--- a/pootle/core/mail.py\n+++ b/pootle/core/mail.py\n@@ -11,7 +11,8 @@\n \n def send_mail(subject, message, from_email, recipient_list,\n fail_silently=False, auth_user=None, auth_password=None,\n- connection=None, html_message=None, headers=None):\n+ connection=None, html_message=None, headers=None,\n+ cc=None, bcc=None):\n \"\"\"Override django send_mail function to allow use of custom email headers.\n \"\"\"\n \n@@ -21,7 +22,8 @@\n \n mail = EmailMultiAlternatives(subject, message,\n from_email, recipient_list,\n- connection=connection, headers=headers)\n+ connection=connection, headers=headers,\n+ cc=cc, bcc=bcc)\n \n if html_message:\n mail.attach_alternative(html_message, 'text/html')\n", "issue": "TP creation email should use BCC\nLet's BCC by default when Pootle is communicating to lists of people. When you create a TP all relevant people are listed in the To field.\n\n", "before_files": [{"content": "# -*- coding: utf-8 -*-\n#\n# Copyright (C) Pootle contributors.\n#\n# This file is a part of the Pootle project. It is distributed under the GPL3\n# or later license. See the LICENSE file for a copy of the license and the\n# AUTHORS file for copyright and authorship information.\n\nfrom django.core.mail import EmailMultiAlternatives, get_connection\n\n\ndef send_mail(subject, message, from_email, recipient_list,\n fail_silently=False, auth_user=None, auth_password=None,\n connection=None, html_message=None, headers=None):\n \"\"\"Override django send_mail function to allow use of custom email headers.\n \"\"\"\n\n connection = connection or get_connection(username=auth_user,\n password=auth_password,\n fail_silently=fail_silently)\n\n mail = EmailMultiAlternatives(subject, message,\n from_email, recipient_list,\n connection=connection, headers=headers)\n\n if html_message:\n mail.attach_alternative(html_message, 'text/html')\n\n return mail.send()\n", "path": "pootle/core/mail.py"}, {"content": "#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n#\n# Copyright (C) Pootle contributors.\n#\n# This file is a part of the Pootle project. It is distributed under the GPL3\n# or later license. See the LICENSE file for a copy of the license and the\n# AUTHORS file for copyright and authorship information.\n\nfrom django.contrib.auth import get_user_model\nfrom django.core.mail import send_mail\nfrom django.db.models import Q\nfrom django.dispatch import receiver\nfrom django.template.loader import render_to_string\nfrom django.utils.translation import ugettext_lazy as _\n\nfrom pootle.core.url_helpers import urljoin\n\nfrom .models import TranslationProject\nfrom .signals import tp_init_failed_async, tp_inited_async\n\n\ndef get_recipients(project):\n User = get_user_model()\n return list(set(User.objects.filter(\n Q(permissionset__positive_permissions__codename=\"administrate\",\n permissionset__directory__pootle_path=project.pootle_path) |\n Q(is_superuser=True)).values_list(\"email\", flat=True)))\n\n\n@receiver(tp_inited_async, sender=TranslationProject)\ndef tp_inited_async(instance, response_url, **kwargs):\n ctx = {\"tp\": instance,\n \"url\": urljoin(response_url, instance.get_absolute_url())}\n message = render_to_string(\n 'projects/admin/email/translation_project_created.txt', ctx)\n subject = _(u\"Translation project (%s) created\" % instance)\n recipients = get_recipients(instance.project)\n send_mail(subject, message, from_email=None,\n recipient_list=recipients, fail_silently=True)\n\n\n@receiver(tp_init_failed_async, sender=TranslationProject)\ndef tp_init_failed_async(instance, **kwargs):\n ctx = {\"tp\": instance}\n message = render_to_string(\n 'projects/admin/email/translation_project_creation_failed.txt', ctx)\n subject = _(u\"Translation project (%s) creation failed\" % instance)\n recipients = get_recipients(instance.project)\n send_mail(subject, message, from_email=None,\n recipient_list=recipients, fail_silently=True)\n", "path": "pootle/apps/pootle_translationproject/receivers.py"}], "after_files": [{"content": "# -*- coding: utf-8 -*-\n#\n# Copyright (C) Pootle contributors.\n#\n# This file is a part of the Pootle project. It is distributed under the GPL3\n# or later license. See the LICENSE file for a copy of the license and the\n# AUTHORS file for copyright and authorship information.\n\nfrom django.core.mail import EmailMultiAlternatives, get_connection\n\n\ndef send_mail(subject, message, from_email, recipient_list,\n fail_silently=False, auth_user=None, auth_password=None,\n connection=None, html_message=None, headers=None,\n cc=None, bcc=None):\n \"\"\"Override django send_mail function to allow use of custom email headers.\n \"\"\"\n\n connection = connection or get_connection(username=auth_user,\n password=auth_password,\n fail_silently=fail_silently)\n\n mail = EmailMultiAlternatives(subject, message,\n from_email, recipient_list,\n connection=connection, headers=headers,\n cc=cc, bcc=bcc)\n\n if html_message:\n mail.attach_alternative(html_message, 'text/html')\n\n return mail.send()\n", "path": "pootle/core/mail.py"}, {"content": "#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n#\n# Copyright (C) Pootle contributors.\n#\n# This file is a part of the Pootle project. It is distributed under the GPL3\n# or later license. See the LICENSE file for a copy of the license and the\n# AUTHORS file for copyright and authorship information.\n\nfrom django.contrib.auth import get_user_model\nfrom django.db.models import Q\nfrom django.dispatch import receiver\nfrom django.template.loader import render_to_string\nfrom django.utils.translation import ugettext_lazy as _\n\nfrom pootle.core.mail import send_mail\nfrom pootle.core.url_helpers import urljoin\n\nfrom .models import TranslationProject\nfrom .signals import tp_init_failed_async, tp_inited_async\n\n\ndef get_recipients(project):\n User = get_user_model()\n return list(set(User.objects.filter(\n Q(permissionset__positive_permissions__codename=\"administrate\",\n permissionset__directory__pootle_path=project.pootle_path) |\n Q(is_superuser=True)).values_list(\"email\", flat=True)))\n\n\n@receiver(tp_inited_async, sender=TranslationProject)\ndef tp_inited_async(instance, response_url, **kwargs):\n ctx = {\"tp\": instance,\n \"url\": urljoin(response_url, instance.get_absolute_url())}\n message = render_to_string(\n 'projects/admin/email/translation_project_created.txt', ctx)\n subject = _(u\"Translation project (%s) created\" % instance)\n recipients = get_recipients(instance.project)\n send_mail(subject, message, from_email=None,\n recipient_list=[], fail_silently=True, bcc=recipients)\n\n\n@receiver(tp_init_failed_async, sender=TranslationProject)\ndef tp_init_failed_async(instance, **kwargs):\n ctx = {\"tp\": instance}\n message = render_to_string(\n 'projects/admin/email/translation_project_creation_failed.txt', ctx)\n subject = _(u\"Translation project (%s) creation failed\" % instance)\n recipients = get_recipients(instance.project)\n send_mail(subject, message, from_email=None,\n recipient_list=[], fail_silently=True, bcc=recipients)\n", "path": "pootle/apps/pootle_translationproject/receivers.py"}]} | 1,144 | 544 |
gh_patches_debug_8480 | rasdani/github-patches | git_diff | elastic__apm-agent-python-580 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
DB interactions not traced when using context manager with psycopg2 connections or cursors
When using a context manager with psycopg2 connections or cursors, db interactions are not captured in spans.
The code below generates a span for `psycopg2.connect`, but not the query:
```
with psycopg2.connect(DSN) as conn:
with conn.cursor() as curs:
curs.execute("SELECT * FROM data.portfolio;")
portfolios = curs.fetchall()
```
whereas the following captures both spans as expected:
```
conn = psycopg2.connect(DSN)
curs = conn.cursor()
curs.execute("SELECT * FROM data.portfolio;")
portfolios = curs.fetchall()
curs.close()
conn.close()
```
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `elasticapm/instrumentation/packages/psycopg2.py`
Content:
```
1 # BSD 3-Clause License
2 #
3 # Copyright (c) 2019, Elasticsearch BV
4 # All rights reserved.
5 #
6 # Redistribution and use in source and binary forms, with or without
7 # modification, are permitted provided that the following conditions are met:
8 #
9 # * Redistributions of source code must retain the above copyright notice, this
10 # list of conditions and the following disclaimer.
11 #
12 # * Redistributions in binary form must reproduce the above copyright notice,
13 # this list of conditions and the following disclaimer in the documentation
14 # and/or other materials provided with the distribution.
15 #
16 # * Neither the name of the copyright holder nor the names of its
17 # contributors may be used to endorse or promote products derived from
18 # this software without specific prior written permission.
19 #
20 # THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
21 # AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
22 # IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
23 # DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE
24 # FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
25 # DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
26 # SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER
27 # CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
28 # OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
29 # OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
30
31 from elasticapm.instrumentation.packages.dbapi2 import (
32 ConnectionProxy,
33 CursorProxy,
34 DbApi2Instrumentation,
35 extract_signature,
36 )
37 from elasticapm.traces import capture_span
38 from elasticapm.utils import default_ports
39
40
41 class PGCursorProxy(CursorProxy):
42 provider_name = "postgresql"
43
44 def _bake_sql(self, sql):
45 # if this is a Composable object, use its `as_string` method
46 # see http://initd.org/psycopg/docs/sql.html
47 if hasattr(sql, "as_string"):
48 return sql.as_string(self.__wrapped__)
49 return sql
50
51 def extract_signature(self, sql):
52 return extract_signature(sql)
53
54
55 class PGConnectionProxy(ConnectionProxy):
56 cursor_proxy = PGCursorProxy
57
58
59 class Psycopg2Instrumentation(DbApi2Instrumentation):
60 name = "psycopg2"
61
62 instrument_list = [("psycopg2", "connect")]
63
64 def call(self, module, method, wrapped, instance, args, kwargs):
65 signature = "psycopg2.connect"
66
67 host = kwargs.get("host")
68 if host:
69 signature += " " + str(host)
70
71 port = kwargs.get("port")
72 if port:
73 port = str(port)
74 if int(port) != default_ports.get("postgresql"):
75 signature += ":" + port
76 else:
77 # Parse connection string and extract host/port
78 pass
79
80 with capture_span(signature, span_type="db", span_subtype="postgresql", span_action="connect"):
81 return PGConnectionProxy(wrapped(*args, **kwargs))
82
83
84 class Psycopg2RegisterTypeInstrumentation(DbApi2Instrumentation):
85 name = "psycopg2-register-type"
86
87 instrument_list = [
88 ("psycopg2.extensions", "register_type"),
89 # specifically instrument `register_json` as it bypasses `register_type`
90 ("psycopg2._json", "register_json"),
91 ]
92
93 def call(self, module, method, wrapped, instance, args, kwargs):
94 if "conn_or_curs" in kwargs and hasattr(kwargs["conn_or_curs"], "__wrapped__"):
95 kwargs["conn_or_curs"] = kwargs["conn_or_curs"].__wrapped__
96 # register_type takes the connection as second argument
97 elif len(args) == 2 and hasattr(args[1], "__wrapped__"):
98 args = (args[0], args[1].__wrapped__)
99 # register_json takes the connection as first argument, and can have
100 # several more arguments
101 elif method == "register_json":
102 if args and hasattr(args[0], "__wrapped__"):
103 args = (args[0].__wrapped__,) + args[1:]
104
105 return wrapped(*args, **kwargs)
106
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/elasticapm/instrumentation/packages/psycopg2.py b/elasticapm/instrumentation/packages/psycopg2.py
--- a/elasticapm/instrumentation/packages/psycopg2.py
+++ b/elasticapm/instrumentation/packages/psycopg2.py
@@ -51,10 +51,16 @@
def extract_signature(self, sql):
return extract_signature(sql)
+ def __enter__(self):
+ return PGCursorProxy(self.__wrapped__.__enter__())
+
class PGConnectionProxy(ConnectionProxy):
cursor_proxy = PGCursorProxy
+ def __enter__(self):
+ return PGConnectionProxy(self.__wrapped__.__enter__())
+
class Psycopg2Instrumentation(DbApi2Instrumentation):
name = "psycopg2"
| {"golden_diff": "diff --git a/elasticapm/instrumentation/packages/psycopg2.py b/elasticapm/instrumentation/packages/psycopg2.py\n--- a/elasticapm/instrumentation/packages/psycopg2.py\n+++ b/elasticapm/instrumentation/packages/psycopg2.py\n@@ -51,10 +51,16 @@\n def extract_signature(self, sql):\n return extract_signature(sql)\n \n+ def __enter__(self):\n+ return PGCursorProxy(self.__wrapped__.__enter__())\n+\n \n class PGConnectionProxy(ConnectionProxy):\n cursor_proxy = PGCursorProxy\n \n+ def __enter__(self):\n+ return PGConnectionProxy(self.__wrapped__.__enter__())\n+\n \n class Psycopg2Instrumentation(DbApi2Instrumentation):\n name = \"psycopg2\"\n", "issue": "DB interactions not traced when using context manager with psycopg2 connections or cursors\nWhen using a context manager with psycopg2 connections or cursors, db interactions are not captured in spans.\r\n\r\nThe code below generates a span for `psycopg2.connect`, but not the query:\r\n```\r\nwith psycopg2.connect(DSN) as conn:\r\n with conn.cursor() as curs:\r\n curs.execute(\"SELECT * FROM data.portfolio;\")\r\n portfolios = curs.fetchall()\r\n```\r\n\r\nwhereas the following captures both spans as expected:\r\n```\r\nconn = psycopg2.connect(DSN)\r\ncurs = conn.cursor()\r\ncurs.execute(\"SELECT * FROM data.portfolio;\")\r\nportfolios = curs.fetchall()\r\ncurs.close()\r\nconn.close()\r\n```\n", "before_files": [{"content": "# BSD 3-Clause License\n#\n# Copyright (c) 2019, Elasticsearch BV\n# All rights reserved.\n#\n# Redistribution and use in source and binary forms, with or without\n# modification, are permitted provided that the following conditions are met:\n#\n# * Redistributions of source code must retain the above copyright notice, this\n# list of conditions and the following disclaimer.\n#\n# * Redistributions in binary form must reproduce the above copyright notice,\n# this list of conditions and the following disclaimer in the documentation\n# and/or other materials provided with the distribution.\n#\n# * Neither the name of the copyright holder nor the names of its\n# contributors may be used to endorse or promote products derived from\n# this software without specific prior written permission.\n#\n# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS \"AS IS\"\n# AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE\n# IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE\n# DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE\n# FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL\n# DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR\n# SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER\n# CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,\n# OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE\n# OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.\n\nfrom elasticapm.instrumentation.packages.dbapi2 import (\n ConnectionProxy,\n CursorProxy,\n DbApi2Instrumentation,\n extract_signature,\n)\nfrom elasticapm.traces import capture_span\nfrom elasticapm.utils import default_ports\n\n\nclass PGCursorProxy(CursorProxy):\n provider_name = \"postgresql\"\n\n def _bake_sql(self, sql):\n # if this is a Composable object, use its `as_string` method\n # see http://initd.org/psycopg/docs/sql.html\n if hasattr(sql, \"as_string\"):\n return sql.as_string(self.__wrapped__)\n return sql\n\n def extract_signature(self, sql):\n return extract_signature(sql)\n\n\nclass PGConnectionProxy(ConnectionProxy):\n cursor_proxy = PGCursorProxy\n\n\nclass Psycopg2Instrumentation(DbApi2Instrumentation):\n name = \"psycopg2\"\n\n instrument_list = [(\"psycopg2\", \"connect\")]\n\n def call(self, module, method, wrapped, instance, args, kwargs):\n signature = \"psycopg2.connect\"\n\n host = kwargs.get(\"host\")\n if host:\n signature += \" \" + str(host)\n\n port = kwargs.get(\"port\")\n if port:\n port = str(port)\n if int(port) != default_ports.get(\"postgresql\"):\n signature += \":\" + port\n else:\n # Parse connection string and extract host/port\n pass\n\n with capture_span(signature, span_type=\"db\", span_subtype=\"postgresql\", span_action=\"connect\"):\n return PGConnectionProxy(wrapped(*args, **kwargs))\n\n\nclass Psycopg2RegisterTypeInstrumentation(DbApi2Instrumentation):\n name = \"psycopg2-register-type\"\n\n instrument_list = [\n (\"psycopg2.extensions\", \"register_type\"),\n # specifically instrument `register_json` as it bypasses `register_type`\n (\"psycopg2._json\", \"register_json\"),\n ]\n\n def call(self, module, method, wrapped, instance, args, kwargs):\n if \"conn_or_curs\" in kwargs and hasattr(kwargs[\"conn_or_curs\"], \"__wrapped__\"):\n kwargs[\"conn_or_curs\"] = kwargs[\"conn_or_curs\"].__wrapped__\n # register_type takes the connection as second argument\n elif len(args) == 2 and hasattr(args[1], \"__wrapped__\"):\n args = (args[0], args[1].__wrapped__)\n # register_json takes the connection as first argument, and can have\n # several more arguments\n elif method == \"register_json\":\n if args and hasattr(args[0], \"__wrapped__\"):\n args = (args[0].__wrapped__,) + args[1:]\n\n return wrapped(*args, **kwargs)\n", "path": "elasticapm/instrumentation/packages/psycopg2.py"}], "after_files": [{"content": "# BSD 3-Clause License\n#\n# Copyright (c) 2019, Elasticsearch BV\n# All rights reserved.\n#\n# Redistribution and use in source and binary forms, with or without\n# modification, are permitted provided that the following conditions are met:\n#\n# * Redistributions of source code must retain the above copyright notice, this\n# list of conditions and the following disclaimer.\n#\n# * Redistributions in binary form must reproduce the above copyright notice,\n# this list of conditions and the following disclaimer in the documentation\n# and/or other materials provided with the distribution.\n#\n# * Neither the name of the copyright holder nor the names of its\n# contributors may be used to endorse or promote products derived from\n# this software without specific prior written permission.\n#\n# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS \"AS IS\"\n# AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE\n# IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE\n# DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE\n# FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL\n# DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR\n# SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER\n# CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,\n# OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE\n# OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.\n\nfrom elasticapm.instrumentation.packages.dbapi2 import (\n ConnectionProxy,\n CursorProxy,\n DbApi2Instrumentation,\n extract_signature,\n)\nfrom elasticapm.traces import capture_span\nfrom elasticapm.utils import default_ports\n\n\nclass PGCursorProxy(CursorProxy):\n provider_name = \"postgresql\"\n\n def _bake_sql(self, sql):\n # if this is a Composable object, use its `as_string` method\n # see http://initd.org/psycopg/docs/sql.html\n if hasattr(sql, \"as_string\"):\n return sql.as_string(self.__wrapped__)\n return sql\n\n def extract_signature(self, sql):\n return extract_signature(sql)\n\n def __enter__(self):\n return PGCursorProxy(self.__wrapped__.__enter__())\n\n\nclass PGConnectionProxy(ConnectionProxy):\n cursor_proxy = PGCursorProxy\n\n def __enter__(self):\n return PGConnectionProxy(self.__wrapped__.__enter__())\n\n\nclass Psycopg2Instrumentation(DbApi2Instrumentation):\n name = \"psycopg2\"\n\n instrument_list = [(\"psycopg2\", \"connect\")]\n\n def call(self, module, method, wrapped, instance, args, kwargs):\n signature = \"psycopg2.connect\"\n\n host = kwargs.get(\"host\")\n if host:\n signature += \" \" + str(host)\n\n port = kwargs.get(\"port\")\n if port:\n port = str(port)\n if int(port) != default_ports.get(\"postgresql\"):\n signature += \":\" + port\n else:\n # Parse connection string and extract host/port\n pass\n\n with capture_span(signature, span_type=\"db\", span_subtype=\"postgresql\", span_action=\"connect\"):\n return PGConnectionProxy(wrapped(*args, **kwargs))\n\n\nclass Psycopg2RegisterTypeInstrumentation(DbApi2Instrumentation):\n name = \"psycopg2-register-type\"\n\n instrument_list = [\n (\"psycopg2.extensions\", \"register_type\"),\n # specifically instrument `register_json` as it bypasses `register_type`\n (\"psycopg2._json\", \"register_json\"),\n ]\n\n def call(self, module, method, wrapped, instance, args, kwargs):\n if \"conn_or_curs\" in kwargs and hasattr(kwargs[\"conn_or_curs\"], \"__wrapped__\"):\n kwargs[\"conn_or_curs\"] = kwargs[\"conn_or_curs\"].__wrapped__\n # register_type takes the connection as second argument\n elif len(args) == 2 and hasattr(args[1], \"__wrapped__\"):\n args = (args[0], args[1].__wrapped__)\n # register_json takes the connection as first argument, and can have\n # several more arguments\n elif method == \"register_json\":\n if args and hasattr(args[0], \"__wrapped__\"):\n args = (args[0].__wrapped__,) + args[1:]\n\n return wrapped(*args, **kwargs)\n", "path": "elasticapm/instrumentation/packages/psycopg2.py"}]} | 1,549 | 174 |
gh_patches_debug_7901 | rasdani/github-patches | git_diff | searxng__searxng-687 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Disabling all engines in a category makes the bang search in general
**How To Reproduce**
1. Disable all engines in one category in the user preferences (e.g. in the files category).
2. Search in the category using the bang syntax (e.g. `!files test`).
**Expected behavior**
No results (and maybe a message that all engines in this category are disabled).
**Observed behavior**
The search is performed in the general category.
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `searx/webadapter.py`
Content:
```
1 from collections import defaultdict
2 from typing import Dict, List, Optional, Tuple
3 from searx.exceptions import SearxParameterException
4 from searx.webutils import VALID_LANGUAGE_CODE
5 from searx.query import RawTextQuery
6 from searx.engines import categories, engines
7 from searx.search import SearchQuery, EngineRef
8 from searx.preferences import Preferences, is_locked
9
10
11 # remove duplicate queries.
12 # FIXME: does not fix "!music !soundcloud", because the categories are 'none' and 'music'
13 def deduplicate_engineref_list(engineref_list: List[EngineRef]) -> List[EngineRef]:
14 engineref_dict = {q.category + '|' + q.name: q for q in engineref_list}
15 return list(engineref_dict.values())
16
17
18 def validate_engineref_list(
19 engineref_list: List[EngineRef], preferences: Preferences
20 ) -> Tuple[List[EngineRef], List[EngineRef], List[EngineRef]]:
21 """Validate query_engines according to the preferences
22
23 Returns:
24 List[EngineRef]: list of existing engines with a validated token
25 List[EngineRef]: list of unknown engine
26 List[EngineRef]: list of engine with invalid token according to the preferences
27 """
28 valid = []
29 unknown = []
30 no_token = []
31 for engineref in engineref_list:
32 if engineref.name not in engines:
33 unknown.append(engineref)
34 continue
35
36 engine = engines[engineref.name]
37 if not preferences.validate_token(engine):
38 no_token.append(engineref)
39 continue
40
41 valid.append(engineref)
42 return valid, unknown, no_token
43
44
45 def parse_pageno(form: Dict[str, str]) -> int:
46 pageno_param = form.get('pageno', '1')
47 if not pageno_param.isdigit() or int(pageno_param) < 1:
48 raise SearxParameterException('pageno', pageno_param)
49 return int(pageno_param)
50
51
52 def parse_lang(preferences: Preferences, form: Dict[str, str], raw_text_query: RawTextQuery) -> str:
53 if is_locked('language'):
54 return preferences.get_value('language')
55 # get language
56 # set specific language if set on request, query or preferences
57 # TODO support search with multible languages
58 if len(raw_text_query.languages):
59 query_lang = raw_text_query.languages[-1]
60 elif 'language' in form:
61 query_lang = form.get('language')
62 else:
63 query_lang = preferences.get_value('language')
64
65 # check language
66 if not VALID_LANGUAGE_CODE.match(query_lang):
67 raise SearxParameterException('language', query_lang)
68
69 return query_lang
70
71
72 def parse_safesearch(preferences: Preferences, form: Dict[str, str]) -> int:
73 if is_locked('safesearch'):
74 return preferences.get_value('safesearch')
75
76 if 'safesearch' in form:
77 query_safesearch = form.get('safesearch')
78 # first check safesearch
79 if not query_safesearch.isdigit():
80 raise SearxParameterException('safesearch', query_safesearch)
81 query_safesearch = int(query_safesearch)
82 else:
83 query_safesearch = preferences.get_value('safesearch')
84
85 # safesearch : second check
86 if query_safesearch < 0 or query_safesearch > 2:
87 raise SearxParameterException('safesearch', query_safesearch)
88
89 return query_safesearch
90
91
92 def parse_time_range(form: Dict[str, str]) -> Optional[str]:
93 query_time_range = form.get('time_range')
94 # check time_range
95 query_time_range = None if query_time_range in ('', 'None') else query_time_range
96 if query_time_range not in (None, 'day', 'week', 'month', 'year'):
97 raise SearxParameterException('time_range', query_time_range)
98 return query_time_range
99
100
101 def parse_timeout(form: Dict[str, str], raw_text_query: RawTextQuery) -> Optional[float]:
102 timeout_limit = raw_text_query.timeout_limit
103 if timeout_limit is None:
104 timeout_limit = form.get('timeout_limit')
105
106 if timeout_limit is None or timeout_limit in ['None', '']:
107 return None
108 try:
109 return float(timeout_limit)
110 except ValueError as e:
111 raise SearxParameterException('timeout_limit', timeout_limit) from e
112
113
114 def parse_category_form(query_categories: List[str], name: str, value: str) -> None:
115 if name == 'categories':
116 query_categories.extend(categ for categ in map(str.strip, value.split(',')) if categ in categories)
117 elif name.startswith('category_'):
118 category = name[9:]
119
120 # if category is not found in list, skip
121 if category not in categories:
122 return
123
124 if value != 'off':
125 # add category to list
126 query_categories.append(category)
127 elif category in query_categories:
128 # remove category from list if property is set to 'off'
129 query_categories.remove(category)
130
131
132 def get_selected_categories(preferences: Preferences, form: Optional[Dict[str, str]]) -> List[str]:
133 selected_categories = []
134
135 if not is_locked('categories') and form is not None:
136 for name, value in form.items():
137 parse_category_form(selected_categories, name, value)
138
139 # if no category is specified for this search,
140 # using user-defined default-configuration which
141 # (is stored in cookie)
142 if not selected_categories:
143 cookie_categories = preferences.get_value('categories')
144 for ccateg in cookie_categories:
145 selected_categories.append(ccateg)
146
147 # if still no category is specified, using general
148 # as default-category
149 if not selected_categories:
150 selected_categories = ['general']
151
152 return selected_categories
153
154
155 def get_engineref_from_category_list(category_list: List[str], disabled_engines: List[str]) -> List[EngineRef]:
156 result = []
157 for categ in category_list:
158 result.extend(
159 EngineRef(engine.name, categ)
160 for engine in categories[categ]
161 if (engine.name, categ) not in disabled_engines
162 )
163 return result
164
165
166 def parse_generic(preferences: Preferences, form: Dict[str, str], disabled_engines: List[str]) -> List[EngineRef]:
167 query_engineref_list = []
168 query_categories = []
169
170 # set categories/engines
171 explicit_engine_list = False
172 if not is_locked('categories'):
173 # parse the form only if the categories are not locked
174 for pd_name, pd in form.items():
175 if pd_name == 'engines':
176 pd_engines = [
177 EngineRef(engine_name, engines[engine_name].categories[0])
178 for engine_name in map(str.strip, pd.split(','))
179 if engine_name in engines
180 ]
181 if pd_engines:
182 query_engineref_list.extend(pd_engines)
183 explicit_engine_list = True
184 else:
185 parse_category_form(query_categories, pd_name, pd)
186
187 if explicit_engine_list:
188 # explicit list of engines with the "engines" parameter in the form
189 if query_categories:
190 # add engines from referenced by the "categories" parameter and the "category_*"" parameters
191 query_engineref_list.extend(get_engineref_from_category_list(query_categories, disabled_engines))
192 else:
193 # no "engines" parameters in the form
194 if not query_categories:
195 # and neither "categories" parameter nor "category_*"" parameters in the form
196 # -> get the categories from the preferences (the cookies or the settings)
197 query_categories = get_selected_categories(preferences, None)
198
199 # using all engines for that search, which are
200 # declared under the specific categories
201 query_engineref_list.extend(get_engineref_from_category_list(query_categories, disabled_engines))
202
203 return query_engineref_list
204
205
206 def parse_engine_data(form):
207 engine_data = defaultdict(dict)
208 for k, v in form.items():
209 if k.startswith("engine_data"):
210 _, engine, key = k.split('-')
211 engine_data[engine][key] = v
212 return engine_data
213
214
215 def get_search_query_from_webapp(
216 preferences: Preferences, form: Dict[str, str]
217 ) -> Tuple[SearchQuery, RawTextQuery, List[EngineRef], List[EngineRef]]:
218 # no text for the query ?
219 if not form.get('q'):
220 raise SearxParameterException('q', '')
221
222 # set blocked engines
223 disabled_engines = preferences.engines.get_disabled()
224
225 # parse query, if tags are set, which change
226 # the serch engine or search-language
227 raw_text_query = RawTextQuery(form['q'], disabled_engines)
228
229 # set query
230 query = raw_text_query.getQuery()
231 query_pageno = parse_pageno(form)
232 query_lang = parse_lang(preferences, form, raw_text_query)
233 query_safesearch = parse_safesearch(preferences, form)
234 query_time_range = parse_time_range(form)
235 query_timeout = parse_timeout(form, raw_text_query)
236 external_bang = raw_text_query.external_bang
237 engine_data = parse_engine_data(form)
238
239 if not is_locked('categories') and raw_text_query.enginerefs and raw_text_query.specific:
240 # if engines are calculated from query,
241 # set categories by using that informations
242 query_engineref_list = raw_text_query.enginerefs
243 else:
244 # otherwise, using defined categories to
245 # calculate which engines should be used
246 query_engineref_list = parse_generic(preferences, form, disabled_engines)
247
248 query_engineref_list = deduplicate_engineref_list(query_engineref_list)
249 query_engineref_list, query_engineref_list_unknown, query_engineref_list_notoken = validate_engineref_list(
250 query_engineref_list, preferences
251 )
252
253 return (
254 SearchQuery(
255 query,
256 query_engineref_list,
257 query_lang,
258 query_safesearch,
259 query_pageno,
260 query_time_range,
261 query_timeout,
262 external_bang=external_bang,
263 engine_data=engine_data,
264 ),
265 raw_text_query,
266 query_engineref_list_unknown,
267 query_engineref_list_notoken,
268 )
269
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/searx/webadapter.py b/searx/webadapter.py
--- a/searx/webadapter.py
+++ b/searx/webadapter.py
@@ -236,7 +236,7 @@
external_bang = raw_text_query.external_bang
engine_data = parse_engine_data(form)
- if not is_locked('categories') and raw_text_query.enginerefs and raw_text_query.specific:
+ if not is_locked('categories') and raw_text_query.specific:
# if engines are calculated from query,
# set categories by using that informations
query_engineref_list = raw_text_query.enginerefs
| {"golden_diff": "diff --git a/searx/webadapter.py b/searx/webadapter.py\n--- a/searx/webadapter.py\n+++ b/searx/webadapter.py\n@@ -236,7 +236,7 @@\n external_bang = raw_text_query.external_bang\n engine_data = parse_engine_data(form)\n \n- if not is_locked('categories') and raw_text_query.enginerefs and raw_text_query.specific:\n+ if not is_locked('categories') and raw_text_query.specific:\n # if engines are calculated from query,\n # set categories by using that informations\n query_engineref_list = raw_text_query.enginerefs\n", "issue": "Disabling all engines in a category makes the bang search in general\n**How To Reproduce**\r\n1. Disable all engines in one category in the user preferences (e.g. in the files category).\r\n2. Search in the category using the bang syntax (e.g. `!files test`).\r\n\r\n**Expected behavior**\r\nNo results (and maybe a message that all engines in this category are disabled).\r\n\r\n**Observed behavior**\r\nThe search is performed in the general category.\n", "before_files": [{"content": "from collections import defaultdict\nfrom typing import Dict, List, Optional, Tuple\nfrom searx.exceptions import SearxParameterException\nfrom searx.webutils import VALID_LANGUAGE_CODE\nfrom searx.query import RawTextQuery\nfrom searx.engines import categories, engines\nfrom searx.search import SearchQuery, EngineRef\nfrom searx.preferences import Preferences, is_locked\n\n\n# remove duplicate queries.\n# FIXME: does not fix \"!music !soundcloud\", because the categories are 'none' and 'music'\ndef deduplicate_engineref_list(engineref_list: List[EngineRef]) -> List[EngineRef]:\n engineref_dict = {q.category + '|' + q.name: q for q in engineref_list}\n return list(engineref_dict.values())\n\n\ndef validate_engineref_list(\n engineref_list: List[EngineRef], preferences: Preferences\n) -> Tuple[List[EngineRef], List[EngineRef], List[EngineRef]]:\n \"\"\"Validate query_engines according to the preferences\n\n Returns:\n List[EngineRef]: list of existing engines with a validated token\n List[EngineRef]: list of unknown engine\n List[EngineRef]: list of engine with invalid token according to the preferences\n \"\"\"\n valid = []\n unknown = []\n no_token = []\n for engineref in engineref_list:\n if engineref.name not in engines:\n unknown.append(engineref)\n continue\n\n engine = engines[engineref.name]\n if not preferences.validate_token(engine):\n no_token.append(engineref)\n continue\n\n valid.append(engineref)\n return valid, unknown, no_token\n\n\ndef parse_pageno(form: Dict[str, str]) -> int:\n pageno_param = form.get('pageno', '1')\n if not pageno_param.isdigit() or int(pageno_param) < 1:\n raise SearxParameterException('pageno', pageno_param)\n return int(pageno_param)\n\n\ndef parse_lang(preferences: Preferences, form: Dict[str, str], raw_text_query: RawTextQuery) -> str:\n if is_locked('language'):\n return preferences.get_value('language')\n # get language\n # set specific language if set on request, query or preferences\n # TODO support search with multible languages\n if len(raw_text_query.languages):\n query_lang = raw_text_query.languages[-1]\n elif 'language' in form:\n query_lang = form.get('language')\n else:\n query_lang = preferences.get_value('language')\n\n # check language\n if not VALID_LANGUAGE_CODE.match(query_lang):\n raise SearxParameterException('language', query_lang)\n\n return query_lang\n\n\ndef parse_safesearch(preferences: Preferences, form: Dict[str, str]) -> int:\n if is_locked('safesearch'):\n return preferences.get_value('safesearch')\n\n if 'safesearch' in form:\n query_safesearch = form.get('safesearch')\n # first check safesearch\n if not query_safesearch.isdigit():\n raise SearxParameterException('safesearch', query_safesearch)\n query_safesearch = int(query_safesearch)\n else:\n query_safesearch = preferences.get_value('safesearch')\n\n # safesearch : second check\n if query_safesearch < 0 or query_safesearch > 2:\n raise SearxParameterException('safesearch', query_safesearch)\n\n return query_safesearch\n\n\ndef parse_time_range(form: Dict[str, str]) -> Optional[str]:\n query_time_range = form.get('time_range')\n # check time_range\n query_time_range = None if query_time_range in ('', 'None') else query_time_range\n if query_time_range not in (None, 'day', 'week', 'month', 'year'):\n raise SearxParameterException('time_range', query_time_range)\n return query_time_range\n\n\ndef parse_timeout(form: Dict[str, str], raw_text_query: RawTextQuery) -> Optional[float]:\n timeout_limit = raw_text_query.timeout_limit\n if timeout_limit is None:\n timeout_limit = form.get('timeout_limit')\n\n if timeout_limit is None or timeout_limit in ['None', '']:\n return None\n try:\n return float(timeout_limit)\n except ValueError as e:\n raise SearxParameterException('timeout_limit', timeout_limit) from e\n\n\ndef parse_category_form(query_categories: List[str], name: str, value: str) -> None:\n if name == 'categories':\n query_categories.extend(categ for categ in map(str.strip, value.split(',')) if categ in categories)\n elif name.startswith('category_'):\n category = name[9:]\n\n # if category is not found in list, skip\n if category not in categories:\n return\n\n if value != 'off':\n # add category to list\n query_categories.append(category)\n elif category in query_categories:\n # remove category from list if property is set to 'off'\n query_categories.remove(category)\n\n\ndef get_selected_categories(preferences: Preferences, form: Optional[Dict[str, str]]) -> List[str]:\n selected_categories = []\n\n if not is_locked('categories') and form is not None:\n for name, value in form.items():\n parse_category_form(selected_categories, name, value)\n\n # if no category is specified for this search,\n # using user-defined default-configuration which\n # (is stored in cookie)\n if not selected_categories:\n cookie_categories = preferences.get_value('categories')\n for ccateg in cookie_categories:\n selected_categories.append(ccateg)\n\n # if still no category is specified, using general\n # as default-category\n if not selected_categories:\n selected_categories = ['general']\n\n return selected_categories\n\n\ndef get_engineref_from_category_list(category_list: List[str], disabled_engines: List[str]) -> List[EngineRef]:\n result = []\n for categ in category_list:\n result.extend(\n EngineRef(engine.name, categ)\n for engine in categories[categ]\n if (engine.name, categ) not in disabled_engines\n )\n return result\n\n\ndef parse_generic(preferences: Preferences, form: Dict[str, str], disabled_engines: List[str]) -> List[EngineRef]:\n query_engineref_list = []\n query_categories = []\n\n # set categories/engines\n explicit_engine_list = False\n if not is_locked('categories'):\n # parse the form only if the categories are not locked\n for pd_name, pd in form.items():\n if pd_name == 'engines':\n pd_engines = [\n EngineRef(engine_name, engines[engine_name].categories[0])\n for engine_name in map(str.strip, pd.split(','))\n if engine_name in engines\n ]\n if pd_engines:\n query_engineref_list.extend(pd_engines)\n explicit_engine_list = True\n else:\n parse_category_form(query_categories, pd_name, pd)\n\n if explicit_engine_list:\n # explicit list of engines with the \"engines\" parameter in the form\n if query_categories:\n # add engines from referenced by the \"categories\" parameter and the \"category_*\"\" parameters\n query_engineref_list.extend(get_engineref_from_category_list(query_categories, disabled_engines))\n else:\n # no \"engines\" parameters in the form\n if not query_categories:\n # and neither \"categories\" parameter nor \"category_*\"\" parameters in the form\n # -> get the categories from the preferences (the cookies or the settings)\n query_categories = get_selected_categories(preferences, None)\n\n # using all engines for that search, which are\n # declared under the specific categories\n query_engineref_list.extend(get_engineref_from_category_list(query_categories, disabled_engines))\n\n return query_engineref_list\n\n\ndef parse_engine_data(form):\n engine_data = defaultdict(dict)\n for k, v in form.items():\n if k.startswith(\"engine_data\"):\n _, engine, key = k.split('-')\n engine_data[engine][key] = v\n return engine_data\n\n\ndef get_search_query_from_webapp(\n preferences: Preferences, form: Dict[str, str]\n) -> Tuple[SearchQuery, RawTextQuery, List[EngineRef], List[EngineRef]]:\n # no text for the query ?\n if not form.get('q'):\n raise SearxParameterException('q', '')\n\n # set blocked engines\n disabled_engines = preferences.engines.get_disabled()\n\n # parse query, if tags are set, which change\n # the serch engine or search-language\n raw_text_query = RawTextQuery(form['q'], disabled_engines)\n\n # set query\n query = raw_text_query.getQuery()\n query_pageno = parse_pageno(form)\n query_lang = parse_lang(preferences, form, raw_text_query)\n query_safesearch = parse_safesearch(preferences, form)\n query_time_range = parse_time_range(form)\n query_timeout = parse_timeout(form, raw_text_query)\n external_bang = raw_text_query.external_bang\n engine_data = parse_engine_data(form)\n\n if not is_locked('categories') and raw_text_query.enginerefs and raw_text_query.specific:\n # if engines are calculated from query,\n # set categories by using that informations\n query_engineref_list = raw_text_query.enginerefs\n else:\n # otherwise, using defined categories to\n # calculate which engines should be used\n query_engineref_list = parse_generic(preferences, form, disabled_engines)\n\n query_engineref_list = deduplicate_engineref_list(query_engineref_list)\n query_engineref_list, query_engineref_list_unknown, query_engineref_list_notoken = validate_engineref_list(\n query_engineref_list, preferences\n )\n\n return (\n SearchQuery(\n query,\n query_engineref_list,\n query_lang,\n query_safesearch,\n query_pageno,\n query_time_range,\n query_timeout,\n external_bang=external_bang,\n engine_data=engine_data,\n ),\n raw_text_query,\n query_engineref_list_unknown,\n query_engineref_list_notoken,\n )\n", "path": "searx/webadapter.py"}], "after_files": [{"content": "from collections import defaultdict\nfrom typing import Dict, List, Optional, Tuple\nfrom searx.exceptions import SearxParameterException\nfrom searx.webutils import VALID_LANGUAGE_CODE\nfrom searx.query import RawTextQuery\nfrom searx.engines import categories, engines\nfrom searx.search import SearchQuery, EngineRef\nfrom searx.preferences import Preferences, is_locked\n\n\n# remove duplicate queries.\n# FIXME: does not fix \"!music !soundcloud\", because the categories are 'none' and 'music'\ndef deduplicate_engineref_list(engineref_list: List[EngineRef]) -> List[EngineRef]:\n engineref_dict = {q.category + '|' + q.name: q for q in engineref_list}\n return list(engineref_dict.values())\n\n\ndef validate_engineref_list(\n engineref_list: List[EngineRef], preferences: Preferences\n) -> Tuple[List[EngineRef], List[EngineRef], List[EngineRef]]:\n \"\"\"Validate query_engines according to the preferences\n\n Returns:\n List[EngineRef]: list of existing engines with a validated token\n List[EngineRef]: list of unknown engine\n List[EngineRef]: list of engine with invalid token according to the preferences\n \"\"\"\n valid = []\n unknown = []\n no_token = []\n for engineref in engineref_list:\n if engineref.name not in engines:\n unknown.append(engineref)\n continue\n\n engine = engines[engineref.name]\n if not preferences.validate_token(engine):\n no_token.append(engineref)\n continue\n\n valid.append(engineref)\n return valid, unknown, no_token\n\n\ndef parse_pageno(form: Dict[str, str]) -> int:\n pageno_param = form.get('pageno', '1')\n if not pageno_param.isdigit() or int(pageno_param) < 1:\n raise SearxParameterException('pageno', pageno_param)\n return int(pageno_param)\n\n\ndef parse_lang(preferences: Preferences, form: Dict[str, str], raw_text_query: RawTextQuery) -> str:\n if is_locked('language'):\n return preferences.get_value('language')\n # get language\n # set specific language if set on request, query or preferences\n # TODO support search with multible languages\n if len(raw_text_query.languages):\n query_lang = raw_text_query.languages[-1]\n elif 'language' in form:\n query_lang = form.get('language')\n else:\n query_lang = preferences.get_value('language')\n\n # check language\n if not VALID_LANGUAGE_CODE.match(query_lang):\n raise SearxParameterException('language', query_lang)\n\n return query_lang\n\n\ndef parse_safesearch(preferences: Preferences, form: Dict[str, str]) -> int:\n if is_locked('safesearch'):\n return preferences.get_value('safesearch')\n\n if 'safesearch' in form:\n query_safesearch = form.get('safesearch')\n # first check safesearch\n if not query_safesearch.isdigit():\n raise SearxParameterException('safesearch', query_safesearch)\n query_safesearch = int(query_safesearch)\n else:\n query_safesearch = preferences.get_value('safesearch')\n\n # safesearch : second check\n if query_safesearch < 0 or query_safesearch > 2:\n raise SearxParameterException('safesearch', query_safesearch)\n\n return query_safesearch\n\n\ndef parse_time_range(form: Dict[str, str]) -> Optional[str]:\n query_time_range = form.get('time_range')\n # check time_range\n query_time_range = None if query_time_range in ('', 'None') else query_time_range\n if query_time_range not in (None, 'day', 'week', 'month', 'year'):\n raise SearxParameterException('time_range', query_time_range)\n return query_time_range\n\n\ndef parse_timeout(form: Dict[str, str], raw_text_query: RawTextQuery) -> Optional[float]:\n timeout_limit = raw_text_query.timeout_limit\n if timeout_limit is None:\n timeout_limit = form.get('timeout_limit')\n\n if timeout_limit is None or timeout_limit in ['None', '']:\n return None\n try:\n return float(timeout_limit)\n except ValueError as e:\n raise SearxParameterException('timeout_limit', timeout_limit) from e\n\n\ndef parse_category_form(query_categories: List[str], name: str, value: str) -> None:\n if name == 'categories':\n query_categories.extend(categ for categ in map(str.strip, value.split(',')) if categ in categories)\n elif name.startswith('category_'):\n category = name[9:]\n\n # if category is not found in list, skip\n if category not in categories:\n return\n\n if value != 'off':\n # add category to list\n query_categories.append(category)\n elif category in query_categories:\n # remove category from list if property is set to 'off'\n query_categories.remove(category)\n\n\ndef get_selected_categories(preferences: Preferences, form: Optional[Dict[str, str]]) -> List[str]:\n selected_categories = []\n\n if not is_locked('categories') and form is not None:\n for name, value in form.items():\n parse_category_form(selected_categories, name, value)\n\n # if no category is specified for this search,\n # using user-defined default-configuration which\n # (is stored in cookie)\n if not selected_categories:\n cookie_categories = preferences.get_value('categories')\n for ccateg in cookie_categories:\n selected_categories.append(ccateg)\n\n # if still no category is specified, using general\n # as default-category\n if not selected_categories:\n selected_categories = ['general']\n\n return selected_categories\n\n\ndef get_engineref_from_category_list(category_list: List[str], disabled_engines: List[str]) -> List[EngineRef]:\n result = []\n for categ in category_list:\n result.extend(\n EngineRef(engine.name, categ)\n for engine in categories[categ]\n if (engine.name, categ) not in disabled_engines\n )\n return result\n\n\ndef parse_generic(preferences: Preferences, form: Dict[str, str], disabled_engines: List[str]) -> List[EngineRef]:\n query_engineref_list = []\n query_categories = []\n\n # set categories/engines\n explicit_engine_list = False\n if not is_locked('categories'):\n # parse the form only if the categories are not locked\n for pd_name, pd in form.items():\n if pd_name == 'engines':\n pd_engines = [\n EngineRef(engine_name, engines[engine_name].categories[0])\n for engine_name in map(str.strip, pd.split(','))\n if engine_name in engines\n ]\n if pd_engines:\n query_engineref_list.extend(pd_engines)\n explicit_engine_list = True\n else:\n parse_category_form(query_categories, pd_name, pd)\n\n if explicit_engine_list:\n # explicit list of engines with the \"engines\" parameter in the form\n if query_categories:\n # add engines from referenced by the \"categories\" parameter and the \"category_*\"\" parameters\n query_engineref_list.extend(get_engineref_from_category_list(query_categories, disabled_engines))\n else:\n # no \"engines\" parameters in the form\n if not query_categories:\n # and neither \"categories\" parameter nor \"category_*\"\" parameters in the form\n # -> get the categories from the preferences (the cookies or the settings)\n query_categories = get_selected_categories(preferences, None)\n\n # using all engines for that search, which are\n # declared under the specific categories\n query_engineref_list.extend(get_engineref_from_category_list(query_categories, disabled_engines))\n\n return query_engineref_list\n\n\ndef parse_engine_data(form):\n engine_data = defaultdict(dict)\n for k, v in form.items():\n if k.startswith(\"engine_data\"):\n _, engine, key = k.split('-')\n engine_data[engine][key] = v\n return engine_data\n\n\ndef get_search_query_from_webapp(\n preferences: Preferences, form: Dict[str, str]\n) -> Tuple[SearchQuery, RawTextQuery, List[EngineRef], List[EngineRef]]:\n # no text for the query ?\n if not form.get('q'):\n raise SearxParameterException('q', '')\n\n # set blocked engines\n disabled_engines = preferences.engines.get_disabled()\n\n # parse query, if tags are set, which change\n # the serch engine or search-language\n raw_text_query = RawTextQuery(form['q'], disabled_engines)\n\n # set query\n query = raw_text_query.getQuery()\n query_pageno = parse_pageno(form)\n query_lang = parse_lang(preferences, form, raw_text_query)\n query_safesearch = parse_safesearch(preferences, form)\n query_time_range = parse_time_range(form)\n query_timeout = parse_timeout(form, raw_text_query)\n external_bang = raw_text_query.external_bang\n engine_data = parse_engine_data(form)\n\n if not is_locked('categories') and raw_text_query.specific:\n # if engines are calculated from query,\n # set categories by using that informations\n query_engineref_list = raw_text_query.enginerefs\n else:\n # otherwise, using defined categories to\n # calculate which engines should be used\n query_engineref_list = parse_generic(preferences, form, disabled_engines)\n\n query_engineref_list = deduplicate_engineref_list(query_engineref_list)\n query_engineref_list, query_engineref_list_unknown, query_engineref_list_notoken = validate_engineref_list(\n query_engineref_list, preferences\n )\n\n return (\n SearchQuery(\n query,\n query_engineref_list,\n query_lang,\n query_safesearch,\n query_pageno,\n query_time_range,\n query_timeout,\n external_bang=external_bang,\n engine_data=engine_data,\n ),\n raw_text_query,\n query_engineref_list_unknown,\n query_engineref_list_notoken,\n )\n", "path": "searx/webadapter.py"}]} | 3,338 | 145 |
gh_patches_debug_15463 | rasdani/github-patches | git_diff | interactions-py__interactions.py-1643 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
[BUG] Recorder Memory Usage
### Library Version
5.11.0
### Describe the Bug
Leaving and rejoining a voice channel with a bot recording causes enormous RAM usage and the bot recorder to stop functioning properly or the bot to crash completely (from out-of-memory). This is because the calculation for how many silence frames to insert is wrong (multiplies by sample rate twice).
### Steps to Reproduce
Use whatever you prefer and start monitoring your system memory usage. Join a voice channel and have a bot join it to record. Talk and then disconnect from the voice channel. Wait 3-5 seconds and rejoin the voice channel. Memory usage will spike into the 10-20 GB range or the bot will crash.
The exact outcome and how long to wait between rejoining the voice channel really depends on the amount of memory available on your system. If you wait too long and don't have enough memory, you likely will see "Error while recording: " in the logs because Python was smart enough to not let you try to allocate that much memory. I have 64 GB of RAM, of which 32 GB is available to the bot, and it will spike up to 25 GB consistently or just crash. When it crashes, you will see "Killed" in the terminal and the exit code will be 137, which is an out-of-memory exit code.
### Expected Results
The memory will not have a substantial spike and the bot will not crash.
### Minimal Reproducible Code
_No response_
### Traceback
_No response_
### Checklist
- [X] I have searched the open issues for duplicates.
- [X] I have shown the entire traceback, if possible.
- [X] I have removed my token from display, if visible.
- [X] I have attempted to debug this myself, and I believe this issue is with the library
### Additional Information
_No response_
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `interactions/api/voice/recorder.py`
Content:
```
1 import asyncio
2 import io
3 import logging
4 import os
5 import shutil
6 import struct
7 import threading
8 import time
9 from asyncio import AbstractEventLoop
10 from collections import defaultdict
11 from typing import TYPE_CHECKING
12
13 import select
14
15 from interactions.api.voice.audio import RawInputAudio
16 from interactions.api.voice.audio_writer import AudioWriter
17 from interactions.api.voice.encryption import Decryption
18 from interactions.api.voice.opus import Decoder
19 from interactions.client.const import logger_name, Missing
20 from interactions.client.utils.input_utils import unpack_helper
21 from interactions.models.discord.snowflake import Snowflake_Type, to_snowflake_list
22
23 if TYPE_CHECKING:
24 from interactions.models.internal.active_voice_state import ActiveVoiceState
25
26 __all__ = ("Recorder",)
27
28 log = logging.getLogger(logger_name)
29
30
31 class Recorder(threading.Thread):
32 def __init__(self, v_state, loop, *, output_dir: str | None = None) -> None:
33 super().__init__()
34 self.daemon = True
35
36 self.state: "ActiveVoiceState" = v_state
37 self.loop: AbstractEventLoop = loop
38 self.decrypter: Decryption = Decryption(self.state.ws.secret)
39 self._decoders: dict[str, Decoder] = defaultdict(Decoder)
40
41 # check if output_dir is a folder not a file
42 if output_dir and not os.path.isdir(output_dir):
43 raise ValueError("output_dir must be a directory")
44
45 self.output_dir = output_dir
46 self.audio: AudioWriter | None = None
47 self.encoding = "mp3"
48 self.recording = False
49 self.used = False
50
51 self.start_time = 0
52 self.user_timestamps = {}
53 self.recording_whitelist: list[Snowflake_Type] = []
54
55 if not shutil.which("ffmpeg"):
56 raise RuntimeError(
57 "Unable to start recorder. FFmpeg was not found. Please add it to your project directory or PATH. (https://ffmpeg.org/)"
58 )
59
60 async def __aenter__(self) -> "Recorder":
61 return self
62
63 async def __aexit__(self, exc_type, exc_val, exc_tb) -> None:
64 await self.stop_recording()
65
66 async def start_recording(self, *user_id: Snowflake_Type, output_dir: str | Missing = Missing) -> None:
67 """
68 Start recording audio from the current channel.
69
70 Args:
71 *user_id: The user_id(s) to record, if not specified everyone will be recorded.
72 output_dir: The directory to save the audio to (overrides the constructor output_dir if specified)
73
74 """
75 if self.used:
76 raise RuntimeError("Cannot reuse a recorder.")
77 self.used = True
78
79 if user_id:
80 self.recording_whitelist = to_snowflake_list(unpack_helper(user_id))
81
82 if output_dir is not Missing:
83 self.output_dir = output_dir
84
85 self.recording = True
86 self.audio = AudioWriter(self, self.state.channel.id)
87 self.start()
88 self.start_time = time.monotonic()
89
90 async def stop_recording(self) -> None:
91 """Stop recording audio from the current channel."""
92 self.recording = False
93
94 def wait() -> None:
95 self.audio.cleanup()
96 self.audio.encode_audio(self.encoding)
97
98 await asyncio.to_thread(wait)
99
100 def decrypt(self, header: bytes, data: bytes) -> bytes:
101 """
102 An alias to call the decryption methods.
103
104 Args:
105 header: The payload header
106 data: The payload data
107 Returns:
108 The decrypted payload
109
110 """
111 # a shorter alias to call
112 return self.decrypter.decrypt(self.state.ws.selected_mode, header, data)
113
114 def get_decoder(self, ssrc) -> Decoder:
115 return self._decoders[ssrc]
116
117 def get_user(self, ssrc: str) -> Snowflake_Type:
118 """
119 Get the corresponding user from a ssrc.
120
121 Args:
122 ssrc: The source to retrieve the user from
123 Returns:
124 A snowflake representing the user
125
126 """
127 return self.state.ws.user_ssrc_map.get(ssrc)["user_id"]
128
129 def get_ssrc(self, user_id: Snowflake_Type) -> str:
130 """
131 Get the corresponding ssrc from a user.
132
133 Args:
134 user_id: The user to retrieve the ssrc from
135 Returns:
136 A string representing the ssrc
137
138 """
139 return next((ssrc for ssrc, user in self.state.ws.user_ssrc_map.items() if user["user_id"] == user_id), None)
140
141 def __enter__(self) -> "Recorder":
142 return self
143
144 @property
145 def output(self) -> dict[int, io.BytesIO | str]:
146 """
147 The output of the recorder.
148
149 Returns:
150 A dictionary of the user_id and the output file.
151 Output file can be a BytesIO or a string (if output_dir is specified)
152
153 """
154 return self.audio.files if self.audio.finished.is_set() else {}
155
156 @property
157 def elapsed_time(self) -> float:
158 return time.monotonic() - self.start_time
159
160 def filter(self, *user_id: Snowflake_Type) -> None:
161 """
162 Filter the users that are being recorded.
163
164 Args:
165 *user_id: The user_id(s) to record
166
167 """
168 if not user_id:
169 self.recording_whitelist = []
170 self.recording_whitelist = to_snowflake_list(unpack_helper(user_id))
171
172 def run(self) -> None:
173 """The recording loop itself."""
174 sock = self.state.ws.socket
175
176 # purge any data that is already in the socket
177 readable, _, _ = select.select([sock], [], [], 0)
178 log.debug("Purging socket buffer")
179 while readable and sock.recv(4096):
180 readable, _, _ = select.select([sock], [], [], 0)
181 log.debug("Socket buffer purged, starting recording")
182
183 with self.audio:
184 while self.recording:
185 ready, _, err = select.select([sock], [], [sock], 0.01)
186 if not ready:
187 if err:
188 log.error("Error while recording: %s", err)
189 continue
190
191 data = sock.recv(4096)
192
193 if 200 <= data[1] <= 204:
194 continue
195
196 try:
197 raw_audio = RawInputAudio(self, data)
198 self.process_data(raw_audio)
199 except Exception as ex:
200 log.error("Error while recording: %s", ex)
201
202 def process_data(self, raw_audio: RawInputAudio) -> None:
203 """
204 Processes incoming audio data and writes it to the corresponding buffer.
205
206 Args:
207 raw_audio: The raw audio that has been received
208
209 """
210 if raw_audio.user_id is None:
211 return # usually the first frame when a user rejoins
212
213 if self.recording_whitelist and raw_audio.user_id not in self.recording_whitelist:
214 return
215
216 decoder = self.get_decoder(raw_audio.ssrc)
217
218 if raw_audio.ssrc not in self.user_timestamps:
219 if last_timestamp := self.audio.last_timestamps.get(raw_audio.user_id, None):
220 diff = raw_audio.timestamp - last_timestamp
221 silence = int(diff * decoder.sample_rate)
222 log.debug(
223 f"{self.state.channel.id}::{raw_audio.user_id} - User rejoined, adding {silence} silence frames ({diff} seconds)"
224 )
225 else:
226 silence = 0
227
228 self.user_timestamps.update({raw_audio.ssrc: raw_audio.timestamp})
229 else:
230 silence = raw_audio.timestamp - self.user_timestamps[raw_audio.ssrc]
231 if silence < 0.1:
232 silence = 0
233 self.user_timestamps[raw_audio.ssrc] = raw_audio.timestamp
234
235 raw_audio.pcm = struct.pack("<h", 0) * int(silence * decoder.sample_rate) * 2 + raw_audio.decoded
236
237 self.audio.write(raw_audio, raw_audio.user_id)
238
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/interactions/api/voice/recorder.py b/interactions/api/voice/recorder.py
--- a/interactions/api/voice/recorder.py
+++ b/interactions/api/voice/recorder.py
@@ -217,10 +217,10 @@
if raw_audio.ssrc not in self.user_timestamps:
if last_timestamp := self.audio.last_timestamps.get(raw_audio.user_id, None):
- diff = raw_audio.timestamp - last_timestamp
- silence = int(diff * decoder.sample_rate)
+ silence = raw_audio.timestamp - last_timestamp
+ frames = int(silence * decoder.sample_rate)
log.debug(
- f"{self.state.channel.id}::{raw_audio.user_id} - User rejoined, adding {silence} silence frames ({diff} seconds)"
+ f"{self.state.channel.id}::{raw_audio.user_id} - User rejoined, adding {frames} silence frames ({silence} seconds)"
)
else:
silence = 0
| {"golden_diff": "diff --git a/interactions/api/voice/recorder.py b/interactions/api/voice/recorder.py\n--- a/interactions/api/voice/recorder.py\n+++ b/interactions/api/voice/recorder.py\n@@ -217,10 +217,10 @@\n \n if raw_audio.ssrc not in self.user_timestamps:\n if last_timestamp := self.audio.last_timestamps.get(raw_audio.user_id, None):\n- diff = raw_audio.timestamp - last_timestamp\n- silence = int(diff * decoder.sample_rate)\n+ silence = raw_audio.timestamp - last_timestamp\n+ frames = int(silence * decoder.sample_rate)\n log.debug(\n- f\"{self.state.channel.id}::{raw_audio.user_id} - User rejoined, adding {silence} silence frames ({diff} seconds)\"\n+ f\"{self.state.channel.id}::{raw_audio.user_id} - User rejoined, adding {frames} silence frames ({silence} seconds)\"\n )\n else:\n silence = 0\n", "issue": "[BUG] Recorder Memory Usage\n### Library Version\n\n5.11.0\n\n### Describe the Bug\n\nLeaving and rejoining a voice channel with a bot recording causes enormous RAM usage and the bot recorder to stop functioning properly or the bot to crash completely (from out-of-memory). This is because the calculation for how many silence frames to insert is wrong (multiplies by sample rate twice).\n\n### Steps to Reproduce\n\nUse whatever you prefer and start monitoring your system memory usage. Join a voice channel and have a bot join it to record. Talk and then disconnect from the voice channel. Wait 3-5 seconds and rejoin the voice channel. Memory usage will spike into the 10-20 GB range or the bot will crash.\r\n\r\nThe exact outcome and how long to wait between rejoining the voice channel really depends on the amount of memory available on your system. If you wait too long and don't have enough memory, you likely will see \"Error while recording: \" in the logs because Python was smart enough to not let you try to allocate that much memory. I have 64 GB of RAM, of which 32 GB is available to the bot, and it will spike up to 25 GB consistently or just crash. When it crashes, you will see \"Killed\" in the terminal and the exit code will be 137, which is an out-of-memory exit code.\n\n### Expected Results\n\nThe memory will not have a substantial spike and the bot will not crash.\n\n### Minimal Reproducible Code\n\n_No response_\n\n### Traceback\n\n_No response_\n\n### Checklist\n\n- [X] I have searched the open issues for duplicates.\n- [X] I have shown the entire traceback, if possible.\n- [X] I have removed my token from display, if visible.\n- [X] I have attempted to debug this myself, and I believe this issue is with the library\n\n### Additional Information\n\n_No response_\n", "before_files": [{"content": "import asyncio\nimport io\nimport logging\nimport os\nimport shutil\nimport struct\nimport threading\nimport time\nfrom asyncio import AbstractEventLoop\nfrom collections import defaultdict\nfrom typing import TYPE_CHECKING\n\nimport select\n\nfrom interactions.api.voice.audio import RawInputAudio\nfrom interactions.api.voice.audio_writer import AudioWriter\nfrom interactions.api.voice.encryption import Decryption\nfrom interactions.api.voice.opus import Decoder\nfrom interactions.client.const import logger_name, Missing\nfrom interactions.client.utils.input_utils import unpack_helper\nfrom interactions.models.discord.snowflake import Snowflake_Type, to_snowflake_list\n\nif TYPE_CHECKING:\n from interactions.models.internal.active_voice_state import ActiveVoiceState\n\n__all__ = (\"Recorder\",)\n\nlog = logging.getLogger(logger_name)\n\n\nclass Recorder(threading.Thread):\n def __init__(self, v_state, loop, *, output_dir: str | None = None) -> None:\n super().__init__()\n self.daemon = True\n\n self.state: \"ActiveVoiceState\" = v_state\n self.loop: AbstractEventLoop = loop\n self.decrypter: Decryption = Decryption(self.state.ws.secret)\n self._decoders: dict[str, Decoder] = defaultdict(Decoder)\n\n # check if output_dir is a folder not a file\n if output_dir and not os.path.isdir(output_dir):\n raise ValueError(\"output_dir must be a directory\")\n\n self.output_dir = output_dir\n self.audio: AudioWriter | None = None\n self.encoding = \"mp3\"\n self.recording = False\n self.used = False\n\n self.start_time = 0\n self.user_timestamps = {}\n self.recording_whitelist: list[Snowflake_Type] = []\n\n if not shutil.which(\"ffmpeg\"):\n raise RuntimeError(\n \"Unable to start recorder. FFmpeg was not found. Please add it to your project directory or PATH. (https://ffmpeg.org/)\"\n )\n\n async def __aenter__(self) -> \"Recorder\":\n return self\n\n async def __aexit__(self, exc_type, exc_val, exc_tb) -> None:\n await self.stop_recording()\n\n async def start_recording(self, *user_id: Snowflake_Type, output_dir: str | Missing = Missing) -> None:\n \"\"\"\n Start recording audio from the current channel.\n\n Args:\n *user_id: The user_id(s) to record, if not specified everyone will be recorded.\n output_dir: The directory to save the audio to (overrides the constructor output_dir if specified)\n\n \"\"\"\n if self.used:\n raise RuntimeError(\"Cannot reuse a recorder.\")\n self.used = True\n\n if user_id:\n self.recording_whitelist = to_snowflake_list(unpack_helper(user_id))\n\n if output_dir is not Missing:\n self.output_dir = output_dir\n\n self.recording = True\n self.audio = AudioWriter(self, self.state.channel.id)\n self.start()\n self.start_time = time.monotonic()\n\n async def stop_recording(self) -> None:\n \"\"\"Stop recording audio from the current channel.\"\"\"\n self.recording = False\n\n def wait() -> None:\n self.audio.cleanup()\n self.audio.encode_audio(self.encoding)\n\n await asyncio.to_thread(wait)\n\n def decrypt(self, header: bytes, data: bytes) -> bytes:\n \"\"\"\n An alias to call the decryption methods.\n\n Args:\n header: The payload header\n data: The payload data\n Returns:\n The decrypted payload\n\n \"\"\"\n # a shorter alias to call\n return self.decrypter.decrypt(self.state.ws.selected_mode, header, data)\n\n def get_decoder(self, ssrc) -> Decoder:\n return self._decoders[ssrc]\n\n def get_user(self, ssrc: str) -> Snowflake_Type:\n \"\"\"\n Get the corresponding user from a ssrc.\n\n Args:\n ssrc: The source to retrieve the user from\n Returns:\n A snowflake representing the user\n\n \"\"\"\n return self.state.ws.user_ssrc_map.get(ssrc)[\"user_id\"]\n\n def get_ssrc(self, user_id: Snowflake_Type) -> str:\n \"\"\"\n Get the corresponding ssrc from a user.\n\n Args:\n user_id: The user to retrieve the ssrc from\n Returns:\n A string representing the ssrc\n\n \"\"\"\n return next((ssrc for ssrc, user in self.state.ws.user_ssrc_map.items() if user[\"user_id\"] == user_id), None)\n\n def __enter__(self) -> \"Recorder\":\n return self\n\n @property\n def output(self) -> dict[int, io.BytesIO | str]:\n \"\"\"\n The output of the recorder.\n\n Returns:\n A dictionary of the user_id and the output file.\n Output file can be a BytesIO or a string (if output_dir is specified)\n\n \"\"\"\n return self.audio.files if self.audio.finished.is_set() else {}\n\n @property\n def elapsed_time(self) -> float:\n return time.monotonic() - self.start_time\n\n def filter(self, *user_id: Snowflake_Type) -> None:\n \"\"\"\n Filter the users that are being recorded.\n\n Args:\n *user_id: The user_id(s) to record\n\n \"\"\"\n if not user_id:\n self.recording_whitelist = []\n self.recording_whitelist = to_snowflake_list(unpack_helper(user_id))\n\n def run(self) -> None:\n \"\"\"The recording loop itself.\"\"\"\n sock = self.state.ws.socket\n\n # purge any data that is already in the socket\n readable, _, _ = select.select([sock], [], [], 0)\n log.debug(\"Purging socket buffer\")\n while readable and sock.recv(4096):\n readable, _, _ = select.select([sock], [], [], 0)\n log.debug(\"Socket buffer purged, starting recording\")\n\n with self.audio:\n while self.recording:\n ready, _, err = select.select([sock], [], [sock], 0.01)\n if not ready:\n if err:\n log.error(\"Error while recording: %s\", err)\n continue\n\n data = sock.recv(4096)\n\n if 200 <= data[1] <= 204:\n continue\n\n try:\n raw_audio = RawInputAudio(self, data)\n self.process_data(raw_audio)\n except Exception as ex:\n log.error(\"Error while recording: %s\", ex)\n\n def process_data(self, raw_audio: RawInputAudio) -> None:\n \"\"\"\n Processes incoming audio data and writes it to the corresponding buffer.\n\n Args:\n raw_audio: The raw audio that has been received\n\n \"\"\"\n if raw_audio.user_id is None:\n return # usually the first frame when a user rejoins\n\n if self.recording_whitelist and raw_audio.user_id not in self.recording_whitelist:\n return\n\n decoder = self.get_decoder(raw_audio.ssrc)\n\n if raw_audio.ssrc not in self.user_timestamps:\n if last_timestamp := self.audio.last_timestamps.get(raw_audio.user_id, None):\n diff = raw_audio.timestamp - last_timestamp\n silence = int(diff * decoder.sample_rate)\n log.debug(\n f\"{self.state.channel.id}::{raw_audio.user_id} - User rejoined, adding {silence} silence frames ({diff} seconds)\"\n )\n else:\n silence = 0\n\n self.user_timestamps.update({raw_audio.ssrc: raw_audio.timestamp})\n else:\n silence = raw_audio.timestamp - self.user_timestamps[raw_audio.ssrc]\n if silence < 0.1:\n silence = 0\n self.user_timestamps[raw_audio.ssrc] = raw_audio.timestamp\n\n raw_audio.pcm = struct.pack(\"<h\", 0) * int(silence * decoder.sample_rate) * 2 + raw_audio.decoded\n\n self.audio.write(raw_audio, raw_audio.user_id)\n", "path": "interactions/api/voice/recorder.py"}], "after_files": [{"content": "import asyncio\nimport io\nimport logging\nimport os\nimport shutil\nimport struct\nimport threading\nimport time\nfrom asyncio import AbstractEventLoop\nfrom collections import defaultdict\nfrom typing import TYPE_CHECKING\n\nimport select\n\nfrom interactions.api.voice.audio import RawInputAudio\nfrom interactions.api.voice.audio_writer import AudioWriter\nfrom interactions.api.voice.encryption import Decryption\nfrom interactions.api.voice.opus import Decoder\nfrom interactions.client.const import logger_name, Missing\nfrom interactions.client.utils.input_utils import unpack_helper\nfrom interactions.models.discord.snowflake import Snowflake_Type, to_snowflake_list\n\nif TYPE_CHECKING:\n from interactions.models.internal.active_voice_state import ActiveVoiceState\n\n__all__ = (\"Recorder\",)\n\nlog = logging.getLogger(logger_name)\n\n\nclass Recorder(threading.Thread):\n def __init__(self, v_state, loop, *, output_dir: str | None = None) -> None:\n super().__init__()\n self.daemon = True\n\n self.state: \"ActiveVoiceState\" = v_state\n self.loop: AbstractEventLoop = loop\n self.decrypter: Decryption = Decryption(self.state.ws.secret)\n self._decoders: dict[str, Decoder] = defaultdict(Decoder)\n\n # check if output_dir is a folder not a file\n if output_dir and not os.path.isdir(output_dir):\n raise ValueError(\"output_dir must be a directory\")\n\n self.output_dir = output_dir\n self.audio: AudioWriter | None = None\n self.encoding = \"mp3\"\n self.recording = False\n self.used = False\n\n self.start_time = 0\n self.user_timestamps = {}\n self.recording_whitelist: list[Snowflake_Type] = []\n\n if not shutil.which(\"ffmpeg\"):\n raise RuntimeError(\n \"Unable to start recorder. FFmpeg was not found. Please add it to your project directory or PATH. (https://ffmpeg.org/)\"\n )\n\n async def __aenter__(self) -> \"Recorder\":\n return self\n\n async def __aexit__(self, exc_type, exc_val, exc_tb) -> None:\n await self.stop_recording()\n\n async def start_recording(self, *user_id: Snowflake_Type, output_dir: str | Missing = Missing) -> None:\n \"\"\"\n Start recording audio from the current channel.\n\n Args:\n *user_id: The user_id(s) to record, if not specified everyone will be recorded.\n output_dir: The directory to save the audio to (overrides the constructor output_dir if specified)\n\n \"\"\"\n if self.used:\n raise RuntimeError(\"Cannot reuse a recorder.\")\n self.used = True\n\n if user_id:\n self.recording_whitelist = to_snowflake_list(unpack_helper(user_id))\n\n if output_dir is not Missing:\n self.output_dir = output_dir\n\n self.recording = True\n self.audio = AudioWriter(self, self.state.channel.id)\n self.start()\n self.start_time = time.monotonic()\n\n async def stop_recording(self) -> None:\n \"\"\"Stop recording audio from the current channel.\"\"\"\n self.recording = False\n\n def wait() -> None:\n self.audio.cleanup()\n self.audio.encode_audio(self.encoding)\n\n await asyncio.to_thread(wait)\n\n def decrypt(self, header: bytes, data: bytes) -> bytes:\n \"\"\"\n An alias to call the decryption methods.\n\n Args:\n header: The payload header\n data: The payload data\n Returns:\n The decrypted payload\n\n \"\"\"\n # a shorter alias to call\n return self.decrypter.decrypt(self.state.ws.selected_mode, header, data)\n\n def get_decoder(self, ssrc) -> Decoder:\n return self._decoders[ssrc]\n\n def get_user(self, ssrc: str) -> Snowflake_Type:\n \"\"\"\n Get the corresponding user from a ssrc.\n\n Args:\n ssrc: The source to retrieve the user from\n Returns:\n A snowflake representing the user\n\n \"\"\"\n return self.state.ws.user_ssrc_map.get(ssrc)[\"user_id\"]\n\n def get_ssrc(self, user_id: Snowflake_Type) -> str:\n \"\"\"\n Get the corresponding ssrc from a user.\n\n Args:\n user_id: The user to retrieve the ssrc from\n Returns:\n A string representing the ssrc\n\n \"\"\"\n return next((ssrc for ssrc, user in self.state.ws.user_ssrc_map.items() if user[\"user_id\"] == user_id), None)\n\n def __enter__(self) -> \"Recorder\":\n return self\n\n @property\n def output(self) -> dict[int, io.BytesIO | str]:\n \"\"\"\n The output of the recorder.\n\n Returns:\n A dictionary of the user_id and the output file.\n Output file can be a BytesIO or a string (if output_dir is specified)\n\n \"\"\"\n return self.audio.files if self.audio.finished.is_set() else {}\n\n @property\n def elapsed_time(self) -> float:\n return time.monotonic() - self.start_time\n\n def filter(self, *user_id: Snowflake_Type) -> None:\n \"\"\"\n Filter the users that are being recorded.\n\n Args:\n *user_id: The user_id(s) to record\n\n \"\"\"\n if not user_id:\n self.recording_whitelist = []\n self.recording_whitelist = to_snowflake_list(unpack_helper(user_id))\n\n def run(self) -> None:\n \"\"\"The recording loop itself.\"\"\"\n sock = self.state.ws.socket\n\n # purge any data that is already in the socket\n readable, _, _ = select.select([sock], [], [], 0)\n log.debug(\"Purging socket buffer\")\n while readable and sock.recv(4096):\n readable, _, _ = select.select([sock], [], [], 0)\n log.debug(\"Socket buffer purged, starting recording\")\n\n with self.audio:\n while self.recording:\n ready, _, err = select.select([sock], [], [sock], 0.01)\n if not ready:\n if err:\n log.error(\"Error while recording: %s\", err)\n continue\n\n data = sock.recv(4096)\n\n if 200 <= data[1] <= 204:\n continue\n\n try:\n raw_audio = RawInputAudio(self, data)\n self.process_data(raw_audio)\n except Exception as ex:\n log.error(\"Error while recording: %s\", ex)\n\n def process_data(self, raw_audio: RawInputAudio) -> None:\n \"\"\"\n Processes incoming audio data and writes it to the corresponding buffer.\n\n Args:\n raw_audio: The raw audio that has been received\n\n \"\"\"\n if raw_audio.user_id is None:\n return # usually the first frame when a user rejoins\n\n if self.recording_whitelist and raw_audio.user_id not in self.recording_whitelist:\n return\n\n decoder = self.get_decoder(raw_audio.ssrc)\n\n if raw_audio.ssrc not in self.user_timestamps:\n if last_timestamp := self.audio.last_timestamps.get(raw_audio.user_id, None):\n silence = raw_audio.timestamp - last_timestamp\n frames = int(silence * decoder.sample_rate)\n log.debug(\n f\"{self.state.channel.id}::{raw_audio.user_id} - User rejoined, adding {frames} silence frames ({silence} seconds)\"\n )\n else:\n silence = 0\n\n self.user_timestamps.update({raw_audio.ssrc: raw_audio.timestamp})\n else:\n silence = raw_audio.timestamp - self.user_timestamps[raw_audio.ssrc]\n if silence < 0.1:\n silence = 0\n self.user_timestamps[raw_audio.ssrc] = raw_audio.timestamp\n\n raw_audio.pcm = struct.pack(\"<h\", 0) * int(silence * decoder.sample_rate) * 2 + raw_audio.decoded\n\n self.audio.write(raw_audio, raw_audio.user_id)\n", "path": "interactions/api/voice/recorder.py"}]} | 3,009 | 221 |
gh_patches_debug_64454 | rasdani/github-patches | git_diff | bokeh__bokeh-1923 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
app_reveal fails importing old plotting stuff
```
(py34devel)[damian@damian-S400CA][slideshow](master)$ python app_reveal.py
Traceback (most recent call last):
File "app_reveal.py", line 19, in <module>
from bokeh.plotting import (annular_wedge, cursession, figure, hold, legend,
ImportError: cannot import name 'annular_wedge'
```
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `examples/embed/slideshow/app_reveal.py`
Content:
```
1 # -*- coding: utf-8 -*-
2 """
3 In this example, we want to show you how you can take isolated blocks of code
4 (featuring different kinds of Bokeh visualizations) and rearrange them in a
5 bigger (encompassing) flask-based application without losing the independence
6 of each example. This is the reason of some weirdness through the code.
7 We are using this "building blocks" approach here because we believe it has some
8 conceptual advantages for people trying to quickly understand, and more
9 importantly, use the embed API, in a more complex way than just a simple script.
10 """
11 import time
12 from threading import Thread
13
14 import numpy as np
15 import scipy.special
16
17 from bokeh.embed import autoload_server
18 from bokeh.models import GlyphRenderer
19 from bokeh.plotting import (annular_wedge, cursession, figure, hold, legend,
20 line, output_server, push, quad, xgrid, ygrid)
21
22 from flask import Flask, render_template
23 app = Flask(__name__)
24
25 @app.route('/')
26 def render_plot():
27 """
28 Get the script tags from each plot object and "insert" them into the template.
29
30 This also starts different threads for each update function, so you can have
31 a non-blocking update.
32 """
33 dist_plot, dist_session = distribution()
34 dist_tag = autoload_server(dist_plot, dist_session)
35
36 anim_plot, anim_session = animated()
37 anim_tag = autoload_server(anim_plot, anim_session)
38 # for update_animation as target we need to pass the anim_plot and anim_session as args
39 thread = Thread(target=update_animation, args=(anim_plot, anim_session))
40 thread.start()
41
42 pop = Population()
43 pop_tag = autoload_server(pop.layout, pop.session)
44 # for update_population as target we need to pass the pop instance as args
45 thread = Thread(target=update_population, args=(pop,))
46 thread.start()
47
48 return render_template('app_plot.html', tag1=dist_tag, tag2=anim_tag, tag3=pop_tag)
49
50
51 def distribution():
52
53 mu, sigma = 0, 0.5
54
55 measured = np.random.normal(mu, sigma, 1000)
56 hist, edges = np.histogram(measured, density=True, bins=20)
57
58 x = np.linspace(-2, 2, 1000)
59 pdf = 1 / (sigma * np.sqrt(2 * np.pi)) * np.exp(-(x - mu) ** 2 / (2 * sigma ** 2))
60 cdf = (1 + scipy.special.erf((x - mu) / np.sqrt(2 * sigma ** 2))) / 2
61
62 output_server("distribution_reveal")
63
64 p = figure(title="Interactive plots",
65 background_fill="#E5E5E5")
66 p.quad(top=hist, bottom=0, left=edges[:-1], right=edges[1:],
67 fill_color="#333333", line_color="#E5E5E5", line_width=3)
68
69 # Use `line` renderers to display the PDF and CDF
70 p.line(x, pdf, line_color="#348abd", line_width=8, alpha=0.7, legend="PDF")
71 p.line(x, cdf, line_color="#7a68a6", line_width=8, alpha=0.7, legend="CDF")
72
73 p.legend.orientation = "top_left"
74
75 p.xaxis.axis_label = 'x'
76 p.xgrid[0].grid_line_color = "white"
77 p.xgrid[0].grid_line_width = 3
78
79 p.yaxis.axis_label = 'Pr(x)'
80 p.ygrid[0].grid_line_color = "white"
81 p.ygrid[0].grid_line_width = 3
82
83 push()
84
85 return p, cursession()
86
87
88 def animated():
89
90 from numpy import pi, cos, sin, linspace
91
92 N = 50 + 1
93 r_base = 8
94 theta = linspace(0, 2 * pi, N)
95 r_x = linspace(0, 6 * pi, N - 1)
96 rmin = r_base - cos(r_x) - 1
97 rmax = r_base + sin(r_x) + 1
98
99 colors = ["FFFFCC", "#C7E9B4", "#7FCDBB", "#41B6C4", "#2C7FB8",
100 "#253494", "#2C7FB8", "#41B6C4", "#7FCDBB", "#C7E9B4"] * 5
101
102 output_server("animated_reveal")
103
104 p = figure(title="Animations", x_range=[-11, 11], y_range=[-11, 11])
105
106 p.annular_wedge(
107 0, 0, rmin, rmax, theta[:-1], theta[1:],
108 inner_radius_units="data",
109 outer_radius_units="data",
110 fill_color=colors,
111 line_color="black",
112 )
113
114 push()
115
116 return p, cursession()
117
118
119 def update_animation(plot, session):
120
121 from numpy import roll
122
123 renderer = plot.select(dict(type=GlyphRenderer))
124 ds = renderer[0].data_source
125
126 while True:
127
128 rmin = ds.data["inner_radius"]
129 rmin = roll(rmin, 1)
130 ds.data["inner_radius"] = rmin
131
132 rmax = ds.data["outer_radius"]
133 rmax = roll(rmax, -1)
134 ds.data["outer_radius"] = rmax
135
136 cursession().store_objects(ds)
137 time.sleep(0.1)
138
139
140 class Population(object):
141
142 year = 2010
143 location = "World"
144
145 def __init__(self):
146 from bokeh.models import ColumnDataSource
147 from bokeh.document import Document
148 from bokeh.session import Session
149 from bokeh.sampledata.population import load_population
150
151 self.document = Document()
152 self.session = Session()
153 self.session.use_doc('population_reveal')
154 self.session.load_document(self.document)
155
156 self.df = load_population()
157 self.source_pyramid = ColumnDataSource(data=dict())
158
159 # just render at the initialization
160 self._render()
161
162 def _render(self):
163 self.pyramid_plot()
164 self.create_layout()
165 self.document.add(self.layout)
166 self.update_pyramid()
167
168 def pyramid_plot(self):
169 from bokeh.models import (Plot, DataRange1d, LinearAxis, Grid,
170 Legend, SingleIntervalTicker)
171 from bokeh.models.glyphs import Quad
172
173 xdr = DataRange1d(sources=[self.source_pyramid.columns("male"),
174 self.source_pyramid.columns("female")])
175 ydr = DataRange1d(sources=[self.source_pyramid.columns("groups")])
176
177 self.plot = Plot(title="Widgets", x_range=xdr, y_range=ydr,
178 plot_width=600, plot_height=600)
179
180 xaxis = LinearAxis()
181 self.plot.add_layout(xaxis, 'below')
182 yaxis = LinearAxis(ticker=SingleIntervalTicker(interval=5))
183 self.plot.add_layout(yaxis, 'left')
184
185 self.plot.add_layout(Grid(dimension=0, ticker=xaxis.ticker))
186 self.plot.add_layout(Grid(dimension=1, ticker=yaxis.ticker))
187
188 male_quad = Quad(left="male", right=0, bottom="groups", top="shifted",
189 fill_color="#3B8686")
190 male_quad_glyph = self.plot.add_glyph(self.source_pyramid, male_quad)
191
192 female_quad = Quad(left=0, right="female", bottom="groups", top="shifted",
193 fill_color="#CFF09E")
194 female_quad_glyph = self.plot.add_glyph(self.source_pyramid, female_quad)
195
196 self.plot.add_layout(Legend(legends=dict(Male=[male_quad_glyph],
197 Female=[female_quad_glyph])))
198
199 def on_year_change(self, obj, attr, old, new):
200 self.year = int(new)
201 self.update_pyramid()
202
203 def on_location_change(self, obj, attr, old, new):
204 self.location = new
205 self.update_pyramid()
206
207 def create_layout(self):
208 from bokeh.models.widgets import Select, HBox, VBox
209
210 years = list(map(str, sorted(self.df.Year.unique())))
211 locations = sorted(self.df.Location.unique())
212
213 year_select = Select(title="Year:", value="2010", options=years)
214 location_select = Select(title="Location:", value="World", options=locations)
215
216 year_select.on_change('value', self.on_year_change)
217 location_select.on_change('value', self.on_location_change)
218
219 controls = HBox(year_select, location_select)
220 self.layout = VBox(controls, self.plot)
221
222 def update_pyramid(self):
223 pyramid = self.df[(self.df.Location == self.location) & (self.df.Year == self.year)]
224
225 male = pyramid[pyramid.Sex == "Male"]
226 female = pyramid[pyramid.Sex == "Female"]
227
228 total = male.Value.sum() + female.Value.sum()
229
230 male_percent = -male.Value / total
231 female_percent = female.Value / total
232
233 groups = male.AgeGrpStart.tolist()
234 shifted = groups[1:] + [groups[-1] + 5]
235
236 self.source_pyramid.data = dict(
237 groups=groups,
238 shifted=shifted,
239 male=male_percent,
240 female=female_percent,
241 )
242 self.session.store_document(self.document)
243
244
245 def update_population(plot):
246 while True:
247 plot.session.load_document(plot.document)
248 time.sleep(0.1)
249
250 if __name__ == '__main__':
251 app.run(debug=True)
252
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/examples/embed/slideshow/app_reveal.py b/examples/embed/slideshow/app_reveal.py
--- a/examples/embed/slideshow/app_reveal.py
+++ b/examples/embed/slideshow/app_reveal.py
@@ -16,8 +16,7 @@
from bokeh.embed import autoload_server
from bokeh.models import GlyphRenderer
-from bokeh.plotting import (annular_wedge, cursession, figure, hold, legend,
- line, output_server, push, quad, xgrid, ygrid)
+from bokeh.plotting import cursession, figure, output_server, push
from flask import Flask, render_template
app = Flask(__name__)
| {"golden_diff": "diff --git a/examples/embed/slideshow/app_reveal.py b/examples/embed/slideshow/app_reveal.py\n--- a/examples/embed/slideshow/app_reveal.py\n+++ b/examples/embed/slideshow/app_reveal.py\n@@ -16,8 +16,7 @@\n \n from bokeh.embed import autoload_server\n from bokeh.models import GlyphRenderer\n-from bokeh.plotting import (annular_wedge, cursession, figure, hold, legend,\n- line, output_server, push, quad, xgrid, ygrid)\n+from bokeh.plotting import cursession, figure, output_server, push\n \n from flask import Flask, render_template\n app = Flask(__name__)\n", "issue": "app_reveal fails importing old plotting stuff\n```\n(py34devel)[damian@damian-S400CA][slideshow](master)$ python app_reveal.py \nTraceback (most recent call last):\n File \"app_reveal.py\", line 19, in <module>\n from bokeh.plotting import (annular_wedge, cursession, figure, hold, legend,\nImportError: cannot import name 'annular_wedge'\n```\n\n", "before_files": [{"content": "# -*- coding: utf-8 -*-\n\"\"\"\nIn this example, we want to show you how you can take isolated blocks of code\n(featuring different kinds of Bokeh visualizations) and rearrange them in a\nbigger (encompassing) flask-based application without losing the independence\nof each example. This is the reason of some weirdness through the code.\nWe are using this \"building blocks\" approach here because we believe it has some\nconceptual advantages for people trying to quickly understand, and more\nimportantly, use the embed API, in a more complex way than just a simple script.\n\"\"\"\nimport time\nfrom threading import Thread\n\nimport numpy as np\nimport scipy.special\n\nfrom bokeh.embed import autoload_server\nfrom bokeh.models import GlyphRenderer\nfrom bokeh.plotting import (annular_wedge, cursession, figure, hold, legend,\n line, output_server, push, quad, xgrid, ygrid)\n\nfrom flask import Flask, render_template\napp = Flask(__name__)\n\[email protected]('/')\ndef render_plot():\n \"\"\"\n Get the script tags from each plot object and \"insert\" them into the template.\n\n This also starts different threads for each update function, so you can have\n a non-blocking update.\n \"\"\"\n dist_plot, dist_session = distribution()\n dist_tag = autoload_server(dist_plot, dist_session)\n\n anim_plot, anim_session = animated()\n anim_tag = autoload_server(anim_plot, anim_session)\n # for update_animation as target we need to pass the anim_plot and anim_session as args\n thread = Thread(target=update_animation, args=(anim_plot, anim_session))\n thread.start()\n\n pop = Population()\n pop_tag = autoload_server(pop.layout, pop.session)\n # for update_population as target we need to pass the pop instance as args\n thread = Thread(target=update_population, args=(pop,))\n thread.start()\n\n return render_template('app_plot.html', tag1=dist_tag, tag2=anim_tag, tag3=pop_tag)\n\n\ndef distribution():\n\n mu, sigma = 0, 0.5\n\n measured = np.random.normal(mu, sigma, 1000)\n hist, edges = np.histogram(measured, density=True, bins=20)\n\n x = np.linspace(-2, 2, 1000)\n pdf = 1 / (sigma * np.sqrt(2 * np.pi)) * np.exp(-(x - mu) ** 2 / (2 * sigma ** 2))\n cdf = (1 + scipy.special.erf((x - mu) / np.sqrt(2 * sigma ** 2))) / 2\n\n output_server(\"distribution_reveal\")\n\n p = figure(title=\"Interactive plots\",\n background_fill=\"#E5E5E5\")\n p.quad(top=hist, bottom=0, left=edges[:-1], right=edges[1:],\n fill_color=\"#333333\", line_color=\"#E5E5E5\", line_width=3)\n\n # Use `line` renderers to display the PDF and CDF\n p.line(x, pdf, line_color=\"#348abd\", line_width=8, alpha=0.7, legend=\"PDF\")\n p.line(x, cdf, line_color=\"#7a68a6\", line_width=8, alpha=0.7, legend=\"CDF\")\n\n p.legend.orientation = \"top_left\"\n\n p.xaxis.axis_label = 'x'\n p.xgrid[0].grid_line_color = \"white\"\n p.xgrid[0].grid_line_width = 3\n\n p.yaxis.axis_label = 'Pr(x)'\n p.ygrid[0].grid_line_color = \"white\"\n p.ygrid[0].grid_line_width = 3\n\n push()\n\n return p, cursession()\n\n\ndef animated():\n\n from numpy import pi, cos, sin, linspace\n\n N = 50 + 1\n r_base = 8\n theta = linspace(0, 2 * pi, N)\n r_x = linspace(0, 6 * pi, N - 1)\n rmin = r_base - cos(r_x) - 1\n rmax = r_base + sin(r_x) + 1\n\n colors = [\"FFFFCC\", \"#C7E9B4\", \"#7FCDBB\", \"#41B6C4\", \"#2C7FB8\",\n \"#253494\", \"#2C7FB8\", \"#41B6C4\", \"#7FCDBB\", \"#C7E9B4\"] * 5\n\n output_server(\"animated_reveal\")\n\n p = figure(title=\"Animations\", x_range=[-11, 11], y_range=[-11, 11])\n\n p.annular_wedge(\n 0, 0, rmin, rmax, theta[:-1], theta[1:],\n inner_radius_units=\"data\",\n outer_radius_units=\"data\",\n fill_color=colors,\n line_color=\"black\",\n )\n\n push()\n\n return p, cursession()\n\n\ndef update_animation(plot, session):\n\n from numpy import roll\n\n renderer = plot.select(dict(type=GlyphRenderer))\n ds = renderer[0].data_source\n\n while True:\n\n rmin = ds.data[\"inner_radius\"]\n rmin = roll(rmin, 1)\n ds.data[\"inner_radius\"] = rmin\n\n rmax = ds.data[\"outer_radius\"]\n rmax = roll(rmax, -1)\n ds.data[\"outer_radius\"] = rmax\n\n cursession().store_objects(ds)\n time.sleep(0.1)\n\n\nclass Population(object):\n\n year = 2010\n location = \"World\"\n\n def __init__(self):\n from bokeh.models import ColumnDataSource\n from bokeh.document import Document\n from bokeh.session import Session\n from bokeh.sampledata.population import load_population\n\n self.document = Document()\n self.session = Session()\n self.session.use_doc('population_reveal')\n self.session.load_document(self.document)\n\n self.df = load_population()\n self.source_pyramid = ColumnDataSource(data=dict())\n\n # just render at the initialization\n self._render()\n\n def _render(self):\n self.pyramid_plot()\n self.create_layout()\n self.document.add(self.layout)\n self.update_pyramid()\n\n def pyramid_plot(self):\n from bokeh.models import (Plot, DataRange1d, LinearAxis, Grid,\n Legend, SingleIntervalTicker)\n from bokeh.models.glyphs import Quad\n\n xdr = DataRange1d(sources=[self.source_pyramid.columns(\"male\"),\n self.source_pyramid.columns(\"female\")])\n ydr = DataRange1d(sources=[self.source_pyramid.columns(\"groups\")])\n\n self.plot = Plot(title=\"Widgets\", x_range=xdr, y_range=ydr,\n plot_width=600, plot_height=600)\n\n xaxis = LinearAxis()\n self.plot.add_layout(xaxis, 'below')\n yaxis = LinearAxis(ticker=SingleIntervalTicker(interval=5))\n self.plot.add_layout(yaxis, 'left')\n\n self.plot.add_layout(Grid(dimension=0, ticker=xaxis.ticker))\n self.plot.add_layout(Grid(dimension=1, ticker=yaxis.ticker))\n\n male_quad = Quad(left=\"male\", right=0, bottom=\"groups\", top=\"shifted\",\n fill_color=\"#3B8686\")\n male_quad_glyph = self.plot.add_glyph(self.source_pyramid, male_quad)\n\n female_quad = Quad(left=0, right=\"female\", bottom=\"groups\", top=\"shifted\",\n fill_color=\"#CFF09E\")\n female_quad_glyph = self.plot.add_glyph(self.source_pyramid, female_quad)\n\n self.plot.add_layout(Legend(legends=dict(Male=[male_quad_glyph],\n Female=[female_quad_glyph])))\n\n def on_year_change(self, obj, attr, old, new):\n self.year = int(new)\n self.update_pyramid()\n\n def on_location_change(self, obj, attr, old, new):\n self.location = new\n self.update_pyramid()\n\n def create_layout(self):\n from bokeh.models.widgets import Select, HBox, VBox\n\n years = list(map(str, sorted(self.df.Year.unique())))\n locations = sorted(self.df.Location.unique())\n\n year_select = Select(title=\"Year:\", value=\"2010\", options=years)\n location_select = Select(title=\"Location:\", value=\"World\", options=locations)\n\n year_select.on_change('value', self.on_year_change)\n location_select.on_change('value', self.on_location_change)\n\n controls = HBox(year_select, location_select)\n self.layout = VBox(controls, self.plot)\n\n def update_pyramid(self):\n pyramid = self.df[(self.df.Location == self.location) & (self.df.Year == self.year)]\n\n male = pyramid[pyramid.Sex == \"Male\"]\n female = pyramid[pyramid.Sex == \"Female\"]\n\n total = male.Value.sum() + female.Value.sum()\n\n male_percent = -male.Value / total\n female_percent = female.Value / total\n\n groups = male.AgeGrpStart.tolist()\n shifted = groups[1:] + [groups[-1] + 5]\n\n self.source_pyramid.data = dict(\n groups=groups,\n shifted=shifted,\n male=male_percent,\n female=female_percent,\n )\n self.session.store_document(self.document)\n\n\ndef update_population(plot):\n while True:\n plot.session.load_document(plot.document)\n time.sleep(0.1)\n\nif __name__ == '__main__':\n app.run(debug=True)\n", "path": "examples/embed/slideshow/app_reveal.py"}], "after_files": [{"content": "# -*- coding: utf-8 -*-\n\"\"\"\nIn this example, we want to show you how you can take isolated blocks of code\n(featuring different kinds of Bokeh visualizations) and rearrange them in a\nbigger (encompassing) flask-based application without losing the independence\nof each example. This is the reason of some weirdness through the code.\nWe are using this \"building blocks\" approach here because we believe it has some\nconceptual advantages for people trying to quickly understand, and more\nimportantly, use the embed API, in a more complex way than just a simple script.\n\"\"\"\nimport time\nfrom threading import Thread\n\nimport numpy as np\nimport scipy.special\n\nfrom bokeh.embed import autoload_server\nfrom bokeh.models import GlyphRenderer\nfrom bokeh.plotting import cursession, figure, output_server, push\n\nfrom flask import Flask, render_template\napp = Flask(__name__)\n\[email protected]('/')\ndef render_plot():\n \"\"\"\n Get the script tags from each plot object and \"insert\" them into the template.\n\n This also starts different threads for each update function, so you can have\n a non-blocking update.\n \"\"\"\n dist_plot, dist_session = distribution()\n dist_tag = autoload_server(dist_plot, dist_session)\n\n anim_plot, anim_session = animated()\n anim_tag = autoload_server(anim_plot, anim_session)\n # for update_animation as target we need to pass the anim_plot and anim_session as args\n thread = Thread(target=update_animation, args=(anim_plot, anim_session))\n thread.start()\n\n pop = Population()\n pop_tag = autoload_server(pop.layout, pop.session)\n # for update_population as target we need to pass the pop instance as args\n thread = Thread(target=update_population, args=(pop,))\n thread.start()\n\n return render_template('app_plot.html', tag1=dist_tag, tag2=anim_tag, tag3=pop_tag)\n\n\ndef distribution():\n\n mu, sigma = 0, 0.5\n\n measured = np.random.normal(mu, sigma, 1000)\n hist, edges = np.histogram(measured, density=True, bins=20)\n\n x = np.linspace(-2, 2, 1000)\n pdf = 1 / (sigma * np.sqrt(2 * np.pi)) * np.exp(-(x - mu) ** 2 / (2 * sigma ** 2))\n cdf = (1 + scipy.special.erf((x - mu) / np.sqrt(2 * sigma ** 2))) / 2\n\n output_server(\"distribution_reveal\")\n\n p = figure(title=\"Interactive plots\",\n background_fill=\"#E5E5E5\")\n p.quad(top=hist, bottom=0, left=edges[:-1], right=edges[1:],\n fill_color=\"#333333\", line_color=\"#E5E5E5\", line_width=3)\n\n # Use `line` renderers to display the PDF and CDF\n p.line(x, pdf, line_color=\"#348abd\", line_width=8, alpha=0.7, legend=\"PDF\")\n p.line(x, cdf, line_color=\"#7a68a6\", line_width=8, alpha=0.7, legend=\"CDF\")\n\n p.legend.orientation = \"top_left\"\n\n p.xaxis.axis_label = 'x'\n p.xgrid[0].grid_line_color = \"white\"\n p.xgrid[0].grid_line_width = 3\n\n p.yaxis.axis_label = 'Pr(x)'\n p.ygrid[0].grid_line_color = \"white\"\n p.ygrid[0].grid_line_width = 3\n\n push()\n\n return p, cursession()\n\n\ndef animated():\n\n from numpy import pi, cos, sin, linspace\n\n N = 50 + 1\n r_base = 8\n theta = linspace(0, 2 * pi, N)\n r_x = linspace(0, 6 * pi, N - 1)\n rmin = r_base - cos(r_x) - 1\n rmax = r_base + sin(r_x) + 1\n\n colors = [\"FFFFCC\", \"#C7E9B4\", \"#7FCDBB\", \"#41B6C4\", \"#2C7FB8\",\n \"#253494\", \"#2C7FB8\", \"#41B6C4\", \"#7FCDBB\", \"#C7E9B4\"] * 5\n\n output_server(\"animated_reveal\")\n\n p = figure(title=\"Animations\", x_range=[-11, 11], y_range=[-11, 11])\n\n p.annular_wedge(\n 0, 0, rmin, rmax, theta[:-1], theta[1:],\n inner_radius_units=\"data\",\n outer_radius_units=\"data\",\n fill_color=colors,\n line_color=\"black\",\n )\n\n push()\n\n return p, cursession()\n\n\ndef update_animation(plot, session):\n\n from numpy import roll\n\n renderer = plot.select(dict(type=GlyphRenderer))\n ds = renderer[0].data_source\n\n while True:\n\n rmin = ds.data[\"inner_radius\"]\n rmin = roll(rmin, 1)\n ds.data[\"inner_radius\"] = rmin\n\n rmax = ds.data[\"outer_radius\"]\n rmax = roll(rmax, -1)\n ds.data[\"outer_radius\"] = rmax\n\n cursession().store_objects(ds)\n time.sleep(0.1)\n\n\nclass Population(object):\n\n year = 2010\n location = \"World\"\n\n def __init__(self):\n from bokeh.models import ColumnDataSource\n from bokeh.document import Document\n from bokeh.session import Session\n from bokeh.sampledata.population import load_population\n\n self.document = Document()\n self.session = Session()\n self.session.use_doc('population_reveal')\n self.session.load_document(self.document)\n\n self.df = load_population()\n self.source_pyramid = ColumnDataSource(data=dict())\n\n # just render at the initialization\n self._render()\n\n def _render(self):\n self.pyramid_plot()\n self.create_layout()\n self.document.add(self.layout)\n self.update_pyramid()\n\n def pyramid_plot(self):\n from bokeh.models import (Plot, DataRange1d, LinearAxis, Grid,\n Legend, SingleIntervalTicker)\n from bokeh.models.glyphs import Quad\n\n xdr = DataRange1d(sources=[self.source_pyramid.columns(\"male\"),\n self.source_pyramid.columns(\"female\")])\n ydr = DataRange1d(sources=[self.source_pyramid.columns(\"groups\")])\n\n self.plot = Plot(title=\"Widgets\", x_range=xdr, y_range=ydr,\n plot_width=600, plot_height=600)\n\n xaxis = LinearAxis()\n self.plot.add_layout(xaxis, 'below')\n yaxis = LinearAxis(ticker=SingleIntervalTicker(interval=5))\n self.plot.add_layout(yaxis, 'left')\n\n self.plot.add_layout(Grid(dimension=0, ticker=xaxis.ticker))\n self.plot.add_layout(Grid(dimension=1, ticker=yaxis.ticker))\n\n male_quad = Quad(left=\"male\", right=0, bottom=\"groups\", top=\"shifted\",\n fill_color=\"#3B8686\")\n male_quad_glyph = self.plot.add_glyph(self.source_pyramid, male_quad)\n\n female_quad = Quad(left=0, right=\"female\", bottom=\"groups\", top=\"shifted\",\n fill_color=\"#CFF09E\")\n female_quad_glyph = self.plot.add_glyph(self.source_pyramid, female_quad)\n\n self.plot.add_layout(Legend(legends=dict(Male=[male_quad_glyph],\n Female=[female_quad_glyph])))\n\n def on_year_change(self, obj, attr, old, new):\n self.year = int(new)\n self.update_pyramid()\n\n def on_location_change(self, obj, attr, old, new):\n self.location = new\n self.update_pyramid()\n\n def create_layout(self):\n from bokeh.models.widgets import Select, HBox, VBox\n\n years = list(map(str, sorted(self.df.Year.unique())))\n locations = sorted(self.df.Location.unique())\n\n year_select = Select(title=\"Year:\", value=\"2010\", options=years)\n location_select = Select(title=\"Location:\", value=\"World\", options=locations)\n\n year_select.on_change('value', self.on_year_change)\n location_select.on_change('value', self.on_location_change)\n\n controls = HBox(year_select, location_select)\n self.layout = VBox(controls, self.plot)\n\n def update_pyramid(self):\n pyramid = self.df[(self.df.Location == self.location) & (self.df.Year == self.year)]\n\n male = pyramid[pyramid.Sex == \"Male\"]\n female = pyramid[pyramid.Sex == \"Female\"]\n\n total = male.Value.sum() + female.Value.sum()\n\n male_percent = -male.Value / total\n female_percent = female.Value / total\n\n groups = male.AgeGrpStart.tolist()\n shifted = groups[1:] + [groups[-1] + 5]\n\n self.source_pyramid.data = dict(\n groups=groups,\n shifted=shifted,\n male=male_percent,\n female=female_percent,\n )\n self.session.store_document(self.document)\n\n\ndef update_population(plot):\n while True:\n plot.session.load_document(plot.document)\n time.sleep(0.1)\n\nif __name__ == '__main__':\n app.run(debug=True)\n", "path": "examples/embed/slideshow/app_reveal.py"}]} | 3,144 | 142 |
gh_patches_debug_11096 | rasdani/github-patches | git_diff | azavea__raster-vision-1958 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Multi-temporal raster source visualizer fails when batch size is 1
https://github.com/azavea/raster-vision/blob/e4e10ad04313bbe5355693ef96f3854f7963f2b1/rastervision_pytorch_learner/rastervision/pytorch_learner/dataset/visualizer/visualizer.py#L122-L127
This code fails when the batch size is 1 because the created `subfigs` doesn't have a `flat` property if there's only one row and one column. Not sure whether this should be fixed upstream in `matplotlib`...
Matplotlib version 3.7.1, rastervision version 0.21.2
```python
import matplotlib.pyplot as plt
fig = plt.figure()
subfigs = fig.subfigures(nrows=2, ncols=1, hspace=0)
subfigs.flat
#> <numpy.flatiter object at 0x5575c63e68f0>
subfigs = fig.subfigures(nrows=1, ncols=1, hspace=0)
subfigs.flat
#> Traceback (most recent call last):
#> File "<string>", line 1, in <module>
#> AttributeError: 'SubFigure' object has no attribute 'flat'
```
<sup>Created at 2023-10-11 17:12:31 CDT by [reprexlite](https://github.com/jayqi/reprexlite) v0.5.0</sup>
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `rastervision_pytorch_learner/rastervision/pytorch_learner/dataset/visualizer/visualizer.py`
Content:
```
1 from typing import (TYPE_CHECKING, Sequence, Optional, List, Dict, Union,
2 Tuple, Any)
3 from abc import ABC, abstractmethod
4
5 import numpy as np
6 import torch
7 from torch import Tensor
8 import albumentations as A
9 from torch.utils.data import DataLoader
10 import matplotlib.pyplot as plt
11
12 from rastervision.pipeline.file_system import make_dir
13 from rastervision.pytorch_learner.utils import (
14 deserialize_albumentation_transform, validate_albumentation_transform,
15 MinMaxNormalize)
16 from rastervision.pytorch_learner.learner_config import (
17 RGBTuple,
18 ChannelInds,
19 ensure_class_colors,
20 validate_channel_display_groups,
21 get_default_channel_display_groups,
22 )
23
24 if TYPE_CHECKING:
25 from torch.utils.data import Dataset
26 from matplotlib.figure import Figure
27
28
29 class Visualizer(ABC):
30 """Base class for plotting samples from computer vision PyTorch Datasets."""
31
32 scale: float = 3.
33
34 def __init__(self,
35 class_names: List[str],
36 class_colors: Optional[List[Union[str, RGBTuple]]] = None,
37 transform: Optional[Dict] = A.to_dict(MinMaxNormalize()),
38 channel_display_groups: Optional[Union[Dict[
39 str, ChannelInds], Sequence[ChannelInds]]] = None):
40 """Constructor.
41
42 Args:
43 class_names: names of classes
44 class_colors: Colors used to display classes. Can be color 3-tuples
45 in list form.
46 transform: An Albumentations transform serialized as a dict that
47 will be applied to each image before it is plotted. Mainly useful
48 for undoing any data transformation that you do not want included in
49 the plot, such as normalization. The default value will shift and scale
50 the image so the values range from 0.0 to 1.0 which is the expected range
51 for the plotting function. This default is useful for cases where the
52 values after normalization are close to zero which makes the plot
53 difficult to see.
54 channel_display_groups: Groups of image channels to display together as a
55 subplot when plotting the data and predictions.
56 Can be a list or tuple of groups (e.g. [(0, 1, 2), (3,)]) or a
57 dict containing title-to-group mappings
58 (e.g. {"RGB": [0, 1, 2], "IR": [3]}),
59 where each group is a list or tuple of channel indices and
60 title is a string that will be used as the title of the subplot
61 for that group.
62 """
63 self.class_names = class_names
64 self.class_colors = ensure_class_colors(self.class_names, class_colors)
65 self.transform = validate_albumentation_transform(transform)
66 self._channel_display_groups = validate_channel_display_groups(
67 channel_display_groups)
68
69 @abstractmethod
70 def plot_xyz(self,
71 axs,
72 x: Tensor,
73 y: Sequence,
74 z: Optional[Sequence] = None,
75 plot_title: bool = True):
76 """Plot image, ground truth labels, and predicted labels.
77
78 Args:
79 axs: matplotlib axes on which to plot
80 x: image
81 y: ground truth labels
82 z: optional predicted labels
83 """
84 pass
85
86 def plot_batch(self,
87 x: Tensor,
88 y: Sequence,
89 output_path: Optional[str] = None,
90 z: Optional[Sequence] = None,
91 batch_limit: Optional[int] = None,
92 show: bool = False):
93 """Plot a whole batch in a grid using plot_xyz.
94
95 Args:
96 x: batch of images
97 y: ground truth labels
98 output_path: local path where to save plot image
99 z: optional predicted labels
100 batch_limit: optional limit on (rendered) batch size
101 """
102 params = self.get_plot_params(
103 x=x, y=y, z=z, output_path=output_path, batch_limit=batch_limit)
104 if params['subplot_args']['nrows'] == 0:
105 return
106
107 if x.ndim == 4:
108 fig, axs = plt.subplots(**params['fig_args'],
109 **params['subplot_args'])
110 plot_xyz_args = params['plot_xyz_args']
111 self._plot_batch(fig, axs, plot_xyz_args, x, y=y, z=z)
112 elif x.ndim == 5:
113 # If a temporal dimension is present, we divide the figure into
114 # multiple subfigures--one for each batch. Then, in each subfigure,
115 # we plot all timesteps as if they were a single batch. To
116 # delineate the boundary b/w batch items, we adopt the convention
117 # of only displaying subplot titles once per batch (above the first
118 # row in each batch).
119 batch_sz, T, *_ = x.shape
120 params['fig_args']['figsize'][1] *= T
121 fig = plt.figure(**params['fig_args'])
122 subfigs = fig.subfigures(nrows=batch_sz, ncols=1, hspace=0.0)
123 subfig_axs = [
124 subfig.subplots(
125 nrows=T, ncols=params['subplot_args']['ncols'])
126 for subfig in subfigs.flat
127 ]
128 for i, axs in enumerate(subfig_axs):
129 plot_xyz_args = [
130 dict(params['plot_xyz_args'][i]) for _ in range(T)
131 ]
132 plot_xyz_args[0]['plot_title'] = True
133 for args in plot_xyz_args[1:]:
134 args['plot_title'] = False
135 _x = x[i]
136 _y = [y[i]] * T
137 _z = None if z is None else [z[i]] * T
138 self._plot_batch(fig, axs, plot_xyz_args, _x, y=_y, z=_z)
139 else:
140 raise ValueError('Expected x to have 4 or 5 dims, but found '
141 f'x.shape: {x.shape}')
142
143 if show:
144 plt.show()
145 if output_path is not None:
146 make_dir(output_path, use_dirname=True)
147 fig.savefig(output_path, bbox_inches='tight', pad_inches=0.2)
148
149 plt.close(fig)
150
151 def _plot_batch(
152 self,
153 fig: 'Figure',
154 axs: Sequence,
155 plot_xyz_args: List[dict],
156 x: Tensor,
157 y: Optional[Sequence] = None,
158 z: Optional[Sequence] = None,
159 ):
160 # (N, c, h, w) --> (N, h, w, c)
161 x = x.permute(0, 2, 3, 1)
162
163 # apply transform, if given
164 if self.transform is not None:
165 tf = deserialize_albumentation_transform(self.transform)
166 imgs = [tf(image=img)['image'] for img in x.numpy()]
167 x = torch.from_numpy(np.stack(imgs))
168
169 for i, row_axs in enumerate(axs):
170 _z = None if z is None else z[i]
171 self.plot_xyz(row_axs, x[i], y[i], z=_z, **plot_xyz_args[i])
172
173 def get_channel_display_groups(
174 self, nb_img_channels: int
175 ) -> Union[Dict[str, ChannelInds], Sequence[ChannelInds]]:
176 # The default channel_display_groups object depends on the number of
177 # channels in the image. This number is not known when the Visualizer
178 # is constructed which is why it needs to be created later.
179 if self._channel_display_groups is not None:
180 return self._channel_display_groups
181 return get_default_channel_display_groups(nb_img_channels)
182
183 def get_collate_fn(self) -> Optional[callable]:
184 """Returns a custom collate_fn to use in DataLoader.
185
186 None is returned if default collate_fn should be used.
187
188 See https://pytorch.org/docs/stable/data.html#working-with-collate-fn
189 """
190 return None
191
192 def get_batch(self, dataset: 'Dataset', batch_sz: int = 4,
193 **kwargs) -> Tuple[Tensor, Any]:
194 """Generate a batch from a dataset.
195
196 This is a convenience method for generating a batch of data to plot.
197
198 Args:
199 dataset (Dataset): A Pytorch Datset.
200 batch_sz (int): Batch size. Defaults to 4.
201 **kwargs: Extra args for :class:`~torch.utils.data.DataLoader`.
202
203 Returns:
204 Tuple[Tensor, Any]: (x, y) tuple where x is images and y is labels.
205 """
206 collate_fn = self.get_collate_fn()
207 dl = DataLoader(dataset, batch_sz, collate_fn=collate_fn, **kwargs)
208 try:
209 x, y = next(iter(dl))
210 except StopIteration:
211 raise ValueError('dataset did not return a batch')
212
213 return x, y
214
215 def get_plot_nrows(self, **kwargs) -> int:
216 x = kwargs['x']
217 batch_limit = kwargs.get('batch_limit')
218 batch_sz = x.shape[0]
219 nrows = min(batch_sz,
220 batch_limit) if batch_limit is not None else batch_sz
221 return nrows
222
223 def get_plot_ncols(self, **kwargs) -> int:
224 x = kwargs['x']
225 nb_img_channels = x.shape[1]
226 ncols = len(self.get_channel_display_groups(nb_img_channels))
227 return ncols
228
229 def get_plot_params(self, **kwargs) -> dict:
230 nrows = self.get_plot_nrows(**kwargs)
231 ncols = self.get_plot_ncols(**kwargs)
232 params = {
233 'fig_args': {
234 'constrained_layout': True,
235 'figsize': np.array((self.scale * ncols, self.scale * nrows)),
236 },
237 'subplot_args': {
238 'nrows': nrows,
239 'ncols': ncols,
240 'squeeze': False
241 },
242 'plot_xyz_args': [{} for _ in range(nrows)]
243 }
244 return params
245
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/rastervision_pytorch_learner/rastervision/pytorch_learner/dataset/visualizer/visualizer.py b/rastervision_pytorch_learner/rastervision/pytorch_learner/dataset/visualizer/visualizer.py
--- a/rastervision_pytorch_learner/rastervision/pytorch_learner/dataset/visualizer/visualizer.py
+++ b/rastervision_pytorch_learner/rastervision/pytorch_learner/dataset/visualizer/visualizer.py
@@ -119,7 +119,8 @@
batch_sz, T, *_ = x.shape
params['fig_args']['figsize'][1] *= T
fig = plt.figure(**params['fig_args'])
- subfigs = fig.subfigures(nrows=batch_sz, ncols=1, hspace=0.0)
+ subfigs = fig.subfigures(
+ nrows=batch_sz, ncols=1, hspace=0.0, squeeze=False)
subfig_axs = [
subfig.subplots(
nrows=T, ncols=params['subplot_args']['ncols'])
| {"golden_diff": "diff --git a/rastervision_pytorch_learner/rastervision/pytorch_learner/dataset/visualizer/visualizer.py b/rastervision_pytorch_learner/rastervision/pytorch_learner/dataset/visualizer/visualizer.py\n--- a/rastervision_pytorch_learner/rastervision/pytorch_learner/dataset/visualizer/visualizer.py\n+++ b/rastervision_pytorch_learner/rastervision/pytorch_learner/dataset/visualizer/visualizer.py\n@@ -119,7 +119,8 @@\n batch_sz, T, *_ = x.shape\n params['fig_args']['figsize'][1] *= T\n fig = plt.figure(**params['fig_args'])\n- subfigs = fig.subfigures(nrows=batch_sz, ncols=1, hspace=0.0)\n+ subfigs = fig.subfigures(\n+ nrows=batch_sz, ncols=1, hspace=0.0, squeeze=False)\n subfig_axs = [\n subfig.subplots(\n nrows=T, ncols=params['subplot_args']['ncols'])\n", "issue": "Multi-temporal raster source visualizer fails when batch size is 1\nhttps://github.com/azavea/raster-vision/blob/e4e10ad04313bbe5355693ef96f3854f7963f2b1/rastervision_pytorch_learner/rastervision/pytorch_learner/dataset/visualizer/visualizer.py#L122-L127\r\n\r\nThis code fails when the batch size is 1 because the created `subfigs` doesn't have a `flat` property if there's only one row and one column. Not sure whether this should be fixed upstream in `matplotlib`...\r\n\r\nMatplotlib version 3.7.1, rastervision version 0.21.2\r\n\r\n```python\r\nimport matplotlib.pyplot as plt\r\nfig = plt.figure()\r\nsubfigs = fig.subfigures(nrows=2, ncols=1, hspace=0)\r\nsubfigs.flat\r\n#> <numpy.flatiter object at 0x5575c63e68f0>\r\n\r\nsubfigs = fig.subfigures(nrows=1, ncols=1, hspace=0)\r\nsubfigs.flat\r\n#> Traceback (most recent call last):\r\n#> File \"<string>\", line 1, in <module>\r\n#> AttributeError: 'SubFigure' object has no attribute 'flat'\r\n```\r\n\r\n<sup>Created at 2023-10-11 17:12:31 CDT by [reprexlite](https://github.com/jayqi/reprexlite) v0.5.0</sup>\r\n\n", "before_files": [{"content": "from typing import (TYPE_CHECKING, Sequence, Optional, List, Dict, Union,\n Tuple, Any)\nfrom abc import ABC, abstractmethod\n\nimport numpy as np\nimport torch\nfrom torch import Tensor\nimport albumentations as A\nfrom torch.utils.data import DataLoader\nimport matplotlib.pyplot as plt\n\nfrom rastervision.pipeline.file_system import make_dir\nfrom rastervision.pytorch_learner.utils import (\n deserialize_albumentation_transform, validate_albumentation_transform,\n MinMaxNormalize)\nfrom rastervision.pytorch_learner.learner_config import (\n RGBTuple,\n ChannelInds,\n ensure_class_colors,\n validate_channel_display_groups,\n get_default_channel_display_groups,\n)\n\nif TYPE_CHECKING:\n from torch.utils.data import Dataset\n from matplotlib.figure import Figure\n\n\nclass Visualizer(ABC):\n \"\"\"Base class for plotting samples from computer vision PyTorch Datasets.\"\"\"\n\n scale: float = 3.\n\n def __init__(self,\n class_names: List[str],\n class_colors: Optional[List[Union[str, RGBTuple]]] = None,\n transform: Optional[Dict] = A.to_dict(MinMaxNormalize()),\n channel_display_groups: Optional[Union[Dict[\n str, ChannelInds], Sequence[ChannelInds]]] = None):\n \"\"\"Constructor.\n\n Args:\n class_names: names of classes\n class_colors: Colors used to display classes. Can be color 3-tuples\n in list form.\n transform: An Albumentations transform serialized as a dict that\n will be applied to each image before it is plotted. Mainly useful\n for undoing any data transformation that you do not want included in\n the plot, such as normalization. The default value will shift and scale\n the image so the values range from 0.0 to 1.0 which is the expected range\n for the plotting function. This default is useful for cases where the\n values after normalization are close to zero which makes the plot\n difficult to see.\n channel_display_groups: Groups of image channels to display together as a\n subplot when plotting the data and predictions.\n Can be a list or tuple of groups (e.g. [(0, 1, 2), (3,)]) or a\n dict containing title-to-group mappings\n (e.g. {\"RGB\": [0, 1, 2], \"IR\": [3]}),\n where each group is a list or tuple of channel indices and\n title is a string that will be used as the title of the subplot\n for that group.\n \"\"\"\n self.class_names = class_names\n self.class_colors = ensure_class_colors(self.class_names, class_colors)\n self.transform = validate_albumentation_transform(transform)\n self._channel_display_groups = validate_channel_display_groups(\n channel_display_groups)\n\n @abstractmethod\n def plot_xyz(self,\n axs,\n x: Tensor,\n y: Sequence,\n z: Optional[Sequence] = None,\n plot_title: bool = True):\n \"\"\"Plot image, ground truth labels, and predicted labels.\n\n Args:\n axs: matplotlib axes on which to plot\n x: image\n y: ground truth labels\n z: optional predicted labels\n \"\"\"\n pass\n\n def plot_batch(self,\n x: Tensor,\n y: Sequence,\n output_path: Optional[str] = None,\n z: Optional[Sequence] = None,\n batch_limit: Optional[int] = None,\n show: bool = False):\n \"\"\"Plot a whole batch in a grid using plot_xyz.\n\n Args:\n x: batch of images\n y: ground truth labels\n output_path: local path where to save plot image\n z: optional predicted labels\n batch_limit: optional limit on (rendered) batch size\n \"\"\"\n params = self.get_plot_params(\n x=x, y=y, z=z, output_path=output_path, batch_limit=batch_limit)\n if params['subplot_args']['nrows'] == 0:\n return\n\n if x.ndim == 4:\n fig, axs = plt.subplots(**params['fig_args'],\n **params['subplot_args'])\n plot_xyz_args = params['plot_xyz_args']\n self._plot_batch(fig, axs, plot_xyz_args, x, y=y, z=z)\n elif x.ndim == 5:\n # If a temporal dimension is present, we divide the figure into\n # multiple subfigures--one for each batch. Then, in each subfigure,\n # we plot all timesteps as if they were a single batch. To\n # delineate the boundary b/w batch items, we adopt the convention\n # of only displaying subplot titles once per batch (above the first\n # row in each batch).\n batch_sz, T, *_ = x.shape\n params['fig_args']['figsize'][1] *= T\n fig = plt.figure(**params['fig_args'])\n subfigs = fig.subfigures(nrows=batch_sz, ncols=1, hspace=0.0)\n subfig_axs = [\n subfig.subplots(\n nrows=T, ncols=params['subplot_args']['ncols'])\n for subfig in subfigs.flat\n ]\n for i, axs in enumerate(subfig_axs):\n plot_xyz_args = [\n dict(params['plot_xyz_args'][i]) for _ in range(T)\n ]\n plot_xyz_args[0]['plot_title'] = True\n for args in plot_xyz_args[1:]:\n args['plot_title'] = False\n _x = x[i]\n _y = [y[i]] * T\n _z = None if z is None else [z[i]] * T\n self._plot_batch(fig, axs, plot_xyz_args, _x, y=_y, z=_z)\n else:\n raise ValueError('Expected x to have 4 or 5 dims, but found '\n f'x.shape: {x.shape}')\n\n if show:\n plt.show()\n if output_path is not None:\n make_dir(output_path, use_dirname=True)\n fig.savefig(output_path, bbox_inches='tight', pad_inches=0.2)\n\n plt.close(fig)\n\n def _plot_batch(\n self,\n fig: 'Figure',\n axs: Sequence,\n plot_xyz_args: List[dict],\n x: Tensor,\n y: Optional[Sequence] = None,\n z: Optional[Sequence] = None,\n ):\n # (N, c, h, w) --> (N, h, w, c)\n x = x.permute(0, 2, 3, 1)\n\n # apply transform, if given\n if self.transform is not None:\n tf = deserialize_albumentation_transform(self.transform)\n imgs = [tf(image=img)['image'] for img in x.numpy()]\n x = torch.from_numpy(np.stack(imgs))\n\n for i, row_axs in enumerate(axs):\n _z = None if z is None else z[i]\n self.plot_xyz(row_axs, x[i], y[i], z=_z, **plot_xyz_args[i])\n\n def get_channel_display_groups(\n self, nb_img_channels: int\n ) -> Union[Dict[str, ChannelInds], Sequence[ChannelInds]]:\n # The default channel_display_groups object depends on the number of\n # channels in the image. This number is not known when the Visualizer\n # is constructed which is why it needs to be created later.\n if self._channel_display_groups is not None:\n return self._channel_display_groups\n return get_default_channel_display_groups(nb_img_channels)\n\n def get_collate_fn(self) -> Optional[callable]:\n \"\"\"Returns a custom collate_fn to use in DataLoader.\n\n None is returned if default collate_fn should be used.\n\n See https://pytorch.org/docs/stable/data.html#working-with-collate-fn\n \"\"\"\n return None\n\n def get_batch(self, dataset: 'Dataset', batch_sz: int = 4,\n **kwargs) -> Tuple[Tensor, Any]:\n \"\"\"Generate a batch from a dataset.\n\n This is a convenience method for generating a batch of data to plot.\n\n Args:\n dataset (Dataset): A Pytorch Datset.\n batch_sz (int): Batch size. Defaults to 4.\n **kwargs: Extra args for :class:`~torch.utils.data.DataLoader`.\n\n Returns:\n Tuple[Tensor, Any]: (x, y) tuple where x is images and y is labels.\n \"\"\"\n collate_fn = self.get_collate_fn()\n dl = DataLoader(dataset, batch_sz, collate_fn=collate_fn, **kwargs)\n try:\n x, y = next(iter(dl))\n except StopIteration:\n raise ValueError('dataset did not return a batch')\n\n return x, y\n\n def get_plot_nrows(self, **kwargs) -> int:\n x = kwargs['x']\n batch_limit = kwargs.get('batch_limit')\n batch_sz = x.shape[0]\n nrows = min(batch_sz,\n batch_limit) if batch_limit is not None else batch_sz\n return nrows\n\n def get_plot_ncols(self, **kwargs) -> int:\n x = kwargs['x']\n nb_img_channels = x.shape[1]\n ncols = len(self.get_channel_display_groups(nb_img_channels))\n return ncols\n\n def get_plot_params(self, **kwargs) -> dict:\n nrows = self.get_plot_nrows(**kwargs)\n ncols = self.get_plot_ncols(**kwargs)\n params = {\n 'fig_args': {\n 'constrained_layout': True,\n 'figsize': np.array((self.scale * ncols, self.scale * nrows)),\n },\n 'subplot_args': {\n 'nrows': nrows,\n 'ncols': ncols,\n 'squeeze': False\n },\n 'plot_xyz_args': [{} for _ in range(nrows)]\n }\n return params\n", "path": "rastervision_pytorch_learner/rastervision/pytorch_learner/dataset/visualizer/visualizer.py"}], "after_files": [{"content": "from typing import (TYPE_CHECKING, Sequence, Optional, List, Dict, Union,\n Tuple, Any)\nfrom abc import ABC, abstractmethod\n\nimport numpy as np\nimport torch\nfrom torch import Tensor\nimport albumentations as A\nfrom torch.utils.data import DataLoader\nimport matplotlib.pyplot as plt\n\nfrom rastervision.pipeline.file_system import make_dir\nfrom rastervision.pytorch_learner.utils import (\n deserialize_albumentation_transform, validate_albumentation_transform,\n MinMaxNormalize)\nfrom rastervision.pytorch_learner.learner_config import (\n RGBTuple,\n ChannelInds,\n ensure_class_colors,\n validate_channel_display_groups,\n get_default_channel_display_groups,\n)\n\nif TYPE_CHECKING:\n from torch.utils.data import Dataset\n from matplotlib.figure import Figure\n\n\nclass Visualizer(ABC):\n \"\"\"Base class for plotting samples from computer vision PyTorch Datasets.\"\"\"\n\n scale: float = 3.\n\n def __init__(self,\n class_names: List[str],\n class_colors: Optional[List[Union[str, RGBTuple]]] = None,\n transform: Optional[Dict] = A.to_dict(MinMaxNormalize()),\n channel_display_groups: Optional[Union[Dict[\n str, ChannelInds], Sequence[ChannelInds]]] = None):\n \"\"\"Constructor.\n\n Args:\n class_names: names of classes\n class_colors: Colors used to display classes. Can be color 3-tuples\n in list form.\n transform: An Albumentations transform serialized as a dict that\n will be applied to each image before it is plotted. Mainly useful\n for undoing any data transformation that you do not want included in\n the plot, such as normalization. The default value will shift and scale\n the image so the values range from 0.0 to 1.0 which is the expected range\n for the plotting function. This default is useful for cases where the\n values after normalization are close to zero which makes the plot\n difficult to see.\n channel_display_groups: Groups of image channels to display together as a\n subplot when plotting the data and predictions.\n Can be a list or tuple of groups (e.g. [(0, 1, 2), (3,)]) or a\n dict containing title-to-group mappings\n (e.g. {\"RGB\": [0, 1, 2], \"IR\": [3]}),\n where each group is a list or tuple of channel indices and\n title is a string that will be used as the title of the subplot\n for that group.\n \"\"\"\n self.class_names = class_names\n self.class_colors = ensure_class_colors(self.class_names, class_colors)\n self.transform = validate_albumentation_transform(transform)\n self._channel_display_groups = validate_channel_display_groups(\n channel_display_groups)\n\n @abstractmethod\n def plot_xyz(self,\n axs,\n x: Tensor,\n y: Sequence,\n z: Optional[Sequence] = None,\n plot_title: bool = True):\n \"\"\"Plot image, ground truth labels, and predicted labels.\n\n Args:\n axs: matplotlib axes on which to plot\n x: image\n y: ground truth labels\n z: optional predicted labels\n \"\"\"\n pass\n\n def plot_batch(self,\n x: Tensor,\n y: Sequence,\n output_path: Optional[str] = None,\n z: Optional[Sequence] = None,\n batch_limit: Optional[int] = None,\n show: bool = False):\n \"\"\"Plot a whole batch in a grid using plot_xyz.\n\n Args:\n x: batch of images\n y: ground truth labels\n output_path: local path where to save plot image\n z: optional predicted labels\n batch_limit: optional limit on (rendered) batch size\n \"\"\"\n params = self.get_plot_params(\n x=x, y=y, z=z, output_path=output_path, batch_limit=batch_limit)\n if params['subplot_args']['nrows'] == 0:\n return\n\n if x.ndim == 4:\n fig, axs = plt.subplots(**params['fig_args'],\n **params['subplot_args'])\n plot_xyz_args = params['plot_xyz_args']\n self._plot_batch(fig, axs, plot_xyz_args, x, y=y, z=z)\n elif x.ndim == 5:\n # If a temporal dimension is present, we divide the figure into\n # multiple subfigures--one for each batch. Then, in each subfigure,\n # we plot all timesteps as if they were a single batch. To\n # delineate the boundary b/w batch items, we adopt the convention\n # of only displaying subplot titles once per batch (above the first\n # row in each batch).\n batch_sz, T, *_ = x.shape\n params['fig_args']['figsize'][1] *= T\n fig = plt.figure(**params['fig_args'])\n subfigs = fig.subfigures(\n nrows=batch_sz, ncols=1, hspace=0.0, squeeze=False)\n subfig_axs = [\n subfig.subplots(\n nrows=T, ncols=params['subplot_args']['ncols'])\n for subfig in subfigs.flat\n ]\n for i, axs in enumerate(subfig_axs):\n plot_xyz_args = [\n dict(params['plot_xyz_args'][i]) for _ in range(T)\n ]\n plot_xyz_args[0]['plot_title'] = True\n for args in plot_xyz_args[1:]:\n args['plot_title'] = False\n _x = x[i]\n _y = [y[i]] * T\n _z = None if z is None else [z[i]] * T\n self._plot_batch(fig, axs, plot_xyz_args, _x, y=_y, z=_z)\n else:\n raise ValueError('Expected x to have 4 or 5 dims, but found '\n f'x.shape: {x.shape}')\n\n if show:\n plt.show()\n if output_path is not None:\n make_dir(output_path, use_dirname=True)\n fig.savefig(output_path, bbox_inches='tight', pad_inches=0.2)\n\n plt.close(fig)\n\n def _plot_batch(\n self,\n fig: 'Figure',\n axs: Sequence,\n plot_xyz_args: List[dict],\n x: Tensor,\n y: Optional[Sequence] = None,\n z: Optional[Sequence] = None,\n ):\n # (N, c, h, w) --> (N, h, w, c)\n x = x.permute(0, 2, 3, 1)\n\n # apply transform, if given\n if self.transform is not None:\n tf = deserialize_albumentation_transform(self.transform)\n imgs = [tf(image=img)['image'] for img in x.numpy()]\n x = torch.from_numpy(np.stack(imgs))\n\n for i, row_axs in enumerate(axs):\n _z = None if z is None else z[i]\n self.plot_xyz(row_axs, x[i], y[i], z=_z, **plot_xyz_args[i])\n\n def get_channel_display_groups(\n self, nb_img_channels: int\n ) -> Union[Dict[str, ChannelInds], Sequence[ChannelInds]]:\n # The default channel_display_groups object depends on the number of\n # channels in the image. This number is not known when the Visualizer\n # is constructed which is why it needs to be created later.\n if self._channel_display_groups is not None:\n return self._channel_display_groups\n return get_default_channel_display_groups(nb_img_channels)\n\n def get_collate_fn(self) -> Optional[callable]:\n \"\"\"Returns a custom collate_fn to use in DataLoader.\n\n None is returned if default collate_fn should be used.\n\n See https://pytorch.org/docs/stable/data.html#working-with-collate-fn\n \"\"\"\n return None\n\n def get_batch(self, dataset: 'Dataset', batch_sz: int = 4,\n **kwargs) -> Tuple[Tensor, Any]:\n \"\"\"Generate a batch from a dataset.\n\n This is a convenience method for generating a batch of data to plot.\n\n Args:\n dataset (Dataset): A Pytorch Datset.\n batch_sz (int): Batch size. Defaults to 4.\n **kwargs: Extra args for :class:`~torch.utils.data.DataLoader`.\n\n Returns:\n Tuple[Tensor, Any]: (x, y) tuple where x is images and y is labels.\n \"\"\"\n collate_fn = self.get_collate_fn()\n dl = DataLoader(dataset, batch_sz, collate_fn=collate_fn, **kwargs)\n try:\n x, y = next(iter(dl))\n except StopIteration:\n raise ValueError('dataset did not return a batch')\n\n return x, y\n\n def get_plot_nrows(self, **kwargs) -> int:\n x = kwargs['x']\n batch_limit = kwargs.get('batch_limit')\n batch_sz = x.shape[0]\n nrows = min(batch_sz,\n batch_limit) if batch_limit is not None else batch_sz\n return nrows\n\n def get_plot_ncols(self, **kwargs) -> int:\n x = kwargs['x']\n nb_img_channels = x.shape[1]\n ncols = len(self.get_channel_display_groups(nb_img_channels))\n return ncols\n\n def get_plot_params(self, **kwargs) -> dict:\n nrows = self.get_plot_nrows(**kwargs)\n ncols = self.get_plot_ncols(**kwargs)\n params = {\n 'fig_args': {\n 'constrained_layout': True,\n 'figsize': np.array((self.scale * ncols, self.scale * nrows)),\n },\n 'subplot_args': {\n 'nrows': nrows,\n 'ncols': ncols,\n 'squeeze': False\n },\n 'plot_xyz_args': [{} for _ in range(nrows)]\n }\n return params\n", "path": "rastervision_pytorch_learner/rastervision/pytorch_learner/dataset/visualizer/visualizer.py"}]} | 3,415 | 242 |
gh_patches_debug_143 | rasdani/github-patches | git_diff | ManimCommunity__manim-126 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Remove argparse from setup.py
https://github.com/ManimCommunity/manim/blob/cf8c5b9938abafba9f6c2c1aeff9e15c8edbfdd1/setup.py#L17
Remove `argparse` from setup.py as it is a default library and need not be mentioned in `requirements.txt` and `setup.py`.
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `setup.py`
Content:
```
1 from setuptools import setup, find_namespace_packages
2
3 setup(
4 name="manimlib",
5 version="0.2.0",
6 description="Animation engine for explanatory math videos",
7 license="MIT",
8 packages=find_namespace_packages(),
9 package_data={ "manim": ["*.tex"] },
10 entry_points={
11 "console_scripts": [
12 "manim=manim.__main__:main",
13 "manimcm=manim.__main__:main",
14 ]
15 },
16 install_requires=[
17 "argparse",
18 "colour",
19 "numpy",
20 "Pillow",
21 "progressbar",
22 "scipy",
23 "tqdm",
24 "pycairo",
25 "pydub",
26 "pygments",
27 "pyreadline; sys_platform == 'win32'",
28 "rich",
29 ],
30 )
31
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/setup.py b/setup.py
--- a/setup.py
+++ b/setup.py
@@ -14,7 +14,6 @@
]
},
install_requires=[
- "argparse",
"colour",
"numpy",
"Pillow",
| {"golden_diff": "diff --git a/setup.py b/setup.py\n--- a/setup.py\n+++ b/setup.py\n@@ -14,7 +14,6 @@\n ]\n },\n install_requires=[\n- \"argparse\",\n \"colour\",\n \"numpy\",\n \"Pillow\",\n", "issue": "Remove argparse from setup.py\nhttps://github.com/ManimCommunity/manim/blob/cf8c5b9938abafba9f6c2c1aeff9e15c8edbfdd1/setup.py#L17\r\nRemove `argparse` from setup.py as it is a default library and need not be mentioned in `requirements.txt` and `setup.py`.\n", "before_files": [{"content": "from setuptools import setup, find_namespace_packages\n\nsetup(\n name=\"manimlib\",\n version=\"0.2.0\",\n description=\"Animation engine for explanatory math videos\",\n license=\"MIT\",\n packages=find_namespace_packages(),\n package_data={ \"manim\": [\"*.tex\"] },\n entry_points={\n \"console_scripts\": [\n \"manim=manim.__main__:main\",\n \"manimcm=manim.__main__:main\",\n ]\n },\n install_requires=[\n \"argparse\",\n \"colour\",\n \"numpy\",\n \"Pillow\",\n \"progressbar\",\n \"scipy\",\n \"tqdm\",\n \"pycairo\",\n \"pydub\",\n \"pygments\",\n \"pyreadline; sys_platform == 'win32'\",\n \"rich\",\n ],\n)\n", "path": "setup.py"}], "after_files": [{"content": "from setuptools import setup, find_namespace_packages\n\nsetup(\n name=\"manimlib\",\n version=\"0.2.0\",\n description=\"Animation engine for explanatory math videos\",\n license=\"MIT\",\n packages=find_namespace_packages(),\n package_data={ \"manim\": [\"*.tex\"] },\n entry_points={\n \"console_scripts\": [\n \"manim=manim.__main__:main\",\n \"manimcm=manim.__main__:main\",\n ]\n },\n install_requires=[\n \"colour\",\n \"numpy\",\n \"Pillow\",\n \"progressbar\",\n \"scipy\",\n \"tqdm\",\n \"pycairo\",\n \"pydub\",\n \"pygments\",\n \"pyreadline; sys_platform == 'win32'\",\n \"rich\",\n ],\n)\n", "path": "setup.py"}]} | 569 | 59 |
gh_patches_debug_9582 | rasdani/github-patches | git_diff | mampfes__hacs_waste_collection_schedule-596 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Error with Service Provider Stadtreinigung Leipzig / stadtreinigung-leipzig.de
Hi everyone,
Since 03.01.2023 (this is where I noticed it), WCS can no longer retrieve data from Stadtwerke Leipzig.
The following error is displayed:
fetch failed for source Stadtreinigung Leipzig: Traceback (most recent call last): File "/config/custom_components/waste_collection_schedule/waste_collection_schedule/source_shell.py", line 134, in fetch entries = self._source.fetch() File "/config/custom_components/waste_collection_schedule/waste_collection_schedule/source/stadtreinigung_leipzig_de.py", line 34, in fetch raise Exception(f"street not found: {self._street}") Exception: street not found: Pflugkstraße
My configuration.yaml:
waste_collection_schedule:
sources:
- name: stadtreinigung_leipzig_de
args:
street: Pflugkstraße
house_number: 1
calendar_title: Abfallkalender
I've been trying around here for a few days, but I can't find a solution to the problem. Is it possible that the API has been changed/defective?
Thanks for your help...
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `custom_components/waste_collection_schedule/waste_collection_schedule/source/stadtreinigung_leipzig_de.py`
Content:
```
1 import json
2 import logging
3
4 import requests
5 from waste_collection_schedule import Collection # type: ignore[attr-defined]
6 from waste_collection_schedule.service.ICS import ICS
7
8 _LOGGER = logging.getLogger(__name__)
9
10 TITLE = "Stadtreinigung Leipzig"
11 DESCRIPTION = "Source for Stadtreinigung Leipzig."
12 URL = "https://stadtreinigung-leipzig.de"
13 TEST_CASES = {"Bahnhofsallee": {"street": "Bahnhofsallee", "house_number": 7}}
14
15
16 class Source:
17 def __init__(self, street, house_number):
18 self._street = street
19 self._house_number = house_number
20 self._ics = ICS()
21
22 def fetch(self):
23 params = {
24 "name": self._street,
25 }
26
27 # get list of streets and house numbers
28 r = requests.get(
29 "https://stadtreinigung-leipzig.de/rest/wastecalendarstreets", params=params
30 )
31
32 data = json.loads(r.text)
33 if len(data["results"]) == 0:
34 raise Exception(f"street not found: {self._street}")
35 street_entry = data["results"].get(self._street)
36 if street_entry is None:
37 raise Exception(f"street not found: {self._street}")
38
39 id = street_entry.get(str(self._house_number))
40 if id is None:
41 raise Exception(f"house_number not found: {self._house_number}")
42
43 # get ics file
44 params = {
45 "position_nos": id,
46 }
47 r = requests.get(
48 "https://stadtreinigung-leipzig.de/wir-kommen-zu-ihnen/abfallkalender/ical.ics",
49 params=params,
50 )
51 dates = self._ics.convert(r.text)
52
53 entries = []
54 for d in dates:
55 entries.append(Collection(d[0], d[1].removesuffix(", ")))
56 return entries
57
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/custom_components/waste_collection_schedule/waste_collection_schedule/source/stadtreinigung_leipzig_de.py b/custom_components/waste_collection_schedule/waste_collection_schedule/source/stadtreinigung_leipzig_de.py
--- a/custom_components/waste_collection_schedule/waste_collection_schedule/source/stadtreinigung_leipzig_de.py
+++ b/custom_components/waste_collection_schedule/waste_collection_schedule/source/stadtreinigung_leipzig_de.py
@@ -21,12 +21,13 @@
def fetch(self):
params = {
- "name": self._street,
+ "old_format": 1,
+ "search": self._street,
}
# get list of streets and house numbers
r = requests.get(
- "https://stadtreinigung-leipzig.de/rest/wastecalendarstreets", params=params
+ "https://stadtreinigung-leipzig.de/rest/Navision/Streets", params=params
)
data = json.loads(r.text)
| {"golden_diff": "diff --git a/custom_components/waste_collection_schedule/waste_collection_schedule/source/stadtreinigung_leipzig_de.py b/custom_components/waste_collection_schedule/waste_collection_schedule/source/stadtreinigung_leipzig_de.py\n--- a/custom_components/waste_collection_schedule/waste_collection_schedule/source/stadtreinigung_leipzig_de.py\n+++ b/custom_components/waste_collection_schedule/waste_collection_schedule/source/stadtreinigung_leipzig_de.py\n@@ -21,12 +21,13 @@\n \n def fetch(self):\n params = {\n- \"name\": self._street,\n+ \"old_format\": 1,\n+ \"search\": self._street,\n }\n \n # get list of streets and house numbers\n r = requests.get(\n- \"https://stadtreinigung-leipzig.de/rest/wastecalendarstreets\", params=params\n+ \"https://stadtreinigung-leipzig.de/rest/Navision/Streets\", params=params\n )\n \n data = json.loads(r.text)\n", "issue": "Error with Service Provider Stadtreinigung Leipzig / stadtreinigung-leipzig.de\nHi everyone,\r\nSince 03.01.2023 (this is where I noticed it), WCS can no longer retrieve data from Stadtwerke Leipzig.\r\nThe following error is displayed:\r\n\r\nfetch failed for source Stadtreinigung Leipzig: Traceback (most recent call last): File \"/config/custom_components/waste_collection_schedule/waste_collection_schedule/source_shell.py\", line 134, in fetch entries = self._source.fetch() File \"/config/custom_components/waste_collection_schedule/waste_collection_schedule/source/stadtreinigung_leipzig_de.py\", line 34, in fetch raise Exception(f\"street not found: {self._street}\") Exception: street not found: Pflugkstra\u00dfe\r\n\r\nMy configuration.yaml:\r\nwaste_collection_schedule:\r\n sources:\r\n - name: stadtreinigung_leipzig_de\r\n args:\r\n street: Pflugkstra\u00dfe\r\n house_number: 1\r\n calendar_title: Abfallkalender\r\n\r\nI've been trying around here for a few days, but I can't find a solution to the problem. Is it possible that the API has been changed/defective?\r\nThanks for your help...\n", "before_files": [{"content": "import json\nimport logging\n\nimport requests\nfrom waste_collection_schedule import Collection # type: ignore[attr-defined]\nfrom waste_collection_schedule.service.ICS import ICS\n\n_LOGGER = logging.getLogger(__name__)\n\nTITLE = \"Stadtreinigung Leipzig\"\nDESCRIPTION = \"Source for Stadtreinigung Leipzig.\"\nURL = \"https://stadtreinigung-leipzig.de\"\nTEST_CASES = {\"Bahnhofsallee\": {\"street\": \"Bahnhofsallee\", \"house_number\": 7}}\n\n\nclass Source:\n def __init__(self, street, house_number):\n self._street = street\n self._house_number = house_number\n self._ics = ICS()\n\n def fetch(self):\n params = {\n \"name\": self._street,\n }\n\n # get list of streets and house numbers\n r = requests.get(\n \"https://stadtreinigung-leipzig.de/rest/wastecalendarstreets\", params=params\n )\n\n data = json.loads(r.text)\n if len(data[\"results\"]) == 0:\n raise Exception(f\"street not found: {self._street}\")\n street_entry = data[\"results\"].get(self._street)\n if street_entry is None:\n raise Exception(f\"street not found: {self._street}\")\n\n id = street_entry.get(str(self._house_number))\n if id is None:\n raise Exception(f\"house_number not found: {self._house_number}\")\n\n # get ics file\n params = {\n \"position_nos\": id,\n }\n r = requests.get(\n \"https://stadtreinigung-leipzig.de/wir-kommen-zu-ihnen/abfallkalender/ical.ics\",\n params=params,\n )\n dates = self._ics.convert(r.text)\n\n entries = []\n for d in dates:\n entries.append(Collection(d[0], d[1].removesuffix(\", \")))\n return entries\n", "path": "custom_components/waste_collection_schedule/waste_collection_schedule/source/stadtreinigung_leipzig_de.py"}], "after_files": [{"content": "import json\nimport logging\n\nimport requests\nfrom waste_collection_schedule import Collection # type: ignore[attr-defined]\nfrom waste_collection_schedule.service.ICS import ICS\n\n_LOGGER = logging.getLogger(__name__)\n\nTITLE = \"Stadtreinigung Leipzig\"\nDESCRIPTION = \"Source for Stadtreinigung Leipzig.\"\nURL = \"https://stadtreinigung-leipzig.de\"\nTEST_CASES = {\"Bahnhofsallee\": {\"street\": \"Bahnhofsallee\", \"house_number\": 7}}\n\n\nclass Source:\n def __init__(self, street, house_number):\n self._street = street\n self._house_number = house_number\n self._ics = ICS()\n\n def fetch(self):\n params = {\n \"old_format\": 1,\n \"search\": self._street,\n }\n\n # get list of streets and house numbers\n r = requests.get(\n \"https://stadtreinigung-leipzig.de/rest/Navision/Streets\", params=params\n )\n\n data = json.loads(r.text)\n if len(data[\"results\"]) == 0:\n raise Exception(f\"street not found: {self._street}\")\n street_entry = data[\"results\"].get(self._street)\n if street_entry is None:\n raise Exception(f\"street not found: {self._street}\")\n\n id = street_entry.get(str(self._house_number))\n if id is None:\n raise Exception(f\"house_number not found: {self._house_number}\")\n\n # get ics file\n params = {\n \"position_nos\": id,\n }\n r = requests.get(\n \"https://stadtreinigung-leipzig.de/wir-kommen-zu-ihnen/abfallkalender/ical.ics\",\n params=params,\n )\n dates = self._ics.convert(r.text)\n\n entries = []\n for d in dates:\n entries.append(Collection(d[0], d[1].removesuffix(\", \")))\n return entries\n", "path": "custom_components/waste_collection_schedule/waste_collection_schedule/source/stadtreinigung_leipzig_de.py"}]} | 1,056 | 219 |
gh_patches_debug_37670 | rasdani/github-patches | git_diff | biolab__orange3-3842 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Replicability in Neural networks and Random forests
Follow up from #3715: Neural networks and Random forests should have a checkbox `Replicable training` or something like this, which would decide whether random seed is fixed (to 0) or "random".
In Neural networks: add the check box.
In Random forest: remove the spin box.
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `Orange/widgets/model/owrandomforest.py`
Content:
```
1 from AnyQt.QtCore import Qt
2
3 from Orange.data import Table
4 from Orange.modelling import RandomForestLearner
5 from Orange.widgets import settings, gui
6 from Orange.widgets.utils.owlearnerwidget import OWBaseLearner
7 from Orange.widgets.utils.widgetpreview import WidgetPreview
8 from Orange.widgets.widget import Msg
9
10
11 class OWRandomForest(OWBaseLearner):
12 name = "Random Forest"
13 description = "Predict using an ensemble of decision trees."
14 icon = "icons/RandomForest.svg"
15 replaces = [
16 "Orange.widgets.classify.owrandomforest.OWRandomForest",
17 "Orange.widgets.regression.owrandomforestregression.OWRandomForestRegression",
18 ]
19 priority = 40
20 keywords = []
21
22 LEARNER = RandomForestLearner
23
24 n_estimators = settings.Setting(10)
25 max_features = settings.Setting(5)
26 use_max_features = settings.Setting(False)
27 random_state = settings.Setting(0)
28 use_random_state = settings.Setting(False)
29 max_depth = settings.Setting(3)
30 use_max_depth = settings.Setting(False)
31 min_samples_split = settings.Setting(5)
32 use_min_samples_split = settings.Setting(True)
33 index_output = settings.Setting(0)
34
35 class Error(OWBaseLearner.Error):
36 not_enough_features = Msg("Insufficient number of attributes ({})")
37
38 def add_main_layout(self):
39 box = gui.vBox(self.controlArea, 'Basic Properties')
40 self.n_estimators_spin = gui.spin(
41 box, self, "n_estimators", minv=1, maxv=10000, controlWidth=80,
42 alignment=Qt.AlignRight, label="Number of trees: ",
43 callback=self.settings_changed)
44 self.max_features_spin = gui.spin(
45 box, self, "max_features", 2, 50, controlWidth=80,
46 label="Number of attributes considered at each split: ",
47 callback=self.settings_changed, checked="use_max_features",
48 checkCallback=self.settings_changed, alignment=Qt.AlignRight,)
49 self.random_state_spin = gui.spin(
50 box, self, "random_state", 0, 2 ** 31 - 1, controlWidth=80,
51 label="Fixed seed for random generator: ", alignment=Qt.AlignRight,
52 callback=self.settings_changed, checked="use_random_state",
53 checkCallback=self.settings_changed)
54
55 box = gui.vBox(self.controlArea, "Growth Control")
56 self.max_depth_spin = gui.spin(
57 box, self, "max_depth", 1, 50, controlWidth=80,
58 label="Limit depth of individual trees: ", alignment=Qt.AlignRight,
59 callback=self.settings_changed, checked="use_max_depth",
60 checkCallback=self.settings_changed)
61 self.min_samples_split_spin = gui.spin(
62 box, self, "min_samples_split", 2, 1000, controlWidth=80,
63 label="Do not split subsets smaller than: ",
64 callback=self.settings_changed, checked="use_min_samples_split",
65 checkCallback=self.settings_changed, alignment=Qt.AlignRight)
66
67 def create_learner(self):
68 common_args = {"n_estimators": self.n_estimators}
69 if self.use_max_features:
70 common_args["max_features"] = self.max_features
71 if self.use_random_state:
72 common_args["random_state"] = self.random_state
73 if self.use_max_depth:
74 common_args["max_depth"] = self.max_depth
75 if self.use_min_samples_split:
76 common_args["min_samples_split"] = self.min_samples_split
77
78 return self.LEARNER(preprocessors=self.preprocessors, **common_args)
79
80 def check_data(self):
81 self.Error.not_enough_features.clear()
82 if super().check_data():
83 n_features = len(self.data.domain.attributes)
84 if self.use_max_features and self.max_features > n_features:
85 self.Error.not_enough_features(n_features)
86 self.valid_data = False
87 return self.valid_data
88
89 def get_learner_parameters(self):
90 """Called by send report to list the parameters of the learner."""
91 return (
92 ("Number of trees", self.n_estimators),
93 ("Maximal number of considered features",
94 self.max_features if self.use_max_features else "unlimited"),
95 ("Fixed random seed", self.use_random_state and self.random_state),
96 ("Maximal tree depth",
97 self.max_depth if self.use_max_depth else "unlimited"),
98 ("Stop splitting nodes with maximum instances",
99 self.min_samples_split if self.use_min_samples_split else "unlimited")
100 )
101
102
103 if __name__ == "__main__": # pragma: no cover
104 WidgetPreview(OWRandomForest).run(Table("iris"))
105
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/Orange/widgets/model/owrandomforest.py b/Orange/widgets/model/owrandomforest.py
--- a/Orange/widgets/model/owrandomforest.py
+++ b/Orange/widgets/model/owrandomforest.py
@@ -24,7 +24,6 @@
n_estimators = settings.Setting(10)
max_features = settings.Setting(5)
use_max_features = settings.Setting(False)
- random_state = settings.Setting(0)
use_random_state = settings.Setting(False)
max_depth = settings.Setting(3)
use_max_depth = settings.Setting(False)
@@ -46,11 +45,9 @@
label="Number of attributes considered at each split: ",
callback=self.settings_changed, checked="use_max_features",
checkCallback=self.settings_changed, alignment=Qt.AlignRight,)
- self.random_state_spin = gui.spin(
- box, self, "random_state", 0, 2 ** 31 - 1, controlWidth=80,
- label="Fixed seed for random generator: ", alignment=Qt.AlignRight,
- callback=self.settings_changed, checked="use_random_state",
- checkCallback=self.settings_changed)
+ self.random_state = gui.checkBox(
+ box, self, "use_random_state", label="Replicable training",
+ callback=self.settings_changed)
box = gui.vBox(self.controlArea, "Growth Control")
self.max_depth_spin = gui.spin(
@@ -69,7 +66,7 @@
if self.use_max_features:
common_args["max_features"] = self.max_features
if self.use_random_state:
- common_args["random_state"] = self.random_state
+ common_args["random_state"] = 0
if self.use_max_depth:
common_args["max_depth"] = self.max_depth
if self.use_min_samples_split:
@@ -92,7 +89,7 @@
("Number of trees", self.n_estimators),
("Maximal number of considered features",
self.max_features if self.use_max_features else "unlimited"),
- ("Fixed random seed", self.use_random_state and self.random_state),
+ ("Replicable training", ["No", "Yes"][self.use_random_state]),
("Maximal tree depth",
self.max_depth if self.use_max_depth else "unlimited"),
("Stop splitting nodes with maximum instances",
| {"golden_diff": "diff --git a/Orange/widgets/model/owrandomforest.py b/Orange/widgets/model/owrandomforest.py\n--- a/Orange/widgets/model/owrandomforest.py\n+++ b/Orange/widgets/model/owrandomforest.py\n@@ -24,7 +24,6 @@\n n_estimators = settings.Setting(10)\n max_features = settings.Setting(5)\n use_max_features = settings.Setting(False)\n- random_state = settings.Setting(0)\n use_random_state = settings.Setting(False)\n max_depth = settings.Setting(3)\n use_max_depth = settings.Setting(False)\n@@ -46,11 +45,9 @@\n label=\"Number of attributes considered at each split: \",\n callback=self.settings_changed, checked=\"use_max_features\",\n checkCallback=self.settings_changed, alignment=Qt.AlignRight,)\n- self.random_state_spin = gui.spin(\n- box, self, \"random_state\", 0, 2 ** 31 - 1, controlWidth=80,\n- label=\"Fixed seed for random generator: \", alignment=Qt.AlignRight,\n- callback=self.settings_changed, checked=\"use_random_state\",\n- checkCallback=self.settings_changed)\n+ self.random_state = gui.checkBox(\n+ box, self, \"use_random_state\", label=\"Replicable training\",\n+ callback=self.settings_changed)\n \n box = gui.vBox(self.controlArea, \"Growth Control\")\n self.max_depth_spin = gui.spin(\n@@ -69,7 +66,7 @@\n if self.use_max_features:\n common_args[\"max_features\"] = self.max_features\n if self.use_random_state:\n- common_args[\"random_state\"] = self.random_state\n+ common_args[\"random_state\"] = 0\n if self.use_max_depth:\n common_args[\"max_depth\"] = self.max_depth\n if self.use_min_samples_split:\n@@ -92,7 +89,7 @@\n (\"Number of trees\", self.n_estimators),\n (\"Maximal number of considered features\",\n self.max_features if self.use_max_features else \"unlimited\"),\n- (\"Fixed random seed\", self.use_random_state and self.random_state),\n+ (\"Replicable training\", [\"No\", \"Yes\"][self.use_random_state]),\n (\"Maximal tree depth\",\n self.max_depth if self.use_max_depth else \"unlimited\"),\n (\"Stop splitting nodes with maximum instances\",\n", "issue": "Replicability in Neural networks and Random forests\nFollow up from #3715: Neural networks and Random forests should have a checkbox `Replicable training` or something like this, which would decide whether random seed is fixed (to 0) or \"random\".\r\n\r\nIn Neural networks: add the check box.\r\n\r\nIn Random forest: remove the spin box.\n", "before_files": [{"content": "from AnyQt.QtCore import Qt\n\nfrom Orange.data import Table\nfrom Orange.modelling import RandomForestLearner\nfrom Orange.widgets import settings, gui\nfrom Orange.widgets.utils.owlearnerwidget import OWBaseLearner\nfrom Orange.widgets.utils.widgetpreview import WidgetPreview\nfrom Orange.widgets.widget import Msg\n\n\nclass OWRandomForest(OWBaseLearner):\n name = \"Random Forest\"\n description = \"Predict using an ensemble of decision trees.\"\n icon = \"icons/RandomForest.svg\"\n replaces = [\n \"Orange.widgets.classify.owrandomforest.OWRandomForest\",\n \"Orange.widgets.regression.owrandomforestregression.OWRandomForestRegression\",\n ]\n priority = 40\n keywords = []\n\n LEARNER = RandomForestLearner\n\n n_estimators = settings.Setting(10)\n max_features = settings.Setting(5)\n use_max_features = settings.Setting(False)\n random_state = settings.Setting(0)\n use_random_state = settings.Setting(False)\n max_depth = settings.Setting(3)\n use_max_depth = settings.Setting(False)\n min_samples_split = settings.Setting(5)\n use_min_samples_split = settings.Setting(True)\n index_output = settings.Setting(0)\n\n class Error(OWBaseLearner.Error):\n not_enough_features = Msg(\"Insufficient number of attributes ({})\")\n\n def add_main_layout(self):\n box = gui.vBox(self.controlArea, 'Basic Properties')\n self.n_estimators_spin = gui.spin(\n box, self, \"n_estimators\", minv=1, maxv=10000, controlWidth=80,\n alignment=Qt.AlignRight, label=\"Number of trees: \",\n callback=self.settings_changed)\n self.max_features_spin = gui.spin(\n box, self, \"max_features\", 2, 50, controlWidth=80,\n label=\"Number of attributes considered at each split: \",\n callback=self.settings_changed, checked=\"use_max_features\",\n checkCallback=self.settings_changed, alignment=Qt.AlignRight,)\n self.random_state_spin = gui.spin(\n box, self, \"random_state\", 0, 2 ** 31 - 1, controlWidth=80,\n label=\"Fixed seed for random generator: \", alignment=Qt.AlignRight,\n callback=self.settings_changed, checked=\"use_random_state\",\n checkCallback=self.settings_changed)\n\n box = gui.vBox(self.controlArea, \"Growth Control\")\n self.max_depth_spin = gui.spin(\n box, self, \"max_depth\", 1, 50, controlWidth=80,\n label=\"Limit depth of individual trees: \", alignment=Qt.AlignRight,\n callback=self.settings_changed, checked=\"use_max_depth\",\n checkCallback=self.settings_changed)\n self.min_samples_split_spin = gui.spin(\n box, self, \"min_samples_split\", 2, 1000, controlWidth=80,\n label=\"Do not split subsets smaller than: \",\n callback=self.settings_changed, checked=\"use_min_samples_split\",\n checkCallback=self.settings_changed, alignment=Qt.AlignRight)\n\n def create_learner(self):\n common_args = {\"n_estimators\": self.n_estimators}\n if self.use_max_features:\n common_args[\"max_features\"] = self.max_features\n if self.use_random_state:\n common_args[\"random_state\"] = self.random_state\n if self.use_max_depth:\n common_args[\"max_depth\"] = self.max_depth\n if self.use_min_samples_split:\n common_args[\"min_samples_split\"] = self.min_samples_split\n\n return self.LEARNER(preprocessors=self.preprocessors, **common_args)\n\n def check_data(self):\n self.Error.not_enough_features.clear()\n if super().check_data():\n n_features = len(self.data.domain.attributes)\n if self.use_max_features and self.max_features > n_features:\n self.Error.not_enough_features(n_features)\n self.valid_data = False\n return self.valid_data\n\n def get_learner_parameters(self):\n \"\"\"Called by send report to list the parameters of the learner.\"\"\"\n return (\n (\"Number of trees\", self.n_estimators),\n (\"Maximal number of considered features\",\n self.max_features if self.use_max_features else \"unlimited\"),\n (\"Fixed random seed\", self.use_random_state and self.random_state),\n (\"Maximal tree depth\",\n self.max_depth if self.use_max_depth else \"unlimited\"),\n (\"Stop splitting nodes with maximum instances\",\n self.min_samples_split if self.use_min_samples_split else \"unlimited\")\n )\n\n\nif __name__ == \"__main__\": # pragma: no cover\n WidgetPreview(OWRandomForest).run(Table(\"iris\"))\n", "path": "Orange/widgets/model/owrandomforest.py"}], "after_files": [{"content": "from AnyQt.QtCore import Qt\n\nfrom Orange.data import Table\nfrom Orange.modelling import RandomForestLearner\nfrom Orange.widgets import settings, gui\nfrom Orange.widgets.utils.owlearnerwidget import OWBaseLearner\nfrom Orange.widgets.utils.widgetpreview import WidgetPreview\nfrom Orange.widgets.widget import Msg\n\n\nclass OWRandomForest(OWBaseLearner):\n name = \"Random Forest\"\n description = \"Predict using an ensemble of decision trees.\"\n icon = \"icons/RandomForest.svg\"\n replaces = [\n \"Orange.widgets.classify.owrandomforest.OWRandomForest\",\n \"Orange.widgets.regression.owrandomforestregression.OWRandomForestRegression\",\n ]\n priority = 40\n keywords = []\n\n LEARNER = RandomForestLearner\n\n n_estimators = settings.Setting(10)\n max_features = settings.Setting(5)\n use_max_features = settings.Setting(False)\n use_random_state = settings.Setting(False)\n max_depth = settings.Setting(3)\n use_max_depth = settings.Setting(False)\n min_samples_split = settings.Setting(5)\n use_min_samples_split = settings.Setting(True)\n index_output = settings.Setting(0)\n\n class Error(OWBaseLearner.Error):\n not_enough_features = Msg(\"Insufficient number of attributes ({})\")\n\n def add_main_layout(self):\n box = gui.vBox(self.controlArea, 'Basic Properties')\n self.n_estimators_spin = gui.spin(\n box, self, \"n_estimators\", minv=1, maxv=10000, controlWidth=80,\n alignment=Qt.AlignRight, label=\"Number of trees: \",\n callback=self.settings_changed)\n self.max_features_spin = gui.spin(\n box, self, \"max_features\", 2, 50, controlWidth=80,\n label=\"Number of attributes considered at each split: \",\n callback=self.settings_changed, checked=\"use_max_features\",\n checkCallback=self.settings_changed, alignment=Qt.AlignRight,)\n self.random_state = gui.checkBox(\n box, self, \"use_random_state\", label=\"Replicable training\",\n callback=self.settings_changed)\n\n box = gui.vBox(self.controlArea, \"Growth Control\")\n self.max_depth_spin = gui.spin(\n box, self, \"max_depth\", 1, 50, controlWidth=80,\n label=\"Limit depth of individual trees: \", alignment=Qt.AlignRight,\n callback=self.settings_changed, checked=\"use_max_depth\",\n checkCallback=self.settings_changed)\n self.min_samples_split_spin = gui.spin(\n box, self, \"min_samples_split\", 2, 1000, controlWidth=80,\n label=\"Do not split subsets smaller than: \",\n callback=self.settings_changed, checked=\"use_min_samples_split\",\n checkCallback=self.settings_changed, alignment=Qt.AlignRight)\n\n def create_learner(self):\n common_args = {\"n_estimators\": self.n_estimators}\n if self.use_max_features:\n common_args[\"max_features\"] = self.max_features\n if self.use_random_state:\n common_args[\"random_state\"] = 0\n if self.use_max_depth:\n common_args[\"max_depth\"] = self.max_depth\n if self.use_min_samples_split:\n common_args[\"min_samples_split\"] = self.min_samples_split\n\n return self.LEARNER(preprocessors=self.preprocessors, **common_args)\n\n def check_data(self):\n self.Error.not_enough_features.clear()\n if super().check_data():\n n_features = len(self.data.domain.attributes)\n if self.use_max_features and self.max_features > n_features:\n self.Error.not_enough_features(n_features)\n self.valid_data = False\n return self.valid_data\n\n def get_learner_parameters(self):\n \"\"\"Called by send report to list the parameters of the learner.\"\"\"\n return (\n (\"Number of trees\", self.n_estimators),\n (\"Maximal number of considered features\",\n self.max_features if self.use_max_features else \"unlimited\"),\n (\"Replicable training\", [\"No\", \"Yes\"][self.use_random_state]),\n (\"Maximal tree depth\",\n self.max_depth if self.use_max_depth else \"unlimited\"),\n (\"Stop splitting nodes with maximum instances\",\n self.min_samples_split if self.use_min_samples_split else \"unlimited\")\n )\n\n\nif __name__ == \"__main__\": # pragma: no cover\n WidgetPreview(OWRandomForest).run(Table(\"iris\"))\n", "path": "Orange/widgets/model/owrandomforest.py"}]} | 1,556 | 516 |
gh_patches_debug_31272 | rasdani/github-patches | git_diff | Lightning-AI__pytorch-lightning-453 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
min_max log_gpu_memory option bug
**Describe the bug**
Setting `log_gpu_memory='min_max'` in `Trainer` leads to the following bug.
```
Traceback (most recent call last):
File "/opt/conda/lib/python3.7/site-packages/pytorch_lightning/trainer/trainer.py", line 347, in fit
self.single_gpu_train(model)
File "/opt/conda/lib/python3.7/site-packages/pytorch_lightning/trainer/dp_mixin.py", line 79, in single_gpu_train
self.run_pretrain_routine(model)
File "/opt/conda/lib/python3.7/site-packages/pytorch_lightning/trainer/trainer.py", line 467, in run_pretrain_routine
self.train()
File "/opt/conda/lib/python3.7/site-packages/pytorch_lightning/trainer/train_loop_mixin.py", line 60, in train
self.run_training_epoch()
File "/opt/conda/lib/python3.7/site-packages/pytorch_lightning/trainer/train_loop_mixin.py", line 126, in run_training_epoch
self.log_metrics(batch_step_metrics, grad_norm_dic)
File "/opt/conda/lib/python3.7/site-packages/pytorch_lightning/trainer/logging_mixin.py", line 20, in log_metrics
mem_map = memory.get_memory_profile(self.log_gpu_memory)
File "/opt/conda/lib/python3.7/site-packages/pytorch_lightning/root_module/memory.py", line 205, in get_memory_profile
for k, v in memory_map:
ValueError: too many values to unpack (expected 2)
```
**To Reproduce**
On current master, execute the following.
```
trainer = Trainer(
...
log_gpu_memory='min_max',
...
)
trainer.fit(model)
```
**Expected behavior**
Log the min/max utilization of gpu memory, as `min_max` option is documented.
**Desktop (please complete the following information):**
- OS: Ubuntu 18.04
- Version: Current master
I am working on this issue. Will submit a PR soon.
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `pytorch_lightning/root_module/memory.py`
Content:
```
1 '''
2 Generates a summary of a model's layers and dimensionality
3 '''
4
5 import gc
6 import subprocess
7
8 import numpy as np
9 import pandas as pd
10 import torch
11
12
13 class ModelSummary(object):
14
15 def __init__(self, model, mode='full'):
16 '''
17 Generates summaries of model layers and dimensions.
18 '''
19 self.model = model
20 self.mode = mode
21 self.in_sizes = []
22 self.out_sizes = []
23
24 self.summarize()
25
26 def __str__(self):
27 return self.summary.__str__()
28
29 def __repr__(self):
30 return self.summary.__str__()
31
32 def named_modules(self):
33 if self.mode == 'full':
34 mods = self.model.named_modules()
35 mods = list(mods)[1:] # do not include root module (LightningModule)
36 elif self.mode == 'top':
37 # the children are the top-level modules
38 mods = self.model.named_children()
39 else:
40 mods = []
41 return list(mods)
42
43 def get_variable_sizes(self):
44 '''Run sample input through each layer to get output sizes'''
45 mods = self.named_modules()
46 in_sizes = []
47 out_sizes = []
48 input_ = self.model.example_input_array
49
50 if self.model.on_gpu:
51 input_ = input_.cuda(0)
52
53 if self.model.trainer.use_amp:
54 input_ = input_.half()
55
56 with torch.no_grad():
57
58 for _, m in mods:
59 if type(input_) is list or type(input_) is tuple: # pragma: no cover
60 out = m(*input_)
61 else:
62 out = m(input_)
63
64 if type(input_) is tuple or type(input_) is list: # pragma: no cover
65 in_size = []
66 for x in input_:
67 if type(x) is list:
68 in_size.append(len(x))
69 else:
70 in_size.append(x.size())
71 else:
72 in_size = np.array(input_.size())
73
74 in_sizes.append(in_size)
75
76 if type(out) is tuple or type(out) is list: # pragma: no cover
77 out_size = np.asarray([x.size() for x in out])
78 else:
79 out_size = np.array(out.size())
80
81 out_sizes.append(out_size)
82 input_ = out
83
84 self.in_sizes = in_sizes
85 self.out_sizes = out_sizes
86 assert len(in_sizes) == len(out_sizes)
87 return
88
89 def get_layer_names(self):
90 '''Collect Layer Names'''
91 mods = self.named_modules()
92 names = []
93 layers = []
94 for name, m in mods:
95 names += [name]
96 layers += [str(m.__class__)]
97
98 layer_types = [x.split('.')[-1][:-2] for x in layers]
99
100 self.layer_names = names
101 self.layer_types = layer_types
102 return
103
104 def get_parameter_sizes(self):
105 '''Get sizes of all parameters in `model`'''
106 mods = self.named_modules()
107 sizes = []
108 for _, m in mods:
109 p = list(m.parameters())
110 modsz = []
111 for j in range(len(p)):
112 modsz.append(np.array(p[j].size()))
113 sizes.append(modsz)
114
115 self.param_sizes = sizes
116 return
117
118 def get_parameter_nums(self):
119 '''Get number of parameters in each layer'''
120 param_nums = []
121 for mod in self.param_sizes:
122 all_params = 0
123 for p in mod:
124 all_params += np.prod(p)
125 param_nums.append(all_params)
126 self.param_nums = param_nums
127 return
128
129 def make_summary(self):
130 '''
131 Makes a summary listing with:
132
133 Layer Name, Layer Type, Input Size, Output Size, Number of Parameters
134 '''
135
136 cols = ['Name', 'Type', 'Params']
137 if self.model.example_input_array is not None:
138 cols.extend(['In_sizes', 'Out_sizes'])
139
140 df = pd.DataFrame(np.zeros((len(self.layer_names), len(cols))))
141 df.columns = cols
142
143 df['Name'] = self.layer_names
144 df['Type'] = self.layer_types
145 df['Params'] = self.param_nums
146 df['Params'] = df['Params'].map(get_human_readable_count)
147
148 if self.model.example_input_array is not None:
149 df['In_sizes'] = self.in_sizes
150 df['Out_sizes'] = self.out_sizes
151
152 self.summary = df
153 return
154
155 def summarize(self):
156 self.get_layer_names()
157 self.get_parameter_sizes()
158 self.get_parameter_nums()
159
160 if self.model.example_input_array is not None:
161 self.get_variable_sizes()
162 self.make_summary()
163
164
165 def print_mem_stack(): # pragma: no cover
166 for obj in gc.get_objects():
167 try:
168 if torch.is_tensor(obj) or (hasattr(obj, 'data') and torch.is_tensor(obj.data)):
169 print(type(obj), obj.size())
170 except Exception:
171 pass
172
173
174 def count_mem_items(): # pragma: no cover
175 nb_params = 0
176 nb_tensors = 0
177 for obj in gc.get_objects():
178 try:
179 if torch.is_tensor(obj) or (hasattr(obj, 'data') and torch.is_tensor(obj.data)):
180 obj_type = str(type(obj))
181 if 'parameter' in obj_type:
182 nb_params += 1
183 else:
184 nb_tensors += 1
185 except Exception:
186 pass
187
188 return nb_params, nb_tensors
189
190
191 def get_memory_profile(mode):
192 """
193 'all' means return memory for all gpus
194 'min_max' means return memory for max and min
195 :param mode:
196 :return:
197 """
198 memory_map = get_gpu_memory_map()
199
200 if mode == 'min_max':
201 min_mem = 1000000
202 min_k = None
203 max_mem = 0
204 max_k = None
205 for k, v in memory_map:
206 if v > max_mem:
207 max_mem = v
208 max_k = k
209 if v < min_mem:
210 min_mem = v
211 min_k = k
212
213 memory_map = {min_k: min_mem, max_k: max_mem}
214
215 return memory_map
216
217
218 def get_gpu_memory_map():
219 """Get the current gpu usage.
220
221 Returns
222 -------
223 usage: dict
224 Keys are device ids as integers.
225 Values are memory usage as integers in MB.
226 """
227 result = subprocess.check_output(
228 [
229 'nvidia-smi', '--query-gpu=memory.used',
230 '--format=csv,nounits,noheader'
231 ], encoding='utf-8')
232 # Convert lines into a dictionary
233 gpu_memory = [int(x) for x in result.strip().split('\n')]
234 gpu_memory_map = {}
235 for k, v in zip(range(len(gpu_memory)), gpu_memory):
236 k = f'gpu_{k}'
237 gpu_memory_map[k] = v
238 return gpu_memory_map
239
240
241 def get_human_readable_count(number):
242 """
243 Abbreviates an integer number with K, M, B, T for thousands, millions,
244 billions and trillions, respectively.
245 Examples:
246 123 -> 123
247 1234 -> 1 K (one thousand)
248 2e6 -> 2 M (two million)
249 3e9 -> 3 B (three billion)
250 4e12 -> 4 T (four trillion)
251 5e15 -> 5,000 T
252 :param number: a positive integer number
253 :returns a string formatted according to the pattern described above.
254 """
255 assert number >= 0
256 labels = [' ', 'K', 'M', 'B', 'T']
257 num_digits = int(np.floor(np.log10(number)) + 1 if number > 0 else 1)
258 num_groups = int(np.ceil(num_digits / 3))
259 num_groups = min(num_groups, len(labels)) # don't abbreviate beyond trillions
260 shift = -3 * (num_groups - 1)
261 number = number * (10 ** shift)
262 index = num_groups - 1
263 return f'{int(number):,d} {labels[index]}'
264
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/pytorch_lightning/root_module/memory.py b/pytorch_lightning/root_module/memory.py
--- a/pytorch_lightning/root_module/memory.py
+++ b/pytorch_lightning/root_module/memory.py
@@ -3,6 +3,7 @@
'''
import gc
+import os
import subprocess
import numpy as np
@@ -198,19 +199,10 @@
memory_map = get_gpu_memory_map()
if mode == 'min_max':
- min_mem = 1000000
- min_k = None
- max_mem = 0
- max_k = None
- for k, v in memory_map:
- if v > max_mem:
- max_mem = v
- max_k = k
- if v < min_mem:
- min_mem = v
- min_k = k
-
- memory_map = {min_k: min_mem, max_k: max_mem}
+ min_index, min_memory = min(memory_map.items(), key=lambda item: item[1])
+ max_index, max_memory = max(memory_map.items(), key=lambda item: item[1])
+
+ memory_map = {min_index: min_memory, max_index: max_memory}
return memory_map
@@ -224,17 +216,18 @@
Keys are device ids as integers.
Values are memory usage as integers in MB.
"""
- result = subprocess.check_output(
+ result = subprocess.run(
[
- 'nvidia-smi', '--query-gpu=memory.used',
- '--format=csv,nounits,noheader'
- ], encoding='utf-8')
+ 'nvidia-smi',
+ '--query-gpu=memory.used',
+ '--format=csv,nounits,noheader',
+ ],
+ encoding='utf-8',
+ capture_output=True,
+ check=True)
# Convert lines into a dictionary
- gpu_memory = [int(x) for x in result.strip().split('\n')]
- gpu_memory_map = {}
- for k, v in zip(range(len(gpu_memory)), gpu_memory):
- k = f'gpu_{k}'
- gpu_memory_map[k] = v
+ gpu_memory = [int(x) for x in result.stdout.strip().split(os.linesep)]
+ gpu_memory_map = {f'gpu_{index}': memory for index, memory in enumerate(gpu_memory)}
return gpu_memory_map
| {"golden_diff": "diff --git a/pytorch_lightning/root_module/memory.py b/pytorch_lightning/root_module/memory.py\n--- a/pytorch_lightning/root_module/memory.py\n+++ b/pytorch_lightning/root_module/memory.py\n@@ -3,6 +3,7 @@\n '''\n \n import gc\n+import os\n import subprocess\n \n import numpy as np\n@@ -198,19 +199,10 @@\n memory_map = get_gpu_memory_map()\n \n if mode == 'min_max':\n- min_mem = 1000000\n- min_k = None\n- max_mem = 0\n- max_k = None\n- for k, v in memory_map:\n- if v > max_mem:\n- max_mem = v\n- max_k = k\n- if v < min_mem:\n- min_mem = v\n- min_k = k\n-\n- memory_map = {min_k: min_mem, max_k: max_mem}\n+ min_index, min_memory = min(memory_map.items(), key=lambda item: item[1])\n+ max_index, max_memory = max(memory_map.items(), key=lambda item: item[1])\n+\n+ memory_map = {min_index: min_memory, max_index: max_memory}\n \n return memory_map\n \n@@ -224,17 +216,18 @@\n Keys are device ids as integers.\n Values are memory usage as integers in MB.\n \"\"\"\n- result = subprocess.check_output(\n+ result = subprocess.run(\n [\n- 'nvidia-smi', '--query-gpu=memory.used',\n- '--format=csv,nounits,noheader'\n- ], encoding='utf-8')\n+ 'nvidia-smi',\n+ '--query-gpu=memory.used',\n+ '--format=csv,nounits,noheader',\n+ ],\n+ encoding='utf-8',\n+ capture_output=True,\n+ check=True)\n # Convert lines into a dictionary\n- gpu_memory = [int(x) for x in result.strip().split('\\n')]\n- gpu_memory_map = {}\n- for k, v in zip(range(len(gpu_memory)), gpu_memory):\n- k = f'gpu_{k}'\n- gpu_memory_map[k] = v\n+ gpu_memory = [int(x) for x in result.stdout.strip().split(os.linesep)]\n+ gpu_memory_map = {f'gpu_{index}': memory for index, memory in enumerate(gpu_memory)}\n return gpu_memory_map\n", "issue": "min_max log_gpu_memory option bug\n**Describe the bug**\r\nSetting `log_gpu_memory='min_max'` in `Trainer` leads to the following bug.\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"/opt/conda/lib/python3.7/site-packages/pytorch_lightning/trainer/trainer.py\", line 347, in fit\r\n self.single_gpu_train(model)\r\n File \"/opt/conda/lib/python3.7/site-packages/pytorch_lightning/trainer/dp_mixin.py\", line 79, in single_gpu_train\r\n self.run_pretrain_routine(model)\r\n File \"/opt/conda/lib/python3.7/site-packages/pytorch_lightning/trainer/trainer.py\", line 467, in run_pretrain_routine\r\n self.train()\r\n File \"/opt/conda/lib/python3.7/site-packages/pytorch_lightning/trainer/train_loop_mixin.py\", line 60, in train\r\n self.run_training_epoch()\r\n File \"/opt/conda/lib/python3.7/site-packages/pytorch_lightning/trainer/train_loop_mixin.py\", line 126, in run_training_epoch\r\n self.log_metrics(batch_step_metrics, grad_norm_dic)\r\n File \"/opt/conda/lib/python3.7/site-packages/pytorch_lightning/trainer/logging_mixin.py\", line 20, in log_metrics\r\n mem_map = memory.get_memory_profile(self.log_gpu_memory)\r\n File \"/opt/conda/lib/python3.7/site-packages/pytorch_lightning/root_module/memory.py\", line 205, in get_memory_profile\r\n for k, v in memory_map:\r\nValueError: too many values to unpack (expected 2)\r\n```\r\n\r\n\r\n**To Reproduce**\r\nOn current master, execute the following.\r\n```\r\n trainer = Trainer(\r\n ...\r\n log_gpu_memory='min_max',\r\n ...\r\n )\r\n trainer.fit(model)\r\n```\r\n\r\n**Expected behavior**\r\nLog the min/max utilization of gpu memory, as `min_max` option is documented.\r\n\r\n**Desktop (please complete the following information):**\r\n - OS: Ubuntu 18.04 \r\n - Version: Current master\r\n\r\nI am working on this issue. Will submit a PR soon.\n", "before_files": [{"content": "'''\nGenerates a summary of a model's layers and dimensionality\n'''\n\nimport gc\nimport subprocess\n\nimport numpy as np\nimport pandas as pd\nimport torch\n\n\nclass ModelSummary(object):\n\n def __init__(self, model, mode='full'):\n '''\n Generates summaries of model layers and dimensions.\n '''\n self.model = model\n self.mode = mode\n self.in_sizes = []\n self.out_sizes = []\n\n self.summarize()\n\n def __str__(self):\n return self.summary.__str__()\n\n def __repr__(self):\n return self.summary.__str__()\n\n def named_modules(self):\n if self.mode == 'full':\n mods = self.model.named_modules()\n mods = list(mods)[1:] # do not include root module (LightningModule)\n elif self.mode == 'top':\n # the children are the top-level modules\n mods = self.model.named_children()\n else:\n mods = []\n return list(mods)\n\n def get_variable_sizes(self):\n '''Run sample input through each layer to get output sizes'''\n mods = self.named_modules()\n in_sizes = []\n out_sizes = []\n input_ = self.model.example_input_array\n\n if self.model.on_gpu:\n input_ = input_.cuda(0)\n\n if self.model.trainer.use_amp:\n input_ = input_.half()\n\n with torch.no_grad():\n\n for _, m in mods:\n if type(input_) is list or type(input_) is tuple: # pragma: no cover\n out = m(*input_)\n else:\n out = m(input_)\n\n if type(input_) is tuple or type(input_) is list: # pragma: no cover\n in_size = []\n for x in input_:\n if type(x) is list:\n in_size.append(len(x))\n else:\n in_size.append(x.size())\n else:\n in_size = np.array(input_.size())\n\n in_sizes.append(in_size)\n\n if type(out) is tuple or type(out) is list: # pragma: no cover\n out_size = np.asarray([x.size() for x in out])\n else:\n out_size = np.array(out.size())\n\n out_sizes.append(out_size)\n input_ = out\n\n self.in_sizes = in_sizes\n self.out_sizes = out_sizes\n assert len(in_sizes) == len(out_sizes)\n return\n\n def get_layer_names(self):\n '''Collect Layer Names'''\n mods = self.named_modules()\n names = []\n layers = []\n for name, m in mods:\n names += [name]\n layers += [str(m.__class__)]\n\n layer_types = [x.split('.')[-1][:-2] for x in layers]\n\n self.layer_names = names\n self.layer_types = layer_types\n return\n\n def get_parameter_sizes(self):\n '''Get sizes of all parameters in `model`'''\n mods = self.named_modules()\n sizes = []\n for _, m in mods:\n p = list(m.parameters())\n modsz = []\n for j in range(len(p)):\n modsz.append(np.array(p[j].size()))\n sizes.append(modsz)\n\n self.param_sizes = sizes\n return\n\n def get_parameter_nums(self):\n '''Get number of parameters in each layer'''\n param_nums = []\n for mod in self.param_sizes:\n all_params = 0\n for p in mod:\n all_params += np.prod(p)\n param_nums.append(all_params)\n self.param_nums = param_nums\n return\n\n def make_summary(self):\n '''\n Makes a summary listing with:\n\n Layer Name, Layer Type, Input Size, Output Size, Number of Parameters\n '''\n\n cols = ['Name', 'Type', 'Params']\n if self.model.example_input_array is not None:\n cols.extend(['In_sizes', 'Out_sizes'])\n\n df = pd.DataFrame(np.zeros((len(self.layer_names), len(cols))))\n df.columns = cols\n\n df['Name'] = self.layer_names\n df['Type'] = self.layer_types\n df['Params'] = self.param_nums\n df['Params'] = df['Params'].map(get_human_readable_count)\n\n if self.model.example_input_array is not None:\n df['In_sizes'] = self.in_sizes\n df['Out_sizes'] = self.out_sizes\n\n self.summary = df\n return\n\n def summarize(self):\n self.get_layer_names()\n self.get_parameter_sizes()\n self.get_parameter_nums()\n\n if self.model.example_input_array is not None:\n self.get_variable_sizes()\n self.make_summary()\n\n\ndef print_mem_stack(): # pragma: no cover\n for obj in gc.get_objects():\n try:\n if torch.is_tensor(obj) or (hasattr(obj, 'data') and torch.is_tensor(obj.data)):\n print(type(obj), obj.size())\n except Exception:\n pass\n\n\ndef count_mem_items(): # pragma: no cover\n nb_params = 0\n nb_tensors = 0\n for obj in gc.get_objects():\n try:\n if torch.is_tensor(obj) or (hasattr(obj, 'data') and torch.is_tensor(obj.data)):\n obj_type = str(type(obj))\n if 'parameter' in obj_type:\n nb_params += 1\n else:\n nb_tensors += 1\n except Exception:\n pass\n\n return nb_params, nb_tensors\n\n\ndef get_memory_profile(mode):\n \"\"\"\n 'all' means return memory for all gpus\n 'min_max' means return memory for max and min\n :param mode:\n :return:\n \"\"\"\n memory_map = get_gpu_memory_map()\n\n if mode == 'min_max':\n min_mem = 1000000\n min_k = None\n max_mem = 0\n max_k = None\n for k, v in memory_map:\n if v > max_mem:\n max_mem = v\n max_k = k\n if v < min_mem:\n min_mem = v\n min_k = k\n\n memory_map = {min_k: min_mem, max_k: max_mem}\n\n return memory_map\n\n\ndef get_gpu_memory_map():\n \"\"\"Get the current gpu usage.\n\n Returns\n -------\n usage: dict\n Keys are device ids as integers.\n Values are memory usage as integers in MB.\n \"\"\"\n result = subprocess.check_output(\n [\n 'nvidia-smi', '--query-gpu=memory.used',\n '--format=csv,nounits,noheader'\n ], encoding='utf-8')\n # Convert lines into a dictionary\n gpu_memory = [int(x) for x in result.strip().split('\\n')]\n gpu_memory_map = {}\n for k, v in zip(range(len(gpu_memory)), gpu_memory):\n k = f'gpu_{k}'\n gpu_memory_map[k] = v\n return gpu_memory_map\n\n\ndef get_human_readable_count(number):\n \"\"\"\n Abbreviates an integer number with K, M, B, T for thousands, millions,\n billions and trillions, respectively.\n Examples:\n 123 -> 123\n 1234 -> 1 K (one thousand)\n 2e6 -> 2 M (two million)\n 3e9 -> 3 B (three billion)\n 4e12 -> 4 T (four trillion)\n 5e15 -> 5,000 T\n :param number: a positive integer number\n :returns a string formatted according to the pattern described above.\n \"\"\"\n assert number >= 0\n labels = [' ', 'K', 'M', 'B', 'T']\n num_digits = int(np.floor(np.log10(number)) + 1 if number > 0 else 1)\n num_groups = int(np.ceil(num_digits / 3))\n num_groups = min(num_groups, len(labels)) # don't abbreviate beyond trillions\n shift = -3 * (num_groups - 1)\n number = number * (10 ** shift)\n index = num_groups - 1\n return f'{int(number):,d} {labels[index]}'\n", "path": "pytorch_lightning/root_module/memory.py"}], "after_files": [{"content": "'''\nGenerates a summary of a model's layers and dimensionality\n'''\n\nimport gc\nimport os\nimport subprocess\n\nimport numpy as np\nimport pandas as pd\nimport torch\n\n\nclass ModelSummary(object):\n\n def __init__(self, model, mode='full'):\n '''\n Generates summaries of model layers and dimensions.\n '''\n self.model = model\n self.mode = mode\n self.in_sizes = []\n self.out_sizes = []\n\n self.summarize()\n\n def __str__(self):\n return self.summary.__str__()\n\n def __repr__(self):\n return self.summary.__str__()\n\n def named_modules(self):\n if self.mode == 'full':\n mods = self.model.named_modules()\n mods = list(mods)[1:] # do not include root module (LightningModule)\n elif self.mode == 'top':\n # the children are the top-level modules\n mods = self.model.named_children()\n else:\n mods = []\n return list(mods)\n\n def get_variable_sizes(self):\n '''Run sample input through each layer to get output sizes'''\n mods = self.named_modules()\n in_sizes = []\n out_sizes = []\n input_ = self.model.example_input_array\n\n if self.model.on_gpu:\n input_ = input_.cuda(0)\n\n if self.model.trainer.use_amp:\n input_ = input_.half()\n\n with torch.no_grad():\n\n for _, m in mods:\n if type(input_) is list or type(input_) is tuple: # pragma: no cover\n out = m(*input_)\n else:\n out = m(input_)\n\n if type(input_) is tuple or type(input_) is list: # pragma: no cover\n in_size = []\n for x in input_:\n if type(x) is list:\n in_size.append(len(x))\n else:\n in_size.append(x.size())\n else:\n in_size = np.array(input_.size())\n\n in_sizes.append(in_size)\n\n if type(out) is tuple or type(out) is list: # pragma: no cover\n out_size = np.asarray([x.size() for x in out])\n else:\n out_size = np.array(out.size())\n\n out_sizes.append(out_size)\n input_ = out\n\n self.in_sizes = in_sizes\n self.out_sizes = out_sizes\n assert len(in_sizes) == len(out_sizes)\n return\n\n def get_layer_names(self):\n '''Collect Layer Names'''\n mods = self.named_modules()\n names = []\n layers = []\n for name, m in mods:\n names += [name]\n layers += [str(m.__class__)]\n\n layer_types = [x.split('.')[-1][:-2] for x in layers]\n\n self.layer_names = names\n self.layer_types = layer_types\n return\n\n def get_parameter_sizes(self):\n '''Get sizes of all parameters in `model`'''\n mods = self.named_modules()\n sizes = []\n for _, m in mods:\n p = list(m.parameters())\n modsz = []\n for j in range(len(p)):\n modsz.append(np.array(p[j].size()))\n sizes.append(modsz)\n\n self.param_sizes = sizes\n return\n\n def get_parameter_nums(self):\n '''Get number of parameters in each layer'''\n param_nums = []\n for mod in self.param_sizes:\n all_params = 0\n for p in mod:\n all_params += np.prod(p)\n param_nums.append(all_params)\n self.param_nums = param_nums\n return\n\n def make_summary(self):\n '''\n Makes a summary listing with:\n\n Layer Name, Layer Type, Input Size, Output Size, Number of Parameters\n '''\n\n cols = ['Name', 'Type', 'Params']\n if self.model.example_input_array is not None:\n cols.extend(['In_sizes', 'Out_sizes'])\n\n df = pd.DataFrame(np.zeros((len(self.layer_names), len(cols))))\n df.columns = cols\n\n df['Name'] = self.layer_names\n df['Type'] = self.layer_types\n df['Params'] = self.param_nums\n df['Params'] = df['Params'].map(get_human_readable_count)\n\n if self.model.example_input_array is not None:\n df['In_sizes'] = self.in_sizes\n df['Out_sizes'] = self.out_sizes\n\n self.summary = df\n return\n\n def summarize(self):\n self.get_layer_names()\n self.get_parameter_sizes()\n self.get_parameter_nums()\n\n if self.model.example_input_array is not None:\n self.get_variable_sizes()\n self.make_summary()\n\n\ndef print_mem_stack(): # pragma: no cover\n for obj in gc.get_objects():\n try:\n if torch.is_tensor(obj) or (hasattr(obj, 'data') and torch.is_tensor(obj.data)):\n print(type(obj), obj.size())\n except Exception:\n pass\n\n\ndef count_mem_items(): # pragma: no cover\n nb_params = 0\n nb_tensors = 0\n for obj in gc.get_objects():\n try:\n if torch.is_tensor(obj) or (hasattr(obj, 'data') and torch.is_tensor(obj.data)):\n obj_type = str(type(obj))\n if 'parameter' in obj_type:\n nb_params += 1\n else:\n nb_tensors += 1\n except Exception:\n pass\n\n return nb_params, nb_tensors\n\n\ndef get_memory_profile(mode):\n \"\"\"\n 'all' means return memory for all gpus\n 'min_max' means return memory for max and min\n :param mode:\n :return:\n \"\"\"\n memory_map = get_gpu_memory_map()\n\n if mode == 'min_max':\n min_index, min_memory = min(memory_map.items(), key=lambda item: item[1])\n max_index, max_memory = max(memory_map.items(), key=lambda item: item[1])\n\n memory_map = {min_index: min_memory, max_index: max_memory}\n\n return memory_map\n\n\ndef get_gpu_memory_map():\n \"\"\"Get the current gpu usage.\n\n Returns\n -------\n usage: dict\n Keys are device ids as integers.\n Values are memory usage as integers in MB.\n \"\"\"\n result = subprocess.run(\n [\n 'nvidia-smi',\n '--query-gpu=memory.used',\n '--format=csv,nounits,noheader',\n ],\n encoding='utf-8',\n capture_output=True,\n check=True)\n # Convert lines into a dictionary\n gpu_memory = [int(x) for x in result.stdout.strip().split(os.linesep)]\n gpu_memory_map = {f'gpu_{index}': memory for index, memory in enumerate(gpu_memory)}\n return gpu_memory_map\n\n\ndef get_human_readable_count(number):\n \"\"\"\n Abbreviates an integer number with K, M, B, T for thousands, millions,\n billions and trillions, respectively.\n Examples:\n 123 -> 123\n 1234 -> 1 K (one thousand)\n 2e6 -> 2 M (two million)\n 3e9 -> 3 B (three billion)\n 4e12 -> 4 T (four trillion)\n 5e15 -> 5,000 T\n :param number: a positive integer number\n :returns a string formatted according to the pattern described above.\n \"\"\"\n assert number >= 0\n labels = [' ', 'K', 'M', 'B', 'T']\n num_digits = int(np.floor(np.log10(number)) + 1 if number > 0 else 1)\n num_groups = int(np.ceil(num_digits / 3))\n num_groups = min(num_groups, len(labels)) # don't abbreviate beyond trillions\n shift = -3 * (num_groups - 1)\n number = number * (10 ** shift)\n index = num_groups - 1\n return f'{int(number):,d} {labels[index]}'\n", "path": "pytorch_lightning/root_module/memory.py"}]} | 3,201 | 546 |
gh_patches_debug_15202 | rasdani/github-patches | git_diff | vega__altair-1265 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
line_percent.py does not work offline
We need all examples to work offline. Currently ``line_percent.py`` uses ``pd.read_json`` from a URL.
The example should probably use a URL plus a filter.
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `altair/examples/boxplot_max_min.py`
Content:
```
1 """
2 Box Plot with Min/Max Whiskers
3 ------------------------------
4 This example shows how to make a basic box plot using US Population data from 2000.
5 """
6 # category: other charts
7 import altair as alt
8 from vega_datasets import data
9
10 source = data.population()
11
12 base = alt.Chart(source)
13
14 # Define aggregate fields
15 lower_box = 'q1(people):Q'
16 lower_whisker = 'min(people):Q'
17 upper_box = 'q3(people):Q'
18 upper_whisker = 'max(people):Q'
19
20 # Compose each layer individually
21 lower_plot = base.mark_rule().encode(
22 y=alt.Y(lower_whisker, title="population"),
23 y2=lower_box,
24 x='age:O'
25 )
26
27 middle_plot = base.mark_bar(size=5.0).encode(
28 y=lower_box,
29 y2=upper_box,
30 x='age:O'
31 )
32
33 upper_plot = base.mark_rule().encode(
34 y=upper_whisker,
35 y2=upper_box,
36 x='age:O'
37 )
38
39 middle_tick = base.mark_tick(
40 color='white',
41 size=5.0
42 ).encode(
43 y='median(people):Q',
44 x='age:O',
45 )
46
47 lower_plot + middle_plot + upper_plot + middle_tick
48
```
Path: `altair/examples/line_percent.py`
Content:
```
1 """
2 Line Chart with Percent axis
3 ----------------------------
4 This example shows how to format the tick labels of the y-axis of a chart as percentages.
5 """
6 # category: line charts
7 import altair as alt
8 import pandas as pd
9 from vega_datasets import data
10
11 source = pd.read_json(data.jobs.url)
12 welders = source[source.job == 'Welder']
13
14 alt.Chart(welders).mark_line().encode(
15 alt.X('year:O'),
16 alt.Y('perc:Q', axis=alt.Axis(format='%')),
17 color='sex:N'
18 )
19
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/altair/examples/boxplot_max_min.py b/altair/examples/boxplot_max_min.py
--- a/altair/examples/boxplot_max_min.py
+++ b/altair/examples/boxplot_max_min.py
@@ -7,7 +7,7 @@
import altair as alt
from vega_datasets import data
-source = data.population()
+source = data.population.url
base = alt.Chart(source)
diff --git a/altair/examples/line_percent.py b/altair/examples/line_percent.py
--- a/altair/examples/line_percent.py
+++ b/altair/examples/line_percent.py
@@ -5,14 +5,14 @@
"""
# category: line charts
import altair as alt
-import pandas as pd
from vega_datasets import data
-source = pd.read_json(data.jobs.url)
-welders = source[source.job == 'Welder']
+source = data.jobs.url
-alt.Chart(welders).mark_line().encode(
+alt.Chart(source).mark_line().encode(
alt.X('year:O'),
alt.Y('perc:Q', axis=alt.Axis(format='%')),
color='sex:N'
+).transform_filter(
+ alt.datum.job == 'Welder'
)
| {"golden_diff": "diff --git a/altair/examples/boxplot_max_min.py b/altair/examples/boxplot_max_min.py\n--- a/altair/examples/boxplot_max_min.py\n+++ b/altair/examples/boxplot_max_min.py\n@@ -7,7 +7,7 @@\n import altair as alt\n from vega_datasets import data\n \n-source = data.population()\n+source = data.population.url\n \n base = alt.Chart(source)\n \ndiff --git a/altair/examples/line_percent.py b/altair/examples/line_percent.py\n--- a/altair/examples/line_percent.py\n+++ b/altair/examples/line_percent.py\n@@ -5,14 +5,14 @@\n \"\"\"\n # category: line charts\n import altair as alt\n-import pandas as pd\n from vega_datasets import data\n \n-source = pd.read_json(data.jobs.url)\n-welders = source[source.job == 'Welder']\n+source = data.jobs.url\n \n-alt.Chart(welders).mark_line().encode(\n+alt.Chart(source).mark_line().encode(\n alt.X('year:O'),\n alt.Y('perc:Q', axis=alt.Axis(format='%')),\n color='sex:N'\n+).transform_filter(\n+ alt.datum.job == 'Welder'\n )\n", "issue": "line_percent.py does not work offline\nWe need all examples to work offline. Currently ``line_percent.py`` uses ``pd.read_json`` from a URL.\r\n\r\nThe example should probably use a URL plus a filter.\n", "before_files": [{"content": "\"\"\"\nBox Plot with Min/Max Whiskers\n------------------------------\nThis example shows how to make a basic box plot using US Population data from 2000.\n\"\"\"\n# category: other charts\nimport altair as alt\nfrom vega_datasets import data\n\nsource = data.population()\n\nbase = alt.Chart(source)\n\n# Define aggregate fields\nlower_box = 'q1(people):Q'\nlower_whisker = 'min(people):Q'\nupper_box = 'q3(people):Q'\nupper_whisker = 'max(people):Q'\n\n# Compose each layer individually\nlower_plot = base.mark_rule().encode(\n y=alt.Y(lower_whisker, title=\"population\"),\n y2=lower_box,\n x='age:O'\n)\n\nmiddle_plot = base.mark_bar(size=5.0).encode(\n y=lower_box,\n y2=upper_box,\n x='age:O'\n)\n\nupper_plot = base.mark_rule().encode(\n y=upper_whisker,\n y2=upper_box,\n x='age:O'\n)\n\nmiddle_tick = base.mark_tick(\n color='white',\n size=5.0\n).encode(\n y='median(people):Q',\n x='age:O',\n)\n\nlower_plot + middle_plot + upper_plot + middle_tick\n", "path": "altair/examples/boxplot_max_min.py"}, {"content": "\"\"\"\nLine Chart with Percent axis\n----------------------------\nThis example shows how to format the tick labels of the y-axis of a chart as percentages.\n\"\"\"\n# category: line charts\nimport altair as alt\nimport pandas as pd\nfrom vega_datasets import data\n\nsource = pd.read_json(data.jobs.url)\nwelders = source[source.job == 'Welder']\n\nalt.Chart(welders).mark_line().encode(\n alt.X('year:O'),\n alt.Y('perc:Q', axis=alt.Axis(format='%')),\n color='sex:N'\n)\n", "path": "altair/examples/line_percent.py"}], "after_files": [{"content": "\"\"\"\nBox Plot with Min/Max Whiskers\n------------------------------\nThis example shows how to make a basic box plot using US Population data from 2000.\n\"\"\"\n# category: other charts\nimport altair as alt\nfrom vega_datasets import data\n\nsource = data.population.url\n\nbase = alt.Chart(source)\n\n# Define aggregate fields\nlower_box = 'q1(people):Q'\nlower_whisker = 'min(people):Q'\nupper_box = 'q3(people):Q'\nupper_whisker = 'max(people):Q'\n\n# Compose each layer individually\nlower_plot = base.mark_rule().encode(\n y=alt.Y(lower_whisker, title=\"population\"),\n y2=lower_box,\n x='age:O'\n)\n\nmiddle_plot = base.mark_bar(size=5.0).encode(\n y=lower_box,\n y2=upper_box,\n x='age:O'\n)\n\nupper_plot = base.mark_rule().encode(\n y=upper_whisker,\n y2=upper_box,\n x='age:O'\n)\n\nmiddle_tick = base.mark_tick(\n color='white',\n size=5.0\n).encode(\n y='median(people):Q',\n x='age:O',\n)\n\nlower_plot + middle_plot + upper_plot + middle_tick\n", "path": "altair/examples/boxplot_max_min.py"}, {"content": "\"\"\"\nLine Chart with Percent axis\n----------------------------\nThis example shows how to format the tick labels of the y-axis of a chart as percentages.\n\"\"\"\n# category: line charts\nimport altair as alt\nfrom vega_datasets import data\n\nsource = data.jobs.url\n\nalt.Chart(source).mark_line().encode(\n alt.X('year:O'),\n alt.Y('perc:Q', axis=alt.Axis(format='%')),\n color='sex:N'\n).transform_filter(\n alt.datum.job == 'Welder'\n)\n", "path": "altair/examples/line_percent.py"}]} | 843 | 270 |
gh_patches_debug_14641 | rasdani/github-patches | git_diff | scrapy__scrapy-5993 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Backward compatibility in utils.conf.build_component_list
There is some code from 2015 in `scrapy.utils.conf.build_component_list` marked as "Backward compatibility for old (base, custom) call signature", which was added in #1586. I couldn't understand after a quick glance why is it "backward compatibility" but if it's something deprecated we should deprecare it properly with a message, and if it's a properly supported code path we should remove the comments.
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `scrapy/utils/conf.py`
Content:
```
1 import numbers
2 import os
3 import sys
4 import warnings
5 from configparser import ConfigParser
6 from operator import itemgetter
7 from pathlib import Path
8 from typing import Any, Dict, List, Optional, Union
9
10 from scrapy.exceptions import ScrapyDeprecationWarning, UsageError
11 from scrapy.settings import BaseSettings
12 from scrapy.utils.deprecate import update_classpath
13 from scrapy.utils.python import without_none_values
14
15
16 def build_component_list(compdict, custom=None, convert=update_classpath):
17 """Compose a component list from a { class: order } dictionary."""
18
19 def _check_components(complist):
20 if len({convert(c) for c in complist}) != len(complist):
21 raise ValueError(
22 f"Some paths in {complist!r} convert to the same object, "
23 "please update your settings"
24 )
25
26 def _map_keys(compdict):
27 if isinstance(compdict, BaseSettings):
28 compbs = BaseSettings()
29 for k, v in compdict.items():
30 prio = compdict.getpriority(k)
31 assert prio is not None
32 if compbs.getpriority(convert(k)) == prio:
33 raise ValueError(
34 f"Some paths in {list(compdict.keys())!r} "
35 "convert to the same "
36 "object, please update your settings"
37 )
38 else:
39 compbs.set(convert(k), v, priority=prio)
40 return compbs
41 _check_components(compdict)
42 return {convert(k): v for k, v in compdict.items()}
43
44 def _validate_values(compdict):
45 """Fail if a value in the components dict is not a real number or None."""
46 for name, value in compdict.items():
47 if value is not None and not isinstance(value, numbers.Real):
48 raise ValueError(
49 f"Invalid value {value} for component {name}, "
50 "please provide a real number or None instead"
51 )
52
53 if isinstance(custom, (list, tuple)):
54 _check_components(custom)
55 return type(custom)(convert(c) for c in custom)
56
57 if custom is not None:
58 compdict.update(custom)
59
60 _validate_values(compdict)
61 compdict = without_none_values(_map_keys(compdict))
62 return [k for k, v in sorted(compdict.items(), key=itemgetter(1))]
63
64
65 def arglist_to_dict(arglist):
66 """Convert a list of arguments like ['arg1=val1', 'arg2=val2', ...] to a
67 dict
68 """
69 return dict(x.split("=", 1) for x in arglist)
70
71
72 def closest_scrapy_cfg(
73 path: Union[str, os.PathLike] = ".",
74 prevpath: Optional[Union[str, os.PathLike]] = None,
75 ) -> str:
76 """Return the path to the closest scrapy.cfg file by traversing the current
77 directory and its parents
78 """
79 if prevpath is not None and str(path) == str(prevpath):
80 return ""
81 path = Path(path).resolve()
82 cfgfile = path / "scrapy.cfg"
83 if cfgfile.exists():
84 return str(cfgfile)
85 return closest_scrapy_cfg(path.parent, path)
86
87
88 def init_env(project="default", set_syspath=True):
89 """Initialize environment to use command-line tool from inside a project
90 dir. This sets the Scrapy settings module and modifies the Python path to
91 be able to locate the project module.
92 """
93 cfg = get_config()
94 if cfg.has_option("settings", project):
95 os.environ["SCRAPY_SETTINGS_MODULE"] = cfg.get("settings", project)
96 closest = closest_scrapy_cfg()
97 if closest:
98 projdir = str(Path(closest).parent)
99 if set_syspath and projdir not in sys.path:
100 sys.path.append(projdir)
101
102
103 def get_config(use_closest=True):
104 """Get Scrapy config file as a ConfigParser"""
105 sources = get_sources(use_closest)
106 cfg = ConfigParser()
107 cfg.read(sources)
108 return cfg
109
110
111 def get_sources(use_closest=True) -> List[str]:
112 xdg_config_home = (
113 os.environ.get("XDG_CONFIG_HOME") or Path("~/.config").expanduser()
114 )
115 sources = [
116 "/etc/scrapy.cfg",
117 r"c:\scrapy\scrapy.cfg",
118 str(Path(xdg_config_home) / "scrapy.cfg"),
119 str(Path("~/.scrapy.cfg").expanduser()),
120 ]
121 if use_closest:
122 sources.append(closest_scrapy_cfg())
123 return sources
124
125
126 def feed_complete_default_values_from_settings(feed, settings):
127 out = feed.copy()
128 out.setdefault("batch_item_count", settings.getint("FEED_EXPORT_BATCH_ITEM_COUNT"))
129 out.setdefault("encoding", settings["FEED_EXPORT_ENCODING"])
130 out.setdefault("fields", settings.getdictorlist("FEED_EXPORT_FIELDS") or None)
131 out.setdefault("store_empty", settings.getbool("FEED_STORE_EMPTY"))
132 out.setdefault("uri_params", settings["FEED_URI_PARAMS"])
133 out.setdefault("item_export_kwargs", {})
134 if settings["FEED_EXPORT_INDENT"] is None:
135 out.setdefault("indent", None)
136 else:
137 out.setdefault("indent", settings.getint("FEED_EXPORT_INDENT"))
138 return out
139
140
141 def feed_process_params_from_cli(
142 settings,
143 output: List[str],
144 output_format=None,
145 overwrite_output: Optional[List[str]] = None,
146 ):
147 """
148 Receives feed export params (from the 'crawl' or 'runspider' commands),
149 checks for inconsistencies in their quantities and returns a dictionary
150 suitable to be used as the FEEDS setting.
151 """
152 valid_output_formats = without_none_values(
153 settings.getwithbase("FEED_EXPORTERS")
154 ).keys()
155
156 def check_valid_format(output_format):
157 if output_format not in valid_output_formats:
158 raise UsageError(
159 f"Unrecognized output format '{output_format}'. "
160 f"Set a supported one ({tuple(valid_output_formats)}) "
161 "after a colon at the end of the output URI (i.e. -o/-O "
162 "<URI>:<FORMAT>) or as a file extension."
163 )
164
165 overwrite = False
166 if overwrite_output:
167 if output:
168 raise UsageError(
169 "Please use only one of -o/--output and -O/--overwrite-output"
170 )
171 if output_format:
172 raise UsageError(
173 "-t/--output-format is a deprecated command line option"
174 " and does not work in combination with -O/--overwrite-output."
175 " To specify a format please specify it after a colon at the end of the"
176 " output URI (i.e. -O <URI>:<FORMAT>)."
177 " Example working in the tutorial: "
178 "scrapy crawl quotes -O quotes.json:json"
179 )
180 output = overwrite_output
181 overwrite = True
182
183 if output_format:
184 if len(output) == 1:
185 check_valid_format(output_format)
186 message = (
187 "The -t/--output-format command line option is deprecated in favor of "
188 "specifying the output format within the output URI using the -o/--output or the"
189 " -O/--overwrite-output option (i.e. -o/-O <URI>:<FORMAT>). See the documentation"
190 " of the -o or -O option or the following examples for more information. "
191 "Examples working in the tutorial: "
192 "scrapy crawl quotes -o quotes.csv:csv or "
193 "scrapy crawl quotes -O quotes.json:json"
194 )
195 warnings.warn(message, ScrapyDeprecationWarning, stacklevel=2)
196 return {output[0]: {"format": output_format}}
197 raise UsageError(
198 "The -t command-line option cannot be used if multiple output "
199 "URIs are specified"
200 )
201
202 result: Dict[str, Dict[str, Any]] = {}
203 for element in output:
204 try:
205 feed_uri, feed_format = element.rsplit(":", 1)
206 check_valid_format(feed_format)
207 except (ValueError, UsageError):
208 feed_uri = element
209 feed_format = Path(element).suffix.replace(".", "")
210 else:
211 if feed_uri == "-":
212 feed_uri = "stdout:"
213 check_valid_format(feed_format)
214 result[feed_uri] = {"format": feed_format}
215 if overwrite:
216 result[feed_uri]["overwrite"] = True
217
218 # FEEDS setting should take precedence over the matching CLI options
219 result.update(settings.getdict("FEEDS"))
220
221 return result
222
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/scrapy/utils/conf.py b/scrapy/utils/conf.py
--- a/scrapy/utils/conf.py
+++ b/scrapy/utils/conf.py
@@ -50,11 +50,17 @@
"please provide a real number or None instead"
)
- if isinstance(custom, (list, tuple)):
- _check_components(custom)
- return type(custom)(convert(c) for c in custom)
-
if custom is not None:
+ warnings.warn(
+ "The 'custom' attribute of build_component_list() is deprecated. "
+ "Please merge its value into 'compdict' manually or change your "
+ "code to use Settings.getwithbase().",
+ category=ScrapyDeprecationWarning,
+ stacklevel=2,
+ )
+ if isinstance(custom, (list, tuple)):
+ _check_components(custom)
+ return type(custom)(convert(c) for c in custom)
compdict.update(custom)
_validate_values(compdict)
| {"golden_diff": "diff --git a/scrapy/utils/conf.py b/scrapy/utils/conf.py\n--- a/scrapy/utils/conf.py\n+++ b/scrapy/utils/conf.py\n@@ -50,11 +50,17 @@\n \"please provide a real number or None instead\"\n )\n \n- if isinstance(custom, (list, tuple)):\n- _check_components(custom)\n- return type(custom)(convert(c) for c in custom)\n-\n if custom is not None:\n+ warnings.warn(\n+ \"The 'custom' attribute of build_component_list() is deprecated. \"\n+ \"Please merge its value into 'compdict' manually or change your \"\n+ \"code to use Settings.getwithbase().\",\n+ category=ScrapyDeprecationWarning,\n+ stacklevel=2,\n+ )\n+ if isinstance(custom, (list, tuple)):\n+ _check_components(custom)\n+ return type(custom)(convert(c) for c in custom)\n compdict.update(custom)\n \n _validate_values(compdict)\n", "issue": "Backward compatibility in utils.conf.build_component_list\nThere is some code from 2015 in `scrapy.utils.conf.build_component_list` marked as \"Backward compatibility for old (base, custom) call signature\", which was added in #1586. I couldn't understand after a quick glance why is it \"backward compatibility\" but if it's something deprecated we should deprecare it properly with a message, and if it's a properly supported code path we should remove the comments.\n", "before_files": [{"content": "import numbers\nimport os\nimport sys\nimport warnings\nfrom configparser import ConfigParser\nfrom operator import itemgetter\nfrom pathlib import Path\nfrom typing import Any, Dict, List, Optional, Union\n\nfrom scrapy.exceptions import ScrapyDeprecationWarning, UsageError\nfrom scrapy.settings import BaseSettings\nfrom scrapy.utils.deprecate import update_classpath\nfrom scrapy.utils.python import without_none_values\n\n\ndef build_component_list(compdict, custom=None, convert=update_classpath):\n \"\"\"Compose a component list from a { class: order } dictionary.\"\"\"\n\n def _check_components(complist):\n if len({convert(c) for c in complist}) != len(complist):\n raise ValueError(\n f\"Some paths in {complist!r} convert to the same object, \"\n \"please update your settings\"\n )\n\n def _map_keys(compdict):\n if isinstance(compdict, BaseSettings):\n compbs = BaseSettings()\n for k, v in compdict.items():\n prio = compdict.getpriority(k)\n assert prio is not None\n if compbs.getpriority(convert(k)) == prio:\n raise ValueError(\n f\"Some paths in {list(compdict.keys())!r} \"\n \"convert to the same \"\n \"object, please update your settings\"\n )\n else:\n compbs.set(convert(k), v, priority=prio)\n return compbs\n _check_components(compdict)\n return {convert(k): v for k, v in compdict.items()}\n\n def _validate_values(compdict):\n \"\"\"Fail if a value in the components dict is not a real number or None.\"\"\"\n for name, value in compdict.items():\n if value is not None and not isinstance(value, numbers.Real):\n raise ValueError(\n f\"Invalid value {value} for component {name}, \"\n \"please provide a real number or None instead\"\n )\n\n if isinstance(custom, (list, tuple)):\n _check_components(custom)\n return type(custom)(convert(c) for c in custom)\n\n if custom is not None:\n compdict.update(custom)\n\n _validate_values(compdict)\n compdict = without_none_values(_map_keys(compdict))\n return [k for k, v in sorted(compdict.items(), key=itemgetter(1))]\n\n\ndef arglist_to_dict(arglist):\n \"\"\"Convert a list of arguments like ['arg1=val1', 'arg2=val2', ...] to a\n dict\n \"\"\"\n return dict(x.split(\"=\", 1) for x in arglist)\n\n\ndef closest_scrapy_cfg(\n path: Union[str, os.PathLike] = \".\",\n prevpath: Optional[Union[str, os.PathLike]] = None,\n) -> str:\n \"\"\"Return the path to the closest scrapy.cfg file by traversing the current\n directory and its parents\n \"\"\"\n if prevpath is not None and str(path) == str(prevpath):\n return \"\"\n path = Path(path).resolve()\n cfgfile = path / \"scrapy.cfg\"\n if cfgfile.exists():\n return str(cfgfile)\n return closest_scrapy_cfg(path.parent, path)\n\n\ndef init_env(project=\"default\", set_syspath=True):\n \"\"\"Initialize environment to use command-line tool from inside a project\n dir. This sets the Scrapy settings module and modifies the Python path to\n be able to locate the project module.\n \"\"\"\n cfg = get_config()\n if cfg.has_option(\"settings\", project):\n os.environ[\"SCRAPY_SETTINGS_MODULE\"] = cfg.get(\"settings\", project)\n closest = closest_scrapy_cfg()\n if closest:\n projdir = str(Path(closest).parent)\n if set_syspath and projdir not in sys.path:\n sys.path.append(projdir)\n\n\ndef get_config(use_closest=True):\n \"\"\"Get Scrapy config file as a ConfigParser\"\"\"\n sources = get_sources(use_closest)\n cfg = ConfigParser()\n cfg.read(sources)\n return cfg\n\n\ndef get_sources(use_closest=True) -> List[str]:\n xdg_config_home = (\n os.environ.get(\"XDG_CONFIG_HOME\") or Path(\"~/.config\").expanduser()\n )\n sources = [\n \"/etc/scrapy.cfg\",\n r\"c:\\scrapy\\scrapy.cfg\",\n str(Path(xdg_config_home) / \"scrapy.cfg\"),\n str(Path(\"~/.scrapy.cfg\").expanduser()),\n ]\n if use_closest:\n sources.append(closest_scrapy_cfg())\n return sources\n\n\ndef feed_complete_default_values_from_settings(feed, settings):\n out = feed.copy()\n out.setdefault(\"batch_item_count\", settings.getint(\"FEED_EXPORT_BATCH_ITEM_COUNT\"))\n out.setdefault(\"encoding\", settings[\"FEED_EXPORT_ENCODING\"])\n out.setdefault(\"fields\", settings.getdictorlist(\"FEED_EXPORT_FIELDS\") or None)\n out.setdefault(\"store_empty\", settings.getbool(\"FEED_STORE_EMPTY\"))\n out.setdefault(\"uri_params\", settings[\"FEED_URI_PARAMS\"])\n out.setdefault(\"item_export_kwargs\", {})\n if settings[\"FEED_EXPORT_INDENT\"] is None:\n out.setdefault(\"indent\", None)\n else:\n out.setdefault(\"indent\", settings.getint(\"FEED_EXPORT_INDENT\"))\n return out\n\n\ndef feed_process_params_from_cli(\n settings,\n output: List[str],\n output_format=None,\n overwrite_output: Optional[List[str]] = None,\n):\n \"\"\"\n Receives feed export params (from the 'crawl' or 'runspider' commands),\n checks for inconsistencies in their quantities and returns a dictionary\n suitable to be used as the FEEDS setting.\n \"\"\"\n valid_output_formats = without_none_values(\n settings.getwithbase(\"FEED_EXPORTERS\")\n ).keys()\n\n def check_valid_format(output_format):\n if output_format not in valid_output_formats:\n raise UsageError(\n f\"Unrecognized output format '{output_format}'. \"\n f\"Set a supported one ({tuple(valid_output_formats)}) \"\n \"after a colon at the end of the output URI (i.e. -o/-O \"\n \"<URI>:<FORMAT>) or as a file extension.\"\n )\n\n overwrite = False\n if overwrite_output:\n if output:\n raise UsageError(\n \"Please use only one of -o/--output and -O/--overwrite-output\"\n )\n if output_format:\n raise UsageError(\n \"-t/--output-format is a deprecated command line option\"\n \" and does not work in combination with -O/--overwrite-output.\"\n \" To specify a format please specify it after a colon at the end of the\"\n \" output URI (i.e. -O <URI>:<FORMAT>).\"\n \" Example working in the tutorial: \"\n \"scrapy crawl quotes -O quotes.json:json\"\n )\n output = overwrite_output\n overwrite = True\n\n if output_format:\n if len(output) == 1:\n check_valid_format(output_format)\n message = (\n \"The -t/--output-format command line option is deprecated in favor of \"\n \"specifying the output format within the output URI using the -o/--output or the\"\n \" -O/--overwrite-output option (i.e. -o/-O <URI>:<FORMAT>). See the documentation\"\n \" of the -o or -O option or the following examples for more information. \"\n \"Examples working in the tutorial: \"\n \"scrapy crawl quotes -o quotes.csv:csv or \"\n \"scrapy crawl quotes -O quotes.json:json\"\n )\n warnings.warn(message, ScrapyDeprecationWarning, stacklevel=2)\n return {output[0]: {\"format\": output_format}}\n raise UsageError(\n \"The -t command-line option cannot be used if multiple output \"\n \"URIs are specified\"\n )\n\n result: Dict[str, Dict[str, Any]] = {}\n for element in output:\n try:\n feed_uri, feed_format = element.rsplit(\":\", 1)\n check_valid_format(feed_format)\n except (ValueError, UsageError):\n feed_uri = element\n feed_format = Path(element).suffix.replace(\".\", \"\")\n else:\n if feed_uri == \"-\":\n feed_uri = \"stdout:\"\n check_valid_format(feed_format)\n result[feed_uri] = {\"format\": feed_format}\n if overwrite:\n result[feed_uri][\"overwrite\"] = True\n\n # FEEDS setting should take precedence over the matching CLI options\n result.update(settings.getdict(\"FEEDS\"))\n\n return result\n", "path": "scrapy/utils/conf.py"}], "after_files": [{"content": "import numbers\nimport os\nimport sys\nimport warnings\nfrom configparser import ConfigParser\nfrom operator import itemgetter\nfrom pathlib import Path\nfrom typing import Any, Dict, List, Optional, Union\n\nfrom scrapy.exceptions import ScrapyDeprecationWarning, UsageError\nfrom scrapy.settings import BaseSettings\nfrom scrapy.utils.deprecate import update_classpath\nfrom scrapy.utils.python import without_none_values\n\n\ndef build_component_list(compdict, custom=None, convert=update_classpath):\n \"\"\"Compose a component list from a { class: order } dictionary.\"\"\"\n\n def _check_components(complist):\n if len({convert(c) for c in complist}) != len(complist):\n raise ValueError(\n f\"Some paths in {complist!r} convert to the same object, \"\n \"please update your settings\"\n )\n\n def _map_keys(compdict):\n if isinstance(compdict, BaseSettings):\n compbs = BaseSettings()\n for k, v in compdict.items():\n prio = compdict.getpriority(k)\n assert prio is not None\n if compbs.getpriority(convert(k)) == prio:\n raise ValueError(\n f\"Some paths in {list(compdict.keys())!r} \"\n \"convert to the same \"\n \"object, please update your settings\"\n )\n else:\n compbs.set(convert(k), v, priority=prio)\n return compbs\n _check_components(compdict)\n return {convert(k): v for k, v in compdict.items()}\n\n def _validate_values(compdict):\n \"\"\"Fail if a value in the components dict is not a real number or None.\"\"\"\n for name, value in compdict.items():\n if value is not None and not isinstance(value, numbers.Real):\n raise ValueError(\n f\"Invalid value {value} for component {name}, \"\n \"please provide a real number or None instead\"\n )\n\n if custom is not None:\n warnings.warn(\n \"The 'custom' attribute of build_component_list() is deprecated. \"\n \"Please merge its value into 'compdict' manually or change your \"\n \"code to use Settings.getwithbase().\",\n category=ScrapyDeprecationWarning,\n stacklevel=2,\n )\n if isinstance(custom, (list, tuple)):\n _check_components(custom)\n return type(custom)(convert(c) for c in custom)\n compdict.update(custom)\n\n _validate_values(compdict)\n compdict = without_none_values(_map_keys(compdict))\n return [k for k, v in sorted(compdict.items(), key=itemgetter(1))]\n\n\ndef arglist_to_dict(arglist):\n \"\"\"Convert a list of arguments like ['arg1=val1', 'arg2=val2', ...] to a\n dict\n \"\"\"\n return dict(x.split(\"=\", 1) for x in arglist)\n\n\ndef closest_scrapy_cfg(\n path: Union[str, os.PathLike] = \".\",\n prevpath: Optional[Union[str, os.PathLike]] = None,\n) -> str:\n \"\"\"Return the path to the closest scrapy.cfg file by traversing the current\n directory and its parents\n \"\"\"\n if prevpath is not None and str(path) == str(prevpath):\n return \"\"\n path = Path(path).resolve()\n cfgfile = path / \"scrapy.cfg\"\n if cfgfile.exists():\n return str(cfgfile)\n return closest_scrapy_cfg(path.parent, path)\n\n\ndef init_env(project=\"default\", set_syspath=True):\n \"\"\"Initialize environment to use command-line tool from inside a project\n dir. This sets the Scrapy settings module and modifies the Python path to\n be able to locate the project module.\n \"\"\"\n cfg = get_config()\n if cfg.has_option(\"settings\", project):\n os.environ[\"SCRAPY_SETTINGS_MODULE\"] = cfg.get(\"settings\", project)\n closest = closest_scrapy_cfg()\n if closest:\n projdir = str(Path(closest).parent)\n if set_syspath and projdir not in sys.path:\n sys.path.append(projdir)\n\n\ndef get_config(use_closest=True):\n \"\"\"Get Scrapy config file as a ConfigParser\"\"\"\n sources = get_sources(use_closest)\n cfg = ConfigParser()\n cfg.read(sources)\n return cfg\n\n\ndef get_sources(use_closest=True) -> List[str]:\n xdg_config_home = (\n os.environ.get(\"XDG_CONFIG_HOME\") or Path(\"~/.config\").expanduser()\n )\n sources = [\n \"/etc/scrapy.cfg\",\n r\"c:\\scrapy\\scrapy.cfg\",\n str(Path(xdg_config_home) / \"scrapy.cfg\"),\n str(Path(\"~/.scrapy.cfg\").expanduser()),\n ]\n if use_closest:\n sources.append(closest_scrapy_cfg())\n return sources\n\n\ndef feed_complete_default_values_from_settings(feed, settings):\n out = feed.copy()\n out.setdefault(\"batch_item_count\", settings.getint(\"FEED_EXPORT_BATCH_ITEM_COUNT\"))\n out.setdefault(\"encoding\", settings[\"FEED_EXPORT_ENCODING\"])\n out.setdefault(\"fields\", settings.getdictorlist(\"FEED_EXPORT_FIELDS\") or None)\n out.setdefault(\"store_empty\", settings.getbool(\"FEED_STORE_EMPTY\"))\n out.setdefault(\"uri_params\", settings[\"FEED_URI_PARAMS\"])\n out.setdefault(\"item_export_kwargs\", {})\n if settings[\"FEED_EXPORT_INDENT\"] is None:\n out.setdefault(\"indent\", None)\n else:\n out.setdefault(\"indent\", settings.getint(\"FEED_EXPORT_INDENT\"))\n return out\n\n\ndef feed_process_params_from_cli(\n settings,\n output: List[str],\n output_format=None,\n overwrite_output: Optional[List[str]] = None,\n):\n \"\"\"\n Receives feed export params (from the 'crawl' or 'runspider' commands),\n checks for inconsistencies in their quantities and returns a dictionary\n suitable to be used as the FEEDS setting.\n \"\"\"\n valid_output_formats = without_none_values(\n settings.getwithbase(\"FEED_EXPORTERS\")\n ).keys()\n\n def check_valid_format(output_format):\n if output_format not in valid_output_formats:\n raise UsageError(\n f\"Unrecognized output format '{output_format}'. \"\n f\"Set a supported one ({tuple(valid_output_formats)}) \"\n \"after a colon at the end of the output URI (i.e. -o/-O \"\n \"<URI>:<FORMAT>) or as a file extension.\"\n )\n\n overwrite = False\n if overwrite_output:\n if output:\n raise UsageError(\n \"Please use only one of -o/--output and -O/--overwrite-output\"\n )\n if output_format:\n raise UsageError(\n \"-t/--output-format is a deprecated command line option\"\n \" and does not work in combination with -O/--overwrite-output.\"\n \" To specify a format please specify it after a colon at the end of the\"\n \" output URI (i.e. -O <URI>:<FORMAT>).\"\n \" Example working in the tutorial: \"\n \"scrapy crawl quotes -O quotes.json:json\"\n )\n output = overwrite_output\n overwrite = True\n\n if output_format:\n if len(output) == 1:\n check_valid_format(output_format)\n message = (\n \"The -t/--output-format command line option is deprecated in favor of \"\n \"specifying the output format within the output URI using the -o/--output or the\"\n \" -O/--overwrite-output option (i.e. -o/-O <URI>:<FORMAT>). See the documentation\"\n \" of the -o or -O option or the following examples for more information. \"\n \"Examples working in the tutorial: \"\n \"scrapy crawl quotes -o quotes.csv:csv or \"\n \"scrapy crawl quotes -O quotes.json:json\"\n )\n warnings.warn(message, ScrapyDeprecationWarning, stacklevel=2)\n return {output[0]: {\"format\": output_format}}\n raise UsageError(\n \"The -t command-line option cannot be used if multiple output \"\n \"URIs are specified\"\n )\n\n result: Dict[str, Dict[str, Any]] = {}\n for element in output:\n try:\n feed_uri, feed_format = element.rsplit(\":\", 1)\n check_valid_format(feed_format)\n except (ValueError, UsageError):\n feed_uri = element\n feed_format = Path(element).suffix.replace(\".\", \"\")\n else:\n if feed_uri == \"-\":\n feed_uri = \"stdout:\"\n check_valid_format(feed_format)\n result[feed_uri] = {\"format\": feed_format}\n if overwrite:\n result[feed_uri][\"overwrite\"] = True\n\n # FEEDS setting should take precedence over the matching CLI options\n result.update(settings.getdict(\"FEEDS\"))\n\n return result\n", "path": "scrapy/utils/conf.py"}]} | 2,737 | 218 |
gh_patches_debug_11084 | rasdani/github-patches | git_diff | python-discord__bot-852 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Disallow editing duration of expired infractions
Currently the command assumes that an infraction is still active if a duration is being edited. It tries to cancel the previous infraction but it will fail with a warning if the infraction already expired. Relevant code can be found here:
https://github.com/python-discord/bot/blob/582ddbb1ca8bab2cb883781911f5f35962330995/bot/cogs/moderation/management.py#L130-L142
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `bot/cogs/moderation/management.py`
Content:
```
1 import logging
2 import textwrap
3 import typing as t
4 from datetime import datetime
5
6 import discord
7 from discord.ext import commands
8 from discord.ext.commands import Context
9
10 from bot import constants
11 from bot.bot import Bot
12 from bot.converters import Expiry, InfractionSearchQuery, allowed_strings, proxy_user
13 from bot.pagination import LinePaginator
14 from bot.utils import time
15 from bot.utils.checks import in_channel_check, with_role_check
16 from . import utils
17 from .infractions import Infractions
18 from .modlog import ModLog
19
20 log = logging.getLogger(__name__)
21
22
23 class ModManagement(commands.Cog):
24 """Management of infractions."""
25
26 category = "Moderation"
27
28 def __init__(self, bot: Bot):
29 self.bot = bot
30
31 @property
32 def mod_log(self) -> ModLog:
33 """Get currently loaded ModLog cog instance."""
34 return self.bot.get_cog("ModLog")
35
36 @property
37 def infractions_cog(self) -> Infractions:
38 """Get currently loaded Infractions cog instance."""
39 return self.bot.get_cog("Infractions")
40
41 # region: Edit infraction commands
42
43 @commands.group(name='infraction', aliases=('infr', 'infractions', 'inf'), invoke_without_command=True)
44 async def infraction_group(self, ctx: Context) -> None:
45 """Infraction manipulation commands."""
46 await ctx.invoke(self.bot.get_command("help"), "infraction")
47
48 @infraction_group.command(name='edit')
49 async def infraction_edit(
50 self,
51 ctx: Context,
52 infraction_id: t.Union[int, allowed_strings("l", "last", "recent")],
53 duration: t.Union[Expiry, allowed_strings("p", "permanent"), None],
54 *,
55 reason: str = None
56 ) -> None:
57 """
58 Edit the duration and/or the reason of an infraction.
59
60 Durations are relative to the time of updating and should be appended with a unit of time.
61 Units (∗case-sensitive):
62 \u2003`y` - years
63 \u2003`m` - months∗
64 \u2003`w` - weeks
65 \u2003`d` - days
66 \u2003`h` - hours
67 \u2003`M` - minutes∗
68 \u2003`s` - seconds
69
70 Use "l", "last", or "recent" as the infraction ID to specify that the most recent infraction
71 authored by the command invoker should be edited.
72
73 Use "p" or "permanent" to mark the infraction as permanent. Alternatively, an ISO 8601
74 timestamp can be provided for the duration.
75 """
76 if duration is None and reason is None:
77 # Unlike UserInputError, the error handler will show a specified message for BadArgument
78 raise commands.BadArgument("Neither a new expiry nor a new reason was specified.")
79
80 # Retrieve the previous infraction for its information.
81 if isinstance(infraction_id, str):
82 params = {
83 "actor__id": ctx.author.id,
84 "ordering": "-inserted_at"
85 }
86 infractions = await self.bot.api_client.get(f"bot/infractions", params=params)
87
88 if infractions:
89 old_infraction = infractions[0]
90 infraction_id = old_infraction["id"]
91 else:
92 await ctx.send(
93 f":x: Couldn't find most recent infraction; you have never given an infraction."
94 )
95 return
96 else:
97 old_infraction = await self.bot.api_client.get(f"bot/infractions/{infraction_id}")
98
99 request_data = {}
100 confirm_messages = []
101 log_text = ""
102
103 if isinstance(duration, str):
104 request_data['expires_at'] = None
105 confirm_messages.append("marked as permanent")
106 elif duration is not None:
107 request_data['expires_at'] = duration.isoformat()
108 expiry = time.format_infraction_with_duration(request_data['expires_at'])
109 confirm_messages.append(f"set to expire on {expiry}")
110 else:
111 confirm_messages.append("expiry unchanged")
112
113 if reason:
114 request_data['reason'] = reason
115 confirm_messages.append("set a new reason")
116 log_text += f"""
117 Previous reason: {old_infraction['reason']}
118 New reason: {reason}
119 """.rstrip()
120 else:
121 confirm_messages.append("reason unchanged")
122
123 # Update the infraction
124 new_infraction = await self.bot.api_client.patch(
125 f'bot/infractions/{infraction_id}',
126 json=request_data,
127 )
128
129 # Re-schedule infraction if the expiration has been updated
130 if 'expires_at' in request_data:
131 # A scheduled task should only exist if the old infraction wasn't permanent
132 if old_infraction['expires_at']:
133 self.infractions_cog.cancel_task(new_infraction['id'])
134
135 # If the infraction was not marked as permanent, schedule a new expiration task
136 if request_data['expires_at']:
137 self.infractions_cog.schedule_task(new_infraction['id'], new_infraction)
138
139 log_text += f"""
140 Previous expiry: {old_infraction['expires_at'] or "Permanent"}
141 New expiry: {new_infraction['expires_at'] or "Permanent"}
142 """.rstrip()
143
144 changes = ' & '.join(confirm_messages)
145 await ctx.send(f":ok_hand: Updated infraction #{infraction_id}: {changes}")
146
147 # Get information about the infraction's user
148 user_id = new_infraction['user']
149 user = ctx.guild.get_member(user_id)
150
151 if user:
152 user_text = f"{user.mention} (`{user.id}`)"
153 thumbnail = user.avatar_url_as(static_format="png")
154 else:
155 user_text = f"`{user_id}`"
156 thumbnail = None
157
158 # The infraction's actor
159 actor_id = new_infraction['actor']
160 actor = ctx.guild.get_member(actor_id) or f"`{actor_id}`"
161
162 await self.mod_log.send_log_message(
163 icon_url=constants.Icons.pencil,
164 colour=discord.Colour.blurple(),
165 title="Infraction edited",
166 thumbnail=thumbnail,
167 text=textwrap.dedent(f"""
168 Member: {user_text}
169 Actor: {actor}
170 Edited by: {ctx.message.author}{log_text}
171 """)
172 )
173
174 # endregion
175 # region: Search infractions
176
177 @infraction_group.group(name="search", invoke_without_command=True)
178 async def infraction_search_group(self, ctx: Context, query: InfractionSearchQuery) -> None:
179 """Searches for infractions in the database."""
180 if isinstance(query, discord.User):
181 await ctx.invoke(self.search_user, query)
182 else:
183 await ctx.invoke(self.search_reason, query)
184
185 @infraction_search_group.command(name="user", aliases=("member", "id"))
186 async def search_user(self, ctx: Context, user: t.Union[discord.User, proxy_user]) -> None:
187 """Search for infractions by member."""
188 infraction_list = await self.bot.api_client.get(
189 'bot/infractions',
190 params={'user__id': str(user.id)}
191 )
192 embed = discord.Embed(
193 title=f"Infractions for {user} ({len(infraction_list)} total)",
194 colour=discord.Colour.orange()
195 )
196 await self.send_infraction_list(ctx, embed, infraction_list)
197
198 @infraction_search_group.command(name="reason", aliases=("match", "regex", "re"))
199 async def search_reason(self, ctx: Context, reason: str) -> None:
200 """Search for infractions by their reason. Use Re2 for matching."""
201 infraction_list = await self.bot.api_client.get(
202 'bot/infractions',
203 params={'search': reason}
204 )
205 embed = discord.Embed(
206 title=f"Infractions matching `{reason}` ({len(infraction_list)} total)",
207 colour=discord.Colour.orange()
208 )
209 await self.send_infraction_list(ctx, embed, infraction_list)
210
211 # endregion
212 # region: Utility functions
213
214 async def send_infraction_list(
215 self,
216 ctx: Context,
217 embed: discord.Embed,
218 infractions: t.Iterable[utils.Infraction]
219 ) -> None:
220 """Send a paginated embed of infractions for the specified user."""
221 if not infractions:
222 await ctx.send(f":warning: No infractions could be found for that query.")
223 return
224
225 lines = tuple(
226 self.infraction_to_string(infraction)
227 for infraction in infractions
228 )
229
230 await LinePaginator.paginate(
231 lines,
232 ctx=ctx,
233 embed=embed,
234 empty=True,
235 max_lines=3,
236 max_size=1000
237 )
238
239 def infraction_to_string(self, infraction: utils.Infraction) -> str:
240 """Convert the infraction object to a string representation."""
241 actor_id = infraction["actor"]
242 guild = self.bot.get_guild(constants.Guild.id)
243 actor = guild.get_member(actor_id)
244 active = infraction["active"]
245 user_id = infraction["user"]
246 hidden = infraction["hidden"]
247 created = time.format_infraction(infraction["inserted_at"])
248
249 if active:
250 remaining = time.until_expiration(infraction["expires_at"]) or "Expired"
251 else:
252 remaining = "Inactive"
253
254 if infraction["expires_at"] is None:
255 expires = "*Permanent*"
256 else:
257 date_from = datetime.strptime(created, time.INFRACTION_FORMAT)
258 expires = time.format_infraction_with_duration(infraction["expires_at"], date_from)
259
260 lines = textwrap.dedent(f"""
261 {"**===============**" if active else "==============="}
262 Status: {"__**Active**__" if active else "Inactive"}
263 User: {self.bot.get_user(user_id)} (`{user_id}`)
264 Type: **{infraction["type"]}**
265 Shadow: {hidden}
266 Reason: {infraction["reason"] or "*None*"}
267 Created: {created}
268 Expires: {expires}
269 Remaining: {remaining}
270 Actor: {actor.mention if actor else actor_id}
271 ID: `{infraction["id"]}`
272 {"**===============**" if active else "==============="}
273 """)
274
275 return lines.strip()
276
277 # endregion
278
279 # This cannot be static (must have a __func__ attribute).
280 def cog_check(self, ctx: Context) -> bool:
281 """Only allow moderators from moderator channels to invoke the commands in this cog."""
282 checks = [
283 with_role_check(ctx, *constants.MODERATION_ROLES),
284 in_channel_check(ctx, *constants.MODERATION_CHANNELS)
285 ]
286 return all(checks)
287
288 # This cannot be static (must have a __func__ attribute).
289 async def cog_command_error(self, ctx: Context, error: Exception) -> None:
290 """Send a notification to the invoking context on a Union failure."""
291 if isinstance(error, commands.BadUnionArgument):
292 if discord.User in error.converters:
293 await ctx.send(str(error.errors[0]))
294 error.handled = True
295
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/bot/cogs/moderation/management.py b/bot/cogs/moderation/management.py
--- a/bot/cogs/moderation/management.py
+++ b/bot/cogs/moderation/management.py
@@ -100,7 +100,12 @@
confirm_messages = []
log_text = ""
- if isinstance(duration, str):
+ if duration is not None and not old_infraction['active']:
+ if reason is None:
+ await ctx.send(":x: Cannot edit the expiration of an expired infraction.")
+ return
+ confirm_messages.append("expiry unchanged (infraction already expired)")
+ elif isinstance(duration, str):
request_data['expires_at'] = None
confirm_messages.append("marked as permanent")
elif duration is not None:
| {"golden_diff": "diff --git a/bot/cogs/moderation/management.py b/bot/cogs/moderation/management.py\n--- a/bot/cogs/moderation/management.py\n+++ b/bot/cogs/moderation/management.py\n@@ -100,7 +100,12 @@\n confirm_messages = []\n log_text = \"\"\n \n- if isinstance(duration, str):\n+ if duration is not None and not old_infraction['active']:\n+ if reason is None:\n+ await ctx.send(\":x: Cannot edit the expiration of an expired infraction.\")\n+ return\n+ confirm_messages.append(\"expiry unchanged (infraction already expired)\")\n+ elif isinstance(duration, str):\n request_data['expires_at'] = None\n confirm_messages.append(\"marked as permanent\")\n elif duration is not None:\n", "issue": "Disallow editing duration of expired infractions\nCurrently the command assumes that an infraction is still active if a duration is being edited. It tries to cancel the previous infraction but it will fail with a warning if the infraction already expired. Relevant code can be found here:\r\nhttps://github.com/python-discord/bot/blob/582ddbb1ca8bab2cb883781911f5f35962330995/bot/cogs/moderation/management.py#L130-L142\n", "before_files": [{"content": "import logging\nimport textwrap\nimport typing as t\nfrom datetime import datetime\n\nimport discord\nfrom discord.ext import commands\nfrom discord.ext.commands import Context\n\nfrom bot import constants\nfrom bot.bot import Bot\nfrom bot.converters import Expiry, InfractionSearchQuery, allowed_strings, proxy_user\nfrom bot.pagination import LinePaginator\nfrom bot.utils import time\nfrom bot.utils.checks import in_channel_check, with_role_check\nfrom . import utils\nfrom .infractions import Infractions\nfrom .modlog import ModLog\n\nlog = logging.getLogger(__name__)\n\n\nclass ModManagement(commands.Cog):\n \"\"\"Management of infractions.\"\"\"\n\n category = \"Moderation\"\n\n def __init__(self, bot: Bot):\n self.bot = bot\n\n @property\n def mod_log(self) -> ModLog:\n \"\"\"Get currently loaded ModLog cog instance.\"\"\"\n return self.bot.get_cog(\"ModLog\")\n\n @property\n def infractions_cog(self) -> Infractions:\n \"\"\"Get currently loaded Infractions cog instance.\"\"\"\n return self.bot.get_cog(\"Infractions\")\n\n # region: Edit infraction commands\n\n @commands.group(name='infraction', aliases=('infr', 'infractions', 'inf'), invoke_without_command=True)\n async def infraction_group(self, ctx: Context) -> None:\n \"\"\"Infraction manipulation commands.\"\"\"\n await ctx.invoke(self.bot.get_command(\"help\"), \"infraction\")\n\n @infraction_group.command(name='edit')\n async def infraction_edit(\n self,\n ctx: Context,\n infraction_id: t.Union[int, allowed_strings(\"l\", \"last\", \"recent\")],\n duration: t.Union[Expiry, allowed_strings(\"p\", \"permanent\"), None],\n *,\n reason: str = None\n ) -> None:\n \"\"\"\n Edit the duration and/or the reason of an infraction.\n\n Durations are relative to the time of updating and should be appended with a unit of time.\n Units (\u2217case-sensitive):\n \\u2003`y` - years\n \\u2003`m` - months\u2217\n \\u2003`w` - weeks\n \\u2003`d` - days\n \\u2003`h` - hours\n \\u2003`M` - minutes\u2217\n \\u2003`s` - seconds\n\n Use \"l\", \"last\", or \"recent\" as the infraction ID to specify that the most recent infraction\n authored by the command invoker should be edited.\n\n Use \"p\" or \"permanent\" to mark the infraction as permanent. Alternatively, an ISO 8601\n timestamp can be provided for the duration.\n \"\"\"\n if duration is None and reason is None:\n # Unlike UserInputError, the error handler will show a specified message for BadArgument\n raise commands.BadArgument(\"Neither a new expiry nor a new reason was specified.\")\n\n # Retrieve the previous infraction for its information.\n if isinstance(infraction_id, str):\n params = {\n \"actor__id\": ctx.author.id,\n \"ordering\": \"-inserted_at\"\n }\n infractions = await self.bot.api_client.get(f\"bot/infractions\", params=params)\n\n if infractions:\n old_infraction = infractions[0]\n infraction_id = old_infraction[\"id\"]\n else:\n await ctx.send(\n f\":x: Couldn't find most recent infraction; you have never given an infraction.\"\n )\n return\n else:\n old_infraction = await self.bot.api_client.get(f\"bot/infractions/{infraction_id}\")\n\n request_data = {}\n confirm_messages = []\n log_text = \"\"\n\n if isinstance(duration, str):\n request_data['expires_at'] = None\n confirm_messages.append(\"marked as permanent\")\n elif duration is not None:\n request_data['expires_at'] = duration.isoformat()\n expiry = time.format_infraction_with_duration(request_data['expires_at'])\n confirm_messages.append(f\"set to expire on {expiry}\")\n else:\n confirm_messages.append(\"expiry unchanged\")\n\n if reason:\n request_data['reason'] = reason\n confirm_messages.append(\"set a new reason\")\n log_text += f\"\"\"\n Previous reason: {old_infraction['reason']}\n New reason: {reason}\n \"\"\".rstrip()\n else:\n confirm_messages.append(\"reason unchanged\")\n\n # Update the infraction\n new_infraction = await self.bot.api_client.patch(\n f'bot/infractions/{infraction_id}',\n json=request_data,\n )\n\n # Re-schedule infraction if the expiration has been updated\n if 'expires_at' in request_data:\n # A scheduled task should only exist if the old infraction wasn't permanent\n if old_infraction['expires_at']:\n self.infractions_cog.cancel_task(new_infraction['id'])\n\n # If the infraction was not marked as permanent, schedule a new expiration task\n if request_data['expires_at']:\n self.infractions_cog.schedule_task(new_infraction['id'], new_infraction)\n\n log_text += f\"\"\"\n Previous expiry: {old_infraction['expires_at'] or \"Permanent\"}\n New expiry: {new_infraction['expires_at'] or \"Permanent\"}\n \"\"\".rstrip()\n\n changes = ' & '.join(confirm_messages)\n await ctx.send(f\":ok_hand: Updated infraction #{infraction_id}: {changes}\")\n\n # Get information about the infraction's user\n user_id = new_infraction['user']\n user = ctx.guild.get_member(user_id)\n\n if user:\n user_text = f\"{user.mention} (`{user.id}`)\"\n thumbnail = user.avatar_url_as(static_format=\"png\")\n else:\n user_text = f\"`{user_id}`\"\n thumbnail = None\n\n # The infraction's actor\n actor_id = new_infraction['actor']\n actor = ctx.guild.get_member(actor_id) or f\"`{actor_id}`\"\n\n await self.mod_log.send_log_message(\n icon_url=constants.Icons.pencil,\n colour=discord.Colour.blurple(),\n title=\"Infraction edited\",\n thumbnail=thumbnail,\n text=textwrap.dedent(f\"\"\"\n Member: {user_text}\n Actor: {actor}\n Edited by: {ctx.message.author}{log_text}\n \"\"\")\n )\n\n # endregion\n # region: Search infractions\n\n @infraction_group.group(name=\"search\", invoke_without_command=True)\n async def infraction_search_group(self, ctx: Context, query: InfractionSearchQuery) -> None:\n \"\"\"Searches for infractions in the database.\"\"\"\n if isinstance(query, discord.User):\n await ctx.invoke(self.search_user, query)\n else:\n await ctx.invoke(self.search_reason, query)\n\n @infraction_search_group.command(name=\"user\", aliases=(\"member\", \"id\"))\n async def search_user(self, ctx: Context, user: t.Union[discord.User, proxy_user]) -> None:\n \"\"\"Search for infractions by member.\"\"\"\n infraction_list = await self.bot.api_client.get(\n 'bot/infractions',\n params={'user__id': str(user.id)}\n )\n embed = discord.Embed(\n title=f\"Infractions for {user} ({len(infraction_list)} total)\",\n colour=discord.Colour.orange()\n )\n await self.send_infraction_list(ctx, embed, infraction_list)\n\n @infraction_search_group.command(name=\"reason\", aliases=(\"match\", \"regex\", \"re\"))\n async def search_reason(self, ctx: Context, reason: str) -> None:\n \"\"\"Search for infractions by their reason. Use Re2 for matching.\"\"\"\n infraction_list = await self.bot.api_client.get(\n 'bot/infractions',\n params={'search': reason}\n )\n embed = discord.Embed(\n title=f\"Infractions matching `{reason}` ({len(infraction_list)} total)\",\n colour=discord.Colour.orange()\n )\n await self.send_infraction_list(ctx, embed, infraction_list)\n\n # endregion\n # region: Utility functions\n\n async def send_infraction_list(\n self,\n ctx: Context,\n embed: discord.Embed,\n infractions: t.Iterable[utils.Infraction]\n ) -> None:\n \"\"\"Send a paginated embed of infractions for the specified user.\"\"\"\n if not infractions:\n await ctx.send(f\":warning: No infractions could be found for that query.\")\n return\n\n lines = tuple(\n self.infraction_to_string(infraction)\n for infraction in infractions\n )\n\n await LinePaginator.paginate(\n lines,\n ctx=ctx,\n embed=embed,\n empty=True,\n max_lines=3,\n max_size=1000\n )\n\n def infraction_to_string(self, infraction: utils.Infraction) -> str:\n \"\"\"Convert the infraction object to a string representation.\"\"\"\n actor_id = infraction[\"actor\"]\n guild = self.bot.get_guild(constants.Guild.id)\n actor = guild.get_member(actor_id)\n active = infraction[\"active\"]\n user_id = infraction[\"user\"]\n hidden = infraction[\"hidden\"]\n created = time.format_infraction(infraction[\"inserted_at\"])\n\n if active:\n remaining = time.until_expiration(infraction[\"expires_at\"]) or \"Expired\"\n else:\n remaining = \"Inactive\"\n\n if infraction[\"expires_at\"] is None:\n expires = \"*Permanent*\"\n else:\n date_from = datetime.strptime(created, time.INFRACTION_FORMAT)\n expires = time.format_infraction_with_duration(infraction[\"expires_at\"], date_from)\n\n lines = textwrap.dedent(f\"\"\"\n {\"**===============**\" if active else \"===============\"}\n Status: {\"__**Active**__\" if active else \"Inactive\"}\n User: {self.bot.get_user(user_id)} (`{user_id}`)\n Type: **{infraction[\"type\"]}**\n Shadow: {hidden}\n Reason: {infraction[\"reason\"] or \"*None*\"}\n Created: {created}\n Expires: {expires}\n Remaining: {remaining}\n Actor: {actor.mention if actor else actor_id}\n ID: `{infraction[\"id\"]}`\n {\"**===============**\" if active else \"===============\"}\n \"\"\")\n\n return lines.strip()\n\n # endregion\n\n # This cannot be static (must have a __func__ attribute).\n def cog_check(self, ctx: Context) -> bool:\n \"\"\"Only allow moderators from moderator channels to invoke the commands in this cog.\"\"\"\n checks = [\n with_role_check(ctx, *constants.MODERATION_ROLES),\n in_channel_check(ctx, *constants.MODERATION_CHANNELS)\n ]\n return all(checks)\n\n # This cannot be static (must have a __func__ attribute).\n async def cog_command_error(self, ctx: Context, error: Exception) -> None:\n \"\"\"Send a notification to the invoking context on a Union failure.\"\"\"\n if isinstance(error, commands.BadUnionArgument):\n if discord.User in error.converters:\n await ctx.send(str(error.errors[0]))\n error.handled = True\n", "path": "bot/cogs/moderation/management.py"}], "after_files": [{"content": "import logging\nimport textwrap\nimport typing as t\nfrom datetime import datetime\n\nimport discord\nfrom discord.ext import commands\nfrom discord.ext.commands import Context\n\nfrom bot import constants\nfrom bot.bot import Bot\nfrom bot.converters import Expiry, InfractionSearchQuery, allowed_strings, proxy_user\nfrom bot.pagination import LinePaginator\nfrom bot.utils import time\nfrom bot.utils.checks import in_channel_check, with_role_check\nfrom . import utils\nfrom .infractions import Infractions\nfrom .modlog import ModLog\n\nlog = logging.getLogger(__name__)\n\n\nclass ModManagement(commands.Cog):\n \"\"\"Management of infractions.\"\"\"\n\n category = \"Moderation\"\n\n def __init__(self, bot: Bot):\n self.bot = bot\n\n @property\n def mod_log(self) -> ModLog:\n \"\"\"Get currently loaded ModLog cog instance.\"\"\"\n return self.bot.get_cog(\"ModLog\")\n\n @property\n def infractions_cog(self) -> Infractions:\n \"\"\"Get currently loaded Infractions cog instance.\"\"\"\n return self.bot.get_cog(\"Infractions\")\n\n # region: Edit infraction commands\n\n @commands.group(name='infraction', aliases=('infr', 'infractions', 'inf'), invoke_without_command=True)\n async def infraction_group(self, ctx: Context) -> None:\n \"\"\"Infraction manipulation commands.\"\"\"\n await ctx.invoke(self.bot.get_command(\"help\"), \"infraction\")\n\n @infraction_group.command(name='edit')\n async def infraction_edit(\n self,\n ctx: Context,\n infraction_id: t.Union[int, allowed_strings(\"l\", \"last\", \"recent\")],\n duration: t.Union[Expiry, allowed_strings(\"p\", \"permanent\"), None],\n *,\n reason: str = None\n ) -> None:\n \"\"\"\n Edit the duration and/or the reason of an infraction.\n\n Durations are relative to the time of updating and should be appended with a unit of time.\n Units (\u2217case-sensitive):\n \\u2003`y` - years\n \\u2003`m` - months\u2217\n \\u2003`w` - weeks\n \\u2003`d` - days\n \\u2003`h` - hours\n \\u2003`M` - minutes\u2217\n \\u2003`s` - seconds\n\n Use \"l\", \"last\", or \"recent\" as the infraction ID to specify that the most recent infraction\n authored by the command invoker should be edited.\n\n Use \"p\" or \"permanent\" to mark the infraction as permanent. Alternatively, an ISO 8601\n timestamp can be provided for the duration.\n \"\"\"\n if duration is None and reason is None:\n # Unlike UserInputError, the error handler will show a specified message for BadArgument\n raise commands.BadArgument(\"Neither a new expiry nor a new reason was specified.\")\n\n # Retrieve the previous infraction for its information.\n if isinstance(infraction_id, str):\n params = {\n \"actor__id\": ctx.author.id,\n \"ordering\": \"-inserted_at\"\n }\n infractions = await self.bot.api_client.get(f\"bot/infractions\", params=params)\n\n if infractions:\n old_infraction = infractions[0]\n infraction_id = old_infraction[\"id\"]\n else:\n await ctx.send(\n f\":x: Couldn't find most recent infraction; you have never given an infraction.\"\n )\n return\n else:\n old_infraction = await self.bot.api_client.get(f\"bot/infractions/{infraction_id}\")\n\n request_data = {}\n confirm_messages = []\n log_text = \"\"\n\n if duration is not None and not old_infraction['active']:\n if reason is None:\n await ctx.send(\":x: Cannot edit the expiration of an expired infraction.\")\n return\n confirm_messages.append(\"expiry unchanged (infraction already expired)\")\n elif isinstance(duration, str):\n request_data['expires_at'] = None\n confirm_messages.append(\"marked as permanent\")\n elif duration is not None:\n request_data['expires_at'] = duration.isoformat()\n expiry = time.format_infraction_with_duration(request_data['expires_at'])\n confirm_messages.append(f\"set to expire on {expiry}\")\n else:\n confirm_messages.append(\"expiry unchanged\")\n\n if reason:\n request_data['reason'] = reason\n confirm_messages.append(\"set a new reason\")\n log_text += f\"\"\"\n Previous reason: {old_infraction['reason']}\n New reason: {reason}\n \"\"\".rstrip()\n else:\n confirm_messages.append(\"reason unchanged\")\n\n # Update the infraction\n new_infraction = await self.bot.api_client.patch(\n f'bot/infractions/{infraction_id}',\n json=request_data,\n )\n\n # Re-schedule infraction if the expiration has been updated\n if 'expires_at' in request_data:\n # A scheduled task should only exist if the old infraction wasn't permanent\n if old_infraction['expires_at']:\n self.infractions_cog.cancel_task(new_infraction['id'])\n\n # If the infraction was not marked as permanent, schedule a new expiration task\n if request_data['expires_at']:\n self.infractions_cog.schedule_task(new_infraction['id'], new_infraction)\n\n log_text += f\"\"\"\n Previous expiry: {old_infraction['expires_at'] or \"Permanent\"}\n New expiry: {new_infraction['expires_at'] or \"Permanent\"}\n \"\"\".rstrip()\n\n changes = ' & '.join(confirm_messages)\n await ctx.send(f\":ok_hand: Updated infraction #{infraction_id}: {changes}\")\n\n # Get information about the infraction's user\n user_id = new_infraction['user']\n user = ctx.guild.get_member(user_id)\n\n if user:\n user_text = f\"{user.mention} (`{user.id}`)\"\n thumbnail = user.avatar_url_as(static_format=\"png\")\n else:\n user_text = f\"`{user_id}`\"\n thumbnail = None\n\n # The infraction's actor\n actor_id = new_infraction['actor']\n actor = ctx.guild.get_member(actor_id) or f\"`{actor_id}`\"\n\n await self.mod_log.send_log_message(\n icon_url=constants.Icons.pencil,\n colour=discord.Colour.blurple(),\n title=\"Infraction edited\",\n thumbnail=thumbnail,\n text=textwrap.dedent(f\"\"\"\n Member: {user_text}\n Actor: {actor}\n Edited by: {ctx.message.author}{log_text}\n \"\"\")\n )\n\n # endregion\n # region: Search infractions\n\n @infraction_group.group(name=\"search\", invoke_without_command=True)\n async def infraction_search_group(self, ctx: Context, query: InfractionSearchQuery) -> None:\n \"\"\"Searches for infractions in the database.\"\"\"\n if isinstance(query, discord.User):\n await ctx.invoke(self.search_user, query)\n else:\n await ctx.invoke(self.search_reason, query)\n\n @infraction_search_group.command(name=\"user\", aliases=(\"member\", \"id\"))\n async def search_user(self, ctx: Context, user: t.Union[discord.User, proxy_user]) -> None:\n \"\"\"Search for infractions by member.\"\"\"\n infraction_list = await self.bot.api_client.get(\n 'bot/infractions',\n params={'user__id': str(user.id)}\n )\n embed = discord.Embed(\n title=f\"Infractions for {user} ({len(infraction_list)} total)\",\n colour=discord.Colour.orange()\n )\n await self.send_infraction_list(ctx, embed, infraction_list)\n\n @infraction_search_group.command(name=\"reason\", aliases=(\"match\", \"regex\", \"re\"))\n async def search_reason(self, ctx: Context, reason: str) -> None:\n \"\"\"Search for infractions by their reason. Use Re2 for matching.\"\"\"\n infraction_list = await self.bot.api_client.get(\n 'bot/infractions',\n params={'search': reason}\n )\n embed = discord.Embed(\n title=f\"Infractions matching `{reason}` ({len(infraction_list)} total)\",\n colour=discord.Colour.orange()\n )\n await self.send_infraction_list(ctx, embed, infraction_list)\n\n # endregion\n # region: Utility functions\n\n async def send_infraction_list(\n self,\n ctx: Context,\n embed: discord.Embed,\n infractions: t.Iterable[utils.Infraction]\n ) -> None:\n \"\"\"Send a paginated embed of infractions for the specified user.\"\"\"\n if not infractions:\n await ctx.send(f\":warning: No infractions could be found for that query.\")\n return\n\n lines = tuple(\n self.infraction_to_string(infraction)\n for infraction in infractions\n )\n\n await LinePaginator.paginate(\n lines,\n ctx=ctx,\n embed=embed,\n empty=True,\n max_lines=3,\n max_size=1000\n )\n\n def infraction_to_string(self, infraction: utils.Infraction) -> str:\n \"\"\"Convert the infraction object to a string representation.\"\"\"\n actor_id = infraction[\"actor\"]\n guild = self.bot.get_guild(constants.Guild.id)\n actor = guild.get_member(actor_id)\n active = infraction[\"active\"]\n user_id = infraction[\"user\"]\n hidden = infraction[\"hidden\"]\n created = time.format_infraction(infraction[\"inserted_at\"])\n\n if active:\n remaining = time.until_expiration(infraction[\"expires_at\"]) or \"Expired\"\n else:\n remaining = \"Inactive\"\n\n if infraction[\"expires_at\"] is None:\n expires = \"*Permanent*\"\n else:\n date_from = datetime.strptime(created, time.INFRACTION_FORMAT)\n expires = time.format_infraction_with_duration(infraction[\"expires_at\"], date_from)\n\n lines = textwrap.dedent(f\"\"\"\n {\"**===============**\" if active else \"===============\"}\n Status: {\"__**Active**__\" if active else \"Inactive\"}\n User: {self.bot.get_user(user_id)} (`{user_id}`)\n Type: **{infraction[\"type\"]}**\n Shadow: {hidden}\n Reason: {infraction[\"reason\"] or \"*None*\"}\n Created: {created}\n Expires: {expires}\n Remaining: {remaining}\n Actor: {actor.mention if actor else actor_id}\n ID: `{infraction[\"id\"]}`\n {\"**===============**\" if active else \"===============\"}\n \"\"\")\n\n return lines.strip()\n\n # endregion\n\n # This cannot be static (must have a __func__ attribute).\n def cog_check(self, ctx: Context) -> bool:\n \"\"\"Only allow moderators from moderator channels to invoke the commands in this cog.\"\"\"\n checks = [\n with_role_check(ctx, *constants.MODERATION_ROLES),\n in_channel_check(ctx, *constants.MODERATION_CHANNELS)\n ]\n return all(checks)\n\n # This cannot be static (must have a __func__ attribute).\n async def cog_command_error(self, ctx: Context, error: Exception) -> None:\n \"\"\"Send a notification to the invoking context on a Union failure.\"\"\"\n if isinstance(error, commands.BadUnionArgument):\n if discord.User in error.converters:\n await ctx.send(str(error.errors[0]))\n error.handled = True\n", "path": "bot/cogs/moderation/management.py"}]} | 3,591 | 174 |
gh_patches_debug_5955 | rasdani/github-patches | git_diff | vispy__vispy-1383 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Wrong install location for docs
From `setup.py`:
```
setup(
[...]
package_data={
'vispy': [op.join('io', '_data', '*'),
op.join('html', 'static', 'js', '*'),
op.join('app', 'tests', 'qt-designer.ui'),
op.join('..', 'doc', '*'),
],
```
This line `op.join('..', 'doc', '*')` is wrong for a system-wide install. It leads to the documentation being install under `dist-packages` or `site-packages`, which is definitely non-standard. IMO, the best would be to just not install the docs yourself, and let the package build system (conda or Debian) handle it.
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `setup.py`
Content:
```
1 # -*- coding: utf-8 -*-
2 # Copyright (c) Vispy Development Team. All Rights Reserved.
3 # Distributed under the (new) BSD License. See LICENSE.txt for more info.
4
5 """ Vispy setup script.
6
7 Steps to do a new release:
8
9 Preparations:
10 * Test on Windows, Linux, Mac
11 * Make release notes
12 * Update API documentation and other docs that need updating.
13 * Install 'twine' package for uploading to PyPI
14
15 Define the version:
16 * update __version__ in __init__.py
17 * tag the tip changeset as version x.x.x; `git tag -a 'vX.Y.Z'`
18
19 Test installation:
20 * clear the build and dist dir (if they exist)
21 * python setup.py sdist
22 * twine register --repository-url https://test.pypi.org/legacy/ dist/*
23 * twine upload --repository-url https://test.pypi.org/legacy/ dist/*
24 * pip install -i https://testpypi.python.org/pypi vispy
25
26 Generate and upload package
27 * python setup.py sdist
28 * twine register dist/*
29 * twine upload dist/*
30
31 Announcing:
32 * It can be worth waiting a day for eager users to report critical bugs
33 * Announce in scipy-user, vispy mailing list, G+
34
35 """
36
37 import os
38 from os import path as op
39 from warnings import warn
40
41 try:
42 # use setuptools namespace, allows for "develop"
43 import setuptools # noqa, analysis:ignore
44 except ImportError:
45 warn("unable to load setuptools. 'setup.py develop' will not work")
46 pass # it's not essential for installation
47 from distutils.core import setup
48
49 name = 'vispy'
50 description = 'Interactive visualization in Python'
51
52
53 # Get version and docstring
54 __version__ = None
55 __doc__ = ''
56 docStatus = 0 # Not started, in progress, done
57 initFile = os.path.join(os.path.dirname(__file__), 'vispy', '__init__.py')
58 for line in open(initFile).readlines():
59 if (line.startswith('version_info') or line.startswith('__version__')):
60 exec(line.strip())
61 elif line.startswith('"""'):
62 if docStatus == 0:
63 docStatus = 1
64 line = line.lstrip('"')
65 elif docStatus == 1:
66 docStatus = 2
67 if docStatus == 1:
68 __doc__ += line
69
70
71 def package_tree(pkgroot):
72 path = os.path.dirname(__file__)
73 subdirs = [os.path.relpath(i[0], path).replace(os.path.sep, '.')
74 for i in os.walk(os.path.join(path, pkgroot))
75 if '__init__.py' in i[2]]
76 return subdirs
77
78
79 setup(
80 name=name,
81 version=__version__,
82 author='Vispy contributors',
83 author_email='[email protected]',
84 license='(new) BSD',
85 url='http://vispy.org',
86 download_url='https://pypi.python.org/pypi/vispy',
87 keywords="visualization OpenGl ES medical imaging 3D plotting "
88 "numpy bigdata",
89 description=description,
90 long_description=__doc__,
91 platforms='any',
92 provides=['vispy'],
93 install_requires=['numpy'],
94 extras_require={
95 'ipython-static': ['ipython'],
96 'ipython-vnc': ['ipython>=2'],
97 'ipython-webgl': ['ipython>=2', 'tornado'],
98 'pyglet': ['pyglet>=1.2'],
99 # 'pyqt4': [], # Why is this on PyPI, but without downloads?
100 # 'pyqt5': [], # Ditto.
101 'pyside': ['PySide'],
102 'sdl2': ['PySDL2'],
103 'wx': ['wxPython'],
104 },
105 packages=package_tree('vispy'),
106 package_dir={
107 'vispy': 'vispy'},
108 package_data={
109 'vispy': [op.join('io', '_data', '*'),
110 op.join('html', 'static', 'js', '*'),
111 op.join('app', 'tests', 'qt-designer.ui'),
112 op.join('..', 'doc', '*'),
113 ],
114
115 'vispy.glsl': ['*.vert','*.frag', "*.glsl"],
116 'vispy.glsl.antialias': ['*.vert','*.frag', "*.glsl"],
117 'vispy.glsl.arrowheads': ['*.vert','*.frag', "*.glsl"],
118 'vispy.glsl.arrows': ['*.vert','*.frag', "*.glsl"],
119 'vispy.glsl.collections': ['*.vert','*.frag', "*.glsl"],
120 'vispy.glsl.colormaps': ['*.vert','*.frag', "*.glsl"],
121 'vispy.glsl.lines': ['*.vert','*.frag', "*.glsl"],
122 'vispy.glsl.markers': ['*.vert','*.frag', "*.glsl"],
123 'vispy.glsl.math': ['*.vert','*.frag', "*.glsl"],
124 'vispy.glsl.misc': ['*.vert','*.frag', "*.glsl"],
125 'vispy.glsl.transforms': ['*.vert','*.frag', "*.glsl"],
126
127 },
128 zip_safe=False,
129 classifiers=[
130 'Development Status :: 3 - Alpha',
131 'Intended Audience :: Science/Research',
132 'Intended Audience :: Education',
133 'Intended Audience :: Developers',
134 'Topic :: Scientific/Engineering :: Visualization',
135 'License :: OSI Approved :: BSD License',
136 'Operating System :: MacOS :: MacOS X',
137 'Operating System :: Microsoft :: Windows',
138 'Operating System :: POSIX',
139 'Programming Language :: Python',
140 'Programming Language :: Python :: 2.7',
141 'Programming Language :: Python :: 3.3',
142 'Programming Language :: Python :: 3.4',
143 'Programming Language :: Python :: 3.5',
144 'Programming Language :: Python :: 3.6',
145 'Framework :: IPython'
146 ],
147 )
148
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/setup.py b/setup.py
--- a/setup.py
+++ b/setup.py
@@ -109,7 +109,6 @@
'vispy': [op.join('io', '_data', '*'),
op.join('html', 'static', 'js', '*'),
op.join('app', 'tests', 'qt-designer.ui'),
- op.join('..', 'doc', '*'),
],
'vispy.glsl': ['*.vert','*.frag', "*.glsl"],
| {"golden_diff": "diff --git a/setup.py b/setup.py\n--- a/setup.py\n+++ b/setup.py\n@@ -109,7 +109,6 @@\n 'vispy': [op.join('io', '_data', '*'),\n op.join('html', 'static', 'js', '*'),\n op.join('app', 'tests', 'qt-designer.ui'),\n- op.join('..', 'doc', '*'),\n ],\n \n 'vispy.glsl': ['*.vert','*.frag', \"*.glsl\"],\n", "issue": "Wrong install location for docs\nFrom `setup.py`:\r\n```\r\nsetup(\r\n [...]\r\n package_data={\r\n 'vispy': [op.join('io', '_data', '*'),\r\n op.join('html', 'static', 'js', '*'),\r\n op.join('app', 'tests', 'qt-designer.ui'),\r\n op.join('..', 'doc', '*'),\r\n ],\r\n```\r\nThis line `op.join('..', 'doc', '*')` is wrong for a system-wide install. It leads to the documentation being install under `dist-packages` or `site-packages`, which is definitely non-standard. IMO, the best would be to just not install the docs yourself, and let the package build system (conda or Debian) handle it.\n", "before_files": [{"content": "# -*- coding: utf-8 -*-\n# Copyright (c) Vispy Development Team. All Rights Reserved.\n# Distributed under the (new) BSD License. See LICENSE.txt for more info.\n\n\"\"\" Vispy setup script.\n\nSteps to do a new release:\n\nPreparations:\n * Test on Windows, Linux, Mac\n * Make release notes\n * Update API documentation and other docs that need updating.\n * Install 'twine' package for uploading to PyPI\n\nDefine the version:\n * update __version__ in __init__.py\n * tag the tip changeset as version x.x.x; `git tag -a 'vX.Y.Z'`\n\nTest installation:\n * clear the build and dist dir (if they exist)\n * python setup.py sdist\n * twine register --repository-url https://test.pypi.org/legacy/ dist/*\n * twine upload --repository-url https://test.pypi.org/legacy/ dist/*\n * pip install -i https://testpypi.python.org/pypi vispy\n\nGenerate and upload package\n * python setup.py sdist\n * twine register dist/*\n * twine upload dist/*\n\nAnnouncing:\n * It can be worth waiting a day for eager users to report critical bugs\n * Announce in scipy-user, vispy mailing list, G+\n\n\"\"\"\n\nimport os\nfrom os import path as op\nfrom warnings import warn\n\ntry:\n # use setuptools namespace, allows for \"develop\"\n import setuptools # noqa, analysis:ignore\nexcept ImportError:\n warn(\"unable to load setuptools. 'setup.py develop' will not work\")\n pass # it's not essential for installation\nfrom distutils.core import setup\n\nname = 'vispy'\ndescription = 'Interactive visualization in Python'\n\n\n# Get version and docstring\n__version__ = None\n__doc__ = ''\ndocStatus = 0 # Not started, in progress, done\ninitFile = os.path.join(os.path.dirname(__file__), 'vispy', '__init__.py')\nfor line in open(initFile).readlines():\n if (line.startswith('version_info') or line.startswith('__version__')):\n exec(line.strip())\n elif line.startswith('\"\"\"'):\n if docStatus == 0:\n docStatus = 1\n line = line.lstrip('\"')\n elif docStatus == 1:\n docStatus = 2\n if docStatus == 1:\n __doc__ += line\n\n\ndef package_tree(pkgroot):\n path = os.path.dirname(__file__)\n subdirs = [os.path.relpath(i[0], path).replace(os.path.sep, '.')\n for i in os.walk(os.path.join(path, pkgroot))\n if '__init__.py' in i[2]]\n return subdirs\n\n\nsetup(\n name=name,\n version=__version__,\n author='Vispy contributors',\n author_email='[email protected]',\n license='(new) BSD',\n url='http://vispy.org',\n download_url='https://pypi.python.org/pypi/vispy',\n keywords=\"visualization OpenGl ES medical imaging 3D plotting \"\n \"numpy bigdata\",\n description=description,\n long_description=__doc__,\n platforms='any',\n provides=['vispy'],\n install_requires=['numpy'],\n extras_require={\n 'ipython-static': ['ipython'],\n 'ipython-vnc': ['ipython>=2'],\n 'ipython-webgl': ['ipython>=2', 'tornado'],\n 'pyglet': ['pyglet>=1.2'],\n # 'pyqt4': [], # Why is this on PyPI, but without downloads?\n # 'pyqt5': [], # Ditto.\n 'pyside': ['PySide'],\n 'sdl2': ['PySDL2'],\n 'wx': ['wxPython'],\n },\n packages=package_tree('vispy'),\n package_dir={\n 'vispy': 'vispy'},\n package_data={\n 'vispy': [op.join('io', '_data', '*'),\n op.join('html', 'static', 'js', '*'),\n op.join('app', 'tests', 'qt-designer.ui'),\n op.join('..', 'doc', '*'),\n ],\n\n 'vispy.glsl': ['*.vert','*.frag', \"*.glsl\"],\n 'vispy.glsl.antialias': ['*.vert','*.frag', \"*.glsl\"],\n 'vispy.glsl.arrowheads': ['*.vert','*.frag', \"*.glsl\"],\n 'vispy.glsl.arrows': ['*.vert','*.frag', \"*.glsl\"],\n 'vispy.glsl.collections': ['*.vert','*.frag', \"*.glsl\"],\n 'vispy.glsl.colormaps': ['*.vert','*.frag', \"*.glsl\"],\n 'vispy.glsl.lines': ['*.vert','*.frag', \"*.glsl\"],\n 'vispy.glsl.markers': ['*.vert','*.frag', \"*.glsl\"],\n 'vispy.glsl.math': ['*.vert','*.frag', \"*.glsl\"],\n 'vispy.glsl.misc': ['*.vert','*.frag', \"*.glsl\"],\n 'vispy.glsl.transforms': ['*.vert','*.frag', \"*.glsl\"],\n\n },\n zip_safe=False,\n classifiers=[\n 'Development Status :: 3 - Alpha',\n 'Intended Audience :: Science/Research',\n 'Intended Audience :: Education',\n 'Intended Audience :: Developers',\n 'Topic :: Scientific/Engineering :: Visualization',\n 'License :: OSI Approved :: BSD License',\n 'Operating System :: MacOS :: MacOS X',\n 'Operating System :: Microsoft :: Windows',\n 'Operating System :: POSIX',\n 'Programming Language :: Python',\n 'Programming Language :: Python :: 2.7',\n 'Programming Language :: Python :: 3.3',\n 'Programming Language :: Python :: 3.4',\n 'Programming Language :: Python :: 3.5',\n 'Programming Language :: Python :: 3.6',\n 'Framework :: IPython'\n ],\n)\n", "path": "setup.py"}], "after_files": [{"content": "# -*- coding: utf-8 -*-\n# Copyright (c) Vispy Development Team. All Rights Reserved.\n# Distributed under the (new) BSD License. See LICENSE.txt for more info.\n\n\"\"\" Vispy setup script.\n\nSteps to do a new release:\n\nPreparations:\n * Test on Windows, Linux, Mac\n * Make release notes\n * Update API documentation and other docs that need updating.\n * Install 'twine' package for uploading to PyPI\n\nDefine the version:\n * update __version__ in __init__.py\n * tag the tip changeset as version x.x.x; `git tag -a 'vX.Y.Z'`\n\nTest installation:\n * clear the build and dist dir (if they exist)\n * python setup.py sdist\n * twine register --repository-url https://test.pypi.org/legacy/ dist/*\n * twine upload --repository-url https://test.pypi.org/legacy/ dist/*\n * pip install -i https://testpypi.python.org/pypi vispy\n\nGenerate and upload package\n * python setup.py sdist\n * twine register dist/*\n * twine upload dist/*\n\nAnnouncing:\n * It can be worth waiting a day for eager users to report critical bugs\n * Announce in scipy-user, vispy mailing list, G+\n\n\"\"\"\n\nimport os\nfrom os import path as op\nfrom warnings import warn\n\ntry:\n # use setuptools namespace, allows for \"develop\"\n import setuptools # noqa, analysis:ignore\nexcept ImportError:\n warn(\"unable to load setuptools. 'setup.py develop' will not work\")\n pass # it's not essential for installation\nfrom distutils.core import setup\n\nname = 'vispy'\ndescription = 'Interactive visualization in Python'\n\n\n# Get version and docstring\n__version__ = None\n__doc__ = ''\ndocStatus = 0 # Not started, in progress, done\ninitFile = os.path.join(os.path.dirname(__file__), 'vispy', '__init__.py')\nfor line in open(initFile).readlines():\n if (line.startswith('version_info') or line.startswith('__version__')):\n exec(line.strip())\n elif line.startswith('\"\"\"'):\n if docStatus == 0:\n docStatus = 1\n line = line.lstrip('\"')\n elif docStatus == 1:\n docStatus = 2\n if docStatus == 1:\n __doc__ += line\n\n\ndef package_tree(pkgroot):\n path = os.path.dirname(__file__)\n subdirs = [os.path.relpath(i[0], path).replace(os.path.sep, '.')\n for i in os.walk(os.path.join(path, pkgroot))\n if '__init__.py' in i[2]]\n return subdirs\n\n\nsetup(\n name=name,\n version=__version__,\n author='Vispy contributors',\n author_email='[email protected]',\n license='(new) BSD',\n url='http://vispy.org',\n download_url='https://pypi.python.org/pypi/vispy',\n keywords=\"visualization OpenGl ES medical imaging 3D plotting \"\n \"numpy bigdata\",\n description=description,\n long_description=__doc__,\n platforms='any',\n provides=['vispy'],\n install_requires=['numpy'],\n extras_require={\n 'ipython-static': ['ipython'],\n 'ipython-vnc': ['ipython>=2'],\n 'ipython-webgl': ['ipython>=2', 'tornado'],\n 'pyglet': ['pyglet>=1.2'],\n # 'pyqt4': [], # Why is this on PyPI, but without downloads?\n # 'pyqt5': [], # Ditto.\n 'pyside': ['PySide'],\n 'sdl2': ['PySDL2'],\n 'wx': ['wxPython'],\n },\n packages=package_tree('vispy'),\n package_dir={\n 'vispy': 'vispy'},\n package_data={\n 'vispy': [op.join('io', '_data', '*'),\n op.join('html', 'static', 'js', '*'),\n op.join('app', 'tests', 'qt-designer.ui'),\n ],\n\n 'vispy.glsl': ['*.vert','*.frag', \"*.glsl\"],\n 'vispy.glsl.antialias': ['*.vert','*.frag', \"*.glsl\"],\n 'vispy.glsl.arrowheads': ['*.vert','*.frag', \"*.glsl\"],\n 'vispy.glsl.arrows': ['*.vert','*.frag', \"*.glsl\"],\n 'vispy.glsl.collections': ['*.vert','*.frag', \"*.glsl\"],\n 'vispy.glsl.colormaps': ['*.vert','*.frag', \"*.glsl\"],\n 'vispy.glsl.lines': ['*.vert','*.frag', \"*.glsl\"],\n 'vispy.glsl.markers': ['*.vert','*.frag', \"*.glsl\"],\n 'vispy.glsl.math': ['*.vert','*.frag', \"*.glsl\"],\n 'vispy.glsl.misc': ['*.vert','*.frag', \"*.glsl\"],\n 'vispy.glsl.transforms': ['*.vert','*.frag', \"*.glsl\"],\n\n },\n zip_safe=False,\n classifiers=[\n 'Development Status :: 3 - Alpha',\n 'Intended Audience :: Science/Research',\n 'Intended Audience :: Education',\n 'Intended Audience :: Developers',\n 'Topic :: Scientific/Engineering :: Visualization',\n 'License :: OSI Approved :: BSD License',\n 'Operating System :: MacOS :: MacOS X',\n 'Operating System :: Microsoft :: Windows',\n 'Operating System :: POSIX',\n 'Programming Language :: Python',\n 'Programming Language :: Python :: 2.7',\n 'Programming Language :: Python :: 3.3',\n 'Programming Language :: Python :: 3.4',\n 'Programming Language :: Python :: 3.5',\n 'Programming Language :: Python :: 3.6',\n 'Framework :: IPython'\n ],\n)\n", "path": "setup.py"}]} | 2,056 | 112 |
gh_patches_debug_12654 | rasdani/github-patches | git_diff | ocadotechnology__aimmo-499 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Game Creator RC initialised with wrong game API URL
The `REPLACE_ME` change in one of the latest PR's has broken the game on minikube level in `minikube.py`. The URL is incorrect so minikube does not work and prohibits testing.
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `setup.py`
Content:
```
1 # -*- coding: utf-8 -*-
2 from setuptools import find_packages, setup
3
4 import versioneer
5
6 setup(
7 name='aimmo',
8 cmdclass=versioneer.get_cmdclass(),
9 packages=find_packages(),
10 include_package_data=True,
11 install_requires=[
12 'django >= 1.8.3, < 1.9.0',
13 'django-autoconfig >= 0.3.6, < 1.0.0',
14 'django-forms-bootstrap',
15 'django-js-reverse',
16 'eventlet',
17 'flask',
18 'flask-socketio',
19 'requests',
20 'six',
21 'pykube',
22 'hypothesis',
23 'flask-cors >= 3.0, < 3.1',
24 'psutil >= 5.4, < 5.5',
25 ],
26 tests_require=[
27 'django-setuptest',
28 'httmock',
29 ],
30 test_suite='setuptest.setuptest.SetupTestSuite',
31 version=versioneer.get_version(),
32 zip_safe=False,
33 )
34
```
Path: `aimmo_runner/shell_api.py`
Content:
```
1 import subprocess
2 import sys
3 import os
4 import stat
5 import errno
6 import platform
7 from subprocess import CalledProcessError
8 from urllib import urlretrieve, urlopen
9
10 BASE_DIR = os.path.abspath(os.path.dirname(os.path.dirname(__file__)))
11 TEST_BIN = os.path.join(BASE_DIR, 'test-bin')
12 OS = platform.system().lower()
13 FILE_SUFFIX = '.exe' if OS == 'windows' else ''
14 KUBECTL = os.path.join(TEST_BIN, 'kubectl%s' % FILE_SUFFIX)
15 MINIKUBE = os.path.join(TEST_BIN, 'minikube%s' % FILE_SUFFIX)
16 FNULL = open(os.devnull, 'w')
17
18 def log(message):
19 sys.stderr.write(message + "\n")
20
21
22 def run_command(args, capture_output=False):
23 try:
24 if capture_output:
25 return subprocess.check_output(args)
26 else:
27 subprocess.check_call(args)
28 except CalledProcessError as e:
29 log('Command failed with exit status %d: %s' % (e.returncode, ' '.join(args)))
30 raise
31
32
33 def run_command_async(args, capture_output=False):
34 if capture_output is True:
35 p = subprocess.Popen(args, stdout=FNULL, stderr=subprocess.STDOUT)
36 else:
37 p = subprocess.Popen(args)
38 return p
39
40
41 def create_test_bin():
42 try:
43 os.makedirs(TEST_BIN)
44 except OSError as err:
45 if err.errno != errno.EEXIST:
46 raise
47
48
49 def binary_exists(filename):
50 # Check if binary is callable on our path
51 try:
52 run_command([filename], True)
53 return True
54 except OSError:
55 return False
56
57
58 def download_exec(url, dest):
59 dest = urlretrieve(url, dest)[0]
60 make_exec(dest)
61
62
63 def make_exec(file):
64 current_stat = os.stat(file)
65 os.chmod(file, current_stat.st_mode | stat.S_IEXEC)
66
67
68 def get_latest_github_version(repo):
69 result = urlopen('https://github.com/%s/releases/latest' % repo)
70 return result.geturl().split('/')[-1]
71
72
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/aimmo_runner/shell_api.py b/aimmo_runner/shell_api.py
--- a/aimmo_runner/shell_api.py
+++ b/aimmo_runner/shell_api.py
@@ -15,6 +15,7 @@
MINIKUBE = os.path.join(TEST_BIN, 'minikube%s' % FILE_SUFFIX)
FNULL = open(os.devnull, 'w')
+
def log(message):
sys.stderr.write(message + "\n")
diff --git a/setup.py b/setup.py
--- a/setup.py
+++ b/setup.py
@@ -26,6 +26,10 @@
tests_require=[
'django-setuptest',
'httmock',
+ 'mock == 2.0.0',
+ 'docker == 2.7.0',
+ 'kubernetes == 4.0.0',
+ 'PyYAML == 3.12',
],
test_suite='setuptest.setuptest.SetupTestSuite',
version=versioneer.get_version(),
| {"golden_diff": "diff --git a/aimmo_runner/shell_api.py b/aimmo_runner/shell_api.py\n--- a/aimmo_runner/shell_api.py\n+++ b/aimmo_runner/shell_api.py\n@@ -15,6 +15,7 @@\n MINIKUBE = os.path.join(TEST_BIN, 'minikube%s' % FILE_SUFFIX)\n FNULL = open(os.devnull, 'w')\n \n+\n def log(message):\n sys.stderr.write(message + \"\\n\")\n \ndiff --git a/setup.py b/setup.py\n--- a/setup.py\n+++ b/setup.py\n@@ -26,6 +26,10 @@\n tests_require=[\n 'django-setuptest',\n 'httmock',\n+ 'mock == 2.0.0',\n+ 'docker == 2.7.0',\n+ 'kubernetes == 4.0.0',\n+ 'PyYAML == 3.12',\n ],\n test_suite='setuptest.setuptest.SetupTestSuite',\n version=versioneer.get_version(),\n", "issue": "Game Creator RC initialised with wrong game API URL\nThe `REPLACE_ME` change in one of the latest PR's has broken the game on minikube level in `minikube.py`. The URL is incorrect so minikube does not work and prohibits testing. \r\n\r\n\n", "before_files": [{"content": "# -*- coding: utf-8 -*-\nfrom setuptools import find_packages, setup\n\nimport versioneer\n\nsetup(\n name='aimmo',\n cmdclass=versioneer.get_cmdclass(),\n packages=find_packages(),\n include_package_data=True,\n install_requires=[\n 'django >= 1.8.3, < 1.9.0',\n 'django-autoconfig >= 0.3.6, < 1.0.0',\n 'django-forms-bootstrap',\n 'django-js-reverse',\n 'eventlet',\n 'flask',\n 'flask-socketio',\n 'requests',\n 'six',\n 'pykube',\n 'hypothesis',\n 'flask-cors >= 3.0, < 3.1',\n 'psutil >= 5.4, < 5.5',\n ],\n tests_require=[\n 'django-setuptest',\n 'httmock',\n ],\n test_suite='setuptest.setuptest.SetupTestSuite',\n version=versioneer.get_version(),\n zip_safe=False,\n)\n", "path": "setup.py"}, {"content": "import subprocess\nimport sys\nimport os\nimport stat\nimport errno\nimport platform\nfrom subprocess import CalledProcessError\nfrom urllib import urlretrieve, urlopen\n\nBASE_DIR = os.path.abspath(os.path.dirname(os.path.dirname(__file__)))\nTEST_BIN = os.path.join(BASE_DIR, 'test-bin')\nOS = platform.system().lower()\nFILE_SUFFIX = '.exe' if OS == 'windows' else ''\nKUBECTL = os.path.join(TEST_BIN, 'kubectl%s' % FILE_SUFFIX)\nMINIKUBE = os.path.join(TEST_BIN, 'minikube%s' % FILE_SUFFIX)\nFNULL = open(os.devnull, 'w')\n\ndef log(message):\n sys.stderr.write(message + \"\\n\")\n\n\ndef run_command(args, capture_output=False):\n try:\n if capture_output:\n return subprocess.check_output(args)\n else:\n subprocess.check_call(args)\n except CalledProcessError as e:\n log('Command failed with exit status %d: %s' % (e.returncode, ' '.join(args)))\n raise\n\n\ndef run_command_async(args, capture_output=False):\n if capture_output is True:\n p = subprocess.Popen(args, stdout=FNULL, stderr=subprocess.STDOUT)\n else:\n p = subprocess.Popen(args)\n return p\n\n\ndef create_test_bin():\n try:\n os.makedirs(TEST_BIN)\n except OSError as err:\n if err.errno != errno.EEXIST:\n raise\n\n\ndef binary_exists(filename):\n # Check if binary is callable on our path\n try:\n run_command([filename], True)\n return True\n except OSError:\n return False\n\n\ndef download_exec(url, dest):\n dest = urlretrieve(url, dest)[0]\n make_exec(dest)\n\n\ndef make_exec(file):\n current_stat = os.stat(file)\n os.chmod(file, current_stat.st_mode | stat.S_IEXEC)\n\n\ndef get_latest_github_version(repo):\n result = urlopen('https://github.com/%s/releases/latest' % repo)\n return result.geturl().split('/')[-1]\n\n", "path": "aimmo_runner/shell_api.py"}], "after_files": [{"content": "# -*- coding: utf-8 -*-\nfrom setuptools import find_packages, setup\n\nimport versioneer\n\nsetup(\n name='aimmo',\n cmdclass=versioneer.get_cmdclass(),\n packages=find_packages(),\n include_package_data=True,\n install_requires=[\n 'django >= 1.8.3, < 1.9.0',\n 'django-autoconfig >= 0.3.6, < 1.0.0',\n 'django-forms-bootstrap',\n 'django-js-reverse',\n 'eventlet',\n 'flask',\n 'flask-socketio',\n 'requests',\n 'six',\n 'pykube',\n 'hypothesis',\n 'flask-cors >= 3.0, < 3.1',\n 'psutil >= 5.4, < 5.5',\n ],\n tests_require=[\n 'django-setuptest',\n 'httmock',\n 'mock == 2.0.0',\n 'docker == 2.7.0',\n 'kubernetes == 4.0.0',\n 'PyYAML == 3.12',\n ],\n test_suite='setuptest.setuptest.SetupTestSuite',\n version=versioneer.get_version(),\n zip_safe=False,\n)\n", "path": "setup.py"}, {"content": "import subprocess\nimport sys\nimport os\nimport stat\nimport errno\nimport platform\nfrom subprocess import CalledProcessError\nfrom urllib import urlretrieve, urlopen\n\nBASE_DIR = os.path.abspath(os.path.dirname(os.path.dirname(__file__)))\nTEST_BIN = os.path.join(BASE_DIR, 'test-bin')\nOS = platform.system().lower()\nFILE_SUFFIX = '.exe' if OS == 'windows' else ''\nKUBECTL = os.path.join(TEST_BIN, 'kubectl%s' % FILE_SUFFIX)\nMINIKUBE = os.path.join(TEST_BIN, 'minikube%s' % FILE_SUFFIX)\nFNULL = open(os.devnull, 'w')\n\n\ndef log(message):\n sys.stderr.write(message + \"\\n\")\n\n\ndef run_command(args, capture_output=False):\n try:\n if capture_output:\n return subprocess.check_output(args)\n else:\n subprocess.check_call(args)\n except CalledProcessError as e:\n log('Command failed with exit status %d: %s' % (e.returncode, ' '.join(args)))\n raise\n\n\ndef run_command_async(args, capture_output=False):\n if capture_output is True:\n p = subprocess.Popen(args, stdout=FNULL, stderr=subprocess.STDOUT)\n else:\n p = subprocess.Popen(args)\n return p\n\n\ndef create_test_bin():\n try:\n os.makedirs(TEST_BIN)\n except OSError as err:\n if err.errno != errno.EEXIST:\n raise\n\n\ndef binary_exists(filename):\n # Check if binary is callable on our path\n try:\n run_command([filename], True)\n return True\n except OSError:\n return False\n\n\ndef download_exec(url, dest):\n dest = urlretrieve(url, dest)[0]\n make_exec(dest)\n\n\ndef make_exec(file):\n current_stat = os.stat(file)\n os.chmod(file, current_stat.st_mode | stat.S_IEXEC)\n\n\ndef get_latest_github_version(repo):\n result = urlopen('https://github.com/%s/releases/latest' % repo)\n return result.geturl().split('/')[-1]\n\n", "path": "aimmo_runner/shell_api.py"}]} | 1,197 | 227 |
gh_patches_debug_2784 | rasdani/github-patches | git_diff | archlinux__archinstall-1954 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
[master] awesome (desktops in general?) don't install packages?
It appears when choosing awesome, install is called: https://github.com/archlinux/archinstall/blob/7326d51161bf6fd7f1c683cf1d7ce09338efe4b7/archinstall/default_profiles/desktops/awesome.py#L23-L24
And super being `XorgProfile`: https://github.com/archlinux/archinstall/blob/7326d51161bf6fd7f1c683cf1d7ce09338efe4b7/archinstall/default_profiles/xorg.py#L1-L21
That class does not have an install so it calls `Profile.install()` which contains: https://github.com/archlinux/archinstall/blob/7326d51161bf6fd7f1c683cf1d7ce09338efe4b7/archinstall/default_profiles/profile.py#L101-L104
Which is just a placeholder?

I haven't ran through all the profiles yet, but have we overlooked something here?
What happened to all the packages per profile when we moved them to the dataclass structure? :)
I obviously missed something in a PR some where hehe
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `archinstall/default_profiles/desktops/awesome.py`
Content:
```
1 from typing import List, Optional, Any, TYPE_CHECKING
2
3 from archinstall.default_profiles.profile import ProfileType
4 from archinstall.default_profiles.xorg import XorgProfile
5
6 if TYPE_CHECKING:
7 from archinstall.lib.installer import Installer
8 _: Any
9
10
11 class AwesomeProfile(XorgProfile):
12 def __init__(self):
13 super().__init__('Awesome', ProfileType.WindowMgr, description='')
14
15 @property
16 def packages(self) -> List[str]:
17 return ['alacritty']
18
19 def preview_text(self) -> Optional[str]:
20 text = str(_('Environment type: {}')).format(self.profile_type.value)
21 return text + '\n' + self.packages_text()
22
23 def install(self, install_session: 'Installer'):
24 super().install(install_session)
25
26 # TODO: Copy a full configuration to ~/.config/awesome/rc.lua instead.
27 with open(f"{install_session.target}/etc/xdg/awesome/rc.lua", 'r') as fh:
28 awesome_lua = fh.read()
29
30 # Replace xterm with alacritty for a smoother experience.
31 awesome_lua = awesome_lua.replace('"xterm"', '"alacritty"')
32
33 with open(f"{install_session.target}/etc/xdg/awesome/rc.lua", 'w') as fh:
34 fh.write(awesome_lua)
35
36 # TODO: Configure the right-click-menu to contain the above packages that were installed. (as a user config)
37
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/archinstall/default_profiles/desktops/awesome.py b/archinstall/default_profiles/desktops/awesome.py
--- a/archinstall/default_profiles/desktops/awesome.py
+++ b/archinstall/default_profiles/desktops/awesome.py
@@ -14,7 +14,10 @@
@property
def packages(self) -> List[str]:
- return ['alacritty']
+ return [
+ 'awesome',
+ 'alacritty'
+ ]
def preview_text(self) -> Optional[str]:
text = str(_('Environment type: {}')).format(self.profile_type.value)
| {"golden_diff": "diff --git a/archinstall/default_profiles/desktops/awesome.py b/archinstall/default_profiles/desktops/awesome.py\n--- a/archinstall/default_profiles/desktops/awesome.py\n+++ b/archinstall/default_profiles/desktops/awesome.py\n@@ -14,7 +14,10 @@\n \n \t@property\n \tdef packages(self) -> List[str]:\n-\t\treturn ['alacritty']\n+\t\treturn [\n+\t\t\t'awesome',\n+\t\t\t'alacritty'\n+\t\t]\n \n \tdef preview_text(self) -> Optional[str]:\n \t\ttext = str(_('Environment type: {}')).format(self.profile_type.value)\n", "issue": "[master] awesome (desktops in general?) don't install packages?\nIt appears when choosing awesome, install is called: https://github.com/archlinux/archinstall/blob/7326d51161bf6fd7f1c683cf1d7ce09338efe4b7/archinstall/default_profiles/desktops/awesome.py#L23-L24\r\n\r\nAnd super being `XorgProfile`: https://github.com/archlinux/archinstall/blob/7326d51161bf6fd7f1c683cf1d7ce09338efe4b7/archinstall/default_profiles/xorg.py#L1-L21\r\n\r\nThat class does not have an install so it calls `Profile.install()` which contains: https://github.com/archlinux/archinstall/blob/7326d51161bf6fd7f1c683cf1d7ce09338efe4b7/archinstall/default_profiles/profile.py#L101-L104\r\nWhich is just a placeholder?\r\n\r\n\r\n\r\nI haven't ran through all the profiles yet, but have we overlooked something here?\r\nWhat happened to all the packages per profile when we moved them to the dataclass structure? :)\r\n\r\nI obviously missed something in a PR some where hehe\n", "before_files": [{"content": "from typing import List, Optional, Any, TYPE_CHECKING\n\nfrom archinstall.default_profiles.profile import ProfileType\nfrom archinstall.default_profiles.xorg import XorgProfile\n\nif TYPE_CHECKING:\n\tfrom archinstall.lib.installer import Installer\n\t_: Any\n\n\nclass AwesomeProfile(XorgProfile):\n\tdef __init__(self):\n\t\tsuper().__init__('Awesome', ProfileType.WindowMgr, description='')\n\n\t@property\n\tdef packages(self) -> List[str]:\n\t\treturn ['alacritty']\n\n\tdef preview_text(self) -> Optional[str]:\n\t\ttext = str(_('Environment type: {}')).format(self.profile_type.value)\n\t\treturn text + '\\n' + self.packages_text()\n\n\tdef install(self, install_session: 'Installer'):\n\t\tsuper().install(install_session)\n\n\t\t# TODO: Copy a full configuration to ~/.config/awesome/rc.lua instead.\n\t\twith open(f\"{install_session.target}/etc/xdg/awesome/rc.lua\", 'r') as fh:\n\t\t\tawesome_lua = fh.read()\n\n\t\t# Replace xterm with alacritty for a smoother experience.\n\t\tawesome_lua = awesome_lua.replace('\"xterm\"', '\"alacritty\"')\n\n\t\twith open(f\"{install_session.target}/etc/xdg/awesome/rc.lua\", 'w') as fh:\n\t\t\tfh.write(awesome_lua)\n\n\t\t# TODO: Configure the right-click-menu to contain the above packages that were installed. (as a user config)\n", "path": "archinstall/default_profiles/desktops/awesome.py"}], "after_files": [{"content": "from typing import List, Optional, Any, TYPE_CHECKING\n\nfrom archinstall.default_profiles.profile import ProfileType\nfrom archinstall.default_profiles.xorg import XorgProfile\n\nif TYPE_CHECKING:\n\tfrom archinstall.lib.installer import Installer\n\t_: Any\n\n\nclass AwesomeProfile(XorgProfile):\n\tdef __init__(self):\n\t\tsuper().__init__('Awesome', ProfileType.WindowMgr, description='')\n\n\t@property\n\tdef packages(self) -> List[str]:\n\t\treturn [\n\t\t\t'awesome',\n\t\t\t'alacritty'\n\t\t]\n\n\tdef preview_text(self) -> Optional[str]:\n\t\ttext = str(_('Environment type: {}')).format(self.profile_type.value)\n\t\treturn text + '\\n' + self.packages_text()\n\n\tdef install(self, install_session: 'Installer'):\n\t\tsuper().install(install_session)\n\n\t\t# TODO: Copy a full configuration to ~/.config/awesome/rc.lua instead.\n\t\twith open(f\"{install_session.target}/etc/xdg/awesome/rc.lua\", 'r') as fh:\n\t\t\tawesome_lua = fh.read()\n\n\t\t# Replace xterm with alacritty for a smoother experience.\n\t\tawesome_lua = awesome_lua.replace('\"xterm\"', '\"alacritty\"')\n\n\t\twith open(f\"{install_session.target}/etc/xdg/awesome/rc.lua\", 'w') as fh:\n\t\t\tfh.write(awesome_lua)\n\n\t\t# TODO: Configure the right-click-menu to contain the above packages that were installed. (as a user config)\n", "path": "archinstall/default_profiles/desktops/awesome.py"}]} | 984 | 133 |
gh_patches_debug_8701 | rasdani/github-patches | git_diff | sublimelsp__LSP-658 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
hover regression
After the latest commits hover popups produce errors, I suspect c9a73907f00972a415e8e979d5dcc96c11ec3f09 is responsible for this
* OS: Windows 10
* language server: pyls
* How you installed LSP: git
* Minimal reproduction steps
1. open a .py file
2. hover over an function name
* Log
```
LSP: --> textDocument/hover
LSP: {'contents': "is_supported_syntax(syntax: str, configs: 'List[ClientConfig]') -> bool"}
Parse Error: #x27;List[ClientConfig]') -> bool</p><p><a href='definition'>Definition</a> | <a href='references'>References</a> | <a href='rename'>Rename</a></p></div></div> code: Unexpected character
Parse Error: 27;List[ClientConfig]') -> bool</p><p><a href='definition'>Definition</a> | <a href='references'>References</a> | <a href='rename'>Rename</a></p></div></div> code: Unexpected character
Parse Error: 7;List[ClientConfig]') -> bool</p><p><a href='definition'>Definition</a> | <a href='references'>References</a> | <a href='rename'>Rename</a></p></div></div> code: Unexpected character
Parse Error: #x27; code: Unknown entity
Parse Error: #x27;) -> bool</p><p><a href='definition'>Definition</a> | <a href='references'>References</a> | <a href='rename'>Rename</a></p></div></div> code: Unexpected character
Parse Error: 27;) -> bool</p><p><a href='definition'>Definition</a> | <a href='references'>References</a> | <a href='rename'>Rename</a></p></div></div> code: Unexpected character
Parse Error: 7;) -> bool</p><p><a href='definition'>Definition</a> | <a href='references'>References</a> | <a href='rename'>Rename</a></p></div></div> code: Unexpected character
Parse Error: #x27; code: Unknown entity
```
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `plugin/hover.py`
Content:
```
1 import mdpopups
2 import sublime
3 import sublime_plugin
4 import webbrowser
5 from html import escape
6 try:
7 from typing import List, Optional, Any, Dict
8 assert List and Optional and Any and Dict
9 except ImportError:
10 pass
11
12 from .core.configurations import is_supported_syntax
13 from .diagnostics import get_point_diagnostics
14 from .core.registry import session_for_view, LspTextCommand
15 from .core.protocol import Request, DiagnosticSeverity
16 from .core.documents import get_document_position
17 from .core.popups import popup_css, popup_class
18 from .core.settings import client_configs
19
20 SUBLIME_WORD_MASK = 515
21 NO_HOVER_SCOPES = 'comment, string'
22
23
24 class HoverHandler(sublime_plugin.ViewEventListener):
25 def __init__(self, view):
26 self.view = view
27
28 @classmethod
29 def is_applicable(cls, settings):
30 syntax = settings.get('syntax')
31 return syntax and is_supported_syntax(syntax, client_configs.all)
32
33 def on_hover(self, point, hover_zone):
34 if hover_zone != sublime.HOVER_TEXT or self.view.is_popup_visible():
35 return
36 self.view.run_command("lsp_hover", {"point": point})
37
38
39 _test_contents = [] # type: List[str]
40
41
42 class_for_severity = {
43 DiagnosticSeverity.Error: 'errors',
44 DiagnosticSeverity.Warning: 'warnings',
45 DiagnosticSeverity.Information: 'info',
46 DiagnosticSeverity.Hint: 'hints'
47 }
48
49
50 class LspHoverCommand(LspTextCommand):
51 def __init__(self, view):
52 super().__init__(view)
53
54 def is_likely_at_symbol(self, point):
55 word_at_sel = self.view.classify(point)
56 return word_at_sel & SUBLIME_WORD_MASK and not self.view.match_selector(point, NO_HOVER_SCOPES)
57
58 def run(self, edit, point=None):
59 if point is None:
60 point = self.view.sel()[0].begin()
61 if self.is_likely_at_symbol(point):
62 self.request_symbol_hover(point)
63 point_diagnostics = get_point_diagnostics(self.view, point)
64 if point_diagnostics:
65 self.show_hover(point, self.diagnostics_content(point_diagnostics))
66
67 def request_symbol_hover(self, point) -> None:
68 session = session_for_view(self.view, point)
69 if session:
70 if session.has_capability('hoverProvider'):
71 document_position = get_document_position(self.view, point)
72 if document_position:
73 if session.client:
74 session.client.send_request(
75 Request.hover(document_position),
76 lambda response: self.handle_response(response, point))
77
78 def handle_response(self, response: 'Optional[Any]', point) -> None:
79 all_content = ""
80
81 point_diagnostics = get_point_diagnostics(self.view, point)
82 if point_diagnostics:
83 all_content += self.diagnostics_content(point_diagnostics)
84
85 all_content += self.hover_content(point, response)
86 all_content += self.symbol_actions_content()
87
88 _test_contents.clear()
89 _test_contents.append(all_content) # for testing only
90 self.show_hover(point, all_content)
91
92 def symbol_actions_content(self):
93 actions = []
94 if self.has_client_with_capability('definitionProvider'):
95 actions.append("<a href='{}'>{}</a>".format('definition', 'Definition'))
96
97 if self.has_client_with_capability('referencesProvider'):
98 actions.append("<a href='{}'>{}</a>".format('references', 'References'))
99
100 if self.has_client_with_capability('renameProvider'):
101 actions.append("<a href='{}'>{}</a>".format('rename', 'Rename'))
102
103 return "<p>" + " | ".join(actions) + "</p>"
104
105 def format_diagnostic(self, diagnostic):
106 if diagnostic.source:
107 return "<pre>[{}] {}</pre>".format(diagnostic.source, escape(diagnostic.message, False))
108 else:
109 return "<pre>{}</pre>".format(escape(diagnostic.message, False))
110
111 def diagnostics_content(self, diagnostics):
112 by_severity = {} # type: Dict[int, List[str]]
113 for diagnostic in diagnostics:
114 by_severity.setdefault(diagnostic.severity, []).append(self.format_diagnostic(diagnostic))
115 formatted = []
116 for severity, items in by_severity.items():
117 formatted.append("<div class='{}'>".format(class_for_severity[severity]))
118 formatted.extend(items)
119 formatted.append("<a href='{}'>{}</a>".format('code-actions',
120 'Code Actions'))
121 formatted.append("</div>")
122
123 return "".join(formatted)
124
125 def hover_content(self, point, response: 'Optional[Any]') -> str:
126 contents = ["No description available."]
127 if isinstance(response, dict):
128 # Flow returns None sometimes
129 # See: https://github.com/flowtype/flow-language-server/issues/51
130 response_content = response.get('contents')
131 if response_content:
132 if isinstance(response_content, list):
133 contents = response_content
134 else:
135 contents = [response_content]
136
137 formatted = []
138 for item in contents:
139 value = ""
140 language = None
141 if isinstance(item, str):
142 value = escape(item)
143 else:
144 value = escape(item.get("value"))
145 language = item.get("language")
146 if language:
147 formatted.append("```{}\n{}\n```\n".format(language, value))
148 else:
149 formatted.append(value)
150
151 return mdpopups.md2html(self.view, "\n".join(formatted))
152
153 def show_hover(self, point, contents):
154 mdpopups.show_popup(
155 self.view,
156 contents,
157 css=popup_css,
158 md=False,
159 flags=sublime.HIDE_ON_MOUSE_MOVE_AWAY,
160 location=point,
161 wrapper_class=popup_class,
162 max_width=800,
163 on_navigate=lambda href: self.on_hover_navigate(href, point))
164
165 def on_hover_navigate(self, href, point):
166 if href == 'definition':
167 self.run_command_from_point(point, "lsp_symbol_definition")
168 elif href == 'references':
169 self.run_command_from_point(point, "lsp_symbol_references")
170 elif href == 'rename':
171 self.run_command_from_point(point, "lsp_symbol_rename")
172 elif href == 'code-actions':
173 self.run_command_from_point(point, "lsp_code_actions")
174 else:
175 webbrowser.open_new_tab(href)
176
177 def run_command_from_point(self, point, command_name):
178 sel = self.view.sel()
179 sel.clear()
180 sel.add(sublime.Region(point, point))
181 self.view.run_command(command_name)
182
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/plugin/hover.py b/plugin/hover.py
--- a/plugin/hover.py
+++ b/plugin/hover.py
@@ -139,9 +139,9 @@
value = ""
language = None
if isinstance(item, str):
- value = escape(item)
+ value = item
else:
- value = escape(item.get("value"))
+ value = item.get("value")
language = item.get("language")
if language:
formatted.append("```{}\n{}\n```\n".format(language, value))
| {"golden_diff": "diff --git a/plugin/hover.py b/plugin/hover.py\n--- a/plugin/hover.py\n+++ b/plugin/hover.py\n@@ -139,9 +139,9 @@\n value = \"\"\n language = None\n if isinstance(item, str):\n- value = escape(item)\n+ value = item\n else:\n- value = escape(item.get(\"value\"))\n+ value = item.get(\"value\")\n language = item.get(\"language\")\n if language:\n formatted.append(\"```{}\\n{}\\n```\\n\".format(language, value))\n", "issue": "hover regression\nAfter the latest commits hover popups produce errors, I suspect c9a73907f00972a415e8e979d5dcc96c11ec3f09 is responsible for this\r\n* OS: Windows 10\r\n* language server: pyls\r\n* How you installed LSP: git\r\n* Minimal reproduction steps\r\n 1. open a .py file\r\n 2. hover over an function name\r\n* Log\r\n```\r\nLSP: --> textDocument/hover\r\nLSP: {'contents': \"is_supported_syntax(syntax: str, configs: 'List[ClientConfig]') -> bool\"}\r\nParse Error: #x27;List[ClientConfig]') -> bool</p><p><a href='definition'>Definition</a> | <a href='references'>References</a> | <a href='rename'>Rename</a></p></div></div> code: Unexpected character\r\nParse Error: 27;List[ClientConfig]') -> bool</p><p><a href='definition'>Definition</a> | <a href='references'>References</a> | <a href='rename'>Rename</a></p></div></div> code: Unexpected character\r\nParse Error: 7;List[ClientConfig]') -> bool</p><p><a href='definition'>Definition</a> | <a href='references'>References</a> | <a href='rename'>Rename</a></p></div></div> code: Unexpected character\r\nParse Error: #x27; code: Unknown entity\r\nParse Error: #x27;) -> bool</p><p><a href='definition'>Definition</a> | <a href='references'>References</a> | <a href='rename'>Rename</a></p></div></div> code: Unexpected character\r\nParse Error: 27;) -> bool</p><p><a href='definition'>Definition</a> | <a href='references'>References</a> | <a href='rename'>Rename</a></p></div></div> code: Unexpected character\r\nParse Error: 7;) -> bool</p><p><a href='definition'>Definition</a> | <a href='references'>References</a> | <a href='rename'>Rename</a></p></div></div> code: Unexpected character\r\nParse Error: #x27; code: Unknown entity\r\n```\n", "before_files": [{"content": "import mdpopups\nimport sublime\nimport sublime_plugin\nimport webbrowser\nfrom html import escape\ntry:\n from typing import List, Optional, Any, Dict\n assert List and Optional and Any and Dict\nexcept ImportError:\n pass\n\nfrom .core.configurations import is_supported_syntax\nfrom .diagnostics import get_point_diagnostics\nfrom .core.registry import session_for_view, LspTextCommand\nfrom .core.protocol import Request, DiagnosticSeverity\nfrom .core.documents import get_document_position\nfrom .core.popups import popup_css, popup_class\nfrom .core.settings import client_configs\n\nSUBLIME_WORD_MASK = 515\nNO_HOVER_SCOPES = 'comment, string'\n\n\nclass HoverHandler(sublime_plugin.ViewEventListener):\n def __init__(self, view):\n self.view = view\n\n @classmethod\n def is_applicable(cls, settings):\n syntax = settings.get('syntax')\n return syntax and is_supported_syntax(syntax, client_configs.all)\n\n def on_hover(self, point, hover_zone):\n if hover_zone != sublime.HOVER_TEXT or self.view.is_popup_visible():\n return\n self.view.run_command(\"lsp_hover\", {\"point\": point})\n\n\n_test_contents = [] # type: List[str]\n\n\nclass_for_severity = {\n DiagnosticSeverity.Error: 'errors',\n DiagnosticSeverity.Warning: 'warnings',\n DiagnosticSeverity.Information: 'info',\n DiagnosticSeverity.Hint: 'hints'\n}\n\n\nclass LspHoverCommand(LspTextCommand):\n def __init__(self, view):\n super().__init__(view)\n\n def is_likely_at_symbol(self, point):\n word_at_sel = self.view.classify(point)\n return word_at_sel & SUBLIME_WORD_MASK and not self.view.match_selector(point, NO_HOVER_SCOPES)\n\n def run(self, edit, point=None):\n if point is None:\n point = self.view.sel()[0].begin()\n if self.is_likely_at_symbol(point):\n self.request_symbol_hover(point)\n point_diagnostics = get_point_diagnostics(self.view, point)\n if point_diagnostics:\n self.show_hover(point, self.diagnostics_content(point_diagnostics))\n\n def request_symbol_hover(self, point) -> None:\n session = session_for_view(self.view, point)\n if session:\n if session.has_capability('hoverProvider'):\n document_position = get_document_position(self.view, point)\n if document_position:\n if session.client:\n session.client.send_request(\n Request.hover(document_position),\n lambda response: self.handle_response(response, point))\n\n def handle_response(self, response: 'Optional[Any]', point) -> None:\n all_content = \"\"\n\n point_diagnostics = get_point_diagnostics(self.view, point)\n if point_diagnostics:\n all_content += self.diagnostics_content(point_diagnostics)\n\n all_content += self.hover_content(point, response)\n all_content += self.symbol_actions_content()\n\n _test_contents.clear()\n _test_contents.append(all_content) # for testing only\n self.show_hover(point, all_content)\n\n def symbol_actions_content(self):\n actions = []\n if self.has_client_with_capability('definitionProvider'):\n actions.append(\"<a href='{}'>{}</a>\".format('definition', 'Definition'))\n\n if self.has_client_with_capability('referencesProvider'):\n actions.append(\"<a href='{}'>{}</a>\".format('references', 'References'))\n\n if self.has_client_with_capability('renameProvider'):\n actions.append(\"<a href='{}'>{}</a>\".format('rename', 'Rename'))\n\n return \"<p>\" + \" | \".join(actions) + \"</p>\"\n\n def format_diagnostic(self, diagnostic):\n if diagnostic.source:\n return \"<pre>[{}] {}</pre>\".format(diagnostic.source, escape(diagnostic.message, False))\n else:\n return \"<pre>{}</pre>\".format(escape(diagnostic.message, False))\n\n def diagnostics_content(self, diagnostics):\n by_severity = {} # type: Dict[int, List[str]]\n for diagnostic in diagnostics:\n by_severity.setdefault(diagnostic.severity, []).append(self.format_diagnostic(diagnostic))\n formatted = []\n for severity, items in by_severity.items():\n formatted.append(\"<div class='{}'>\".format(class_for_severity[severity]))\n formatted.extend(items)\n formatted.append(\"<a href='{}'>{}</a>\".format('code-actions',\n 'Code Actions'))\n formatted.append(\"</div>\")\n\n return \"\".join(formatted)\n\n def hover_content(self, point, response: 'Optional[Any]') -> str:\n contents = [\"No description available.\"]\n if isinstance(response, dict):\n # Flow returns None sometimes\n # See: https://github.com/flowtype/flow-language-server/issues/51\n response_content = response.get('contents')\n if response_content:\n if isinstance(response_content, list):\n contents = response_content\n else:\n contents = [response_content]\n\n formatted = []\n for item in contents:\n value = \"\"\n language = None\n if isinstance(item, str):\n value = escape(item)\n else:\n value = escape(item.get(\"value\"))\n language = item.get(\"language\")\n if language:\n formatted.append(\"```{}\\n{}\\n```\\n\".format(language, value))\n else:\n formatted.append(value)\n\n return mdpopups.md2html(self.view, \"\\n\".join(formatted))\n\n def show_hover(self, point, contents):\n mdpopups.show_popup(\n self.view,\n contents,\n css=popup_css,\n md=False,\n flags=sublime.HIDE_ON_MOUSE_MOVE_AWAY,\n location=point,\n wrapper_class=popup_class,\n max_width=800,\n on_navigate=lambda href: self.on_hover_navigate(href, point))\n\n def on_hover_navigate(self, href, point):\n if href == 'definition':\n self.run_command_from_point(point, \"lsp_symbol_definition\")\n elif href == 'references':\n self.run_command_from_point(point, \"lsp_symbol_references\")\n elif href == 'rename':\n self.run_command_from_point(point, \"lsp_symbol_rename\")\n elif href == 'code-actions':\n self.run_command_from_point(point, \"lsp_code_actions\")\n else:\n webbrowser.open_new_tab(href)\n\n def run_command_from_point(self, point, command_name):\n sel = self.view.sel()\n sel.clear()\n sel.add(sublime.Region(point, point))\n self.view.run_command(command_name)\n", "path": "plugin/hover.py"}], "after_files": [{"content": "import mdpopups\nimport sublime\nimport sublime_plugin\nimport webbrowser\nfrom html import escape\ntry:\n from typing import List, Optional, Any, Dict\n assert List and Optional and Any and Dict\nexcept ImportError:\n pass\n\nfrom .core.configurations import is_supported_syntax\nfrom .diagnostics import get_point_diagnostics\nfrom .core.registry import session_for_view, LspTextCommand\nfrom .core.protocol import Request, DiagnosticSeverity\nfrom .core.documents import get_document_position\nfrom .core.popups import popup_css, popup_class\nfrom .core.settings import client_configs\n\nSUBLIME_WORD_MASK = 515\nNO_HOVER_SCOPES = 'comment, string'\n\n\nclass HoverHandler(sublime_plugin.ViewEventListener):\n def __init__(self, view):\n self.view = view\n\n @classmethod\n def is_applicable(cls, settings):\n syntax = settings.get('syntax')\n return syntax and is_supported_syntax(syntax, client_configs.all)\n\n def on_hover(self, point, hover_zone):\n if hover_zone != sublime.HOVER_TEXT or self.view.is_popup_visible():\n return\n self.view.run_command(\"lsp_hover\", {\"point\": point})\n\n\n_test_contents = [] # type: List[str]\n\n\nclass_for_severity = {\n DiagnosticSeverity.Error: 'errors',\n DiagnosticSeverity.Warning: 'warnings',\n DiagnosticSeverity.Information: 'info',\n DiagnosticSeverity.Hint: 'hints'\n}\n\n\nclass LspHoverCommand(LspTextCommand):\n def __init__(self, view):\n super().__init__(view)\n\n def is_likely_at_symbol(self, point):\n word_at_sel = self.view.classify(point)\n return word_at_sel & SUBLIME_WORD_MASK and not self.view.match_selector(point, NO_HOVER_SCOPES)\n\n def run(self, edit, point=None):\n if point is None:\n point = self.view.sel()[0].begin()\n if self.is_likely_at_symbol(point):\n self.request_symbol_hover(point)\n point_diagnostics = get_point_diagnostics(self.view, point)\n if point_diagnostics:\n self.show_hover(point, self.diagnostics_content(point_diagnostics))\n\n def request_symbol_hover(self, point) -> None:\n session = session_for_view(self.view, point)\n if session:\n if session.has_capability('hoverProvider'):\n document_position = get_document_position(self.view, point)\n if document_position:\n if session.client:\n session.client.send_request(\n Request.hover(document_position),\n lambda response: self.handle_response(response, point))\n\n def handle_response(self, response: 'Optional[Any]', point) -> None:\n all_content = \"\"\n\n point_diagnostics = get_point_diagnostics(self.view, point)\n if point_diagnostics:\n all_content += self.diagnostics_content(point_diagnostics)\n\n all_content += self.hover_content(point, response)\n all_content += self.symbol_actions_content()\n\n _test_contents.clear()\n _test_contents.append(all_content) # for testing only\n self.show_hover(point, all_content)\n\n def symbol_actions_content(self):\n actions = []\n if self.has_client_with_capability('definitionProvider'):\n actions.append(\"<a href='{}'>{}</a>\".format('definition', 'Definition'))\n\n if self.has_client_with_capability('referencesProvider'):\n actions.append(\"<a href='{}'>{}</a>\".format('references', 'References'))\n\n if self.has_client_with_capability('renameProvider'):\n actions.append(\"<a href='{}'>{}</a>\".format('rename', 'Rename'))\n\n return \"<p>\" + \" | \".join(actions) + \"</p>\"\n\n def format_diagnostic(self, diagnostic):\n if diagnostic.source:\n return \"<pre>[{}] {}</pre>\".format(diagnostic.source, escape(diagnostic.message, False))\n else:\n return \"<pre>{}</pre>\".format(escape(diagnostic.message, False))\n\n def diagnostics_content(self, diagnostics):\n by_severity = {} # type: Dict[int, List[str]]\n for diagnostic in diagnostics:\n by_severity.setdefault(diagnostic.severity, []).append(self.format_diagnostic(diagnostic))\n formatted = []\n for severity, items in by_severity.items():\n formatted.append(\"<div class='{}'>\".format(class_for_severity[severity]))\n formatted.extend(items)\n formatted.append(\"<a href='{}'>{}</a>\".format('code-actions',\n 'Code Actions'))\n formatted.append(\"</div>\")\n\n return \"\".join(formatted)\n\n def hover_content(self, point, response: 'Optional[Any]') -> str:\n contents = [\"No description available.\"]\n if isinstance(response, dict):\n # Flow returns None sometimes\n # See: https://github.com/flowtype/flow-language-server/issues/51\n response_content = response.get('contents')\n if response_content:\n if isinstance(response_content, list):\n contents = response_content\n else:\n contents = [response_content]\n\n formatted = []\n for item in contents:\n value = \"\"\n language = None\n if isinstance(item, str):\n value = item\n else:\n value = item.get(\"value\")\n language = item.get(\"language\")\n if language:\n formatted.append(\"```{}\\n{}\\n```\\n\".format(language, value))\n else:\n formatted.append(value)\n\n return mdpopups.md2html(self.view, \"\\n\".join(formatted))\n\n def show_hover(self, point, contents):\n mdpopups.show_popup(\n self.view,\n contents,\n css=popup_css,\n md=False,\n flags=sublime.HIDE_ON_MOUSE_MOVE_AWAY,\n location=point,\n wrapper_class=popup_class,\n max_width=800,\n on_navigate=lambda href: self.on_hover_navigate(href, point))\n\n def on_hover_navigate(self, href, point):\n if href == 'definition':\n self.run_command_from_point(point, \"lsp_symbol_definition\")\n elif href == 'references':\n self.run_command_from_point(point, \"lsp_symbol_references\")\n elif href == 'rename':\n self.run_command_from_point(point, \"lsp_symbol_rename\")\n elif href == 'code-actions':\n self.run_command_from_point(point, \"lsp_code_actions\")\n else:\n webbrowser.open_new_tab(href)\n\n def run_command_from_point(self, point, command_name):\n sel = self.view.sel()\n sel.clear()\n sel.add(sublime.Region(point, point))\n self.view.run_command(command_name)\n", "path": "plugin/hover.py"}]} | 2,654 | 123 |
gh_patches_debug_24635 | rasdani/github-patches | git_diff | plone__Products.CMFPlone-1438 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
plone-upgrade to 5.0.3 shows plain text as result
This is caused by https://github.com/plone/plone.app.upgrade/pull/67 by @vangheem, though this change looks fine to me.
With the above change, when running `@@plone-upgrade`, even with dry-run selected, the result page is shown as text: you see plain html. Very strange. Reported here: https://community.plone.org/t/plone-5-0-3-soft-released/1699/4
When I empty the registry.xml, keeping only the main `registry` tags for safety, it all works fine. Keeping one of the two changed records, it again shows as text.
To check it:
- Use current coredev 5.0
- Create a Plone Site.
- Simulate a Plone 5.0.2 site: in portal_setup, Upgrades, select Products.CMFPlone:plone, and run the to502 upgrade profile.
- Go to @@plone-upgrade, optionally select dry-run, and run the upgrade.
Result: it will show as plain text.
BTW, afterwards, all is fine: the migration has succeeded and it looks like all pages show up fine.
Any idea?
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `Products/CMFPlone/resources/exportimport/bundles.py`
Content:
```
1 from plone.registry.interfaces import IRegistry
2 from zope.component import queryUtility
3
4 from ..browser.combine import combine_bundles
5
6
7 def combine(context):
8
9 logger = context.getLogger('bundles')
10 registry = queryUtility(IRegistry)
11
12 if registry is None:
13 logger.info("Cannot find registry")
14 return
15
16 body = context.readDataFile('registry.xml')
17 if body and "IBundleRegistry" in body:
18 site = context.getSite()
19 combine_bundles(site)
20
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/Products/CMFPlone/resources/exportimport/bundles.py b/Products/CMFPlone/resources/exportimport/bundles.py
--- a/Products/CMFPlone/resources/exportimport/bundles.py
+++ b/Products/CMFPlone/resources/exportimport/bundles.py
@@ -1,5 +1,6 @@
from plone.registry.interfaces import IRegistry
from zope.component import queryUtility
+from zope.globalrequest import getRequest
from ..browser.combine import combine_bundles
@@ -16,4 +17,20 @@
body = context.readDataFile('registry.xml')
if body and "IBundleRegistry" in body:
site = context.getSite()
+ # Calling combine_bundles will have as side effect that the
+ # Content-Type header of the response is set to application/javascript,
+ # which we do not want. So we reset it to the original at the end.
+ site = context.getSite()
+ request = getattr(site, 'REQUEST', getRequest())
+ if request is not None:
+ # Easily happens in tests.
+ orig_header = request.response.getHeader('Content-Type')
combine_bundles(site)
+ if request is not None:
+ new_header = request.response.getHeader('Content-Type')
+ if new_header != orig_header:
+ if orig_header is None:
+ # Setting it to None would result in the string 'None'.
+ # So pick a saner one.
+ orig_header = 'text/html'
+ request.response.setHeader('Content-Type', orig_header)
| {"golden_diff": "diff --git a/Products/CMFPlone/resources/exportimport/bundles.py b/Products/CMFPlone/resources/exportimport/bundles.py\n--- a/Products/CMFPlone/resources/exportimport/bundles.py\n+++ b/Products/CMFPlone/resources/exportimport/bundles.py\n@@ -1,5 +1,6 @@\n from plone.registry.interfaces import IRegistry\n from zope.component import queryUtility\n+from zope.globalrequest import getRequest\n \n from ..browser.combine import combine_bundles\n \n@@ -16,4 +17,20 @@\n body = context.readDataFile('registry.xml')\n if body and \"IBundleRegistry\" in body:\n site = context.getSite()\n+ # Calling combine_bundles will have as side effect that the\n+ # Content-Type header of the response is set to application/javascript,\n+ # which we do not want. So we reset it to the original at the end.\n+ site = context.getSite()\n+ request = getattr(site, 'REQUEST', getRequest())\n+ if request is not None:\n+ # Easily happens in tests.\n+ orig_header = request.response.getHeader('Content-Type')\n combine_bundles(site)\n+ if request is not None:\n+ new_header = request.response.getHeader('Content-Type')\n+ if new_header != orig_header:\n+ if orig_header is None:\n+ # Setting it to None would result in the string 'None'.\n+ # So pick a saner one.\n+ orig_header = 'text/html'\n+ request.response.setHeader('Content-Type', orig_header)\n", "issue": "plone-upgrade to 5.0.3 shows plain text as result\nThis is caused by https://github.com/plone/plone.app.upgrade/pull/67 by @vangheem, though this change looks fine to me.\n\nWith the above change, when running `@@plone-upgrade`, even with dry-run selected, the result page is shown as text: you see plain html. Very strange. Reported here: https://community.plone.org/t/plone-5-0-3-soft-released/1699/4\nWhen I empty the registry.xml, keeping only the main `registry` tags for safety, it all works fine. Keeping one of the two changed records, it again shows as text.\n\nTo check it:\n- Use current coredev 5.0\n- Create a Plone Site.\n- Simulate a Plone 5.0.2 site: in portal_setup, Upgrades, select Products.CMFPlone:plone, and run the to502 upgrade profile.\n- Go to @@plone-upgrade, optionally select dry-run, and run the upgrade.\n\nResult: it will show as plain text.\nBTW, afterwards, all is fine: the migration has succeeded and it looks like all pages show up fine.\n\nAny idea?\n\n", "before_files": [{"content": "from plone.registry.interfaces import IRegistry\nfrom zope.component import queryUtility\n\nfrom ..browser.combine import combine_bundles\n\n\ndef combine(context):\n\n logger = context.getLogger('bundles')\n registry = queryUtility(IRegistry)\n\n if registry is None:\n logger.info(\"Cannot find registry\")\n return\n\n body = context.readDataFile('registry.xml')\n if body and \"IBundleRegistry\" in body:\n site = context.getSite()\n combine_bundles(site)\n", "path": "Products/CMFPlone/resources/exportimport/bundles.py"}], "after_files": [{"content": "from plone.registry.interfaces import IRegistry\nfrom zope.component import queryUtility\nfrom zope.globalrequest import getRequest\n\nfrom ..browser.combine import combine_bundles\n\n\ndef combine(context):\n\n logger = context.getLogger('bundles')\n registry = queryUtility(IRegistry)\n\n if registry is None:\n logger.info(\"Cannot find registry\")\n return\n\n body = context.readDataFile('registry.xml')\n if body and \"IBundleRegistry\" in body:\n site = context.getSite()\n # Calling combine_bundles will have as side effect that the\n # Content-Type header of the response is set to application/javascript,\n # which we do not want. So we reset it to the original at the end.\n site = context.getSite()\n request = getattr(site, 'REQUEST', getRequest())\n if request is not None:\n # Easily happens in tests.\n orig_header = request.response.getHeader('Content-Type')\n combine_bundles(site)\n if request is not None:\n new_header = request.response.getHeader('Content-Type')\n if new_header != orig_header:\n if orig_header is None:\n # Setting it to None would result in the string 'None'.\n # So pick a saner one.\n orig_header = 'text/html'\n request.response.setHeader('Content-Type', orig_header)\n", "path": "Products/CMFPlone/resources/exportimport/bundles.py"}]} | 672 | 340 |
gh_patches_debug_19129 | rasdani/github-patches | git_diff | DataDog__dd-trace-py-3799 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Starlette/Fastapi: endpoint duration includes the duration of background tasks
### Which version of dd-trace-py are you using?
ddtrace==0.55.4
### Which version of pip are you using?
21.2.4
### Which version of the libraries are you using?
fastapi==0.68.2
starlette==0.14.2
### How can we reproduce your problem?
this would be a minimal proof of concept `app.py`, running through `ddtrace-run uvicorn app:app`
```
import asyncio
from ddtrace import tracer
from fastapi import FastAPI, BackgroundTasks
app = FastAPI()
async def some_background_task():
with tracer.start_span("some_background_task", activate=True):
tracer.context_provider.activate(None)
await asyncio.sleep(10)
@app.get("/")
async def main(background_tasks: BackgroundTasks):
background_tasks.add_task(some_background_task)
return "ok"
```
### What is the result that you get?
The duration of `/` is reported to be 10s, while the browser immediately receives the response.
`some_background_task` is also reported with a duration of 10s.
### What is the result that you expected?
I would expect that the reported endpoint duration matches the time it took to get the response, and that the background task is reported separately. Please don't mind that `tracer.context_provider.activate(None)` might be redundant here, adding it here to show what I have tried.
FastAPI's `add_task` actually comes from starlette https://www.starlette.io/background/
I can understand why the endpoint duration includes the background task, this is the definition of starlette's `Response.__call__`:
https://github.com/encode/starlette/blob/ada99beee530e7b841ce320bc6e66f6dbd9ad781/starlette/responses.py#L159
```
async def __call__(self, scope: Scope, receive: Receive, send: Send) -> None:
await send(
{
"type": "http.response.start",
"status": self.status_code,
"headers": self.raw_headers,
}
)
await send({"type": "http.response.body", "body": self.body})
if self.background is not None:
await self.background()
```
The response header and body is sent, but the function itself is not finished until all background tasks have been processed.
I believe that this is not what users of ddtrace would expect: the background tasks are used to return a response early without waiting for background operations to finish ; the reported endpoint duration should correspond to when the body was sent
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `ddtrace/contrib/asgi/middleware.py`
Content:
```
1 import sys
2 from typing import TYPE_CHECKING
3
4 import ddtrace
5 from ddtrace import config
6 from ddtrace.constants import ANALYTICS_SAMPLE_RATE_KEY
7 from ddtrace.ext import SpanTypes
8 from ddtrace.ext import http
9
10 from .. import trace_utils
11 from ...internal.compat import reraise
12 from ...internal.logger import get_logger
13 from .utils import guarantee_single_callable
14
15
16 if TYPE_CHECKING:
17 from typing import Any
18 from typing import Mapping
19 from typing import Optional
20
21 from ddtrace import Span
22
23
24 log = get_logger(__name__)
25
26 config._add(
27 "asgi",
28 dict(service_name=config._get_service(default="asgi"), request_span_name="asgi.request", distributed_tracing=True),
29 )
30
31 ASGI_VERSION = "asgi.version"
32 ASGI_SPEC_VERSION = "asgi.spec_version"
33
34
35 def bytes_to_str(str_or_bytes):
36 return str_or_bytes.decode() if isinstance(str_or_bytes, bytes) else str_or_bytes
37
38
39 def _extract_versions_from_scope(scope, integration_config):
40 tags = {}
41
42 http_version = scope.get("http_version")
43 if http_version:
44 tags[http.VERSION] = http_version
45
46 scope_asgi = scope.get("asgi")
47
48 if scope_asgi and "version" in scope_asgi:
49 tags[ASGI_VERSION] = scope_asgi["version"]
50
51 if scope_asgi and "spec_version" in scope_asgi:
52 tags[ASGI_SPEC_VERSION] = scope_asgi["spec_version"]
53
54 return tags
55
56
57 def _extract_headers(scope):
58 headers = scope.get("headers")
59 if headers:
60 # headers: (Iterable[[byte string, byte string]])
61 return dict((bytes_to_str(k), bytes_to_str(v)) for (k, v) in headers)
62 return {}
63
64
65 def _default_handle_exception_span(exc, span):
66 """Default handler for exception for span"""
67 span.set_tag(http.STATUS_CODE, 500)
68
69
70 def span_from_scope(scope):
71 # type: (Mapping[str, Any]) -> Optional[Span]
72 """Retrieve the top-level ASGI span from the scope."""
73 return scope.get("datadog", {}).get("request_spans", [None])[0]
74
75
76 class TraceMiddleware:
77 """
78 ASGI application middleware that traces the requests.
79 Args:
80 app: The ASGI application.
81 tracer: Custom tracer. Defaults to the global tracer.
82 """
83
84 def __init__(
85 self,
86 app,
87 tracer=None,
88 integration_config=config.asgi,
89 handle_exception_span=_default_handle_exception_span,
90 span_modifier=None,
91 ):
92 self.app = guarantee_single_callable(app)
93 self.tracer = tracer or ddtrace.tracer
94 self.integration_config = integration_config
95 self.handle_exception_span = handle_exception_span
96 self.span_modifier = span_modifier
97
98 async def __call__(self, scope, receive, send):
99 if scope["type"] != "http":
100 return await self.app(scope, receive, send)
101
102 try:
103 headers = _extract_headers(scope)
104 except Exception:
105 log.warning("failed to decode headers for distributed tracing", exc_info=True)
106 headers = {}
107 else:
108 trace_utils.activate_distributed_headers(
109 self.tracer, int_config=self.integration_config, request_headers=headers
110 )
111
112 resource = "{} {}".format(scope["method"], scope["path"])
113
114 span = self.tracer.trace(
115 name=self.integration_config.get("request_span_name", "asgi.request"),
116 service=trace_utils.int_service(None, self.integration_config),
117 resource=resource,
118 span_type=SpanTypes.WEB,
119 )
120
121 if "datadog" not in scope:
122 scope["datadog"] = {"request_spans": [span]}
123 else:
124 scope["datadog"]["request_spans"].append(span)
125
126 if self.span_modifier:
127 self.span_modifier(span, scope)
128
129 sample_rate = self.integration_config.get_analytics_sample_rate(use_global_config=True)
130 if sample_rate is not None:
131 span.set_tag(ANALYTICS_SAMPLE_RATE_KEY, sample_rate)
132
133 method = scope.get("method")
134 server = scope.get("server")
135 if server and len(server) == 2:
136 port = server[1]
137 server_host = server[0] + (":" + str(port) if port is not None and port != 80 else "")
138 full_path = scope.get("root_path", "") + scope.get("path", "")
139 url = scope.get("scheme", "http") + "://" + server_host + full_path
140 else:
141 url = None
142
143 if self.integration_config.trace_query_string:
144 query_string = scope.get("query_string")
145 if len(query_string) > 0:
146 query_string = bytes_to_str(query_string)
147 else:
148 query_string = None
149
150 trace_utils.set_http_meta(
151 span, self.integration_config, method=method, url=url, query=query_string, request_headers=headers
152 )
153
154 tags = _extract_versions_from_scope(scope, self.integration_config)
155 span.set_tags(tags)
156
157 async def wrapped_send(message):
158 if span and message.get("type") == "http.response.start" and "status" in message:
159 status_code = message["status"]
160 else:
161 status_code = None
162
163 if "headers" in message:
164 response_headers = message["headers"]
165 else:
166 response_headers = None
167
168 trace_utils.set_http_meta(
169 span, self.integration_config, status_code=status_code, response_headers=response_headers
170 )
171
172 return await send(message)
173
174 try:
175 return await self.app(scope, receive, wrapped_send)
176 except Exception as exc:
177 (exc_type, exc_val, exc_tb) = sys.exc_info()
178 span.set_exc_info(exc_type, exc_val, exc_tb)
179 self.handle_exception_span(exc, span)
180 reraise(exc_type, exc_val, exc_tb)
181 finally:
182 try:
183 del scope["datadog"]["request_span"]
184 except KeyError:
185 pass
186 span.finish()
187
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/ddtrace/contrib/asgi/middleware.py b/ddtrace/contrib/asgi/middleware.py
--- a/ddtrace/contrib/asgi/middleware.py
+++ b/ddtrace/contrib/asgi/middleware.py
@@ -169,7 +169,13 @@
span, self.integration_config, status_code=status_code, response_headers=response_headers
)
- return await send(message)
+ try:
+ return await send(message)
+ finally:
+ # Per asgi spec, "more_body" is used if there is still data to send
+ # Close the span if "http.response.body" has no more data left to send in the response.
+ if message.get("type") == "http.response.body" and not message.get("more_body", False):
+ span.finish()
try:
return await self.app(scope, receive, wrapped_send)
@@ -183,4 +189,5 @@
del scope["datadog"]["request_span"]
except KeyError:
pass
+
span.finish()
| {"golden_diff": "diff --git a/ddtrace/contrib/asgi/middleware.py b/ddtrace/contrib/asgi/middleware.py\n--- a/ddtrace/contrib/asgi/middleware.py\n+++ b/ddtrace/contrib/asgi/middleware.py\n@@ -169,7 +169,13 @@\n span, self.integration_config, status_code=status_code, response_headers=response_headers\n )\n \n- return await send(message)\n+ try:\n+ return await send(message)\n+ finally:\n+ # Per asgi spec, \"more_body\" is used if there is still data to send\n+ # Close the span if \"http.response.body\" has no more data left to send in the response.\n+ if message.get(\"type\") == \"http.response.body\" and not message.get(\"more_body\", False):\n+ span.finish()\n \n try:\n return await self.app(scope, receive, wrapped_send)\n@@ -183,4 +189,5 @@\n del scope[\"datadog\"][\"request_span\"]\n except KeyError:\n pass\n+\n span.finish()\n", "issue": "Starlette/Fastapi: endpoint duration includes the duration of background tasks\n### Which version of dd-trace-py are you using?\r\n\r\nddtrace==0.55.4\r\n\r\n### Which version of pip are you using?\r\n\r\n21.2.4\r\n\r\n\r\n### Which version of the libraries are you using?\r\n\r\nfastapi==0.68.2\r\nstarlette==0.14.2\r\n\r\n### How can we reproduce your problem?\r\n\r\nthis would be a minimal proof of concept `app.py`, running through `ddtrace-run uvicorn app:app`\r\n\r\n```\r\nimport asyncio\r\n\r\nfrom ddtrace import tracer\r\nfrom fastapi import FastAPI, BackgroundTasks\r\n\r\napp = FastAPI()\r\n\r\n\r\nasync def some_background_task():\r\n with tracer.start_span(\"some_background_task\", activate=True):\r\n tracer.context_provider.activate(None)\r\n await asyncio.sleep(10)\r\n\r\n\r\[email protected](\"/\")\r\nasync def main(background_tasks: BackgroundTasks):\r\n background_tasks.add_task(some_background_task)\r\n return \"ok\"\r\n\r\n```\r\n\r\n### What is the result that you get?\r\n\r\nThe duration of `/` is reported to be 10s, while the browser immediately receives the response.\r\n`some_background_task` is also reported with a duration of 10s.\r\n\r\n### What is the result that you expected?\r\n\r\nI would expect that the reported endpoint duration matches the time it took to get the response, and that the background task is reported separately. Please don't mind that `tracer.context_provider.activate(None)` might be redundant here, adding it here to show what I have tried.\r\n\r\nFastAPI's `add_task` actually comes from starlette https://www.starlette.io/background/\r\n\r\nI can understand why the endpoint duration includes the background task, this is the definition of starlette's `Response.__call__`:\r\n\r\nhttps://github.com/encode/starlette/blob/ada99beee530e7b841ce320bc6e66f6dbd9ad781/starlette/responses.py#L159\r\n```\r\n async def __call__(self, scope: Scope, receive: Receive, send: Send) -> None:\r\n await send(\r\n {\r\n \"type\": \"http.response.start\",\r\n \"status\": self.status_code,\r\n \"headers\": self.raw_headers,\r\n }\r\n )\r\n await send({\"type\": \"http.response.body\", \"body\": self.body})\r\n\r\n if self.background is not None:\r\n await self.background()\r\n```\r\n\r\nThe response header and body is sent, but the function itself is not finished until all background tasks have been processed.\r\n\r\nI believe that this is not what users of ddtrace would expect: the background tasks are used to return a response early without waiting for background operations to finish ; the reported endpoint duration should correspond to when the body was sent\r\n\n", "before_files": [{"content": "import sys\nfrom typing import TYPE_CHECKING\n\nimport ddtrace\nfrom ddtrace import config\nfrom ddtrace.constants import ANALYTICS_SAMPLE_RATE_KEY\nfrom ddtrace.ext import SpanTypes\nfrom ddtrace.ext import http\n\nfrom .. import trace_utils\nfrom ...internal.compat import reraise\nfrom ...internal.logger import get_logger\nfrom .utils import guarantee_single_callable\n\n\nif TYPE_CHECKING:\n from typing import Any\n from typing import Mapping\n from typing import Optional\n\n from ddtrace import Span\n\n\nlog = get_logger(__name__)\n\nconfig._add(\n \"asgi\",\n dict(service_name=config._get_service(default=\"asgi\"), request_span_name=\"asgi.request\", distributed_tracing=True),\n)\n\nASGI_VERSION = \"asgi.version\"\nASGI_SPEC_VERSION = \"asgi.spec_version\"\n\n\ndef bytes_to_str(str_or_bytes):\n return str_or_bytes.decode() if isinstance(str_or_bytes, bytes) else str_or_bytes\n\n\ndef _extract_versions_from_scope(scope, integration_config):\n tags = {}\n\n http_version = scope.get(\"http_version\")\n if http_version:\n tags[http.VERSION] = http_version\n\n scope_asgi = scope.get(\"asgi\")\n\n if scope_asgi and \"version\" in scope_asgi:\n tags[ASGI_VERSION] = scope_asgi[\"version\"]\n\n if scope_asgi and \"spec_version\" in scope_asgi:\n tags[ASGI_SPEC_VERSION] = scope_asgi[\"spec_version\"]\n\n return tags\n\n\ndef _extract_headers(scope):\n headers = scope.get(\"headers\")\n if headers:\n # headers: (Iterable[[byte string, byte string]])\n return dict((bytes_to_str(k), bytes_to_str(v)) for (k, v) in headers)\n return {}\n\n\ndef _default_handle_exception_span(exc, span):\n \"\"\"Default handler for exception for span\"\"\"\n span.set_tag(http.STATUS_CODE, 500)\n\n\ndef span_from_scope(scope):\n # type: (Mapping[str, Any]) -> Optional[Span]\n \"\"\"Retrieve the top-level ASGI span from the scope.\"\"\"\n return scope.get(\"datadog\", {}).get(\"request_spans\", [None])[0]\n\n\nclass TraceMiddleware:\n \"\"\"\n ASGI application middleware that traces the requests.\n Args:\n app: The ASGI application.\n tracer: Custom tracer. Defaults to the global tracer.\n \"\"\"\n\n def __init__(\n self,\n app,\n tracer=None,\n integration_config=config.asgi,\n handle_exception_span=_default_handle_exception_span,\n span_modifier=None,\n ):\n self.app = guarantee_single_callable(app)\n self.tracer = tracer or ddtrace.tracer\n self.integration_config = integration_config\n self.handle_exception_span = handle_exception_span\n self.span_modifier = span_modifier\n\n async def __call__(self, scope, receive, send):\n if scope[\"type\"] != \"http\":\n return await self.app(scope, receive, send)\n\n try:\n headers = _extract_headers(scope)\n except Exception:\n log.warning(\"failed to decode headers for distributed tracing\", exc_info=True)\n headers = {}\n else:\n trace_utils.activate_distributed_headers(\n self.tracer, int_config=self.integration_config, request_headers=headers\n )\n\n resource = \"{} {}\".format(scope[\"method\"], scope[\"path\"])\n\n span = self.tracer.trace(\n name=self.integration_config.get(\"request_span_name\", \"asgi.request\"),\n service=trace_utils.int_service(None, self.integration_config),\n resource=resource,\n span_type=SpanTypes.WEB,\n )\n\n if \"datadog\" not in scope:\n scope[\"datadog\"] = {\"request_spans\": [span]}\n else:\n scope[\"datadog\"][\"request_spans\"].append(span)\n\n if self.span_modifier:\n self.span_modifier(span, scope)\n\n sample_rate = self.integration_config.get_analytics_sample_rate(use_global_config=True)\n if sample_rate is not None:\n span.set_tag(ANALYTICS_SAMPLE_RATE_KEY, sample_rate)\n\n method = scope.get(\"method\")\n server = scope.get(\"server\")\n if server and len(server) == 2:\n port = server[1]\n server_host = server[0] + (\":\" + str(port) if port is not None and port != 80 else \"\")\n full_path = scope.get(\"root_path\", \"\") + scope.get(\"path\", \"\")\n url = scope.get(\"scheme\", \"http\") + \"://\" + server_host + full_path\n else:\n url = None\n\n if self.integration_config.trace_query_string:\n query_string = scope.get(\"query_string\")\n if len(query_string) > 0:\n query_string = bytes_to_str(query_string)\n else:\n query_string = None\n\n trace_utils.set_http_meta(\n span, self.integration_config, method=method, url=url, query=query_string, request_headers=headers\n )\n\n tags = _extract_versions_from_scope(scope, self.integration_config)\n span.set_tags(tags)\n\n async def wrapped_send(message):\n if span and message.get(\"type\") == \"http.response.start\" and \"status\" in message:\n status_code = message[\"status\"]\n else:\n status_code = None\n\n if \"headers\" in message:\n response_headers = message[\"headers\"]\n else:\n response_headers = None\n\n trace_utils.set_http_meta(\n span, self.integration_config, status_code=status_code, response_headers=response_headers\n )\n\n return await send(message)\n\n try:\n return await self.app(scope, receive, wrapped_send)\n except Exception as exc:\n (exc_type, exc_val, exc_tb) = sys.exc_info()\n span.set_exc_info(exc_type, exc_val, exc_tb)\n self.handle_exception_span(exc, span)\n reraise(exc_type, exc_val, exc_tb)\n finally:\n try:\n del scope[\"datadog\"][\"request_span\"]\n except KeyError:\n pass\n span.finish()\n", "path": "ddtrace/contrib/asgi/middleware.py"}], "after_files": [{"content": "import sys\nfrom typing import TYPE_CHECKING\n\nimport ddtrace\nfrom ddtrace import config\nfrom ddtrace.constants import ANALYTICS_SAMPLE_RATE_KEY\nfrom ddtrace.ext import SpanTypes\nfrom ddtrace.ext import http\n\nfrom .. import trace_utils\nfrom ...internal.compat import reraise\nfrom ...internal.logger import get_logger\nfrom .utils import guarantee_single_callable\n\n\nif TYPE_CHECKING:\n from typing import Any\n from typing import Mapping\n from typing import Optional\n\n from ddtrace import Span\n\n\nlog = get_logger(__name__)\n\nconfig._add(\n \"asgi\",\n dict(service_name=config._get_service(default=\"asgi\"), request_span_name=\"asgi.request\", distributed_tracing=True),\n)\n\nASGI_VERSION = \"asgi.version\"\nASGI_SPEC_VERSION = \"asgi.spec_version\"\n\n\ndef bytes_to_str(str_or_bytes):\n return str_or_bytes.decode() if isinstance(str_or_bytes, bytes) else str_or_bytes\n\n\ndef _extract_versions_from_scope(scope, integration_config):\n tags = {}\n\n http_version = scope.get(\"http_version\")\n if http_version:\n tags[http.VERSION] = http_version\n\n scope_asgi = scope.get(\"asgi\")\n\n if scope_asgi and \"version\" in scope_asgi:\n tags[ASGI_VERSION] = scope_asgi[\"version\"]\n\n if scope_asgi and \"spec_version\" in scope_asgi:\n tags[ASGI_SPEC_VERSION] = scope_asgi[\"spec_version\"]\n\n return tags\n\n\ndef _extract_headers(scope):\n headers = scope.get(\"headers\")\n if headers:\n # headers: (Iterable[[byte string, byte string]])\n return dict((bytes_to_str(k), bytes_to_str(v)) for (k, v) in headers)\n return {}\n\n\ndef _default_handle_exception_span(exc, span):\n \"\"\"Default handler for exception for span\"\"\"\n span.set_tag(http.STATUS_CODE, 500)\n\n\ndef span_from_scope(scope):\n # type: (Mapping[str, Any]) -> Optional[Span]\n \"\"\"Retrieve the top-level ASGI span from the scope.\"\"\"\n return scope.get(\"datadog\", {}).get(\"request_spans\", [None])[0]\n\n\nclass TraceMiddleware:\n \"\"\"\n ASGI application middleware that traces the requests.\n Args:\n app: The ASGI application.\n tracer: Custom tracer. Defaults to the global tracer.\n \"\"\"\n\n def __init__(\n self,\n app,\n tracer=None,\n integration_config=config.asgi,\n handle_exception_span=_default_handle_exception_span,\n span_modifier=None,\n ):\n self.app = guarantee_single_callable(app)\n self.tracer = tracer or ddtrace.tracer\n self.integration_config = integration_config\n self.handle_exception_span = handle_exception_span\n self.span_modifier = span_modifier\n\n async def __call__(self, scope, receive, send):\n if scope[\"type\"] != \"http\":\n return await self.app(scope, receive, send)\n\n try:\n headers = _extract_headers(scope)\n except Exception:\n log.warning(\"failed to decode headers for distributed tracing\", exc_info=True)\n headers = {}\n else:\n trace_utils.activate_distributed_headers(\n self.tracer, int_config=self.integration_config, request_headers=headers\n )\n\n resource = \"{} {}\".format(scope[\"method\"], scope[\"path\"])\n\n span = self.tracer.trace(\n name=self.integration_config.get(\"request_span_name\", \"asgi.request\"),\n service=trace_utils.int_service(None, self.integration_config),\n resource=resource,\n span_type=SpanTypes.WEB,\n )\n\n if \"datadog\" not in scope:\n scope[\"datadog\"] = {\"request_spans\": [span]}\n else:\n scope[\"datadog\"][\"request_spans\"].append(span)\n\n if self.span_modifier:\n self.span_modifier(span, scope)\n\n sample_rate = self.integration_config.get_analytics_sample_rate(use_global_config=True)\n if sample_rate is not None:\n span.set_tag(ANALYTICS_SAMPLE_RATE_KEY, sample_rate)\n\n method = scope.get(\"method\")\n server = scope.get(\"server\")\n if server and len(server) == 2:\n port = server[1]\n server_host = server[0] + (\":\" + str(port) if port is not None and port != 80 else \"\")\n full_path = scope.get(\"root_path\", \"\") + scope.get(\"path\", \"\")\n url = scope.get(\"scheme\", \"http\") + \"://\" + server_host + full_path\n else:\n url = None\n\n if self.integration_config.trace_query_string:\n query_string = scope.get(\"query_string\")\n if len(query_string) > 0:\n query_string = bytes_to_str(query_string)\n else:\n query_string = None\n\n trace_utils.set_http_meta(\n span, self.integration_config, method=method, url=url, query=query_string, request_headers=headers\n )\n\n tags = _extract_versions_from_scope(scope, self.integration_config)\n span.set_tags(tags)\n\n async def wrapped_send(message):\n if span and message.get(\"type\") == \"http.response.start\" and \"status\" in message:\n status_code = message[\"status\"]\n else:\n status_code = None\n\n if \"headers\" in message:\n response_headers = message[\"headers\"]\n else:\n response_headers = None\n\n trace_utils.set_http_meta(\n span, self.integration_config, status_code=status_code, response_headers=response_headers\n )\n\n try:\n return await send(message)\n finally:\n # Per asgi spec, \"more_body\" is used if there is still data to send\n # Close the span if \"http.response.body\" has no more data left to send in the response.\n if message.get(\"type\") == \"http.response.body\" and not message.get(\"more_body\", False):\n span.finish()\n\n try:\n return await self.app(scope, receive, wrapped_send)\n except Exception as exc:\n (exc_type, exc_val, exc_tb) = sys.exc_info()\n span.set_exc_info(exc_type, exc_val, exc_tb)\n self.handle_exception_span(exc, span)\n reraise(exc_type, exc_val, exc_tb)\n finally:\n try:\n del scope[\"datadog\"][\"request_span\"]\n except KeyError:\n pass\n\n span.finish()\n", "path": "ddtrace/contrib/asgi/middleware.py"}]} | 2,604 | 232 |
gh_patches_debug_28087 | rasdani/github-patches | git_diff | sktime__sktime-1493 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
[BUG] ValueError when fitting Prophet and X.index.name=='ds'
**Describe the bug**
The following error is raised when trying to fit Prophet using X where the name of X's DateTimeIndex is named "ds":
ValueError: 'ds' is both an index level and a column label, which is ambiguous.
It happens in the _merge_X method in the _ProphetAdapter, line 193, because after running line 191, X contains a column and an index called "ds".
A suggested solution would be to add this oneliner before line 189 of .forecasting.base.adapters._fbprophet.py:
```python
X.index.name = "index"
```
Note: This assumes that X never has a MultiIndex. Not sure if that ever happens though..
This way, the user is still allowed to call X's index "ds". Since "ds" is the standard name that Prophet users use for their DateTime column, this is the most intuitive and least restrictive solution.
**To Reproduce**
```python
import pandas as pd
from sktime.datasets import load_airline
from sktime.forecasting.fbprophet import Prophet
from sktime.forecasting.model_selection import ForecastingGridSearchCV
from sktime.forecasting.model_selection import ExpandingWindowSplitter
y_train = load_airline().to_timestamp(freq='M')
# Commenting out the following line makes everything run smoothly
y_train.index.name = 'ds'
X_train = pd.DataFrame({'x1':y_train+200, 'x2':y_train+100})
forecaster = Prophet(yearly_seasonality=False, weekly_seasonality=False, daily_seasonality=False)
forecaster.fit(y_train, X_train)
```
**Expected behavior**
No error
**Additional context**
**Versions**
<details>
System:
python: 3.9.7 | packaged by conda-forge | (default, Sep 23 2021, 07:24:41) [MSC v.1916 64 bit (AMD64)]
executable: C:\Users\............\env\python.exe
machine: Windows-10-10.0.19043-SP0
Python dependencies:
pip: 21.2.4
setuptools: 58.1.0
sklearn: 1.0
sktime: 0.7.0
statsmodels: 0.12.2
numpy: 1.21.2
scipy: 1.7.1
Cython: 0.29.24
pandas: 1.3.3
matplotlib: 3.4.3
joblib: 1.0.1
numba: 0.53.1
pmdarima: 1.8.3
tsfresh: None
</details>
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `sktime/utils/validation/series.py`
Content:
```
1 #!/usr/bin/env python3 -u
2 # -*- coding: utf-8 -*-
3
4 """Functions for checking input data."""
5
6 __author__ = ["Markus Löning", "Drishti Bhasin"]
7 __all__ = [
8 "check_series",
9 "check_time_index",
10 "check_equal_time_index",
11 "check_consistent_index_type",
12 ]
13 import pandas as pd
14 import numpy as np
15
16 # We currently support the following types for input data and time index types.
17 VALID_DATA_TYPES = (pd.DataFrame, pd.Series, np.ndarray)
18 VALID_INDEX_TYPES = (pd.Int64Index, pd.RangeIndex, pd.PeriodIndex, pd.DatetimeIndex)
19
20
21 def _check_is_univariate(y, var_name="input"):
22 """Check if series is univariate."""
23 if isinstance(y, pd.DataFrame):
24 nvars = y.shape[1]
25 if nvars > 1:
26 raise ValueError(
27 f"{var_name} must be univariate, but found {nvars} variables."
28 )
29 if isinstance(y, np.ndarray) and y.ndim > 1 and y.shape[1] > 1:
30 raise ValueError(
31 f"{var_name} must be univariate, but found np.ndarray with more than "
32 "one column"
33 )
34
35
36 def _check_is_multivariate(Z, var_name="input"):
37 """Check if series is multivariate."""
38 if isinstance(Z, pd.Series):
39 raise ValueError(f"{var_name} must have 2 or more variables, but found 1.")
40 if isinstance(Z, pd.DataFrame):
41 nvars = Z.shape[1]
42 if nvars < 2:
43 raise ValueError(
44 f"{var_name} must have 2 or more variables, but found {nvars}."
45 )
46 if isinstance(Z, np.ndarray):
47 if Z.ndim == 1 or (Z.ndim == 2 and Z.shape[1] == 1):
48 raise ValueError(f"{var_name} must have 2 or more variables, but found 1.")
49
50
51 def check_series(
52 Z,
53 enforce_univariate=False,
54 enforce_multivariate=False,
55 allow_empty=False,
56 allow_numpy=True,
57 allow_None=True,
58 enforce_index_type=None,
59 var_name="input",
60 ):
61 """Validate input data to be a valid mtype for Series.
62
63 Parameters
64 ----------
65 Z : pd.Series, pd.DataFrame, np.ndarray, or None
66 Univariate or multivariate time series.
67 enforce_univariate : bool, default = False
68 If True, multivariate Z will raise an error.
69 enforce_multivariate: bool, default = False
70 If True, univariate Z will raise an error.
71 allow_empty : bool, default = False
72 whether a container with zero samples is allowed
73 allow_numpy : bool, default = True
74 whether no error is raised if Z is in a valid numpy.ndarray format
75 allow_None : bool, default = True
76 whether no error is raised if Z is None
77 enforce_index_type : type, default = None
78 type of time index
79 var_name : str, default = "input" - variable name printed in error messages
80
81 Returns
82 -------
83 Z : pd.Series, pd.DataFrame, np.ndarray, or None
84 Validated time series - a reference to the input Z
85
86 Raises
87 ------
88 TypeError - if Z is not in a valid type or format for scitype Series
89 if enforce_univariate is True:
90 ValueError if Z has 2 or more columns
91 if enforce_multivariate is True:
92 ValueError if Z has 1 column
93 if allow_numpy is false:
94 TypeError - if Z is of type np.ndarray
95 if allow_empty is false:
96 ValueError - if Z has length 0
97 if allow_None is false:
98 ValueError - if Z is None
99 if enforce_index_type is not None and Z is pandas type:
100 ValueError - if Z has index type other than enforce_index_type
101 """
102 if Z is None:
103 if allow_None:
104 return Z
105 else:
106 raise ValueError(var_name + " cannot be None")
107
108 # Check if pandas series or numpy array
109 if not allow_numpy:
110 valid_data_types = tuple(
111 filter(lambda x: x is not np.ndarray, VALID_DATA_TYPES)
112 )
113 else:
114 valid_data_types = VALID_DATA_TYPES
115
116 if not isinstance(Z, valid_data_types):
117 raise TypeError(
118 f"{var_name} must be a one of {valid_data_types}, but found type: {type(Z)}"
119 )
120
121 if enforce_univariate and enforce_multivariate:
122 raise ValueError(
123 "`enforce_univariate` and `enforce_multivariate` cannot both be set to "
124 "True."
125 )
126
127 if enforce_univariate:
128 _check_is_univariate(Z, var_name=var_name)
129
130 if enforce_multivariate:
131 _check_is_multivariate(Z, var_name=var_name)
132
133 # check time index if input data is not an NumPy ndarray
134 if not isinstance(Z, np.ndarray):
135 check_time_index(
136 Z.index,
137 allow_empty=allow_empty,
138 enforce_index_type=enforce_index_type,
139 var_name=var_name,
140 )
141
142 return Z
143
144
145 def check_time_index(
146 index, allow_empty=False, enforce_index_type=None, var_name="input"
147 ):
148 """Check time index.
149
150 Parameters
151 ----------
152 index : pd.Index or np.array
153 Time index
154 allow_empty : bool, optional (default=False)
155 If False, empty `index` raises an error.
156 enforce_index_type : type, optional (default=None)
157 type of time index
158 var_name : str, default = "input" - variable name printed in error messages
159
160 Returns
161 -------
162 time_index : pd.Index
163 Validated time index - a reference to the input index
164 """
165 if isinstance(index, np.ndarray):
166 index = pd.Index(index)
167
168 # We here check for type equality because isinstance does not
169 # work reliably because index types inherit from each other.
170 if not type(index) in VALID_INDEX_TYPES:
171 raise NotImplementedError(
172 f"{type(index)} is not supported for {var_name}, use "
173 f"one of {VALID_INDEX_TYPES} instead."
174 )
175
176 if enforce_index_type and type(index) is not enforce_index_type:
177 raise NotImplementedError(
178 f"{type(index)} is not supported for {var_name}, use "
179 f"type: {enforce_index_type} instead."
180 )
181
182 # Check time index is ordered in time
183 if not index.is_monotonic:
184 raise ValueError(
185 f"The (time) index of {var_name} must be sorted monotonically increasing, "
186 f"but found: {index}"
187 )
188
189 # Check that index is not empty
190 if not allow_empty and len(index) < 1:
191 raise ValueError(
192 f"{var_name} must contain at least some values, but found none."
193 )
194
195 return index
196
197
198 def check_equal_time_index(*ys):
199 """Check that time series have the same (time) indices.
200
201 Parameters
202 ----------
203 *ys : tuple of pd.Series, pd.DataFrame or np.ndarray, or None
204 One or more time series
205
206 Raises
207 ------
208 ValueError
209 If there are at least two no=-None entries of ys
210 of which pandas indices are not the same
211 np.ndarray are considered having integer range index on axis 0
212 """
213 # None entries are ignored
214 y_not_None = [y for y in ys if y is not None]
215
216 # if there is no or just one element, there is nothing to compare
217 if len(y_not_None) < 2:
218 return None
219
220 # only validate indices if data is passed as pd.Series
221 if isinstance(y_not_None[0], np.ndarray):
222 first_index = pd.Index(range(len(y_not_None[0])))
223 else:
224 first_index = y_not_None[0].index
225
226 check_time_index(first_index)
227
228 for y in y_not_None[1:]:
229 if isinstance(y, np.ndarray):
230 y_index = pd.Index(y)
231 else:
232 y_index = y.index
233
234 check_time_index(y_index)
235
236 if not first_index.equals(y_index):
237 raise ValueError("Some (time) indices are not the same.")
238
239
240 def _is_int_index(index):
241 """Check if index type is one of pd.RangeIndex or pd.Int64Index."""
242 return type(index) in (pd.Int64Index, pd.RangeIndex)
243
244
245 def check_consistent_index_type(a, b):
246 """Check that two indices have consistent types.
247
248 Parameters
249 ----------
250 a : pd.Index
251 Index being checked for consistency
252 b : pd.Index
253 Index being checked for consistency
254
255 Raises
256 ------
257 TypeError
258 If index types are inconsistent
259 """
260 msg = (
261 "Found series with inconsistent index types, please make sure all "
262 "series have the same index type."
263 )
264
265 if _is_int_index(a):
266 if not _is_int_index(b):
267 raise TypeError(msg)
268
269 else:
270 # check types, note that isinstance() does not work here because index
271 # types inherit from each other, hence we check for type equality
272 if not type(a) is type(b): # noqa
273 raise TypeError(msg)
274
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/sktime/utils/validation/series.py b/sktime/utils/validation/series.py
--- a/sktime/utils/validation/series.py
+++ b/sktime/utils/validation/series.py
@@ -10,8 +10,8 @@
"check_equal_time_index",
"check_consistent_index_type",
]
-import pandas as pd
import numpy as np
+import pandas as pd
# We currently support the following types for input data and time index types.
VALID_DATA_TYPES = (pd.DataFrame, pd.Series, np.ndarray)
@@ -56,6 +56,7 @@
allow_numpy=True,
allow_None=True,
enforce_index_type=None,
+ allow_index_names=False,
var_name="input",
):
"""Validate input data to be a valid mtype for Series.
@@ -76,6 +77,8 @@
whether no error is raised if Z is None
enforce_index_type : type, default = None
type of time index
+ allow_index_names : bool, default = False
+ If False, names of Z.index will be set to None
var_name : str, default = "input" - variable name printed in error messages
Returns
@@ -139,6 +142,9 @@
var_name=var_name,
)
+ if not allow_index_names and not isinstance(Z, np.ndarray):
+ Z.index.names = [None for name in Z.index.names]
+
return Z
| {"golden_diff": "diff --git a/sktime/utils/validation/series.py b/sktime/utils/validation/series.py\n--- a/sktime/utils/validation/series.py\n+++ b/sktime/utils/validation/series.py\n@@ -10,8 +10,8 @@\n \"check_equal_time_index\",\n \"check_consistent_index_type\",\n ]\n-import pandas as pd\n import numpy as np\n+import pandas as pd\n \n # We currently support the following types for input data and time index types.\n VALID_DATA_TYPES = (pd.DataFrame, pd.Series, np.ndarray)\n@@ -56,6 +56,7 @@\n allow_numpy=True,\n allow_None=True,\n enforce_index_type=None,\n+ allow_index_names=False,\n var_name=\"input\",\n ):\n \"\"\"Validate input data to be a valid mtype for Series.\n@@ -76,6 +77,8 @@\n whether no error is raised if Z is None\n enforce_index_type : type, default = None\n type of time index\n+ allow_index_names : bool, default = False\n+ If False, names of Z.index will be set to None\n var_name : str, default = \"input\" - variable name printed in error messages\n \n Returns\n@@ -139,6 +142,9 @@\n var_name=var_name,\n )\n \n+ if not allow_index_names and not isinstance(Z, np.ndarray):\n+ Z.index.names = [None for name in Z.index.names]\n+\n return Z\n", "issue": "[BUG] ValueError when fitting Prophet and X.index.name=='ds' \n**Describe the bug**\r\nThe following error is raised when trying to fit Prophet using X where the name of X's DateTimeIndex is named \"ds\":\r\nValueError: 'ds' is both an index level and a column label, which is ambiguous.\r\n\r\nIt happens in the _merge_X method in the _ProphetAdapter, line 193, because after running line 191, X contains a column and an index called \"ds\".\r\n\r\nA suggested solution would be to add this oneliner before line 189 of .forecasting.base.adapters._fbprophet.py:\r\n```python\r\nX.index.name = \"index\"\r\n```\r\nNote: This assumes that X never has a MultiIndex. Not sure if that ever happens though..\r\n\r\nThis way, the user is still allowed to call X's index \"ds\". Since \"ds\" is the standard name that Prophet users use for their DateTime column, this is the most intuitive and least restrictive solution.\r\n\r\n**To Reproduce**\r\n```python\r\nimport pandas as pd\r\nfrom sktime.datasets import load_airline\r\nfrom sktime.forecasting.fbprophet import Prophet\r\nfrom sktime.forecasting.model_selection import ForecastingGridSearchCV\r\nfrom sktime.forecasting.model_selection import ExpandingWindowSplitter\r\n\r\ny_train = load_airline().to_timestamp(freq='M')\r\n\r\n# Commenting out the following line makes everything run smoothly\r\ny_train.index.name = 'ds'\r\n\r\nX_train = pd.DataFrame({'x1':y_train+200, 'x2':y_train+100})\r\n\r\nforecaster = Prophet(yearly_seasonality=False, weekly_seasonality=False, daily_seasonality=False)\r\nforecaster.fit(y_train, X_train)\r\n```\r\n\r\n**Expected behavior**\r\nNo error\r\n\r\n**Additional context**\r\n\r\n\r\n**Versions**\r\n<details>\r\n\r\nSystem:\r\n python: 3.9.7 | packaged by conda-forge | (default, Sep 23 2021, 07:24:41) [MSC v.1916 64 bit (AMD64)]\r\nexecutable: C:\\Users\\............\\env\\python.exe\r\n machine: Windows-10-10.0.19043-SP0\r\n\r\nPython dependencies:\r\n pip: 21.2.4\r\n setuptools: 58.1.0\r\n sklearn: 1.0\r\n sktime: 0.7.0\r\n statsmodels: 0.12.2\r\n numpy: 1.21.2\r\n scipy: 1.7.1\r\n Cython: 0.29.24\r\n pandas: 1.3.3\r\n matplotlib: 3.4.3\r\n joblib: 1.0.1\r\n numba: 0.53.1\r\n pmdarima: 1.8.3\r\n tsfresh: None\r\n</details>\r\n\n", "before_files": [{"content": "#!/usr/bin/env python3 -u\n# -*- coding: utf-8 -*-\n\n\"\"\"Functions for checking input data.\"\"\"\n\n__author__ = [\"Markus L\u00f6ning\", \"Drishti Bhasin\"]\n__all__ = [\n \"check_series\",\n \"check_time_index\",\n \"check_equal_time_index\",\n \"check_consistent_index_type\",\n]\nimport pandas as pd\nimport numpy as np\n\n# We currently support the following types for input data and time index types.\nVALID_DATA_TYPES = (pd.DataFrame, pd.Series, np.ndarray)\nVALID_INDEX_TYPES = (pd.Int64Index, pd.RangeIndex, pd.PeriodIndex, pd.DatetimeIndex)\n\n\ndef _check_is_univariate(y, var_name=\"input\"):\n \"\"\"Check if series is univariate.\"\"\"\n if isinstance(y, pd.DataFrame):\n nvars = y.shape[1]\n if nvars > 1:\n raise ValueError(\n f\"{var_name} must be univariate, but found {nvars} variables.\"\n )\n if isinstance(y, np.ndarray) and y.ndim > 1 and y.shape[1] > 1:\n raise ValueError(\n f\"{var_name} must be univariate, but found np.ndarray with more than \"\n \"one column\"\n )\n\n\ndef _check_is_multivariate(Z, var_name=\"input\"):\n \"\"\"Check if series is multivariate.\"\"\"\n if isinstance(Z, pd.Series):\n raise ValueError(f\"{var_name} must have 2 or more variables, but found 1.\")\n if isinstance(Z, pd.DataFrame):\n nvars = Z.shape[1]\n if nvars < 2:\n raise ValueError(\n f\"{var_name} must have 2 or more variables, but found {nvars}.\"\n )\n if isinstance(Z, np.ndarray):\n if Z.ndim == 1 or (Z.ndim == 2 and Z.shape[1] == 1):\n raise ValueError(f\"{var_name} must have 2 or more variables, but found 1.\")\n\n\ndef check_series(\n Z,\n enforce_univariate=False,\n enforce_multivariate=False,\n allow_empty=False,\n allow_numpy=True,\n allow_None=True,\n enforce_index_type=None,\n var_name=\"input\",\n):\n \"\"\"Validate input data to be a valid mtype for Series.\n\n Parameters\n ----------\n Z : pd.Series, pd.DataFrame, np.ndarray, or None\n Univariate or multivariate time series.\n enforce_univariate : bool, default = False\n If True, multivariate Z will raise an error.\n enforce_multivariate: bool, default = False\n If True, univariate Z will raise an error.\n allow_empty : bool, default = False\n whether a container with zero samples is allowed\n allow_numpy : bool, default = True\n whether no error is raised if Z is in a valid numpy.ndarray format\n allow_None : bool, default = True\n whether no error is raised if Z is None\n enforce_index_type : type, default = None\n type of time index\n var_name : str, default = \"input\" - variable name printed in error messages\n\n Returns\n -------\n Z : pd.Series, pd.DataFrame, np.ndarray, or None\n Validated time series - a reference to the input Z\n\n Raises\n ------\n TypeError - if Z is not in a valid type or format for scitype Series\n if enforce_univariate is True:\n ValueError if Z has 2 or more columns\n if enforce_multivariate is True:\n ValueError if Z has 1 column\n if allow_numpy is false:\n TypeError - if Z is of type np.ndarray\n if allow_empty is false:\n ValueError - if Z has length 0\n if allow_None is false:\n ValueError - if Z is None\n if enforce_index_type is not None and Z is pandas type:\n ValueError - if Z has index type other than enforce_index_type\n \"\"\"\n if Z is None:\n if allow_None:\n return Z\n else:\n raise ValueError(var_name + \" cannot be None\")\n\n # Check if pandas series or numpy array\n if not allow_numpy:\n valid_data_types = tuple(\n filter(lambda x: x is not np.ndarray, VALID_DATA_TYPES)\n )\n else:\n valid_data_types = VALID_DATA_TYPES\n\n if not isinstance(Z, valid_data_types):\n raise TypeError(\n f\"{var_name} must be a one of {valid_data_types}, but found type: {type(Z)}\"\n )\n\n if enforce_univariate and enforce_multivariate:\n raise ValueError(\n \"`enforce_univariate` and `enforce_multivariate` cannot both be set to \"\n \"True.\"\n )\n\n if enforce_univariate:\n _check_is_univariate(Z, var_name=var_name)\n\n if enforce_multivariate:\n _check_is_multivariate(Z, var_name=var_name)\n\n # check time index if input data is not an NumPy ndarray\n if not isinstance(Z, np.ndarray):\n check_time_index(\n Z.index,\n allow_empty=allow_empty,\n enforce_index_type=enforce_index_type,\n var_name=var_name,\n )\n\n return Z\n\n\ndef check_time_index(\n index, allow_empty=False, enforce_index_type=None, var_name=\"input\"\n):\n \"\"\"Check time index.\n\n Parameters\n ----------\n index : pd.Index or np.array\n Time index\n allow_empty : bool, optional (default=False)\n If False, empty `index` raises an error.\n enforce_index_type : type, optional (default=None)\n type of time index\n var_name : str, default = \"input\" - variable name printed in error messages\n\n Returns\n -------\n time_index : pd.Index\n Validated time index - a reference to the input index\n \"\"\"\n if isinstance(index, np.ndarray):\n index = pd.Index(index)\n\n # We here check for type equality because isinstance does not\n # work reliably because index types inherit from each other.\n if not type(index) in VALID_INDEX_TYPES:\n raise NotImplementedError(\n f\"{type(index)} is not supported for {var_name}, use \"\n f\"one of {VALID_INDEX_TYPES} instead.\"\n )\n\n if enforce_index_type and type(index) is not enforce_index_type:\n raise NotImplementedError(\n f\"{type(index)} is not supported for {var_name}, use \"\n f\"type: {enforce_index_type} instead.\"\n )\n\n # Check time index is ordered in time\n if not index.is_monotonic:\n raise ValueError(\n f\"The (time) index of {var_name} must be sorted monotonically increasing, \"\n f\"but found: {index}\"\n )\n\n # Check that index is not empty\n if not allow_empty and len(index) < 1:\n raise ValueError(\n f\"{var_name} must contain at least some values, but found none.\"\n )\n\n return index\n\n\ndef check_equal_time_index(*ys):\n \"\"\"Check that time series have the same (time) indices.\n\n Parameters\n ----------\n *ys : tuple of pd.Series, pd.DataFrame or np.ndarray, or None\n One or more time series\n\n Raises\n ------\n ValueError\n If there are at least two no=-None entries of ys\n of which pandas indices are not the same\n np.ndarray are considered having integer range index on axis 0\n \"\"\"\n # None entries are ignored\n y_not_None = [y for y in ys if y is not None]\n\n # if there is no or just one element, there is nothing to compare\n if len(y_not_None) < 2:\n return None\n\n # only validate indices if data is passed as pd.Series\n if isinstance(y_not_None[0], np.ndarray):\n first_index = pd.Index(range(len(y_not_None[0])))\n else:\n first_index = y_not_None[0].index\n\n check_time_index(first_index)\n\n for y in y_not_None[1:]:\n if isinstance(y, np.ndarray):\n y_index = pd.Index(y)\n else:\n y_index = y.index\n\n check_time_index(y_index)\n\n if not first_index.equals(y_index):\n raise ValueError(\"Some (time) indices are not the same.\")\n\n\ndef _is_int_index(index):\n \"\"\"Check if index type is one of pd.RangeIndex or pd.Int64Index.\"\"\"\n return type(index) in (pd.Int64Index, pd.RangeIndex)\n\n\ndef check_consistent_index_type(a, b):\n \"\"\"Check that two indices have consistent types.\n\n Parameters\n ----------\n a : pd.Index\n Index being checked for consistency\n b : pd.Index\n Index being checked for consistency\n\n Raises\n ------\n TypeError\n If index types are inconsistent\n \"\"\"\n msg = (\n \"Found series with inconsistent index types, please make sure all \"\n \"series have the same index type.\"\n )\n\n if _is_int_index(a):\n if not _is_int_index(b):\n raise TypeError(msg)\n\n else:\n # check types, note that isinstance() does not work here because index\n # types inherit from each other, hence we check for type equality\n if not type(a) is type(b): # noqa\n raise TypeError(msg)\n", "path": "sktime/utils/validation/series.py"}], "after_files": [{"content": "#!/usr/bin/env python3 -u\n# -*- coding: utf-8 -*-\n\n\"\"\"Functions for checking input data.\"\"\"\n\n__author__ = [\"Markus L\u00f6ning\", \"Drishti Bhasin\"]\n__all__ = [\n \"check_series\",\n \"check_time_index\",\n \"check_equal_time_index\",\n \"check_consistent_index_type\",\n]\nimport numpy as np\nimport pandas as pd\n\n# We currently support the following types for input data and time index types.\nVALID_DATA_TYPES = (pd.DataFrame, pd.Series, np.ndarray)\nVALID_INDEX_TYPES = (pd.Int64Index, pd.RangeIndex, pd.PeriodIndex, pd.DatetimeIndex)\n\n\ndef _check_is_univariate(y, var_name=\"input\"):\n \"\"\"Check if series is univariate.\"\"\"\n if isinstance(y, pd.DataFrame):\n nvars = y.shape[1]\n if nvars > 1:\n raise ValueError(\n f\"{var_name} must be univariate, but found {nvars} variables.\"\n )\n if isinstance(y, np.ndarray) and y.ndim > 1 and y.shape[1] > 1:\n raise ValueError(\n f\"{var_name} must be univariate, but found np.ndarray with more than \"\n \"one column\"\n )\n\n\ndef _check_is_multivariate(Z, var_name=\"input\"):\n \"\"\"Check if series is multivariate.\"\"\"\n if isinstance(Z, pd.Series):\n raise ValueError(f\"{var_name} must have 2 or more variables, but found 1.\")\n if isinstance(Z, pd.DataFrame):\n nvars = Z.shape[1]\n if nvars < 2:\n raise ValueError(\n f\"{var_name} must have 2 or more variables, but found {nvars}.\"\n )\n if isinstance(Z, np.ndarray):\n if Z.ndim == 1 or (Z.ndim == 2 and Z.shape[1] == 1):\n raise ValueError(f\"{var_name} must have 2 or more variables, but found 1.\")\n\n\ndef check_series(\n Z,\n enforce_univariate=False,\n enforce_multivariate=False,\n allow_empty=False,\n allow_numpy=True,\n allow_None=True,\n enforce_index_type=None,\n allow_index_names=False,\n var_name=\"input\",\n):\n \"\"\"Validate input data to be a valid mtype for Series.\n\n Parameters\n ----------\n Z : pd.Series, pd.DataFrame, np.ndarray, or None\n Univariate or multivariate time series.\n enforce_univariate : bool, default = False\n If True, multivariate Z will raise an error.\n enforce_multivariate: bool, default = False\n If True, univariate Z will raise an error.\n allow_empty : bool, default = False\n whether a container with zero samples is allowed\n allow_numpy : bool, default = True\n whether no error is raised if Z is in a valid numpy.ndarray format\n allow_None : bool, default = True\n whether no error is raised if Z is None\n enforce_index_type : type, default = None\n type of time index\n allow_index_names : bool, default = False\n If False, names of Z.index will be set to None\n var_name : str, default = \"input\" - variable name printed in error messages\n\n Returns\n -------\n Z : pd.Series, pd.DataFrame, np.ndarray, or None\n Validated time series - a reference to the input Z\n\n Raises\n ------\n TypeError - if Z is not in a valid type or format for scitype Series\n if enforce_univariate is True:\n ValueError if Z has 2 or more columns\n if enforce_multivariate is True:\n ValueError if Z has 1 column\n if allow_numpy is false:\n TypeError - if Z is of type np.ndarray\n if allow_empty is false:\n ValueError - if Z has length 0\n if allow_None is false:\n ValueError - if Z is None\n if enforce_index_type is not None and Z is pandas type:\n ValueError - if Z has index type other than enforce_index_type\n \"\"\"\n if Z is None:\n if allow_None:\n return Z\n else:\n raise ValueError(var_name + \" cannot be None\")\n\n # Check if pandas series or numpy array\n if not allow_numpy:\n valid_data_types = tuple(\n filter(lambda x: x is not np.ndarray, VALID_DATA_TYPES)\n )\n else:\n valid_data_types = VALID_DATA_TYPES\n\n if not isinstance(Z, valid_data_types):\n raise TypeError(\n f\"{var_name} must be a one of {valid_data_types}, but found type: {type(Z)}\"\n )\n\n if enforce_univariate and enforce_multivariate:\n raise ValueError(\n \"`enforce_univariate` and `enforce_multivariate` cannot both be set to \"\n \"True.\"\n )\n\n if enforce_univariate:\n _check_is_univariate(Z, var_name=var_name)\n\n if enforce_multivariate:\n _check_is_multivariate(Z, var_name=var_name)\n\n # check time index if input data is not an NumPy ndarray\n if not isinstance(Z, np.ndarray):\n check_time_index(\n Z.index,\n allow_empty=allow_empty,\n enforce_index_type=enforce_index_type,\n var_name=var_name,\n )\n\n if not allow_index_names and not isinstance(Z, np.ndarray):\n Z.index.names = [None for name in Z.index.names]\n\n return Z\n\n\ndef check_time_index(\n index, allow_empty=False, enforce_index_type=None, var_name=\"input\"\n):\n \"\"\"Check time index.\n\n Parameters\n ----------\n index : pd.Index or np.array\n Time index\n allow_empty : bool, optional (default=False)\n If False, empty `index` raises an error.\n enforce_index_type : type, optional (default=None)\n type of time index\n var_name : str, default = \"input\" - variable name printed in error messages\n\n Returns\n -------\n time_index : pd.Index\n Validated time index - a reference to the input index\n \"\"\"\n if isinstance(index, np.ndarray):\n index = pd.Index(index)\n\n # We here check for type equality because isinstance does not\n # work reliably because index types inherit from each other.\n if not type(index) in VALID_INDEX_TYPES:\n raise NotImplementedError(\n f\"{type(index)} is not supported for {var_name}, use \"\n f\"one of {VALID_INDEX_TYPES} instead.\"\n )\n\n if enforce_index_type and type(index) is not enforce_index_type:\n raise NotImplementedError(\n f\"{type(index)} is not supported for {var_name}, use \"\n f\"type: {enforce_index_type} instead.\"\n )\n\n # Check time index is ordered in time\n if not index.is_monotonic:\n raise ValueError(\n f\"The (time) index of {var_name} must be sorted monotonically increasing, \"\n f\"but found: {index}\"\n )\n\n # Check that index is not empty\n if not allow_empty and len(index) < 1:\n raise ValueError(\n f\"{var_name} must contain at least some values, but found none.\"\n )\n\n return index\n\n\ndef check_equal_time_index(*ys):\n \"\"\"Check that time series have the same (time) indices.\n\n Parameters\n ----------\n *ys : tuple of pd.Series, pd.DataFrame or np.ndarray, or None\n One or more time series\n\n Raises\n ------\n ValueError\n If there are at least two no=-None entries of ys\n of which pandas indices are not the same\n np.ndarray are considered having integer range index on axis 0\n \"\"\"\n # None entries are ignored\n y_not_None = [y for y in ys if y is not None]\n\n # if there is no or just one element, there is nothing to compare\n if len(y_not_None) < 2:\n return None\n\n # only validate indices if data is passed as pd.Series\n if isinstance(y_not_None[0], np.ndarray):\n first_index = pd.Index(range(len(y_not_None[0])))\n else:\n first_index = y_not_None[0].index\n\n check_time_index(first_index)\n\n for y in y_not_None[1:]:\n if isinstance(y, np.ndarray):\n y_index = pd.Index(y)\n else:\n y_index = y.index\n\n check_time_index(y_index)\n\n if not first_index.equals(y_index):\n raise ValueError(\"Some (time) indices are not the same.\")\n\n\ndef _is_int_index(index):\n \"\"\"Check if index type is one of pd.RangeIndex or pd.Int64Index.\"\"\"\n return type(index) in (pd.Int64Index, pd.RangeIndex)\n\n\ndef check_consistent_index_type(a, b):\n \"\"\"Check that two indices have consistent types.\n\n Parameters\n ----------\n a : pd.Index\n Index being checked for consistency\n b : pd.Index\n Index being checked for consistency\n\n Raises\n ------\n TypeError\n If index types are inconsistent\n \"\"\"\n msg = (\n \"Found series with inconsistent index types, please make sure all \"\n \"series have the same index type.\"\n )\n\n if _is_int_index(a):\n if not _is_int_index(b):\n raise TypeError(msg)\n\n else:\n # check types, note that isinstance() does not work here because index\n # types inherit from each other, hence we check for type equality\n if not type(a) is type(b): # noqa\n raise TypeError(msg)\n", "path": "sktime/utils/validation/series.py"}]} | 3,624 | 321 |
gh_patches_debug_9715 | rasdani/github-patches | git_diff | OCA__server-tools-74 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
[7.0] [base_optional_quick_create] AttributeError: 'NoneType' object has no attribute 'name_create'
Error at starting before a migration if a model has been removed
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `base_optional_quick_create/model.py`
Content:
```
1 # -*- coding: utf-8 -*-
2 ##############################################################################
3 #
4 # Copyright (C) 2013 Agile Business Group sagl (<http://www.agilebg.com>)
5 #
6 # This program is free software: you can redistribute it and/or modify
7 # it under the terms of the GNU Affero General Public License as published
8 # by the Free Software Foundation, either version 3 of the License, or
9 # (at your option) any later version.
10 #
11 # This program is distributed in the hope that it will be useful,
12 # but WITHOUT ANY WARRANTY; without even the implied warranty of
13 # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
14 # GNU Affero General Public License for more details.
15 #
16 # You should have received a copy of the GNU Affero General Public License
17 # along with this program. If not, see <http://www.gnu.org/licenses/>.
18 #
19 ##############################################################################
20
21 from openerp.osv import orm, fields
22 from openerp import SUPERUSER_ID
23 from openerp.tools.translate import _
24
25
26 class ir_model(orm.Model):
27
28 _inherit = 'ir.model'
29
30 _columns = {
31 'avoid_quick_create': fields.boolean('Avoid quick create'),
32 }
33
34 def _wrap_name_create(self, old_create, model):
35 def wrapper(cr, uid, name, context=None):
36 raise orm.except_orm(_('Error'),
37 _("Can't create quickly. "
38 "Opening create form"))
39 return wrapper
40
41 def _register_hook(self, cr, ids=None):
42 if ids is None:
43 ids = self.search(cr, SUPERUSER_ID, [])
44 for model in self.browse(cr, SUPERUSER_ID, ids):
45 if model.avoid_quick_create:
46 model_name = model.model
47 model_obj = self.pool.get(model_name)
48 if not hasattr(model_obj, 'check_quick_create'):
49 model_obj.name_create = self._wrap_name_create(
50 model_obj.name_create,
51 model_name)
52 model_obj.check_quick_create = True
53 return True
54
55 def create(self, cr, uid, vals, context=None):
56 res_id = super(ir_model, self).create(cr, uid, vals, context=context)
57 self._register_hook(cr, [res_id])
58 return res_id
59
60 def write(self, cr, uid, ids, vals, context=None):
61 if isinstance(ids, (int, long)):
62 ids = [ids]
63 super(ir_model, self).write(cr, uid, ids, vals, context=context)
64 self._register_hook(cr, ids)
65 return True
66
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/base_optional_quick_create/model.py b/base_optional_quick_create/model.py
--- a/base_optional_quick_create/model.py
+++ b/base_optional_quick_create/model.py
@@ -45,7 +45,7 @@
if model.avoid_quick_create:
model_name = model.model
model_obj = self.pool.get(model_name)
- if not hasattr(model_obj, 'check_quick_create'):
+ if model_obj and not hasattr(model_obj, 'check_quick_create'):
model_obj.name_create = self._wrap_name_create(
model_obj.name_create,
model_name)
| {"golden_diff": "diff --git a/base_optional_quick_create/model.py b/base_optional_quick_create/model.py\n--- a/base_optional_quick_create/model.py\n+++ b/base_optional_quick_create/model.py\n@@ -45,7 +45,7 @@\n if model.avoid_quick_create:\n model_name = model.model\n model_obj = self.pool.get(model_name)\n- if not hasattr(model_obj, 'check_quick_create'):\n+ if model_obj and not hasattr(model_obj, 'check_quick_create'):\n model_obj.name_create = self._wrap_name_create(\n model_obj.name_create,\n model_name)\n", "issue": "[7.0] [base_optional_quick_create] AttributeError: 'NoneType' object has no attribute 'name_create'\nError at starting before a migration if a model has been removed\n\n", "before_files": [{"content": "# -*- coding: utf-8 -*-\n##############################################################################\n#\n# Copyright (C) 2013 Agile Business Group sagl (<http://www.agilebg.com>)\n#\n# This program is free software: you can redistribute it and/or modify\n# it under the terms of the GNU Affero General Public License as published\n# by the Free Software Foundation, either version 3 of the License, or\n# (at your option) any later version.\n#\n# This program is distributed in the hope that it will be useful,\n# but WITHOUT ANY WARRANTY; without even the implied warranty of\n# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the\n# GNU Affero General Public License for more details.\n#\n# You should have received a copy of the GNU Affero General Public License\n# along with this program. If not, see <http://www.gnu.org/licenses/>.\n#\n##############################################################################\n\nfrom openerp.osv import orm, fields\nfrom openerp import SUPERUSER_ID\nfrom openerp.tools.translate import _\n\n\nclass ir_model(orm.Model):\n\n _inherit = 'ir.model'\n\n _columns = {\n 'avoid_quick_create': fields.boolean('Avoid quick create'),\n }\n\n def _wrap_name_create(self, old_create, model):\n def wrapper(cr, uid, name, context=None):\n raise orm.except_orm(_('Error'),\n _(\"Can't create quickly. \"\n \"Opening create form\"))\n return wrapper\n\n def _register_hook(self, cr, ids=None):\n if ids is None:\n ids = self.search(cr, SUPERUSER_ID, [])\n for model in self.browse(cr, SUPERUSER_ID, ids):\n if model.avoid_quick_create:\n model_name = model.model\n model_obj = self.pool.get(model_name)\n if not hasattr(model_obj, 'check_quick_create'):\n model_obj.name_create = self._wrap_name_create(\n model_obj.name_create,\n model_name)\n model_obj.check_quick_create = True\n return True\n\n def create(self, cr, uid, vals, context=None):\n res_id = super(ir_model, self).create(cr, uid, vals, context=context)\n self._register_hook(cr, [res_id])\n return res_id\n\n def write(self, cr, uid, ids, vals, context=None):\n if isinstance(ids, (int, long)):\n ids = [ids]\n super(ir_model, self).write(cr, uid, ids, vals, context=context)\n self._register_hook(cr, ids)\n return True\n", "path": "base_optional_quick_create/model.py"}], "after_files": [{"content": "# -*- coding: utf-8 -*-\n##############################################################################\n#\n# Copyright (C) 2013 Agile Business Group sagl (<http://www.agilebg.com>)\n#\n# This program is free software: you can redistribute it and/or modify\n# it under the terms of the GNU Affero General Public License as published\n# by the Free Software Foundation, either version 3 of the License, or\n# (at your option) any later version.\n#\n# This program is distributed in the hope that it will be useful,\n# but WITHOUT ANY WARRANTY; without even the implied warranty of\n# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the\n# GNU Affero General Public License for more details.\n#\n# You should have received a copy of the GNU Affero General Public License\n# along with this program. If not, see <http://www.gnu.org/licenses/>.\n#\n##############################################################################\n\nfrom openerp.osv import orm, fields\nfrom openerp import SUPERUSER_ID\nfrom openerp.tools.translate import _\n\n\nclass ir_model(orm.Model):\n\n _inherit = 'ir.model'\n\n _columns = {\n 'avoid_quick_create': fields.boolean('Avoid quick create'),\n }\n\n def _wrap_name_create(self, old_create, model):\n def wrapper(cr, uid, name, context=None):\n raise orm.except_orm(_('Error'),\n _(\"Can't create quickly. \"\n \"Opening create form\"))\n return wrapper\n\n def _register_hook(self, cr, ids=None):\n if ids is None:\n ids = self.search(cr, SUPERUSER_ID, [])\n for model in self.browse(cr, SUPERUSER_ID, ids):\n if model.avoid_quick_create:\n model_name = model.model\n model_obj = self.pool.get(model_name)\n if model_obj and not hasattr(model_obj, 'check_quick_create'):\n model_obj.name_create = self._wrap_name_create(\n model_obj.name_create,\n model_name)\n model_obj.check_quick_create = True\n return True\n\n def create(self, cr, uid, vals, context=None):\n res_id = super(ir_model, self).create(cr, uid, vals, context=context)\n self._register_hook(cr, [res_id])\n return res_id\n\n def write(self, cr, uid, ids, vals, context=None):\n if isinstance(ids, (int, long)):\n ids = [ids]\n super(ir_model, self).write(cr, uid, ids, vals, context=context)\n self._register_hook(cr, ids)\n return True\n", "path": "base_optional_quick_create/model.py"}]} | 967 | 124 |
gh_patches_debug_27735 | rasdani/github-patches | git_diff | e-valuation__EvaP-1263 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Remove ViewTest where possible
Right now we have a `class ViewTest`, for which there is one subclass for each view that we have.
For views that we have tested properly, it provides no additional value and I I propose to replace it with the original `WebTest`.
Originally I proposed to remove it altogether and copypaste its test to all the test cases that wouldn't have any valuable test otherwise. @janno42 convinced me to leave it there and rename it to `WebTestWith200Check` instead.
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `evap/evaluation/migrations/0002_initial_data.py`
Content:
```
1 from django.db import migrations
2 from django.contrib.auth.models import Group
3
4
5 def insert_emailtemplates(apps, _schema_editor):
6 emailtemplates = [
7 ("Lecturer Review Notice", "[EvaP] New Course ready for approval"),
8 ("Student Reminder", "[EvaP] Evaluation period is ending"),
9 ("Publishing Notice", "[EvaP] A course has been published"),
10 ("Login Key Created", "[EvaP] A login key was created"),
11 ]
12
13 EmailTemplate = apps.get_model("evaluation", "EmailTemplate")
14
15 for name, subject in emailtemplates:
16 if not EmailTemplate.objects.filter(name=name).exists():
17 EmailTemplate.objects.create(name=name, subject=subject, body="")
18
19 Group.objects.create(name="Staff")
20
21
22 class Migration(migrations.Migration):
23
24 dependencies = [
25 ('evaluation', '0001_initial'),
26 ]
27
28 operations = [
29 migrations.RunPython(insert_emailtemplates),
30 ]
31
```
Path: `evap/grades/migrations/0002_initial_data.py`
Content:
```
1 from django.db import migrations
2 from django.contrib.auth.models import Group
3
4
5 def add_group(_apps, _schema_editor):
6 Group.objects.create(name="Grade publisher")
7
8
9 class Migration(migrations.Migration):
10
11 dependencies = [
12 ('grades', '0001_initial'),
13 ]
14
15 operations = [
16 migrations.RunPython(add_group),
17 ]
18
```
Path: `evap/evaluation/migrations/0055_reviewer_group.py`
Content:
```
1 from django.contrib.auth.models import Group
2 from django.db import migrations
3
4
5 def add_group(_apps, _schema_editor):
6 Group.objects.create(name="Reviewer")
7
8
9 def delete_group(_apps, _schema_editor):
10 Group.objects.get(name="Reviewer").delete()
11
12
13 class Migration(migrations.Migration):
14
15 dependencies = [
16 ('evaluation', '0054_userprofile_language'),
17 ]
18
19 operations = [
20 migrations.RunPython(add_group, reverse_code=delete_group),
21 ]
22
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/evap/evaluation/migrations/0002_initial_data.py b/evap/evaluation/migrations/0002_initial_data.py
--- a/evap/evaluation/migrations/0002_initial_data.py
+++ b/evap/evaluation/migrations/0002_initial_data.py
@@ -1,5 +1,4 @@
from django.db import migrations
-from django.contrib.auth.models import Group
def insert_emailtemplates(apps, _schema_editor):
@@ -16,6 +15,7 @@
if not EmailTemplate.objects.filter(name=name).exists():
EmailTemplate.objects.create(name=name, subject=subject, body="")
+ Group = apps.get_model("auth", "Group")
Group.objects.create(name="Staff")
diff --git a/evap/evaluation/migrations/0055_reviewer_group.py b/evap/evaluation/migrations/0055_reviewer_group.py
--- a/evap/evaluation/migrations/0055_reviewer_group.py
+++ b/evap/evaluation/migrations/0055_reviewer_group.py
@@ -1,12 +1,13 @@
-from django.contrib.auth.models import Group
from django.db import migrations
-def add_group(_apps, _schema_editor):
+def add_group(apps, _schema_editor):
+ Group = apps.get_model("auth", "Group")
Group.objects.create(name="Reviewer")
-def delete_group(_apps, _schema_editor):
+def delete_group(apps, _schema_editor):
+ Group = apps.get_model("auth", "Group")
Group.objects.get(name="Reviewer").delete()
diff --git a/evap/grades/migrations/0002_initial_data.py b/evap/grades/migrations/0002_initial_data.py
--- a/evap/grades/migrations/0002_initial_data.py
+++ b/evap/grades/migrations/0002_initial_data.py
@@ -1,8 +1,8 @@
from django.db import migrations
-from django.contrib.auth.models import Group
-def add_group(_apps, _schema_editor):
+def add_group(apps, _schema_editor):
+ Group = apps.get_model("auth", "Group")
Group.objects.create(name="Grade publisher")
| {"golden_diff": "diff --git a/evap/evaluation/migrations/0002_initial_data.py b/evap/evaluation/migrations/0002_initial_data.py\n--- a/evap/evaluation/migrations/0002_initial_data.py\n+++ b/evap/evaluation/migrations/0002_initial_data.py\n@@ -1,5 +1,4 @@\n from django.db import migrations\n-from django.contrib.auth.models import Group\n \n \n def insert_emailtemplates(apps, _schema_editor):\n@@ -16,6 +15,7 @@\n if not EmailTemplate.objects.filter(name=name).exists():\n EmailTemplate.objects.create(name=name, subject=subject, body=\"\")\n \n+ Group = apps.get_model(\"auth\", \"Group\")\n Group.objects.create(name=\"Staff\")\n \n \ndiff --git a/evap/evaluation/migrations/0055_reviewer_group.py b/evap/evaluation/migrations/0055_reviewer_group.py\n--- a/evap/evaluation/migrations/0055_reviewer_group.py\n+++ b/evap/evaluation/migrations/0055_reviewer_group.py\n@@ -1,12 +1,13 @@\n-from django.contrib.auth.models import Group\n from django.db import migrations\n \n \n-def add_group(_apps, _schema_editor):\n+def add_group(apps, _schema_editor):\n+ Group = apps.get_model(\"auth\", \"Group\")\n Group.objects.create(name=\"Reviewer\")\n \n \n-def delete_group(_apps, _schema_editor):\n+def delete_group(apps, _schema_editor):\n+ Group = apps.get_model(\"auth\", \"Group\")\n Group.objects.get(name=\"Reviewer\").delete()\n \n \ndiff --git a/evap/grades/migrations/0002_initial_data.py b/evap/grades/migrations/0002_initial_data.py\n--- a/evap/grades/migrations/0002_initial_data.py\n+++ b/evap/grades/migrations/0002_initial_data.py\n@@ -1,8 +1,8 @@\n from django.db import migrations\n-from django.contrib.auth.models import Group\n \n \n-def add_group(_apps, _schema_editor):\n+def add_group(apps, _schema_editor):\n+ Group = apps.get_model(\"auth\", \"Group\")\n Group.objects.create(name=\"Grade publisher\")\n", "issue": "Remove ViewTest where possible\nRight now we have a `class ViewTest`, for which there is one subclass for each view that we have.\r\n\r\nFor views that we have tested properly, it provides no additional value and I I propose to replace it with the original `WebTest`. \r\n\r\nOriginally I proposed to remove it altogether and copypaste its test to all the test cases that wouldn't have any valuable test otherwise. @janno42 convinced me to leave it there and rename it to `WebTestWith200Check` instead.\n", "before_files": [{"content": "from django.db import migrations\nfrom django.contrib.auth.models import Group\n\n\ndef insert_emailtemplates(apps, _schema_editor):\n emailtemplates = [\n (\"Lecturer Review Notice\", \"[EvaP] New Course ready for approval\"),\n (\"Student Reminder\", \"[EvaP] Evaluation period is ending\"),\n (\"Publishing Notice\", \"[EvaP] A course has been published\"),\n (\"Login Key Created\", \"[EvaP] A login key was created\"),\n ]\n\n EmailTemplate = apps.get_model(\"evaluation\", \"EmailTemplate\")\n\n for name, subject in emailtemplates:\n if not EmailTemplate.objects.filter(name=name).exists():\n EmailTemplate.objects.create(name=name, subject=subject, body=\"\")\n\n Group.objects.create(name=\"Staff\")\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ('evaluation', '0001_initial'),\n ]\n\n operations = [\n migrations.RunPython(insert_emailtemplates),\n ]\n", "path": "evap/evaluation/migrations/0002_initial_data.py"}, {"content": "from django.db import migrations\nfrom django.contrib.auth.models import Group\n\n\ndef add_group(_apps, _schema_editor):\n Group.objects.create(name=\"Grade publisher\")\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ('grades', '0001_initial'),\n ]\n\n operations = [\n migrations.RunPython(add_group),\n ]\n", "path": "evap/grades/migrations/0002_initial_data.py"}, {"content": "from django.contrib.auth.models import Group\nfrom django.db import migrations\n\n\ndef add_group(_apps, _schema_editor):\n Group.objects.create(name=\"Reviewer\")\n\n\ndef delete_group(_apps, _schema_editor):\n Group.objects.get(name=\"Reviewer\").delete()\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ('evaluation', '0054_userprofile_language'),\n ]\n\n operations = [\n migrations.RunPython(add_group, reverse_code=delete_group),\n ]\n", "path": "evap/evaluation/migrations/0055_reviewer_group.py"}], "after_files": [{"content": "from django.db import migrations\n\n\ndef insert_emailtemplates(apps, _schema_editor):\n emailtemplates = [\n (\"Lecturer Review Notice\", \"[EvaP] New Course ready for approval\"),\n (\"Student Reminder\", \"[EvaP] Evaluation period is ending\"),\n (\"Publishing Notice\", \"[EvaP] A course has been published\"),\n (\"Login Key Created\", \"[EvaP] A login key was created\"),\n ]\n\n EmailTemplate = apps.get_model(\"evaluation\", \"EmailTemplate\")\n\n for name, subject in emailtemplates:\n if not EmailTemplate.objects.filter(name=name).exists():\n EmailTemplate.objects.create(name=name, subject=subject, body=\"\")\n\n Group = apps.get_model(\"auth\", \"Group\")\n Group.objects.create(name=\"Staff\")\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ('evaluation', '0001_initial'),\n ]\n\n operations = [\n migrations.RunPython(insert_emailtemplates),\n ]\n", "path": "evap/evaluation/migrations/0002_initial_data.py"}, {"content": "from django.db import migrations\n\n\ndef add_group(apps, _schema_editor):\n Group = apps.get_model(\"auth\", \"Group\")\n Group.objects.create(name=\"Grade publisher\")\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ('grades', '0001_initial'),\n ]\n\n operations = [\n migrations.RunPython(add_group),\n ]\n", "path": "evap/grades/migrations/0002_initial_data.py"}, {"content": "from django.db import migrations\n\n\ndef add_group(apps, _schema_editor):\n Group = apps.get_model(\"auth\", \"Group\")\n Group.objects.create(name=\"Reviewer\")\n\n\ndef delete_group(apps, _schema_editor):\n Group = apps.get_model(\"auth\", \"Group\")\n Group.objects.get(name=\"Reviewer\").delete()\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ('evaluation', '0054_userprofile_language'),\n ]\n\n operations = [\n migrations.RunPython(add_group, reverse_code=delete_group),\n ]\n", "path": "evap/evaluation/migrations/0055_reviewer_group.py"}]} | 929 | 500 |
gh_patches_debug_35202 | rasdani/github-patches | git_diff | openfun__marsha-2158 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Configure sentry in standalone site
## Feature Request
**Is your feature request related to a problem or unsupported use case? Please describe.**
Sentry is not configured in the standalone site and we have no feedback is there are errors.
**Describe the solution you'd like**
Find a solution to configure sentry. Maybe by fetching th configuration on a dedicated api endpoint
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `src/backend/marsha/urls.py`
Content:
```
1 """Marsha URLs configuration."""
2 import re
3
4 from django.conf import settings
5 from django.contrib import admin
6 from django.urls import include, path, re_path, register_converter
7 from django.views.decorators.cache import cache_page
8
9 from rest_framework.routers import DefaultRouter
10
11 from marsha.core import models
12 from marsha.core.api import (
13 ChallengeAuthenticationView,
14 DocumentViewSet,
15 LiveSessionViewSet,
16 OrganizationViewSet,
17 PlaylistAccessViewSet,
18 PlaylistViewSet,
19 PortabilityResourceViewSet,
20 SharedLiveMediaViewSet,
21 ThumbnailViewSet,
22 TimedTextTrackViewSet,
23 UserViewSet,
24 VideoViewSet,
25 XAPIStatementView,
26 pairing_challenge,
27 recording_slices_manifest,
28 recording_slices_state,
29 update_state,
30 )
31 from marsha.core.api.lti_user_association import LtiUserAssociationViewSet
32 from marsha.core.urls.converters import XAPIResourceKindConverter
33 from marsha.core.utils.lti_select_utils import get_lti_select_resources
34 from marsha.core.views import (
35 DocumentLTIView,
36 DocumentView,
37 LTIConfigView,
38 LTIRespondView,
39 LTISelectView,
40 RemindersCancelView,
41 SiteView,
42 VideoLTIView,
43 VideoView,
44 )
45 from marsha.development.api import local_document_upload, local_video_upload
46
47
48 register_converter(XAPIResourceKindConverter, "xapi_resource_kind")
49
50 LTI_SELECT_ROUTE_PATTERN = (
51 rf"lti/select/((?P<resource_kind>{'|'.join(get_lti_select_resources().keys())})/)?$"
52 )
53
54 router = DefaultRouter()
55 router.register(models.Video.RESOURCE_NAME, VideoViewSet, basename="videos")
56 router.register(models.Document.RESOURCE_NAME, DocumentViewSet, basename="documents")
57 router.register(
58 models.TimedTextTrack.RESOURCE_NAME,
59 TimedTextTrackViewSet,
60 basename="timed_text_tracks",
61 )
62 router.register(models.Thumbnail.RESOURCE_NAME, ThumbnailViewSet, basename="thumbnails")
63 router.register("organizations", OrganizationViewSet, basename="organizations")
64 router.register("playlists", PlaylistViewSet, basename="playlists")
65 router.register(
66 "playlist-accesses", PlaylistAccessViewSet, basename="playlist_accesses"
67 )
68 router.register(
69 "portability-requests", PortabilityResourceViewSet, basename="portability_requests"
70 )
71 router.register("users", UserViewSet, basename="users")
72 router.register(
73 "lti-user-associations", LtiUserAssociationViewSet, basename="lti_user_associations"
74 )
75 router.register(
76 models.SharedLiveMedia.RESOURCE_NAME,
77 SharedLiveMediaViewSet,
78 basename="sharedlivemedias",
79 )
80
81 # Old routes to remove
82 router.register(
83 models.LiveSession.RESOURCE_NAME,
84 LiveSessionViewSet,
85 basename="live_sessions",
86 )
87
88 # Video related resources (for nested routes)
89 video_related_router = DefaultRouter()
90 video_related_router.register(
91 models.LiveSession.RESOURCE_NAME,
92 LiveSessionViewSet,
93 basename="live_sessions",
94 )
95
96 urlpatterns = [
97 # Account
98 path("account/", include("marsha.account.urls")),
99 # Admin
100 path("admin/", admin.site.urls),
101 # LTI
102 path("lti/config.xml", LTIConfigView.as_view(), name="config_lti_view"),
103 re_path(
104 LTI_SELECT_ROUTE_PATTERN,
105 LTISelectView.as_view(),
106 name="select_lti_view",
107 ),
108 path("lti/respond/", LTIRespondView.as_view(), name="respond_lti_view"),
109 path("lti/videos/<uuid:uuid>", VideoLTIView.as_view(), name="video_lti_view"),
110 path(
111 "lti/documents/<uuid:uuid>", DocumentLTIView.as_view(), name="document_lti_view"
112 ),
113 # Public resources
114 path("videos/<uuid:uuid>", VideoView.as_view(), name="video_direct_access"),
115 path("documents/<uuid:uuid>", DocumentView.as_view(), name="document_public"),
116 # API
117 path(
118 "api/auth/challenge/",
119 ChallengeAuthenticationView.as_view(),
120 name="api_authentication",
121 ),
122 path("api/pairing-challenge", pairing_challenge, name="pairing_challenge"),
123 path("api/update-state", update_state, name="update_state"),
124 path(
125 "api/recording-slices-manifest",
126 recording_slices_manifest,
127 name="recording_slices_manifest",
128 ),
129 path(
130 "api/recording-slices-state",
131 recording_slices_state,
132 name="recording_slices_state",
133 ),
134 path("api/", include(router.urls)),
135 path(
136 f"api/{models.Video.RESOURCE_NAME}/<uuid:video_id>/",
137 include(video_related_router.urls),
138 ),
139 path(
140 "reminders/cancel/<str:pk>/<str:key>",
141 RemindersCancelView.as_view(),
142 name="reminders_cancel",
143 ),
144 path(
145 "xapi/<xapi_resource_kind:resource_kind>/<uuid:resource_id>/",
146 XAPIStatementView.as_view(),
147 name="xapi",
148 ),
149 ]
150
151 if settings.BBB_ENABLED:
152 urlpatterns += [path("", include("marsha.bbb.urls"))]
153
154 if settings.DEPOSIT_ENABLED:
155 urlpatterns += [path("", include("marsha.deposit.urls"))]
156
157 if settings.MARKDOWN_ENABLED:
158 urlpatterns += [path("", include("marsha.markdown.urls"))]
159
160 if settings.DEBUG:
161 urlpatterns += [path("", include("marsha.development.urls"))]
162
163 if "dummy" in settings.STORAGE_BACKEND:
164 urlpatterns += [
165 path(
166 "e2e/api/video-upload/<uuid:uuid>",
167 local_video_upload,
168 name="local-video-upload",
169 ),
170 path(
171 "api/document-upload/<uuid:uuid>",
172 local_document_upload,
173 name="local-document-upload",
174 ),
175 ]
176
177 static_path = re.escape(settings.STATIC_URL.lstrip("/"))
178 media_path = re.escape(settings.MEDIA_URL.lstrip("/"))
179 SITE_IGNORE_PREFIX = "|".join([static_path, media_path])
180
181 urlpatterns += [
182 re_path(
183 # Catch all URLs that are not handled before
184 # and which do not regard static files or media files
185 rf"^(?!{SITE_IGNORE_PREFIX}).*",
186 cache_page(86400, key_prefix=settings.RELEASE)(SiteView.as_view()),
187 name="site",
188 ),
189 ]
190
```
Path: `src/backend/marsha/core/api/base.py`
Content:
```
1 """Declare API endpoints with Django RestFramework viewsets."""
2
3 from django.shortcuts import get_object_or_404
4
5 from rest_framework.decorators import api_view
6 from rest_framework.response import Response
7
8 from .. import defaults, serializers
9 from ..models import Video
10 from ..simple_jwt.tokens import ResourceAccessToken
11 from ..utils.api_utils import get_uploadable_models_s3_mapping, validate_signature
12
13
14 class ObjectPkMixin:
15 """
16 Get the object primary key from the URL path.
17
18 This is useful to avoid making extra requests using view.get_object() on
19 a ViewSet when we only need the object's id, which is available in the URL.
20 """
21
22 def get_object_pk(self):
23 """Get the object primary key from the URL path."""
24 lookup_url_kwarg = self.lookup_url_kwarg or self.lookup_field
25 return self.kwargs.get(lookup_url_kwarg)
26
27
28 class ObjectRelatedMixin:
29 """
30 Get the related video belonging to the current object.
31
32 Using view.get_object is permissions is not possible anymore due to
33 infinite recursion between has_permission and has_object_permission call.
34 """
35
36 def get_related_object(self):
37 """Get the video related to the current object."""
38 queryset = self.filter_queryset(self.get_queryset())
39 return queryset.get(pk=self.get_object_pk())
40
41
42 @api_view(["POST"])
43 def update_state(request):
44 """View handling AWS POST request to update the state of an object by key.
45
46 Parameters
47 ----------
48 request : Type[django.http.request.HttpRequest]
49 The request on the API endpoint, it should contain a payload with the following fields:
50 - key: the key of an object in the source bucket as delivered in the upload policy,
51 - state: state of the upload, should be either "ready" or "error",
52 - extraParameters: Dict containing arbitrary data sent from AWS Lambda.
53
54 Returns
55 -------
56 Type[rest_framework.response.Response]
57 HttpResponse acknowledging the success or failure of the state update operation.
58
59 """
60 msg = request.body
61 serializer = serializers.UpdateStateSerializer(data=request.data)
62
63 if serializer.is_valid() is not True:
64 return Response(serializer.errors, status=400)
65
66 # Check if the provided signature is valid against any secret in our list
67 if not validate_signature(request.headers.get("X-Marsha-Signature"), msg):
68 return Response("Forbidden", status=403)
69
70 # Retrieve the elements from the key
71 key_elements = serializer.get_key_elements()
72
73 # Update the object targeted by the "object_id" and "resource_id"
74 model_from_s3_identifier = get_uploadable_models_s3_mapping()
75 model = model_from_s3_identifier[key_elements["model_name"]]
76
77 extra_parameters = serializer.validated_data["extraParameters"]
78 if (
79 serializer.validated_data["state"] == defaults.READY
80 and hasattr(model, "extension")
81 and "extension" not in extra_parameters
82 ):
83 # The extension is part of the s3 key name and added in this key
84 # when generated by the initiate upload
85 extra_parameters["extension"] = key_elements.get("extension")
86
87 try:
88 object_instance = model.objects.get(id=key_elements["object_id"])
89 except model.DoesNotExist:
90 return Response({"success": False}, status=404)
91
92 object_instance.update_upload_state(
93 upload_state=serializer.validated_data["state"],
94 uploaded_on=key_elements.get("uploaded_on")
95 if serializer.validated_data["state"] == defaults.READY
96 else None,
97 **extra_parameters,
98 )
99
100 return Response({"success": True})
101
102
103 @api_view(["POST"])
104 def recording_slices_manifest(request):
105 """View handling AWS POST request to set a manifest on a record slice.
106
107 Parameters
108 ----------
109 request : Type[django.http.request.HttpRequest]
110 The request on the API endpoint, it should contain a payload with the following fields:
111 - video_id: the pk of a video.
112 - harvest_job_id: the id of the harvest job.
113 - manifest_key: the manifest key of the record slice.
114
115 Returns
116 -------
117 Type[rest_framework.response.Response]
118 HttpResponse containing the current harvest status of all recording slices.
119
120 """
121 # Check if the provided signature is valid against any secret in our list
122 if not validate_signature(request.headers.get("X-Marsha-Signature"), request.body):
123 return Response("Forbidden", status=403)
124
125 video = get_object_or_404(Video, pk=request.data["video_id"])
126 video.set_recording_slice_manifest_key(
127 request.data["harvest_job_id"], request.data["manifest_key"]
128 )
129 return Response({"success": True})
130
131
132 @api_view(["POST"])
133 def recording_slices_state(request):
134 """View handling AWS POST request to check each record slice harvest status by video pk.
135
136 Parameters
137 ----------
138 request : Type[django.http.request.HttpRequest]
139 The request on the API endpoint, it should contain a payload with the following fields:
140 - video_id: the pk of a video.
141
142 Returns
143 -------
144 Type[rest_framework.response.Response]
145 HttpResponse containing the current harvest status of all recording slices.
146
147 """
148 # Check if the provided signature is valid against any secret in our list
149 if not validate_signature(request.headers.get("X-Marsha-Signature"), request.body):
150 return Response("Forbidden", status=403)
151
152 video = get_object_or_404(Video, pk=request.data["video_id"])
153 return Response(video.get_recording_slices_state())
154
155
156 class APIViewMixin:
157 """
158 Mixin to enhance the base DRF APIView.
159
160 Must be used on every Marsha API views (APIView, Viewset, ...)
161
162 This provides a way to separate request information between `request.user`
163 and `request.resource`
164 """
165
166 def check_permissions(self, request):
167 """
168 Add resource attribute to the request when we are in a resource context.
169
170 Note, we use the `check_permissions` method to do it because:
171 - we need the authentication to be done to know whether we have a user
172 token or a resource token
173 - AND we also need to be before the call to `check_permissions` and we
174 can't do it in `self.initial`
175 - PLUS we can't be in `perform_authentication` because "metadata" views
176 partially duplicates the request (see `clone_request`) and our
177 `resource` won't be passed along. But `check_permissions`is called
178 after the cloning so this works.
179
180 Also, we can't alter the `request.user` without messing up with the
181 middleware called with `process_response` so in case of a resource,
182 we keep it as a TokenResource...
183 """
184 request.resource = None
185 if isinstance(request.auth, ResourceAccessToken): # otherwise, nothing to do
186 request.resource = request.user
187
188 super().check_permissions(request)
189
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/src/backend/marsha/core/api/base.py b/src/backend/marsha/core/api/base.py
--- a/src/backend/marsha/core/api/base.py
+++ b/src/backend/marsha/core/api/base.py
@@ -1,11 +1,14 @@
"""Declare API endpoints with Django RestFramework viewsets."""
-
+from django.conf import settings
+from django.http import JsonResponse
from django.shortcuts import get_object_or_404
from rest_framework.decorators import api_view
from rest_framework.response import Response
+from waffle import switch_is_active
from .. import defaults, serializers
+from ..defaults import SENTRY
from ..models import Video
from ..simple_jwt.tokens import ResourceAccessToken
from ..utils.api_utils import get_uploadable_models_s3_mapping, validate_signature
@@ -153,6 +156,30 @@
return Response(video.get_recording_slices_state())
+@api_view(["GET"])
+def get_frontend_configuration(request):
+ """View handling GET request to get the frontend configuration.
+
+ Parameters
+ ----------
+ request : Type[django.http.request.HttpRequest]
+ The request on the API endpoint.
+
+ Returns
+ -------
+ Type[rest_framework.response.Response]
+ HttpResponse containing the frontend configuration.
+
+ """
+ return JsonResponse(
+ {
+ "environment": settings.ENVIRONMENT,
+ "release": settings.RELEASE,
+ "sentry_dsn": settings.SENTRY_DSN if switch_is_active(SENTRY) else None,
+ }
+ )
+
+
class APIViewMixin:
"""
Mixin to enhance the base DRF APIView.
diff --git a/src/backend/marsha/urls.py b/src/backend/marsha/urls.py
--- a/src/backend/marsha/urls.py
+++ b/src/backend/marsha/urls.py
@@ -23,6 +23,7 @@
UserViewSet,
VideoViewSet,
XAPIStatementView,
+ get_frontend_configuration,
pairing_challenge,
recording_slices_manifest,
recording_slices_state,
@@ -131,6 +132,7 @@
recording_slices_state,
name="recording_slices_state",
),
+ path("api/config/", get_frontend_configuration, name="sentry_config"),
path("api/", include(router.urls)),
path(
f"api/{models.Video.RESOURCE_NAME}/<uuid:video_id>/",
| {"golden_diff": "diff --git a/src/backend/marsha/core/api/base.py b/src/backend/marsha/core/api/base.py\n--- a/src/backend/marsha/core/api/base.py\n+++ b/src/backend/marsha/core/api/base.py\n@@ -1,11 +1,14 @@\n \"\"\"Declare API endpoints with Django RestFramework viewsets.\"\"\"\n-\n+from django.conf import settings\n+from django.http import JsonResponse\n from django.shortcuts import get_object_or_404\n \n from rest_framework.decorators import api_view\n from rest_framework.response import Response\n+from waffle import switch_is_active\n \n from .. import defaults, serializers\n+from ..defaults import SENTRY\n from ..models import Video\n from ..simple_jwt.tokens import ResourceAccessToken\n from ..utils.api_utils import get_uploadable_models_s3_mapping, validate_signature\n@@ -153,6 +156,30 @@\n return Response(video.get_recording_slices_state())\n \n \n+@api_view([\"GET\"])\n+def get_frontend_configuration(request):\n+ \"\"\"View handling GET request to get the frontend configuration.\n+\n+ Parameters\n+ ----------\n+ request : Type[django.http.request.HttpRequest]\n+ The request on the API endpoint.\n+\n+ Returns\n+ -------\n+ Type[rest_framework.response.Response]\n+ HttpResponse containing the frontend configuration.\n+\n+ \"\"\"\n+ return JsonResponse(\n+ {\n+ \"environment\": settings.ENVIRONMENT,\n+ \"release\": settings.RELEASE,\n+ \"sentry_dsn\": settings.SENTRY_DSN if switch_is_active(SENTRY) else None,\n+ }\n+ )\n+\n+\n class APIViewMixin:\n \"\"\"\n Mixin to enhance the base DRF APIView.\ndiff --git a/src/backend/marsha/urls.py b/src/backend/marsha/urls.py\n--- a/src/backend/marsha/urls.py\n+++ b/src/backend/marsha/urls.py\n@@ -23,6 +23,7 @@\n UserViewSet,\n VideoViewSet,\n XAPIStatementView,\n+ get_frontend_configuration,\n pairing_challenge,\n recording_slices_manifest,\n recording_slices_state,\n@@ -131,6 +132,7 @@\n recording_slices_state,\n name=\"recording_slices_state\",\n ),\n+ path(\"api/config/\", get_frontend_configuration, name=\"sentry_config\"),\n path(\"api/\", include(router.urls)),\n path(\n f\"api/{models.Video.RESOURCE_NAME}/<uuid:video_id>/\",\n", "issue": "Configure sentry in standalone site\n## Feature Request\r\n\r\n**Is your feature request related to a problem or unsupported use case? Please describe.**\r\n\r\nSentry is not configured in the standalone site and we have no feedback is there are errors.\r\n\r\n**Describe the solution you'd like**\r\n\r\nFind a solution to configure sentry. Maybe by fetching th configuration on a dedicated api endpoint\r\n\r\n\n", "before_files": [{"content": "\"\"\"Marsha URLs configuration.\"\"\"\nimport re\n\nfrom django.conf import settings\nfrom django.contrib import admin\nfrom django.urls import include, path, re_path, register_converter\nfrom django.views.decorators.cache import cache_page\n\nfrom rest_framework.routers import DefaultRouter\n\nfrom marsha.core import models\nfrom marsha.core.api import (\n ChallengeAuthenticationView,\n DocumentViewSet,\n LiveSessionViewSet,\n OrganizationViewSet,\n PlaylistAccessViewSet,\n PlaylistViewSet,\n PortabilityResourceViewSet,\n SharedLiveMediaViewSet,\n ThumbnailViewSet,\n TimedTextTrackViewSet,\n UserViewSet,\n VideoViewSet,\n XAPIStatementView,\n pairing_challenge,\n recording_slices_manifest,\n recording_slices_state,\n update_state,\n)\nfrom marsha.core.api.lti_user_association import LtiUserAssociationViewSet\nfrom marsha.core.urls.converters import XAPIResourceKindConverter\nfrom marsha.core.utils.lti_select_utils import get_lti_select_resources\nfrom marsha.core.views import (\n DocumentLTIView,\n DocumentView,\n LTIConfigView,\n LTIRespondView,\n LTISelectView,\n RemindersCancelView,\n SiteView,\n VideoLTIView,\n VideoView,\n)\nfrom marsha.development.api import local_document_upload, local_video_upload\n\n\nregister_converter(XAPIResourceKindConverter, \"xapi_resource_kind\")\n\nLTI_SELECT_ROUTE_PATTERN = (\n rf\"lti/select/((?P<resource_kind>{'|'.join(get_lti_select_resources().keys())})/)?$\"\n)\n\nrouter = DefaultRouter()\nrouter.register(models.Video.RESOURCE_NAME, VideoViewSet, basename=\"videos\")\nrouter.register(models.Document.RESOURCE_NAME, DocumentViewSet, basename=\"documents\")\nrouter.register(\n models.TimedTextTrack.RESOURCE_NAME,\n TimedTextTrackViewSet,\n basename=\"timed_text_tracks\",\n)\nrouter.register(models.Thumbnail.RESOURCE_NAME, ThumbnailViewSet, basename=\"thumbnails\")\nrouter.register(\"organizations\", OrganizationViewSet, basename=\"organizations\")\nrouter.register(\"playlists\", PlaylistViewSet, basename=\"playlists\")\nrouter.register(\n \"playlist-accesses\", PlaylistAccessViewSet, basename=\"playlist_accesses\"\n)\nrouter.register(\n \"portability-requests\", PortabilityResourceViewSet, basename=\"portability_requests\"\n)\nrouter.register(\"users\", UserViewSet, basename=\"users\")\nrouter.register(\n \"lti-user-associations\", LtiUserAssociationViewSet, basename=\"lti_user_associations\"\n)\nrouter.register(\n models.SharedLiveMedia.RESOURCE_NAME,\n SharedLiveMediaViewSet,\n basename=\"sharedlivemedias\",\n)\n\n# Old routes to remove\nrouter.register(\n models.LiveSession.RESOURCE_NAME,\n LiveSessionViewSet,\n basename=\"live_sessions\",\n)\n\n# Video related resources (for nested routes)\nvideo_related_router = DefaultRouter()\nvideo_related_router.register(\n models.LiveSession.RESOURCE_NAME,\n LiveSessionViewSet,\n basename=\"live_sessions\",\n)\n\nurlpatterns = [\n # Account\n path(\"account/\", include(\"marsha.account.urls\")),\n # Admin\n path(\"admin/\", admin.site.urls),\n # LTI\n path(\"lti/config.xml\", LTIConfigView.as_view(), name=\"config_lti_view\"),\n re_path(\n LTI_SELECT_ROUTE_PATTERN,\n LTISelectView.as_view(),\n name=\"select_lti_view\",\n ),\n path(\"lti/respond/\", LTIRespondView.as_view(), name=\"respond_lti_view\"),\n path(\"lti/videos/<uuid:uuid>\", VideoLTIView.as_view(), name=\"video_lti_view\"),\n path(\n \"lti/documents/<uuid:uuid>\", DocumentLTIView.as_view(), name=\"document_lti_view\"\n ),\n # Public resources\n path(\"videos/<uuid:uuid>\", VideoView.as_view(), name=\"video_direct_access\"),\n path(\"documents/<uuid:uuid>\", DocumentView.as_view(), name=\"document_public\"),\n # API\n path(\n \"api/auth/challenge/\",\n ChallengeAuthenticationView.as_view(),\n name=\"api_authentication\",\n ),\n path(\"api/pairing-challenge\", pairing_challenge, name=\"pairing_challenge\"),\n path(\"api/update-state\", update_state, name=\"update_state\"),\n path(\n \"api/recording-slices-manifest\",\n recording_slices_manifest,\n name=\"recording_slices_manifest\",\n ),\n path(\n \"api/recording-slices-state\",\n recording_slices_state,\n name=\"recording_slices_state\",\n ),\n path(\"api/\", include(router.urls)),\n path(\n f\"api/{models.Video.RESOURCE_NAME}/<uuid:video_id>/\",\n include(video_related_router.urls),\n ),\n path(\n \"reminders/cancel/<str:pk>/<str:key>\",\n RemindersCancelView.as_view(),\n name=\"reminders_cancel\",\n ),\n path(\n \"xapi/<xapi_resource_kind:resource_kind>/<uuid:resource_id>/\",\n XAPIStatementView.as_view(),\n name=\"xapi\",\n ),\n]\n\nif settings.BBB_ENABLED:\n urlpatterns += [path(\"\", include(\"marsha.bbb.urls\"))]\n\nif settings.DEPOSIT_ENABLED:\n urlpatterns += [path(\"\", include(\"marsha.deposit.urls\"))]\n\nif settings.MARKDOWN_ENABLED:\n urlpatterns += [path(\"\", include(\"marsha.markdown.urls\"))]\n\nif settings.DEBUG:\n urlpatterns += [path(\"\", include(\"marsha.development.urls\"))]\n\nif \"dummy\" in settings.STORAGE_BACKEND:\n urlpatterns += [\n path(\n \"e2e/api/video-upload/<uuid:uuid>\",\n local_video_upload,\n name=\"local-video-upload\",\n ),\n path(\n \"api/document-upload/<uuid:uuid>\",\n local_document_upload,\n name=\"local-document-upload\",\n ),\n ]\n\nstatic_path = re.escape(settings.STATIC_URL.lstrip(\"/\"))\nmedia_path = re.escape(settings.MEDIA_URL.lstrip(\"/\"))\nSITE_IGNORE_PREFIX = \"|\".join([static_path, media_path])\n\nurlpatterns += [\n re_path(\n # Catch all URLs that are not handled before\n # and which do not regard static files or media files\n rf\"^(?!{SITE_IGNORE_PREFIX}).*\",\n cache_page(86400, key_prefix=settings.RELEASE)(SiteView.as_view()),\n name=\"site\",\n ),\n]\n", "path": "src/backend/marsha/urls.py"}, {"content": "\"\"\"Declare API endpoints with Django RestFramework viewsets.\"\"\"\n\nfrom django.shortcuts import get_object_or_404\n\nfrom rest_framework.decorators import api_view\nfrom rest_framework.response import Response\n\nfrom .. import defaults, serializers\nfrom ..models import Video\nfrom ..simple_jwt.tokens import ResourceAccessToken\nfrom ..utils.api_utils import get_uploadable_models_s3_mapping, validate_signature\n\n\nclass ObjectPkMixin:\n \"\"\"\n Get the object primary key from the URL path.\n\n This is useful to avoid making extra requests using view.get_object() on\n a ViewSet when we only need the object's id, which is available in the URL.\n \"\"\"\n\n def get_object_pk(self):\n \"\"\"Get the object primary key from the URL path.\"\"\"\n lookup_url_kwarg = self.lookup_url_kwarg or self.lookup_field\n return self.kwargs.get(lookup_url_kwarg)\n\n\nclass ObjectRelatedMixin:\n \"\"\"\n Get the related video belonging to the current object.\n\n Using view.get_object is permissions is not possible anymore due to\n infinite recursion between has_permission and has_object_permission call.\n \"\"\"\n\n def get_related_object(self):\n \"\"\"Get the video related to the current object.\"\"\"\n queryset = self.filter_queryset(self.get_queryset())\n return queryset.get(pk=self.get_object_pk())\n\n\n@api_view([\"POST\"])\ndef update_state(request):\n \"\"\"View handling AWS POST request to update the state of an object by key.\n\n Parameters\n ----------\n request : Type[django.http.request.HttpRequest]\n The request on the API endpoint, it should contain a payload with the following fields:\n - key: the key of an object in the source bucket as delivered in the upload policy,\n - state: state of the upload, should be either \"ready\" or \"error\",\n - extraParameters: Dict containing arbitrary data sent from AWS Lambda.\n\n Returns\n -------\n Type[rest_framework.response.Response]\n HttpResponse acknowledging the success or failure of the state update operation.\n\n \"\"\"\n msg = request.body\n serializer = serializers.UpdateStateSerializer(data=request.data)\n\n if serializer.is_valid() is not True:\n return Response(serializer.errors, status=400)\n\n # Check if the provided signature is valid against any secret in our list\n if not validate_signature(request.headers.get(\"X-Marsha-Signature\"), msg):\n return Response(\"Forbidden\", status=403)\n\n # Retrieve the elements from the key\n key_elements = serializer.get_key_elements()\n\n # Update the object targeted by the \"object_id\" and \"resource_id\"\n model_from_s3_identifier = get_uploadable_models_s3_mapping()\n model = model_from_s3_identifier[key_elements[\"model_name\"]]\n\n extra_parameters = serializer.validated_data[\"extraParameters\"]\n if (\n serializer.validated_data[\"state\"] == defaults.READY\n and hasattr(model, \"extension\")\n and \"extension\" not in extra_parameters\n ):\n # The extension is part of the s3 key name and added in this key\n # when generated by the initiate upload\n extra_parameters[\"extension\"] = key_elements.get(\"extension\")\n\n try:\n object_instance = model.objects.get(id=key_elements[\"object_id\"])\n except model.DoesNotExist:\n return Response({\"success\": False}, status=404)\n\n object_instance.update_upload_state(\n upload_state=serializer.validated_data[\"state\"],\n uploaded_on=key_elements.get(\"uploaded_on\")\n if serializer.validated_data[\"state\"] == defaults.READY\n else None,\n **extra_parameters,\n )\n\n return Response({\"success\": True})\n\n\n@api_view([\"POST\"])\ndef recording_slices_manifest(request):\n \"\"\"View handling AWS POST request to set a manifest on a record slice.\n\n Parameters\n ----------\n request : Type[django.http.request.HttpRequest]\n The request on the API endpoint, it should contain a payload with the following fields:\n - video_id: the pk of a video.\n - harvest_job_id: the id of the harvest job.\n - manifest_key: the manifest key of the record slice.\n\n Returns\n -------\n Type[rest_framework.response.Response]\n HttpResponse containing the current harvest status of all recording slices.\n\n \"\"\"\n # Check if the provided signature is valid against any secret in our list\n if not validate_signature(request.headers.get(\"X-Marsha-Signature\"), request.body):\n return Response(\"Forbidden\", status=403)\n\n video = get_object_or_404(Video, pk=request.data[\"video_id\"])\n video.set_recording_slice_manifest_key(\n request.data[\"harvest_job_id\"], request.data[\"manifest_key\"]\n )\n return Response({\"success\": True})\n\n\n@api_view([\"POST\"])\ndef recording_slices_state(request):\n \"\"\"View handling AWS POST request to check each record slice harvest status by video pk.\n\n Parameters\n ----------\n request : Type[django.http.request.HttpRequest]\n The request on the API endpoint, it should contain a payload with the following fields:\n - video_id: the pk of a video.\n\n Returns\n -------\n Type[rest_framework.response.Response]\n HttpResponse containing the current harvest status of all recording slices.\n\n \"\"\"\n # Check if the provided signature is valid against any secret in our list\n if not validate_signature(request.headers.get(\"X-Marsha-Signature\"), request.body):\n return Response(\"Forbidden\", status=403)\n\n video = get_object_or_404(Video, pk=request.data[\"video_id\"])\n return Response(video.get_recording_slices_state())\n\n\nclass APIViewMixin:\n \"\"\"\n Mixin to enhance the base DRF APIView.\n\n Must be used on every Marsha API views (APIView, Viewset, ...)\n\n This provides a way to separate request information between `request.user`\n and `request.resource`\n \"\"\"\n\n def check_permissions(self, request):\n \"\"\"\n Add resource attribute to the request when we are in a resource context.\n\n Note, we use the `check_permissions` method to do it because:\n - we need the authentication to be done to know whether we have a user\n token or a resource token\n - AND we also need to be before the call to `check_permissions` and we\n can't do it in `self.initial`\n - PLUS we can't be in `perform_authentication` because \"metadata\" views\n partially duplicates the request (see `clone_request`) and our\n `resource` won't be passed along. But `check_permissions`is called\n after the cloning so this works.\n\n Also, we can't alter the `request.user` without messing up with the\n middleware called with `process_response` so in case of a resource,\n we keep it as a TokenResource...\n \"\"\"\n request.resource = None\n if isinstance(request.auth, ResourceAccessToken): # otherwise, nothing to do\n request.resource = request.user\n\n super().check_permissions(request)\n", "path": "src/backend/marsha/core/api/base.py"}], "after_files": [{"content": "\"\"\"Marsha URLs configuration.\"\"\"\nimport re\n\nfrom django.conf import settings\nfrom django.contrib import admin\nfrom django.urls import include, path, re_path, register_converter\nfrom django.views.decorators.cache import cache_page\n\nfrom rest_framework.routers import DefaultRouter\n\nfrom marsha.core import models\nfrom marsha.core.api import (\n ChallengeAuthenticationView,\n DocumentViewSet,\n LiveSessionViewSet,\n OrganizationViewSet,\n PlaylistAccessViewSet,\n PlaylistViewSet,\n PortabilityResourceViewSet,\n SharedLiveMediaViewSet,\n ThumbnailViewSet,\n TimedTextTrackViewSet,\n UserViewSet,\n VideoViewSet,\n XAPIStatementView,\n get_frontend_configuration,\n pairing_challenge,\n recording_slices_manifest,\n recording_slices_state,\n update_state,\n)\nfrom marsha.core.api.lti_user_association import LtiUserAssociationViewSet\nfrom marsha.core.urls.converters import XAPIResourceKindConverter\nfrom marsha.core.utils.lti_select_utils import get_lti_select_resources\nfrom marsha.core.views import (\n DocumentLTIView,\n DocumentView,\n LTIConfigView,\n LTIRespondView,\n LTISelectView,\n RemindersCancelView,\n SiteView,\n VideoLTIView,\n VideoView,\n)\nfrom marsha.development.api import local_document_upload, local_video_upload\n\n\nregister_converter(XAPIResourceKindConverter, \"xapi_resource_kind\")\n\nLTI_SELECT_ROUTE_PATTERN = (\n rf\"lti/select/((?P<resource_kind>{'|'.join(get_lti_select_resources().keys())})/)?$\"\n)\n\nrouter = DefaultRouter()\nrouter.register(models.Video.RESOURCE_NAME, VideoViewSet, basename=\"videos\")\nrouter.register(models.Document.RESOURCE_NAME, DocumentViewSet, basename=\"documents\")\nrouter.register(\n models.TimedTextTrack.RESOURCE_NAME,\n TimedTextTrackViewSet,\n basename=\"timed_text_tracks\",\n)\nrouter.register(models.Thumbnail.RESOURCE_NAME, ThumbnailViewSet, basename=\"thumbnails\")\nrouter.register(\"organizations\", OrganizationViewSet, basename=\"organizations\")\nrouter.register(\"playlists\", PlaylistViewSet, basename=\"playlists\")\nrouter.register(\n \"playlist-accesses\", PlaylistAccessViewSet, basename=\"playlist_accesses\"\n)\nrouter.register(\n \"portability-requests\", PortabilityResourceViewSet, basename=\"portability_requests\"\n)\nrouter.register(\"users\", UserViewSet, basename=\"users\")\nrouter.register(\n \"lti-user-associations\", LtiUserAssociationViewSet, basename=\"lti_user_associations\"\n)\nrouter.register(\n models.SharedLiveMedia.RESOURCE_NAME,\n SharedLiveMediaViewSet,\n basename=\"sharedlivemedias\",\n)\n\n# Old routes to remove\nrouter.register(\n models.LiveSession.RESOURCE_NAME,\n LiveSessionViewSet,\n basename=\"live_sessions\",\n)\n\n# Video related resources (for nested routes)\nvideo_related_router = DefaultRouter()\nvideo_related_router.register(\n models.LiveSession.RESOURCE_NAME,\n LiveSessionViewSet,\n basename=\"live_sessions\",\n)\n\nurlpatterns = [\n # Account\n path(\"account/\", include(\"marsha.account.urls\")),\n # Admin\n path(\"admin/\", admin.site.urls),\n # LTI\n path(\"lti/config.xml\", LTIConfigView.as_view(), name=\"config_lti_view\"),\n re_path(\n LTI_SELECT_ROUTE_PATTERN,\n LTISelectView.as_view(),\n name=\"select_lti_view\",\n ),\n path(\"lti/respond/\", LTIRespondView.as_view(), name=\"respond_lti_view\"),\n path(\"lti/videos/<uuid:uuid>\", VideoLTIView.as_view(), name=\"video_lti_view\"),\n path(\n \"lti/documents/<uuid:uuid>\", DocumentLTIView.as_view(), name=\"document_lti_view\"\n ),\n # Public resources\n path(\"videos/<uuid:uuid>\", VideoView.as_view(), name=\"video_direct_access\"),\n path(\"documents/<uuid:uuid>\", DocumentView.as_view(), name=\"document_public\"),\n # API\n path(\n \"api/auth/challenge/\",\n ChallengeAuthenticationView.as_view(),\n name=\"api_authentication\",\n ),\n path(\"api/pairing-challenge\", pairing_challenge, name=\"pairing_challenge\"),\n path(\"api/update-state\", update_state, name=\"update_state\"),\n path(\n \"api/recording-slices-manifest\",\n recording_slices_manifest,\n name=\"recording_slices_manifest\",\n ),\n path(\n \"api/recording-slices-state\",\n recording_slices_state,\n name=\"recording_slices_state\",\n ),\n path(\"api/config/\", get_frontend_configuration, name=\"sentry_config\"),\n path(\"api/\", include(router.urls)),\n path(\n f\"api/{models.Video.RESOURCE_NAME}/<uuid:video_id>/\",\n include(video_related_router.urls),\n ),\n path(\n \"reminders/cancel/<str:pk>/<str:key>\",\n RemindersCancelView.as_view(),\n name=\"reminders_cancel\",\n ),\n path(\n \"xapi/<xapi_resource_kind:resource_kind>/<uuid:resource_id>/\",\n XAPIStatementView.as_view(),\n name=\"xapi\",\n ),\n]\n\nif settings.BBB_ENABLED:\n urlpatterns += [path(\"\", include(\"marsha.bbb.urls\"))]\n\nif settings.DEPOSIT_ENABLED:\n urlpatterns += [path(\"\", include(\"marsha.deposit.urls\"))]\n\nif settings.MARKDOWN_ENABLED:\n urlpatterns += [path(\"\", include(\"marsha.markdown.urls\"))]\n\nif settings.DEBUG:\n urlpatterns += [path(\"\", include(\"marsha.development.urls\"))]\n\nif \"dummy\" in settings.STORAGE_BACKEND:\n urlpatterns += [\n path(\n \"e2e/api/video-upload/<uuid:uuid>\",\n local_video_upload,\n name=\"local-video-upload\",\n ),\n path(\n \"api/document-upload/<uuid:uuid>\",\n local_document_upload,\n name=\"local-document-upload\",\n ),\n ]\n\nstatic_path = re.escape(settings.STATIC_URL.lstrip(\"/\"))\nmedia_path = re.escape(settings.MEDIA_URL.lstrip(\"/\"))\nSITE_IGNORE_PREFIX = \"|\".join([static_path, media_path])\n\nurlpatterns += [\n re_path(\n # Catch all URLs that are not handled before\n # and which do not regard static files or media files\n rf\"^(?!{SITE_IGNORE_PREFIX}).*\",\n cache_page(86400, key_prefix=settings.RELEASE)(SiteView.as_view()),\n name=\"site\",\n ),\n]\n", "path": "src/backend/marsha/urls.py"}, {"content": "\"\"\"Declare API endpoints with Django RestFramework viewsets.\"\"\"\nfrom django.conf import settings\nfrom django.http import JsonResponse\nfrom django.shortcuts import get_object_or_404\n\nfrom rest_framework.decorators import api_view\nfrom rest_framework.response import Response\nfrom waffle import switch_is_active\n\nfrom .. import defaults, serializers\nfrom ..defaults import SENTRY\nfrom ..models import Video\nfrom ..simple_jwt.tokens import ResourceAccessToken\nfrom ..utils.api_utils import get_uploadable_models_s3_mapping, validate_signature\n\n\nclass ObjectPkMixin:\n \"\"\"\n Get the object primary key from the URL path.\n\n This is useful to avoid making extra requests using view.get_object() on\n a ViewSet when we only need the object's id, which is available in the URL.\n \"\"\"\n\n def get_object_pk(self):\n \"\"\"Get the object primary key from the URL path.\"\"\"\n lookup_url_kwarg = self.lookup_url_kwarg or self.lookup_field\n return self.kwargs.get(lookup_url_kwarg)\n\n\nclass ObjectRelatedMixin:\n \"\"\"\n Get the related video belonging to the current object.\n\n Using view.get_object is permissions is not possible anymore due to\n infinite recursion between has_permission and has_object_permission call.\n \"\"\"\n\n def get_related_object(self):\n \"\"\"Get the video related to the current object.\"\"\"\n queryset = self.filter_queryset(self.get_queryset())\n return queryset.get(pk=self.get_object_pk())\n\n\n@api_view([\"POST\"])\ndef update_state(request):\n \"\"\"View handling AWS POST request to update the state of an object by key.\n\n Parameters\n ----------\n request : Type[django.http.request.HttpRequest]\n The request on the API endpoint, it should contain a payload with the following fields:\n - key: the key of an object in the source bucket as delivered in the upload policy,\n - state: state of the upload, should be either \"ready\" or \"error\",\n - extraParameters: Dict containing arbitrary data sent from AWS Lambda.\n\n Returns\n -------\n Type[rest_framework.response.Response]\n HttpResponse acknowledging the success or failure of the state update operation.\n\n \"\"\"\n msg = request.body\n serializer = serializers.UpdateStateSerializer(data=request.data)\n\n if serializer.is_valid() is not True:\n return Response(serializer.errors, status=400)\n\n # Check if the provided signature is valid against any secret in our list\n if not validate_signature(request.headers.get(\"X-Marsha-Signature\"), msg):\n return Response(\"Forbidden\", status=403)\n\n # Retrieve the elements from the key\n key_elements = serializer.get_key_elements()\n\n # Update the object targeted by the \"object_id\" and \"resource_id\"\n model_from_s3_identifier = get_uploadable_models_s3_mapping()\n model = model_from_s3_identifier[key_elements[\"model_name\"]]\n\n extra_parameters = serializer.validated_data[\"extraParameters\"]\n if (\n serializer.validated_data[\"state\"] == defaults.READY\n and hasattr(model, \"extension\")\n and \"extension\" not in extra_parameters\n ):\n # The extension is part of the s3 key name and added in this key\n # when generated by the initiate upload\n extra_parameters[\"extension\"] = key_elements.get(\"extension\")\n\n try:\n object_instance = model.objects.get(id=key_elements[\"object_id\"])\n except model.DoesNotExist:\n return Response({\"success\": False}, status=404)\n\n object_instance.update_upload_state(\n upload_state=serializer.validated_data[\"state\"],\n uploaded_on=key_elements.get(\"uploaded_on\")\n if serializer.validated_data[\"state\"] == defaults.READY\n else None,\n **extra_parameters,\n )\n\n return Response({\"success\": True})\n\n\n@api_view([\"POST\"])\ndef recording_slices_manifest(request):\n \"\"\"View handling AWS POST request to set a manifest on a record slice.\n\n Parameters\n ----------\n request : Type[django.http.request.HttpRequest]\n The request on the API endpoint, it should contain a payload with the following fields:\n - video_id: the pk of a video.\n - harvest_job_id: the id of the harvest job.\n - manifest_key: the manifest key of the record slice.\n\n Returns\n -------\n Type[rest_framework.response.Response]\n HttpResponse containing the current harvest status of all recording slices.\n\n \"\"\"\n # Check if the provided signature is valid against any secret in our list\n if not validate_signature(request.headers.get(\"X-Marsha-Signature\"), request.body):\n return Response(\"Forbidden\", status=403)\n\n video = get_object_or_404(Video, pk=request.data[\"video_id\"])\n video.set_recording_slice_manifest_key(\n request.data[\"harvest_job_id\"], request.data[\"manifest_key\"]\n )\n return Response({\"success\": True})\n\n\n@api_view([\"POST\"])\ndef recording_slices_state(request):\n \"\"\"View handling AWS POST request to check each record slice harvest status by video pk.\n\n Parameters\n ----------\n request : Type[django.http.request.HttpRequest]\n The request on the API endpoint, it should contain a payload with the following fields:\n - video_id: the pk of a video.\n\n Returns\n -------\n Type[rest_framework.response.Response]\n HttpResponse containing the current harvest status of all recording slices.\n\n \"\"\"\n # Check if the provided signature is valid against any secret in our list\n if not validate_signature(request.headers.get(\"X-Marsha-Signature\"), request.body):\n return Response(\"Forbidden\", status=403)\n\n video = get_object_or_404(Video, pk=request.data[\"video_id\"])\n return Response(video.get_recording_slices_state())\n\n\n@api_view([\"GET\"])\ndef get_frontend_configuration(request):\n \"\"\"View handling GET request to get the frontend configuration.\n\n Parameters\n ----------\n request : Type[django.http.request.HttpRequest]\n The request on the API endpoint.\n\n Returns\n -------\n Type[rest_framework.response.Response]\n HttpResponse containing the frontend configuration.\n\n \"\"\"\n return JsonResponse(\n {\n \"environment\": settings.ENVIRONMENT,\n \"release\": settings.RELEASE,\n \"sentry_dsn\": settings.SENTRY_DSN if switch_is_active(SENTRY) else None,\n }\n )\n\n\nclass APIViewMixin:\n \"\"\"\n Mixin to enhance the base DRF APIView.\n\n Must be used on every Marsha API views (APIView, Viewset, ...)\n\n This provides a way to separate request information between `request.user`\n and `request.resource`\n \"\"\"\n\n def check_permissions(self, request):\n \"\"\"\n Add resource attribute to the request when we are in a resource context.\n\n Note, we use the `check_permissions` method to do it because:\n - we need the authentication to be done to know whether we have a user\n token or a resource token\n - AND we also need to be before the call to `check_permissions` and we\n can't do it in `self.initial`\n - PLUS we can't be in `perform_authentication` because \"metadata\" views\n partially duplicates the request (see `clone_request`) and our\n `resource` won't be passed along. But `check_permissions`is called\n after the cloning so this works.\n\n Also, we can't alter the `request.user` without messing up with the\n middleware called with `process_response` so in case of a resource,\n we keep it as a TokenResource...\n \"\"\"\n request.resource = None\n if isinstance(request.auth, ResourceAccessToken): # otherwise, nothing to do\n request.resource = request.user\n\n super().check_permissions(request)\n", "path": "src/backend/marsha/core/api/base.py"}]} | 4,086 | 524 |
gh_patches_debug_8092 | rasdani/github-patches | git_diff | vega__altair-1907 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Interval Selection Example Bug
I am having trouble with the the [Interval Selection Example](https://altair-viz.github.io/gallery/interval_selection.html).

```python
import altair as alt
from vega_datasets import data
source = data.sp500.url
brush = alt.selection(type='interval', encodings=['x'])
upper = alt.Chart(source).mark_area().encode(
alt.X('date:T', scale=alt.Scale(domain=brush)),
y='price:Q'
).properties(
width=600,
height=200
)
lower = upper.properties(
height=60
).add_selection(brush)
upper & lower
```
The example looks correct so I am unsure what is causing this behavior.
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `altair/examples/interval_selection.py`
Content:
```
1 """
2 Interval Selection Example
3 ==========================
4
5 This is an example of creating a stacked chart for which the domain of the
6 top chart can be selected by interacting with the bottom chart.
7 """
8 # category: area charts
9 import altair as alt
10 from vega_datasets import data
11
12 source = data.sp500.url
13
14 brush = alt.selection(type='interval', encodings=['x'])
15
16 upper = alt.Chart(source).mark_area().encode(
17 alt.X('date:T', scale=alt.Scale(domain=brush)),
18 y='price:Q'
19 ).properties(
20 width=600,
21 height=200
22 )
23
24 lower = upper.properties(
25 height=60
26 ).add_selection(brush)
27
28 upper & lower
29
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/altair/examples/interval_selection.py b/altair/examples/interval_selection.py
--- a/altair/examples/interval_selection.py
+++ b/altair/examples/interval_selection.py
@@ -13,15 +13,19 @@
brush = alt.selection(type='interval', encodings=['x'])
-upper = alt.Chart(source).mark_area().encode(
- alt.X('date:T', scale=alt.Scale(domain=brush)),
- y='price:Q'
+base = alt.Chart(source).mark_area().encode(
+ x = 'date:T',
+ y = 'price:Q'
).properties(
width=600,
height=200
)
-lower = upper.properties(
+upper = base.encode(
+ alt.X('date:T', scale=alt.Scale(domain=brush))
+)
+
+lower = base.properties(
height=60
).add_selection(brush)
| {"golden_diff": "diff --git a/altair/examples/interval_selection.py b/altair/examples/interval_selection.py\n--- a/altair/examples/interval_selection.py\n+++ b/altair/examples/interval_selection.py\n@@ -13,15 +13,19 @@\n \n brush = alt.selection(type='interval', encodings=['x'])\n \n-upper = alt.Chart(source).mark_area().encode(\n- alt.X('date:T', scale=alt.Scale(domain=brush)),\n- y='price:Q'\n+base = alt.Chart(source).mark_area().encode(\n+ x = 'date:T',\n+ y = 'price:Q'\n ).properties(\n width=600,\n height=200\n )\n \n-lower = upper.properties(\n+upper = base.encode(\n+ alt.X('date:T', scale=alt.Scale(domain=brush))\n+)\n+\n+lower = base.properties(\n height=60\n ).add_selection(brush)\n", "issue": "Interval Selection Example Bug\nI am having trouble with the the [Interval Selection Example](https://altair-viz.github.io/gallery/interval_selection.html). \r\n\r\n\r\n```python\r\nimport altair as alt\r\nfrom vega_datasets import data\r\n\r\nsource = data.sp500.url\r\n\r\nbrush = alt.selection(type='interval', encodings=['x'])\r\n\r\nupper = alt.Chart(source).mark_area().encode(\r\n alt.X('date:T', scale=alt.Scale(domain=brush)),\r\n y='price:Q'\r\n).properties(\r\n width=600,\r\n height=200\r\n)\r\n\r\nlower = upper.properties(\r\n height=60\r\n).add_selection(brush)\r\n\r\nupper & lower\r\n```\r\n\r\nThe example looks correct so I am unsure what is causing this behavior. \n", "before_files": [{"content": "\"\"\"\nInterval Selection Example\n==========================\n\nThis is an example of creating a stacked chart for which the domain of the\ntop chart can be selected by interacting with the bottom chart.\n\"\"\"\n# category: area charts\nimport altair as alt\nfrom vega_datasets import data\n\nsource = data.sp500.url\n\nbrush = alt.selection(type='interval', encodings=['x'])\n\nupper = alt.Chart(source).mark_area().encode(\n alt.X('date:T', scale=alt.Scale(domain=brush)),\n y='price:Q'\n).properties(\n width=600,\n height=200\n)\n\nlower = upper.properties(\n height=60\n).add_selection(brush)\n\nupper & lower\n", "path": "altair/examples/interval_selection.py"}], "after_files": [{"content": "\"\"\"\nInterval Selection Example\n==========================\n\nThis is an example of creating a stacked chart for which the domain of the\ntop chart can be selected by interacting with the bottom chart.\n\"\"\"\n# category: area charts\nimport altair as alt\nfrom vega_datasets import data\n\nsource = data.sp500.url\n\nbrush = alt.selection(type='interval', encodings=['x'])\n\nbase = alt.Chart(source).mark_area().encode(\n x = 'date:T',\n y = 'price:Q'\n).properties(\n width=600,\n height=200\n)\n\nupper = base.encode(\n alt.X('date:T', scale=alt.Scale(domain=brush))\n)\n\nlower = base.properties(\n height=60\n).add_selection(brush)\n\nupper & lower\n", "path": "altair/examples/interval_selection.py"}]} | 687 | 205 |
gh_patches_debug_21929 | rasdani/github-patches | git_diff | Lightning-Universe__lightning-flash-210 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
NLTK being loaded on image classifcation
## 🐛 Bug
<!-- A clear and concise description of what the bug is. -->
### To Reproduce
```python
from flash.data import labels_from_csv
from flash.vision import ImageClassificationData
from flash.vision import ImageClassifier
from flash import Trainer
```
<!-- If you have a code sample, error messages, stack traces, please provide it here as well -->
[nltk_data] Error loading punkt: <urlopen error [Errno -3] Temporary
[nltk_data] failure in name resolution>
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `flash/text/seq2seq/summarization/metric.py`
Content:
```
1 # Copyright The PyTorch Lightning team.
2 #
3 # Licensed under the Apache License, Version 2.0 (the "License");
4 # you may not use this file except in compliance with the License.
5 # You may obtain a copy of the License at
6 #
7 # http://www.apache.org/licenses/LICENSE-2.0
8 #
9 # Unless required by applicable law or agreed to in writing, software
10 # distributed under the License is distributed on an "AS IS" BASIS,
11 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12 # See the License for the specific language governing permissions and
13 # limitations under the License.
14 from typing import Dict, List, Tuple
15
16 import numpy as np
17 from rouge_score import rouge_scorer, scoring
18 from rouge_score.scoring import AggregateScore, Score
19 from torch import tensor
20 from torchmetrics import Metric
21
22 from flash.text.seq2seq.summarization.utils import add_newline_to_end_of_each_sentence
23
24
25 class RougeMetric(Metric):
26 """
27 Metric used for automatic summarization. https://www.aclweb.org/anthology/W04-1013/
28
29 Example:
30
31 >>> target = "Is your name John".split()
32 >>> preds = "My name is John".split()
33 >>> rouge = RougeMetric()
34 >>> from pprint import pprint
35 >>> pprint(rouge(preds, target)) # doctest: +NORMALIZE_WHITESPACE
36 {'rouge1_fmeasure': 0.25,
37 'rouge1_precision': 0.25,
38 'rouge1_recall': 0.25,
39 'rouge2_fmeasure': 0.0,
40 'rouge2_precision': 0.0,
41 'rouge2_recall': 0.0,
42 'rougeL_fmeasure': 0.25,
43 'rougeL_precision': 0.25,
44 'rougeL_recall': 0.25,
45 'rougeLsum_fmeasure': 0.25,
46 'rougeLsum_precision': 0.25,
47 'rougeLsum_recall': 0.25}
48 """
49
50 def __init__(
51 self,
52 rouge_newline_sep: bool = False,
53 use_stemmer: bool = False,
54 rouge_keys: Tuple[str] = ("rouge1", "rouge2", "rougeL", "rougeLsum"),
55 ):
56 super().__init__()
57 self.rouge_newline_sep = rouge_newline_sep
58 self.rouge_keys = rouge_keys
59 self.use_stemmer = use_stemmer
60 self.aggregator = RougeBatchAggregator()
61 self.scorer = rouge_scorer.RougeScorer(rouge_keys, use_stemmer=self.use_stemmer)
62
63 for key in rouge_keys:
64 self.add_state(key, [])
65
66 def update(self, pred_lns: List[str], tgt_lns: List[str]):
67 for pred, tgt in zip(pred_lns, tgt_lns):
68 # rougeLsum expects "\n" separated sentences within a summary
69 if self.rouge_newline_sep:
70 pred = add_newline_to_end_of_each_sentence(pred)
71 tgt = add_newline_to_end_of_each_sentence(tgt)
72 results = self.scorer.score(pred, tgt)
73 for key, score in results.items():
74 score = tensor([score.precision, score.recall, score.fmeasure])
75 getattr(self, key).append(score)
76
77 def compute(self) -> Dict[str, float]:
78 scores = {key: getattr(self, key) for key in self.rouge_keys}
79 self.aggregator.add_scores(scores)
80 result = self.aggregator.aggregate()
81 return format_rouge_results(result)
82
83 def __hash__(self):
84 # override to hash list objects.
85 # this is a bug in the upstream pytorch release.
86 hash_vals = [self.__class__.__name__]
87
88 for key in self._defaults.keys():
89 value = getattr(self, key)
90 if isinstance(value, list):
91 value = tuple(value)
92 hash_vals.append(value)
93
94 return hash(tuple(hash_vals))
95
96
97 class RougeBatchAggregator(scoring.BootstrapAggregator):
98 """
99 Aggregates rouge scores and provides confidence intervals.
100 """
101
102 def aggregate(self):
103 """
104 Override function to wrap the final results in `Score` objects.
105 This is due to the scores being replaced with a list of torch tensors.
106 """
107 result = {}
108 for score_type, scores in self._scores.items():
109 # Stack scores into a 2-d matrix of (sample, measure).
110 score_matrix = np.vstack(tuple(scores))
111 # Percentiles are returned as (interval, measure).
112 percentiles = self._bootstrap_resample(score_matrix)
113 # Extract the three intervals (low, mid, high).
114 intervals = tuple((Score(*percentiles[j, :]) for j in range(3)))
115 result[score_type] = AggregateScore(low=intervals[0], mid=intervals[1], high=intervals[2])
116 return result
117
118 def add_scores(self, scores):
119 self._scores = scores
120
121
122 def format_rouge_results(result: Dict[str, AggregateScore], decimal_places: int = 4) -> Dict[str, float]:
123 flattened_result = {}
124 for rouge_key, rouge_aggregate_score in result.items():
125 for stat in ["precision", "recall", "fmeasure"]:
126 mid = rouge_aggregate_score.mid
127 score = round(getattr(mid, stat), decimal_places)
128 flattened_result[f"{rouge_key}_{stat}"] = score
129 return flattened_result
130
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/flash/text/seq2seq/summarization/metric.py b/flash/text/seq2seq/summarization/metric.py
--- a/flash/text/seq2seq/summarization/metric.py
+++ b/flash/text/seq2seq/summarization/metric.py
@@ -19,7 +19,7 @@
from torch import tensor
from torchmetrics import Metric
-from flash.text.seq2seq.summarization.utils import add_newline_to_end_of_each_sentence
+from flash.text.seq2seq import summarization
class RougeMetric(Metric):
@@ -67,8 +67,8 @@
for pred, tgt in zip(pred_lns, tgt_lns):
# rougeLsum expects "\n" separated sentences within a summary
if self.rouge_newline_sep:
- pred = add_newline_to_end_of_each_sentence(pred)
- tgt = add_newline_to_end_of_each_sentence(tgt)
+ pred = summarization.utils.add_newline_to_end_of_each_sentence(pred)
+ tgt = summarization.utils.add_newline_to_end_of_each_sentence(tgt)
results = self.scorer.score(pred, tgt)
for key, score in results.items():
score = tensor([score.precision, score.recall, score.fmeasure])
| {"golden_diff": "diff --git a/flash/text/seq2seq/summarization/metric.py b/flash/text/seq2seq/summarization/metric.py\n--- a/flash/text/seq2seq/summarization/metric.py\n+++ b/flash/text/seq2seq/summarization/metric.py\n@@ -19,7 +19,7 @@\n from torch import tensor\n from torchmetrics import Metric\n \n-from flash.text.seq2seq.summarization.utils import add_newline_to_end_of_each_sentence\n+from flash.text.seq2seq import summarization\n \n \n class RougeMetric(Metric):\n@@ -67,8 +67,8 @@\n for pred, tgt in zip(pred_lns, tgt_lns):\n # rougeLsum expects \"\\n\" separated sentences within a summary\n if self.rouge_newline_sep:\n- pred = add_newline_to_end_of_each_sentence(pred)\n- tgt = add_newline_to_end_of_each_sentence(tgt)\n+ pred = summarization.utils.add_newline_to_end_of_each_sentence(pred)\n+ tgt = summarization.utils.add_newline_to_end_of_each_sentence(tgt)\n results = self.scorer.score(pred, tgt)\n for key, score in results.items():\n score = tensor([score.precision, score.recall, score.fmeasure])\n", "issue": "NLTK being loaded on image classifcation\n## \ud83d\udc1b Bug\r\n\r\n<!-- A clear and concise description of what the bug is. -->\r\n\r\n### To Reproduce\r\n\r\n\r\n```python\r\nfrom flash.data import labels_from_csv\r\nfrom flash.vision import ImageClassificationData\r\nfrom flash.vision import ImageClassifier\r\nfrom flash import Trainer\r\n```\r\n<!-- If you have a code sample, error messages, stack traces, please provide it here as well -->\r\n\r\n[nltk_data] Error loading punkt: <urlopen error [Errno -3] Temporary\r\n[nltk_data] failure in name resolution>\r\n\r\n\n", "before_files": [{"content": "# Copyright The PyTorch Lightning team.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\nfrom typing import Dict, List, Tuple\n\nimport numpy as np\nfrom rouge_score import rouge_scorer, scoring\nfrom rouge_score.scoring import AggregateScore, Score\nfrom torch import tensor\nfrom torchmetrics import Metric\n\nfrom flash.text.seq2seq.summarization.utils import add_newline_to_end_of_each_sentence\n\n\nclass RougeMetric(Metric):\n \"\"\"\n Metric used for automatic summarization. https://www.aclweb.org/anthology/W04-1013/\n\n Example:\n\n >>> target = \"Is your name John\".split()\n >>> preds = \"My name is John\".split()\n >>> rouge = RougeMetric()\n >>> from pprint import pprint\n >>> pprint(rouge(preds, target)) # doctest: +NORMALIZE_WHITESPACE\n {'rouge1_fmeasure': 0.25,\n 'rouge1_precision': 0.25,\n 'rouge1_recall': 0.25,\n 'rouge2_fmeasure': 0.0,\n 'rouge2_precision': 0.0,\n 'rouge2_recall': 0.0,\n 'rougeL_fmeasure': 0.25,\n 'rougeL_precision': 0.25,\n 'rougeL_recall': 0.25,\n 'rougeLsum_fmeasure': 0.25,\n 'rougeLsum_precision': 0.25,\n 'rougeLsum_recall': 0.25}\n \"\"\"\n\n def __init__(\n self,\n rouge_newline_sep: bool = False,\n use_stemmer: bool = False,\n rouge_keys: Tuple[str] = (\"rouge1\", \"rouge2\", \"rougeL\", \"rougeLsum\"),\n ):\n super().__init__()\n self.rouge_newline_sep = rouge_newline_sep\n self.rouge_keys = rouge_keys\n self.use_stemmer = use_stemmer\n self.aggregator = RougeBatchAggregator()\n self.scorer = rouge_scorer.RougeScorer(rouge_keys, use_stemmer=self.use_stemmer)\n\n for key in rouge_keys:\n self.add_state(key, [])\n\n def update(self, pred_lns: List[str], tgt_lns: List[str]):\n for pred, tgt in zip(pred_lns, tgt_lns):\n # rougeLsum expects \"\\n\" separated sentences within a summary\n if self.rouge_newline_sep:\n pred = add_newline_to_end_of_each_sentence(pred)\n tgt = add_newline_to_end_of_each_sentence(tgt)\n results = self.scorer.score(pred, tgt)\n for key, score in results.items():\n score = tensor([score.precision, score.recall, score.fmeasure])\n getattr(self, key).append(score)\n\n def compute(self) -> Dict[str, float]:\n scores = {key: getattr(self, key) for key in self.rouge_keys}\n self.aggregator.add_scores(scores)\n result = self.aggregator.aggregate()\n return format_rouge_results(result)\n\n def __hash__(self):\n # override to hash list objects.\n # this is a bug in the upstream pytorch release.\n hash_vals = [self.__class__.__name__]\n\n for key in self._defaults.keys():\n value = getattr(self, key)\n if isinstance(value, list):\n value = tuple(value)\n hash_vals.append(value)\n\n return hash(tuple(hash_vals))\n\n\nclass RougeBatchAggregator(scoring.BootstrapAggregator):\n \"\"\"\n Aggregates rouge scores and provides confidence intervals.\n \"\"\"\n\n def aggregate(self):\n \"\"\"\n Override function to wrap the final results in `Score` objects.\n This is due to the scores being replaced with a list of torch tensors.\n \"\"\"\n result = {}\n for score_type, scores in self._scores.items():\n # Stack scores into a 2-d matrix of (sample, measure).\n score_matrix = np.vstack(tuple(scores))\n # Percentiles are returned as (interval, measure).\n percentiles = self._bootstrap_resample(score_matrix)\n # Extract the three intervals (low, mid, high).\n intervals = tuple((Score(*percentiles[j, :]) for j in range(3)))\n result[score_type] = AggregateScore(low=intervals[0], mid=intervals[1], high=intervals[2])\n return result\n\n def add_scores(self, scores):\n self._scores = scores\n\n\ndef format_rouge_results(result: Dict[str, AggregateScore], decimal_places: int = 4) -> Dict[str, float]:\n flattened_result = {}\n for rouge_key, rouge_aggregate_score in result.items():\n for stat in [\"precision\", \"recall\", \"fmeasure\"]:\n mid = rouge_aggregate_score.mid\n score = round(getattr(mid, stat), decimal_places)\n flattened_result[f\"{rouge_key}_{stat}\"] = score\n return flattened_result\n", "path": "flash/text/seq2seq/summarization/metric.py"}], "after_files": [{"content": "# Copyright The PyTorch Lightning team.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\nfrom typing import Dict, List, Tuple\n\nimport numpy as np\nfrom rouge_score import rouge_scorer, scoring\nfrom rouge_score.scoring import AggregateScore, Score\nfrom torch import tensor\nfrom torchmetrics import Metric\n\nfrom flash.text.seq2seq import summarization\n\n\nclass RougeMetric(Metric):\n \"\"\"\n Metric used for automatic summarization. https://www.aclweb.org/anthology/W04-1013/\n\n Example:\n\n >>> target = \"Is your name John\".split()\n >>> preds = \"My name is John\".split()\n >>> rouge = RougeMetric()\n >>> from pprint import pprint\n >>> pprint(rouge(preds, target)) # doctest: +NORMALIZE_WHITESPACE\n {'rouge1_fmeasure': 0.25,\n 'rouge1_precision': 0.25,\n 'rouge1_recall': 0.25,\n 'rouge2_fmeasure': 0.0,\n 'rouge2_precision': 0.0,\n 'rouge2_recall': 0.0,\n 'rougeL_fmeasure': 0.25,\n 'rougeL_precision': 0.25,\n 'rougeL_recall': 0.25,\n 'rougeLsum_fmeasure': 0.25,\n 'rougeLsum_precision': 0.25,\n 'rougeLsum_recall': 0.25}\n \"\"\"\n\n def __init__(\n self,\n rouge_newline_sep: bool = False,\n use_stemmer: bool = False,\n rouge_keys: Tuple[str] = (\"rouge1\", \"rouge2\", \"rougeL\", \"rougeLsum\"),\n ):\n super().__init__()\n self.rouge_newline_sep = rouge_newline_sep\n self.rouge_keys = rouge_keys\n self.use_stemmer = use_stemmer\n self.aggregator = RougeBatchAggregator()\n self.scorer = rouge_scorer.RougeScorer(rouge_keys, use_stemmer=self.use_stemmer)\n\n for key in rouge_keys:\n self.add_state(key, [])\n\n def update(self, pred_lns: List[str], tgt_lns: List[str]):\n for pred, tgt in zip(pred_lns, tgt_lns):\n # rougeLsum expects \"\\n\" separated sentences within a summary\n if self.rouge_newline_sep:\n pred = summarization.utils.add_newline_to_end_of_each_sentence(pred)\n tgt = summarization.utils.add_newline_to_end_of_each_sentence(tgt)\n results = self.scorer.score(pred, tgt)\n for key, score in results.items():\n score = tensor([score.precision, score.recall, score.fmeasure])\n getattr(self, key).append(score)\n\n def compute(self) -> Dict[str, float]:\n scores = {key: getattr(self, key) for key in self.rouge_keys}\n self.aggregator.add_scores(scores)\n result = self.aggregator.aggregate()\n return format_rouge_results(result)\n\n def __hash__(self):\n # override to hash list objects.\n # this is a bug in the upstream pytorch release.\n hash_vals = [self.__class__.__name__]\n\n for key in self._defaults.keys():\n value = getattr(self, key)\n if isinstance(value, list):\n value = tuple(value)\n hash_vals.append(value)\n\n return hash(tuple(hash_vals))\n\n\nclass RougeBatchAggregator(scoring.BootstrapAggregator):\n \"\"\"\n Aggregates rouge scores and provides confidence intervals.\n \"\"\"\n\n def aggregate(self):\n \"\"\"\n Override function to wrap the final results in `Score` objects.\n This is due to the scores being replaced with a list of torch tensors.\n \"\"\"\n result = {}\n for score_type, scores in self._scores.items():\n # Stack scores into a 2-d matrix of (sample, measure).\n score_matrix = np.vstack(tuple(scores))\n # Percentiles are returned as (interval, measure).\n percentiles = self._bootstrap_resample(score_matrix)\n # Extract the three intervals (low, mid, high).\n intervals = tuple((Score(*percentiles[j, :]) for j in range(3)))\n result[score_type] = AggregateScore(low=intervals[0], mid=intervals[1], high=intervals[2])\n return result\n\n def add_scores(self, scores):\n self._scores = scores\n\n\ndef format_rouge_results(result: Dict[str, AggregateScore], decimal_places: int = 4) -> Dict[str, float]:\n flattened_result = {}\n for rouge_key, rouge_aggregate_score in result.items():\n for stat in [\"precision\", \"recall\", \"fmeasure\"]:\n mid = rouge_aggregate_score.mid\n score = round(getattr(mid, stat), decimal_places)\n flattened_result[f\"{rouge_key}_{stat}\"] = score\n return flattened_result\n", "path": "flash/text/seq2seq/summarization/metric.py"}]} | 1,885 | 282 |
gh_patches_debug_34681 | rasdani/github-patches | git_diff | sql-machine-learning__elasticdl-323 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
submit master pod using k8s python client instead of kubectl
use k8s python client to submit master pod instead of using the command below
`os.system('kubectl run ...')`
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `elasticdl/client/client.py`
Content:
```
1 import os
2 import inspect
3 import shutil
4 import time
5 import getpass
6 from string import Template
7 import docker
8
9
10 def run(model_class, train_data_dir=None,
11 num_epoch=1, minibatch_size=10,
12 record_per_task=100, num_worker=1, grads_to_wait=2):
13 m_path, m_file = _getModelFile()
14 m_file_in_docker = "/model/" + m_file
15 timestamp = int(round(time.time() * 1000))
16 _build_docker_image(m_path, m_file, m_file_in_docker, timestamp)
17 yaml_file = _generate_yaml(m_file_in_docker, model_class.__name__, train_data_dir=train_data_dir,
18 num_epoch=num_epoch, minibatch_size=minibatch_size,
19 record_per_task=record_per_task, num_worker=num_worker,
20 grads_to_wait=grads_to_wait, timestamp=timestamp)
21 _submit(yaml_file)
22
23 def _getModelFile():
24 m_file = inspect.currentframe().f_back.f_back.f_code.co_filename
25 m_path = os.path.abspath(os.path.dirname(m_file))
26 return m_path, m_file
27
28 def _build_docker_image(m_path, m_file, m_file_in_docker, timestamp):
29 d_path = os.path.abspath(os.path.dirname(
30 inspect.currentframe().f_back.f_code.co_filename))
31 new_dfile = m_path + "/Dockerfile"
32 shutil.copyfile(d_path + "/../Dockerfile.dev", new_dfile)
33
34 with open(new_dfile, 'a') as df:
35 df.write("COPY " + m_file + " " + m_file_in_docker)
36 client = docker.APIClient(base_url='unix://var/run/docker.sock')
37 for line in client.build(dockerfile='Dockerfile', path='.', tag='elasticdl:dev_' + str(timestamp)):
38 print(str(line, encoding = "utf-8"))
39
40 # TODO: upload docker image to docker hub.
41
42 def _generate_yaml(m_file, m_class,
43 train_data_dir=None, num_epoch=1,
44 minibatch_size=10, record_per_task=100,
45 num_worker=1, grads_to_wait=2, timestamp=1):
46 YAML_TEMPLATE = """
47 apiVersion: v1
48 kind: Pod
49 metadata:
50 name: elasticdl-master-$timestamp
51 labels:
52 purpose: test-command
53 spec:
54 containers:
55 - name: elasticdl-master-$timestamp
56 image: elasticdl:dev_$timestamp
57 command: ["python"]
58 args: ["-m", "elasticdl.master.main",
59 "--model-file", "$m_file",
60 "--num_worker", "$num_worker",
61 "--worker_image", "elasticdl:dev_$timestamp",
62 "--job_name", "elasticdl-$timestamp",
63 "--model-class", "$m_class",
64 "--train_data_dir", "$train_data_dir",
65 "--num_epoch", "$num_epoch",
66 "--grads_to_wait", "$grads_to_wait",
67 "--minibatch_size", "$minibatch_size",
68 "--record_per_task", "$record_per_task"]
69 imagePullPolicy: Never
70 env:
71 - name: MY_POD_IP
72 valueFrom:
73 fieldRef:
74 fieldPath: status.podIP
75 restartPolicy: Never
76 """
77 t = Template(YAML_TEMPLATE)
78 yaml_file = 'job_desc.yaml'
79 with open(yaml_file, "w") as yaml:
80 yaml.write(t.substitute(m_file=m_file, m_class=m_class,
81 train_data_dir=train_data_dir,
82 timestamp=timestamp, num_worker=num_worker, num_epoch=num_epoch,
83 minibatch_size=minibatch_size, record_per_task=record_per_task,
84 user=getpass.getuser(), grads_to_wait=grads_to_wait))
85 return yaml_file
86
87 def _submit(yaml_file):
88 os.system('kubectl create -f ' + yaml_file)
89
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/elasticdl/client/client.py b/elasticdl/client/client.py
--- a/elasticdl/client/client.py
+++ b/elasticdl/client/client.py
@@ -5,6 +5,9 @@
import getpass
from string import Template
import docker
+import yaml
+from kubernetes.client.apis import core_v1_api
+from kubernetes import config
def run(model_class, train_data_dir=None,
@@ -14,11 +17,11 @@
m_file_in_docker = "/model/" + m_file
timestamp = int(round(time.time() * 1000))
_build_docker_image(m_path, m_file, m_file_in_docker, timestamp)
- yaml_file = _generate_yaml(m_file_in_docker, model_class.__name__, train_data_dir=train_data_dir,
+ yaml_content = _generate_yaml(m_file_in_docker, model_class.__name__, train_data_dir=train_data_dir,
num_epoch=num_epoch, minibatch_size=minibatch_size,
record_per_task=record_per_task, num_worker=num_worker,
grads_to_wait=grads_to_wait, timestamp=timestamp)
- _submit(yaml_file)
+ _submit(yaml_content)
def _getModelFile():
m_file = inspect.currentframe().f_back.f_back.f_code.co_filename
@@ -75,14 +78,15 @@
restartPolicy: Never
"""
t = Template(YAML_TEMPLATE)
- yaml_file = 'job_desc.yaml'
- with open(yaml_file, "w") as yaml:
- yaml.write(t.substitute(m_file=m_file, m_class=m_class,
- train_data_dir=train_data_dir,
- timestamp=timestamp, num_worker=num_worker, num_epoch=num_epoch,
- minibatch_size=minibatch_size, record_per_task=record_per_task,
- user=getpass.getuser(), grads_to_wait=grads_to_wait))
- return yaml_file
+ return t.substitute(m_file=m_file, m_class=m_class,
+ train_data_dir=train_data_dir,
+ timestamp=timestamp, num_worker=num_worker, num_epoch=num_epoch,
+ minibatch_size=minibatch_size, record_per_task=record_per_task,
+ user=getpass.getuser(), grads_to_wait=grads_to_wait)
-def _submit(yaml_file):
- os.system('kubectl create -f ' + yaml_file)
+def _submit(yaml_content):
+ config.load_kube_config()
+ pod_desc = yaml.safe_load(yaml_content)
+ api = core_v1_api.CoreV1Api()
+ resp = api.create_namespaced_pod(body=pod_desc, namespace='default')
+ print("Pod created. status='%s'" % str(resp.status))
| {"golden_diff": "diff --git a/elasticdl/client/client.py b/elasticdl/client/client.py\n--- a/elasticdl/client/client.py\n+++ b/elasticdl/client/client.py\n@@ -5,6 +5,9 @@\n import getpass\n from string import Template\n import docker\n+import yaml\n+from kubernetes.client.apis import core_v1_api\n+from kubernetes import config\n \n \n def run(model_class, train_data_dir=None, \n@@ -14,11 +17,11 @@\n m_file_in_docker = \"/model/\" + m_file \n timestamp = int(round(time.time() * 1000))\n _build_docker_image(m_path, m_file, m_file_in_docker, timestamp)\n- yaml_file = _generate_yaml(m_file_in_docker, model_class.__name__, train_data_dir=train_data_dir, \n+ yaml_content = _generate_yaml(m_file_in_docker, model_class.__name__, train_data_dir=train_data_dir, \n num_epoch=num_epoch, minibatch_size=minibatch_size, \n record_per_task=record_per_task, num_worker=num_worker, \n grads_to_wait=grads_to_wait, timestamp=timestamp)\n- _submit(yaml_file)\n+ _submit(yaml_content)\n \n def _getModelFile():\n m_file = inspect.currentframe().f_back.f_back.f_code.co_filename\n@@ -75,14 +78,15 @@\n restartPolicy: Never\n \"\"\"\n t = Template(YAML_TEMPLATE)\n- yaml_file = 'job_desc.yaml'\n- with open(yaml_file, \"w\") as yaml:\n- yaml.write(t.substitute(m_file=m_file, m_class=m_class, \n- train_data_dir=train_data_dir, \n- timestamp=timestamp, num_worker=num_worker, num_epoch=num_epoch,\n- minibatch_size=minibatch_size, record_per_task=record_per_task,\n- user=getpass.getuser(), grads_to_wait=grads_to_wait))\n- return yaml_file\n+ return t.substitute(m_file=m_file, m_class=m_class, \n+ train_data_dir=train_data_dir, \n+ timestamp=timestamp, num_worker=num_worker, num_epoch=num_epoch,\n+ minibatch_size=minibatch_size, record_per_task=record_per_task,\n+ user=getpass.getuser(), grads_to_wait=grads_to_wait)\n \n-def _submit(yaml_file):\n- os.system('kubectl create -f ' + yaml_file)\n+def _submit(yaml_content):\n+ config.load_kube_config()\n+ pod_desc = yaml.safe_load(yaml_content)\n+ api = core_v1_api.CoreV1Api()\n+ resp = api.create_namespaced_pod(body=pod_desc, namespace='default')\n+ print(\"Pod created. status='%s'\" % str(resp.status))\n", "issue": "submit master pod using k8s python client instead of kubectl \nuse k8s python client to submit master pod instead of using the command below\r\n`os.system('kubectl run ...')`\n", "before_files": [{"content": "import os\nimport inspect\nimport shutil\nimport time\nimport getpass\nfrom string import Template\nimport docker\n\n\ndef run(model_class, train_data_dir=None, \n num_epoch=1, minibatch_size=10, \n record_per_task=100, num_worker=1, grads_to_wait=2):\n m_path, m_file = _getModelFile()\n m_file_in_docker = \"/model/\" + m_file \n timestamp = int(round(time.time() * 1000))\n _build_docker_image(m_path, m_file, m_file_in_docker, timestamp)\n yaml_file = _generate_yaml(m_file_in_docker, model_class.__name__, train_data_dir=train_data_dir, \n num_epoch=num_epoch, minibatch_size=minibatch_size, \n record_per_task=record_per_task, num_worker=num_worker, \n grads_to_wait=grads_to_wait, timestamp=timestamp)\n _submit(yaml_file)\n\ndef _getModelFile():\n m_file = inspect.currentframe().f_back.f_back.f_code.co_filename\n m_path = os.path.abspath(os.path.dirname(m_file))\n return m_path, m_file\n\ndef _build_docker_image(m_path, m_file, m_file_in_docker, timestamp):\n d_path = os.path.abspath(os.path.dirname(\n inspect.currentframe().f_back.f_code.co_filename))\n new_dfile = m_path + \"/Dockerfile\"\n shutil.copyfile(d_path + \"/../Dockerfile.dev\", new_dfile)\n\n with open(new_dfile, 'a') as df:\n df.write(\"COPY \" + m_file + \" \" + m_file_in_docker)\n client = docker.APIClient(base_url='unix://var/run/docker.sock') \n for line in client.build(dockerfile='Dockerfile', path='.', tag='elasticdl:dev_' + str(timestamp)):\n print(str(line, encoding = \"utf-8\"))\n\n # TODO: upload docker image to docker hub.\n\ndef _generate_yaml(m_file, m_class,\n train_data_dir=None, num_epoch=1,\n minibatch_size=10, record_per_task=100, \n num_worker=1, grads_to_wait=2, timestamp=1):\n YAML_TEMPLATE = \"\"\"\n apiVersion: v1\n kind: Pod\n metadata:\n name: elasticdl-master-$timestamp\n labels:\n purpose: test-command\n spec:\n containers:\n - name: elasticdl-master-$timestamp\n image: elasticdl:dev_$timestamp\n command: [\"python\"]\n args: [\"-m\", \"elasticdl.master.main\",\n \"--model-file\", \"$m_file\",\n \"--num_worker\", \"$num_worker\",\n \"--worker_image\", \"elasticdl:dev_$timestamp\",\n \"--job_name\", \"elasticdl-$timestamp\",\n \"--model-class\", \"$m_class\",\n \"--train_data_dir\", \"$train_data_dir\",\n \"--num_epoch\", \"$num_epoch\",\n \"--grads_to_wait\", \"$grads_to_wait\",\n \"--minibatch_size\", \"$minibatch_size\",\n \"--record_per_task\", \"$record_per_task\"]\n imagePullPolicy: Never\n env:\n - name: MY_POD_IP\n valueFrom:\n fieldRef:\n fieldPath: status.podIP\n restartPolicy: Never\n \"\"\"\n t = Template(YAML_TEMPLATE)\n yaml_file = 'job_desc.yaml'\n with open(yaml_file, \"w\") as yaml:\n yaml.write(t.substitute(m_file=m_file, m_class=m_class, \n train_data_dir=train_data_dir, \n timestamp=timestamp, num_worker=num_worker, num_epoch=num_epoch,\n minibatch_size=minibatch_size, record_per_task=record_per_task,\n user=getpass.getuser(), grads_to_wait=grads_to_wait))\n return yaml_file\n\ndef _submit(yaml_file):\n os.system('kubectl create -f ' + yaml_file)\n", "path": "elasticdl/client/client.py"}], "after_files": [{"content": "import os\nimport inspect\nimport shutil\nimport time\nimport getpass\nfrom string import Template\nimport docker\nimport yaml\nfrom kubernetes.client.apis import core_v1_api\nfrom kubernetes import config\n\n\ndef run(model_class, train_data_dir=None, \n num_epoch=1, minibatch_size=10, \n record_per_task=100, num_worker=1, grads_to_wait=2):\n m_path, m_file = _getModelFile()\n m_file_in_docker = \"/model/\" + m_file \n timestamp = int(round(time.time() * 1000))\n _build_docker_image(m_path, m_file, m_file_in_docker, timestamp)\n yaml_content = _generate_yaml(m_file_in_docker, model_class.__name__, train_data_dir=train_data_dir, \n num_epoch=num_epoch, minibatch_size=minibatch_size, \n record_per_task=record_per_task, num_worker=num_worker, \n grads_to_wait=grads_to_wait, timestamp=timestamp)\n _submit(yaml_content)\n\ndef _getModelFile():\n m_file = inspect.currentframe().f_back.f_back.f_code.co_filename\n m_path = os.path.abspath(os.path.dirname(m_file))\n return m_path, m_file\n\ndef _build_docker_image(m_path, m_file, m_file_in_docker, timestamp):\n d_path = os.path.abspath(os.path.dirname(\n inspect.currentframe().f_back.f_code.co_filename))\n new_dfile = m_path + \"/Dockerfile\"\n shutil.copyfile(d_path + \"/../Dockerfile.dev\", new_dfile)\n\n with open(new_dfile, 'a') as df:\n df.write(\"COPY \" + m_file + \" \" + m_file_in_docker)\n client = docker.APIClient(base_url='unix://var/run/docker.sock') \n for line in client.build(dockerfile='Dockerfile', path='.', tag='elasticdl:dev_' + str(timestamp)):\n print(str(line, encoding = \"utf-8\"))\n\n # TODO: upload docker image to docker hub.\n\ndef _generate_yaml(m_file, m_class,\n train_data_dir=None, num_epoch=1,\n minibatch_size=10, record_per_task=100, \n num_worker=1, grads_to_wait=2, timestamp=1):\n YAML_TEMPLATE = \"\"\"\n apiVersion: v1\n kind: Pod\n metadata:\n name: elasticdl-master-$timestamp\n labels:\n purpose: test-command\n spec:\n containers:\n - name: elasticdl-master-$timestamp\n image: elasticdl:dev_$timestamp\n command: [\"python\"]\n args: [\"-m\", \"elasticdl.master.main\",\n \"--model-file\", \"$m_file\",\n \"--num_worker\", \"$num_worker\",\n \"--worker_image\", \"elasticdl:dev_$timestamp\",\n \"--job_name\", \"elasticdl-$timestamp\",\n \"--model-class\", \"$m_class\",\n \"--train_data_dir\", \"$train_data_dir\",\n \"--num_epoch\", \"$num_epoch\",\n \"--grads_to_wait\", \"$grads_to_wait\",\n \"--minibatch_size\", \"$minibatch_size\",\n \"--record_per_task\", \"$record_per_task\"]\n imagePullPolicy: Never\n env:\n - name: MY_POD_IP\n valueFrom:\n fieldRef:\n fieldPath: status.podIP\n restartPolicy: Never\n \"\"\"\n t = Template(YAML_TEMPLATE)\n return t.substitute(m_file=m_file, m_class=m_class, \n train_data_dir=train_data_dir, \n timestamp=timestamp, num_worker=num_worker, num_epoch=num_epoch,\n minibatch_size=minibatch_size, record_per_task=record_per_task,\n user=getpass.getuser(), grads_to_wait=grads_to_wait)\n\ndef _submit(yaml_content):\n config.load_kube_config()\n pod_desc = yaml.safe_load(yaml_content)\n api = core_v1_api.CoreV1Api()\n resp = api.create_namespaced_pod(body=pod_desc, namespace='default')\n print(\"Pod created. status='%s'\" % str(resp.status))\n", "path": "elasticdl/client/client.py"}]} | 1,317 | 608 |
gh_patches_debug_27572 | rasdani/github-patches | git_diff | cookiecutter__cookiecutter-1358 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Add log message to get_user_config function
there should be debug message for cases when default config overwrites user_config.
Currently, it is done silently.
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `cookiecutter/config.py`
Content:
```
1 # -*- coding: utf-8 -*-
2
3 """Global configuration handling."""
4
5 from __future__ import unicode_literals
6 import copy
7 import logging
8 import os
9 import io
10 import collections
11
12 import poyo
13
14 from cookiecutter.exceptions import ConfigDoesNotExistException
15 from cookiecutter.exceptions import InvalidConfiguration
16
17
18 logger = logging.getLogger(__name__)
19
20 USER_CONFIG_PATH = os.path.expanduser('~/.cookiecutterrc')
21
22 BUILTIN_ABBREVIATIONS = {
23 'gh': 'https://github.com/{0}.git',
24 'gl': 'https://gitlab.com/{0}.git',
25 'bb': 'https://bitbucket.org/{0}',
26 }
27
28 DEFAULT_CONFIG = {
29 'cookiecutters_dir': os.path.expanduser('~/.cookiecutters/'),
30 'replay_dir': os.path.expanduser('~/.cookiecutter_replay/'),
31 'default_context': collections.OrderedDict([]),
32 'abbreviations': BUILTIN_ABBREVIATIONS,
33 }
34
35
36 def _expand_path(path):
37 """Expand both environment variables and user home in the given path."""
38 path = os.path.expandvars(path)
39 path = os.path.expanduser(path)
40 return path
41
42
43 def merge_configs(default, overwrite):
44 """Recursively update a dict with the key/value pair of another.
45
46 Dict values that are dictionaries themselves will be updated, whilst
47 preserving existing keys.
48 """
49 new_config = copy.deepcopy(default)
50
51 for k, v in overwrite.items():
52 # Make sure to preserve existing items in
53 # nested dicts, for example `abbreviations`
54 if isinstance(v, dict):
55 new_config[k] = merge_configs(default[k], v)
56 else:
57 new_config[k] = v
58
59 return new_config
60
61
62 def get_config(config_path):
63 """Retrieve the config from the specified path, returning a config dict."""
64 if not os.path.exists(config_path):
65 raise ConfigDoesNotExistException
66
67 logger.debug('config_path is %s', config_path)
68 with io.open(config_path, encoding='utf-8') as file_handle:
69 try:
70 yaml_dict = poyo.parse_string(file_handle.read())
71 except poyo.exceptions.PoyoException as e:
72 raise InvalidConfiguration(
73 'Unable to parse YAML file {}. Error: {}'
74 ''.format(config_path, e)
75 )
76
77 config_dict = merge_configs(DEFAULT_CONFIG, yaml_dict)
78
79 raw_replay_dir = config_dict['replay_dir']
80 config_dict['replay_dir'] = _expand_path(raw_replay_dir)
81
82 raw_cookies_dir = config_dict['cookiecutters_dir']
83 config_dict['cookiecutters_dir'] = _expand_path(raw_cookies_dir)
84
85 return config_dict
86
87
88 def get_user_config(config_file=None, default_config=False):
89 """Return the user config as a dict.
90
91 If ``default_config`` is True, ignore ``config_file`` and return default
92 values for the config parameters.
93
94 If a path to a ``config_file`` is given, that is different from the default
95 location, load the user config from that.
96
97 Otherwise look up the config file path in the ``COOKIECUTTER_CONFIG``
98 environment variable. If set, load the config from this path. This will
99 raise an error if the specified path is not valid.
100
101 If the environment variable is not set, try the default config file path
102 before falling back to the default config values.
103 """
104 # Do NOT load a config. Return defaults instead.
105 if default_config:
106 return copy.copy(DEFAULT_CONFIG)
107
108 # Load the given config file
109 if config_file and config_file is not USER_CONFIG_PATH:
110 return get_config(config_file)
111
112 try:
113 # Does the user set up a config environment variable?
114 env_config_file = os.environ['COOKIECUTTER_CONFIG']
115 except KeyError:
116 # Load an optional user config if it exists
117 # otherwise return the defaults
118 if os.path.exists(USER_CONFIG_PATH):
119 return get_config(USER_CONFIG_PATH)
120 else:
121 return copy.copy(DEFAULT_CONFIG)
122 else:
123 # There is a config environment variable. Try to load it.
124 # Do not check for existence, so invalid file paths raise an error.
125 return get_config(env_config_file)
126
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/cookiecutter/config.py b/cookiecutter/config.py
--- a/cookiecutter/config.py
+++ b/cookiecutter/config.py
@@ -103,10 +103,12 @@
"""
# Do NOT load a config. Return defaults instead.
if default_config:
+ logger.debug("Force ignoring user config with default_config switch.")
return copy.copy(DEFAULT_CONFIG)
# Load the given config file
if config_file and config_file is not USER_CONFIG_PATH:
+ logger.debug("Loading custom config from %s.", config_file)
return get_config(config_file)
try:
@@ -116,10 +118,13 @@
# Load an optional user config if it exists
# otherwise return the defaults
if os.path.exists(USER_CONFIG_PATH):
+ logger.debug("Loading config from %s.", USER_CONFIG_PATH)
return get_config(USER_CONFIG_PATH)
else:
+ logger.debug("User config not found. Loading default config.")
return copy.copy(DEFAULT_CONFIG)
else:
# There is a config environment variable. Try to load it.
# Do not check for existence, so invalid file paths raise an error.
+ logger.debug("User config not found or not specified. Loading default config.")
return get_config(env_config_file)
| {"golden_diff": "diff --git a/cookiecutter/config.py b/cookiecutter/config.py\n--- a/cookiecutter/config.py\n+++ b/cookiecutter/config.py\n@@ -103,10 +103,12 @@\n \"\"\"\n # Do NOT load a config. Return defaults instead.\n if default_config:\n+ logger.debug(\"Force ignoring user config with default_config switch.\")\n return copy.copy(DEFAULT_CONFIG)\n \n # Load the given config file\n if config_file and config_file is not USER_CONFIG_PATH:\n+ logger.debug(\"Loading custom config from %s.\", config_file)\n return get_config(config_file)\n \n try:\n@@ -116,10 +118,13 @@\n # Load an optional user config if it exists\n # otherwise return the defaults\n if os.path.exists(USER_CONFIG_PATH):\n+ logger.debug(\"Loading config from %s.\", USER_CONFIG_PATH)\n return get_config(USER_CONFIG_PATH)\n else:\n+ logger.debug(\"User config not found. Loading default config.\")\n return copy.copy(DEFAULT_CONFIG)\n else:\n # There is a config environment variable. Try to load it.\n # Do not check for existence, so invalid file paths raise an error.\n+ logger.debug(\"User config not found or not specified. Loading default config.\")\n return get_config(env_config_file)\n", "issue": "Add log message to get_user_config function\nthere should be debug message for cases when default config overwrites user_config. \r\nCurrently, it is done silently.\n", "before_files": [{"content": "# -*- coding: utf-8 -*-\n\n\"\"\"Global configuration handling.\"\"\"\n\nfrom __future__ import unicode_literals\nimport copy\nimport logging\nimport os\nimport io\nimport collections\n\nimport poyo\n\nfrom cookiecutter.exceptions import ConfigDoesNotExistException\nfrom cookiecutter.exceptions import InvalidConfiguration\n\n\nlogger = logging.getLogger(__name__)\n\nUSER_CONFIG_PATH = os.path.expanduser('~/.cookiecutterrc')\n\nBUILTIN_ABBREVIATIONS = {\n 'gh': 'https://github.com/{0}.git',\n 'gl': 'https://gitlab.com/{0}.git',\n 'bb': 'https://bitbucket.org/{0}',\n}\n\nDEFAULT_CONFIG = {\n 'cookiecutters_dir': os.path.expanduser('~/.cookiecutters/'),\n 'replay_dir': os.path.expanduser('~/.cookiecutter_replay/'),\n 'default_context': collections.OrderedDict([]),\n 'abbreviations': BUILTIN_ABBREVIATIONS,\n}\n\n\ndef _expand_path(path):\n \"\"\"Expand both environment variables and user home in the given path.\"\"\"\n path = os.path.expandvars(path)\n path = os.path.expanduser(path)\n return path\n\n\ndef merge_configs(default, overwrite):\n \"\"\"Recursively update a dict with the key/value pair of another.\n\n Dict values that are dictionaries themselves will be updated, whilst\n preserving existing keys.\n \"\"\"\n new_config = copy.deepcopy(default)\n\n for k, v in overwrite.items():\n # Make sure to preserve existing items in\n # nested dicts, for example `abbreviations`\n if isinstance(v, dict):\n new_config[k] = merge_configs(default[k], v)\n else:\n new_config[k] = v\n\n return new_config\n\n\ndef get_config(config_path):\n \"\"\"Retrieve the config from the specified path, returning a config dict.\"\"\"\n if not os.path.exists(config_path):\n raise ConfigDoesNotExistException\n\n logger.debug('config_path is %s', config_path)\n with io.open(config_path, encoding='utf-8') as file_handle:\n try:\n yaml_dict = poyo.parse_string(file_handle.read())\n except poyo.exceptions.PoyoException as e:\n raise InvalidConfiguration(\n 'Unable to parse YAML file {}. Error: {}'\n ''.format(config_path, e)\n )\n\n config_dict = merge_configs(DEFAULT_CONFIG, yaml_dict)\n\n raw_replay_dir = config_dict['replay_dir']\n config_dict['replay_dir'] = _expand_path(raw_replay_dir)\n\n raw_cookies_dir = config_dict['cookiecutters_dir']\n config_dict['cookiecutters_dir'] = _expand_path(raw_cookies_dir)\n\n return config_dict\n\n\ndef get_user_config(config_file=None, default_config=False):\n \"\"\"Return the user config as a dict.\n\n If ``default_config`` is True, ignore ``config_file`` and return default\n values for the config parameters.\n\n If a path to a ``config_file`` is given, that is different from the default\n location, load the user config from that.\n\n Otherwise look up the config file path in the ``COOKIECUTTER_CONFIG``\n environment variable. If set, load the config from this path. This will\n raise an error if the specified path is not valid.\n\n If the environment variable is not set, try the default config file path\n before falling back to the default config values.\n \"\"\"\n # Do NOT load a config. Return defaults instead.\n if default_config:\n return copy.copy(DEFAULT_CONFIG)\n\n # Load the given config file\n if config_file and config_file is not USER_CONFIG_PATH:\n return get_config(config_file)\n\n try:\n # Does the user set up a config environment variable?\n env_config_file = os.environ['COOKIECUTTER_CONFIG']\n except KeyError:\n # Load an optional user config if it exists\n # otherwise return the defaults\n if os.path.exists(USER_CONFIG_PATH):\n return get_config(USER_CONFIG_PATH)\n else:\n return copy.copy(DEFAULT_CONFIG)\n else:\n # There is a config environment variable. Try to load it.\n # Do not check for existence, so invalid file paths raise an error.\n return get_config(env_config_file)\n", "path": "cookiecutter/config.py"}], "after_files": [{"content": "# -*- coding: utf-8 -*-\n\n\"\"\"Global configuration handling.\"\"\"\n\nfrom __future__ import unicode_literals\nimport copy\nimport logging\nimport os\nimport io\nimport collections\n\nimport poyo\n\nfrom cookiecutter.exceptions import ConfigDoesNotExistException\nfrom cookiecutter.exceptions import InvalidConfiguration\n\n\nlogger = logging.getLogger(__name__)\n\nUSER_CONFIG_PATH = os.path.expanduser('~/.cookiecutterrc')\n\nBUILTIN_ABBREVIATIONS = {\n 'gh': 'https://github.com/{0}.git',\n 'gl': 'https://gitlab.com/{0}.git',\n 'bb': 'https://bitbucket.org/{0}',\n}\n\nDEFAULT_CONFIG = {\n 'cookiecutters_dir': os.path.expanduser('~/.cookiecutters/'),\n 'replay_dir': os.path.expanduser('~/.cookiecutter_replay/'),\n 'default_context': collections.OrderedDict([]),\n 'abbreviations': BUILTIN_ABBREVIATIONS,\n}\n\n\ndef _expand_path(path):\n \"\"\"Expand both environment variables and user home in the given path.\"\"\"\n path = os.path.expandvars(path)\n path = os.path.expanduser(path)\n return path\n\n\ndef merge_configs(default, overwrite):\n \"\"\"Recursively update a dict with the key/value pair of another.\n\n Dict values that are dictionaries themselves will be updated, whilst\n preserving existing keys.\n \"\"\"\n new_config = copy.deepcopy(default)\n\n for k, v in overwrite.items():\n # Make sure to preserve existing items in\n # nested dicts, for example `abbreviations`\n if isinstance(v, dict):\n new_config[k] = merge_configs(default[k], v)\n else:\n new_config[k] = v\n\n return new_config\n\n\ndef get_config(config_path):\n \"\"\"Retrieve the config from the specified path, returning a config dict.\"\"\"\n if not os.path.exists(config_path):\n raise ConfigDoesNotExistException\n\n logger.debug('config_path is %s', config_path)\n with io.open(config_path, encoding='utf-8') as file_handle:\n try:\n yaml_dict = poyo.parse_string(file_handle.read())\n except poyo.exceptions.PoyoException as e:\n raise InvalidConfiguration(\n 'Unable to parse YAML file {}. Error: {}'\n ''.format(config_path, e)\n )\n\n config_dict = merge_configs(DEFAULT_CONFIG, yaml_dict)\n\n raw_replay_dir = config_dict['replay_dir']\n config_dict['replay_dir'] = _expand_path(raw_replay_dir)\n\n raw_cookies_dir = config_dict['cookiecutters_dir']\n config_dict['cookiecutters_dir'] = _expand_path(raw_cookies_dir)\n\n return config_dict\n\n\ndef get_user_config(config_file=None, default_config=False):\n \"\"\"Return the user config as a dict.\n\n If ``default_config`` is True, ignore ``config_file`` and return default\n values for the config parameters.\n\n If a path to a ``config_file`` is given, that is different from the default\n location, load the user config from that.\n\n Otherwise look up the config file path in the ``COOKIECUTTER_CONFIG``\n environment variable. If set, load the config from this path. This will\n raise an error if the specified path is not valid.\n\n If the environment variable is not set, try the default config file path\n before falling back to the default config values.\n \"\"\"\n # Do NOT load a config. Return defaults instead.\n if default_config:\n logger.debug(\"Force ignoring user config with default_config switch.\")\n return copy.copy(DEFAULT_CONFIG)\n\n # Load the given config file\n if config_file and config_file is not USER_CONFIG_PATH:\n logger.debug(\"Loading custom config from %s.\", config_file)\n return get_config(config_file)\n\n try:\n # Does the user set up a config environment variable?\n env_config_file = os.environ['COOKIECUTTER_CONFIG']\n except KeyError:\n # Load an optional user config if it exists\n # otherwise return the defaults\n if os.path.exists(USER_CONFIG_PATH):\n logger.debug(\"Loading config from %s.\", USER_CONFIG_PATH)\n return get_config(USER_CONFIG_PATH)\n else:\n logger.debug(\"User config not found. Loading default config.\")\n return copy.copy(DEFAULT_CONFIG)\n else:\n # There is a config environment variable. Try to load it.\n # Do not check for existence, so invalid file paths raise an error.\n logger.debug(\"User config not found or not specified. Loading default config.\")\n return get_config(env_config_file)\n", "path": "cookiecutter/config.py"}]} | 1,469 | 288 |
gh_patches_debug_27158 | rasdani/github-patches | git_diff | archlinux__archinstall-702 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
[2.3.0-RC1] Automatic partitioning does not fill drive with btrfs and no encryption

[My installation log](https://github.com/archlinux/archinstall/files/7500204/install.log.txt)
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `archinstall/lib/disk/user_guides.py`
Content:
```
1 import logging
2 from .helpers import sort_block_devices_based_on_performance, select_largest_device, select_disk_larger_than_or_close_to
3 from ..output import log
4
5 def suggest_single_disk_layout(block_device, default_filesystem=None):
6 if not default_filesystem:
7 from ..user_interaction import ask_for_main_filesystem_format
8 default_filesystem = ask_for_main_filesystem_format()
9
10 MIN_SIZE_TO_ALLOW_HOME_PART = 40 # Gb
11
12 layout = {
13 block_device.path : {
14 "wipe" : True,
15 "partitions" : []
16 }
17 }
18
19 layout[block_device.path]['partitions'].append({
20 # Boot
21 "type" : "primary",
22 "start" : "1MiB",
23 "size" : "513MiB",
24 "boot" : True,
25 "encrypted" : False,
26 "format" : True,
27 "mountpoint" : "/boot",
28 "filesystem" : {
29 "format" : "fat32"
30 }
31 })
32 layout[block_device.path]['partitions'].append({
33 # Root
34 "type" : "primary",
35 "start" : "513MiB",
36 "encrypted" : False,
37 "format" : True,
38 "size" : "100%" if block_device.size < MIN_SIZE_TO_ALLOW_HOME_PART else f"{min(block_device.size, 20)*1024}MiB",
39 "mountpoint" : "/",
40 "filesystem" : {
41 "format" : default_filesystem
42 }
43 })
44
45 if default_filesystem == 'btrfs' and input('Would you like to use BTRFS subvolumes? (Y/n): ').strip().lower() in ('', 'y', 'yes'):
46 if input('Do you want to use a recommended structure? (Y/n): ').strip().lower() in ('', 'y', 'yes'):
47 # https://btrfs.wiki.kernel.org/index.php/FAQ
48 # https://unix.stackexchange.com/questions/246976/btrfs-subvolume-uuid-clash
49 # https://github.com/classy-giraffe/easy-arch/blob/main/easy-arch.sh
50 layout[block_device.path]['partitions'][1]['btrfs'] = {
51 "subvolumes" : {
52 "@home" : "/home",
53 "@log" : "/var/log",
54 "@pkgs" : "/var/cache/pacman/pkg",
55 "@.snapshots" : "/.snapshots"
56 }
57 }
58 else:
59 pass # ... implement a guided setup
60
61 elif block_device.size >= MIN_SIZE_TO_ALLOW_HOME_PART:
62 # If we don't want to use subvolumes,
63 # But we want to be able to re-use data between re-installs..
64 # A second partition for /home would be nice if we have the space for it
65 layout[block_device.path]['partitions'].append({
66 # Home
67 "type" : "primary",
68 "encrypted" : False,
69 "format" : True,
70 "start" : f"{min(block_device.size*0.2, 20)*1024}MiB",
71 "size" : "100%",
72 "mountpoint" : "/home",
73 "filesystem" : {
74 "format" : default_filesystem
75 }
76 })
77
78 return layout
79
80
81 def suggest_multi_disk_layout(block_devices, default_filesystem=None):
82 if not default_filesystem:
83 from ..user_interaction import ask_for_main_filesystem_format
84 default_filesystem = ask_for_main_filesystem_format()
85
86 # Not really a rock solid foundation of information to stand on, but it's a start:
87 # https://www.reddit.com/r/btrfs/comments/m287gp/partition_strategy_for_two_physical_disks/
88 # https://www.reddit.com/r/btrfs/comments/9us4hr/what_is_your_btrfs_partitionsubvolumes_scheme/
89
90 MIN_SIZE_TO_ALLOW_HOME_PART = 40 # Gb
91 ARCH_LINUX_INSTALLED_SIZE = 20 # Gb, rough estimate taking in to account user desktops etc. TODO: Catch user packages to detect size?
92
93 block_devices = sort_block_devices_based_on_performance(block_devices).keys()
94
95 home_device = select_largest_device(block_devices, gigabytes=MIN_SIZE_TO_ALLOW_HOME_PART)
96 root_device = select_disk_larger_than_or_close_to(block_devices, gigabytes=ARCH_LINUX_INSTALLED_SIZE, filter_out=[home_device])
97
98 log(f"Suggesting multi-disk-layout using {len(block_devices)} disks, where {root_device} will be /root and {home_device} will be /home", level=logging.DEBUG)
99
100 layout = {
101 root_device.path : {
102 "wipe" : True,
103 "partitions" : []
104 },
105 home_device.path : {
106 "wipe" : True,
107 "partitions" : []
108 },
109 }
110
111 layout[root_device.path]['partitions'].append({
112 # Boot
113 "type" : "primary",
114 "start" : "1MiB",
115 "size" : "513MiB",
116 "boot" : True,
117 "encrypted" : False,
118 "format" : True,
119 "mountpoint" : "/boot",
120 "filesystem" : {
121 "format" : "fat32"
122 }
123 })
124 layout[root_device.path]['partitions'].append({
125 # Root
126 "type" : "primary",
127 "start" : "513MiB",
128 "encrypted" : False,
129 "format" : True,
130 "size" : "100%",
131 "mountpoint" : "/",
132 "filesystem" : {
133 "format" : default_filesystem
134 }
135 })
136
137 layout[home_device.path]['partitions'].append({
138 # Home
139 "type" : "primary",
140 "encrypted" : False,
141 "format" : True,
142 "start" : "4MiB",
143 "size" : "100%",
144 "mountpoint" : "/home",
145 "filesystem" : {
146 "format" : default_filesystem
147 }
148 })
149
150 return layout
151
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/archinstall/lib/disk/user_guides.py b/archinstall/lib/disk/user_guides.py
--- a/archinstall/lib/disk/user_guides.py
+++ b/archinstall/lib/disk/user_guides.py
@@ -8,6 +8,10 @@
default_filesystem = ask_for_main_filesystem_format()
MIN_SIZE_TO_ALLOW_HOME_PART = 40 # Gb
+ using_subvolumes = False
+
+ if default_filesystem == 'btrfs':
+ using_subvolumes = input('Would you like to use BTRFS subvolumes? (Y/n): ').strip().lower() in ('', 'y', 'yes')
layout = {
block_device.path : {
@@ -35,14 +39,14 @@
"start" : "513MiB",
"encrypted" : False,
"format" : True,
- "size" : "100%" if block_device.size < MIN_SIZE_TO_ALLOW_HOME_PART else f"{min(block_device.size, 20)*1024}MiB",
+ "size" : "100%" if (using_subvolumes or block_device.size < MIN_SIZE_TO_ALLOW_HOME_PART) else f"{min(block_device.size, 20)*1024}MiB",
"mountpoint" : "/",
"filesystem" : {
"format" : default_filesystem
}
})
- if default_filesystem == 'btrfs' and input('Would you like to use BTRFS subvolumes? (Y/n): ').strip().lower() in ('', 'y', 'yes'):
+ if default_filesystem == 'btrfs' and using_subvolumes:
if input('Do you want to use a recommended structure? (Y/n): ').strip().lower() in ('', 'y', 'yes'):
# https://btrfs.wiki.kernel.org/index.php/FAQ
# https://unix.stackexchange.com/questions/246976/btrfs-subvolume-uuid-clash
| {"golden_diff": "diff --git a/archinstall/lib/disk/user_guides.py b/archinstall/lib/disk/user_guides.py\n--- a/archinstall/lib/disk/user_guides.py\n+++ b/archinstall/lib/disk/user_guides.py\n@@ -8,6 +8,10 @@\n \t\tdefault_filesystem = ask_for_main_filesystem_format()\n \t\t\n \tMIN_SIZE_TO_ALLOW_HOME_PART = 40 # Gb\n+\tusing_subvolumes = False\n+\n+\tif default_filesystem == 'btrfs':\n+\t\tusing_subvolumes = input('Would you like to use BTRFS subvolumes? (Y/n): ').strip().lower() in ('', 'y', 'yes')\n \n \tlayout = {\n \t\tblock_device.path : {\n@@ -35,14 +39,14 @@\n \t\t\"start\" : \"513MiB\",\n \t\t\"encrypted\" : False,\n \t\t\"format\" : True,\n-\t\t\"size\" : \"100%\" if block_device.size < MIN_SIZE_TO_ALLOW_HOME_PART else f\"{min(block_device.size, 20)*1024}MiB\",\n+\t\t\"size\" : \"100%\" if (using_subvolumes or block_device.size < MIN_SIZE_TO_ALLOW_HOME_PART) else f\"{min(block_device.size, 20)*1024}MiB\",\n \t\t\"mountpoint\" : \"/\",\n \t\t\"filesystem\" : {\n \t\t\t\"format\" : default_filesystem\n \t\t}\n \t})\n \n-\tif default_filesystem == 'btrfs' and input('Would you like to use BTRFS subvolumes? (Y/n): ').strip().lower() in ('', 'y', 'yes'):\n+\tif default_filesystem == 'btrfs' and using_subvolumes:\n \t\tif input('Do you want to use a recommended structure? (Y/n): ').strip().lower() in ('', 'y', 'yes'):\n \t\t\t# https://btrfs.wiki.kernel.org/index.php/FAQ\n \t\t\t# https://unix.stackexchange.com/questions/246976/btrfs-subvolume-uuid-clash\n", "issue": "[2.3.0-RC1] Automatic partitioning does not fill drive with btrfs and no encryption\n\r\n\r\n[My installation log](https://github.com/archlinux/archinstall/files/7500204/install.log.txt)\r\n\n", "before_files": [{"content": "import logging\nfrom .helpers import sort_block_devices_based_on_performance, select_largest_device, select_disk_larger_than_or_close_to\nfrom ..output import log\n\ndef suggest_single_disk_layout(block_device, default_filesystem=None):\n\tif not default_filesystem:\n\t\tfrom ..user_interaction import ask_for_main_filesystem_format\n\t\tdefault_filesystem = ask_for_main_filesystem_format()\n\t\t\n\tMIN_SIZE_TO_ALLOW_HOME_PART = 40 # Gb\n\n\tlayout = {\n\t\tblock_device.path : {\n\t\t\t\"wipe\" : True,\n\t\t\t\"partitions\" : []\n\t\t}\n\t}\n\n\tlayout[block_device.path]['partitions'].append({\n\t\t# Boot\n\t\t\"type\" : \"primary\",\n\t\t\"start\" : \"1MiB\",\n\t\t\"size\" : \"513MiB\",\n\t\t\"boot\" : True,\n\t\t\"encrypted\" : False,\n\t\t\"format\" : True,\n\t\t\"mountpoint\" : \"/boot\",\n\t\t\"filesystem\" : {\n\t\t\t\"format\" : \"fat32\"\n\t\t}\n\t})\n\tlayout[block_device.path]['partitions'].append({\n\t\t# Root\n\t\t\"type\" : \"primary\",\n\t\t\"start\" : \"513MiB\",\n\t\t\"encrypted\" : False,\n\t\t\"format\" : True,\n\t\t\"size\" : \"100%\" if block_device.size < MIN_SIZE_TO_ALLOW_HOME_PART else f\"{min(block_device.size, 20)*1024}MiB\",\n\t\t\"mountpoint\" : \"/\",\n\t\t\"filesystem\" : {\n\t\t\t\"format\" : default_filesystem\n\t\t}\n\t})\n\n\tif default_filesystem == 'btrfs' and input('Would you like to use BTRFS subvolumes? (Y/n): ').strip().lower() in ('', 'y', 'yes'):\n\t\tif input('Do you want to use a recommended structure? (Y/n): ').strip().lower() in ('', 'y', 'yes'):\n\t\t\t# https://btrfs.wiki.kernel.org/index.php/FAQ\n\t\t\t# https://unix.stackexchange.com/questions/246976/btrfs-subvolume-uuid-clash\n\t\t\t# https://github.com/classy-giraffe/easy-arch/blob/main/easy-arch.sh\n\t\t\tlayout[block_device.path]['partitions'][1]['btrfs'] = {\n\t\t\t\t\"subvolumes\" : {\n\t\t\t\t\t\"@home\" : \"/home\",\n\t\t\t\t\t\"@log\" : \"/var/log\",\n\t\t\t\t\t\"@pkgs\" : \"/var/cache/pacman/pkg\",\n\t\t\t\t\t\"@.snapshots\" : \"/.snapshots\"\n\t\t\t\t}\n\t\t\t}\n\t\telse:\n\t\t\tpass # ... implement a guided setup\n\n\telif block_device.size >= MIN_SIZE_TO_ALLOW_HOME_PART:\n\t\t# If we don't want to use subvolumes,\n\t\t# But we want to be able to re-use data between re-installs..\n\t\t# A second partition for /home would be nice if we have the space for it\n\t\tlayout[block_device.path]['partitions'].append({\n\t\t\t# Home\n\t\t\t\"type\" : \"primary\",\n\t\t\t\"encrypted\" : False,\n\t\t\t\"format\" : True,\n\t\t\t\"start\" : f\"{min(block_device.size*0.2, 20)*1024}MiB\",\n\t\t\t\"size\" : \"100%\",\n\t\t\t\"mountpoint\" : \"/home\",\n\t\t\t\"filesystem\" : {\n\t\t\t\t\"format\" : default_filesystem\n\t\t\t}\n\t\t})\n\n\treturn layout\n\n\ndef suggest_multi_disk_layout(block_devices, default_filesystem=None):\n\tif not default_filesystem:\n\t\tfrom ..user_interaction import ask_for_main_filesystem_format\n\t\tdefault_filesystem = ask_for_main_filesystem_format()\n\n\t# Not really a rock solid foundation of information to stand on, but it's a start:\n\t# https://www.reddit.com/r/btrfs/comments/m287gp/partition_strategy_for_two_physical_disks/\n\t# https://www.reddit.com/r/btrfs/comments/9us4hr/what_is_your_btrfs_partitionsubvolumes_scheme/\n\n\tMIN_SIZE_TO_ALLOW_HOME_PART = 40 # Gb\n\tARCH_LINUX_INSTALLED_SIZE = 20 # Gb, rough estimate taking in to account user desktops etc. TODO: Catch user packages to detect size?\n\n\tblock_devices = sort_block_devices_based_on_performance(block_devices).keys()\n\n\thome_device = select_largest_device(block_devices, gigabytes=MIN_SIZE_TO_ALLOW_HOME_PART)\n\troot_device = select_disk_larger_than_or_close_to(block_devices, gigabytes=ARCH_LINUX_INSTALLED_SIZE, filter_out=[home_device])\n\n\tlog(f\"Suggesting multi-disk-layout using {len(block_devices)} disks, where {root_device} will be /root and {home_device} will be /home\", level=logging.DEBUG)\n\n\tlayout = {\n\t\troot_device.path : {\n\t\t\t\"wipe\" : True,\n\t\t\t\"partitions\" : []\n\t\t},\n\t\thome_device.path : {\n\t\t\t\"wipe\" : True,\n\t\t\t\"partitions\" : []\n\t\t},\n\t}\n\n\tlayout[root_device.path]['partitions'].append({\n\t\t# Boot\n\t\t\"type\" : \"primary\",\n\t\t\"start\" : \"1MiB\",\n\t\t\"size\" : \"513MiB\",\n\t\t\"boot\" : True,\n\t\t\"encrypted\" : False,\n\t\t\"format\" : True,\n\t\t\"mountpoint\" : \"/boot\",\n\t\t\"filesystem\" : {\n\t\t\t\"format\" : \"fat32\"\n\t\t}\n\t})\n\tlayout[root_device.path]['partitions'].append({\n\t\t# Root\n\t\t\"type\" : \"primary\",\n\t\t\"start\" : \"513MiB\",\n\t\t\"encrypted\" : False,\n\t\t\"format\" : True,\n\t\t\"size\" : \"100%\",\n\t\t\"mountpoint\" : \"/\",\n\t\t\"filesystem\" : {\n\t\t\t\"format\" : default_filesystem\n\t\t}\n\t})\n\n\tlayout[home_device.path]['partitions'].append({\n\t\t# Home\n\t\t\"type\" : \"primary\",\n\t\t\"encrypted\" : False,\n\t\t\"format\" : True,\n\t\t\"start\" : \"4MiB\",\n\t\t\"size\" : \"100%\",\n\t\t\"mountpoint\" : \"/home\",\n\t\t\"filesystem\" : {\n\t\t\t\"format\" : default_filesystem\n\t\t}\n\t})\n\n\treturn layout\n", "path": "archinstall/lib/disk/user_guides.py"}], "after_files": [{"content": "import logging\nfrom .helpers import sort_block_devices_based_on_performance, select_largest_device, select_disk_larger_than_or_close_to\nfrom ..output import log\n\ndef suggest_single_disk_layout(block_device, default_filesystem=None):\n\tif not default_filesystem:\n\t\tfrom ..user_interaction import ask_for_main_filesystem_format\n\t\tdefault_filesystem = ask_for_main_filesystem_format()\n\t\t\n\tMIN_SIZE_TO_ALLOW_HOME_PART = 40 # Gb\n\tusing_subvolumes = False\n\n\tif default_filesystem == 'btrfs':\n\t\tusing_subvolumes = input('Would you like to use BTRFS subvolumes? (Y/n): ').strip().lower() in ('', 'y', 'yes')\n\n\tlayout = {\n\t\tblock_device.path : {\n\t\t\t\"wipe\" : True,\n\t\t\t\"partitions\" : []\n\t\t}\n\t}\n\n\tlayout[block_device.path]['partitions'].append({\n\t\t# Boot\n\t\t\"type\" : \"primary\",\n\t\t\"start\" : \"1MiB\",\n\t\t\"size\" : \"513MiB\",\n\t\t\"boot\" : True,\n\t\t\"encrypted\" : False,\n\t\t\"format\" : True,\n\t\t\"mountpoint\" : \"/boot\",\n\t\t\"filesystem\" : {\n\t\t\t\"format\" : \"fat32\"\n\t\t}\n\t})\n\tlayout[block_device.path]['partitions'].append({\n\t\t# Root\n\t\t\"type\" : \"primary\",\n\t\t\"start\" : \"513MiB\",\n\t\t\"encrypted\" : False,\n\t\t\"format\" : True,\n\t\t\"size\" : \"100%\" if (using_subvolumes or block_device.size < MIN_SIZE_TO_ALLOW_HOME_PART) else f\"{min(block_device.size, 20)*1024}MiB\",\n\t\t\"mountpoint\" : \"/\",\n\t\t\"filesystem\" : {\n\t\t\t\"format\" : default_filesystem\n\t\t}\n\t})\n\n\tif default_filesystem == 'btrfs' and using_subvolumes:\n\t\tif input('Do you want to use a recommended structure? (Y/n): ').strip().lower() in ('', 'y', 'yes'):\n\t\t\t# https://btrfs.wiki.kernel.org/index.php/FAQ\n\t\t\t# https://unix.stackexchange.com/questions/246976/btrfs-subvolume-uuid-clash\n\t\t\t# https://github.com/classy-giraffe/easy-arch/blob/main/easy-arch.sh\n\t\t\tlayout[block_device.path]['partitions'][1]['btrfs'] = {\n\t\t\t\t\"subvolumes\" : {\n\t\t\t\t\t\"@home\" : \"/home\",\n\t\t\t\t\t\"@log\" : \"/var/log\",\n\t\t\t\t\t\"@pkgs\" : \"/var/cache/pacman/pkg\",\n\t\t\t\t\t\"@.snapshots\" : \"/.snapshots\"\n\t\t\t\t}\n\t\t\t}\n\t\telse:\n\t\t\tpass # ... implement a guided setup\n\n\telif block_device.size >= MIN_SIZE_TO_ALLOW_HOME_PART:\n\t\t# If we don't want to use subvolumes,\n\t\t# But we want to be able to re-use data between re-installs..\n\t\t# A second partition for /home would be nice if we have the space for it\n\t\tlayout[block_device.path]['partitions'].append({\n\t\t\t# Home\n\t\t\t\"type\" : \"primary\",\n\t\t\t\"encrypted\" : False,\n\t\t\t\"format\" : True,\n\t\t\t\"start\" : f\"{min(block_device.size*0.2, 20)*1024}MiB\",\n\t\t\t\"size\" : \"100%\",\n\t\t\t\"mountpoint\" : \"/home\",\n\t\t\t\"filesystem\" : {\n\t\t\t\t\"format\" : default_filesystem\n\t\t\t}\n\t\t})\n\n\treturn layout\n\n\ndef suggest_multi_disk_layout(block_devices, default_filesystem=None):\n\tif not default_filesystem:\n\t\tfrom ..user_interaction import ask_for_main_filesystem_format\n\t\tdefault_filesystem = ask_for_main_filesystem_format()\n\n\t# Not really a rock solid foundation of information to stand on, but it's a start:\n\t# https://www.reddit.com/r/btrfs/comments/m287gp/partition_strategy_for_two_physical_disks/\n\t# https://www.reddit.com/r/btrfs/comments/9us4hr/what_is_your_btrfs_partitionsubvolumes_scheme/\n\n\tMIN_SIZE_TO_ALLOW_HOME_PART = 40 # Gb\n\tARCH_LINUX_INSTALLED_SIZE = 20 # Gb, rough estimate taking in to account user desktops etc. TODO: Catch user packages to detect size?\n\n\tblock_devices = sort_block_devices_based_on_performance(block_devices).keys()\n\n\thome_device = select_largest_device(block_devices, gigabytes=MIN_SIZE_TO_ALLOW_HOME_PART)\n\troot_device = select_disk_larger_than_or_close_to(block_devices, gigabytes=ARCH_LINUX_INSTALLED_SIZE, filter_out=[home_device])\n\n\tlog(f\"Suggesting multi-disk-layout using {len(block_devices)} disks, where {root_device} will be /root and {home_device} will be /home\", level=logging.DEBUG)\n\n\tlayout = {\n\t\troot_device.path : {\n\t\t\t\"wipe\" : True,\n\t\t\t\"partitions\" : []\n\t\t},\n\t\thome_device.path : {\n\t\t\t\"wipe\" : True,\n\t\t\t\"partitions\" : []\n\t\t},\n\t}\n\n\tlayout[root_device.path]['partitions'].append({\n\t\t# Boot\n\t\t\"type\" : \"primary\",\n\t\t\"start\" : \"1MiB\",\n\t\t\"size\" : \"513MiB\",\n\t\t\"boot\" : True,\n\t\t\"encrypted\" : False,\n\t\t\"format\" : True,\n\t\t\"mountpoint\" : \"/boot\",\n\t\t\"filesystem\" : {\n\t\t\t\"format\" : \"fat32\"\n\t\t}\n\t})\n\tlayout[root_device.path]['partitions'].append({\n\t\t# Root\n\t\t\"type\" : \"primary\",\n\t\t\"start\" : \"513MiB\",\n\t\t\"encrypted\" : False,\n\t\t\"format\" : True,\n\t\t\"size\" : \"100%\",\n\t\t\"mountpoint\" : \"/\",\n\t\t\"filesystem\" : {\n\t\t\t\"format\" : default_filesystem\n\t\t}\n\t})\n\n\tlayout[home_device.path]['partitions'].append({\n\t\t# Home\n\t\t\"type\" : \"primary\",\n\t\t\"encrypted\" : False,\n\t\t\"format\" : True,\n\t\t\"start\" : \"4MiB\",\n\t\t\"size\" : \"100%\",\n\t\t\"mountpoint\" : \"/home\",\n\t\t\"filesystem\" : {\n\t\t\t\"format\" : default_filesystem\n\t\t}\n\t})\n\n\treturn layout\n", "path": "archinstall/lib/disk/user_guides.py"}]} | 2,156 | 458 |
gh_patches_debug_14365 | rasdani/github-patches | git_diff | comic__grand-challenge.org-1084 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Create a simple interface for fetching datatypes schemas on grand-challenge
**Problem**
The grand-challenge datatypes are currently only stored in the [gc-api](https://github.com/DIAGNijmegen/rse-gcapi/tree/master/gcapi/schemas) repository. However, the information is required by other libraries as well. Duplication of this information seems bad.
**Solution**
* [x] It would be nice to have this information in a central location like grand-challenge and provide a simple GET interface to allow the libraries /comic/evalutils and /DIAGNijmegen/rse-gcapi/ to fetch and cache this information (only a few kbs) from grand-challenge.
* [x] The answer type schemas should be added to the generated schema
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `app/grandchallenge/reader_studies/serializers.py`
Content:
```
1 from rest_framework.exceptions import ValidationError
2 from rest_framework.fields import CharField
3 from rest_framework.relations import HyperlinkedRelatedField, SlugRelatedField
4 from rest_framework.serializers import (
5 HyperlinkedModelSerializer,
6 SerializerMethodField,
7 )
8
9 from grandchallenge.api.swagger import swagger_schema_fields_for_charfield
10 from grandchallenge.cases.models import Image
11 from grandchallenge.reader_studies.models import Answer, Question, ReaderStudy
12
13
14 class QuestionSerializer(HyperlinkedModelSerializer):
15 answer_type = CharField(source="get_answer_type_display")
16 reader_study = HyperlinkedRelatedField(
17 view_name="api:reader-study-detail", read_only=True
18 )
19 form_direction = CharField(source="get_direction_display")
20 image_port = CharField(source="get_image_port_display")
21
22 class Meta:
23 model = Question
24 fields = (
25 "answer_type",
26 "api_url",
27 "form_direction",
28 "help_text",
29 "image_port",
30 "pk",
31 "question_text",
32 "reader_study",
33 "required",
34 )
35 swagger_schema_fields = swagger_schema_fields_for_charfield(
36 answer_type=model._meta.get_field("answer_type"),
37 form_direction=model._meta.get_field(
38 "direction"
39 ), # model.direction gets remapped
40 image_port=model._meta.get_field("image_port"),
41 )
42
43
44 class ReaderStudySerializer(HyperlinkedModelSerializer):
45 questions = QuestionSerializer(many=True, read_only=True)
46 hanging_list_images = SerializerMethodField()
47
48 class Meta:
49 model = ReaderStudy
50 fields = (
51 "api_url",
52 "description",
53 "hanging_list_images",
54 "is_valid",
55 "pk",
56 "questions",
57 "title",
58 )
59
60 def get_hanging_list_images(self, obj: ReaderStudy):
61 """Used by hanging_list_images serializer field."""
62 return obj.get_hanging_list_images_for_user(
63 user=self.context["request"].user
64 )
65
66
67 class AnswerSerializer(HyperlinkedModelSerializer):
68 creator = SlugRelatedField(read_only=True, slug_field="username")
69 question = HyperlinkedRelatedField(
70 view_name="api:reader-studies-question-detail",
71 queryset=Question.objects.all(),
72 )
73 images = HyperlinkedRelatedField(
74 many=True, queryset=Image.objects.all(), view_name="api:image-detail"
75 )
76
77 def validate(self, attrs):
78 question = attrs["question"]
79 images = attrs["images"]
80 answer = attrs["answer"]
81 creator = self.context.get("request").user
82
83 if not question.reader_study.is_reader(user=creator):
84 raise ValidationError("This user is not a reader for this study.")
85
86 if not question.is_answer_valid(answer=answer):
87 raise ValidationError(
88 f"You answer is not the correct type. "
89 f"{question.get_answer_type_display()} expected, "
90 f"{type(answer)} found."
91 )
92
93 if len(images) == 0:
94 raise ValidationError(
95 "You must specify the images that this answer corresponds to."
96 )
97
98 reader_study_images = question.reader_study.images.all()
99 for im in images:
100 if im not in reader_study_images:
101 raise ValidationError(
102 f"Image {im} does not belong to this reader study."
103 )
104
105 if Answer.objects.filter(
106 creator=creator, question=question, images__in=images
107 ).exists():
108 raise ValidationError(
109 f"User {creator} has already answered this question "
110 f"for at least 1 of these images."
111 )
112
113 return attrs
114
115 class Meta:
116 model = Answer
117 fields = (
118 "answer",
119 "api_url",
120 "created",
121 "creator",
122 "images",
123 "pk",
124 "question",
125 )
126
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/app/grandchallenge/reader_studies/serializers.py b/app/grandchallenge/reader_studies/serializers.py
--- a/app/grandchallenge/reader_studies/serializers.py
+++ b/app/grandchallenge/reader_studies/serializers.py
@@ -8,7 +8,12 @@
from grandchallenge.api.swagger import swagger_schema_fields_for_charfield
from grandchallenge.cases.models import Image
-from grandchallenge.reader_studies.models import Answer, Question, ReaderStudy
+from grandchallenge.reader_studies.models import (
+ ANSWER_TYPE_SCHEMA,
+ Answer,
+ Question,
+ ReaderStudy,
+)
class QuestionSerializer(HyperlinkedModelSerializer):
@@ -123,3 +128,6 @@
"pk",
"question",
)
+ swagger_schema_fields = {
+ "properties": {"answer": {"title": "Answer", **ANSWER_TYPE_SCHEMA}}
+ }
| {"golden_diff": "diff --git a/app/grandchallenge/reader_studies/serializers.py b/app/grandchallenge/reader_studies/serializers.py\n--- a/app/grandchallenge/reader_studies/serializers.py\n+++ b/app/grandchallenge/reader_studies/serializers.py\n@@ -8,7 +8,12 @@\n \n from grandchallenge.api.swagger import swagger_schema_fields_for_charfield\n from grandchallenge.cases.models import Image\n-from grandchallenge.reader_studies.models import Answer, Question, ReaderStudy\n+from grandchallenge.reader_studies.models import (\n+ ANSWER_TYPE_SCHEMA,\n+ Answer,\n+ Question,\n+ ReaderStudy,\n+)\n \n \n class QuestionSerializer(HyperlinkedModelSerializer):\n@@ -123,3 +128,6 @@\n \"pk\",\n \"question\",\n )\n+ swagger_schema_fields = {\n+ \"properties\": {\"answer\": {\"title\": \"Answer\", **ANSWER_TYPE_SCHEMA}}\n+ }\n", "issue": "Create a simple interface for fetching datatypes schemas on grand-challenge \n**Problem**\r\nThe grand-challenge datatypes are currently only stored in the [gc-api](https://github.com/DIAGNijmegen/rse-gcapi/tree/master/gcapi/schemas) repository. However, the information is required by other libraries as well. Duplication of this information seems bad.\r\n\r\n**Solution**\r\n* [x] It would be nice to have this information in a central location like grand-challenge and provide a simple GET interface to allow the libraries /comic/evalutils and /DIAGNijmegen/rse-gcapi/ to fetch and cache this information (only a few kbs) from grand-challenge.\r\n* [x] The answer type schemas should be added to the generated schema\r\n\n", "before_files": [{"content": "from rest_framework.exceptions import ValidationError\nfrom rest_framework.fields import CharField\nfrom rest_framework.relations import HyperlinkedRelatedField, SlugRelatedField\nfrom rest_framework.serializers import (\n HyperlinkedModelSerializer,\n SerializerMethodField,\n)\n\nfrom grandchallenge.api.swagger import swagger_schema_fields_for_charfield\nfrom grandchallenge.cases.models import Image\nfrom grandchallenge.reader_studies.models import Answer, Question, ReaderStudy\n\n\nclass QuestionSerializer(HyperlinkedModelSerializer):\n answer_type = CharField(source=\"get_answer_type_display\")\n reader_study = HyperlinkedRelatedField(\n view_name=\"api:reader-study-detail\", read_only=True\n )\n form_direction = CharField(source=\"get_direction_display\")\n image_port = CharField(source=\"get_image_port_display\")\n\n class Meta:\n model = Question\n fields = (\n \"answer_type\",\n \"api_url\",\n \"form_direction\",\n \"help_text\",\n \"image_port\",\n \"pk\",\n \"question_text\",\n \"reader_study\",\n \"required\",\n )\n swagger_schema_fields = swagger_schema_fields_for_charfield(\n answer_type=model._meta.get_field(\"answer_type\"),\n form_direction=model._meta.get_field(\n \"direction\"\n ), # model.direction gets remapped\n image_port=model._meta.get_field(\"image_port\"),\n )\n\n\nclass ReaderStudySerializer(HyperlinkedModelSerializer):\n questions = QuestionSerializer(many=True, read_only=True)\n hanging_list_images = SerializerMethodField()\n\n class Meta:\n model = ReaderStudy\n fields = (\n \"api_url\",\n \"description\",\n \"hanging_list_images\",\n \"is_valid\",\n \"pk\",\n \"questions\",\n \"title\",\n )\n\n def get_hanging_list_images(self, obj: ReaderStudy):\n \"\"\"Used by hanging_list_images serializer field.\"\"\"\n return obj.get_hanging_list_images_for_user(\n user=self.context[\"request\"].user\n )\n\n\nclass AnswerSerializer(HyperlinkedModelSerializer):\n creator = SlugRelatedField(read_only=True, slug_field=\"username\")\n question = HyperlinkedRelatedField(\n view_name=\"api:reader-studies-question-detail\",\n queryset=Question.objects.all(),\n )\n images = HyperlinkedRelatedField(\n many=True, queryset=Image.objects.all(), view_name=\"api:image-detail\"\n )\n\n def validate(self, attrs):\n question = attrs[\"question\"]\n images = attrs[\"images\"]\n answer = attrs[\"answer\"]\n creator = self.context.get(\"request\").user\n\n if not question.reader_study.is_reader(user=creator):\n raise ValidationError(\"This user is not a reader for this study.\")\n\n if not question.is_answer_valid(answer=answer):\n raise ValidationError(\n f\"You answer is not the correct type. \"\n f\"{question.get_answer_type_display()} expected, \"\n f\"{type(answer)} found.\"\n )\n\n if len(images) == 0:\n raise ValidationError(\n \"You must specify the images that this answer corresponds to.\"\n )\n\n reader_study_images = question.reader_study.images.all()\n for im in images:\n if im not in reader_study_images:\n raise ValidationError(\n f\"Image {im} does not belong to this reader study.\"\n )\n\n if Answer.objects.filter(\n creator=creator, question=question, images__in=images\n ).exists():\n raise ValidationError(\n f\"User {creator} has already answered this question \"\n f\"for at least 1 of these images.\"\n )\n\n return attrs\n\n class Meta:\n model = Answer\n fields = (\n \"answer\",\n \"api_url\",\n \"created\",\n \"creator\",\n \"images\",\n \"pk\",\n \"question\",\n )\n", "path": "app/grandchallenge/reader_studies/serializers.py"}], "after_files": [{"content": "from rest_framework.exceptions import ValidationError\nfrom rest_framework.fields import CharField\nfrom rest_framework.relations import HyperlinkedRelatedField, SlugRelatedField\nfrom rest_framework.serializers import (\n HyperlinkedModelSerializer,\n SerializerMethodField,\n)\n\nfrom grandchallenge.api.swagger import swagger_schema_fields_for_charfield\nfrom grandchallenge.cases.models import Image\nfrom grandchallenge.reader_studies.models import (\n ANSWER_TYPE_SCHEMA,\n Answer,\n Question,\n ReaderStudy,\n)\n\n\nclass QuestionSerializer(HyperlinkedModelSerializer):\n answer_type = CharField(source=\"get_answer_type_display\")\n reader_study = HyperlinkedRelatedField(\n view_name=\"api:reader-study-detail\", read_only=True\n )\n form_direction = CharField(source=\"get_direction_display\")\n image_port = CharField(source=\"get_image_port_display\")\n\n class Meta:\n model = Question\n fields = (\n \"answer_type\",\n \"api_url\",\n \"form_direction\",\n \"help_text\",\n \"image_port\",\n \"pk\",\n \"question_text\",\n \"reader_study\",\n \"required\",\n )\n swagger_schema_fields = swagger_schema_fields_for_charfield(\n answer_type=model._meta.get_field(\"answer_type\"),\n form_direction=model._meta.get_field(\n \"direction\"\n ), # model.direction gets remapped\n image_port=model._meta.get_field(\"image_port\"),\n )\n\n\nclass ReaderStudySerializer(HyperlinkedModelSerializer):\n questions = QuestionSerializer(many=True, read_only=True)\n hanging_list_images = SerializerMethodField()\n\n class Meta:\n model = ReaderStudy\n fields = (\n \"api_url\",\n \"description\",\n \"hanging_list_images\",\n \"is_valid\",\n \"pk\",\n \"questions\",\n \"title\",\n )\n\n def get_hanging_list_images(self, obj: ReaderStudy):\n \"\"\"Used by hanging_list_images serializer field.\"\"\"\n return obj.get_hanging_list_images_for_user(\n user=self.context[\"request\"].user\n )\n\n\nclass AnswerSerializer(HyperlinkedModelSerializer):\n creator = SlugRelatedField(read_only=True, slug_field=\"username\")\n question = HyperlinkedRelatedField(\n view_name=\"api:reader-studies-question-detail\",\n queryset=Question.objects.all(),\n )\n images = HyperlinkedRelatedField(\n many=True, queryset=Image.objects.all(), view_name=\"api:image-detail\"\n )\n\n def validate(self, attrs):\n question = attrs[\"question\"]\n images = attrs[\"images\"]\n answer = attrs[\"answer\"]\n creator = self.context.get(\"request\").user\n\n if not question.reader_study.is_reader(user=creator):\n raise ValidationError(\"This user is not a reader for this study.\")\n\n if not question.is_answer_valid(answer=answer):\n raise ValidationError(\n f\"You answer is not the correct type. \"\n f\"{question.get_answer_type_display()} expected, \"\n f\"{type(answer)} found.\"\n )\n\n if len(images) == 0:\n raise ValidationError(\n \"You must specify the images that this answer corresponds to.\"\n )\n\n reader_study_images = question.reader_study.images.all()\n for im in images:\n if im not in reader_study_images:\n raise ValidationError(\n f\"Image {im} does not belong to this reader study.\"\n )\n\n if Answer.objects.filter(\n creator=creator, question=question, images__in=images\n ).exists():\n raise ValidationError(\n f\"User {creator} has already answered this question \"\n f\"for at least 1 of these images.\"\n )\n\n return attrs\n\n class Meta:\n model = Answer\n fields = (\n \"answer\",\n \"api_url\",\n \"created\",\n \"creator\",\n \"images\",\n \"pk\",\n \"question\",\n )\n swagger_schema_fields = {\n \"properties\": {\"answer\": {\"title\": \"Answer\", **ANSWER_TYPE_SCHEMA}}\n }\n", "path": "app/grandchallenge/reader_studies/serializers.py"}]} | 1,497 | 205 |
gh_patches_debug_56573 | rasdani/github-patches | git_diff | TabbycatDebate__tabbycat-1056 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Reply-to address warning linking to incorrect URL
Bug report on: 2.3.0A
The "not specified reply-to" warning on the Email Participant pages links to an incorrect URL for the notification preferences returning a 404 error.
Linked to: https://asd-2019-jo.herokuapp.com/jo2019central/admin/options/notifications/
Correct: https://asd-2019-jo.herokuapp.com/jo2019central/admin/options/email/


UI Request: Add way to select/deselect all participants for the "email participants page"
UI feature request for 2.3.0A
The "Email Participants" page should have a way to select/deselect all participants from the page.
A screenshot has been included below of the Email Participants Adjudicator Assignments page.

@tienne-B thanks a lot for the new notification features, they make using emails tonnes easier and we are making good use of them already. Apologies for the stream of issues following this one.
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `tabbycat/notifications/views.py`
Content:
```
1 from datetime import datetime
2 import json
3 from smtplib import SMTPException
4
5 from asgiref.sync import async_to_sync
6 from channels.layers import get_channel_layer
7 from django.contrib import messages
8 from django.db.models import Prefetch, Q
9 from django.http import HttpResponse
10 from django.urls import reverse_lazy
11 from django.utils import timezone
12 from django.utils.translation import gettext as _, gettext_lazy, ngettext
13 from django.views.generic.base import View
14 from django.views.generic.edit import FormView
15
16 from participants.models import Person
17 from tournaments.mixins import RoundMixin, TournamentMixin
18 from utils.misc import reverse_tournament
19 from utils.mixins import AdministratorMixin
20 from utils.tables import TabbycatTableBuilder
21 from utils.views import VueTableTemplateView
22
23 from .forms import BasicEmailForm, TestEmailForm
24 from .models import EmailStatus, SentMessageRecord
25
26
27 class TestEmailView(AdministratorMixin, FormView):
28 form_class = TestEmailForm
29 template_name = 'test_email.html'
30 success_url = reverse_lazy('notifications-test-email')
31 view_role = ""
32
33 def form_valid(self, form):
34 host = self.request.get_host()
35 try:
36 recipient = form.send_email(host)
37 except (ConnectionError, SMTPException) as e:
38 messages.error(self.request,
39 _("There was an error sending the test email: %(error)s") % {'error': str(e)})
40 else:
41 messages.success(self.request,
42 _("A test email has been sent to %(recipient)s.") % {'recipient': recipient})
43 return super().form_valid(form)
44
45
46 class EmailStatusView(AdministratorMixin, TournamentMixin, VueTableTemplateView):
47 page_title = gettext_lazy("Email Statuses")
48 page_emoji = '📤'
49 template_name = 'email_statuses.html'
50
51 tables_orientation = 'rows'
52
53 def _create_status_timeline(self, status):
54 statuses = []
55 for s in status:
56 text = _("%(status)s @ %(time)s") % {'status': s.get_event_display(), 'time': s.timestamp}
57 statuses.append({
58 'text': '<span class="%s">%s</span>' % (self._get_event_class(s.event), text)
59 })
60 return statuses
61
62 def _get_event_class(self, event):
63 return {
64 EmailStatus.EVENT_TYPE_BOUNCED: 'text-warning',
65 EmailStatus.EVENT_TYPE_DROPPED: 'text-warning',
66 EmailStatus.EVENT_TYPE_SPAM: 'text-warning',
67 EmailStatus.EVENT_TYPE_DEFERRED: 'text-warning',
68 EmailStatus.EVENT_TYPE_PROCESSED: 'text-info',
69 EmailStatus.EVENT_TYPE_DELIVERED: 'text-info',
70 EmailStatus.EVENT_TYPE_OPENED: 'text-success',
71 EmailStatus.EVENT_TYPE_CLICKED: 'text-success',
72 EmailStatus.EVENT_TYPE_UNSUBSCRIBED: None,
73 EmailStatus.EVENT_TYPE_ASM_UNSUBSCRIBED: None,
74 EmailStatus.EVENT_TYPE_ASM_RESUBSCRIBED: None
75 }[event]
76
77 def get_tables(self):
78 tables = []
79 notifications = self.tournament.bulknotification_set.select_related('round').prefetch_related(
80 Prefetch('sentmessagerecord_set', queryset=SentMessageRecord.objects.select_related('recipient').prefetch_related('emailstatus_set')))
81
82 for n in notifications:
83 emails = n.sentmessagerecord_set.all()
84
85 subtitle = n.round.name if n.round is not None else _("@ %s") % timezone.localtime(n.timestamp).strftime("%a, %d %b %Y %H:%M:%S")
86 table = TabbycatTableBuilder(view=self, title=n.get_event_display().title(), subtitle=subtitle)
87
88 # Create arrays for columns
89 emails_status = []
90 emails_time = []
91 for e in emails:
92 status = e.emailstatus_set.all()
93 if status.count() == 0:
94 na_email = {'text': _("N/A"), 'class': 'text-muted'}
95 emails_status.append(na_email)
96 emails_time.append(na_email)
97 continue
98
99 first_status = status.first()
100 status_cell = {
101 "text": first_status.get_event_display(),
102 "class": self._get_event_class(first_status.event),
103 "popover": {"title": _("Timeline"), "content": self._create_status_timeline(status)}
104 }
105 emails_status.append(status_cell)
106 emails_time.append(first_status.timestamp)
107
108 table.add_column({'key': 'name', 'tooltip': _("Participant"), 'icon': 'user'}, [e.recipient.name for e in emails])
109 table.add_column({'key': 'name', 'title': _("Status")}, emails_status)
110 table.add_column({'key': 'name', 'title': _("Time")}, emails_time)
111
112 tables.append(table)
113
114 return tables
115
116
117 class EmailEventWebhookView(TournamentMixin, View):
118
119 def post(self, request, *args, **kwargs):
120 if kwargs['key'] is not self.tournament.pref('email_hook_key'):
121 return HttpResponse(status=404) # 404: Not Found
122
123 data = json.loads(request.body)
124
125 records = SentMessageRecord.objects.filter(message_id__in=[obj['smtp-id'] for obj in data])
126 record_lookup = {smr.message_id: smr.id for smr in records}
127 statuses = []
128
129 for obj in data:
130 dt = datetime.fromtimestamp(obj['timestamp'])
131 timestamp = timezone.make_aware(dt, timezone.utc)
132 email_id = record_lookup.get(obj['smtp-id'], None)
133 if email_id is None:
134 continue
135 statuses.append(EmailStatus(email_id=email_id, timestamp=timestamp, event=obj['event'], data=obj))
136
137 EmailStatus.objects.bulk_create(statuses)
138
139 return HttpResponse(status=201) # 201: Created
140
141
142 class BaseSelectPeopleEmailView(AdministratorMixin, TournamentMixin, VueTableTemplateView, FormView):
143 template_name = "email_participants.html"
144 page_title = gettext_lazy("Email Participants")
145 page_emoji = '📤'
146
147 form_class = BasicEmailForm
148
149 def get_context_data(self, **kwargs):
150 context = super().get_context_data(**kwargs)
151 context['sg_webhook'] = EmailStatus.objects.filter(email__notification__tournament=self.tournament).exists()
152 return context
153
154 def get_default_send_queryset(self):
155 return self.get_queryset().filter(email__isnull=False).exclude(email__exact="")
156
157 def get_queryset(self):
158 """All the people from the tournament who could receive the message"""
159 queryset_filter = Q(speaker__team__tournament=self.tournament) | Q(adjudicator__tournament=self.tournament)
160 if self.tournament.pref('share_adjs'):
161 queryset_filter |= Q(adjudicator__tournament__isnull=True)
162
163 return Person.objects.filter(queryset_filter).select_related('speaker', 'adjudicator')
164
165 def default_send(self, p, default_send_queryset=None):
166 """Whether the person should be emailed by default"""
167 return p in default_send_queryset
168
169 def get_table(self, mixed_participants=False):
170 table = TabbycatTableBuilder(view=self, sort_key='name')
171
172 queryset = self.get_queryset()
173 default_send_queryset = self.get_default_send_queryset()
174
175 table.add_column({'key': 'send', 'title': _("Send Email")}, [{
176 'component': 'check-cell',
177 'checked': self.default_send(p, default_send_queryset),
178 'id': p.id,
179 'name': 'recipients',
180 'value': p.id,
181 'noSave': True,
182 'type': 'adj' if mixed_participants and hasattr(p, 'adjudicator') else 'spk'
183 } for p in queryset])
184
185 table.add_column({'key': 'name', 'tooltip': _("Participant"), 'icon': 'user'}, [{
186 'text': p.name,
187 'class': 'no-wrap' if len(p.name) < 20 else ''
188 } for p in queryset])
189
190 table.add_column({'key': 'email', 'tooltip': _("Email Address"), 'icon': 'mail'}, [{
191 'text': p.email if p.email else _("Not Provided"),
192 'class': 'small' if p.email else 'small text-warning'
193 } for p in queryset])
194
195 return table
196
197
198 class RoleColumnMixin:
199 """Mixin to have a column Adjudicator/Speaker for email"""
200
201 def get_table(self, mixed_participants=True):
202 table = super().get_table(mixed_participants)
203
204 table.add_column({'key': 'role', 'title': _("Role")}, [{
205 'text': _("Adjudicator") if hasattr(p, 'adjudicator') else _("Speaker")
206 } for p in self.get_queryset()])
207
208 return table
209
210 def get_context_data(self, **kwargs):
211 context = super().get_context_data(**kwargs)
212 context['categories'] = [
213 {'id': 'spk', 'name': _("Speakers")},
214 {'id': 'adj', 'name': _("Adjudicators")}
215 ]
216 return context
217
218
219 class CustomEmailCreateView(RoleColumnMixin, BaseSelectPeopleEmailView):
220
221 def get_success_url(self):
222 return reverse_tournament('notifications-email', self.tournament)
223
224 def default_send(self, p, default_send_queryset):
225 return False
226
227 def post(self, request, *args, **kwargs):
228 people = Person.objects.filter(id__in=list(map(int, request.POST.getlist('recipients'))))
229
230 async_to_sync(get_channel_layer().send)("notifications", {
231 "type": "email_custom",
232 "subject": request.POST['subject_line'],
233 "body": request.POST['message_body'],
234 "tournament": self.tournament.id,
235 "send_to": [(p.id, p.email) for p in people]
236 })
237
238 messages.success(request, ngettext(
239 "%(count)s email has been queued for sending.",
240 "%(count)s emails have been queued for sending.",
241 len(people)
242 ) % {'count': len(people)})
243 return super().post(request, *args, **kwargs)
244
245
246 class TemplateEmailCreateView(BaseSelectPeopleEmailView):
247
248 def get_initial(self):
249 initial = super().get_initial()
250 initial['subject_line'] = self.tournament.pref(self.subject_template)
251 initial['message_body'] = self.tournament.pref(self.message_template)
252
253 return initial
254
255 def post(self, request, *args, **kwargs):
256 self.tournament.preferences[self.subject_template] = request.POST['subject_line']
257 self.tournament.preferences[self.message_template] = request.POST['message_body']
258 email_recipients = list(map(int, request.POST.getlist('recipients')))
259
260 async_to_sync(get_channel_layer().send)("notifications", {
261 "type": "email",
262 "message": self.event,
263 "extra": self.get_extra(),
264 "send_to": email_recipients,
265 "subject": request.POST['subject_line'],
266 "body": request.POST['message_body']
267 })
268
269 messages.success(request, ngettext(
270 "%(count)s email has been queued for sending.",
271 "%(count)s emails have been queued for sending.",
272 len(email_recipients)
273 ) % {'count': len(email_recipients)})
274 return super().post(request, *args, **kwargs)
275
276
277 class TournamentTemplateEmailCreateView(TemplateEmailCreateView):
278
279 def get_default_send_queryset(self):
280 return super().get_default_send_queryset().exclude(
281 sentmessagerecord__notification__event=self.event, sentmessagerecord__notification__tournament=self.tournament)
282
283 def get_extra(self):
284 extra = {'tournament_id': self.tournament.id}
285 return extra
286
287
288 class RoundTemplateEmailCreateView(TemplateEmailCreateView, RoundMixin):
289
290 def get_default_send_queryset(self):
291 return super().get_default_send_queryset().exclude(
292 sentmessagerecord__notification__event=self.event, sentmessagerecord__notification__round=self.round)
293
294 def get_extra(self):
295 extra = {'round_id': self.round.id}
296 return extra
297
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/tabbycat/notifications/views.py b/tabbycat/notifications/views.py
--- a/tabbycat/notifications/views.py
+++ b/tabbycat/notifications/views.py
@@ -149,6 +149,8 @@
def get_context_data(self, **kwargs):
context = super().get_context_data(**kwargs)
context['sg_webhook'] = EmailStatus.objects.filter(email__notification__tournament=self.tournament).exists()
+
+ context['categories'] = [{'id': 'spk', 'name': "Email"}]
return context
def get_default_send_queryset(self):
| {"golden_diff": "diff --git a/tabbycat/notifications/views.py b/tabbycat/notifications/views.py\n--- a/tabbycat/notifications/views.py\n+++ b/tabbycat/notifications/views.py\n@@ -149,6 +149,8 @@\n def get_context_data(self, **kwargs):\n context = super().get_context_data(**kwargs)\n context['sg_webhook'] = EmailStatus.objects.filter(email__notification__tournament=self.tournament).exists()\n+\n+ context['categories'] = [{'id': 'spk', 'name': \"Email\"}]\n return context\n \n def get_default_send_queryset(self):\n", "issue": "Reply-to address warning linking to incorrect URL\nBug report on: 2.3.0A\r\n\r\nThe \"not specified reply-to\" warning on the Email Participant pages links to an incorrect URL for the notification preferences returning a 404 error.\r\n\r\nLinked to: https://asd-2019-jo.herokuapp.com/jo2019central/admin/options/notifications/\r\nCorrect: https://asd-2019-jo.herokuapp.com/jo2019central/admin/options/email/\r\n\r\n\r\n\r\n\r\n\nUI Request: Add way to select/deselect all participants for the \"email participants page\"\nUI feature request for 2.3.0A\r\n\r\nThe \"Email Participants\" page should have a way to select/deselect all participants from the page.\r\n\r\nA screenshot has been included below of the Email Participants Adjudicator Assignments page. \r\n\r\n\r\n\r\n@tienne-B thanks a lot for the new notification features, they make using emails tonnes easier and we are making good use of them already. Apologies for the stream of issues following this one. \n", "before_files": [{"content": "from datetime import datetime\nimport json\nfrom smtplib import SMTPException\n\nfrom asgiref.sync import async_to_sync\nfrom channels.layers import get_channel_layer\nfrom django.contrib import messages\nfrom django.db.models import Prefetch, Q\nfrom django.http import HttpResponse\nfrom django.urls import reverse_lazy\nfrom django.utils import timezone\nfrom django.utils.translation import gettext as _, gettext_lazy, ngettext\nfrom django.views.generic.base import View\nfrom django.views.generic.edit import FormView\n\nfrom participants.models import Person\nfrom tournaments.mixins import RoundMixin, TournamentMixin\nfrom utils.misc import reverse_tournament\nfrom utils.mixins import AdministratorMixin\nfrom utils.tables import TabbycatTableBuilder\nfrom utils.views import VueTableTemplateView\n\nfrom .forms import BasicEmailForm, TestEmailForm\nfrom .models import EmailStatus, SentMessageRecord\n\n\nclass TestEmailView(AdministratorMixin, FormView):\n form_class = TestEmailForm\n template_name = 'test_email.html'\n success_url = reverse_lazy('notifications-test-email')\n view_role = \"\"\n\n def form_valid(self, form):\n host = self.request.get_host()\n try:\n recipient = form.send_email(host)\n except (ConnectionError, SMTPException) as e:\n messages.error(self.request,\n _(\"There was an error sending the test email: %(error)s\") % {'error': str(e)})\n else:\n messages.success(self.request,\n _(\"A test email has been sent to %(recipient)s.\") % {'recipient': recipient})\n return super().form_valid(form)\n\n\nclass EmailStatusView(AdministratorMixin, TournamentMixin, VueTableTemplateView):\n page_title = gettext_lazy(\"Email Statuses\")\n page_emoji = '\ud83d\udce4'\n template_name = 'email_statuses.html'\n\n tables_orientation = 'rows'\n\n def _create_status_timeline(self, status):\n statuses = []\n for s in status:\n text = _(\"%(status)s @ %(time)s\") % {'status': s.get_event_display(), 'time': s.timestamp}\n statuses.append({\n 'text': '<span class=\"%s\">%s</span>' % (self._get_event_class(s.event), text)\n })\n return statuses\n\n def _get_event_class(self, event):\n return {\n EmailStatus.EVENT_TYPE_BOUNCED: 'text-warning',\n EmailStatus.EVENT_TYPE_DROPPED: 'text-warning',\n EmailStatus.EVENT_TYPE_SPAM: 'text-warning',\n EmailStatus.EVENT_TYPE_DEFERRED: 'text-warning',\n EmailStatus.EVENT_TYPE_PROCESSED: 'text-info',\n EmailStatus.EVENT_TYPE_DELIVERED: 'text-info',\n EmailStatus.EVENT_TYPE_OPENED: 'text-success',\n EmailStatus.EVENT_TYPE_CLICKED: 'text-success',\n EmailStatus.EVENT_TYPE_UNSUBSCRIBED: None,\n EmailStatus.EVENT_TYPE_ASM_UNSUBSCRIBED: None,\n EmailStatus.EVENT_TYPE_ASM_RESUBSCRIBED: None\n }[event]\n\n def get_tables(self):\n tables = []\n notifications = self.tournament.bulknotification_set.select_related('round').prefetch_related(\n Prefetch('sentmessagerecord_set', queryset=SentMessageRecord.objects.select_related('recipient').prefetch_related('emailstatus_set')))\n\n for n in notifications:\n emails = n.sentmessagerecord_set.all()\n\n subtitle = n.round.name if n.round is not None else _(\"@ %s\") % timezone.localtime(n.timestamp).strftime(\"%a, %d %b %Y %H:%M:%S\")\n table = TabbycatTableBuilder(view=self, title=n.get_event_display().title(), subtitle=subtitle)\n\n # Create arrays for columns\n emails_status = []\n emails_time = []\n for e in emails:\n status = e.emailstatus_set.all()\n if status.count() == 0:\n na_email = {'text': _(\"N/A\"), 'class': 'text-muted'}\n emails_status.append(na_email)\n emails_time.append(na_email)\n continue\n\n first_status = status.first()\n status_cell = {\n \"text\": first_status.get_event_display(),\n \"class\": self._get_event_class(first_status.event),\n \"popover\": {\"title\": _(\"Timeline\"), \"content\": self._create_status_timeline(status)}\n }\n emails_status.append(status_cell)\n emails_time.append(first_status.timestamp)\n\n table.add_column({'key': 'name', 'tooltip': _(\"Participant\"), 'icon': 'user'}, [e.recipient.name for e in emails])\n table.add_column({'key': 'name', 'title': _(\"Status\")}, emails_status)\n table.add_column({'key': 'name', 'title': _(\"Time\")}, emails_time)\n\n tables.append(table)\n\n return tables\n\n\nclass EmailEventWebhookView(TournamentMixin, View):\n\n def post(self, request, *args, **kwargs):\n if kwargs['key'] is not self.tournament.pref('email_hook_key'):\n return HttpResponse(status=404) # 404: Not Found\n\n data = json.loads(request.body)\n\n records = SentMessageRecord.objects.filter(message_id__in=[obj['smtp-id'] for obj in data])\n record_lookup = {smr.message_id: smr.id for smr in records}\n statuses = []\n\n for obj in data:\n dt = datetime.fromtimestamp(obj['timestamp'])\n timestamp = timezone.make_aware(dt, timezone.utc)\n email_id = record_lookup.get(obj['smtp-id'], None)\n if email_id is None:\n continue\n statuses.append(EmailStatus(email_id=email_id, timestamp=timestamp, event=obj['event'], data=obj))\n\n EmailStatus.objects.bulk_create(statuses)\n\n return HttpResponse(status=201) # 201: Created\n\n\nclass BaseSelectPeopleEmailView(AdministratorMixin, TournamentMixin, VueTableTemplateView, FormView):\n template_name = \"email_participants.html\"\n page_title = gettext_lazy(\"Email Participants\")\n page_emoji = '\ud83d\udce4'\n\n form_class = BasicEmailForm\n\n def get_context_data(self, **kwargs):\n context = super().get_context_data(**kwargs)\n context['sg_webhook'] = EmailStatus.objects.filter(email__notification__tournament=self.tournament).exists()\n return context\n\n def get_default_send_queryset(self):\n return self.get_queryset().filter(email__isnull=False).exclude(email__exact=\"\")\n\n def get_queryset(self):\n \"\"\"All the people from the tournament who could receive the message\"\"\"\n queryset_filter = Q(speaker__team__tournament=self.tournament) | Q(adjudicator__tournament=self.tournament)\n if self.tournament.pref('share_adjs'):\n queryset_filter |= Q(adjudicator__tournament__isnull=True)\n\n return Person.objects.filter(queryset_filter).select_related('speaker', 'adjudicator')\n\n def default_send(self, p, default_send_queryset=None):\n \"\"\"Whether the person should be emailed by default\"\"\"\n return p in default_send_queryset\n\n def get_table(self, mixed_participants=False):\n table = TabbycatTableBuilder(view=self, sort_key='name')\n\n queryset = self.get_queryset()\n default_send_queryset = self.get_default_send_queryset()\n\n table.add_column({'key': 'send', 'title': _(\"Send Email\")}, [{\n 'component': 'check-cell',\n 'checked': self.default_send(p, default_send_queryset),\n 'id': p.id,\n 'name': 'recipients',\n 'value': p.id,\n 'noSave': True,\n 'type': 'adj' if mixed_participants and hasattr(p, 'adjudicator') else 'spk'\n } for p in queryset])\n\n table.add_column({'key': 'name', 'tooltip': _(\"Participant\"), 'icon': 'user'}, [{\n 'text': p.name,\n 'class': 'no-wrap' if len(p.name) < 20 else ''\n } for p in queryset])\n\n table.add_column({'key': 'email', 'tooltip': _(\"Email Address\"), 'icon': 'mail'}, [{\n 'text': p.email if p.email else _(\"Not Provided\"),\n 'class': 'small' if p.email else 'small text-warning'\n } for p in queryset])\n\n return table\n\n\nclass RoleColumnMixin:\n \"\"\"Mixin to have a column Adjudicator/Speaker for email\"\"\"\n\n def get_table(self, mixed_participants=True):\n table = super().get_table(mixed_participants)\n\n table.add_column({'key': 'role', 'title': _(\"Role\")}, [{\n 'text': _(\"Adjudicator\") if hasattr(p, 'adjudicator') else _(\"Speaker\")\n } for p in self.get_queryset()])\n\n return table\n\n def get_context_data(self, **kwargs):\n context = super().get_context_data(**kwargs)\n context['categories'] = [\n {'id': 'spk', 'name': _(\"Speakers\")},\n {'id': 'adj', 'name': _(\"Adjudicators\")}\n ]\n return context\n\n\nclass CustomEmailCreateView(RoleColumnMixin, BaseSelectPeopleEmailView):\n\n def get_success_url(self):\n return reverse_tournament('notifications-email', self.tournament)\n\n def default_send(self, p, default_send_queryset):\n return False\n\n def post(self, request, *args, **kwargs):\n people = Person.objects.filter(id__in=list(map(int, request.POST.getlist('recipients'))))\n\n async_to_sync(get_channel_layer().send)(\"notifications\", {\n \"type\": \"email_custom\",\n \"subject\": request.POST['subject_line'],\n \"body\": request.POST['message_body'],\n \"tournament\": self.tournament.id,\n \"send_to\": [(p.id, p.email) for p in people]\n })\n\n messages.success(request, ngettext(\n \"%(count)s email has been queued for sending.\",\n \"%(count)s emails have been queued for sending.\",\n len(people)\n ) % {'count': len(people)})\n return super().post(request, *args, **kwargs)\n\n\nclass TemplateEmailCreateView(BaseSelectPeopleEmailView):\n\n def get_initial(self):\n initial = super().get_initial()\n initial['subject_line'] = self.tournament.pref(self.subject_template)\n initial['message_body'] = self.tournament.pref(self.message_template)\n\n return initial\n\n def post(self, request, *args, **kwargs):\n self.tournament.preferences[self.subject_template] = request.POST['subject_line']\n self.tournament.preferences[self.message_template] = request.POST['message_body']\n email_recipients = list(map(int, request.POST.getlist('recipients')))\n\n async_to_sync(get_channel_layer().send)(\"notifications\", {\n \"type\": \"email\",\n \"message\": self.event,\n \"extra\": self.get_extra(),\n \"send_to\": email_recipients,\n \"subject\": request.POST['subject_line'],\n \"body\": request.POST['message_body']\n })\n\n messages.success(request, ngettext(\n \"%(count)s email has been queued for sending.\",\n \"%(count)s emails have been queued for sending.\",\n len(email_recipients)\n ) % {'count': len(email_recipients)})\n return super().post(request, *args, **kwargs)\n\n\nclass TournamentTemplateEmailCreateView(TemplateEmailCreateView):\n\n def get_default_send_queryset(self):\n return super().get_default_send_queryset().exclude(\n sentmessagerecord__notification__event=self.event, sentmessagerecord__notification__tournament=self.tournament)\n\n def get_extra(self):\n extra = {'tournament_id': self.tournament.id}\n return extra\n\n\nclass RoundTemplateEmailCreateView(TemplateEmailCreateView, RoundMixin):\n\n def get_default_send_queryset(self):\n return super().get_default_send_queryset().exclude(\n sentmessagerecord__notification__event=self.event, sentmessagerecord__notification__round=self.round)\n\n def get_extra(self):\n extra = {'round_id': self.round.id}\n return extra\n", "path": "tabbycat/notifications/views.py"}], "after_files": [{"content": "from datetime import datetime\nimport json\nfrom smtplib import SMTPException\n\nfrom asgiref.sync import async_to_sync\nfrom channels.layers import get_channel_layer\nfrom django.contrib import messages\nfrom django.db.models import Prefetch, Q\nfrom django.http import HttpResponse\nfrom django.urls import reverse_lazy\nfrom django.utils import timezone\nfrom django.utils.translation import gettext as _, gettext_lazy, ngettext\nfrom django.views.generic.base import View\nfrom django.views.generic.edit import FormView\n\nfrom participants.models import Person\nfrom tournaments.mixins import RoundMixin, TournamentMixin\nfrom utils.misc import reverse_tournament\nfrom utils.mixins import AdministratorMixin\nfrom utils.tables import TabbycatTableBuilder\nfrom utils.views import VueTableTemplateView\n\nfrom .forms import BasicEmailForm, TestEmailForm\nfrom .models import EmailStatus, SentMessageRecord\n\n\nclass TestEmailView(AdministratorMixin, FormView):\n form_class = TestEmailForm\n template_name = 'test_email.html'\n success_url = reverse_lazy('notifications-test-email')\n view_role = \"\"\n\n def form_valid(self, form):\n host = self.request.get_host()\n try:\n recipient = form.send_email(host)\n except (ConnectionError, SMTPException) as e:\n messages.error(self.request,\n _(\"There was an error sending the test email: %(error)s\") % {'error': str(e)})\n else:\n messages.success(self.request,\n _(\"A test email has been sent to %(recipient)s.\") % {'recipient': recipient})\n return super().form_valid(form)\n\n\nclass EmailStatusView(AdministratorMixin, TournamentMixin, VueTableTemplateView):\n page_title = gettext_lazy(\"Email Statuses\")\n page_emoji = '\ud83d\udce4'\n template_name = 'email_statuses.html'\n\n tables_orientation = 'rows'\n\n def _create_status_timeline(self, status):\n statuses = []\n for s in status:\n text = _(\"%(status)s @ %(time)s\") % {'status': s.get_event_display(), 'time': s.timestamp}\n statuses.append({\n 'text': '<span class=\"%s\">%s</span>' % (self._get_event_class(s.event), text)\n })\n return statuses\n\n def _get_event_class(self, event):\n return {\n EmailStatus.EVENT_TYPE_BOUNCED: 'text-warning',\n EmailStatus.EVENT_TYPE_DROPPED: 'text-warning',\n EmailStatus.EVENT_TYPE_SPAM: 'text-warning',\n EmailStatus.EVENT_TYPE_DEFERRED: 'text-warning',\n EmailStatus.EVENT_TYPE_PROCESSED: 'text-info',\n EmailStatus.EVENT_TYPE_DELIVERED: 'text-info',\n EmailStatus.EVENT_TYPE_OPENED: 'text-success',\n EmailStatus.EVENT_TYPE_CLICKED: 'text-success',\n EmailStatus.EVENT_TYPE_UNSUBSCRIBED: None,\n EmailStatus.EVENT_TYPE_ASM_UNSUBSCRIBED: None,\n EmailStatus.EVENT_TYPE_ASM_RESUBSCRIBED: None\n }[event]\n\n def get_tables(self):\n tables = []\n notifications = self.tournament.bulknotification_set.select_related('round').prefetch_related(\n Prefetch('sentmessagerecord_set', queryset=SentMessageRecord.objects.select_related('recipient').prefetch_related('emailstatus_set')))\n\n for n in notifications:\n emails = n.sentmessagerecord_set.all()\n\n subtitle = n.round.name if n.round is not None else _(\"@ %s\") % timezone.localtime(n.timestamp).strftime(\"%a, %d %b %Y %H:%M:%S\")\n table = TabbycatTableBuilder(view=self, title=n.get_event_display().title(), subtitle=subtitle)\n\n # Create arrays for columns\n emails_status = []\n emails_time = []\n for e in emails:\n status = e.emailstatus_set.all()\n if status.count() == 0:\n na_email = {'text': _(\"N/A\"), 'class': 'text-muted'}\n emails_status.append(na_email)\n emails_time.append(na_email)\n continue\n\n first_status = status.first()\n status_cell = {\n \"text\": first_status.get_event_display(),\n \"class\": self._get_event_class(first_status.event),\n \"popover\": {\"title\": _(\"Timeline\"), \"content\": self._create_status_timeline(status)}\n }\n emails_status.append(status_cell)\n emails_time.append(first_status.timestamp)\n\n table.add_column({'key': 'name', 'tooltip': _(\"Participant\"), 'icon': 'user'}, [e.recipient.name for e in emails])\n table.add_column({'key': 'name', 'title': _(\"Status\")}, emails_status)\n table.add_column({'key': 'name', 'title': _(\"Time\")}, emails_time)\n\n tables.append(table)\n\n return tables\n\n\nclass EmailEventWebhookView(TournamentMixin, View):\n\n def post(self, request, *args, **kwargs):\n if kwargs['key'] is not self.tournament.pref('email_hook_key'):\n return HttpResponse(status=404) # 404: Not Found\n\n data = json.loads(request.body)\n\n records = SentMessageRecord.objects.filter(message_id__in=[obj['smtp-id'] for obj in data])\n record_lookup = {smr.message_id: smr.id for smr in records}\n statuses = []\n\n for obj in data:\n dt = datetime.fromtimestamp(obj['timestamp'])\n timestamp = timezone.make_aware(dt, timezone.utc)\n email_id = record_lookup.get(obj['smtp-id'], None)\n if email_id is None:\n continue\n statuses.append(EmailStatus(email_id=email_id, timestamp=timestamp, event=obj['event'], data=obj))\n\n EmailStatus.objects.bulk_create(statuses)\n\n return HttpResponse(status=201) # 201: Created\n\n\nclass BaseSelectPeopleEmailView(AdministratorMixin, TournamentMixin, VueTableTemplateView, FormView):\n template_name = \"email_participants.html\"\n page_title = gettext_lazy(\"Email Participants\")\n page_emoji = '\ud83d\udce4'\n\n form_class = BasicEmailForm\n\n def get_context_data(self, **kwargs):\n context = super().get_context_data(**kwargs)\n context['sg_webhook'] = EmailStatus.objects.filter(email__notification__tournament=self.tournament).exists()\n\n context['categories'] = [{'id': 'spk', 'name': \"Email\"}]\n return context\n\n def get_default_send_queryset(self):\n return self.get_queryset().filter(email__isnull=False).exclude(email__exact=\"\")\n\n def get_queryset(self):\n \"\"\"All the people from the tournament who could receive the message\"\"\"\n queryset_filter = Q(speaker__team__tournament=self.tournament) | Q(adjudicator__tournament=self.tournament)\n if self.tournament.pref('share_adjs'):\n queryset_filter |= Q(adjudicator__tournament__isnull=True)\n\n return Person.objects.filter(queryset_filter).select_related('speaker', 'adjudicator')\n\n def default_send(self, p, default_send_queryset=None):\n \"\"\"Whether the person should be emailed by default\"\"\"\n return p in default_send_queryset\n\n def get_table(self, mixed_participants=False):\n table = TabbycatTableBuilder(view=self, sort_key='name')\n\n queryset = self.get_queryset()\n default_send_queryset = self.get_default_send_queryset()\n\n table.add_column({'key': 'send', 'title': _(\"Send Email\")}, [{\n 'component': 'check-cell',\n 'checked': self.default_send(p, default_send_queryset),\n 'id': p.id,\n 'name': 'recipients',\n 'value': p.id,\n 'noSave': True,\n 'type': 'adj' if mixed_participants and hasattr(p, 'adjudicator') else 'spk'\n } for p in queryset])\n\n table.add_column({'key': 'name', 'tooltip': _(\"Participant\"), 'icon': 'user'}, [{\n 'text': p.name,\n 'class': 'no-wrap' if len(p.name) < 20 else ''\n } for p in queryset])\n\n table.add_column({'key': 'email', 'tooltip': _(\"Email Address\"), 'icon': 'mail'}, [{\n 'text': p.email if p.email else _(\"Not Provided\"),\n 'class': 'small' if p.email else 'small text-warning'\n } for p in queryset])\n\n return table\n\n\nclass RoleColumnMixin:\n \"\"\"Mixin to have a column Adjudicator/Speaker for email\"\"\"\n\n def get_table(self, mixed_participants=True):\n table = super().get_table(mixed_participants)\n\n table.add_column({'key': 'role', 'title': _(\"Role\")}, [{\n 'text': _(\"Adjudicator\") if hasattr(p, 'adjudicator') else _(\"Speaker\")\n } for p in self.get_queryset()])\n\n return table\n\n def get_context_data(self, **kwargs):\n context = super().get_context_data(**kwargs)\n context['categories'] = [\n {'id': 'spk', 'name': _(\"Speakers\")},\n {'id': 'adj', 'name': _(\"Adjudicators\")}\n ]\n return context\n\n\nclass CustomEmailCreateView(RoleColumnMixin, BaseSelectPeopleEmailView):\n\n def get_success_url(self):\n return reverse_tournament('notifications-email', self.tournament)\n\n def default_send(self, p, default_send_queryset):\n return False\n\n def post(self, request, *args, **kwargs):\n people = Person.objects.filter(id__in=list(map(int, request.POST.getlist('recipients'))))\n\n async_to_sync(get_channel_layer().send)(\"notifications\", {\n \"type\": \"email_custom\",\n \"subject\": request.POST['subject_line'],\n \"body\": request.POST['message_body'],\n \"tournament\": self.tournament.id,\n \"send_to\": [(p.id, p.email) for p in people]\n })\n\n messages.success(request, ngettext(\n \"%(count)s email has been queued for sending.\",\n \"%(count)s emails have been queued for sending.\",\n len(people)\n ) % {'count': len(people)})\n return super().post(request, *args, **kwargs)\n\n\nclass TemplateEmailCreateView(BaseSelectPeopleEmailView):\n\n def get_initial(self):\n initial = super().get_initial()\n initial['subject_line'] = self.tournament.pref(self.subject_template)\n initial['message_body'] = self.tournament.pref(self.message_template)\n\n return initial\n\n def post(self, request, *args, **kwargs):\n self.tournament.preferences[self.subject_template] = request.POST['subject_line']\n self.tournament.preferences[self.message_template] = request.POST['message_body']\n email_recipients = list(map(int, request.POST.getlist('recipients')))\n\n async_to_sync(get_channel_layer().send)(\"notifications\", {\n \"type\": \"email\",\n \"message\": self.event,\n \"extra\": self.get_extra(),\n \"send_to\": email_recipients,\n \"subject\": request.POST['subject_line'],\n \"body\": request.POST['message_body']\n })\n\n messages.success(request, ngettext(\n \"%(count)s email has been queued for sending.\",\n \"%(count)s emails have been queued for sending.\",\n len(email_recipients)\n ) % {'count': len(email_recipients)})\n return super().post(request, *args, **kwargs)\n\n\nclass TournamentTemplateEmailCreateView(TemplateEmailCreateView):\n\n def get_default_send_queryset(self):\n return super().get_default_send_queryset().exclude(\n sentmessagerecord__notification__event=self.event, sentmessagerecord__notification__tournament=self.tournament)\n\n def get_extra(self):\n extra = {'tournament_id': self.tournament.id}\n return extra\n\n\nclass RoundTemplateEmailCreateView(TemplateEmailCreateView, RoundMixin):\n\n def get_default_send_queryset(self):\n return super().get_default_send_queryset().exclude(\n sentmessagerecord__notification__event=self.event, sentmessagerecord__notification__round=self.round)\n\n def get_extra(self):\n extra = {'round_id': self.round.id}\n return extra\n", "path": "tabbycat/notifications/views.py"}]} | 4,018 | 132 |
gh_patches_debug_12030 | rasdani/github-patches | git_diff | pytorch__vision-8256 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
`to_image` does not handle numpy 2D arrays
### 🐛 Describe the bug
[`to_image`](https://github.com/pytorch/vision/blob/806dba678d5b01f6e8a46f7c48fdf8c09369a267/torchvision/transforms/v2/functional/_type_conversion.py#L11) should be able to handle [numpy arrays](https://numpy.org/doc/stable/reference/generated/numpy.array.html) with shape `(H, W)`. This corresponds to the previous behavior of [`to_tensor`](https://github.com/pytorch/vision/blob/806dba678d5b01f6e8a46f7c48fdf8c09369a267/torchvision/transforms/functional.py#L149). Running the following:
```python
import numpy as np
from torchvision.transforms.v2.functional import to_image
img_npy = np.random.randint(0, 256, (224, 224), dtype=np.uint8)
to_image(img_npy)
```
results in error:
```
Traceback (most recent call last):
File "<stdin>", line 1, in <module>
File "/home/mantasu/programs/anaconda/envs/glasses-detector-312/lib/python3.12/site-packages/torchvision/transforms/v2/functional/_type_conversion.py", line 14, in to_image
output = torch.from_numpy(inpt).permute((2, 0, 1)).contiguous()
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
RuntimeError: permute(sparse_coo): number of dimensions in the tensor input does not match the length of the desired ordering of dimensions i.e. input.dim() = 2 is not equal to len(dims) = 3
```
PIL grayscale images are handled correctly:
```python
from PIL import Image
img_pil = Image.fromarray(img_npy)
print(to_image(img_pil).shape) # (1, 224, 224)
```
### Versions
```
PyTorch version: 2.2.0
Is debug build: False
CUDA used to build PyTorch: 12.1
ROCM used to build PyTorch: N/A
OS: Ubuntu 22.04.3 LTS (x86_64)
GCC version: (Ubuntu 11.4.0-1ubuntu1~22.04) 11.4.0
Clang version: Could not collect
CMake version: Could not collect
Libc version: glibc-2.35
Python version: 3.11.7 | packaged by conda-forge | (main, Dec 23 2023, 14:43:09) [GCC 12.3.0] (64-bit runtime)
Python platform: Linux-5.15.133.1-microsoft-standard-WSL2-x86_64-with-glibc2.35
Is CUDA available: True
CUDA runtime version: Could not collect
CUDA_MODULE_LOADING set to: LAZY
GPU models and configuration: GPU 0: NVIDIA GeForce RTX 3080 Ti Laptop GPU
Nvidia driver version: 546.33
cuDNN version: Probably one of the following:
/usr/lib/x86_64-linux-gnu/libcudnn.so.8.9.7
/usr/lib/x86_64-linux-gnu/libcudnn_adv_infer.so.8.9.7
/usr/lib/x86_64-linux-gnu/libcudnn_adv_train.so.8.9.7
/usr/lib/x86_64-linux-gnu/libcudnn_cnn_infer.so.8.9.7
/usr/lib/x86_64-linux-gnu/libcudnn_cnn_train.so.8.9.7
/usr/lib/x86_64-linux-gnu/libcudnn_ops_infer.so.8.9.7
/usr/lib/x86_64-linux-gnu/libcudnn_ops_train.so.8.9.7
HIP runtime version: N/A
MIOpen runtime version: N/A
Is XNNPACK available: True
CPU:
Architecture: x86_64
CPU op-mode(s): 32-bit, 64-bit
Address sizes: 39 bits physical, 48 bits virtual
Byte Order: Little Endian
CPU(s): 24
On-line CPU(s) list: 0-23
Vendor ID: GenuineIntel
Model name: 12th Gen Intel(R) Core(TM) i7-12800HX
CPU family: 6
Model: 151
Thread(s) per core: 2
Core(s) per socket: 12
Socket(s): 1
Stepping: 2
BogoMIPS: 4607.99
Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ss ht syscall nx pdpe1gb rdtscp lm constant_tsc rep_good nopl xtopology tsc_reliable nonstop_tsc cpuid pni pclmulqdq vmx ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand hypervisor lahf_lm abm 3dnowprefetch ssbd ibrs ibpb stibp ibrs_enhanced tpr_shadow vnmi ept vpid ept_ad fsgsbase tsc_adjust bmi1 avx2 smep bmi2 erms invpcid rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 xsaves avx_vnni umip waitpkg gfni vaes vpclmulqdq rdpid movdiri movdir64b fsrm serialize flush_l1d arch_capabilities
Virtualization: VT-x
Hypervisor vendor: Microsoft
Virtualization type: full
L1d cache: 576 KiB (12 instances)
L1i cache: 384 KiB (12 instances)
L2 cache: 15 MiB (12 instances)
L3 cache: 25 MiB (1 instance)
Vulnerability Gather data sampling: Not affected
Vulnerability Itlb multihit: Not affected
Vulnerability L1tf: Not affected
Vulnerability Mds: Not affected
Vulnerability Meltdown: Not affected
Vulnerability Mmio stale data: Not affected
Vulnerability Retbleed: Mitigation; Enhanced IBRS
Vulnerability Spec rstack overflow: Not affected
Vulnerability Spec store bypass: Mitigation; Speculative Store Bypass disabled via prctl and seccomp
Vulnerability Spectre v1: Mitigation; usercopy/swapgs barriers and __user pointer sanitization
Vulnerability Spectre v2: Mitigation; Enhanced IBRS, IBPB conditional, RSB filling, PBRSB-eIBRS SW sequence
Vulnerability Srbds: Not affected
Vulnerability Tsx async abort: Not affected
Versions of relevant libraries:
[pip3] numpy==1.26.3
[pip3] torch==2.2.0
[pip3] torchaudio==2.2.0
[pip3] torchvision==0.17.0
[pip3] triton==2.2.0
[conda] blas 1.0 mkl conda-forge
[conda] ffmpeg 4.3 hf484d3e_0 pytorch
[conda] libjpeg-turbo 2.0.0 h9bf148f_0 pytorch
[conda] mkl 2023.1.0 h213fc3f_46344
[conda] numpy 1.26.3 py311h64a7726_0 conda-forge
[conda] pytorch 2.2.0 py3.11_cuda12.1_cudnn8.9.2_0 pytorch
[conda] pytorch-cuda 12.1 ha16c6d3_5 pytorch
[conda] pytorch-mutex 1.0 cuda pytorch
[conda] torchaudio 2.2.0 py311_cu121 pytorch
[conda] torchtriton 2.2.0 py311 pytorch
[conda] torchvision 0.17.0 py311_cu121 pytorch
```
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `torchvision/transforms/v2/functional/_type_conversion.py`
Content:
```
1 from typing import Union
2
3 import numpy as np
4 import PIL.Image
5 import torch
6 from torchvision import tv_tensors
7 from torchvision.transforms import functional as _F
8
9
10 @torch.jit.unused
11 def to_image(inpt: Union[torch.Tensor, PIL.Image.Image, np.ndarray]) -> tv_tensors.Image:
12 """See :class:`~torchvision.transforms.v2.ToImage` for details."""
13 if isinstance(inpt, np.ndarray):
14 output = torch.from_numpy(inpt).permute((2, 0, 1)).contiguous()
15 elif isinstance(inpt, PIL.Image.Image):
16 output = pil_to_tensor(inpt)
17 elif isinstance(inpt, torch.Tensor):
18 output = inpt
19 else:
20 raise TypeError(
21 f"Input can either be a pure Tensor, a numpy array, or a PIL image, but got {type(inpt)} instead."
22 )
23 return tv_tensors.Image(output)
24
25
26 to_pil_image = _F.to_pil_image
27 pil_to_tensor = _F.pil_to_tensor
28
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/torchvision/transforms/v2/functional/_type_conversion.py b/torchvision/transforms/v2/functional/_type_conversion.py
--- a/torchvision/transforms/v2/functional/_type_conversion.py
+++ b/torchvision/transforms/v2/functional/_type_conversion.py
@@ -11,7 +11,7 @@
def to_image(inpt: Union[torch.Tensor, PIL.Image.Image, np.ndarray]) -> tv_tensors.Image:
"""See :class:`~torchvision.transforms.v2.ToImage` for details."""
if isinstance(inpt, np.ndarray):
- output = torch.from_numpy(inpt).permute((2, 0, 1)).contiguous()
+ output = torch.from_numpy(np.atleast_3d(inpt)).permute((2, 0, 1)).contiguous()
elif isinstance(inpt, PIL.Image.Image):
output = pil_to_tensor(inpt)
elif isinstance(inpt, torch.Tensor):
| {"golden_diff": "diff --git a/torchvision/transforms/v2/functional/_type_conversion.py b/torchvision/transforms/v2/functional/_type_conversion.py\n--- a/torchvision/transforms/v2/functional/_type_conversion.py\n+++ b/torchvision/transforms/v2/functional/_type_conversion.py\n@@ -11,7 +11,7 @@\n def to_image(inpt: Union[torch.Tensor, PIL.Image.Image, np.ndarray]) -> tv_tensors.Image:\n \"\"\"See :class:`~torchvision.transforms.v2.ToImage` for details.\"\"\"\n if isinstance(inpt, np.ndarray):\n- output = torch.from_numpy(inpt).permute((2, 0, 1)).contiguous()\n+ output = torch.from_numpy(np.atleast_3d(inpt)).permute((2, 0, 1)).contiguous()\n elif isinstance(inpt, PIL.Image.Image):\n output = pil_to_tensor(inpt)\n elif isinstance(inpt, torch.Tensor):\n", "issue": "`to_image` does not handle numpy 2D arrays\n### \ud83d\udc1b Describe the bug\r\n\r\n[`to_image`](https://github.com/pytorch/vision/blob/806dba678d5b01f6e8a46f7c48fdf8c09369a267/torchvision/transforms/v2/functional/_type_conversion.py#L11) should be able to handle [numpy arrays](https://numpy.org/doc/stable/reference/generated/numpy.array.html) with shape `(H, W)`. This corresponds to the previous behavior of [`to_tensor`](https://github.com/pytorch/vision/blob/806dba678d5b01f6e8a46f7c48fdf8c09369a267/torchvision/transforms/functional.py#L149). Running the following:\r\n\r\n\r\n```python\r\nimport numpy as np\r\nfrom torchvision.transforms.v2.functional import to_image\r\n\r\nimg_npy = np.random.randint(0, 256, (224, 224), dtype=np.uint8)\r\nto_image(img_npy)\r\n```\r\n\r\nresults in error:\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"/home/mantasu/programs/anaconda/envs/glasses-detector-312/lib/python3.12/site-packages/torchvision/transforms/v2/functional/_type_conversion.py\", line 14, in to_image\r\n output = torch.from_numpy(inpt).permute((2, 0, 1)).contiguous()\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\nRuntimeError: permute(sparse_coo): number of dimensions in the tensor input does not match the length of the desired ordering of dimensions i.e. input.dim() = 2 is not equal to len(dims) = 3\r\n```\r\n\r\n\r\nPIL grayscale images are handled correctly:\r\n```python\r\nfrom PIL import Image\r\n\r\nimg_pil = Image.fromarray(img_npy)\r\nprint(to_image(img_pil).shape) # (1, 224, 224)\r\n```\r\n\r\n\r\n\r\n### Versions\r\n\r\n```\r\nPyTorch version: 2.2.0\r\nIs debug build: False\r\nCUDA used to build PyTorch: 12.1\r\nROCM used to build PyTorch: N/A\r\n\r\nOS: Ubuntu 22.04.3 LTS (x86_64)\r\nGCC version: (Ubuntu 11.4.0-1ubuntu1~22.04) 11.4.0\r\nClang version: Could not collect\r\nCMake version: Could not collect\r\nLibc version: glibc-2.35\r\n\r\nPython version: 3.11.7 | packaged by conda-forge | (main, Dec 23 2023, 14:43:09) [GCC 12.3.0] (64-bit runtime)\r\nPython platform: Linux-5.15.133.1-microsoft-standard-WSL2-x86_64-with-glibc2.35\r\nIs CUDA available: True\r\nCUDA runtime version: Could not collect\r\nCUDA_MODULE_LOADING set to: LAZY\r\nGPU models and configuration: GPU 0: NVIDIA GeForce RTX 3080 Ti Laptop GPU\r\nNvidia driver version: 546.33\r\ncuDNN version: Probably one of the following:\r\n/usr/lib/x86_64-linux-gnu/libcudnn.so.8.9.7\r\n/usr/lib/x86_64-linux-gnu/libcudnn_adv_infer.so.8.9.7\r\n/usr/lib/x86_64-linux-gnu/libcudnn_adv_train.so.8.9.7\r\n/usr/lib/x86_64-linux-gnu/libcudnn_cnn_infer.so.8.9.7\r\n/usr/lib/x86_64-linux-gnu/libcudnn_cnn_train.so.8.9.7\r\n/usr/lib/x86_64-linux-gnu/libcudnn_ops_infer.so.8.9.7\r\n/usr/lib/x86_64-linux-gnu/libcudnn_ops_train.so.8.9.7\r\nHIP runtime version: N/A\r\nMIOpen runtime version: N/A\r\nIs XNNPACK available: True\r\n\r\nCPU:\r\nArchitecture: x86_64\r\nCPU op-mode(s): 32-bit, 64-bit\r\nAddress sizes: 39 bits physical, 48 bits virtual\r\nByte Order: Little Endian\r\nCPU(s): 24\r\nOn-line CPU(s) list: 0-23\r\nVendor ID: GenuineIntel\r\nModel name: 12th Gen Intel(R) Core(TM) i7-12800HX\r\nCPU family: 6\r\nModel: 151\r\nThread(s) per core: 2\r\nCore(s) per socket: 12\r\nSocket(s): 1\r\nStepping: 2\r\nBogoMIPS: 4607.99\r\nFlags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ss ht syscall nx pdpe1gb rdtscp lm constant_tsc rep_good nopl xtopology tsc_reliable nonstop_tsc cpuid pni pclmulqdq vmx ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand hypervisor lahf_lm abm 3dnowprefetch ssbd ibrs ibpb stibp ibrs_enhanced tpr_shadow vnmi ept vpid ept_ad fsgsbase tsc_adjust bmi1 avx2 smep bmi2 erms invpcid rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 xsaves avx_vnni umip waitpkg gfni vaes vpclmulqdq rdpid movdiri movdir64b fsrm serialize flush_l1d arch_capabilities\r\nVirtualization: VT-x\r\nHypervisor vendor: Microsoft\r\nVirtualization type: full\r\nL1d cache: 576 KiB (12 instances)\r\nL1i cache: 384 KiB (12 instances)\r\nL2 cache: 15 MiB (12 instances)\r\nL3 cache: 25 MiB (1 instance)\r\nVulnerability Gather data sampling: Not affected\r\nVulnerability Itlb multihit: Not affected\r\nVulnerability L1tf: Not affected\r\nVulnerability Mds: Not affected\r\nVulnerability Meltdown: Not affected\r\nVulnerability Mmio stale data: Not affected\r\nVulnerability Retbleed: Mitigation; Enhanced IBRS\r\nVulnerability Spec rstack overflow: Not affected\r\nVulnerability Spec store bypass: Mitigation; Speculative Store Bypass disabled via prctl and seccomp\r\nVulnerability Spectre v1: Mitigation; usercopy/swapgs barriers and __user pointer sanitization\r\nVulnerability Spectre v2: Mitigation; Enhanced IBRS, IBPB conditional, RSB filling, PBRSB-eIBRS SW sequence\r\nVulnerability Srbds: Not affected\r\nVulnerability Tsx async abort: Not affected\r\n\r\nVersions of relevant libraries:\r\n[pip3] numpy==1.26.3\r\n[pip3] torch==2.2.0\r\n[pip3] torchaudio==2.2.0\r\n[pip3] torchvision==0.17.0\r\n[pip3] triton==2.2.0\r\n[conda] blas 1.0 mkl conda-forge\r\n[conda] ffmpeg 4.3 hf484d3e_0 pytorch\r\n[conda] libjpeg-turbo 2.0.0 h9bf148f_0 pytorch\r\n[conda] mkl 2023.1.0 h213fc3f_46344 \r\n[conda] numpy 1.26.3 py311h64a7726_0 conda-forge\r\n[conda] pytorch 2.2.0 py3.11_cuda12.1_cudnn8.9.2_0 pytorch\r\n[conda] pytorch-cuda 12.1 ha16c6d3_5 pytorch\r\n[conda] pytorch-mutex 1.0 cuda pytorch\r\n[conda] torchaudio 2.2.0 py311_cu121 pytorch\r\n[conda] torchtriton 2.2.0 py311 pytorch\r\n[conda] torchvision 0.17.0 py311_cu121 pytorch\r\n```\n", "before_files": [{"content": "from typing import Union\n\nimport numpy as np\nimport PIL.Image\nimport torch\nfrom torchvision import tv_tensors\nfrom torchvision.transforms import functional as _F\n\n\[email protected]\ndef to_image(inpt: Union[torch.Tensor, PIL.Image.Image, np.ndarray]) -> tv_tensors.Image:\n \"\"\"See :class:`~torchvision.transforms.v2.ToImage` for details.\"\"\"\n if isinstance(inpt, np.ndarray):\n output = torch.from_numpy(inpt).permute((2, 0, 1)).contiguous()\n elif isinstance(inpt, PIL.Image.Image):\n output = pil_to_tensor(inpt)\n elif isinstance(inpt, torch.Tensor):\n output = inpt\n else:\n raise TypeError(\n f\"Input can either be a pure Tensor, a numpy array, or a PIL image, but got {type(inpt)} instead.\"\n )\n return tv_tensors.Image(output)\n\n\nto_pil_image = _F.to_pil_image\npil_to_tensor = _F.pil_to_tensor\n", "path": "torchvision/transforms/v2/functional/_type_conversion.py"}], "after_files": [{"content": "from typing import Union\n\nimport numpy as np\nimport PIL.Image\nimport torch\nfrom torchvision import tv_tensors\nfrom torchvision.transforms import functional as _F\n\n\[email protected]\ndef to_image(inpt: Union[torch.Tensor, PIL.Image.Image, np.ndarray]) -> tv_tensors.Image:\n \"\"\"See :class:`~torchvision.transforms.v2.ToImage` for details.\"\"\"\n if isinstance(inpt, np.ndarray):\n output = torch.from_numpy(np.atleast_3d(inpt)).permute((2, 0, 1)).contiguous()\n elif isinstance(inpt, PIL.Image.Image):\n output = pil_to_tensor(inpt)\n elif isinstance(inpt, torch.Tensor):\n output = inpt\n else:\n raise TypeError(\n f\"Input can either be a pure Tensor, a numpy array, or a PIL image, but got {type(inpt)} instead.\"\n )\n return tv_tensors.Image(output)\n\n\nto_pil_image = _F.to_pil_image\npil_to_tensor = _F.pil_to_tensor\n", "path": "torchvision/transforms/v2/functional/_type_conversion.py"}]} | 2,557 | 207 |
gh_patches_debug_36244 | rasdani/github-patches | git_diff | mars-project__mars-632 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
[ENH] More accurate memory stats using memory.stat in cgroup
Given kernel doc https://www.kernel.org/doc/Documentation/cgroup-v1/memory.txt, we may use ``/sys/fs/cgroup/memory/memory.stat`` to provide more accurate memory size estimations in Docker. Note that parsing this file is not supported in ``psutil`` now, thus we may parse it ourselves.
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `mars/resource.py`
Content:
```
1 # -*- coding: utf-8 -*-
2 # Copyright 1999-2018 Alibaba Group Holding Ltd.
3 #
4 # Licensed under the Apache License, Version 2.0 (the "License");
5 # you may not use this file except in compliance with the License.
6 # You may obtain a copy of the License at
7 #
8 # http://www.apache.org/licenses/LICENSE-2.0
9 #
10 # Unless required by applicable law or agreed to in writing, software
11 # distributed under the License is distributed on an "AS IS" BASIS,
12 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13 # See the License for the specific language governing permissions and
14 # limitations under the License.
15
16 import os
17 import subprocess # nosec
18 import sys
19 import time
20 from collections import namedtuple
21
22 import psutil
23
24 from .lib import nvutils
25
26 _proc = psutil.Process()
27 _timer = getattr(time, 'monotonic', time.time)
28
29 _cpu_use_process_stat = bool(int(os.environ.get('MARS_CPU_USE_PROCESS_STAT', '0').strip('"')))
30 _mem_use_process_stat = bool(int(os.environ.get('MARS_MEM_USE_PROCESS_STAT', '0').strip('"')))
31
32 if 'MARS_USE_PROCESS_STAT' in os.environ:
33 _cpu_use_process_stat = _mem_use_process_stat = \
34 bool(int(os.environ['MARS_USE_PROCESS_STAT'].strip('"')))
35
36 if 'MARS_CPU_TOTAL' in os.environ:
37 _cpu_total = int(os.environ['MARS_CPU_TOTAL'].strip('"'))
38 else:
39 _cpu_total = psutil.cpu_count(logical=True)
40
41 if 'MARS_MEMORY_TOTAL' in os.environ:
42 _mem_total = int(os.environ['MARS_MEMORY_TOTAL'].strip('"'))
43 else:
44 _mem_total = None
45
46 _virt_memory_stat = namedtuple('virtual_memory', 'total available percent used free')
47
48 _shm_path = [pt.mountpoint for pt in psutil.disk_partitions(all=True)
49 if pt.mountpoint in ('/tmp', '/dev/shm') and pt.fstype == 'tmpfs']
50 if not _shm_path:
51 _shm_path = None
52 else:
53 _shm_path = _shm_path[0]
54
55
56 def virtual_memory():
57 sys_mem = psutil.virtual_memory()
58 if not _mem_use_process_stat:
59 total = sys_mem.total
60 used = sys_mem.used + getattr(sys_mem, 'shared', 0)
61 available = sys_mem.available
62 free = sys_mem.free
63 percent = 100.0 * (total - available) / total
64 return _virt_memory_stat(total, available, percent, used, free)
65 else:
66 used = 0
67 for p in psutil.process_iter():
68 try:
69 used += p.memory_info().rss
70 except (psutil.NoSuchProcess, psutil.AccessDenied):
71 pass
72
73 if _shm_path:
74 shm_stats = psutil.disk_usage(_shm_path)
75 used += shm_stats.used
76
77 total = min(_mem_total or sys_mem.total, sys_mem.total)
78 # TODO sys_mem.available does not work in container
79 # available = min(sys_mem.available, total - used)
80 available = total - used
81 free = min(sys_mem.free, total - used)
82 percent = 100.0 * (total - available) / total
83 return _virt_memory_stat(total, available, percent, used, free)
84
85
86 def cpu_count():
87 return _cpu_total
88
89
90 _last_cpu_measure = None
91
92
93 def _take_process_cpu_snapshot():
94 num_cpus = cpu_count() or 1
95
96 def timer():
97 return _timer() * num_cpus
98
99 processes = [p for p in psutil.process_iter() if p.pid != _proc.pid]
100
101 pts = dict()
102 sts = dict()
103 for p in processes:
104 try:
105 pts[p.pid] = p.cpu_times()
106 sts[p.pid] = timer()
107 except (psutil.NoSuchProcess, psutil.AccessDenied):
108 pass
109
110 pts[_proc.pid] = _proc.cpu_times()
111 sts[_proc.pid] = timer()
112 return pts, sts
113
114
115 def cpu_percent():
116 global _last_cpu_measure
117 if not _cpu_use_process_stat:
118 return sum(psutil.cpu_percent(percpu=True))
119
120 num_cpus = cpu_count() or 1
121 pts, sts = _take_process_cpu_snapshot()
122
123 if _last_cpu_measure is None:
124 _last_cpu_measure = (pts, sts)
125 return None
126
127 old_pts, old_sts = _last_cpu_measure
128
129 percents = []
130 for pid in pts:
131 if pid not in old_pts:
132 continue
133 pt1 = old_pts[pid]
134 pt2 = pts[pid]
135 delta_proc = (pt2.user - pt1.user) + (pt2.system - pt1.system)
136 delta_time = sts[pid] - old_sts[pid]
137
138 try:
139 overall_cpus_percent = (delta_proc / delta_time) * 100
140 except ZeroDivisionError:
141 percents.append(0.0)
142 else:
143 single_cpu_percent = overall_cpus_percent * num_cpus
144 percents.append(single_cpu_percent)
145 _last_cpu_measure = (pts, sts)
146 return round(sum(percents), 1)
147
148
149 def disk_usage(d):
150 return psutil.disk_usage(d)
151
152
153 def iowait():
154 cpu_percent = psutil.cpu_times_percent()
155 try:
156 return cpu_percent.iowait
157 except AttributeError:
158 return None
159
160
161 _last_disk_io_meta = None
162 _win_diskperf_called = False
163
164
165 def disk_io_usage():
166 global _last_disk_io_meta, _win_diskperf_called
167
168 # Needed by psutil.disk_io_counters() under newer version of Windows.
169 # diskperf -y need to be called or no disk information can be found.
170 if sys.platform == 'win32' and not _win_diskperf_called: # pragma: no cover
171 CREATE_NO_WINDOW = 0x08000000
172 try:
173 proc = subprocess.Popen(['diskperf', '-y'], shell=False,
174 creationflags=CREATE_NO_WINDOW) # nosec
175 proc.wait()
176 except (subprocess.CalledProcessError, OSError):
177 pass
178 _win_diskperf_called = True
179
180 disk_counters = psutil.disk_io_counters()
181 tst = time.time()
182
183 read_bytes = disk_counters.read_bytes
184 write_bytes = disk_counters.write_bytes
185 if _last_disk_io_meta is None:
186 _last_disk_io_meta = (read_bytes, write_bytes, tst)
187 return None
188
189 last_read_bytes, last_write_bytes, last_time = _last_disk_io_meta
190 delta_time = tst - last_time
191 read_speed = (read_bytes - last_read_bytes) / delta_time
192 write_speed = (write_bytes - last_write_bytes) / delta_time
193
194 _last_disk_io_meta = (read_bytes, write_bytes, tst)
195 return read_speed, write_speed
196
197
198 _last_net_io_meta = None
199
200
201 def net_io_usage():
202 global _last_net_io_meta
203
204 net_counters = psutil.net_io_counters()
205 tst = time.time()
206
207 send_bytes = net_counters.bytes_sent
208 recv_bytes = net_counters.bytes_recv
209 if _last_net_io_meta is None:
210 _last_net_io_meta = (send_bytes, recv_bytes, tst)
211 return None
212
213 last_send_bytes, last_recv_bytes, last_time = _last_net_io_meta
214 delta_time = tst - last_time
215 recv_speed = (recv_bytes - last_recv_bytes) / delta_time
216 send_speed = (send_bytes - last_send_bytes) / delta_time
217
218 _last_net_io_meta = (send_bytes, recv_bytes, tst)
219 return recv_speed, send_speed
220
221
222 _cuda_info = namedtuple('cuda_info', 'driver_version cuda_version products gpu_count')
223 _cuda_card_stat = namedtuple('cuda_card_stat', 'product_name gpu_usage temperature fb_mem_info')
224
225
226 def cuda_info(): # pragma: no cover
227 driver_info = nvutils.get_driver_info()
228 if not driver_info:
229 return
230 gpu_count = nvutils.get_device_count()
231 return _cuda_info(
232 driver_version=driver_info.driver_version,
233 cuda_version=driver_info.cuda_version,
234 products=[nvutils.get_device_info(idx).name for idx in range(gpu_count)],
235 gpu_count=gpu_count,
236 )
237
238
239 def cuda_card_stats(): # pragma: no cover
240 infos = []
241 device_count = nvutils.get_device_count()
242 if not device_count:
243 return
244 for device_idx in range(device_count):
245 device_info = nvutils.get_device_info(device_idx)
246 device_status = nvutils.get_device_status(device_idx)
247
248 infos.append(_cuda_card_stat(
249 product_name=device_info.name,
250 gpu_usage=device_status.gpu_util,
251 temperature=device_status.temperature,
252 fb_mem_info=_virt_memory_stat(
253 total=device_status.fb_total_mem, used=device_status.fb_used_mem,
254 free=device_status.fb_free_mem, available=device_status.fb_free_mem,
255 percent=device_status.mem_util,
256 )
257 ))
258 return infos
259
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/mars/resource.py b/mars/resource.py
--- a/mars/resource.py
+++ b/mars/resource.py
@@ -23,22 +23,25 @@
from .lib import nvutils
+CGROUP_MEM_STAT_FILE = '/sys/fs/cgroup/memory/memory.stat'
+
_proc = psutil.Process()
_timer = getattr(time, 'monotonic', time.time)
_cpu_use_process_stat = bool(int(os.environ.get('MARS_CPU_USE_PROCESS_STAT', '0').strip('"')))
_mem_use_process_stat = bool(int(os.environ.get('MARS_MEM_USE_PROCESS_STAT', '0').strip('"')))
+_mem_use_cgroup_stat = bool(int(os.environ.get('MARS_MEM_USE_CGROUP_STAT', '0').strip('"')))
if 'MARS_USE_PROCESS_STAT' in os.environ:
_cpu_use_process_stat = _mem_use_process_stat = \
bool(int(os.environ['MARS_USE_PROCESS_STAT'].strip('"')))
-if 'MARS_CPU_TOTAL' in os.environ:
+if _cpu_use_process_stat and 'MARS_CPU_TOTAL' in os.environ:
_cpu_total = int(os.environ['MARS_CPU_TOTAL'].strip('"'))
else:
_cpu_total = psutil.cpu_count(logical=True)
-if 'MARS_MEMORY_TOTAL' in os.environ:
+if _mem_use_process_stat and 'MARS_MEMORY_TOTAL' in os.environ:
_mem_total = int(os.environ['MARS_MEMORY_TOTAL'].strip('"'))
else:
_mem_total = None
@@ -53,9 +56,28 @@
_shm_path = _shm_path[0]
+def _read_cgroup_stat_file():
+ with open(CGROUP_MEM_STAT_FILE, 'r') as cg_file:
+ contents = cg_file.read()
+ kvs = dict()
+ for l in contents.splitlines():
+ parts = l.split(' ')
+ if len(parts) == 2:
+ kvs[parts[0]] = int(parts[1])
+ return kvs
+
+
def virtual_memory():
sys_mem = psutil.virtual_memory()
- if not _mem_use_process_stat:
+ if _mem_use_cgroup_stat:
+ # see section 5.5 in https://www.kernel.org/doc/Documentation/cgroup-v1/memory.txt
+ cgroup_mem_info = _read_cgroup_stat_file()
+ total = cgroup_mem_info['hierarchical_memory_limit']
+ used = cgroup_mem_info['cache'] + cgroup_mem_info['rss'] + cgroup_mem_info.get('swap', 0)
+ available = free = total - used
+ percent = 100.0 * (total - available) / total
+ return _virt_memory_stat(total, available, percent, used, free)
+ elif not _mem_use_process_stat:
total = sys_mem.total
used = sys_mem.used + getattr(sys_mem, 'shared', 0)
available = sys_mem.available
| {"golden_diff": "diff --git a/mars/resource.py b/mars/resource.py\n--- a/mars/resource.py\n+++ b/mars/resource.py\n@@ -23,22 +23,25 @@\n \n from .lib import nvutils\n \n+CGROUP_MEM_STAT_FILE = '/sys/fs/cgroup/memory/memory.stat'\n+\n _proc = psutil.Process()\n _timer = getattr(time, 'monotonic', time.time)\n \n _cpu_use_process_stat = bool(int(os.environ.get('MARS_CPU_USE_PROCESS_STAT', '0').strip('\"')))\n _mem_use_process_stat = bool(int(os.environ.get('MARS_MEM_USE_PROCESS_STAT', '0').strip('\"')))\n+_mem_use_cgroup_stat = bool(int(os.environ.get('MARS_MEM_USE_CGROUP_STAT', '0').strip('\"')))\n \n if 'MARS_USE_PROCESS_STAT' in os.environ:\n _cpu_use_process_stat = _mem_use_process_stat = \\\n bool(int(os.environ['MARS_USE_PROCESS_STAT'].strip('\"')))\n \n-if 'MARS_CPU_TOTAL' in os.environ:\n+if _cpu_use_process_stat and 'MARS_CPU_TOTAL' in os.environ:\n _cpu_total = int(os.environ['MARS_CPU_TOTAL'].strip('\"'))\n else:\n _cpu_total = psutil.cpu_count(logical=True)\n \n-if 'MARS_MEMORY_TOTAL' in os.environ:\n+if _mem_use_process_stat and 'MARS_MEMORY_TOTAL' in os.environ:\n _mem_total = int(os.environ['MARS_MEMORY_TOTAL'].strip('\"'))\n else:\n _mem_total = None\n@@ -53,9 +56,28 @@\n _shm_path = _shm_path[0]\n \n \n+def _read_cgroup_stat_file():\n+ with open(CGROUP_MEM_STAT_FILE, 'r') as cg_file:\n+ contents = cg_file.read()\n+ kvs = dict()\n+ for l in contents.splitlines():\n+ parts = l.split(' ')\n+ if len(parts) == 2:\n+ kvs[parts[0]] = int(parts[1])\n+ return kvs\n+\n+\n def virtual_memory():\n sys_mem = psutil.virtual_memory()\n- if not _mem_use_process_stat:\n+ if _mem_use_cgroup_stat:\n+ # see section 5.5 in https://www.kernel.org/doc/Documentation/cgroup-v1/memory.txt\n+ cgroup_mem_info = _read_cgroup_stat_file()\n+ total = cgroup_mem_info['hierarchical_memory_limit']\n+ used = cgroup_mem_info['cache'] + cgroup_mem_info['rss'] + cgroup_mem_info.get('swap', 0)\n+ available = free = total - used\n+ percent = 100.0 * (total - available) / total\n+ return _virt_memory_stat(total, available, percent, used, free)\n+ elif not _mem_use_process_stat:\n total = sys_mem.total\n used = sys_mem.used + getattr(sys_mem, 'shared', 0)\n available = sys_mem.available\n", "issue": "[ENH] More accurate memory stats using memory.stat in cgroup\nGiven kernel doc https://www.kernel.org/doc/Documentation/cgroup-v1/memory.txt, we may use ``/sys/fs/cgroup/memory/memory.stat`` to provide more accurate memory size estimations in Docker. Note that parsing this file is not supported in ``psutil`` now, thus we may parse it ourselves.\n", "before_files": [{"content": "# -*- coding: utf-8 -*-\n# Copyright 1999-2018 Alibaba Group Holding Ltd.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nimport os\nimport subprocess # nosec\nimport sys\nimport time\nfrom collections import namedtuple\n\nimport psutil\n\nfrom .lib import nvutils\n\n_proc = psutil.Process()\n_timer = getattr(time, 'monotonic', time.time)\n\n_cpu_use_process_stat = bool(int(os.environ.get('MARS_CPU_USE_PROCESS_STAT', '0').strip('\"')))\n_mem_use_process_stat = bool(int(os.environ.get('MARS_MEM_USE_PROCESS_STAT', '0').strip('\"')))\n\nif 'MARS_USE_PROCESS_STAT' in os.environ:\n _cpu_use_process_stat = _mem_use_process_stat = \\\n bool(int(os.environ['MARS_USE_PROCESS_STAT'].strip('\"')))\n\nif 'MARS_CPU_TOTAL' in os.environ:\n _cpu_total = int(os.environ['MARS_CPU_TOTAL'].strip('\"'))\nelse:\n _cpu_total = psutil.cpu_count(logical=True)\n\nif 'MARS_MEMORY_TOTAL' in os.environ:\n _mem_total = int(os.environ['MARS_MEMORY_TOTAL'].strip('\"'))\nelse:\n _mem_total = None\n\n_virt_memory_stat = namedtuple('virtual_memory', 'total available percent used free')\n\n_shm_path = [pt.mountpoint for pt in psutil.disk_partitions(all=True)\n if pt.mountpoint in ('/tmp', '/dev/shm') and pt.fstype == 'tmpfs']\nif not _shm_path:\n _shm_path = None\nelse:\n _shm_path = _shm_path[0]\n\n\ndef virtual_memory():\n sys_mem = psutil.virtual_memory()\n if not _mem_use_process_stat:\n total = sys_mem.total\n used = sys_mem.used + getattr(sys_mem, 'shared', 0)\n available = sys_mem.available\n free = sys_mem.free\n percent = 100.0 * (total - available) / total\n return _virt_memory_stat(total, available, percent, used, free)\n else:\n used = 0\n for p in psutil.process_iter():\n try:\n used += p.memory_info().rss\n except (psutil.NoSuchProcess, psutil.AccessDenied):\n pass\n\n if _shm_path:\n shm_stats = psutil.disk_usage(_shm_path)\n used += shm_stats.used\n\n total = min(_mem_total or sys_mem.total, sys_mem.total)\n # TODO sys_mem.available does not work in container\n # available = min(sys_mem.available, total - used)\n available = total - used\n free = min(sys_mem.free, total - used)\n percent = 100.0 * (total - available) / total\n return _virt_memory_stat(total, available, percent, used, free)\n\n\ndef cpu_count():\n return _cpu_total\n\n\n_last_cpu_measure = None\n\n\ndef _take_process_cpu_snapshot():\n num_cpus = cpu_count() or 1\n\n def timer():\n return _timer() * num_cpus\n\n processes = [p for p in psutil.process_iter() if p.pid != _proc.pid]\n\n pts = dict()\n sts = dict()\n for p in processes:\n try:\n pts[p.pid] = p.cpu_times()\n sts[p.pid] = timer()\n except (psutil.NoSuchProcess, psutil.AccessDenied):\n pass\n\n pts[_proc.pid] = _proc.cpu_times()\n sts[_proc.pid] = timer()\n return pts, sts\n\n\ndef cpu_percent():\n global _last_cpu_measure\n if not _cpu_use_process_stat:\n return sum(psutil.cpu_percent(percpu=True))\n\n num_cpus = cpu_count() or 1\n pts, sts = _take_process_cpu_snapshot()\n\n if _last_cpu_measure is None:\n _last_cpu_measure = (pts, sts)\n return None\n\n old_pts, old_sts = _last_cpu_measure\n\n percents = []\n for pid in pts:\n if pid not in old_pts:\n continue\n pt1 = old_pts[pid]\n pt2 = pts[pid]\n delta_proc = (pt2.user - pt1.user) + (pt2.system - pt1.system)\n delta_time = sts[pid] - old_sts[pid]\n\n try:\n overall_cpus_percent = (delta_proc / delta_time) * 100\n except ZeroDivisionError:\n percents.append(0.0)\n else:\n single_cpu_percent = overall_cpus_percent * num_cpus\n percents.append(single_cpu_percent)\n _last_cpu_measure = (pts, sts)\n return round(sum(percents), 1)\n\n\ndef disk_usage(d):\n return psutil.disk_usage(d)\n\n\ndef iowait():\n cpu_percent = psutil.cpu_times_percent()\n try:\n return cpu_percent.iowait\n except AttributeError:\n return None\n\n\n_last_disk_io_meta = None\n_win_diskperf_called = False\n\n\ndef disk_io_usage():\n global _last_disk_io_meta, _win_diskperf_called\n\n # Needed by psutil.disk_io_counters() under newer version of Windows.\n # diskperf -y need to be called or no disk information can be found.\n if sys.platform == 'win32' and not _win_diskperf_called: # pragma: no cover\n CREATE_NO_WINDOW = 0x08000000\n try:\n proc = subprocess.Popen(['diskperf', '-y'], shell=False,\n creationflags=CREATE_NO_WINDOW) # nosec\n proc.wait()\n except (subprocess.CalledProcessError, OSError):\n pass\n _win_diskperf_called = True\n\n disk_counters = psutil.disk_io_counters()\n tst = time.time()\n\n read_bytes = disk_counters.read_bytes\n write_bytes = disk_counters.write_bytes\n if _last_disk_io_meta is None:\n _last_disk_io_meta = (read_bytes, write_bytes, tst)\n return None\n\n last_read_bytes, last_write_bytes, last_time = _last_disk_io_meta\n delta_time = tst - last_time\n read_speed = (read_bytes - last_read_bytes) / delta_time\n write_speed = (write_bytes - last_write_bytes) / delta_time\n\n _last_disk_io_meta = (read_bytes, write_bytes, tst)\n return read_speed, write_speed\n\n\n_last_net_io_meta = None\n\n\ndef net_io_usage():\n global _last_net_io_meta\n\n net_counters = psutil.net_io_counters()\n tst = time.time()\n\n send_bytes = net_counters.bytes_sent\n recv_bytes = net_counters.bytes_recv\n if _last_net_io_meta is None:\n _last_net_io_meta = (send_bytes, recv_bytes, tst)\n return None\n\n last_send_bytes, last_recv_bytes, last_time = _last_net_io_meta\n delta_time = tst - last_time\n recv_speed = (recv_bytes - last_recv_bytes) / delta_time\n send_speed = (send_bytes - last_send_bytes) / delta_time\n\n _last_net_io_meta = (send_bytes, recv_bytes, tst)\n return recv_speed, send_speed\n\n\n_cuda_info = namedtuple('cuda_info', 'driver_version cuda_version products gpu_count')\n_cuda_card_stat = namedtuple('cuda_card_stat', 'product_name gpu_usage temperature fb_mem_info')\n\n\ndef cuda_info(): # pragma: no cover\n driver_info = nvutils.get_driver_info()\n if not driver_info:\n return\n gpu_count = nvutils.get_device_count()\n return _cuda_info(\n driver_version=driver_info.driver_version,\n cuda_version=driver_info.cuda_version,\n products=[nvutils.get_device_info(idx).name for idx in range(gpu_count)],\n gpu_count=gpu_count,\n )\n\n\ndef cuda_card_stats(): # pragma: no cover\n infos = []\n device_count = nvutils.get_device_count()\n if not device_count:\n return\n for device_idx in range(device_count):\n device_info = nvutils.get_device_info(device_idx)\n device_status = nvutils.get_device_status(device_idx)\n\n infos.append(_cuda_card_stat(\n product_name=device_info.name,\n gpu_usage=device_status.gpu_util,\n temperature=device_status.temperature,\n fb_mem_info=_virt_memory_stat(\n total=device_status.fb_total_mem, used=device_status.fb_used_mem,\n free=device_status.fb_free_mem, available=device_status.fb_free_mem,\n percent=device_status.mem_util,\n )\n ))\n return infos\n", "path": "mars/resource.py"}], "after_files": [{"content": "# -*- coding: utf-8 -*-\n# Copyright 1999-2018 Alibaba Group Holding Ltd.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nimport os\nimport subprocess # nosec\nimport sys\nimport time\nfrom collections import namedtuple\n\nimport psutil\n\nfrom .lib import nvutils\n\nCGROUP_MEM_STAT_FILE = '/sys/fs/cgroup/memory/memory.stat'\n\n_proc = psutil.Process()\n_timer = getattr(time, 'monotonic', time.time)\n\n_cpu_use_process_stat = bool(int(os.environ.get('MARS_CPU_USE_PROCESS_STAT', '0').strip('\"')))\n_mem_use_process_stat = bool(int(os.environ.get('MARS_MEM_USE_PROCESS_STAT', '0').strip('\"')))\n_mem_use_cgroup_stat = bool(int(os.environ.get('MARS_MEM_USE_CGROUP_STAT', '0').strip('\"')))\n\nif 'MARS_USE_PROCESS_STAT' in os.environ:\n _cpu_use_process_stat = _mem_use_process_stat = \\\n bool(int(os.environ['MARS_USE_PROCESS_STAT'].strip('\"')))\n\nif _cpu_use_process_stat and 'MARS_CPU_TOTAL' in os.environ:\n _cpu_total = int(os.environ['MARS_CPU_TOTAL'].strip('\"'))\nelse:\n _cpu_total = psutil.cpu_count(logical=True)\n\nif _mem_use_process_stat and 'MARS_MEMORY_TOTAL' in os.environ:\n _mem_total = int(os.environ['MARS_MEMORY_TOTAL'].strip('\"'))\nelse:\n _mem_total = None\n\n_virt_memory_stat = namedtuple('virtual_memory', 'total available percent used free')\n\n_shm_path = [pt.mountpoint for pt in psutil.disk_partitions(all=True)\n if pt.mountpoint in ('/tmp', '/dev/shm') and pt.fstype == 'tmpfs']\nif not _shm_path:\n _shm_path = None\nelse:\n _shm_path = _shm_path[0]\n\n\ndef _read_cgroup_stat_file():\n with open(CGROUP_MEM_STAT_FILE, 'r') as cg_file:\n contents = cg_file.read()\n kvs = dict()\n for l in contents.splitlines():\n parts = l.split(' ')\n if len(parts) == 2:\n kvs[parts[0]] = int(parts[1])\n return kvs\n\n\ndef virtual_memory():\n sys_mem = psutil.virtual_memory()\n if _mem_use_cgroup_stat:\n # see section 5.5 in https://www.kernel.org/doc/Documentation/cgroup-v1/memory.txt\n cgroup_mem_info = _read_cgroup_stat_file()\n total = cgroup_mem_info['hierarchical_memory_limit']\n used = cgroup_mem_info['cache'] + cgroup_mem_info['rss'] + cgroup_mem_info.get('swap', 0)\n available = free = total - used\n percent = 100.0 * (total - available) / total\n return _virt_memory_stat(total, available, percent, used, free)\n elif not _mem_use_process_stat:\n total = sys_mem.total\n used = sys_mem.used + getattr(sys_mem, 'shared', 0)\n available = sys_mem.available\n free = sys_mem.free\n percent = 100.0 * (total - available) / total\n return _virt_memory_stat(total, available, percent, used, free)\n else:\n used = 0\n for p in psutil.process_iter():\n try:\n used += p.memory_info().rss\n except (psutil.NoSuchProcess, psutil.AccessDenied):\n pass\n\n if _shm_path:\n shm_stats = psutil.disk_usage(_shm_path)\n used += shm_stats.used\n\n total = min(_mem_total or sys_mem.total, sys_mem.total)\n # TODO sys_mem.available does not work in container\n # available = min(sys_mem.available, total - used)\n available = total - used\n free = min(sys_mem.free, total - used)\n percent = 100.0 * (total - available) / total\n return _virt_memory_stat(total, available, percent, used, free)\n\n\ndef cpu_count():\n return _cpu_total\n\n\n_last_cpu_measure = None\n\n\ndef _take_process_cpu_snapshot():\n num_cpus = cpu_count() or 1\n\n def timer():\n return _timer() * num_cpus\n\n processes = [p for p in psutil.process_iter() if p.pid != _proc.pid]\n\n pts = dict()\n sts = dict()\n for p in processes:\n try:\n pts[p.pid] = p.cpu_times()\n sts[p.pid] = timer()\n except (psutil.NoSuchProcess, psutil.AccessDenied):\n pass\n\n pts[_proc.pid] = _proc.cpu_times()\n sts[_proc.pid] = timer()\n return pts, sts\n\n\ndef cpu_percent():\n global _last_cpu_measure\n if not _cpu_use_process_stat:\n return sum(psutil.cpu_percent(percpu=True))\n\n num_cpus = cpu_count() or 1\n pts, sts = _take_process_cpu_snapshot()\n\n if _last_cpu_measure is None:\n _last_cpu_measure = (pts, sts)\n return None\n\n old_pts, old_sts = _last_cpu_measure\n\n percents = []\n for pid in pts:\n if pid not in old_pts:\n continue\n pt1 = old_pts[pid]\n pt2 = pts[pid]\n delta_proc = (pt2.user - pt1.user) + (pt2.system - pt1.system)\n delta_time = sts[pid] - old_sts[pid]\n\n try:\n overall_cpus_percent = (delta_proc / delta_time) * 100\n except ZeroDivisionError:\n percents.append(0.0)\n else:\n single_cpu_percent = overall_cpus_percent * num_cpus\n percents.append(single_cpu_percent)\n _last_cpu_measure = (pts, sts)\n return round(sum(percents), 1)\n\n\ndef disk_usage(d):\n return psutil.disk_usage(d)\n\n\ndef iowait():\n cpu_percent = psutil.cpu_times_percent()\n try:\n return cpu_percent.iowait\n except AttributeError:\n return None\n\n\n_last_disk_io_meta = None\n_win_diskperf_called = False\n\n\ndef disk_io_usage():\n global _last_disk_io_meta, _win_diskperf_called\n\n # Needed by psutil.disk_io_counters() under newer version of Windows.\n # diskperf -y need to be called or no disk information can be found.\n if sys.platform == 'win32' and not _win_diskperf_called: # pragma: no cover\n CREATE_NO_WINDOW = 0x08000000\n try:\n proc = subprocess.Popen(['diskperf', '-y'], shell=False,\n creationflags=CREATE_NO_WINDOW) # nosec\n proc.wait()\n except (subprocess.CalledProcessError, OSError):\n pass\n _win_diskperf_called = True\n\n disk_counters = psutil.disk_io_counters()\n tst = time.time()\n\n read_bytes = disk_counters.read_bytes\n write_bytes = disk_counters.write_bytes\n if _last_disk_io_meta is None:\n _last_disk_io_meta = (read_bytes, write_bytes, tst)\n return None\n\n last_read_bytes, last_write_bytes, last_time = _last_disk_io_meta\n delta_time = tst - last_time\n read_speed = (read_bytes - last_read_bytes) / delta_time\n write_speed = (write_bytes - last_write_bytes) / delta_time\n\n _last_disk_io_meta = (read_bytes, write_bytes, tst)\n return read_speed, write_speed\n\n\n_last_net_io_meta = None\n\n\ndef net_io_usage():\n global _last_net_io_meta\n\n net_counters = psutil.net_io_counters()\n tst = time.time()\n\n send_bytes = net_counters.bytes_sent\n recv_bytes = net_counters.bytes_recv\n if _last_net_io_meta is None:\n _last_net_io_meta = (send_bytes, recv_bytes, tst)\n return None\n\n last_send_bytes, last_recv_bytes, last_time = _last_net_io_meta\n delta_time = tst - last_time\n recv_speed = (recv_bytes - last_recv_bytes) / delta_time\n send_speed = (send_bytes - last_send_bytes) / delta_time\n\n _last_net_io_meta = (send_bytes, recv_bytes, tst)\n return recv_speed, send_speed\n\n\n_cuda_info = namedtuple('cuda_info', 'driver_version cuda_version products gpu_count')\n_cuda_card_stat = namedtuple('cuda_card_stat', 'product_name gpu_usage temperature fb_mem_info')\n\n\ndef cuda_info(): # pragma: no cover\n driver_info = nvutils.get_driver_info()\n if not driver_info:\n return\n gpu_count = nvutils.get_device_count()\n return _cuda_info(\n driver_version=driver_info.driver_version,\n cuda_version=driver_info.cuda_version,\n products=[nvutils.get_device_info(idx).name for idx in range(gpu_count)],\n gpu_count=gpu_count,\n )\n\n\ndef cuda_card_stats(): # pragma: no cover\n infos = []\n device_count = nvutils.get_device_count()\n if not device_count:\n return\n for device_idx in range(device_count):\n device_info = nvutils.get_device_info(device_idx)\n device_status = nvutils.get_device_status(device_idx)\n\n infos.append(_cuda_card_stat(\n product_name=device_info.name,\n gpu_usage=device_status.gpu_util,\n temperature=device_status.temperature,\n fb_mem_info=_virt_memory_stat(\n total=device_status.fb_total_mem, used=device_status.fb_used_mem,\n free=device_status.fb_free_mem, available=device_status.fb_free_mem,\n percent=device_status.mem_util,\n )\n ))\n return infos\n", "path": "mars/resource.py"}]} | 2,987 | 646 |
gh_patches_debug_27914 | rasdani/github-patches | git_diff | localstack__localstack-1460 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Kinesis putRecords error
It seems that putRecords causes the following error:
```
Starting mock Kinesis (http port 4568)...
Starting mock S3 (http port 4572)...
Starting mock Firehose service (http port 4573)...
Starting mock Lambda service (http port 4574)...
Listening at http://:::4565
* Running on http://0.0.0.0:4563/ (Press CTRL+C to quit)
127.0.0.1 - - [08/May/2018 13:52:25] "GET / HTTP/1.1" 200 -
Ready.
r127.0.0.1 - - [08/May/2018 13:56:43] "PUT /prd1541-qa1-vf-sms-send-stream-archive HTTP/1.1" 200 -
127.0.0.1 - - [08/May/2018 13:56:43] "HEAD /prd1541-qa1-vf-sms-send-stream-archive HTTP/1.1" 200 -
2018-05-08T13:56:43:ERROR:localstack.services.generic_proxy: Error forwarding request: 'Records' Traceback (most recent call last):
File "/opt/code/localstack/localstack/services/generic_proxy.py", line 215, in forward
updated_response = self.proxy.update_listener.return_response(**kwargs)
File "/opt/code/localstack/localstack/services/kinesis/kinesis_listener.py", line 49, in return_response
response_records = response_body['Records']
KeyError: 'Records'
2018-05-08T13:56:43:ERROR:localstack.services.generic_proxy: Error forwarding request: 'Records' Traceback (most recent call last):
File "/opt/code/localstack/localstack/services/generic_proxy.py", line 215, in forward
updated_response = self.proxy.update_listener.return_response(**kwargs)
File "/opt/code/localstack/localstack/services/kinesis/kinesis_listener.py", line 49, in return_response
response_records = response_body['Records']
KeyError: 'Records'
2018-05-08T13:56:43:ERROR:localstack.services.generic_proxy: Error forwarding request: 'Records' Traceback (most recent call last):
File "/opt/code/localstack/localstack/services/generic_proxy.py", line 215, in forward
updated_response = self.proxy.update_listener.return_response(**kwargs)
File "/opt/code/localstack/localstack/services/kinesis/kinesis_listener.py", line 49, in return_response
response_records = response_body['Records']
KeyError: 'Records'
```
I'm using the latest localstack image and my `Kinesis.putRecords` call goes through the Javascript AWS SDK:
```
...
const result = await client.putRecords({
StreamName,
Records: records.map(data => ({
Data: new Buffer(JSON.stringify(data)),
PartitionKey: 'testPartition'
}))
}).promise();
...
```
It seems like there's a mismatch int the expected format in `kinesis_listener.py`.
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `localstack/services/kinesis/kinesis_listener.py`
Content:
```
1 import json
2 import random
3 from requests.models import Response
4 from localstack import config
5 from localstack.utils.common import to_str
6 from localstack.utils.analytics import event_publisher
7 from localstack.services.awslambda import lambda_api
8 from localstack.services.generic_proxy import ProxyListener
9
10 # action headers
11 ACTION_PREFIX = 'Kinesis_20131202'
12 ACTION_PUT_RECORD = '%s.PutRecord' % ACTION_PREFIX
13 ACTION_PUT_RECORDS = '%s.PutRecords' % ACTION_PREFIX
14 ACTION_CREATE_STREAM = '%s.CreateStream' % ACTION_PREFIX
15 ACTION_DELETE_STREAM = '%s.DeleteStream' % ACTION_PREFIX
16 ACTION_UPDATE_SHARD_COUNT = '%s.UpdateShardCount' % ACTION_PREFIX
17
18
19 class ProxyListenerKinesis(ProxyListener):
20
21 def forward_request(self, method, path, data, headers):
22 data = json.loads(to_str(data))
23 action = headers.get('X-Amz-Target')
24
25 if action == '%s.DescribeStreamSummary' % ACTION_PREFIX:
26 stream_arn = data.get('StreamARN') or data['StreamName']
27 # TODO fix values below
28 result = {
29 'StreamDescriptionSummary': {
30 'ConsumerCount': 0,
31 'EnhancedMonitoring': [],
32 'KeyId': 'string',
33 'OpenShardCount': 0,
34 'RetentionPeriodHours': 1,
35 'StreamARN': stream_arn,
36 # 'StreamCreationTimestamp': number,
37 'StreamName': data['StreamName'],
38 'StreamStatus': 'ACTIVE'
39 }
40 }
41 return result
42 if action == '%s.DescribeStreamConsumer' % ACTION_PREFIX:
43 consumer_arn = data.get('ConsumerARN') or data['ConsumerName']
44 consumer_name = data.get('ConsumerName') or data['ConsumerARN']
45 result = {
46 'ConsumerDescription': {
47 'ConsumerARN': consumer_arn,
48 # 'ConsumerCreationTimestamp': number,
49 'ConsumerName': consumer_name,
50 'ConsumerStatus': 'ACTIVE',
51 'StreamARN': data.get('StreamARN')
52 }
53 }
54 return result
55
56 if random.random() < config.KINESIS_ERROR_PROBABILITY:
57 action = headers.get('X-Amz-Target')
58 if action in [ACTION_PUT_RECORD, ACTION_PUT_RECORDS]:
59 return kinesis_error_response(data, action)
60 return True
61
62 def return_response(self, method, path, data, headers, response):
63 action = headers.get('X-Amz-Target')
64 data = json.loads(to_str(data))
65
66 records = []
67 if action in (ACTION_CREATE_STREAM, ACTION_DELETE_STREAM):
68 event_type = (event_publisher.EVENT_KINESIS_CREATE_STREAM if action == ACTION_CREATE_STREAM
69 else event_publisher.EVENT_KINESIS_DELETE_STREAM)
70 payload = {'n': event_publisher.get_hash(data.get('StreamName'))}
71 if action == ACTION_CREATE_STREAM:
72 payload['s'] = data.get('ShardCount')
73 event_publisher.fire_event(event_type, payload=payload)
74 elif action == ACTION_PUT_RECORD:
75 response_body = json.loads(to_str(response.content))
76 event_record = {
77 'data': data['Data'],
78 'partitionKey': data['PartitionKey'],
79 'sequenceNumber': response_body.get('SequenceNumber')
80 }
81 event_records = [event_record]
82 stream_name = data['StreamName']
83 lambda_api.process_kinesis_records(event_records, stream_name)
84 elif action == ACTION_PUT_RECORDS:
85 event_records = []
86 response_body = json.loads(to_str(response.content))
87 response_records = response_body['Records']
88 records = data['Records']
89 for index in range(0, len(records)):
90 record = records[index]
91 event_record = {
92 'data': record['Data'],
93 'partitionKey': record['PartitionKey'],
94 'sequenceNumber': response_records[index].get('SequenceNumber')
95 }
96 event_records.append(event_record)
97 stream_name = data['StreamName']
98 lambda_api.process_kinesis_records(event_records, stream_name)
99 elif action == ACTION_UPDATE_SHARD_COUNT:
100 # Currently kinesalite, which backs the Kinesis implementation for localstack, does
101 # not support UpdateShardCount:
102 # https://github.com/mhart/kinesalite/issues/61
103 #
104 # [Terraform](https://www.terraform.io) makes the call to UpdateShardCount when it
105 # applies Kinesis resources. A Terraform run fails when this is not present.
106 #
107 # The code that follows just returns a successful response, bypassing the 400
108 # response that kinesalite returns.
109 #
110 response = Response()
111 response.status_code = 200
112 content = {
113 'CurrentShardCount': 1,
114 'StreamName': data['StreamName'],
115 'TargetShardCount': data['TargetShardCount']
116 }
117 response.encoding = 'UTF-8'
118 response._content = json.dumps(content)
119 return response
120
121
122 # instantiate listener
123 UPDATE_KINESIS = ProxyListenerKinesis()
124
125
126 def kinesis_error_response(data, action):
127 error_response = Response()
128
129 if action == ACTION_PUT_RECORD:
130 error_response.status_code = 400
131 content = {
132 'ErrorCode': 'ProvisionedThroughputExceededException',
133 'ErrorMessage': 'Rate exceeded for shard X in stream Y under account Z.'
134 }
135 else:
136 error_response.status_code = 200
137 content = {'FailedRecordCount': 1, 'Records': []}
138 for record in data.get('Records', []):
139 content['Records'].append({
140 'ErrorCode': 'ProvisionedThroughputExceededException',
141 'ErrorMessage': 'Rate exceeded for shard X in stream Y under account Z.'
142 })
143
144 error_response._content = json.dumps(content)
145 return error_response
146
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/localstack/services/kinesis/kinesis_listener.py b/localstack/services/kinesis/kinesis_listener.py
--- a/localstack/services/kinesis/kinesis_listener.py
+++ b/localstack/services/kinesis/kinesis_listener.py
@@ -84,18 +84,19 @@
elif action == ACTION_PUT_RECORDS:
event_records = []
response_body = json.loads(to_str(response.content))
- response_records = response_body['Records']
- records = data['Records']
- for index in range(0, len(records)):
- record = records[index]
- event_record = {
- 'data': record['Data'],
- 'partitionKey': record['PartitionKey'],
- 'sequenceNumber': response_records[index].get('SequenceNumber')
- }
- event_records.append(event_record)
- stream_name = data['StreamName']
- lambda_api.process_kinesis_records(event_records, stream_name)
+ if 'Records' in response_body:
+ response_records = response_body['Records']
+ records = data['Records']
+ for index in range(0, len(records)):
+ record = records[index]
+ event_record = {
+ 'data': record['Data'],
+ 'partitionKey': record['PartitionKey'],
+ 'sequenceNumber': response_records[index].get('SequenceNumber')
+ }
+ event_records.append(event_record)
+ stream_name = data['StreamName']
+ lambda_api.process_kinesis_records(event_records, stream_name)
elif action == ACTION_UPDATE_SHARD_COUNT:
# Currently kinesalite, which backs the Kinesis implementation for localstack, does
# not support UpdateShardCount:
| {"golden_diff": "diff --git a/localstack/services/kinesis/kinesis_listener.py b/localstack/services/kinesis/kinesis_listener.py\n--- a/localstack/services/kinesis/kinesis_listener.py\n+++ b/localstack/services/kinesis/kinesis_listener.py\n@@ -84,18 +84,19 @@\n elif action == ACTION_PUT_RECORDS:\n event_records = []\n response_body = json.loads(to_str(response.content))\n- response_records = response_body['Records']\n- records = data['Records']\n- for index in range(0, len(records)):\n- record = records[index]\n- event_record = {\n- 'data': record['Data'],\n- 'partitionKey': record['PartitionKey'],\n- 'sequenceNumber': response_records[index].get('SequenceNumber')\n- }\n- event_records.append(event_record)\n- stream_name = data['StreamName']\n- lambda_api.process_kinesis_records(event_records, stream_name)\n+ if 'Records' in response_body:\n+ response_records = response_body['Records']\n+ records = data['Records']\n+ for index in range(0, len(records)):\n+ record = records[index]\n+ event_record = {\n+ 'data': record['Data'],\n+ 'partitionKey': record['PartitionKey'],\n+ 'sequenceNumber': response_records[index].get('SequenceNumber')\n+ }\n+ event_records.append(event_record)\n+ stream_name = data['StreamName']\n+ lambda_api.process_kinesis_records(event_records, stream_name)\n elif action == ACTION_UPDATE_SHARD_COUNT:\n # Currently kinesalite, which backs the Kinesis implementation for localstack, does\n # not support UpdateShardCount:\n", "issue": "Kinesis putRecords error\nIt seems that putRecords causes the following error:\r\n\r\n```\r\nStarting mock Kinesis (http port 4568)...\r\nStarting mock S3 (http port 4572)...\r\nStarting mock Firehose service (http port 4573)...\r\nStarting mock Lambda service (http port 4574)...\r\nListening at http://:::4565\r\n* Running on http://0.0.0.0:4563/ (Press CTRL+C to quit)\r\n127.0.0.1 - - [08/May/2018 13:52:25] \"GET / HTTP/1.1\" 200 -\r\nReady.\r\nr127.0.0.1 - - [08/May/2018 13:56:43] \"PUT /prd1541-qa1-vf-sms-send-stream-archive HTTP/1.1\" 200 -\r\n127.0.0.1 - - [08/May/2018 13:56:43] \"HEAD /prd1541-qa1-vf-sms-send-stream-archive HTTP/1.1\" 200 -\r\n2018-05-08T13:56:43:ERROR:localstack.services.generic_proxy: Error forwarding request: 'Records' Traceback (most recent call last):\r\n File \"/opt/code/localstack/localstack/services/generic_proxy.py\", line 215, in forward\r\n updated_response = self.proxy.update_listener.return_response(**kwargs)\r\n File \"/opt/code/localstack/localstack/services/kinesis/kinesis_listener.py\", line 49, in return_response\r\n response_records = response_body['Records']\r\nKeyError: 'Records'\r\n\r\n2018-05-08T13:56:43:ERROR:localstack.services.generic_proxy: Error forwarding request: 'Records' Traceback (most recent call last):\r\n File \"/opt/code/localstack/localstack/services/generic_proxy.py\", line 215, in forward\r\n updated_response = self.proxy.update_listener.return_response(**kwargs)\r\n File \"/opt/code/localstack/localstack/services/kinesis/kinesis_listener.py\", line 49, in return_response\r\n response_records = response_body['Records']\r\nKeyError: 'Records'\r\n\r\n2018-05-08T13:56:43:ERROR:localstack.services.generic_proxy: Error forwarding request: 'Records' Traceback (most recent call last):\r\n File \"/opt/code/localstack/localstack/services/generic_proxy.py\", line 215, in forward\r\n updated_response = self.proxy.update_listener.return_response(**kwargs)\r\n File \"/opt/code/localstack/localstack/services/kinesis/kinesis_listener.py\", line 49, in return_response\r\n response_records = response_body['Records']\r\nKeyError: 'Records'\r\n```\r\n\r\nI'm using the latest localstack image and my `Kinesis.putRecords` call goes through the Javascript AWS SDK: \r\n\r\n```\r\n...\r\n const result = await client.putRecords({\r\n StreamName,\r\n Records: records.map(data => ({\r\n Data: new Buffer(JSON.stringify(data)),\r\n PartitionKey: 'testPartition'\r\n }))\r\n }).promise();\r\n...\r\n```\r\n\r\nIt seems like there's a mismatch int the expected format in `kinesis_listener.py`.\n", "before_files": [{"content": "import json\nimport random\nfrom requests.models import Response\nfrom localstack import config\nfrom localstack.utils.common import to_str\nfrom localstack.utils.analytics import event_publisher\nfrom localstack.services.awslambda import lambda_api\nfrom localstack.services.generic_proxy import ProxyListener\n\n# action headers\nACTION_PREFIX = 'Kinesis_20131202'\nACTION_PUT_RECORD = '%s.PutRecord' % ACTION_PREFIX\nACTION_PUT_RECORDS = '%s.PutRecords' % ACTION_PREFIX\nACTION_CREATE_STREAM = '%s.CreateStream' % ACTION_PREFIX\nACTION_DELETE_STREAM = '%s.DeleteStream' % ACTION_PREFIX\nACTION_UPDATE_SHARD_COUNT = '%s.UpdateShardCount' % ACTION_PREFIX\n\n\nclass ProxyListenerKinesis(ProxyListener):\n\n def forward_request(self, method, path, data, headers):\n data = json.loads(to_str(data))\n action = headers.get('X-Amz-Target')\n\n if action == '%s.DescribeStreamSummary' % ACTION_PREFIX:\n stream_arn = data.get('StreamARN') or data['StreamName']\n # TODO fix values below\n result = {\n 'StreamDescriptionSummary': {\n 'ConsumerCount': 0,\n 'EnhancedMonitoring': [],\n 'KeyId': 'string',\n 'OpenShardCount': 0,\n 'RetentionPeriodHours': 1,\n 'StreamARN': stream_arn,\n # 'StreamCreationTimestamp': number,\n 'StreamName': data['StreamName'],\n 'StreamStatus': 'ACTIVE'\n }\n }\n return result\n if action == '%s.DescribeStreamConsumer' % ACTION_PREFIX:\n consumer_arn = data.get('ConsumerARN') or data['ConsumerName']\n consumer_name = data.get('ConsumerName') or data['ConsumerARN']\n result = {\n 'ConsumerDescription': {\n 'ConsumerARN': consumer_arn,\n # 'ConsumerCreationTimestamp': number,\n 'ConsumerName': consumer_name,\n 'ConsumerStatus': 'ACTIVE',\n 'StreamARN': data.get('StreamARN')\n }\n }\n return result\n\n if random.random() < config.KINESIS_ERROR_PROBABILITY:\n action = headers.get('X-Amz-Target')\n if action in [ACTION_PUT_RECORD, ACTION_PUT_RECORDS]:\n return kinesis_error_response(data, action)\n return True\n\n def return_response(self, method, path, data, headers, response):\n action = headers.get('X-Amz-Target')\n data = json.loads(to_str(data))\n\n records = []\n if action in (ACTION_CREATE_STREAM, ACTION_DELETE_STREAM):\n event_type = (event_publisher.EVENT_KINESIS_CREATE_STREAM if action == ACTION_CREATE_STREAM\n else event_publisher.EVENT_KINESIS_DELETE_STREAM)\n payload = {'n': event_publisher.get_hash(data.get('StreamName'))}\n if action == ACTION_CREATE_STREAM:\n payload['s'] = data.get('ShardCount')\n event_publisher.fire_event(event_type, payload=payload)\n elif action == ACTION_PUT_RECORD:\n response_body = json.loads(to_str(response.content))\n event_record = {\n 'data': data['Data'],\n 'partitionKey': data['PartitionKey'],\n 'sequenceNumber': response_body.get('SequenceNumber')\n }\n event_records = [event_record]\n stream_name = data['StreamName']\n lambda_api.process_kinesis_records(event_records, stream_name)\n elif action == ACTION_PUT_RECORDS:\n event_records = []\n response_body = json.loads(to_str(response.content))\n response_records = response_body['Records']\n records = data['Records']\n for index in range(0, len(records)):\n record = records[index]\n event_record = {\n 'data': record['Data'],\n 'partitionKey': record['PartitionKey'],\n 'sequenceNumber': response_records[index].get('SequenceNumber')\n }\n event_records.append(event_record)\n stream_name = data['StreamName']\n lambda_api.process_kinesis_records(event_records, stream_name)\n elif action == ACTION_UPDATE_SHARD_COUNT:\n # Currently kinesalite, which backs the Kinesis implementation for localstack, does\n # not support UpdateShardCount:\n # https://github.com/mhart/kinesalite/issues/61\n #\n # [Terraform](https://www.terraform.io) makes the call to UpdateShardCount when it\n # applies Kinesis resources. A Terraform run fails when this is not present.\n #\n # The code that follows just returns a successful response, bypassing the 400\n # response that kinesalite returns.\n #\n response = Response()\n response.status_code = 200\n content = {\n 'CurrentShardCount': 1,\n 'StreamName': data['StreamName'],\n 'TargetShardCount': data['TargetShardCount']\n }\n response.encoding = 'UTF-8'\n response._content = json.dumps(content)\n return response\n\n\n# instantiate listener\nUPDATE_KINESIS = ProxyListenerKinesis()\n\n\ndef kinesis_error_response(data, action):\n error_response = Response()\n\n if action == ACTION_PUT_RECORD:\n error_response.status_code = 400\n content = {\n 'ErrorCode': 'ProvisionedThroughputExceededException',\n 'ErrorMessage': 'Rate exceeded for shard X in stream Y under account Z.'\n }\n else:\n error_response.status_code = 200\n content = {'FailedRecordCount': 1, 'Records': []}\n for record in data.get('Records', []):\n content['Records'].append({\n 'ErrorCode': 'ProvisionedThroughputExceededException',\n 'ErrorMessage': 'Rate exceeded for shard X in stream Y under account Z.'\n })\n\n error_response._content = json.dumps(content)\n return error_response\n", "path": "localstack/services/kinesis/kinesis_listener.py"}], "after_files": [{"content": "import json\nimport random\nfrom requests.models import Response\nfrom localstack import config\nfrom localstack.utils.common import to_str\nfrom localstack.utils.analytics import event_publisher\nfrom localstack.services.awslambda import lambda_api\nfrom localstack.services.generic_proxy import ProxyListener\n\n# action headers\nACTION_PREFIX = 'Kinesis_20131202'\nACTION_PUT_RECORD = '%s.PutRecord' % ACTION_PREFIX\nACTION_PUT_RECORDS = '%s.PutRecords' % ACTION_PREFIX\nACTION_CREATE_STREAM = '%s.CreateStream' % ACTION_PREFIX\nACTION_DELETE_STREAM = '%s.DeleteStream' % ACTION_PREFIX\nACTION_UPDATE_SHARD_COUNT = '%s.UpdateShardCount' % ACTION_PREFIX\n\n\nclass ProxyListenerKinesis(ProxyListener):\n\n def forward_request(self, method, path, data, headers):\n data = json.loads(to_str(data))\n action = headers.get('X-Amz-Target')\n\n if action == '%s.DescribeStreamSummary' % ACTION_PREFIX:\n stream_arn = data.get('StreamARN') or data['StreamName']\n # TODO fix values below\n result = {\n 'StreamDescriptionSummary': {\n 'ConsumerCount': 0,\n 'EnhancedMonitoring': [],\n 'KeyId': 'string',\n 'OpenShardCount': 0,\n 'RetentionPeriodHours': 1,\n 'StreamARN': stream_arn,\n # 'StreamCreationTimestamp': number,\n 'StreamName': data['StreamName'],\n 'StreamStatus': 'ACTIVE'\n }\n }\n return result\n if action == '%s.DescribeStreamConsumer' % ACTION_PREFIX:\n consumer_arn = data.get('ConsumerARN') or data['ConsumerName']\n consumer_name = data.get('ConsumerName') or data['ConsumerARN']\n result = {\n 'ConsumerDescription': {\n 'ConsumerARN': consumer_arn,\n # 'ConsumerCreationTimestamp': number,\n 'ConsumerName': consumer_name,\n 'ConsumerStatus': 'ACTIVE',\n 'StreamARN': data.get('StreamARN')\n }\n }\n return result\n\n if random.random() < config.KINESIS_ERROR_PROBABILITY:\n action = headers.get('X-Amz-Target')\n if action in [ACTION_PUT_RECORD, ACTION_PUT_RECORDS]:\n return kinesis_error_response(data, action)\n return True\n\n def return_response(self, method, path, data, headers, response):\n action = headers.get('X-Amz-Target')\n data = json.loads(to_str(data))\n\n records = []\n if action in (ACTION_CREATE_STREAM, ACTION_DELETE_STREAM):\n event_type = (event_publisher.EVENT_KINESIS_CREATE_STREAM if action == ACTION_CREATE_STREAM\n else event_publisher.EVENT_KINESIS_DELETE_STREAM)\n payload = {'n': event_publisher.get_hash(data.get('StreamName'))}\n if action == ACTION_CREATE_STREAM:\n payload['s'] = data.get('ShardCount')\n event_publisher.fire_event(event_type, payload=payload)\n elif action == ACTION_PUT_RECORD:\n response_body = json.loads(to_str(response.content))\n event_record = {\n 'data': data['Data'],\n 'partitionKey': data['PartitionKey'],\n 'sequenceNumber': response_body.get('SequenceNumber')\n }\n event_records = [event_record]\n stream_name = data['StreamName']\n lambda_api.process_kinesis_records(event_records, stream_name)\n elif action == ACTION_PUT_RECORDS:\n event_records = []\n response_body = json.loads(to_str(response.content))\n if 'Records' in response_body:\n response_records = response_body['Records']\n records = data['Records']\n for index in range(0, len(records)):\n record = records[index]\n event_record = {\n 'data': record['Data'],\n 'partitionKey': record['PartitionKey'],\n 'sequenceNumber': response_records[index].get('SequenceNumber')\n }\n event_records.append(event_record)\n stream_name = data['StreamName']\n lambda_api.process_kinesis_records(event_records, stream_name)\n elif action == ACTION_UPDATE_SHARD_COUNT:\n # Currently kinesalite, which backs the Kinesis implementation for localstack, does\n # not support UpdateShardCount:\n # https://github.com/mhart/kinesalite/issues/61\n #\n # [Terraform](https://www.terraform.io) makes the call to UpdateShardCount when it\n # applies Kinesis resources. A Terraform run fails when this is not present.\n #\n # The code that follows just returns a successful response, bypassing the 400\n # response that kinesalite returns.\n #\n response = Response()\n response.status_code = 200\n content = {\n 'CurrentShardCount': 1,\n 'StreamName': data['StreamName'],\n 'TargetShardCount': data['TargetShardCount']\n }\n response.encoding = 'UTF-8'\n response._content = json.dumps(content)\n return response\n\n\n# instantiate listener\nUPDATE_KINESIS = ProxyListenerKinesis()\n\n\ndef kinesis_error_response(data, action):\n error_response = Response()\n\n if action == ACTION_PUT_RECORD:\n error_response.status_code = 400\n content = {\n 'ErrorCode': 'ProvisionedThroughputExceededException',\n 'ErrorMessage': 'Rate exceeded for shard X in stream Y under account Z.'\n }\n else:\n error_response.status_code = 200\n content = {'FailedRecordCount': 1, 'Records': []}\n for record in data.get('Records', []):\n content['Records'].append({\n 'ErrorCode': 'ProvisionedThroughputExceededException',\n 'ErrorMessage': 'Rate exceeded for shard X in stream Y under account Z.'\n })\n\n error_response._content = json.dumps(content)\n return error_response\n", "path": "localstack/services/kinesis/kinesis_listener.py"}]} | 2,582 | 362 |
gh_patches_debug_25795 | rasdani/github-patches | git_diff | pytorch__vision-6644 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Cannot Download PCAM dataset for val split
### 🐛 Describe the bug
```
import torchvision.datasets as datasets
transform_train = transforms.Compose([
transforms.ColorJitter(),
transforms.RandomHorizontalFlip(),
transforms.RandomVerticalFlip(),
transforms.Resize(args.input_size),
transforms.ToTensor(),
transforms.Normalize([0.485, 0.456, 0.406], [0.229, 0.224, 0.225])
])
dataset_train = datasets.PCAM(
root = args.data_path,
split = 'val',
transform = transform_train,
download=True
)
```
The error is like:
`RuntimeError: The MD5 checksum of the download file /localscratch2/xuezhiyu/datasets/pcam/camelyonpatch_level_2_split_valid_x.h5.gz does not match the one on record.Please delete the file and try again. If the issue persists, please report this to torchvision at https://github.com/pytorch/vision/issues.
`
### Versions
Collecting environment information...
PyTorch version: 1.12.1+cu113
Is debug build: False
CUDA used to build PyTorch: 11.3
ROCM used to build PyTorch: N/A
OS: Ubuntu 20.04.3 LTS (x86_64)
GCC version: (Ubuntu 9.3.0-17ubuntu1~20.04) 9.3.0
Clang version: Could not collect
CMake version: version 3.16.3
Libc version: glibc-2.31
Python version: 3.8.8 (default, Apr 13 2021, 19:58:26) [GCC 7.3.0] (64-bit runtime)
Python platform: Linux-5.13.0-27-generic-x86_64-with-glibc2.10
Is CUDA available: True
CUDA runtime version: 11.1.105
GPU models and configuration:
GPU 0: NVIDIA RTX A5000
GPU 1: NVIDIA RTX A5000
GPU 2: NVIDIA RTX A5000
GPU 3: NVIDIA RTX A5000
Nvidia driver version: 470.57.02
cuDNN version: Probably one of the following:
/usr/lib/x86_64-linux-gnu/libcudnn.so.8.3.2
/usr/lib/x86_64-linux-gnu/libcudnn_adv_infer.so.8.3.2
/usr/lib/x86_64-linux-gnu/libcudnn_adv_train.so.8.3.2
/usr/lib/x86_64-linux-gnu/libcudnn_cnn_infer.so.8.3.2
/usr/lib/x86_64-linux-gnu/libcudnn_cnn_train.so.8.3.2
/usr/lib/x86_64-linux-gnu/libcudnn_ops_infer.so.8.3.2
/usr/lib/x86_64-linux-gnu/libcudnn_ops_train.so.8.3.2
/usr/local/cuda-10.1/targets/x86_64-linux/lib/libcudnn.so.8.0.5
/usr/local/cuda-10.1/targets/x86_64-linux/lib/libcudnn_adv_infer.so.8.0.5
/usr/local/cuda-10.1/targets/x86_64-linux/lib/libcudnn_adv_train.so.8.0.5
/usr/local/cuda-10.1/targets/x86_64-linux/lib/libcudnn_cnn_infer.so.8.0.5
/usr/local/cuda-10.1/targets/x86_64-linux/lib/libcudnn_cnn_train.so.8.0.5
/usr/local/cuda-10.1/targets/x86_64-linux/lib/libcudnn_ops_infer.so.8.0.5
/usr/local/cuda-10.1/targets/x86_64-linux/lib/libcudnn_ops_train.so.8.0.5
/usr/local/cuda-10.2/targets/x86_64-linux/lib/libcudnn.so.8.3.1
/usr/local/cuda-10.2/targets/x86_64-linux/lib/libcudnn_adv_infer.so.8.3.1
/usr/local/cuda-10.2/targets/x86_64-linux/lib/libcudnn_adv_train.so.8.3.1
/usr/local/cuda-10.2/targets/x86_64-linux/lib/libcudnn_cnn_infer.so.8.3.1
/usr/local/cuda-10.2/targets/x86_64-linux/lib/libcudnn_cnn_train.so.8.3.1
/usr/local/cuda-10.2/targets/x86_64-linux/lib/libcudnn_ops_infer.so.8.3.1
/usr/local/cuda-10.2/targets/x86_64-linux/lib/libcudnn_ops_train.so.8.3.1
/usr/local/cuda-11.4/targets/x86_64-linux/lib/libcudnn.so.8.2.4
/usr/local/cuda-11.4/targets/x86_64-linux/lib/libcudnn_adv_infer.so.8.2.4
/usr/local/cuda-11.4/targets/x86_64-linux/lib/libcudnn_adv_train.so.8.2.4
/usr/local/cuda-11.4/targets/x86_64-linux/lib/libcudnn_cnn_infer.so.8.2.4
/usr/local/cuda-11.4/targets/x86_64-linux/lib/libcudnn_cnn_train.so.8.2.4
/usr/local/cuda-11.4/targets/x86_64-linux/lib/libcudnn_ops_infer.so.8.2.4
/usr/local/cuda-11.4/targets/x86_64-linux/lib/libcudnn_ops_train.so.8.2.4
HIP runtime version: N/A
MIOpen runtime version: N/A
Is XNNPACK available: True
Versions of relevant libraries:
[pip3] mypy-extensions==0.4.3
[pip3] numpy==1.20.1
[pip3] numpydoc==1.1.0
[pip3] torch==1.12.1+cu113
[pip3] torchaudio==0.12.1+cu113
[pip3] torchmultimodal==0.1.0a0
[pip3] torchvision==0.13.1+cu113
[conda] blas 1.0 mkl
[conda] mkl 2021.2.0 h06a4308_296
[conda] mkl-service 2.3.0 py38h27cfd23_1
[conda] mkl_fft 1.3.0 py38h42c9631_2
[conda] mkl_random 1.2.1 py38ha9443f7_2
[conda] numpy 1.20.1 py38h93e21f0_0
[conda] numpy-base 1.20.1 py38h7d8b39e_0
[conda] numpydoc 1.1.0 pyhd3eb1b0_1
[conda] torch 1.12.1+cu113 pypi_0 pypi
[conda] torchaudio 0.12.1+cu113 pypi_0 pypi
[conda] torchmultimodal 0.1.0a0 dev_0 <develop>
[conda] torchvision 0.13.1+cu113 pypi_0 pypi
cc @pmeier
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `torchvision/datasets/pcam.py`
Content:
```
1 import pathlib
2 from typing import Any, Callable, Optional, Tuple
3
4 from PIL import Image
5
6 from .utils import _decompress, download_file_from_google_drive, verify_str_arg
7 from .vision import VisionDataset
8
9
10 class PCAM(VisionDataset):
11 """`PCAM Dataset <https://github.com/basveeling/pcam>`_.
12
13 The PatchCamelyon dataset is a binary classification dataset with 327,680
14 color images (96px x 96px), extracted from histopathologic scans of lymph node
15 sections. Each image is annotated with a binary label indicating presence of
16 metastatic tissue.
17
18 This dataset requires the ``h5py`` package which you can install with ``pip install h5py``.
19
20 Args:
21 root (string): Root directory of the dataset.
22 split (string, optional): The dataset split, supports ``"train"`` (default), ``"test"`` or ``"val"``.
23 transform (callable, optional): A function/transform that takes in a PIL image and returns a transformed
24 version. E.g, ``transforms.RandomCrop``.
25 target_transform (callable, optional): A function/transform that takes in the target and transforms it.
26 download (bool, optional): If True, downloads the dataset from the internet and puts it into ``root/pcam``. If
27 dataset is already downloaded, it is not downloaded again.
28 """
29
30 _FILES = {
31 "train": {
32 "images": (
33 "camelyonpatch_level_2_split_train_x.h5", # Data file name
34 "1Ka0XfEMiwgCYPdTI-vv6eUElOBnKFKQ2", # Google Drive ID
35 "1571f514728f59376b705fc836ff4b63", # md5 hash
36 ),
37 "targets": (
38 "camelyonpatch_level_2_split_train_y.h5",
39 "1269yhu3pZDP8UYFQs-NYs3FPwuK-nGSG",
40 "35c2d7259d906cfc8143347bb8e05be7",
41 ),
42 },
43 "test": {
44 "images": (
45 "camelyonpatch_level_2_split_test_x.h5",
46 "1qV65ZqZvWzuIVthK8eVDhIwrbnsJdbg_",
47 "d5b63470df7cfa627aeec8b9dc0c066e",
48 ),
49 "targets": (
50 "camelyonpatch_level_2_split_test_y.h5",
51 "17BHrSrwWKjYsOgTMmoqrIjDy6Fa2o_gP",
52 "2b85f58b927af9964a4c15b8f7e8f179",
53 ),
54 },
55 "val": {
56 "images": (
57 "camelyonpatch_level_2_split_valid_x.h5",
58 "1hgshYGWK8V-eGRy8LToWJJgDU_rXWVJ3",
59 "d8c2d60d490dbd479f8199bdfa0cf6ec",
60 ),
61 "targets": (
62 "camelyonpatch_level_2_split_valid_y.h5",
63 "1bH8ZRbhSVAhScTS0p9-ZzGnX91cHT3uO",
64 "60a7035772fbdb7f34eb86d4420cf66a",
65 ),
66 },
67 }
68
69 def __init__(
70 self,
71 root: str,
72 split: str = "train",
73 transform: Optional[Callable] = None,
74 target_transform: Optional[Callable] = None,
75 download: bool = False,
76 ):
77 try:
78 import h5py
79
80 self.h5py = h5py
81 except ImportError:
82 raise RuntimeError(
83 "h5py is not found. This dataset needs to have h5py installed: please run pip install h5py"
84 )
85
86 self._split = verify_str_arg(split, "split", ("train", "test", "val"))
87
88 super().__init__(root, transform=transform, target_transform=target_transform)
89 self._base_folder = pathlib.Path(self.root) / "pcam"
90
91 if download:
92 self._download()
93
94 if not self._check_exists():
95 raise RuntimeError("Dataset not found. You can use download=True to download it")
96
97 def __len__(self) -> int:
98 images_file = self._FILES[self._split]["images"][0]
99 with self.h5py.File(self._base_folder / images_file) as images_data:
100 return images_data["x"].shape[0]
101
102 def __getitem__(self, idx: int) -> Tuple[Any, Any]:
103 images_file = self._FILES[self._split]["images"][0]
104 with self.h5py.File(self._base_folder / images_file) as images_data:
105 image = Image.fromarray(images_data["x"][idx]).convert("RGB")
106
107 targets_file = self._FILES[self._split]["targets"][0]
108 with self.h5py.File(self._base_folder / targets_file) as targets_data:
109 target = int(targets_data["y"][idx, 0, 0, 0]) # shape is [num_images, 1, 1, 1]
110
111 if self.transform:
112 image = self.transform(image)
113 if self.target_transform:
114 target = self.target_transform(target)
115
116 return image, target
117
118 def _check_exists(self) -> bool:
119 images_file = self._FILES[self._split]["images"][0]
120 targets_file = self._FILES[self._split]["targets"][0]
121 return all(self._base_folder.joinpath(h5_file).exists() for h5_file in (images_file, targets_file))
122
123 def _download(self) -> None:
124 if self._check_exists():
125 return
126
127 for file_name, file_id, md5 in self._FILES[self._split].values():
128 archive_name = file_name + ".gz"
129 download_file_from_google_drive(file_id, str(self._base_folder), filename=archive_name, md5=md5)
130 _decompress(str(self._base_folder / archive_name))
131
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/torchvision/datasets/pcam.py b/torchvision/datasets/pcam.py
--- a/torchvision/datasets/pcam.py
+++ b/torchvision/datasets/pcam.py
@@ -44,24 +44,24 @@
"images": (
"camelyonpatch_level_2_split_test_x.h5",
"1qV65ZqZvWzuIVthK8eVDhIwrbnsJdbg_",
- "d5b63470df7cfa627aeec8b9dc0c066e",
+ "d8c2d60d490dbd479f8199bdfa0cf6ec",
),
"targets": (
"camelyonpatch_level_2_split_test_y.h5",
"17BHrSrwWKjYsOgTMmoqrIjDy6Fa2o_gP",
- "2b85f58b927af9964a4c15b8f7e8f179",
+ "60a7035772fbdb7f34eb86d4420cf66a",
),
},
"val": {
"images": (
"camelyonpatch_level_2_split_valid_x.h5",
"1hgshYGWK8V-eGRy8LToWJJgDU_rXWVJ3",
- "d8c2d60d490dbd479f8199bdfa0cf6ec",
+ "d5b63470df7cfa627aeec8b9dc0c066e",
),
"targets": (
"camelyonpatch_level_2_split_valid_y.h5",
"1bH8ZRbhSVAhScTS0p9-ZzGnX91cHT3uO",
- "60a7035772fbdb7f34eb86d4420cf66a",
+ "2b85f58b927af9964a4c15b8f7e8f179",
),
},
}
| {"golden_diff": "diff --git a/torchvision/datasets/pcam.py b/torchvision/datasets/pcam.py\n--- a/torchvision/datasets/pcam.py\n+++ b/torchvision/datasets/pcam.py\n@@ -44,24 +44,24 @@\n \"images\": (\n \"camelyonpatch_level_2_split_test_x.h5\",\n \"1qV65ZqZvWzuIVthK8eVDhIwrbnsJdbg_\",\n- \"d5b63470df7cfa627aeec8b9dc0c066e\",\n+ \"d8c2d60d490dbd479f8199bdfa0cf6ec\",\n ),\n \"targets\": (\n \"camelyonpatch_level_2_split_test_y.h5\",\n \"17BHrSrwWKjYsOgTMmoqrIjDy6Fa2o_gP\",\n- \"2b85f58b927af9964a4c15b8f7e8f179\",\n+ \"60a7035772fbdb7f34eb86d4420cf66a\",\n ),\n },\n \"val\": {\n \"images\": (\n \"camelyonpatch_level_2_split_valid_x.h5\",\n \"1hgshYGWK8V-eGRy8LToWJJgDU_rXWVJ3\",\n- \"d8c2d60d490dbd479f8199bdfa0cf6ec\",\n+ \"d5b63470df7cfa627aeec8b9dc0c066e\",\n ),\n \"targets\": (\n \"camelyonpatch_level_2_split_valid_y.h5\",\n \"1bH8ZRbhSVAhScTS0p9-ZzGnX91cHT3uO\",\n- \"60a7035772fbdb7f34eb86d4420cf66a\",\n+ \"2b85f58b927af9964a4c15b8f7e8f179\",\n ),\n },\n }\n", "issue": "Cannot Download PCAM dataset for val split\n### \ud83d\udc1b Describe the bug\r\n\r\n\r\n```\r\n\r\nimport torchvision.datasets as datasets\r\ntransform_train = transforms.Compose([\r\n transforms.ColorJitter(),\r\n transforms.RandomHorizontalFlip(),\r\n transforms.RandomVerticalFlip(),\r\n transforms.Resize(args.input_size),\r\n transforms.ToTensor(),\r\n transforms.Normalize([0.485, 0.456, 0.406], [0.229, 0.224, 0.225]) \r\n ]) \r\n\r\ndataset_train = datasets.PCAM(\r\n root = args.data_path,\r\n split = 'val',\r\n transform = transform_train,\r\n download=True\r\n)\r\n```\r\n\r\nThe error is like:\r\n\r\n\r\n`RuntimeError: The MD5 checksum of the download file /localscratch2/xuezhiyu/datasets/pcam/camelyonpatch_level_2_split_valid_x.h5.gz does not match the one on record.Please delete the file and try again. If the issue persists, please report this to torchvision at https://github.com/pytorch/vision/issues.\r\n`\r\n### Versions\r\n\r\nCollecting environment information...\r\nPyTorch version: 1.12.1+cu113\r\nIs debug build: False\r\nCUDA used to build PyTorch: 11.3\r\nROCM used to build PyTorch: N/A\r\n\r\nOS: Ubuntu 20.04.3 LTS (x86_64)\r\nGCC version: (Ubuntu 9.3.0-17ubuntu1~20.04) 9.3.0\r\nClang version: Could not collect\r\nCMake version: version 3.16.3\r\nLibc version: glibc-2.31\r\n\r\nPython version: 3.8.8 (default, Apr 13 2021, 19:58:26) [GCC 7.3.0] (64-bit runtime)\r\nPython platform: Linux-5.13.0-27-generic-x86_64-with-glibc2.10\r\nIs CUDA available: True\r\nCUDA runtime version: 11.1.105\r\nGPU models and configuration: \r\nGPU 0: NVIDIA RTX A5000\r\nGPU 1: NVIDIA RTX A5000\r\nGPU 2: NVIDIA RTX A5000\r\nGPU 3: NVIDIA RTX A5000\r\n\r\nNvidia driver version: 470.57.02\r\ncuDNN version: Probably one of the following:\r\n/usr/lib/x86_64-linux-gnu/libcudnn.so.8.3.2\r\n/usr/lib/x86_64-linux-gnu/libcudnn_adv_infer.so.8.3.2\r\n/usr/lib/x86_64-linux-gnu/libcudnn_adv_train.so.8.3.2\r\n/usr/lib/x86_64-linux-gnu/libcudnn_cnn_infer.so.8.3.2\r\n/usr/lib/x86_64-linux-gnu/libcudnn_cnn_train.so.8.3.2\r\n/usr/lib/x86_64-linux-gnu/libcudnn_ops_infer.so.8.3.2\r\n/usr/lib/x86_64-linux-gnu/libcudnn_ops_train.so.8.3.2\r\n/usr/local/cuda-10.1/targets/x86_64-linux/lib/libcudnn.so.8.0.5\r\n/usr/local/cuda-10.1/targets/x86_64-linux/lib/libcudnn_adv_infer.so.8.0.5\r\n/usr/local/cuda-10.1/targets/x86_64-linux/lib/libcudnn_adv_train.so.8.0.5\r\n/usr/local/cuda-10.1/targets/x86_64-linux/lib/libcudnn_cnn_infer.so.8.0.5\r\n/usr/local/cuda-10.1/targets/x86_64-linux/lib/libcudnn_cnn_train.so.8.0.5\r\n/usr/local/cuda-10.1/targets/x86_64-linux/lib/libcudnn_ops_infer.so.8.0.5\r\n/usr/local/cuda-10.1/targets/x86_64-linux/lib/libcudnn_ops_train.so.8.0.5\r\n/usr/local/cuda-10.2/targets/x86_64-linux/lib/libcudnn.so.8.3.1\r\n/usr/local/cuda-10.2/targets/x86_64-linux/lib/libcudnn_adv_infer.so.8.3.1\r\n/usr/local/cuda-10.2/targets/x86_64-linux/lib/libcudnn_adv_train.so.8.3.1\r\n/usr/local/cuda-10.2/targets/x86_64-linux/lib/libcudnn_cnn_infer.so.8.3.1\r\n/usr/local/cuda-10.2/targets/x86_64-linux/lib/libcudnn_cnn_train.so.8.3.1\r\n/usr/local/cuda-10.2/targets/x86_64-linux/lib/libcudnn_ops_infer.so.8.3.1\r\n/usr/local/cuda-10.2/targets/x86_64-linux/lib/libcudnn_ops_train.so.8.3.1\r\n/usr/local/cuda-11.4/targets/x86_64-linux/lib/libcudnn.so.8.2.4\r\n/usr/local/cuda-11.4/targets/x86_64-linux/lib/libcudnn_adv_infer.so.8.2.4\r\n/usr/local/cuda-11.4/targets/x86_64-linux/lib/libcudnn_adv_train.so.8.2.4\r\n/usr/local/cuda-11.4/targets/x86_64-linux/lib/libcudnn_cnn_infer.so.8.2.4\r\n/usr/local/cuda-11.4/targets/x86_64-linux/lib/libcudnn_cnn_train.so.8.2.4\r\n/usr/local/cuda-11.4/targets/x86_64-linux/lib/libcudnn_ops_infer.so.8.2.4\r\n/usr/local/cuda-11.4/targets/x86_64-linux/lib/libcudnn_ops_train.so.8.2.4\r\nHIP runtime version: N/A\r\nMIOpen runtime version: N/A\r\nIs XNNPACK available: True\r\n\r\nVersions of relevant libraries:\r\n[pip3] mypy-extensions==0.4.3\r\n[pip3] numpy==1.20.1\r\n[pip3] numpydoc==1.1.0\r\n[pip3] torch==1.12.1+cu113\r\n[pip3] torchaudio==0.12.1+cu113\r\n[pip3] torchmultimodal==0.1.0a0\r\n[pip3] torchvision==0.13.1+cu113\r\n[conda] blas 1.0 mkl \r\n[conda] mkl 2021.2.0 h06a4308_296 \r\n[conda] mkl-service 2.3.0 py38h27cfd23_1 \r\n[conda] mkl_fft 1.3.0 py38h42c9631_2 \r\n[conda] mkl_random 1.2.1 py38ha9443f7_2 \r\n[conda] numpy 1.20.1 py38h93e21f0_0 \r\n[conda] numpy-base 1.20.1 py38h7d8b39e_0 \r\n[conda] numpydoc 1.1.0 pyhd3eb1b0_1 \r\n[conda] torch 1.12.1+cu113 pypi_0 pypi\r\n[conda] torchaudio 0.12.1+cu113 pypi_0 pypi\r\n[conda] torchmultimodal 0.1.0a0 dev_0 <develop>\r\n[conda] torchvision 0.13.1+cu113 pypi_0 pypi\r\n\n\ncc @pmeier\n", "before_files": [{"content": "import pathlib\nfrom typing import Any, Callable, Optional, Tuple\n\nfrom PIL import Image\n\nfrom .utils import _decompress, download_file_from_google_drive, verify_str_arg\nfrom .vision import VisionDataset\n\n\nclass PCAM(VisionDataset):\n \"\"\"`PCAM Dataset <https://github.com/basveeling/pcam>`_.\n\n The PatchCamelyon dataset is a binary classification dataset with 327,680\n color images (96px x 96px), extracted from histopathologic scans of lymph node\n sections. Each image is annotated with a binary label indicating presence of\n metastatic tissue.\n\n This dataset requires the ``h5py`` package which you can install with ``pip install h5py``.\n\n Args:\n root (string): Root directory of the dataset.\n split (string, optional): The dataset split, supports ``\"train\"`` (default), ``\"test\"`` or ``\"val\"``.\n transform (callable, optional): A function/transform that takes in a PIL image and returns a transformed\n version. E.g, ``transforms.RandomCrop``.\n target_transform (callable, optional): A function/transform that takes in the target and transforms it.\n download (bool, optional): If True, downloads the dataset from the internet and puts it into ``root/pcam``. If\n dataset is already downloaded, it is not downloaded again.\n \"\"\"\n\n _FILES = {\n \"train\": {\n \"images\": (\n \"camelyonpatch_level_2_split_train_x.h5\", # Data file name\n \"1Ka0XfEMiwgCYPdTI-vv6eUElOBnKFKQ2\", # Google Drive ID\n \"1571f514728f59376b705fc836ff4b63\", # md5 hash\n ),\n \"targets\": (\n \"camelyonpatch_level_2_split_train_y.h5\",\n \"1269yhu3pZDP8UYFQs-NYs3FPwuK-nGSG\",\n \"35c2d7259d906cfc8143347bb8e05be7\",\n ),\n },\n \"test\": {\n \"images\": (\n \"camelyonpatch_level_2_split_test_x.h5\",\n \"1qV65ZqZvWzuIVthK8eVDhIwrbnsJdbg_\",\n \"d5b63470df7cfa627aeec8b9dc0c066e\",\n ),\n \"targets\": (\n \"camelyonpatch_level_2_split_test_y.h5\",\n \"17BHrSrwWKjYsOgTMmoqrIjDy6Fa2o_gP\",\n \"2b85f58b927af9964a4c15b8f7e8f179\",\n ),\n },\n \"val\": {\n \"images\": (\n \"camelyonpatch_level_2_split_valid_x.h5\",\n \"1hgshYGWK8V-eGRy8LToWJJgDU_rXWVJ3\",\n \"d8c2d60d490dbd479f8199bdfa0cf6ec\",\n ),\n \"targets\": (\n \"camelyonpatch_level_2_split_valid_y.h5\",\n \"1bH8ZRbhSVAhScTS0p9-ZzGnX91cHT3uO\",\n \"60a7035772fbdb7f34eb86d4420cf66a\",\n ),\n },\n }\n\n def __init__(\n self,\n root: str,\n split: str = \"train\",\n transform: Optional[Callable] = None,\n target_transform: Optional[Callable] = None,\n download: bool = False,\n ):\n try:\n import h5py\n\n self.h5py = h5py\n except ImportError:\n raise RuntimeError(\n \"h5py is not found. This dataset needs to have h5py installed: please run pip install h5py\"\n )\n\n self._split = verify_str_arg(split, \"split\", (\"train\", \"test\", \"val\"))\n\n super().__init__(root, transform=transform, target_transform=target_transform)\n self._base_folder = pathlib.Path(self.root) / \"pcam\"\n\n if download:\n self._download()\n\n if not self._check_exists():\n raise RuntimeError(\"Dataset not found. You can use download=True to download it\")\n\n def __len__(self) -> int:\n images_file = self._FILES[self._split][\"images\"][0]\n with self.h5py.File(self._base_folder / images_file) as images_data:\n return images_data[\"x\"].shape[0]\n\n def __getitem__(self, idx: int) -> Tuple[Any, Any]:\n images_file = self._FILES[self._split][\"images\"][0]\n with self.h5py.File(self._base_folder / images_file) as images_data:\n image = Image.fromarray(images_data[\"x\"][idx]).convert(\"RGB\")\n\n targets_file = self._FILES[self._split][\"targets\"][0]\n with self.h5py.File(self._base_folder / targets_file) as targets_data:\n target = int(targets_data[\"y\"][idx, 0, 0, 0]) # shape is [num_images, 1, 1, 1]\n\n if self.transform:\n image = self.transform(image)\n if self.target_transform:\n target = self.target_transform(target)\n\n return image, target\n\n def _check_exists(self) -> bool:\n images_file = self._FILES[self._split][\"images\"][0]\n targets_file = self._FILES[self._split][\"targets\"][0]\n return all(self._base_folder.joinpath(h5_file).exists() for h5_file in (images_file, targets_file))\n\n def _download(self) -> None:\n if self._check_exists():\n return\n\n for file_name, file_id, md5 in self._FILES[self._split].values():\n archive_name = file_name + \".gz\"\n download_file_from_google_drive(file_id, str(self._base_folder), filename=archive_name, md5=md5)\n _decompress(str(self._base_folder / archive_name))\n", "path": "torchvision/datasets/pcam.py"}], "after_files": [{"content": "import pathlib\nfrom typing import Any, Callable, Optional, Tuple\n\nfrom PIL import Image\n\nfrom .utils import _decompress, download_file_from_google_drive, verify_str_arg\nfrom .vision import VisionDataset\n\n\nclass PCAM(VisionDataset):\n \"\"\"`PCAM Dataset <https://github.com/basveeling/pcam>`_.\n\n The PatchCamelyon dataset is a binary classification dataset with 327,680\n color images (96px x 96px), extracted from histopathologic scans of lymph node\n sections. Each image is annotated with a binary label indicating presence of\n metastatic tissue.\n\n This dataset requires the ``h5py`` package which you can install with ``pip install h5py``.\n\n Args:\n root (string): Root directory of the dataset.\n split (string, optional): The dataset split, supports ``\"train\"`` (default), ``\"test\"`` or ``\"val\"``.\n transform (callable, optional): A function/transform that takes in a PIL image and returns a transformed\n version. E.g, ``transforms.RandomCrop``.\n target_transform (callable, optional): A function/transform that takes in the target and transforms it.\n download (bool, optional): If True, downloads the dataset from the internet and puts it into ``root/pcam``. If\n dataset is already downloaded, it is not downloaded again.\n \"\"\"\n\n _FILES = {\n \"train\": {\n \"images\": (\n \"camelyonpatch_level_2_split_train_x.h5\", # Data file name\n \"1Ka0XfEMiwgCYPdTI-vv6eUElOBnKFKQ2\", # Google Drive ID\n \"1571f514728f59376b705fc836ff4b63\", # md5 hash\n ),\n \"targets\": (\n \"camelyonpatch_level_2_split_train_y.h5\",\n \"1269yhu3pZDP8UYFQs-NYs3FPwuK-nGSG\",\n \"35c2d7259d906cfc8143347bb8e05be7\",\n ),\n },\n \"test\": {\n \"images\": (\n \"camelyonpatch_level_2_split_test_x.h5\",\n \"1qV65ZqZvWzuIVthK8eVDhIwrbnsJdbg_\",\n \"d8c2d60d490dbd479f8199bdfa0cf6ec\",\n ),\n \"targets\": (\n \"camelyonpatch_level_2_split_test_y.h5\",\n \"17BHrSrwWKjYsOgTMmoqrIjDy6Fa2o_gP\",\n \"60a7035772fbdb7f34eb86d4420cf66a\",\n ),\n },\n \"val\": {\n \"images\": (\n \"camelyonpatch_level_2_split_valid_x.h5\",\n \"1hgshYGWK8V-eGRy8LToWJJgDU_rXWVJ3\",\n \"d5b63470df7cfa627aeec8b9dc0c066e\",\n ),\n \"targets\": (\n \"camelyonpatch_level_2_split_valid_y.h5\",\n \"1bH8ZRbhSVAhScTS0p9-ZzGnX91cHT3uO\",\n \"2b85f58b927af9964a4c15b8f7e8f179\",\n ),\n },\n }\n\n def __init__(\n self,\n root: str,\n split: str = \"train\",\n transform: Optional[Callable] = None,\n target_transform: Optional[Callable] = None,\n download: bool = False,\n ):\n try:\n import h5py\n\n self.h5py = h5py\n except ImportError:\n raise RuntimeError(\n \"h5py is not found. This dataset needs to have h5py installed: please run pip install h5py\"\n )\n\n self._split = verify_str_arg(split, \"split\", (\"train\", \"test\", \"val\"))\n\n super().__init__(root, transform=transform, target_transform=target_transform)\n self._base_folder = pathlib.Path(self.root) / \"pcam\"\n\n if download:\n self._download()\n\n if not self._check_exists():\n raise RuntimeError(\"Dataset not found. You can use download=True to download it\")\n\n def __len__(self) -> int:\n images_file = self._FILES[self._split][\"images\"][0]\n with self.h5py.File(self._base_folder / images_file) as images_data:\n return images_data[\"x\"].shape[0]\n\n def __getitem__(self, idx: int) -> Tuple[Any, Any]:\n images_file = self._FILES[self._split][\"images\"][0]\n with self.h5py.File(self._base_folder / images_file) as images_data:\n image = Image.fromarray(images_data[\"x\"][idx]).convert(\"RGB\")\n\n targets_file = self._FILES[self._split][\"targets\"][0]\n with self.h5py.File(self._base_folder / targets_file) as targets_data:\n target = int(targets_data[\"y\"][idx, 0, 0, 0]) # shape is [num_images, 1, 1, 1]\n\n if self.transform:\n image = self.transform(image)\n if self.target_transform:\n target = self.target_transform(target)\n\n return image, target\n\n def _check_exists(self) -> bool:\n images_file = self._FILES[self._split][\"images\"][0]\n targets_file = self._FILES[self._split][\"targets\"][0]\n return all(self._base_folder.joinpath(h5_file).exists() for h5_file in (images_file, targets_file))\n\n def _download(self) -> None:\n if self._check_exists():\n return\n\n for file_name, file_id, md5 in self._FILES[self._split].values():\n archive_name = file_name + \".gz\"\n download_file_from_google_drive(file_id, str(self._base_folder), filename=archive_name, md5=md5)\n _decompress(str(self._base_folder / archive_name))\n", "path": "torchvision/datasets/pcam.py"}]} | 3,863 | 530 |
gh_patches_debug_17748 | rasdani/github-patches | git_diff | canonical__microk8s-2148 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Indentation error in yaml output of microk8s.status
The output of `microk8s.status` is
```
microk8s:
running: False
high-availability:
enabled: False
message: microk8s is not running. Use microk8s inspect for a deeper inspection.
```
which confuses some parsers (e.g. the built-in Python 3.8) due to the extraneous space before `message`.
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `scripts/wrappers/status.py`
Content:
```
1 #!/usr/bin/python3
2 import os
3 import argparse
4
5 from common.utils import (
6 exit_if_no_permission,
7 exit_if_stopped,
8 is_cluster_locked,
9 is_ha_enabled,
10 get_dqlite_info,
11 wait_for_ready,
12 is_cluster_ready,
13 get_available_addons,
14 get_current_arch,
15 get_addon_by_name,
16 kubectl_get,
17 kubectl_get_clusterroles,
18 )
19
20
21 def is_enabled(addon, item):
22 if addon in item:
23 return True
24 else:
25 filepath = os.path.expandvars(addon)
26 return os.path.isfile(filepath)
27
28 return False
29
30
31 def print_short(isReady, enabled_addons, disabled_addons):
32 if isReady:
33 print("microk8s is running")
34 print("addons:")
35 if enabled_addons and len(enabled_addons) > 0:
36 for enabled in enabled_addons:
37 print("{}: enabled".format(enabled["name"]))
38 if disabled_addons and len(disabled_addons) > 0:
39 for disabled in disabled_addons:
40 print("{}: disabled".format(disabled["name"]))
41 else:
42 print("microk8s is not running. Use microk8s inspect for a deeper inspection.")
43
44
45 def print_pretty(isReady, enabled_addons, disabled_addons):
46 console_formatter = "{:>3} {:<20} # {}"
47 if isReady:
48 print("microk8s is running")
49 if not is_ha_enabled():
50 print("high-availability: no")
51 else:
52 info = get_dqlite_info()
53 if ha_cluster_formed(info):
54 print("high-availability: yes")
55 else:
56 print("high-availability: no")
57
58 masters = "none"
59 standby = "none"
60 for node in info:
61 if node[1] == "voter":
62 if masters == "none":
63 masters = "{}".format(node[0])
64 else:
65 masters = "{} {}".format(masters, node[0])
66 if node[1] == "standby":
67 if standby == "none":
68 standby = "{}".format(node[0])
69 else:
70 standby = "{} {}".format(standby, node[0])
71
72 print("{:>2}{} {}".format("", "datastore master nodes:", masters))
73 print("{:>2}{} {}".format("", "datastore standby nodes:", standby))
74
75 print("addons:")
76 if enabled_addons and len(enabled_addons) > 0:
77 print("{:>2}{}".format("", "enabled:"))
78 for enabled in enabled_addons:
79 print(console_formatter.format("", enabled["name"], enabled["description"]))
80 if disabled_addons and len(disabled_addons) > 0:
81 print("{:>2}{}".format("", "disabled:"))
82 for disabled in disabled_addons:
83 print(console_formatter.format("", disabled["name"], disabled["description"]))
84 else:
85 print("microk8s is not running. Use microk8s inspect for a deeper inspection.")
86
87
88 def print_short_yaml(isReady, enabled_addons, disabled_addons):
89 print("microk8s:")
90 print("{:>2}{} {}".format("", "running:", isReady))
91
92 if isReady:
93 print("addons:")
94 for enabled in enabled_addons:
95 print(" {}: enabled".format(enabled["name"]))
96
97 for disabled in disabled_addons:
98 print(" {}: disabled".format(disabled["name"]))
99 else:
100 print(
101 "{:>2} {} {}".format(
102 "",
103 "message:",
104 "microk8s is not running. Use microk8s inspect for a deeper inspection.",
105 )
106 )
107
108
109 def print_yaml(isReady, enabled_addons, disabled_addons):
110 print("microk8s:")
111 print("{:>2}{} {}".format("", "running:", isReady))
112
113 print("{:>2}".format("high-availability:"))
114 ha_enabled = is_ha_enabled()
115 print("{:>2}{} {}".format("", "enabled:", ha_enabled))
116 if ha_enabled:
117 info = get_dqlite_info()
118 print("{:>2}{}".format("", "nodes:"))
119 for node in info:
120 print("{:>6}address: {:<1}".format("- ", node[0]))
121 print("{:>6}role: {:<1}".format("", node[1]))
122
123 if isReady:
124 print("{:>2}".format("addons:"))
125 for enabled in enabled_addons:
126 print("{:>4}name: {:<1}".format("- ", enabled["name"]))
127 print("{:>4}description: {:<1}".format("", enabled["description"]))
128 print("{:>4}version: {:<1}".format("", enabled["version"]))
129 print("{:>4}status: enabled".format(""))
130
131 for disabled in disabled_addons:
132 print("{:>4}name: {:<1}".format("- ", disabled["name"]))
133 print("{:>4}description: {:<1}".format("", disabled["description"]))
134 print("{:>4}version: {:<1}".format("", disabled["version"]))
135 print("{:>4}status: disabled".format(""))
136 else:
137 print(
138 "{:>2} {} {}".format(
139 "",
140 "message:",
141 "microk8s is not running. Use microk8s inspect for a deeper inspection.",
142 )
143 )
144
145
146 def print_addon_status(enabled):
147 if len(enabled) > 0:
148 print("enabled")
149 else:
150 print("disabled")
151
152
153 def get_status(available_addons, isReady):
154 enabled = []
155 disabled = []
156 if isReady:
157 kube_output = kubectl_get("all")
158 cluster_output = kubectl_get_clusterroles()
159 kube_output = kube_output + cluster_output
160 for addon in available_addons:
161 found = False
162 for row in kube_output.split("\n"):
163 if is_enabled(addon["check_status"], row):
164 enabled.append(addon)
165 found = True
166 break
167 if not found:
168 disabled.append(addon)
169
170 return enabled, disabled
171
172
173 def ha_cluster_formed(info):
174 voters = 0
175 for node in info:
176 if node[1] == "voter":
177 voters += 1
178 ha_formed = False
179 if voters > 2:
180 ha_formed = True
181 return ha_formed
182
183
184 if __name__ == "__main__":
185 exit_if_no_permission()
186 exit_if_stopped()
187 is_cluster_locked()
188
189 # initiate the parser with a description
190 parser = argparse.ArgumentParser(
191 description="Microk8s cluster status check.", prog="microk8s status"
192 )
193 parser.add_argument(
194 "--format",
195 help="print cluster and addon status, output can be in yaml, pretty or short",
196 default="pretty",
197 choices={"pretty", "yaml", "short"},
198 )
199 parser.add_argument(
200 "-w", "--wait-ready", action="store_true", help="wait until the cluster is in ready state"
201 )
202 parser.add_argument(
203 "-t",
204 "--timeout",
205 help="specify a timeout in seconds when waiting for the cluster to be ready.",
206 type=int,
207 default=0,
208 )
209 parser.add_argument("-a", "--addon", help="check the status of an addon.", default="all")
210 parser.add_argument(
211 "--yaml", action="store_true", help="DEPRECATED, use '--format yaml' instead"
212 )
213
214 # read arguments from the command line
215 args = parser.parse_args()
216
217 wait_ready = args.wait_ready
218 timeout = args.timeout
219 yaml_short = args.yaml
220
221 if wait_ready:
222 isReady = wait_for_ready(timeout)
223 else:
224 isReady = is_cluster_ready()
225
226 available_addons = get_available_addons(get_current_arch())
227
228 if args.addon != "all":
229 available_addons = get_addon_by_name(available_addons, args.addon)
230
231 enabled, disabled = get_status(available_addons, isReady)
232
233 if args.addon != "all":
234 print_addon_status(enabled)
235 else:
236 if args.format == "yaml":
237 print_yaml(isReady, enabled, disabled)
238 elif args.format == "short":
239 print_short(isReady, enabled, disabled)
240 else:
241 if yaml_short:
242 print_short_yaml(isReady, enabled, disabled)
243 else:
244 print_pretty(isReady, enabled, disabled)
245
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/scripts/wrappers/status.py b/scripts/wrappers/status.py
--- a/scripts/wrappers/status.py
+++ b/scripts/wrappers/status.py
@@ -98,7 +98,7 @@
print(" {}: disabled".format(disabled["name"]))
else:
print(
- "{:>2} {} {}".format(
+ "{:>2}{} {}".format(
"",
"message:",
"microk8s is not running. Use microk8s inspect for a deeper inspection.",
@@ -135,7 +135,7 @@
print("{:>4}status: disabled".format(""))
else:
print(
- "{:>2} {} {}".format(
+ "{:>2}{} {}".format(
"",
"message:",
"microk8s is not running. Use microk8s inspect for a deeper inspection.",
| {"golden_diff": "diff --git a/scripts/wrappers/status.py b/scripts/wrappers/status.py\n--- a/scripts/wrappers/status.py\n+++ b/scripts/wrappers/status.py\n@@ -98,7 +98,7 @@\n print(\" {}: disabled\".format(disabled[\"name\"]))\n else:\n print(\n- \"{:>2} {} {}\".format(\n+ \"{:>2}{} {}\".format(\n \"\",\n \"message:\",\n \"microk8s is not running. Use microk8s inspect for a deeper inspection.\",\n@@ -135,7 +135,7 @@\n print(\"{:>4}status: disabled\".format(\"\"))\n else:\n print(\n- \"{:>2} {} {}\".format(\n+ \"{:>2}{} {}\".format(\n \"\",\n \"message:\",\n \"microk8s is not running. Use microk8s inspect for a deeper inspection.\",\n", "issue": "Indentation error in yaml output of microk8s.status\nThe output of `microk8s.status` is\r\n\r\n```\r\nmicrok8s:\r\n running: False\r\nhigh-availability:\r\n enabled: False\r\n message: microk8s is not running. Use microk8s inspect for a deeper inspection.\r\n```\r\n\r\nwhich confuses some parsers (e.g. the built-in Python 3.8) due to the extraneous space before `message`.\n", "before_files": [{"content": "#!/usr/bin/python3\nimport os\nimport argparse\n\nfrom common.utils import (\n exit_if_no_permission,\n exit_if_stopped,\n is_cluster_locked,\n is_ha_enabled,\n get_dqlite_info,\n wait_for_ready,\n is_cluster_ready,\n get_available_addons,\n get_current_arch,\n get_addon_by_name,\n kubectl_get,\n kubectl_get_clusterroles,\n)\n\n\ndef is_enabled(addon, item):\n if addon in item:\n return True\n else:\n filepath = os.path.expandvars(addon)\n return os.path.isfile(filepath)\n\n return False\n\n\ndef print_short(isReady, enabled_addons, disabled_addons):\n if isReady:\n print(\"microk8s is running\")\n print(\"addons:\")\n if enabled_addons and len(enabled_addons) > 0:\n for enabled in enabled_addons:\n print(\"{}: enabled\".format(enabled[\"name\"]))\n if disabled_addons and len(disabled_addons) > 0:\n for disabled in disabled_addons:\n print(\"{}: disabled\".format(disabled[\"name\"]))\n else:\n print(\"microk8s is not running. Use microk8s inspect for a deeper inspection.\")\n\n\ndef print_pretty(isReady, enabled_addons, disabled_addons):\n console_formatter = \"{:>3} {:<20} # {}\"\n if isReady:\n print(\"microk8s is running\")\n if not is_ha_enabled():\n print(\"high-availability: no\")\n else:\n info = get_dqlite_info()\n if ha_cluster_formed(info):\n print(\"high-availability: yes\")\n else:\n print(\"high-availability: no\")\n\n masters = \"none\"\n standby = \"none\"\n for node in info:\n if node[1] == \"voter\":\n if masters == \"none\":\n masters = \"{}\".format(node[0])\n else:\n masters = \"{} {}\".format(masters, node[0])\n if node[1] == \"standby\":\n if standby == \"none\":\n standby = \"{}\".format(node[0])\n else:\n standby = \"{} {}\".format(standby, node[0])\n\n print(\"{:>2}{} {}\".format(\"\", \"datastore master nodes:\", masters))\n print(\"{:>2}{} {}\".format(\"\", \"datastore standby nodes:\", standby))\n\n print(\"addons:\")\n if enabled_addons and len(enabled_addons) > 0:\n print(\"{:>2}{}\".format(\"\", \"enabled:\"))\n for enabled in enabled_addons:\n print(console_formatter.format(\"\", enabled[\"name\"], enabled[\"description\"]))\n if disabled_addons and len(disabled_addons) > 0:\n print(\"{:>2}{}\".format(\"\", \"disabled:\"))\n for disabled in disabled_addons:\n print(console_formatter.format(\"\", disabled[\"name\"], disabled[\"description\"]))\n else:\n print(\"microk8s is not running. Use microk8s inspect for a deeper inspection.\")\n\n\ndef print_short_yaml(isReady, enabled_addons, disabled_addons):\n print(\"microk8s:\")\n print(\"{:>2}{} {}\".format(\"\", \"running:\", isReady))\n\n if isReady:\n print(\"addons:\")\n for enabled in enabled_addons:\n print(\" {}: enabled\".format(enabled[\"name\"]))\n\n for disabled in disabled_addons:\n print(\" {}: disabled\".format(disabled[\"name\"]))\n else:\n print(\n \"{:>2} {} {}\".format(\n \"\",\n \"message:\",\n \"microk8s is not running. Use microk8s inspect for a deeper inspection.\",\n )\n )\n\n\ndef print_yaml(isReady, enabled_addons, disabled_addons):\n print(\"microk8s:\")\n print(\"{:>2}{} {}\".format(\"\", \"running:\", isReady))\n\n print(\"{:>2}\".format(\"high-availability:\"))\n ha_enabled = is_ha_enabled()\n print(\"{:>2}{} {}\".format(\"\", \"enabled:\", ha_enabled))\n if ha_enabled:\n info = get_dqlite_info()\n print(\"{:>2}{}\".format(\"\", \"nodes:\"))\n for node in info:\n print(\"{:>6}address: {:<1}\".format(\"- \", node[0]))\n print(\"{:>6}role: {:<1}\".format(\"\", node[1]))\n\n if isReady:\n print(\"{:>2}\".format(\"addons:\"))\n for enabled in enabled_addons:\n print(\"{:>4}name: {:<1}\".format(\"- \", enabled[\"name\"]))\n print(\"{:>4}description: {:<1}\".format(\"\", enabled[\"description\"]))\n print(\"{:>4}version: {:<1}\".format(\"\", enabled[\"version\"]))\n print(\"{:>4}status: enabled\".format(\"\"))\n\n for disabled in disabled_addons:\n print(\"{:>4}name: {:<1}\".format(\"- \", disabled[\"name\"]))\n print(\"{:>4}description: {:<1}\".format(\"\", disabled[\"description\"]))\n print(\"{:>4}version: {:<1}\".format(\"\", disabled[\"version\"]))\n print(\"{:>4}status: disabled\".format(\"\"))\n else:\n print(\n \"{:>2} {} {}\".format(\n \"\",\n \"message:\",\n \"microk8s is not running. Use microk8s inspect for a deeper inspection.\",\n )\n )\n\n\ndef print_addon_status(enabled):\n if len(enabled) > 0:\n print(\"enabled\")\n else:\n print(\"disabled\")\n\n\ndef get_status(available_addons, isReady):\n enabled = []\n disabled = []\n if isReady:\n kube_output = kubectl_get(\"all\")\n cluster_output = kubectl_get_clusterroles()\n kube_output = kube_output + cluster_output\n for addon in available_addons:\n found = False\n for row in kube_output.split(\"\\n\"):\n if is_enabled(addon[\"check_status\"], row):\n enabled.append(addon)\n found = True\n break\n if not found:\n disabled.append(addon)\n\n return enabled, disabled\n\n\ndef ha_cluster_formed(info):\n voters = 0\n for node in info:\n if node[1] == \"voter\":\n voters += 1\n ha_formed = False\n if voters > 2:\n ha_formed = True\n return ha_formed\n\n\nif __name__ == \"__main__\":\n exit_if_no_permission()\n exit_if_stopped()\n is_cluster_locked()\n\n # initiate the parser with a description\n parser = argparse.ArgumentParser(\n description=\"Microk8s cluster status check.\", prog=\"microk8s status\"\n )\n parser.add_argument(\n \"--format\",\n help=\"print cluster and addon status, output can be in yaml, pretty or short\",\n default=\"pretty\",\n choices={\"pretty\", \"yaml\", \"short\"},\n )\n parser.add_argument(\n \"-w\", \"--wait-ready\", action=\"store_true\", help=\"wait until the cluster is in ready state\"\n )\n parser.add_argument(\n \"-t\",\n \"--timeout\",\n help=\"specify a timeout in seconds when waiting for the cluster to be ready.\",\n type=int,\n default=0,\n )\n parser.add_argument(\"-a\", \"--addon\", help=\"check the status of an addon.\", default=\"all\")\n parser.add_argument(\n \"--yaml\", action=\"store_true\", help=\"DEPRECATED, use '--format yaml' instead\"\n )\n\n # read arguments from the command line\n args = parser.parse_args()\n\n wait_ready = args.wait_ready\n timeout = args.timeout\n yaml_short = args.yaml\n\n if wait_ready:\n isReady = wait_for_ready(timeout)\n else:\n isReady = is_cluster_ready()\n\n available_addons = get_available_addons(get_current_arch())\n\n if args.addon != \"all\":\n available_addons = get_addon_by_name(available_addons, args.addon)\n\n enabled, disabled = get_status(available_addons, isReady)\n\n if args.addon != \"all\":\n print_addon_status(enabled)\n else:\n if args.format == \"yaml\":\n print_yaml(isReady, enabled, disabled)\n elif args.format == \"short\":\n print_short(isReady, enabled, disabled)\n else:\n if yaml_short:\n print_short_yaml(isReady, enabled, disabled)\n else:\n print_pretty(isReady, enabled, disabled)\n", "path": "scripts/wrappers/status.py"}], "after_files": [{"content": "#!/usr/bin/python3\nimport os\nimport argparse\n\nfrom common.utils import (\n exit_if_no_permission,\n exit_if_stopped,\n is_cluster_locked,\n is_ha_enabled,\n get_dqlite_info,\n wait_for_ready,\n is_cluster_ready,\n get_available_addons,\n get_current_arch,\n get_addon_by_name,\n kubectl_get,\n kubectl_get_clusterroles,\n)\n\n\ndef is_enabled(addon, item):\n if addon in item:\n return True\n else:\n filepath = os.path.expandvars(addon)\n return os.path.isfile(filepath)\n\n return False\n\n\ndef print_short(isReady, enabled_addons, disabled_addons):\n if isReady:\n print(\"microk8s is running\")\n print(\"addons:\")\n if enabled_addons and len(enabled_addons) > 0:\n for enabled in enabled_addons:\n print(\"{}: enabled\".format(enabled[\"name\"]))\n if disabled_addons and len(disabled_addons) > 0:\n for disabled in disabled_addons:\n print(\"{}: disabled\".format(disabled[\"name\"]))\n else:\n print(\"microk8s is not running. Use microk8s inspect for a deeper inspection.\")\n\n\ndef print_pretty(isReady, enabled_addons, disabled_addons):\n console_formatter = \"{:>3} {:<20} # {}\"\n if isReady:\n print(\"microk8s is running\")\n if not is_ha_enabled():\n print(\"high-availability: no\")\n else:\n info = get_dqlite_info()\n if ha_cluster_formed(info):\n print(\"high-availability: yes\")\n else:\n print(\"high-availability: no\")\n\n masters = \"none\"\n standby = \"none\"\n for node in info:\n if node[1] == \"voter\":\n if masters == \"none\":\n masters = \"{}\".format(node[0])\n else:\n masters = \"{} {}\".format(masters, node[0])\n if node[1] == \"standby\":\n if standby == \"none\":\n standby = \"{}\".format(node[0])\n else:\n standby = \"{} {}\".format(standby, node[0])\n\n print(\"{:>2}{} {}\".format(\"\", \"datastore master nodes:\", masters))\n print(\"{:>2}{} {}\".format(\"\", \"datastore standby nodes:\", standby))\n\n print(\"addons:\")\n if enabled_addons and len(enabled_addons) > 0:\n print(\"{:>2}{}\".format(\"\", \"enabled:\"))\n for enabled in enabled_addons:\n print(console_formatter.format(\"\", enabled[\"name\"], enabled[\"description\"]))\n if disabled_addons and len(disabled_addons) > 0:\n print(\"{:>2}{}\".format(\"\", \"disabled:\"))\n for disabled in disabled_addons:\n print(console_formatter.format(\"\", disabled[\"name\"], disabled[\"description\"]))\n else:\n print(\"microk8s is not running. Use microk8s inspect for a deeper inspection.\")\n\n\ndef print_short_yaml(isReady, enabled_addons, disabled_addons):\n print(\"microk8s:\")\n print(\"{:>2}{} {}\".format(\"\", \"running:\", isReady))\n\n if isReady:\n print(\"addons:\")\n for enabled in enabled_addons:\n print(\" {}: enabled\".format(enabled[\"name\"]))\n\n for disabled in disabled_addons:\n print(\" {}: disabled\".format(disabled[\"name\"]))\n else:\n print(\n \"{:>2}{} {}\".format(\n \"\",\n \"message:\",\n \"microk8s is not running. Use microk8s inspect for a deeper inspection.\",\n )\n )\n\n\ndef print_yaml(isReady, enabled_addons, disabled_addons):\n print(\"microk8s:\")\n print(\"{:>2}{} {}\".format(\"\", \"running:\", isReady))\n\n print(\"{:>2}\".format(\"high-availability:\"))\n ha_enabled = is_ha_enabled()\n print(\"{:>2}{} {}\".format(\"\", \"enabled:\", ha_enabled))\n if ha_enabled:\n info = get_dqlite_info()\n print(\"{:>2}{}\".format(\"\", \"nodes:\"))\n for node in info:\n print(\"{:>6}address: {:<1}\".format(\"- \", node[0]))\n print(\"{:>6}role: {:<1}\".format(\"\", node[1]))\n\n if isReady:\n print(\"{:>2}\".format(\"addons:\"))\n for enabled in enabled_addons:\n print(\"{:>4}name: {:<1}\".format(\"- \", enabled[\"name\"]))\n print(\"{:>4}description: {:<1}\".format(\"\", enabled[\"description\"]))\n print(\"{:>4}version: {:<1}\".format(\"\", enabled[\"version\"]))\n print(\"{:>4}status: enabled\".format(\"\"))\n\n for disabled in disabled_addons:\n print(\"{:>4}name: {:<1}\".format(\"- \", disabled[\"name\"]))\n print(\"{:>4}description: {:<1}\".format(\"\", disabled[\"description\"]))\n print(\"{:>4}version: {:<1}\".format(\"\", disabled[\"version\"]))\n print(\"{:>4}status: disabled\".format(\"\"))\n else:\n print(\n \"{:>2}{} {}\".format(\n \"\",\n \"message:\",\n \"microk8s is not running. Use microk8s inspect for a deeper inspection.\",\n )\n )\n\n\ndef print_addon_status(enabled):\n if len(enabled) > 0:\n print(\"enabled\")\n else:\n print(\"disabled\")\n\n\ndef get_status(available_addons, isReady):\n enabled = []\n disabled = []\n if isReady:\n kube_output = kubectl_get(\"all\")\n cluster_output = kubectl_get_clusterroles()\n kube_output = kube_output + cluster_output\n for addon in available_addons:\n found = False\n for row in kube_output.split(\"\\n\"):\n if is_enabled(addon[\"check_status\"], row):\n enabled.append(addon)\n found = True\n break\n if not found:\n disabled.append(addon)\n\n return enabled, disabled\n\n\ndef ha_cluster_formed(info):\n voters = 0\n for node in info:\n if node[1] == \"voter\":\n voters += 1\n ha_formed = False\n if voters > 2:\n ha_formed = True\n return ha_formed\n\n\nif __name__ == \"__main__\":\n exit_if_no_permission()\n exit_if_stopped()\n is_cluster_locked()\n\n # initiate the parser with a description\n parser = argparse.ArgumentParser(\n description=\"Microk8s cluster status check.\", prog=\"microk8s status\"\n )\n parser.add_argument(\n \"--format\",\n help=\"print cluster and addon status, output can be in yaml, pretty or short\",\n default=\"pretty\",\n choices={\"pretty\", \"yaml\", \"short\"},\n )\n parser.add_argument(\n \"-w\", \"--wait-ready\", action=\"store_true\", help=\"wait until the cluster is in ready state\"\n )\n parser.add_argument(\n \"-t\",\n \"--timeout\",\n help=\"specify a timeout in seconds when waiting for the cluster to be ready.\",\n type=int,\n default=0,\n )\n parser.add_argument(\"-a\", \"--addon\", help=\"check the status of an addon.\", default=\"all\")\n parser.add_argument(\n \"--yaml\", action=\"store_true\", help=\"DEPRECATED, use '--format yaml' instead\"\n )\n\n # read arguments from the command line\n args = parser.parse_args()\n\n wait_ready = args.wait_ready\n timeout = args.timeout\n yaml_short = args.yaml\n\n if wait_ready:\n isReady = wait_for_ready(timeout)\n else:\n isReady = is_cluster_ready()\n\n available_addons = get_available_addons(get_current_arch())\n\n if args.addon != \"all\":\n available_addons = get_addon_by_name(available_addons, args.addon)\n\n enabled, disabled = get_status(available_addons, isReady)\n\n if args.addon != \"all\":\n print_addon_status(enabled)\n else:\n if args.format == \"yaml\":\n print_yaml(isReady, enabled, disabled)\n elif args.format == \"short\":\n print_short(isReady, enabled, disabled)\n else:\n if yaml_short:\n print_short_yaml(isReady, enabled, disabled)\n else:\n print_pretty(isReady, enabled, disabled)\n", "path": "scripts/wrappers/status.py"}]} | 2,803 | 193 |
gh_patches_debug_14371 | rasdani/github-patches | git_diff | aio-libs__aiohttp-6164 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
tests_require: add trustme
It is required since https://github.com/aio-libs/aiohttp/pull/3487.
<!-- Thank you for your contribution! -->
## What do these changes do?
<!-- Please give a short brief about these changes. -->
## Are there changes in behavior for the user?
<!-- Outline any notable behaviour for the end users. -->
## Related issue number
<!-- Are there any issues opened that will be resolved by merging this change? -->
## Checklist
- [ ] I think the code is well written
- [ ] Unit tests for the changes exist
- [ ] Documentation reflects the changes
- [ ] If you provide code modification, please add yourself to `CONTRIBUTORS.txt`
* The format is <Name> <Surname>.
* Please keep alphabetical order, the file is sorted by names.
- [ ] Add a new news fragment into the `CHANGES` folder
* name it `<issue_id>.<type>` for example (588.bugfix)
* if you don't have an `issue_id` change it to the pr id after creating the pr
* ensure type is one of the following:
* `.feature`: Signifying a new feature.
* `.bugfix`: Signifying a bug fix.
* `.doc`: Signifying a documentation improvement.
* `.removal`: Signifying a deprecation or removal of public API.
* `.misc`: A ticket has been closed, but it is not of interest to users.
* Make sure to use full sentences with correct case and punctuation, for example: "Fix issue with non-ascii contents in doctest text files."
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `setup.py`
Content:
```
1 import pathlib
2 import re
3 import sys
4 from distutils.command.build_ext import build_ext
5 from distutils.errors import CCompilerError, DistutilsExecError, DistutilsPlatformError
6
7 from setuptools import Extension, setup
8
9 if sys.version_info < (3, 6):
10 raise RuntimeError("aiohttp 3.7+ requires Python 3.6+")
11
12 here = pathlib.Path(__file__).parent
13
14 if (here / ".git").exists() and not (here / "vendor/http-parser/README.md").exists():
15 print("Install submodules when building from git clone", file=sys.stderr)
16 print("Hint:", file=sys.stderr)
17 print(" git submodule update --init", file=sys.stderr)
18 sys.exit(2)
19
20
21 # NOTE: makefile cythonizes all Cython modules
22
23 extensions = [
24 Extension("aiohttp._websocket", ["aiohttp/_websocket.c"]),
25 Extension(
26 "aiohttp._http_parser",
27 [
28 "aiohttp/_http_parser.c",
29 "vendor/http-parser/http_parser.c",
30 "aiohttp/_find_header.c",
31 ],
32 define_macros=[("HTTP_PARSER_STRICT", 0)],
33 ),
34 Extension("aiohttp._helpers", ["aiohttp/_helpers.c"]),
35 Extension("aiohttp._http_writer", ["aiohttp/_http_writer.c"]),
36 ]
37
38
39 class BuildFailed(Exception):
40 pass
41
42
43 class ve_build_ext(build_ext):
44 # This class allows C extension building to fail.
45
46 def run(self):
47 try:
48 build_ext.run(self)
49 except (DistutilsPlatformError, FileNotFoundError):
50 raise BuildFailed()
51
52 def build_extension(self, ext):
53 try:
54 build_ext.build_extension(self, ext)
55 except (CCompilerError, DistutilsExecError, DistutilsPlatformError, ValueError):
56 raise BuildFailed()
57
58
59 txt = (here / "aiohttp" / "__init__.py").read_text("utf-8")
60 try:
61 version = re.findall(r'^__version__ = "([^"]+)"\r?$', txt, re.M)[0]
62 except IndexError:
63 raise RuntimeError("Unable to determine version.")
64
65 install_requires = [
66 "attrs>=17.3.0",
67 "charset-normalizer>=2.0,<3.0",
68 "multidict>=4.5,<7.0",
69 "async_timeout>=4.0.0a3,<5.0",
70 'asynctest==0.13.0; python_version<"3.8"',
71 "yarl>=1.0,<2.0",
72 'idna-ssl>=1.0; python_version<"3.7"',
73 'typing_extensions>=3.7.4; python_version<"3.8"',
74 "frozenlist>=1.1.1",
75 "aiosignal>=1.1.2",
76 ]
77
78
79 def read(f):
80 return (here / f).read_text("utf-8").strip()
81
82
83 NEEDS_PYTEST = {"pytest", "test"}.intersection(sys.argv)
84 pytest_runner = ["pytest-runner"] if NEEDS_PYTEST else []
85
86 tests_require = [
87 "pytest",
88 "gunicorn",
89 "pytest-timeout",
90 "async-generator",
91 "pytest-xdist",
92 ]
93
94
95 args = dict(
96 name="aiohttp",
97 version=version,
98 description="Async http client/server framework (asyncio)",
99 long_description=read("README.rst"),
100 long_description_content_type="text/x-rst",
101 classifiers=[
102 "License :: OSI Approved :: Apache Software License",
103 "Intended Audience :: Developers",
104 "Programming Language :: Python",
105 "Programming Language :: Python :: 3",
106 "Programming Language :: Python :: 3.6",
107 "Programming Language :: Python :: 3.7",
108 "Programming Language :: Python :: 3.8",
109 "Programming Language :: Python :: 3.9",
110 "Programming Language :: Python :: 3.10",
111 "Development Status :: 5 - Production/Stable",
112 "Operating System :: POSIX",
113 "Operating System :: MacOS :: MacOS X",
114 "Operating System :: Microsoft :: Windows",
115 "Topic :: Internet :: WWW/HTTP",
116 "Framework :: AsyncIO",
117 ],
118 author="Nikolay Kim",
119 author_email="[email protected]",
120 maintainer=", ".join(
121 (
122 "Nikolay Kim <[email protected]>",
123 "Andrew Svetlov <[email protected]>",
124 )
125 ),
126 maintainer_email="[email protected]",
127 url="https://github.com/aio-libs/aiohttp",
128 project_urls={
129 "Chat: Gitter": "https://gitter.im/aio-libs/Lobby",
130 "CI: GitHub Actions": "https://github.com/aio-libs/aiohttp/actions?query=workflow%3ACI", # noqa
131 "Coverage: codecov": "https://codecov.io/github/aio-libs/aiohttp",
132 "Docs: RTD": "https://docs.aiohttp.org",
133 "GitHub: issues": "https://github.com/aio-libs/aiohttp/issues",
134 "GitHub: repo": "https://github.com/aio-libs/aiohttp",
135 },
136 license="Apache 2",
137 packages=["aiohttp"],
138 python_requires=">=3.6",
139 install_requires=install_requires,
140 extras_require={
141 "speedups": [
142 "aiodns",
143 "Brotli",
144 "cchardet",
145 ],
146 },
147 tests_require=tests_require,
148 setup_requires=pytest_runner,
149 include_package_data=True,
150 ext_modules=extensions,
151 cmdclass=dict(build_ext=ve_build_ext),
152 )
153
154 try:
155 setup(**args)
156 except BuildFailed:
157 print("************************************************************")
158 print("Cannot compile C accelerator module, use pure python version")
159 print("************************************************************")
160 del args["ext_modules"]
161 del args["cmdclass"]
162 setup(**args)
163
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/setup.py b/setup.py
--- a/setup.py
+++ b/setup.py
@@ -80,18 +80,6 @@
return (here / f).read_text("utf-8").strip()
-NEEDS_PYTEST = {"pytest", "test"}.intersection(sys.argv)
-pytest_runner = ["pytest-runner"] if NEEDS_PYTEST else []
-
-tests_require = [
- "pytest",
- "gunicorn",
- "pytest-timeout",
- "async-generator",
- "pytest-xdist",
-]
-
-
args = dict(
name="aiohttp",
version=version,
@@ -144,8 +132,6 @@
"cchardet",
],
},
- tests_require=tests_require,
- setup_requires=pytest_runner,
include_package_data=True,
ext_modules=extensions,
cmdclass=dict(build_ext=ve_build_ext),
| {"golden_diff": "diff --git a/setup.py b/setup.py\n--- a/setup.py\n+++ b/setup.py\n@@ -80,18 +80,6 @@\n return (here / f).read_text(\"utf-8\").strip()\n \n \n-NEEDS_PYTEST = {\"pytest\", \"test\"}.intersection(sys.argv)\n-pytest_runner = [\"pytest-runner\"] if NEEDS_PYTEST else []\n-\n-tests_require = [\n- \"pytest\",\n- \"gunicorn\",\n- \"pytest-timeout\",\n- \"async-generator\",\n- \"pytest-xdist\",\n-]\n-\n-\n args = dict(\n name=\"aiohttp\",\n version=version,\n@@ -144,8 +132,6 @@\n \"cchardet\",\n ],\n },\n- tests_require=tests_require,\n- setup_requires=pytest_runner,\n include_package_data=True,\n ext_modules=extensions,\n cmdclass=dict(build_ext=ve_build_ext),\n", "issue": "tests_require: add trustme\nIt is required since https://github.com/aio-libs/aiohttp/pull/3487.\r\n\r\n<!-- Thank you for your contribution! -->\r\n\r\n## What do these changes do?\r\n\r\n<!-- Please give a short brief about these changes. -->\r\n\r\n## Are there changes in behavior for the user?\r\n\r\n<!-- Outline any notable behaviour for the end users. -->\r\n\r\n## Related issue number\r\n\r\n<!-- Are there any issues opened that will be resolved by merging this change? -->\r\n\r\n## Checklist\r\n\r\n- [ ] I think the code is well written\r\n- [ ] Unit tests for the changes exist\r\n- [ ] Documentation reflects the changes\r\n- [ ] If you provide code modification, please add yourself to `CONTRIBUTORS.txt`\r\n * The format is <Name> <Surname>.\r\n * Please keep alphabetical order, the file is sorted by names. \r\n- [ ] Add a new news fragment into the `CHANGES` folder\r\n * name it `<issue_id>.<type>` for example (588.bugfix)\r\n * if you don't have an `issue_id` change it to the pr id after creating the pr\r\n * ensure type is one of the following:\r\n * `.feature`: Signifying a new feature.\r\n * `.bugfix`: Signifying a bug fix.\r\n * `.doc`: Signifying a documentation improvement.\r\n * `.removal`: Signifying a deprecation or removal of public API.\r\n * `.misc`: A ticket has been closed, but it is not of interest to users.\r\n * Make sure to use full sentences with correct case and punctuation, for example: \"Fix issue with non-ascii contents in doctest text files.\"\r\n\n", "before_files": [{"content": "import pathlib\nimport re\nimport sys\nfrom distutils.command.build_ext import build_ext\nfrom distutils.errors import CCompilerError, DistutilsExecError, DistutilsPlatformError\n\nfrom setuptools import Extension, setup\n\nif sys.version_info < (3, 6):\n raise RuntimeError(\"aiohttp 3.7+ requires Python 3.6+\")\n\nhere = pathlib.Path(__file__).parent\n\nif (here / \".git\").exists() and not (here / \"vendor/http-parser/README.md\").exists():\n print(\"Install submodules when building from git clone\", file=sys.stderr)\n print(\"Hint:\", file=sys.stderr)\n print(\" git submodule update --init\", file=sys.stderr)\n sys.exit(2)\n\n\n# NOTE: makefile cythonizes all Cython modules\n\nextensions = [\n Extension(\"aiohttp._websocket\", [\"aiohttp/_websocket.c\"]),\n Extension(\n \"aiohttp._http_parser\",\n [\n \"aiohttp/_http_parser.c\",\n \"vendor/http-parser/http_parser.c\",\n \"aiohttp/_find_header.c\",\n ],\n define_macros=[(\"HTTP_PARSER_STRICT\", 0)],\n ),\n Extension(\"aiohttp._helpers\", [\"aiohttp/_helpers.c\"]),\n Extension(\"aiohttp._http_writer\", [\"aiohttp/_http_writer.c\"]),\n]\n\n\nclass BuildFailed(Exception):\n pass\n\n\nclass ve_build_ext(build_ext):\n # This class allows C extension building to fail.\n\n def run(self):\n try:\n build_ext.run(self)\n except (DistutilsPlatformError, FileNotFoundError):\n raise BuildFailed()\n\n def build_extension(self, ext):\n try:\n build_ext.build_extension(self, ext)\n except (CCompilerError, DistutilsExecError, DistutilsPlatformError, ValueError):\n raise BuildFailed()\n\n\ntxt = (here / \"aiohttp\" / \"__init__.py\").read_text(\"utf-8\")\ntry:\n version = re.findall(r'^__version__ = \"([^\"]+)\"\\r?$', txt, re.M)[0]\nexcept IndexError:\n raise RuntimeError(\"Unable to determine version.\")\n\ninstall_requires = [\n \"attrs>=17.3.0\",\n \"charset-normalizer>=2.0,<3.0\",\n \"multidict>=4.5,<7.0\",\n \"async_timeout>=4.0.0a3,<5.0\",\n 'asynctest==0.13.0; python_version<\"3.8\"',\n \"yarl>=1.0,<2.0\",\n 'idna-ssl>=1.0; python_version<\"3.7\"',\n 'typing_extensions>=3.7.4; python_version<\"3.8\"',\n \"frozenlist>=1.1.1\",\n \"aiosignal>=1.1.2\",\n]\n\n\ndef read(f):\n return (here / f).read_text(\"utf-8\").strip()\n\n\nNEEDS_PYTEST = {\"pytest\", \"test\"}.intersection(sys.argv)\npytest_runner = [\"pytest-runner\"] if NEEDS_PYTEST else []\n\ntests_require = [\n \"pytest\",\n \"gunicorn\",\n \"pytest-timeout\",\n \"async-generator\",\n \"pytest-xdist\",\n]\n\n\nargs = dict(\n name=\"aiohttp\",\n version=version,\n description=\"Async http client/server framework (asyncio)\",\n long_description=read(\"README.rst\"),\n long_description_content_type=\"text/x-rst\",\n classifiers=[\n \"License :: OSI Approved :: Apache Software License\",\n \"Intended Audience :: Developers\",\n \"Programming Language :: Python\",\n \"Programming Language :: Python :: 3\",\n \"Programming Language :: Python :: 3.6\",\n \"Programming Language :: Python :: 3.7\",\n \"Programming Language :: Python :: 3.8\",\n \"Programming Language :: Python :: 3.9\",\n \"Programming Language :: Python :: 3.10\",\n \"Development Status :: 5 - Production/Stable\",\n \"Operating System :: POSIX\",\n \"Operating System :: MacOS :: MacOS X\",\n \"Operating System :: Microsoft :: Windows\",\n \"Topic :: Internet :: WWW/HTTP\",\n \"Framework :: AsyncIO\",\n ],\n author=\"Nikolay Kim\",\n author_email=\"[email protected]\",\n maintainer=\", \".join(\n (\n \"Nikolay Kim <[email protected]>\",\n \"Andrew Svetlov <[email protected]>\",\n )\n ),\n maintainer_email=\"[email protected]\",\n url=\"https://github.com/aio-libs/aiohttp\",\n project_urls={\n \"Chat: Gitter\": \"https://gitter.im/aio-libs/Lobby\",\n \"CI: GitHub Actions\": \"https://github.com/aio-libs/aiohttp/actions?query=workflow%3ACI\", # noqa\n \"Coverage: codecov\": \"https://codecov.io/github/aio-libs/aiohttp\",\n \"Docs: RTD\": \"https://docs.aiohttp.org\",\n \"GitHub: issues\": \"https://github.com/aio-libs/aiohttp/issues\",\n \"GitHub: repo\": \"https://github.com/aio-libs/aiohttp\",\n },\n license=\"Apache 2\",\n packages=[\"aiohttp\"],\n python_requires=\">=3.6\",\n install_requires=install_requires,\n extras_require={\n \"speedups\": [\n \"aiodns\",\n \"Brotli\",\n \"cchardet\",\n ],\n },\n tests_require=tests_require,\n setup_requires=pytest_runner,\n include_package_data=True,\n ext_modules=extensions,\n cmdclass=dict(build_ext=ve_build_ext),\n)\n\ntry:\n setup(**args)\nexcept BuildFailed:\n print(\"************************************************************\")\n print(\"Cannot compile C accelerator module, use pure python version\")\n print(\"************************************************************\")\n del args[\"ext_modules\"]\n del args[\"cmdclass\"]\n setup(**args)\n", "path": "setup.py"}], "after_files": [{"content": "import pathlib\nimport re\nimport sys\nfrom distutils.command.build_ext import build_ext\nfrom distutils.errors import CCompilerError, DistutilsExecError, DistutilsPlatformError\n\nfrom setuptools import Extension, setup\n\nif sys.version_info < (3, 6):\n raise RuntimeError(\"aiohttp 3.7+ requires Python 3.6+\")\n\nhere = pathlib.Path(__file__).parent\n\nif (here / \".git\").exists() and not (here / \"vendor/http-parser/README.md\").exists():\n print(\"Install submodules when building from git clone\", file=sys.stderr)\n print(\"Hint:\", file=sys.stderr)\n print(\" git submodule update --init\", file=sys.stderr)\n sys.exit(2)\n\n\n# NOTE: makefile cythonizes all Cython modules\n\nextensions = [\n Extension(\"aiohttp._websocket\", [\"aiohttp/_websocket.c\"]),\n Extension(\n \"aiohttp._http_parser\",\n [\n \"aiohttp/_http_parser.c\",\n \"vendor/http-parser/http_parser.c\",\n \"aiohttp/_find_header.c\",\n ],\n define_macros=[(\"HTTP_PARSER_STRICT\", 0)],\n ),\n Extension(\"aiohttp._helpers\", [\"aiohttp/_helpers.c\"]),\n Extension(\"aiohttp._http_writer\", [\"aiohttp/_http_writer.c\"]),\n]\n\n\nclass BuildFailed(Exception):\n pass\n\n\nclass ve_build_ext(build_ext):\n # This class allows C extension building to fail.\n\n def run(self):\n try:\n build_ext.run(self)\n except (DistutilsPlatformError, FileNotFoundError):\n raise BuildFailed()\n\n def build_extension(self, ext):\n try:\n build_ext.build_extension(self, ext)\n except (CCompilerError, DistutilsExecError, DistutilsPlatformError, ValueError):\n raise BuildFailed()\n\n\ntxt = (here / \"aiohttp\" / \"__init__.py\").read_text(\"utf-8\")\ntry:\n version = re.findall(r'^__version__ = \"([^\"]+)\"\\r?$', txt, re.M)[0]\nexcept IndexError:\n raise RuntimeError(\"Unable to determine version.\")\n\ninstall_requires = [\n \"attrs>=17.3.0\",\n \"charset-normalizer>=2.0,<3.0\",\n \"multidict>=4.5,<7.0\",\n \"async_timeout>=4.0.0a3,<5.0\",\n 'asynctest==0.13.0; python_version<\"3.8\"',\n \"yarl>=1.0,<2.0\",\n 'idna-ssl>=1.0; python_version<\"3.7\"',\n 'typing_extensions>=3.7.4; python_version<\"3.8\"',\n \"frozenlist>=1.1.1\",\n \"aiosignal>=1.1.2\",\n]\n\n\ndef read(f):\n return (here / f).read_text(\"utf-8\").strip()\n\n\nargs = dict(\n name=\"aiohttp\",\n version=version,\n description=\"Async http client/server framework (asyncio)\",\n long_description=read(\"README.rst\"),\n long_description_content_type=\"text/x-rst\",\n classifiers=[\n \"License :: OSI Approved :: Apache Software License\",\n \"Intended Audience :: Developers\",\n \"Programming Language :: Python\",\n \"Programming Language :: Python :: 3\",\n \"Programming Language :: Python :: 3.6\",\n \"Programming Language :: Python :: 3.7\",\n \"Programming Language :: Python :: 3.8\",\n \"Programming Language :: Python :: 3.9\",\n \"Programming Language :: Python :: 3.10\",\n \"Development Status :: 5 - Production/Stable\",\n \"Operating System :: POSIX\",\n \"Operating System :: MacOS :: MacOS X\",\n \"Operating System :: Microsoft :: Windows\",\n \"Topic :: Internet :: WWW/HTTP\",\n \"Framework :: AsyncIO\",\n ],\n author=\"Nikolay Kim\",\n author_email=\"[email protected]\",\n maintainer=\", \".join(\n (\n \"Nikolay Kim <[email protected]>\",\n \"Andrew Svetlov <[email protected]>\",\n )\n ),\n maintainer_email=\"[email protected]\",\n url=\"https://github.com/aio-libs/aiohttp\",\n project_urls={\n \"Chat: Gitter\": \"https://gitter.im/aio-libs/Lobby\",\n \"CI: GitHub Actions\": \"https://github.com/aio-libs/aiohttp/actions?query=workflow%3ACI\", # noqa\n \"Coverage: codecov\": \"https://codecov.io/github/aio-libs/aiohttp\",\n \"Docs: RTD\": \"https://docs.aiohttp.org\",\n \"GitHub: issues\": \"https://github.com/aio-libs/aiohttp/issues\",\n \"GitHub: repo\": \"https://github.com/aio-libs/aiohttp\",\n },\n license=\"Apache 2\",\n packages=[\"aiohttp\"],\n python_requires=\">=3.6\",\n install_requires=install_requires,\n extras_require={\n \"speedups\": [\n \"aiodns\",\n \"Brotli\",\n \"cchardet\",\n ],\n },\n include_package_data=True,\n ext_modules=extensions,\n cmdclass=dict(build_ext=ve_build_ext),\n)\n\ntry:\n setup(**args)\nexcept BuildFailed:\n print(\"************************************************************\")\n print(\"Cannot compile C accelerator module, use pure python version\")\n print(\"************************************************************\")\n del args[\"ext_modules\"]\n del args[\"cmdclass\"]\n setup(**args)\n", "path": "setup.py"}]} | 2,295 | 204 |
gh_patches_debug_20065 | rasdani/github-patches | git_diff | zigpy__zha-device-handlers-1287 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
[Device Support Request] New manufacturerName for DDS238-2 Zigbee
In one chat I got information that from December 2021 this device now is sending with a new manufacturerName attribute: _TZE200_ewxhg6o9. The new version also contains the fix the switch issue (see the details in the #994 thread) and maybe something else, like EnergyFactor attribute.
I can update the quirck but I have no physically the new version of the DDS238-2 device to test, so if someone has the new version of device and can support me with the testing, I can help with the contribution of this update.
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `zhaquirks/tuya/ts0601_din_power.py`
Content:
```
1 """Tuya Din Power Meter."""
2 from zigpy.profiles import zha
3 import zigpy.types as t
4 from zigpy.zcl.clusters.general import Basic, Groups, Ota, Scenes, Time
5 from zigpy.zcl.clusters.homeautomation import ElectricalMeasurement
6 from zigpy.zcl.clusters.smartenergy import Metering
7
8 from zhaquirks import Bus, LocalDataCluster
9 from zhaquirks.const import (
10 DEVICE_TYPE,
11 ENDPOINTS,
12 INPUT_CLUSTERS,
13 MODELS_INFO,
14 OUTPUT_CLUSTERS,
15 PROFILE_ID,
16 )
17 from zhaquirks.tuya import TuyaManufClusterAttributes, TuyaOnOff, TuyaSwitch
18
19 TUYA_TOTAL_ENERGY_ATTR = 0x0211
20 TUYA_CURRENT_ATTR = 0x0212
21 TUYA_POWER_ATTR = 0x0213
22 TUYA_VOLTAGE_ATTR = 0x0214
23 TUYA_DIN_SWITCH_ATTR = 0x0101
24
25 SWITCH_EVENT = "switch_event"
26
27
28 class TuyaManufClusterDinPower(TuyaManufClusterAttributes):
29 """Manufacturer Specific Cluster of the Tuya Power Meter device."""
30
31 manufacturer_attributes = {
32 TUYA_TOTAL_ENERGY_ATTR: ("energy", t.uint16_t),
33 TUYA_CURRENT_ATTR: ("current", t.int16s),
34 TUYA_POWER_ATTR: ("power", t.uint16_t),
35 TUYA_VOLTAGE_ATTR: ("voltage", t.uint16_t),
36 TUYA_DIN_SWITCH_ATTR: ("switch", t.uint8_t),
37 }
38
39 def _update_attribute(self, attrid, value):
40 super()._update_attribute(attrid, value)
41 if attrid == TUYA_TOTAL_ENERGY_ATTR:
42 self.endpoint.smartenergy_metering.energy_reported(value / 100)
43 elif attrid == TUYA_CURRENT_ATTR:
44 self.endpoint.electrical_measurement.current_reported(value)
45 elif attrid == TUYA_POWER_ATTR:
46 self.endpoint.electrical_measurement.power_reported(value / 10)
47 elif attrid == TUYA_VOLTAGE_ATTR:
48 self.endpoint.electrical_measurement.voltage_reported(value / 10)
49 elif attrid == TUYA_DIN_SWITCH_ATTR:
50 self.endpoint.device.switch_bus.listener_event(SWITCH_EVENT, attrid, value)
51
52
53 class TuyaPowerMeasurement(LocalDataCluster, ElectricalMeasurement):
54 """Custom class for power, voltage and current measurement."""
55
56 cluster_id = ElectricalMeasurement.cluster_id
57
58 POWER_ID = 0x050B
59 VOLTAGE_ID = 0x0505
60 CURRENT_ID = 0x0508
61
62 AC_CURRENT_MULTIPLIER = 0x0602
63 AC_CURRENT_DIVISOR = 0x0603
64
65 _CONSTANT_ATTRIBUTES = {AC_CURRENT_MULTIPLIER: 1, AC_CURRENT_DIVISOR: 1000}
66
67 def voltage_reported(self, value):
68 """Voltage reported."""
69 self._update_attribute(self.VOLTAGE_ID, value)
70
71 def power_reported(self, value):
72 """Power reported."""
73 self._update_attribute(self.POWER_ID, value)
74
75 def current_reported(self, value):
76 """Ampers reported."""
77 self._update_attribute(self.CURRENT_ID, value)
78
79
80 class TuyaElectricalMeasurement(LocalDataCluster, Metering):
81 """Custom class for total energy measurement."""
82
83 cluster_id = Metering.cluster_id
84 CURRENT_ID = 0x0000
85 POWER_WATT = 0x0000
86
87 """Setting unit of measurement."""
88 _CONSTANT_ATTRIBUTES = {0x0300: POWER_WATT}
89
90 def energy_reported(self, value):
91 """Summation Energy reported."""
92 self._update_attribute(self.CURRENT_ID, value)
93
94
95 class TuyaPowerMeter(TuyaSwitch):
96 """Tuya power meter device."""
97
98 def __init__(self, *args, **kwargs):
99 """Init device."""
100 self.switch_bus = Bus()
101 super().__init__(*args, **kwargs)
102
103 signature = {
104 # "node_descriptor": "<NodeDescriptor byte1=1 byte2=64 mac_capability_flags=142 manufacturer_code=4098
105 # maximum_buffer_size=82 maximum_incoming_transfer_size=82 server_mask=11264
106 # maximum_outgoing_transfer_size=82 descriptor_capability_field=0>",
107 # device_version=1
108 # input_clusters=[0x0000, 0x0004, 0x0005, 0xef00]
109 # output_clusters=[0x000a, 0x0019]
110 MODELS_INFO: [
111 ("_TZE200_byzdayie", "TS0601"),
112 ],
113 ENDPOINTS: {
114 # <SimpleDescriptor endpoint=1 profile=260 device_type=51
115 # device_version=1
116 # input_clusters=[0, 4, 5, 61184]
117 # output_clusters=[10, 25]>
118 1: {
119 PROFILE_ID: zha.PROFILE_ID,
120 DEVICE_TYPE: zha.DeviceType.SMART_PLUG,
121 INPUT_CLUSTERS: [
122 Basic.cluster_id,
123 Groups.cluster_id,
124 Scenes.cluster_id,
125 TuyaManufClusterAttributes.cluster_id,
126 ],
127 OUTPUT_CLUSTERS: [Time.cluster_id, Ota.cluster_id],
128 }
129 },
130 }
131
132 replacement = {
133 ENDPOINTS: {
134 1: {
135 PROFILE_ID: zha.PROFILE_ID,
136 DEVICE_TYPE: zha.DeviceType.SMART_PLUG,
137 INPUT_CLUSTERS: [
138 Basic.cluster_id,
139 Groups.cluster_id,
140 Scenes.cluster_id,
141 TuyaManufClusterDinPower,
142 TuyaPowerMeasurement,
143 TuyaElectricalMeasurement,
144 TuyaOnOff,
145 ],
146 OUTPUT_CLUSTERS: [Time.cluster_id, Ota.cluster_id],
147 }
148 }
149 }
150
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/zhaquirks/tuya/ts0601_din_power.py b/zhaquirks/tuya/ts0601_din_power.py
--- a/zhaquirks/tuya/ts0601_din_power.py
+++ b/zhaquirks/tuya/ts0601_din_power.py
@@ -47,7 +47,9 @@
elif attrid == TUYA_VOLTAGE_ATTR:
self.endpoint.electrical_measurement.voltage_reported(value / 10)
elif attrid == TUYA_DIN_SWITCH_ATTR:
- self.endpoint.device.switch_bus.listener_event(SWITCH_EVENT, attrid, value)
+ self.endpoint.device.switch_bus.listener_event(
+ SWITCH_EVENT, self.endpoint.endpoint_id, value
+ )
class TuyaPowerMeasurement(LocalDataCluster, ElectricalMeasurement):
@@ -109,6 +111,7 @@
# output_clusters=[0x000a, 0x0019]
MODELS_INFO: [
("_TZE200_byzdayie", "TS0601"),
+ ("_TZE200_ewxhg6o9", "TS0601"),
],
ENDPOINTS: {
# <SimpleDescriptor endpoint=1 profile=260 device_type=51
| {"golden_diff": "diff --git a/zhaquirks/tuya/ts0601_din_power.py b/zhaquirks/tuya/ts0601_din_power.py\n--- a/zhaquirks/tuya/ts0601_din_power.py\n+++ b/zhaquirks/tuya/ts0601_din_power.py\n@@ -47,7 +47,9 @@\n elif attrid == TUYA_VOLTAGE_ATTR:\n self.endpoint.electrical_measurement.voltage_reported(value / 10)\n elif attrid == TUYA_DIN_SWITCH_ATTR:\n- self.endpoint.device.switch_bus.listener_event(SWITCH_EVENT, attrid, value)\n+ self.endpoint.device.switch_bus.listener_event(\n+ SWITCH_EVENT, self.endpoint.endpoint_id, value\n+ )\n \n \n class TuyaPowerMeasurement(LocalDataCluster, ElectricalMeasurement):\n@@ -109,6 +111,7 @@\n # output_clusters=[0x000a, 0x0019]\n MODELS_INFO: [\n (\"_TZE200_byzdayie\", \"TS0601\"),\n+ (\"_TZE200_ewxhg6o9\", \"TS0601\"),\n ],\n ENDPOINTS: {\n # <SimpleDescriptor endpoint=1 profile=260 device_type=51\n", "issue": "[Device Support Request] New manufacturerName for DDS238-2 Zigbee\nIn one chat I got information that from December 2021 this device now is sending with a new manufacturerName attribute: _TZE200_ewxhg6o9. The new version also contains the fix the switch issue (see the details in the #994 thread) and maybe something else, like EnergyFactor attribute.\r\nI can update the quirck but I have no physically the new version of the DDS238-2 device to test, so if someone has the new version of device and can support me with the testing, I can help with the contribution of this update.\n", "before_files": [{"content": "\"\"\"Tuya Din Power Meter.\"\"\"\nfrom zigpy.profiles import zha\nimport zigpy.types as t\nfrom zigpy.zcl.clusters.general import Basic, Groups, Ota, Scenes, Time\nfrom zigpy.zcl.clusters.homeautomation import ElectricalMeasurement\nfrom zigpy.zcl.clusters.smartenergy import Metering\n\nfrom zhaquirks import Bus, LocalDataCluster\nfrom zhaquirks.const import (\n DEVICE_TYPE,\n ENDPOINTS,\n INPUT_CLUSTERS,\n MODELS_INFO,\n OUTPUT_CLUSTERS,\n PROFILE_ID,\n)\nfrom zhaquirks.tuya import TuyaManufClusterAttributes, TuyaOnOff, TuyaSwitch\n\nTUYA_TOTAL_ENERGY_ATTR = 0x0211\nTUYA_CURRENT_ATTR = 0x0212\nTUYA_POWER_ATTR = 0x0213\nTUYA_VOLTAGE_ATTR = 0x0214\nTUYA_DIN_SWITCH_ATTR = 0x0101\n\nSWITCH_EVENT = \"switch_event\"\n\n\nclass TuyaManufClusterDinPower(TuyaManufClusterAttributes):\n \"\"\"Manufacturer Specific Cluster of the Tuya Power Meter device.\"\"\"\n\n manufacturer_attributes = {\n TUYA_TOTAL_ENERGY_ATTR: (\"energy\", t.uint16_t),\n TUYA_CURRENT_ATTR: (\"current\", t.int16s),\n TUYA_POWER_ATTR: (\"power\", t.uint16_t),\n TUYA_VOLTAGE_ATTR: (\"voltage\", t.uint16_t),\n TUYA_DIN_SWITCH_ATTR: (\"switch\", t.uint8_t),\n }\n\n def _update_attribute(self, attrid, value):\n super()._update_attribute(attrid, value)\n if attrid == TUYA_TOTAL_ENERGY_ATTR:\n self.endpoint.smartenergy_metering.energy_reported(value / 100)\n elif attrid == TUYA_CURRENT_ATTR:\n self.endpoint.electrical_measurement.current_reported(value)\n elif attrid == TUYA_POWER_ATTR:\n self.endpoint.electrical_measurement.power_reported(value / 10)\n elif attrid == TUYA_VOLTAGE_ATTR:\n self.endpoint.electrical_measurement.voltage_reported(value / 10)\n elif attrid == TUYA_DIN_SWITCH_ATTR:\n self.endpoint.device.switch_bus.listener_event(SWITCH_EVENT, attrid, value)\n\n\nclass TuyaPowerMeasurement(LocalDataCluster, ElectricalMeasurement):\n \"\"\"Custom class for power, voltage and current measurement.\"\"\"\n\n cluster_id = ElectricalMeasurement.cluster_id\n\n POWER_ID = 0x050B\n VOLTAGE_ID = 0x0505\n CURRENT_ID = 0x0508\n\n AC_CURRENT_MULTIPLIER = 0x0602\n AC_CURRENT_DIVISOR = 0x0603\n\n _CONSTANT_ATTRIBUTES = {AC_CURRENT_MULTIPLIER: 1, AC_CURRENT_DIVISOR: 1000}\n\n def voltage_reported(self, value):\n \"\"\"Voltage reported.\"\"\"\n self._update_attribute(self.VOLTAGE_ID, value)\n\n def power_reported(self, value):\n \"\"\"Power reported.\"\"\"\n self._update_attribute(self.POWER_ID, value)\n\n def current_reported(self, value):\n \"\"\"Ampers reported.\"\"\"\n self._update_attribute(self.CURRENT_ID, value)\n\n\nclass TuyaElectricalMeasurement(LocalDataCluster, Metering):\n \"\"\"Custom class for total energy measurement.\"\"\"\n\n cluster_id = Metering.cluster_id\n CURRENT_ID = 0x0000\n POWER_WATT = 0x0000\n\n \"\"\"Setting unit of measurement.\"\"\"\n _CONSTANT_ATTRIBUTES = {0x0300: POWER_WATT}\n\n def energy_reported(self, value):\n \"\"\"Summation Energy reported.\"\"\"\n self._update_attribute(self.CURRENT_ID, value)\n\n\nclass TuyaPowerMeter(TuyaSwitch):\n \"\"\"Tuya power meter device.\"\"\"\n\n def __init__(self, *args, **kwargs):\n \"\"\"Init device.\"\"\"\n self.switch_bus = Bus()\n super().__init__(*args, **kwargs)\n\n signature = {\n # \"node_descriptor\": \"<NodeDescriptor byte1=1 byte2=64 mac_capability_flags=142 manufacturer_code=4098\n # maximum_buffer_size=82 maximum_incoming_transfer_size=82 server_mask=11264\n # maximum_outgoing_transfer_size=82 descriptor_capability_field=0>\",\n # device_version=1\n # input_clusters=[0x0000, 0x0004, 0x0005, 0xef00]\n # output_clusters=[0x000a, 0x0019]\n MODELS_INFO: [\n (\"_TZE200_byzdayie\", \"TS0601\"),\n ],\n ENDPOINTS: {\n # <SimpleDescriptor endpoint=1 profile=260 device_type=51\n # device_version=1\n # input_clusters=[0, 4, 5, 61184]\n # output_clusters=[10, 25]>\n 1: {\n PROFILE_ID: zha.PROFILE_ID,\n DEVICE_TYPE: zha.DeviceType.SMART_PLUG,\n INPUT_CLUSTERS: [\n Basic.cluster_id,\n Groups.cluster_id,\n Scenes.cluster_id,\n TuyaManufClusterAttributes.cluster_id,\n ],\n OUTPUT_CLUSTERS: [Time.cluster_id, Ota.cluster_id],\n }\n },\n }\n\n replacement = {\n ENDPOINTS: {\n 1: {\n PROFILE_ID: zha.PROFILE_ID,\n DEVICE_TYPE: zha.DeviceType.SMART_PLUG,\n INPUT_CLUSTERS: [\n Basic.cluster_id,\n Groups.cluster_id,\n Scenes.cluster_id,\n TuyaManufClusterDinPower,\n TuyaPowerMeasurement,\n TuyaElectricalMeasurement,\n TuyaOnOff,\n ],\n OUTPUT_CLUSTERS: [Time.cluster_id, Ota.cluster_id],\n }\n }\n }\n", "path": "zhaquirks/tuya/ts0601_din_power.py"}], "after_files": [{"content": "\"\"\"Tuya Din Power Meter.\"\"\"\nfrom zigpy.profiles import zha\nimport zigpy.types as t\nfrom zigpy.zcl.clusters.general import Basic, Groups, Ota, Scenes, Time\nfrom zigpy.zcl.clusters.homeautomation import ElectricalMeasurement\nfrom zigpy.zcl.clusters.smartenergy import Metering\n\nfrom zhaquirks import Bus, LocalDataCluster\nfrom zhaquirks.const import (\n DEVICE_TYPE,\n ENDPOINTS,\n INPUT_CLUSTERS,\n MODELS_INFO,\n OUTPUT_CLUSTERS,\n PROFILE_ID,\n)\nfrom zhaquirks.tuya import TuyaManufClusterAttributes, TuyaOnOff, TuyaSwitch\n\nTUYA_TOTAL_ENERGY_ATTR = 0x0211\nTUYA_CURRENT_ATTR = 0x0212\nTUYA_POWER_ATTR = 0x0213\nTUYA_VOLTAGE_ATTR = 0x0214\nTUYA_DIN_SWITCH_ATTR = 0x0101\n\nSWITCH_EVENT = \"switch_event\"\n\n\nclass TuyaManufClusterDinPower(TuyaManufClusterAttributes):\n \"\"\"Manufacturer Specific Cluster of the Tuya Power Meter device.\"\"\"\n\n manufacturer_attributes = {\n TUYA_TOTAL_ENERGY_ATTR: (\"energy\", t.uint16_t),\n TUYA_CURRENT_ATTR: (\"current\", t.int16s),\n TUYA_POWER_ATTR: (\"power\", t.uint16_t),\n TUYA_VOLTAGE_ATTR: (\"voltage\", t.uint16_t),\n TUYA_DIN_SWITCH_ATTR: (\"switch\", t.uint8_t),\n }\n\n def _update_attribute(self, attrid, value):\n super()._update_attribute(attrid, value)\n if attrid == TUYA_TOTAL_ENERGY_ATTR:\n self.endpoint.smartenergy_metering.energy_reported(value / 100)\n elif attrid == TUYA_CURRENT_ATTR:\n self.endpoint.electrical_measurement.current_reported(value)\n elif attrid == TUYA_POWER_ATTR:\n self.endpoint.electrical_measurement.power_reported(value / 10)\n elif attrid == TUYA_VOLTAGE_ATTR:\n self.endpoint.electrical_measurement.voltage_reported(value / 10)\n elif attrid == TUYA_DIN_SWITCH_ATTR:\n self.endpoint.device.switch_bus.listener_event(\n SWITCH_EVENT, self.endpoint.endpoint_id, value\n )\n\n\nclass TuyaPowerMeasurement(LocalDataCluster, ElectricalMeasurement):\n \"\"\"Custom class for power, voltage and current measurement.\"\"\"\n\n cluster_id = ElectricalMeasurement.cluster_id\n\n POWER_ID = 0x050B\n VOLTAGE_ID = 0x0505\n CURRENT_ID = 0x0508\n\n AC_CURRENT_MULTIPLIER = 0x0602\n AC_CURRENT_DIVISOR = 0x0603\n\n _CONSTANT_ATTRIBUTES = {AC_CURRENT_MULTIPLIER: 1, AC_CURRENT_DIVISOR: 1000}\n\n def voltage_reported(self, value):\n \"\"\"Voltage reported.\"\"\"\n self._update_attribute(self.VOLTAGE_ID, value)\n\n def power_reported(self, value):\n \"\"\"Power reported.\"\"\"\n self._update_attribute(self.POWER_ID, value)\n\n def current_reported(self, value):\n \"\"\"Ampers reported.\"\"\"\n self._update_attribute(self.CURRENT_ID, value)\n\n\nclass TuyaElectricalMeasurement(LocalDataCluster, Metering):\n \"\"\"Custom class for total energy measurement.\"\"\"\n\n cluster_id = Metering.cluster_id\n CURRENT_ID = 0x0000\n POWER_WATT = 0x0000\n\n \"\"\"Setting unit of measurement.\"\"\"\n _CONSTANT_ATTRIBUTES = {0x0300: POWER_WATT}\n\n def energy_reported(self, value):\n \"\"\"Summation Energy reported.\"\"\"\n self._update_attribute(self.CURRENT_ID, value)\n\n\nclass TuyaPowerMeter(TuyaSwitch):\n \"\"\"Tuya power meter device.\"\"\"\n\n def __init__(self, *args, **kwargs):\n \"\"\"Init device.\"\"\"\n self.switch_bus = Bus()\n super().__init__(*args, **kwargs)\n\n signature = {\n # \"node_descriptor\": \"<NodeDescriptor byte1=1 byte2=64 mac_capability_flags=142 manufacturer_code=4098\n # maximum_buffer_size=82 maximum_incoming_transfer_size=82 server_mask=11264\n # maximum_outgoing_transfer_size=82 descriptor_capability_field=0>\",\n # device_version=1\n # input_clusters=[0x0000, 0x0004, 0x0005, 0xef00]\n # output_clusters=[0x000a, 0x0019]\n MODELS_INFO: [\n (\"_TZE200_byzdayie\", \"TS0601\"),\n (\"_TZE200_ewxhg6o9\", \"TS0601\"),\n ],\n ENDPOINTS: {\n # <SimpleDescriptor endpoint=1 profile=260 device_type=51\n # device_version=1\n # input_clusters=[0, 4, 5, 61184]\n # output_clusters=[10, 25]>\n 1: {\n PROFILE_ID: zha.PROFILE_ID,\n DEVICE_TYPE: zha.DeviceType.SMART_PLUG,\n INPUT_CLUSTERS: [\n Basic.cluster_id,\n Groups.cluster_id,\n Scenes.cluster_id,\n TuyaManufClusterAttributes.cluster_id,\n ],\n OUTPUT_CLUSTERS: [Time.cluster_id, Ota.cluster_id],\n }\n },\n }\n\n replacement = {\n ENDPOINTS: {\n 1: {\n PROFILE_ID: zha.PROFILE_ID,\n DEVICE_TYPE: zha.DeviceType.SMART_PLUG,\n INPUT_CLUSTERS: [\n Basic.cluster_id,\n Groups.cluster_id,\n Scenes.cluster_id,\n TuyaManufClusterDinPower,\n TuyaPowerMeasurement,\n TuyaElectricalMeasurement,\n TuyaOnOff,\n ],\n OUTPUT_CLUSTERS: [Time.cluster_id, Ota.cluster_id],\n }\n }\n }\n", "path": "zhaquirks/tuya/ts0601_din_power.py"}]} | 2,086 | 293 |
gh_patches_debug_21733 | rasdani/github-patches | git_diff | getredash__redash-3619 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Support for Presto username and password
Currently the Presto query runner supports username only. We should support password as well.
This probably requires upgrading the PyHive library.
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `redash/query_runner/presto.py`
Content:
```
1 from redash.query_runner import *
2 from redash.utils import json_dumps, json_loads
3
4 import logging
5 logger = logging.getLogger(__name__)
6
7 from collections import defaultdict
8
9 try:
10 from pyhive import presto
11 from pyhive.exc import DatabaseError
12 enabled = True
13
14 except ImportError:
15 enabled = False
16
17 PRESTO_TYPES_MAPPING = {
18 "integer": TYPE_INTEGER,
19 "tinyint": TYPE_INTEGER,
20 "smallint": TYPE_INTEGER,
21 "long": TYPE_INTEGER,
22 "bigint": TYPE_INTEGER,
23 "float": TYPE_FLOAT,
24 "double": TYPE_FLOAT,
25 "boolean": TYPE_BOOLEAN,
26 "string": TYPE_STRING,
27 "varchar": TYPE_STRING,
28 "date": TYPE_DATE,
29 }
30
31
32 class Presto(BaseQueryRunner):
33 noop_query = 'SHOW TABLES'
34
35 @classmethod
36 def configuration_schema(cls):
37 return {
38 'type': 'object',
39 'properties': {
40 'host': {
41 'type': 'string'
42 },
43 'protocol': {
44 'type': 'string',
45 'default': 'http'
46 },
47 'port': {
48 'type': 'number'
49 },
50 'schema': {
51 'type': 'string'
52 },
53 'catalog': {
54 'type': 'string'
55 },
56 'username': {
57 'type': 'string'
58 },
59 },
60 'order': ['host', 'protocol', 'port', 'username', 'schema', 'catalog'],
61 'required': ['host']
62 }
63
64 @classmethod
65 def enabled(cls):
66 return enabled
67
68 @classmethod
69 def type(cls):
70 return "presto"
71
72 def get_schema(self, get_stats=False):
73 schema = {}
74 query = """
75 SELECT table_schema, table_name, column_name
76 FROM information_schema.columns
77 WHERE table_schema NOT IN ('pg_catalog', 'information_schema')
78 """
79
80 results, error = self.run_query(query, None)
81
82 if error is not None:
83 raise Exception("Failed getting schema.")
84
85 results = json_loads(results)
86
87 for row in results['rows']:
88 table_name = '{}.{}'.format(row['table_schema'], row['table_name'])
89
90 if table_name not in schema:
91 schema[table_name] = {'name': table_name, 'columns': []}
92
93 schema[table_name]['columns'].append(row['column_name'])
94
95 return schema.values()
96
97 def run_query(self, query, user):
98 connection = presto.connect(
99 host=self.configuration.get('host', ''),
100 port=self.configuration.get('port', 8080),
101 protocol=self.configuration.get('protocol', 'http'),
102 username=self.configuration.get('username', 'redash'),
103 catalog=self.configuration.get('catalog', 'hive'),
104 schema=self.configuration.get('schema', 'default'))
105
106 cursor = connection.cursor()
107
108
109 try:
110 cursor.execute(query)
111 column_tuples = [(i[0], PRESTO_TYPES_MAPPING.get(i[1], None)) for i in cursor.description]
112 columns = self.fetch_columns(column_tuples)
113 rows = [dict(zip(([c['name'] for c in columns]), r)) for i, r in enumerate(cursor.fetchall())]
114 data = {'columns': columns, 'rows': rows}
115 json_data = json_dumps(data)
116 error = None
117 except DatabaseError as db:
118 json_data = None
119 default_message = 'Unspecified DatabaseError: {0}'.format(db.message)
120 if isinstance(db.message, dict):
121 message = db.message.get('failureInfo', {'message', None}).get('message')
122 else:
123 message = None
124 error = default_message if message is None else message
125 except (KeyboardInterrupt, InterruptException) as e:
126 cursor.cancel()
127 error = "Query cancelled by user."
128 json_data = None
129 except Exception as ex:
130 json_data = None
131 error = ex.message
132 if not isinstance(error, basestring):
133 error = unicode(error)
134
135 return json_data, error
136
137 register(Presto)
138
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/redash/query_runner/presto.py b/redash/query_runner/presto.py
--- a/redash/query_runner/presto.py
+++ b/redash/query_runner/presto.py
@@ -56,8 +56,11 @@
'username': {
'type': 'string'
},
+ 'password': {
+ 'type': 'string'
+ },
},
- 'order': ['host', 'protocol', 'port', 'username', 'schema', 'catalog'],
+ 'order': ['host', 'protocol', 'port', 'username', 'password', 'schema', 'catalog'],
'required': ['host']
}
@@ -100,6 +103,7 @@
port=self.configuration.get('port', 8080),
protocol=self.configuration.get('protocol', 'http'),
username=self.configuration.get('username', 'redash'),
+ password=self.configuration.get('password', ''),
catalog=self.configuration.get('catalog', 'hive'),
schema=self.configuration.get('schema', 'default'))
| {"golden_diff": "diff --git a/redash/query_runner/presto.py b/redash/query_runner/presto.py\n--- a/redash/query_runner/presto.py\n+++ b/redash/query_runner/presto.py\n@@ -56,8 +56,11 @@\n 'username': {\n 'type': 'string'\n },\n+ 'password': {\n+ 'type': 'string'\n+ },\n },\n- 'order': ['host', 'protocol', 'port', 'username', 'schema', 'catalog'],\n+ 'order': ['host', 'protocol', 'port', 'username', 'password', 'schema', 'catalog'],\n 'required': ['host']\n }\n \n@@ -100,6 +103,7 @@\n port=self.configuration.get('port', 8080),\n protocol=self.configuration.get('protocol', 'http'),\n username=self.configuration.get('username', 'redash'),\n+ password=self.configuration.get('password', ''),\n catalog=self.configuration.get('catalog', 'hive'),\n schema=self.configuration.get('schema', 'default'))\n", "issue": "Support for Presto username and password\nCurrently the Presto query runner supports username only. We should support password as well.\r\n\r\nThis probably requires upgrading the PyHive library.\n", "before_files": [{"content": "from redash.query_runner import *\nfrom redash.utils import json_dumps, json_loads\n\nimport logging\nlogger = logging.getLogger(__name__)\n\nfrom collections import defaultdict\n\ntry:\n from pyhive import presto\n from pyhive.exc import DatabaseError\n enabled = True\n\nexcept ImportError:\n enabled = False\n\nPRESTO_TYPES_MAPPING = {\n \"integer\": TYPE_INTEGER,\n \"tinyint\": TYPE_INTEGER,\n \"smallint\": TYPE_INTEGER,\n \"long\": TYPE_INTEGER,\n \"bigint\": TYPE_INTEGER,\n \"float\": TYPE_FLOAT,\n \"double\": TYPE_FLOAT,\n \"boolean\": TYPE_BOOLEAN,\n \"string\": TYPE_STRING,\n \"varchar\": TYPE_STRING,\n \"date\": TYPE_DATE,\n}\n\n\nclass Presto(BaseQueryRunner):\n noop_query = 'SHOW TABLES'\n\n @classmethod\n def configuration_schema(cls):\n return {\n 'type': 'object',\n 'properties': {\n 'host': {\n 'type': 'string'\n },\n 'protocol': {\n 'type': 'string',\n 'default': 'http'\n },\n 'port': {\n 'type': 'number'\n },\n 'schema': {\n 'type': 'string'\n },\n 'catalog': {\n 'type': 'string'\n },\n 'username': {\n 'type': 'string'\n },\n },\n 'order': ['host', 'protocol', 'port', 'username', 'schema', 'catalog'],\n 'required': ['host']\n }\n\n @classmethod\n def enabled(cls):\n return enabled\n\n @classmethod\n def type(cls):\n return \"presto\"\n\n def get_schema(self, get_stats=False):\n schema = {}\n query = \"\"\"\n SELECT table_schema, table_name, column_name\n FROM information_schema.columns\n WHERE table_schema NOT IN ('pg_catalog', 'information_schema')\n \"\"\"\n\n results, error = self.run_query(query, None)\n\n if error is not None:\n raise Exception(\"Failed getting schema.\")\n\n results = json_loads(results)\n\n for row in results['rows']:\n table_name = '{}.{}'.format(row['table_schema'], row['table_name'])\n\n if table_name not in schema:\n schema[table_name] = {'name': table_name, 'columns': []}\n\n schema[table_name]['columns'].append(row['column_name'])\n\n return schema.values()\n\n def run_query(self, query, user):\n connection = presto.connect(\n host=self.configuration.get('host', ''),\n port=self.configuration.get('port', 8080),\n protocol=self.configuration.get('protocol', 'http'),\n username=self.configuration.get('username', 'redash'),\n catalog=self.configuration.get('catalog', 'hive'),\n schema=self.configuration.get('schema', 'default'))\n\n cursor = connection.cursor()\n\n\n try:\n cursor.execute(query)\n column_tuples = [(i[0], PRESTO_TYPES_MAPPING.get(i[1], None)) for i in cursor.description]\n columns = self.fetch_columns(column_tuples)\n rows = [dict(zip(([c['name'] for c in columns]), r)) for i, r in enumerate(cursor.fetchall())]\n data = {'columns': columns, 'rows': rows}\n json_data = json_dumps(data)\n error = None\n except DatabaseError as db:\n json_data = None\n default_message = 'Unspecified DatabaseError: {0}'.format(db.message)\n if isinstance(db.message, dict):\n message = db.message.get('failureInfo', {'message', None}).get('message')\n else:\n message = None\n error = default_message if message is None else message\n except (KeyboardInterrupt, InterruptException) as e:\n cursor.cancel()\n error = \"Query cancelled by user.\"\n json_data = None\n except Exception as ex:\n json_data = None\n error = ex.message\n if not isinstance(error, basestring):\n error = unicode(error)\n\n return json_data, error\n\nregister(Presto)\n", "path": "redash/query_runner/presto.py"}], "after_files": [{"content": "from redash.query_runner import *\nfrom redash.utils import json_dumps, json_loads\n\nimport logging\nlogger = logging.getLogger(__name__)\n\nfrom collections import defaultdict\n\ntry:\n from pyhive import presto\n from pyhive.exc import DatabaseError\n enabled = True\n\nexcept ImportError:\n enabled = False\n\nPRESTO_TYPES_MAPPING = {\n \"integer\": TYPE_INTEGER,\n \"tinyint\": TYPE_INTEGER,\n \"smallint\": TYPE_INTEGER,\n \"long\": TYPE_INTEGER,\n \"bigint\": TYPE_INTEGER,\n \"float\": TYPE_FLOAT,\n \"double\": TYPE_FLOAT,\n \"boolean\": TYPE_BOOLEAN,\n \"string\": TYPE_STRING,\n \"varchar\": TYPE_STRING,\n \"date\": TYPE_DATE,\n}\n\n\nclass Presto(BaseQueryRunner):\n noop_query = 'SHOW TABLES'\n\n @classmethod\n def configuration_schema(cls):\n return {\n 'type': 'object',\n 'properties': {\n 'host': {\n 'type': 'string'\n },\n 'protocol': {\n 'type': 'string',\n 'default': 'http'\n },\n 'port': {\n 'type': 'number'\n },\n 'schema': {\n 'type': 'string'\n },\n 'catalog': {\n 'type': 'string'\n },\n 'username': {\n 'type': 'string'\n },\n 'password': {\n 'type': 'string'\n },\n },\n 'order': ['host', 'protocol', 'port', 'username', 'password', 'schema', 'catalog'],\n 'required': ['host']\n }\n\n @classmethod\n def enabled(cls):\n return enabled\n\n @classmethod\n def type(cls):\n return \"presto\"\n\n def get_schema(self, get_stats=False):\n schema = {}\n query = \"\"\"\n SELECT table_schema, table_name, column_name\n FROM information_schema.columns\n WHERE table_schema NOT IN ('pg_catalog', 'information_schema')\n \"\"\"\n\n results, error = self.run_query(query, None)\n\n if error is not None:\n raise Exception(\"Failed getting schema.\")\n\n results = json_loads(results)\n\n for row in results['rows']:\n table_name = '{}.{}'.format(row['table_schema'], row['table_name'])\n\n if table_name not in schema:\n schema[table_name] = {'name': table_name, 'columns': []}\n\n schema[table_name]['columns'].append(row['column_name'])\n\n return schema.values()\n\n def run_query(self, query, user):\n connection = presto.connect(\n host=self.configuration.get('host', ''),\n port=self.configuration.get('port', 8080),\n protocol=self.configuration.get('protocol', 'http'),\n username=self.configuration.get('username', 'redash'),\n password=self.configuration.get('password', ''),\n catalog=self.configuration.get('catalog', 'hive'),\n schema=self.configuration.get('schema', 'default'))\n\n cursor = connection.cursor()\n\n\n try:\n cursor.execute(query)\n column_tuples = [(i[0], PRESTO_TYPES_MAPPING.get(i[1], None)) for i in cursor.description]\n columns = self.fetch_columns(column_tuples)\n rows = [dict(zip(([c['name'] for c in columns]), r)) for i, r in enumerate(cursor.fetchall())]\n data = {'columns': columns, 'rows': rows}\n json_data = json_dumps(data)\n error = None\n except DatabaseError as db:\n json_data = None\n default_message = 'Unspecified DatabaseError: {0}'.format(db.message)\n if isinstance(db.message, dict):\n message = db.message.get('failureInfo', {'message', None}).get('message')\n else:\n message = None\n error = default_message if message is None else message\n except (KeyboardInterrupt, InterruptException) as e:\n cursor.cancel()\n error = \"Query cancelled by user.\"\n json_data = None\n except Exception as ex:\n json_data = None\n error = ex.message\n if not isinstance(error, basestring):\n error = unicode(error)\n\n return json_data, error\n\nregister(Presto)\n", "path": "redash/query_runner/presto.py"}]} | 1,471 | 228 |
gh_patches_debug_21440 | rasdani/github-patches | git_diff | UTNkar__moore-120 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Application drafts in limbo
<!-- Do you want to ask a question? Are you looking for support? The system administrator can help you: [email protected] -->
### Prerequisites
* [x] Put an X between the brackets on this line if you have done all of the
following:
* Reproduced the problem with clear cache.
* (If running the application locally:) Made sure your running the newest version on the development branch
* Checked that your issue isn't already filed: https://github.com/UTNkar/moore/issues
### Description
The problem has not been reproduced but it's the kind of problem that probably will occur any way.
When a user starts an application draft and forgets to submit it the person can not be appointed to the position with the overturn function. If the application for a position has been saved as draft when the application period ends the application becomes stuck in limbo. The group administrator can before the application period has ended see that there is a draft for the person. When the application period ends the draft is not visible among the submitted applications (very resonable). The problem occurs when the administrator wants to appoint that person anyway with the overturn function then an error message occurs saying. "You can not appoint this person since an application has been submitted". This should not be the case, a draft application should not be seen as a submitted application by the overturn function. The user can not see or delete the draft application after the application period has ended.
Quick fix, an application system administrator can access the applications and set the status to submitted.
### Steps to Reproduce
1. Apply to an position and save the application as draft
2. Wait for the application time to end
3. Go to appoint for the given position
4. Try overturn with the user who saved the application as draft
5. Error message occurs
<!-- Please select the appropriate "topic category"/blue and "issue type"/yellow label -->
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `website/involvement/forms.py`
Content:
```
1 from django import forms
2 from django.contrib.auth import get_user_model
3 from django.utils.translation import ugettext_lazy as _
4
5 from involvement.models import Application, Reference
6 from utils.forms import AdvancedModelMultipleChoiceField
7
8
9 class ApplicationForm(forms.ModelForm):
10 class Meta:
11 model = Application
12 exclude = ['position', 'applicant']
13 widgets = {
14 'cover_letter': forms.Textarea(attrs={'style': 'height: 200px',
15 'class': 'form-control'}),
16 'qualifications': forms.Textarea(attrs={'style': 'height: 200px',
17 'class': 'form-control'}),
18 }
19
20 def clean_status(self):
21 status = self.cleaned_data['status']
22 if status not in ['draft', 'submitted'] \
23 or (self.initial['status'] == 'submitted'
24 and status == 'draft'):
25 raise forms.ValidationError(_('The submitted status was invalid.'))
26 return status
27
28
29 ReferenceFormSet = forms.inlineformset_factory(
30 Application,
31 Reference,
32 fields=('name', 'position', 'email', 'phone_number', 'comment'),
33 widgets={
34 'name': forms.TextInput(attrs={'class': 'form-control'}),
35 'position': forms.TextInput(attrs={'class': 'form-control'}),
36 'email': forms.TextInput(attrs={'class': 'form-control'}),
37 'phone_number': forms.TextInput(attrs={'class': 'form-control'}),
38 'comment': forms.TextInput(attrs={'class': 'form-control'}),
39 },
40 extra=0,
41 )
42
43
44 class ApprovalForm(forms.ModelForm):
45 status = forms.ChoiceField(
46 choices=(
47 ('submitted', '---------'),
48 ('approved', _('Approved')),
49 ('disapproved', _('Disapproved')),
50 ),
51 )
52
53 class Meta:
54 model = Application
55 fields = []
56
57 def clean_status(self):
58 status = self.cleaned_data['status']
59 if status not in ['submitted', 'approved', 'disapproved']:
60 raise forms.ValidationError(_('The submitted status was invalid.'))
61 return status
62
63 def save(self, commit=True):
64 self.instance.status = self.cleaned_data['status']
65
66 super(ApprovalForm, self).save(commit)
67
68
69 class AppointmentForm(forms.Form):
70 appoint = AdvancedModelMultipleChoiceField(
71 Application.objects.none(),
72 widget=forms.CheckboxSelectMultiple(),
73 required=False,
74 )
75 overturn = forms.CharField(
76 required=False,
77 label=_('Overturn'),
78 help_text=_('Enter a comma separated list of users you want to '
79 'appoint to the position, even though did not apply for '
80 'the position.')
81 )
82
83 def __init__(self, position, *args, **kwargs):
84 super(AppointmentForm, self).__init__(*args, **kwargs)
85 self.position = position
86 self.fields['appoint'].queryset = position.applications.filter(
87 status__in=['submitted', 'approved', 'appointed', 'turned_down']
88 )
89 self.initial['appoint'] = position.applications.filter(
90 status='appointed'
91 )
92
93 def clean_overturn(self):
94 string = self.cleaned_data['overturn']
95 string = string.replace(' ', '')
96 if string == '':
97 return []
98 else:
99 users = string.split(',')
100 for u in users:
101 if not get_user_model().objects.filter(
102 username=u
103 ).exists():
104 raise forms.ValidationError(
105 _('No user with the username %(user)s exists.'),
106 params={'user': u},
107 )
108 elif self.position.applications.filter(
109 applicant__username=u
110 ).exists():
111 raise forms.ValidationError(
112 _('User %(user)s already applied for this position '
113 'and can not be appointed through the overturn '
114 'field.'),
115 params={'user': u},
116 )
117 return users
118
119 def clean(self):
120 super(AppointmentForm, self).clean()
121 appoint = self.cleaned_data.get('appoint', [])
122 overturn = self.cleaned_data.get('overturn', [])
123 nr_appointment = len(appoint) + len(overturn)
124 if nr_appointment > self.position.appointments:
125 raise forms.ValidationError(
126 _('You cannot appoint %(current)s applicants. The maximum '
127 'for this position is %(max)s.'),
128 params={
129 'current': nr_appointment,
130 'max': self.position.appointments,
131 },
132 )
133 return self.cleaned_data
134
135 def save(self):
136 for application in self.fields['appoint'].queryset:
137 if application in self.cleaned_data['appoint']:
138 application.status = 'appointed'
139 else:
140 application.status = 'turned_down'
141 application.save()
142
143 for user in self.cleaned_data['overturn']:
144 user = get_user_model().objects.get(
145 username=user
146 )
147 Application.objects.create(
148 position=self.position,
149 applicant=user,
150 status='appointed',
151 )
152
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/website/involvement/forms.py b/website/involvement/forms.py
--- a/website/involvement/forms.py
+++ b/website/involvement/forms.py
@@ -107,6 +107,8 @@
)
elif self.position.applications.filter(
applicant__username=u
+ ).exclude(
+ status='draft'
).exists():
raise forms.ValidationError(
_('User %(user)s already applied for this position '
@@ -144,8 +146,11 @@
user = get_user_model().objects.get(
username=user
)
- Application.objects.create(
+ appl, created = Application.objects.get_or_create(
position=self.position,
applicant=user,
- status='appointed',
+ defaults={'status': 'appointed'}
)
+ if not created:
+ appl.status = 'appointed'
+ appl.save()
| {"golden_diff": "diff --git a/website/involvement/forms.py b/website/involvement/forms.py\n--- a/website/involvement/forms.py\n+++ b/website/involvement/forms.py\n@@ -107,6 +107,8 @@\n )\n elif self.position.applications.filter(\n applicant__username=u\n+ ).exclude(\n+ status='draft'\n ).exists():\n raise forms.ValidationError(\n _('User %(user)s already applied for this position '\n@@ -144,8 +146,11 @@\n user = get_user_model().objects.get(\n username=user\n )\n- Application.objects.create(\n+ appl, created = Application.objects.get_or_create(\n position=self.position,\n applicant=user,\n- status='appointed',\n+ defaults={'status': 'appointed'}\n )\n+ if not created:\n+ appl.status = 'appointed'\n+ appl.save()\n", "issue": "Application drafts in limbo\n<!-- Do you want to ask a question? Are you looking for support? The system administrator can help you: [email protected] -->\r\n\r\n### Prerequisites\r\n\r\n* [x] Put an X between the brackets on this line if you have done all of the\r\nfollowing:\r\n * Reproduced the problem with clear cache.\r\n * (If running the application locally:) Made sure your running the newest version on the development branch\r\n * Checked that your issue isn't already filed: https://github.com/UTNkar/moore/issues\r\n\r\n### Description\r\n\r\nThe problem has not been reproduced but it's the kind of problem that probably will occur any way.\r\n\r\nWhen a user starts an application draft and forgets to submit it the person can not be appointed to the position with the overturn function. If the application for a position has been saved as draft when the application period ends the application becomes stuck in limbo. The group administrator can before the application period has ended see that there is a draft for the person. When the application period ends the draft is not visible among the submitted applications (very resonable). The problem occurs when the administrator wants to appoint that person anyway with the overturn function then an error message occurs saying. \"You can not appoint this person since an application has been submitted\". This should not be the case, a draft application should not be seen as a submitted application by the overturn function. The user can not see or delete the draft application after the application period has ended.\r\n\r\nQuick fix, an application system administrator can access the applications and set the status to submitted. \r\n\r\n### Steps to Reproduce\r\n\r\n1. Apply to an position and save the application as draft\r\n2. Wait for the application time to end\r\n3. Go to appoint for the given position\r\n4. Try overturn with the user who saved the application as draft\r\n5. Error message occurs \r\n\r\n<!-- Please select the appropriate \"topic category\"/blue and \"issue type\"/yellow label -->\n", "before_files": [{"content": "from django import forms\nfrom django.contrib.auth import get_user_model\nfrom django.utils.translation import ugettext_lazy as _\n\nfrom involvement.models import Application, Reference\nfrom utils.forms import AdvancedModelMultipleChoiceField\n\n\nclass ApplicationForm(forms.ModelForm):\n class Meta:\n model = Application\n exclude = ['position', 'applicant']\n widgets = {\n 'cover_letter': forms.Textarea(attrs={'style': 'height: 200px',\n 'class': 'form-control'}),\n 'qualifications': forms.Textarea(attrs={'style': 'height: 200px',\n 'class': 'form-control'}),\n }\n\n def clean_status(self):\n status = self.cleaned_data['status']\n if status not in ['draft', 'submitted'] \\\n or (self.initial['status'] == 'submitted'\n and status == 'draft'):\n raise forms.ValidationError(_('The submitted status was invalid.'))\n return status\n\n\nReferenceFormSet = forms.inlineformset_factory(\n Application,\n Reference,\n fields=('name', 'position', 'email', 'phone_number', 'comment'),\n widgets={\n 'name': forms.TextInput(attrs={'class': 'form-control'}),\n 'position': forms.TextInput(attrs={'class': 'form-control'}),\n 'email': forms.TextInput(attrs={'class': 'form-control'}),\n 'phone_number': forms.TextInput(attrs={'class': 'form-control'}),\n 'comment': forms.TextInput(attrs={'class': 'form-control'}),\n },\n extra=0,\n)\n\n\nclass ApprovalForm(forms.ModelForm):\n status = forms.ChoiceField(\n choices=(\n ('submitted', '---------'),\n ('approved', _('Approved')),\n ('disapproved', _('Disapproved')),\n ),\n )\n\n class Meta:\n model = Application\n fields = []\n\n def clean_status(self):\n status = self.cleaned_data['status']\n if status not in ['submitted', 'approved', 'disapproved']:\n raise forms.ValidationError(_('The submitted status was invalid.'))\n return status\n\n def save(self, commit=True):\n self.instance.status = self.cleaned_data['status']\n\n super(ApprovalForm, self).save(commit)\n\n\nclass AppointmentForm(forms.Form):\n appoint = AdvancedModelMultipleChoiceField(\n Application.objects.none(),\n widget=forms.CheckboxSelectMultiple(),\n required=False,\n )\n overturn = forms.CharField(\n required=False,\n label=_('Overturn'),\n help_text=_('Enter a comma separated list of users you want to '\n 'appoint to the position, even though did not apply for '\n 'the position.')\n )\n\n def __init__(self, position, *args, **kwargs):\n super(AppointmentForm, self).__init__(*args, **kwargs)\n self.position = position\n self.fields['appoint'].queryset = position.applications.filter(\n status__in=['submitted', 'approved', 'appointed', 'turned_down']\n )\n self.initial['appoint'] = position.applications.filter(\n status='appointed'\n )\n\n def clean_overturn(self):\n string = self.cleaned_data['overturn']\n string = string.replace(' ', '')\n if string == '':\n return []\n else:\n users = string.split(',')\n for u in users:\n if not get_user_model().objects.filter(\n username=u\n ).exists():\n raise forms.ValidationError(\n _('No user with the username %(user)s exists.'),\n params={'user': u},\n )\n elif self.position.applications.filter(\n applicant__username=u\n ).exists():\n raise forms.ValidationError(\n _('User %(user)s already applied for this position '\n 'and can not be appointed through the overturn '\n 'field.'),\n params={'user': u},\n )\n return users\n\n def clean(self):\n super(AppointmentForm, self).clean()\n appoint = self.cleaned_data.get('appoint', [])\n overturn = self.cleaned_data.get('overturn', [])\n nr_appointment = len(appoint) + len(overturn)\n if nr_appointment > self.position.appointments:\n raise forms.ValidationError(\n _('You cannot appoint %(current)s applicants. The maximum '\n 'for this position is %(max)s.'),\n params={\n 'current': nr_appointment,\n 'max': self.position.appointments,\n },\n )\n return self.cleaned_data\n\n def save(self):\n for application in self.fields['appoint'].queryset:\n if application in self.cleaned_data['appoint']:\n application.status = 'appointed'\n else:\n application.status = 'turned_down'\n application.save()\n\n for user in self.cleaned_data['overturn']:\n user = get_user_model().objects.get(\n username=user\n )\n Application.objects.create(\n position=self.position,\n applicant=user,\n status='appointed',\n )\n", "path": "website/involvement/forms.py"}], "after_files": [{"content": "from django import forms\nfrom django.contrib.auth import get_user_model\nfrom django.utils.translation import ugettext_lazy as _\n\nfrom involvement.models import Application, Reference\nfrom utils.forms import AdvancedModelMultipleChoiceField\n\n\nclass ApplicationForm(forms.ModelForm):\n class Meta:\n model = Application\n exclude = ['position', 'applicant']\n widgets = {\n 'cover_letter': forms.Textarea(attrs={'style': 'height: 200px',\n 'class': 'form-control'}),\n 'qualifications': forms.Textarea(attrs={'style': 'height: 200px',\n 'class': 'form-control'}),\n }\n\n def clean_status(self):\n status = self.cleaned_data['status']\n if status not in ['draft', 'submitted'] \\\n or (self.initial['status'] == 'submitted'\n and status == 'draft'):\n raise forms.ValidationError(_('The submitted status was invalid.'))\n return status\n\n\nReferenceFormSet = forms.inlineformset_factory(\n Application,\n Reference,\n fields=('name', 'position', 'email', 'phone_number', 'comment'),\n widgets={\n 'name': forms.TextInput(attrs={'class': 'form-control'}),\n 'position': forms.TextInput(attrs={'class': 'form-control'}),\n 'email': forms.TextInput(attrs={'class': 'form-control'}),\n 'phone_number': forms.TextInput(attrs={'class': 'form-control'}),\n 'comment': forms.TextInput(attrs={'class': 'form-control'}),\n },\n extra=0,\n)\n\n\nclass ApprovalForm(forms.ModelForm):\n status = forms.ChoiceField(\n choices=(\n ('submitted', '---------'),\n ('approved', _('Approved')),\n ('disapproved', _('Disapproved')),\n ),\n )\n\n class Meta:\n model = Application\n fields = []\n\n def clean_status(self):\n status = self.cleaned_data['status']\n if status not in ['submitted', 'approved', 'disapproved']:\n raise forms.ValidationError(_('The submitted status was invalid.'))\n return status\n\n def save(self, commit=True):\n self.instance.status = self.cleaned_data['status']\n\n super(ApprovalForm, self).save(commit)\n\n\nclass AppointmentForm(forms.Form):\n appoint = AdvancedModelMultipleChoiceField(\n Application.objects.none(),\n widget=forms.CheckboxSelectMultiple(),\n required=False,\n )\n overturn = forms.CharField(\n required=False,\n label=_('Overturn'),\n help_text=_('Enter a comma separated list of users you want to '\n 'appoint to the position, even though did not apply for '\n 'the position.')\n )\n\n def __init__(self, position, *args, **kwargs):\n super(AppointmentForm, self).__init__(*args, **kwargs)\n self.position = position\n self.fields['appoint'].queryset = position.applications.filter(\n status__in=['submitted', 'approved', 'appointed', 'turned_down']\n )\n self.initial['appoint'] = position.applications.filter(\n status='appointed'\n )\n\n def clean_overturn(self):\n string = self.cleaned_data['overturn']\n string = string.replace(' ', '')\n if string == '':\n return []\n else:\n users = string.split(',')\n for u in users:\n if not get_user_model().objects.filter(\n username=u\n ).exists():\n raise forms.ValidationError(\n _('No user with the username %(user)s exists.'),\n params={'user': u},\n )\n elif self.position.applications.filter(\n applicant__username=u\n ).exclude(\n status='draft'\n ).exists():\n raise forms.ValidationError(\n _('User %(user)s already applied for this position '\n 'and can not be appointed through the overturn '\n 'field.'),\n params={'user': u},\n )\n return users\n\n def clean(self):\n super(AppointmentForm, self).clean()\n appoint = self.cleaned_data.get('appoint', [])\n overturn = self.cleaned_data.get('overturn', [])\n nr_appointment = len(appoint) + len(overturn)\n if nr_appointment > self.position.appointments:\n raise forms.ValidationError(\n _('You cannot appoint %(current)s applicants. The maximum '\n 'for this position is %(max)s.'),\n params={\n 'current': nr_appointment,\n 'max': self.position.appointments,\n },\n )\n return self.cleaned_data\n\n def save(self):\n for application in self.fields['appoint'].queryset:\n if application in self.cleaned_data['appoint']:\n application.status = 'appointed'\n else:\n application.status = 'turned_down'\n application.save()\n\n for user in self.cleaned_data['overturn']:\n user = get_user_model().objects.get(\n username=user\n )\n appl, created = Application.objects.get_or_create(\n position=self.position,\n applicant=user,\n defaults={'status': 'appointed'}\n )\n if not created:\n appl.status = 'appointed'\n appl.save()\n", "path": "website/involvement/forms.py"}]} | 2,019 | 196 |
gh_patches_debug_4296 | rasdani/github-patches | git_diff | beetbox__beets-1181 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
ftintitle: does not gracefully handle duplicate artist name
Using ftintitle on one of my tracks, it seemed to get tripped up and not be able to fix it. I've tracked it down to being a problem with the fact that the Album Artist's name is in the Artist field twice.
```
Artist: The Roots feat. Talib Kweli / The Roots
Album Artist: The Roots
```
When trying to find the album artist in the artist field, it does a string split using the album artist as a separator. This returns a list with the following values `['', 'feat. Talib Kweli / ', '']`.
The code that tries to find the `feat_part` is then only expecting a two element list, but is instead given a three. It then checks if the `-1th` element, `2` in this case, is blank. If it's not, it extracts the featured artist.
If it is blank, it goes on to assume the featured artist must be on the left-hand side of the split and checks element `0`.
Both elements `0` and `2` are blank, so no featured part is found.
I've thought of two solutions, but am not sure which one would make more sense
- Attempt to remove duplicate album artists from the artist string before splitting
- Add another/change the current case to iterate over the split parts to find the first non-blank item
Either way, these methods would presumably still leave the trailing slash on the `feat. Talib Kweli /` and add the extraneous trailing slash to the track title. This, I'm not quite sure how to handle if at all.
Thoughts?
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `beetsplug/ftintitle.py`
Content:
```
1 # This file is part of beets.
2 # Copyright 2013, Verrus, <github.com/Verrus/beets-plugin-featInTitle>
3 #
4 # Permission is hereby granted, free of charge, to any person obtaining
5 # a copy of this software and associated documentation files (the
6 # "Software"), to deal in the Software without restriction, including
7 # without limitation the rights to use, copy, modify, merge, publish,
8 # distribute, sublicense, and/or sell copies of the Software, and to
9 # permit persons to whom the Software is furnished to do so, subject to
10 # the following conditions:
11 #
12 # The above copyright notice and this permission notice shall be
13 # included in all copies or substantial portions of the Software.
14
15 """Moves "featured" artists to the title from the artist field.
16 """
17 from beets import plugins
18 from beets import ui
19 from beets.util import displayable_path
20 from beets import config
21 import logging
22 import re
23
24 log = logging.getLogger('beets')
25
26
27 def split_on_feat(artist):
28 """Given an artist string, split the "main" artist from any artist
29 on the right-hand side of a string like "feat". Return the main
30 artist, which is always a string, and the featuring artist, which
31 may be a string or None if none is present.
32 """
33 # split on the first "feat".
34 regex = re.compile(plugins.feat_tokens(), re.IGNORECASE)
35 parts = [s.strip() for s in regex.split(artist, 1)]
36 if len(parts) == 1:
37 return parts[0], None
38 else:
39 return tuple(parts)
40
41
42 def contains_feat(title):
43 """Determine whether the title contains a "featured" marker.
44 """
45 return bool(re.search(plugins.feat_tokens(), title, flags=re.IGNORECASE))
46
47
48 def update_metadata(item, feat_part, drop_feat, loglevel=logging.DEBUG):
49 """Choose how to add new artists to the title and set the new
50 metadata. Also, print out messages about any changes that are made.
51 If `drop_feat` is set, then do not add the artist to the title; just
52 remove it from the artist field.
53 """
54 # In all cases, update the artist fields.
55 log.log(loglevel, u'artist: {0} -> {1}'.format(
56 item.artist, item.albumartist))
57 item.artist = item.albumartist
58 if item.artist_sort:
59 # Just strip the featured artist from the sort name.
60 item.artist_sort, _ = split_on_feat(item.artist_sort)
61
62 # Only update the title if it does not already contain a featured
63 # artist and if we do not drop featuring information.
64 if not drop_feat and not contains_feat(item.title):
65 new_title = u"{0} feat. {1}".format(item.title, feat_part)
66 log.log(loglevel, u'title: {0} -> {1}'.format(item.title, new_title))
67 item.title = new_title
68
69
70 def ft_in_title(item, drop_feat, loglevel=logging.DEBUG):
71 """Look for featured artists in the item's artist fields and move
72 them to the title.
73 """
74 artist = item.artist.strip()
75 albumartist = item.albumartist.strip()
76
77 # Check whether there is a featured artist on this track and the
78 # artist field does not exactly match the album artist field. In
79 # that case, we attempt to move the featured artist to the title.
80 _, featured = split_on_feat(artist)
81 if featured and albumartist != artist and albumartist:
82 log.log(loglevel, displayable_path(item.path))
83 feat_part = None
84
85 # Look for the album artist in the artist field. If it's not
86 # present, give up.
87 albumartist_split = artist.split(albumartist)
88 if len(albumartist_split) <= 1:
89 log.log(loglevel, 'album artist not present in artist')
90
91 # If the last element of the split (the right-hand side of the
92 # album artist) is nonempty, then it probably contains the
93 # featured artist.
94 elif albumartist_split[-1] != '':
95 # Extract the featured artist from the right-hand side.
96 _, feat_part = split_on_feat(albumartist_split[-1])
97
98 # Otherwise, if there's nothing on the right-hand side, look for a
99 # featuring artist on the left-hand side.
100 else:
101 lhs, rhs = split_on_feat(albumartist_split[0])
102 if rhs:
103 feat_part = lhs
104
105 # If we have a featuring artist, move it to the title.
106 if feat_part:
107 update_metadata(item, feat_part, drop_feat, loglevel)
108 else:
109 log.log(loglevel, u'no featuring artists found')
110
111
112 class FtInTitlePlugin(plugins.BeetsPlugin):
113 def __init__(self):
114 super(FtInTitlePlugin, self).__init__()
115
116 self.config.add({
117 'auto': True,
118 'drop': False,
119 })
120
121 self._command = ui.Subcommand(
122 'ftintitle',
123 help='move featured artists to the title field')
124
125 self._command.parser.add_option(
126 '-d', '--drop', dest='drop',
127 action='store_true', default=False,
128 help='drop featuring from artists and ignore title update')
129
130 if self.config['auto']:
131 self.import_stages = [self.imported]
132
133 def commands(self):
134
135 def func(lib, opts, args):
136 self.config.set_args(opts)
137 drop_feat = self.config['drop'].get(bool)
138 write = config['import']['write'].get(bool)
139
140 for item in lib.items(ui.decargs(args)):
141 ft_in_title(item, drop_feat, logging.INFO)
142 item.store()
143 if write:
144 item.try_write()
145
146 self._command.func = func
147 return [self._command]
148
149 def imported(self, session, task):
150 """Import hook for moving featuring artist automatically.
151 """
152 drop_feat = self.config['drop'].get(bool)
153
154 for item in task.imported_items():
155 ft_in_title(item, drop_feat, logging.DEBUG)
156 item.store()
157
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/beetsplug/ftintitle.py b/beetsplug/ftintitle.py
--- a/beetsplug/ftintitle.py
+++ b/beetsplug/ftintitle.py
@@ -83,7 +83,7 @@
# Look for the album artist in the artist field. If it's not
# present, give up.
- albumartist_split = artist.split(albumartist)
+ albumartist_split = artist.split(albumartist, 1)
if len(albumartist_split) <= 1:
ui.print_('album artist not present in artist')
| {"golden_diff": "diff --git a/beetsplug/ftintitle.py b/beetsplug/ftintitle.py\n--- a/beetsplug/ftintitle.py\n+++ b/beetsplug/ftintitle.py\n@@ -83,7 +83,7 @@\n \n # Look for the album artist in the artist field. If it's not\n # present, give up.\n- albumartist_split = artist.split(albumartist)\n+ albumartist_split = artist.split(albumartist, 1)\n if len(albumartist_split) <= 1:\n ui.print_('album artist not present in artist')\n", "issue": "ftintitle: does not gracefully handle duplicate artist name\nUsing ftintitle on one of my tracks, it seemed to get tripped up and not be able to fix it. I've tracked it down to being a problem with the fact that the Album Artist's name is in the Artist field twice.\n\n```\nArtist: The Roots feat. Talib Kweli / The Roots\nAlbum Artist: The Roots\n```\n\nWhen trying to find the album artist in the artist field, it does a string split using the album artist as a separator. This returns a list with the following values `['', 'feat. Talib Kweli / ', '']`.\n\nThe code that tries to find the `feat_part` is then only expecting a two element list, but is instead given a three. It then checks if the `-1th` element, `2` in this case, is blank. If it's not, it extracts the featured artist. \n\nIf it is blank, it goes on to assume the featured artist must be on the left-hand side of the split and checks element `0`.\n\nBoth elements `0` and `2` are blank, so no featured part is found.\n\nI've thought of two solutions, but am not sure which one would make more sense\n- Attempt to remove duplicate album artists from the artist string before splitting\n- Add another/change the current case to iterate over the split parts to find the first non-blank item\n\nEither way, these methods would presumably still leave the trailing slash on the `feat. Talib Kweli /` and add the extraneous trailing slash to the track title. This, I'm not quite sure how to handle if at all.\n\nThoughts?\n\n", "before_files": [{"content": "# This file is part of beets.\n# Copyright 2013, Verrus, <github.com/Verrus/beets-plugin-featInTitle>\n#\n# Permission is hereby granted, free of charge, to any person obtaining\n# a copy of this software and associated documentation files (the\n# \"Software\"), to deal in the Software without restriction, including\n# without limitation the rights to use, copy, modify, merge, publish,\n# distribute, sublicense, and/or sell copies of the Software, and to\n# permit persons to whom the Software is furnished to do so, subject to\n# the following conditions:\n#\n# The above copyright notice and this permission notice shall be\n# included in all copies or substantial portions of the Software.\n\n\"\"\"Moves \"featured\" artists to the title from the artist field.\n\"\"\"\nfrom beets import plugins\nfrom beets import ui\nfrom beets.util import displayable_path\nfrom beets import config\nimport logging\nimport re\n\nlog = logging.getLogger('beets')\n\n\ndef split_on_feat(artist):\n \"\"\"Given an artist string, split the \"main\" artist from any artist\n on the right-hand side of a string like \"feat\". Return the main\n artist, which is always a string, and the featuring artist, which\n may be a string or None if none is present.\n \"\"\"\n # split on the first \"feat\".\n regex = re.compile(plugins.feat_tokens(), re.IGNORECASE)\n parts = [s.strip() for s in regex.split(artist, 1)]\n if len(parts) == 1:\n return parts[0], None\n else:\n return tuple(parts)\n\n\ndef contains_feat(title):\n \"\"\"Determine whether the title contains a \"featured\" marker.\n \"\"\"\n return bool(re.search(plugins.feat_tokens(), title, flags=re.IGNORECASE))\n\n\ndef update_metadata(item, feat_part, drop_feat, loglevel=logging.DEBUG):\n \"\"\"Choose how to add new artists to the title and set the new\n metadata. Also, print out messages about any changes that are made.\n If `drop_feat` is set, then do not add the artist to the title; just\n remove it from the artist field.\n \"\"\"\n # In all cases, update the artist fields.\n log.log(loglevel, u'artist: {0} -> {1}'.format(\n item.artist, item.albumartist))\n item.artist = item.albumartist\n if item.artist_sort:\n # Just strip the featured artist from the sort name.\n item.artist_sort, _ = split_on_feat(item.artist_sort)\n\n # Only update the title if it does not already contain a featured\n # artist and if we do not drop featuring information.\n if not drop_feat and not contains_feat(item.title):\n new_title = u\"{0} feat. {1}\".format(item.title, feat_part)\n log.log(loglevel, u'title: {0} -> {1}'.format(item.title, new_title))\n item.title = new_title\n\n\ndef ft_in_title(item, drop_feat, loglevel=logging.DEBUG):\n \"\"\"Look for featured artists in the item's artist fields and move\n them to the title.\n \"\"\"\n artist = item.artist.strip()\n albumartist = item.albumartist.strip()\n\n # Check whether there is a featured artist on this track and the\n # artist field does not exactly match the album artist field. In\n # that case, we attempt to move the featured artist to the title.\n _, featured = split_on_feat(artist)\n if featured and albumartist != artist and albumartist:\n log.log(loglevel, displayable_path(item.path))\n feat_part = None\n\n # Look for the album artist in the artist field. If it's not\n # present, give up.\n albumartist_split = artist.split(albumartist)\n if len(albumartist_split) <= 1:\n log.log(loglevel, 'album artist not present in artist')\n\n # If the last element of the split (the right-hand side of the\n # album artist) is nonempty, then it probably contains the\n # featured artist.\n elif albumartist_split[-1] != '':\n # Extract the featured artist from the right-hand side.\n _, feat_part = split_on_feat(albumartist_split[-1])\n\n # Otherwise, if there's nothing on the right-hand side, look for a\n # featuring artist on the left-hand side.\n else:\n lhs, rhs = split_on_feat(albumartist_split[0])\n if rhs:\n feat_part = lhs\n\n # If we have a featuring artist, move it to the title.\n if feat_part:\n update_metadata(item, feat_part, drop_feat, loglevel)\n else:\n log.log(loglevel, u'no featuring artists found')\n\n\nclass FtInTitlePlugin(plugins.BeetsPlugin):\n def __init__(self):\n super(FtInTitlePlugin, self).__init__()\n\n self.config.add({\n 'auto': True,\n 'drop': False,\n })\n\n self._command = ui.Subcommand(\n 'ftintitle',\n help='move featured artists to the title field')\n\n self._command.parser.add_option(\n '-d', '--drop', dest='drop',\n action='store_true', default=False,\n help='drop featuring from artists and ignore title update')\n\n if self.config['auto']:\n self.import_stages = [self.imported]\n\n def commands(self):\n\n def func(lib, opts, args):\n self.config.set_args(opts)\n drop_feat = self.config['drop'].get(bool)\n write = config['import']['write'].get(bool)\n\n for item in lib.items(ui.decargs(args)):\n ft_in_title(item, drop_feat, logging.INFO)\n item.store()\n if write:\n item.try_write()\n\n self._command.func = func\n return [self._command]\n\n def imported(self, session, task):\n \"\"\"Import hook for moving featuring artist automatically.\n \"\"\"\n drop_feat = self.config['drop'].get(bool)\n\n for item in task.imported_items():\n ft_in_title(item, drop_feat, logging.DEBUG)\n item.store()\n", "path": "beetsplug/ftintitle.py"}], "after_files": [{"content": "# This file is part of beets.\n# Copyright 2013, Verrus, <github.com/Verrus/beets-plugin-featInTitle>\n#\n# Permission is hereby granted, free of charge, to any person obtaining\n# a copy of this software and associated documentation files (the\n# \"Software\"), to deal in the Software without restriction, including\n# without limitation the rights to use, copy, modify, merge, publish,\n# distribute, sublicense, and/or sell copies of the Software, and to\n# permit persons to whom the Software is furnished to do so, subject to\n# the following conditions:\n#\n# The above copyright notice and this permission notice shall be\n# included in all copies or substantial portions of the Software.\n\n\"\"\"Moves \"featured\" artists to the title from the artist field.\n\"\"\"\nfrom beets import plugins\nfrom beets import ui\nfrom beets.util import displayable_path\nfrom beets import config\nimport logging\nimport re\n\nlog = logging.getLogger('beets')\n\n\ndef split_on_feat(artist):\n \"\"\"Given an artist string, split the \"main\" artist from any artist\n on the right-hand side of a string like \"feat\". Return the main\n artist, which is always a string, and the featuring artist, which\n may be a string or None if none is present.\n \"\"\"\n # split on the first \"feat\".\n regex = re.compile(plugins.feat_tokens(), re.IGNORECASE)\n parts = [s.strip() for s in regex.split(artist, 1)]\n if len(parts) == 1:\n return parts[0], None\n else:\n return tuple(parts)\n\n\ndef contains_feat(title):\n \"\"\"Determine whether the title contains a \"featured\" marker.\n \"\"\"\n return bool(re.search(plugins.feat_tokens(), title, flags=re.IGNORECASE))\n\n\ndef update_metadata(item, feat_part, drop_feat):\n \"\"\"Choose how to add new artists to the title and set the new\n metadata. Also, print out messages about any changes that are made.\n If `drop_feat` is set, then do not add the artist to the title; just\n remove it from the artist field.\n \"\"\"\n # In all cases, update the artist fields.\n ui.print_(u'artist: {0} -> {1}'.format(item.artist, item.albumartist))\n item.artist = item.albumartist\n if item.artist_sort:\n # Just strip the featured artist from the sort name.\n item.artist_sort, _ = split_on_feat(item.artist_sort)\n\n # Only update the title if it does not already contain a featured\n # artist and if we do not drop featuring information.\n if not drop_feat and not contains_feat(item.title):\n new_title = u\"{0} feat. {1}\".format(item.title, feat_part)\n ui.print_(u'title: {0} -> {1}'.format(item.title, new_title))\n item.title = new_title\n\n\ndef ft_in_title(item, drop_feat):\n \"\"\"Look for featured artists in the item's artist fields and move\n them to the title.\n \"\"\"\n artist = item.artist.strip()\n albumartist = item.albumartist.strip()\n\n # Check whether there is a featured artist on this track and the\n # artist field does not exactly match the album artist field. In\n # that case, we attempt to move the featured artist to the title.\n _, featured = split_on_feat(artist)\n if featured and albumartist != artist and albumartist:\n ui.print_(displayable_path(item.path))\n feat_part = None\n\n # Look for the album artist in the artist field. If it's not\n # present, give up.\n albumartist_split = artist.split(albumartist, 1)\n if len(albumartist_split) <= 1:\n ui.print_('album artist not present in artist')\n\n # If the last element of the split (the right-hand side of the\n # album artist) is nonempty, then it probably contains the\n # featured artist.\n elif albumartist_split[-1] != '':\n # Extract the featured artist from the right-hand side.\n _, feat_part = split_on_feat(albumartist_split[-1])\n\n # Otherwise, if there's nothing on the right-hand side, look for a\n # featuring artist on the left-hand side.\n else:\n lhs, rhs = split_on_feat(albumartist_split[0])\n if rhs:\n feat_part = lhs\n\n # If we have a featuring artist, move it to the title.\n if feat_part:\n update_metadata(item, feat_part, drop_feat)\n else:\n ui.print_(u'no featuring artists found')\n\n ui.print_()\n\n\nclass FtInTitlePlugin(plugins.BeetsPlugin):\n def __init__(self):\n super(FtInTitlePlugin, self).__init__()\n\n self.config.add({\n 'auto': True,\n 'drop': False,\n })\n\n self._command = ui.Subcommand(\n 'ftintitle',\n help='move featured artists to the title field')\n\n self._command.parser.add_option(\n '-d', '--drop', dest='drop',\n action='store_true', default=False,\n help='drop featuring from artists and ignore title update')\n\n if self.config['auto']:\n self.import_stages = [self.imported]\n\n def commands(self):\n\n def func(lib, opts, args):\n self.config.set_args(opts)\n drop_feat = self.config['drop'].get(bool)\n write = config['import']['write'].get(bool)\n\n for item in lib.items(ui.decargs(args)):\n ft_in_title(item, drop_feat)\n item.store()\n if write:\n item.try_write()\n\n self._command.func = func\n return [self._command]\n\n def imported(self, session, task):\n \"\"\"Import hook for moving featuring artist automatically.\n \"\"\"\n drop_feat = self.config['drop'].get(bool)\n\n for item in task.imported_items():\n ft_in_title(item, drop_feat)\n item.store()\n", "path": "beetsplug/ftintitle.py"}]} | 2,296 | 128 |
gh_patches_debug_1064 | rasdani/github-patches | git_diff | scikit-hep__pyhf-1220 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
pytest v6.2.0 causing test_optim_with_value to fail
# Description
`v0.5.4` `bump2version` changes were swept into `master` 2020-12-12 with f824afe and the CI on `master` succeeded. Later that day [`pytest` `v6.2.0`](https://github.com/pytest-dev/pytest/releases/tag/6.2.0) was released and the nightly scheduled CI failed on
```pytb
_______________________ test_optim_with_value[jax-mu=1] ________________________
backend = (<pyhf.tensor.jax_backend.jax_backend object at 0x7f6bf92def50>, None)
source = {'bindata': {'bkg': [100.0, 150.0], 'bkgsys_dn': [98, 100], 'bkgsys_up': [102, 190], 'data': [120.0, 180.0], ...}, 'binning': [2, -0.5, 1.5]}
spec = {'channels': [{'name': 'singlechannel', 'samples': [{'data': [30.0, 95.0], 'modifiers': [{...}], 'name': 'signal'}, {'data': [100.0, 150.0], 'modifiers': [{...}], 'name': 'background'}]}]}
mu = 1.0
@pytest.mark.parametrize('mu', [1.0], ids=['mu=1'])
def test_optim_with_value(backend, source, spec, mu):
pdf = pyhf.Model(spec)
data = source['bindata']['data'] + pdf.config.auxdata
init_pars = pdf.config.suggested_init()
par_bounds = pdf.config.suggested_bounds()
optim = pyhf.optimizer
result = optim.minimize(pyhf.infer.mle.twice_nll, data, pdf, init_pars, par_bounds)
assert pyhf.tensorlib.tolist(result)
result, fitted_val = optim.minimize(
pyhf.infer.mle.twice_nll,
data,
pdf,
init_pars,
par_bounds,
fixed_vals=[(pdf.config.poi_index, mu)],
return_fitted_val=True,
)
assert pyhf.tensorlib.tolist(result)
assert pyhf.tensorlib.shape(fitted_val) == ()
> assert pytest.approx(17.52954975, rel=1e-5) == fitted_val
E assert 17.52954975 ± 1.8e-04 == DeviceArray(17.52954975, dtype=float64)
E + where 17.52954975 ± 1.8e-04 = <function approx at 0x7f6cc1747f80>(17.52954975, rel=1e-05)
E + where <function approx at 0x7f6cc1747f80> = pytest.approx
tests/test_optim.py:383: AssertionError
```
Diffing the installed libraries between the two (in [f824afe_install.txt](https://github.com/scikit-hep/pyhf/files/5684241/f824afe_install.txt) and [failing_install.txt](https://github.com/scikit-hep/pyhf/files/5684242/failing_install.txt)) shows that the relevant change is `pytest`
```
$ diff f824afe_install.txt failing_install.txt
33a34
> importlib-metadata 3.1.1
83c84
< py 1.9.0
---
> py 1.10.0
96c97
< pytest 6.1.2
---
> pytest 6.2.0
143a145
> zipp 3.4.0
```
This is confirmed as if
```diff
--- a/setup.py
+++ b/setup.py
@@ -29,7 +29,7 @@
+ extras_require['contrib']
+ extras_require['shellcomplete']
+ [
- 'pytest~=6.0',
+ 'pytest~=6.1.0',
'pytest-cov>=2.5.1',
'pytest-mock',
'pytest-benchmark[histogram]',
```
the [CI installs `v6.1.2` and passes](https://github.com/scikit-hep/pyhf/actions/runs/418404132).
This behavior is confusing as the only mention of `pytest.approx`in the [`v6.2.0` release notes](https://github.com/pytest-dev/pytest/releases/tag/6.2.0) is under "Improvements"
> 7710: Use strict equality comparison for non-numeric types in pytest.approx instead of
raising TypeError.
>
> This was the undocumented behavior before 3.7, but is now officially a supported feature.
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `setup.py`
Content:
```
1 from setuptools import setup
2
3 extras_require = {
4 'shellcomplete': ['click_completion'],
5 'tensorflow': [
6 'tensorflow~=2.2.0', # TensorFlow minor releases are as volatile as major
7 'tensorflow-probability~=0.10.0',
8 ],
9 'torch': ['torch~=1.2'],
10 'jax': ['jax~=0.2.4', 'jaxlib~=0.1.56'],
11 'xmlio': ['uproot3~=3.14'], # Future proof against uproot4 API changes
12 'minuit': ['iminuit~=1.5.3'],
13 }
14 extras_require['backends'] = sorted(
15 set(
16 extras_require['tensorflow']
17 + extras_require['torch']
18 + extras_require['jax']
19 + extras_require['minuit']
20 )
21 )
22 extras_require['contrib'] = sorted({'matplotlib', 'requests'})
23 extras_require['lint'] = sorted({'flake8', 'black'})
24
25 extras_require['test'] = sorted(
26 set(
27 extras_require['backends']
28 + extras_require['xmlio']
29 + extras_require['contrib']
30 + extras_require['shellcomplete']
31 + [
32 'pytest~=6.0',
33 'pytest-cov>=2.5.1',
34 'pytest-mock',
35 'pytest-benchmark[histogram]',
36 'pytest-console-scripts',
37 'pytest-mpl',
38 'pydocstyle',
39 'coverage>=4.0', # coveralls
40 'papermill~=2.0',
41 'nteract-scrapbook~=0.2',
42 'jupyter',
43 'graphviz',
44 'jsonpatch',
45 ]
46 )
47 )
48 extras_require['docs'] = sorted(
49 {
50 'sphinx>=3.1.2',
51 'sphinxcontrib-bibtex',
52 'sphinx-click',
53 'sphinx_rtd_theme',
54 'nbsphinx',
55 'ipywidgets',
56 'sphinx-issues',
57 'sphinx-copybutton>0.2.9',
58 }
59 )
60 extras_require['develop'] = sorted(
61 set(
62 extras_require['docs']
63 + extras_require['lint']
64 + extras_require['test']
65 + [
66 'nbdime',
67 'bump2version',
68 'ipython',
69 'pre-commit',
70 'check-manifest',
71 'codemetapy>=0.3.4',
72 'twine',
73 ]
74 )
75 )
76 extras_require['complete'] = sorted(set(sum(extras_require.values(), [])))
77
78
79 setup(
80 extras_require=extras_require,
81 use_scm_version=lambda: {'local_scheme': lambda version: ''},
82 )
83
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/setup.py b/setup.py
--- a/setup.py
+++ b/setup.py
@@ -48,7 +48,7 @@
extras_require['docs'] = sorted(
{
'sphinx>=3.1.2',
- 'sphinxcontrib-bibtex',
+ 'sphinxcontrib-bibtex~=1.0',
'sphinx-click',
'sphinx_rtd_theme',
'nbsphinx',
| {"golden_diff": "diff --git a/setup.py b/setup.py\n--- a/setup.py\n+++ b/setup.py\n@@ -48,7 +48,7 @@\n extras_require['docs'] = sorted(\n {\n 'sphinx>=3.1.2',\n- 'sphinxcontrib-bibtex',\n+ 'sphinxcontrib-bibtex~=1.0',\n 'sphinx-click',\n 'sphinx_rtd_theme',\n 'nbsphinx',\n", "issue": "pytest v6.2.0 causing test_optim_with_value to fail\n# Description\r\n\r\n`v0.5.4` `bump2version` changes were swept into `master` 2020-12-12 with f824afe and the CI on `master` succeeded. Later that day [`pytest` `v6.2.0`](https://github.com/pytest-dev/pytest/releases/tag/6.2.0) was released and the nightly scheduled CI failed on \r\n\r\n```pytb\r\n_______________________ test_optim_with_value[jax-mu=1] ________________________\r\n\r\nbackend = (<pyhf.tensor.jax_backend.jax_backend object at 0x7f6bf92def50>, None)\r\nsource = {'bindata': {'bkg': [100.0, 150.0], 'bkgsys_dn': [98, 100], 'bkgsys_up': [102, 190], 'data': [120.0, 180.0], ...}, 'binning': [2, -0.5, 1.5]}\r\nspec = {'channels': [{'name': 'singlechannel', 'samples': [{'data': [30.0, 95.0], 'modifiers': [{...}], 'name': 'signal'}, {'data': [100.0, 150.0], 'modifiers': [{...}], 'name': 'background'}]}]}\r\nmu = 1.0\r\n\r\n @pytest.mark.parametrize('mu', [1.0], ids=['mu=1'])\r\n def test_optim_with_value(backend, source, spec, mu):\r\n pdf = pyhf.Model(spec)\r\n data = source['bindata']['data'] + pdf.config.auxdata\r\n \r\n init_pars = pdf.config.suggested_init()\r\n par_bounds = pdf.config.suggested_bounds()\r\n \r\n optim = pyhf.optimizer\r\n \r\n result = optim.minimize(pyhf.infer.mle.twice_nll, data, pdf, init_pars, par_bounds)\r\n assert pyhf.tensorlib.tolist(result)\r\n \r\n result, fitted_val = optim.minimize(\r\n pyhf.infer.mle.twice_nll,\r\n data,\r\n pdf,\r\n init_pars,\r\n par_bounds,\r\n fixed_vals=[(pdf.config.poi_index, mu)],\r\n return_fitted_val=True,\r\n )\r\n assert pyhf.tensorlib.tolist(result)\r\n assert pyhf.tensorlib.shape(fitted_val) == ()\r\n> assert pytest.approx(17.52954975, rel=1e-5) == fitted_val\r\nE assert 17.52954975 \u00b1 1.8e-04 == DeviceArray(17.52954975, dtype=float64)\r\nE + where 17.52954975 \u00b1 1.8e-04 = <function approx at 0x7f6cc1747f80>(17.52954975, rel=1e-05)\r\nE + where <function approx at 0x7f6cc1747f80> = pytest.approx\r\n\r\ntests/test_optim.py:383: AssertionError\r\n```\r\n\r\nDiffing the installed libraries between the two (in [f824afe_install.txt](https://github.com/scikit-hep/pyhf/files/5684241/f824afe_install.txt) and [failing_install.txt](https://github.com/scikit-hep/pyhf/files/5684242/failing_install.txt)) shows that the relevant change is `pytest`\r\n\r\n```\r\n$ diff f824afe_install.txt failing_install.txt \r\n33a34\r\n> importlib-metadata 3.1.1\r\n83c84\r\n< py 1.9.0\r\n---\r\n> py 1.10.0\r\n96c97\r\n< pytest 6.1.2\r\n---\r\n> pytest 6.2.0\r\n143a145\r\n> zipp 3.4.0\r\n```\r\n\r\nThis is confirmed as if\r\n\r\n```diff\r\n--- a/setup.py\r\n+++ b/setup.py\r\n@@ -29,7 +29,7 @@\r\n + extras_require['contrib']\r\n + extras_require['shellcomplete']\r\n + [\r\n- 'pytest~=6.0',\r\n+ 'pytest~=6.1.0',\r\n 'pytest-cov>=2.5.1',\r\n 'pytest-mock',\r\n 'pytest-benchmark[histogram]',\r\n```\r\n\r\nthe [CI installs `v6.1.2` and passes](https://github.com/scikit-hep/pyhf/actions/runs/418404132).\r\n\r\nThis behavior is confusing as the only mention of `pytest.approx`in the [`v6.2.0` release notes](https://github.com/pytest-dev/pytest/releases/tag/6.2.0) is under \"Improvements\"\r\n\r\n> 7710: Use strict equality comparison for non-numeric types in pytest.approx instead of\r\nraising TypeError.\r\n>\r\n> This was the undocumented behavior before 3.7, but is now officially a supported feature.\n", "before_files": [{"content": "from setuptools import setup\n\nextras_require = {\n 'shellcomplete': ['click_completion'],\n 'tensorflow': [\n 'tensorflow~=2.2.0', # TensorFlow minor releases are as volatile as major\n 'tensorflow-probability~=0.10.0',\n ],\n 'torch': ['torch~=1.2'],\n 'jax': ['jax~=0.2.4', 'jaxlib~=0.1.56'],\n 'xmlio': ['uproot3~=3.14'], # Future proof against uproot4 API changes\n 'minuit': ['iminuit~=1.5.3'],\n}\nextras_require['backends'] = sorted(\n set(\n extras_require['tensorflow']\n + extras_require['torch']\n + extras_require['jax']\n + extras_require['minuit']\n )\n)\nextras_require['contrib'] = sorted({'matplotlib', 'requests'})\nextras_require['lint'] = sorted({'flake8', 'black'})\n\nextras_require['test'] = sorted(\n set(\n extras_require['backends']\n + extras_require['xmlio']\n + extras_require['contrib']\n + extras_require['shellcomplete']\n + [\n 'pytest~=6.0',\n 'pytest-cov>=2.5.1',\n 'pytest-mock',\n 'pytest-benchmark[histogram]',\n 'pytest-console-scripts',\n 'pytest-mpl',\n 'pydocstyle',\n 'coverage>=4.0', # coveralls\n 'papermill~=2.0',\n 'nteract-scrapbook~=0.2',\n 'jupyter',\n 'graphviz',\n 'jsonpatch',\n ]\n )\n)\nextras_require['docs'] = sorted(\n {\n 'sphinx>=3.1.2',\n 'sphinxcontrib-bibtex',\n 'sphinx-click',\n 'sphinx_rtd_theme',\n 'nbsphinx',\n 'ipywidgets',\n 'sphinx-issues',\n 'sphinx-copybutton>0.2.9',\n }\n)\nextras_require['develop'] = sorted(\n set(\n extras_require['docs']\n + extras_require['lint']\n + extras_require['test']\n + [\n 'nbdime',\n 'bump2version',\n 'ipython',\n 'pre-commit',\n 'check-manifest',\n 'codemetapy>=0.3.4',\n 'twine',\n ]\n )\n)\nextras_require['complete'] = sorted(set(sum(extras_require.values(), [])))\n\n\nsetup(\n extras_require=extras_require,\n use_scm_version=lambda: {'local_scheme': lambda version: ''},\n)\n", "path": "setup.py"}], "after_files": [{"content": "from setuptools import setup\n\nextras_require = {\n 'shellcomplete': ['click_completion'],\n 'tensorflow': [\n 'tensorflow~=2.2.0', # TensorFlow minor releases are as volatile as major\n 'tensorflow-probability~=0.10.0',\n ],\n 'torch': ['torch~=1.2'],\n 'jax': ['jax~=0.2.4', 'jaxlib~=0.1.56'],\n 'xmlio': ['uproot3~=3.14'], # Future proof against uproot4 API changes\n 'minuit': ['iminuit~=1.5.3'],\n}\nextras_require['backends'] = sorted(\n set(\n extras_require['tensorflow']\n + extras_require['torch']\n + extras_require['jax']\n + extras_require['minuit']\n )\n)\nextras_require['contrib'] = sorted({'matplotlib', 'requests'})\nextras_require['lint'] = sorted({'flake8', 'black'})\n\nextras_require['test'] = sorted(\n set(\n extras_require['backends']\n + extras_require['xmlio']\n + extras_require['contrib']\n + extras_require['shellcomplete']\n + [\n 'pytest~=6.0',\n 'pytest-cov>=2.5.1',\n 'pytest-mock',\n 'pytest-benchmark[histogram]',\n 'pytest-console-scripts',\n 'pytest-mpl',\n 'pydocstyle',\n 'coverage>=4.0', # coveralls\n 'papermill~=2.0',\n 'nteract-scrapbook~=0.2',\n 'jupyter',\n 'graphviz',\n 'jsonpatch',\n ]\n )\n)\nextras_require['docs'] = sorted(\n {\n 'sphinx>=3.1.2',\n 'sphinxcontrib-bibtex~=1.0',\n 'sphinx-click',\n 'sphinx_rtd_theme',\n 'nbsphinx',\n 'ipywidgets',\n 'sphinx-issues',\n 'sphinx-copybutton>0.2.9',\n }\n)\nextras_require['develop'] = sorted(\n set(\n extras_require['docs']\n + extras_require['lint']\n + extras_require['test']\n + [\n 'nbdime',\n 'bump2version',\n 'ipython',\n 'pre-commit',\n 'check-manifest',\n 'codemetapy>=0.3.4',\n 'twine',\n ]\n )\n)\nextras_require['complete'] = sorted(set(sum(extras_require.values(), [])))\n\n\nsetup(\n extras_require=extras_require,\n use_scm_version=lambda: {'local_scheme': lambda version: ''},\n)\n", "path": "setup.py"}]} | 2,137 | 97 |
gh_patches_debug_15908 | rasdani/github-patches | git_diff | mkdocs__mkdocs-288 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
If the mkdocs.yml is completely empty there is a traceback
```
Traceback (most recent call last):
File "/home/dougalmatthews/.virtualenvs/mkdocs/bin/mkdocs", line 9, in <module>
load_entry_point('mkdocs==0.11.1', 'console_scripts', 'mkdocs')()
File "/home/dougalmatthews/.virtualenvs/mkdocs/lib/python3.4/site-packages/mkdocs/main.py", line 60, in run_main
main(cmd, args=sys.argv[2:], options=dict(opts))
File "/home/dougalmatthews/.virtualenvs/mkdocs/lib/python3.4/site-packages/mkdocs/main.py", line 32, in main
config = load_config(options=options)
File "/home/dougalmatthews/.virtualenvs/mkdocs/lib/python3.4/site-packages/mkdocs/config.py", line 82, in load_config
user_config.update(options)
AttributeError: 'NoneType' object has no attribute 'update'
```
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `mkdocs/config.py`
Content:
```
1 # coding: utf-8
2
3 from mkdocs import utils
4 from mkdocs.compat import urlparse
5 from mkdocs.exceptions import ConfigurationError
6
7 import os
8 import yaml
9
10 DEFAULT_CONFIG = {
11 'site_name': None,
12 'pages': None,
13
14 'site_url': None,
15 'site_description': None,
16 'site_author': None,
17 'site_favicon': None,
18
19 'theme': 'mkdocs',
20 'docs_dir': 'docs',
21 'site_dir': 'site',
22 'theme_dir': None,
23
24 'copyright': None,
25 'google_analytics': None,
26
27 # The address on which to serve the livereloading docs server.
28 'dev_addr': '127.0.0.1:8000',
29
30 # If `True`, use `<page_name>/index.hmtl` style files with hyperlinks to the directory.
31 # If `False`, use `<page_name>.html style file with hyperlinks to the file.
32 # True generates nicer URLs, but False is useful if browsing the output on a filesystem.
33 'use_directory_urls': True,
34
35 # Specify a link to the project source repo to be included
36 # in the documentation pages.
37 'repo_url': None,
38
39 # A name to use for the link to the project source repo.
40 # Default: If repo_url is unset then None, otherwise
41 # "GitHub" or "Bitbucket" for known url or Hostname for unknown urls.
42 'repo_name': None,
43
44 # Specify which css or javascript files from the docs
45 # directionary should be additionally included in the site.
46 # Default: List of all .css and .js files in the docs dir.
47 'extra_css': None,
48 'extra_javascript': None,
49
50 # Determine if the site should include the nav and next/prev elements.
51 # Default: True if the site has more than one page, False otherwise.
52 'include_nav': None,
53 'include_next_prev': None,
54
55 # PyMarkdown extension names.
56 'markdown_extensions': (),
57
58 # Determine if the site should generate a json search index and include
59 # search elements in the theme. - TODO
60 'include_search': False,
61
62 # Determine if the site should include a 404.html page.
63 # TODO: Implment this. Make this None, have it True if a 404.html
64 # template exists in the theme or docs dir.
65 'include_404': False,
66
67 # Determine if the site should include a sitemap.xml page.
68 # TODO: Implement this. Make this None, have it True if a sitemap.xml
69 # template exists in the theme or docs dir.
70 'include_sitemap': False,
71 }
72
73
74 def load_config(filename='mkdocs.yml', options=None):
75 options = options or {}
76 if 'config' in options:
77 filename = options['config']
78 if not os.path.exists(filename):
79 raise ConfigurationError("Config file '%s' does not exist." % filename)
80 with open(filename, 'r') as fp:
81 user_config = yaml.load(fp)
82 user_config.update(options)
83 return validate_config(user_config)
84
85
86 def validate_config(user_config):
87 config = DEFAULT_CONFIG.copy()
88 config.update(user_config)
89
90 if not config['site_name']:
91 raise ConfigurationError("Config must contain 'site_name' setting.")
92
93 # If not specified, then the 'pages' config simply includes all
94 # markdown files in the docs dir, without generating any header items
95 # for them.
96 pages = []
97 extra_css = []
98 extra_javascript = []
99 for (dirpath, dirnames, filenames) in os.walk(config['docs_dir']):
100 for filename in sorted(filenames):
101 fullpath = os.path.join(dirpath, filename)
102 relpath = os.path.relpath(fullpath, config['docs_dir'])
103
104 if utils.is_markdown_file(filename):
105 # index pages should always be the first listed page.
106 if os.path.splitext(relpath)[0] == 'index':
107 pages.insert(0, relpath)
108 else:
109 pages.append(relpath)
110 elif utils.is_css_file(filename):
111 extra_css.append(relpath)
112 elif utils.is_javascript_file(filename):
113 extra_javascript.append(relpath)
114
115 if config['pages'] is None:
116 config['pages'] = pages
117
118 if config['extra_css'] is None:
119 config['extra_css'] = extra_css
120
121 if config['extra_javascript'] is None:
122 config['extra_javascript'] = extra_javascript
123
124 package_dir = os.path.dirname(__file__)
125 theme_dir = [os.path.join(package_dir, 'themes', config['theme'])]
126
127 if config['theme_dir'] is not None:
128 theme_dir.insert(0, config['theme_dir'])
129
130 config['theme_dir'] = theme_dir
131
132 if config['repo_url'] is not None and config['repo_name'] is None:
133 repo_host = urlparse(config['repo_url']).netloc.lower()
134 if repo_host == 'github.com':
135 config['repo_name'] = 'GitHub'
136 elif repo_host == 'bitbucket.com':
137 config['repo_name'] = 'Bitbucket'
138 else:
139 config['repo_name'] = repo_host.split('.')[0].title()
140
141 if config['include_next_prev'] is None:
142 config['include_next_prev'] = len(config['pages']) > 1
143
144 if config['include_nav'] is None:
145 config['include_nav'] = len(config['pages']) > 1
146
147 # To Do:
148
149 # The docs dir must exist.
150 # The theme dir must exist.
151 # Ensure 'theme' is one of 'mkdocs', 'readthedocs', 'custom'
152 # A homepage 'index' must exist.
153 # The theme 'base.html' file must exist.
154 # Cannot set repo_name without setting repo_url.
155 # Cannot set 'include_next_prev: true' when only one page exists.
156 # Cannot set 'include_nav: true' when only one page exists.
157 # Error if any config keys provided that are not in the DEFAULT_CONFIG.
158
159 return config
160
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/mkdocs/config.py b/mkdocs/config.py
--- a/mkdocs/config.py
+++ b/mkdocs/config.py
@@ -74,11 +74,13 @@
def load_config(filename='mkdocs.yml', options=None):
options = options or {}
if 'config' in options:
- filename = options['config']
+ filename = options.pop('config')
if not os.path.exists(filename):
raise ConfigurationError("Config file '%s' does not exist." % filename)
with open(filename, 'r') as fp:
user_config = yaml.load(fp)
+ if not isinstance(user_config, dict):
+ raise ConfigurationError("The mkdocs.yml file is invalid. See http://www.mkdocs.org/user-guide/configuration/ for more information.")
user_config.update(options)
return validate_config(user_config)
| {"golden_diff": "diff --git a/mkdocs/config.py b/mkdocs/config.py\n--- a/mkdocs/config.py\n+++ b/mkdocs/config.py\n@@ -74,11 +74,13 @@\n def load_config(filename='mkdocs.yml', options=None):\n options = options or {}\n if 'config' in options:\n- filename = options['config']\n+ filename = options.pop('config')\n if not os.path.exists(filename):\n raise ConfigurationError(\"Config file '%s' does not exist.\" % filename)\n with open(filename, 'r') as fp:\n user_config = yaml.load(fp)\n+ if not isinstance(user_config, dict):\n+ raise ConfigurationError(\"The mkdocs.yml file is invalid. See http://www.mkdocs.org/user-guide/configuration/ for more information.\")\n user_config.update(options)\n return validate_config(user_config)\n", "issue": "If the mkdocs.yml is completely empty there is a traceback\n```\nTraceback (most recent call last):\n File \"/home/dougalmatthews/.virtualenvs/mkdocs/bin/mkdocs\", line 9, in <module>\n load_entry_point('mkdocs==0.11.1', 'console_scripts', 'mkdocs')()\n File \"/home/dougalmatthews/.virtualenvs/mkdocs/lib/python3.4/site-packages/mkdocs/main.py\", line 60, in run_main\n main(cmd, args=sys.argv[2:], options=dict(opts))\n File \"/home/dougalmatthews/.virtualenvs/mkdocs/lib/python3.4/site-packages/mkdocs/main.py\", line 32, in main\n config = load_config(options=options)\n File \"/home/dougalmatthews/.virtualenvs/mkdocs/lib/python3.4/site-packages/mkdocs/config.py\", line 82, in load_config\n user_config.update(options)\nAttributeError: 'NoneType' object has no attribute 'update'\n```\n\n", "before_files": [{"content": "# coding: utf-8\n\nfrom mkdocs import utils\nfrom mkdocs.compat import urlparse\nfrom mkdocs.exceptions import ConfigurationError\n\nimport os\nimport yaml\n\nDEFAULT_CONFIG = {\n 'site_name': None,\n 'pages': None,\n\n 'site_url': None,\n 'site_description': None,\n 'site_author': None,\n 'site_favicon': None,\n\n 'theme': 'mkdocs',\n 'docs_dir': 'docs',\n 'site_dir': 'site',\n 'theme_dir': None,\n\n 'copyright': None,\n 'google_analytics': None,\n\n # The address on which to serve the livereloading docs server.\n 'dev_addr': '127.0.0.1:8000',\n\n # If `True`, use `<page_name>/index.hmtl` style files with hyperlinks to the directory.\n # If `False`, use `<page_name>.html style file with hyperlinks to the file.\n # True generates nicer URLs, but False is useful if browsing the output on a filesystem.\n 'use_directory_urls': True,\n\n # Specify a link to the project source repo to be included\n # in the documentation pages.\n 'repo_url': None,\n\n # A name to use for the link to the project source repo.\n # Default: If repo_url is unset then None, otherwise\n # \"GitHub\" or \"Bitbucket\" for known url or Hostname for unknown urls.\n 'repo_name': None,\n\n # Specify which css or javascript files from the docs\n # directionary should be additionally included in the site.\n # Default: List of all .css and .js files in the docs dir.\n 'extra_css': None,\n 'extra_javascript': None,\n\n # Determine if the site should include the nav and next/prev elements.\n # Default: True if the site has more than one page, False otherwise.\n 'include_nav': None,\n 'include_next_prev': None,\n\n # PyMarkdown extension names.\n 'markdown_extensions': (),\n\n # Determine if the site should generate a json search index and include\n # search elements in the theme. - TODO\n 'include_search': False,\n\n # Determine if the site should include a 404.html page.\n # TODO: Implment this. Make this None, have it True if a 404.html\n # template exists in the theme or docs dir.\n 'include_404': False,\n\n # Determine if the site should include a sitemap.xml page.\n # TODO: Implement this. Make this None, have it True if a sitemap.xml\n # template exists in the theme or docs dir.\n 'include_sitemap': False,\n}\n\n\ndef load_config(filename='mkdocs.yml', options=None):\n options = options or {}\n if 'config' in options:\n filename = options['config']\n if not os.path.exists(filename):\n raise ConfigurationError(\"Config file '%s' does not exist.\" % filename)\n with open(filename, 'r') as fp:\n user_config = yaml.load(fp)\n user_config.update(options)\n return validate_config(user_config)\n\n\ndef validate_config(user_config):\n config = DEFAULT_CONFIG.copy()\n config.update(user_config)\n\n if not config['site_name']:\n raise ConfigurationError(\"Config must contain 'site_name' setting.\")\n\n # If not specified, then the 'pages' config simply includes all\n # markdown files in the docs dir, without generating any header items\n # for them.\n pages = []\n extra_css = []\n extra_javascript = []\n for (dirpath, dirnames, filenames) in os.walk(config['docs_dir']):\n for filename in sorted(filenames):\n fullpath = os.path.join(dirpath, filename)\n relpath = os.path.relpath(fullpath, config['docs_dir'])\n\n if utils.is_markdown_file(filename):\n # index pages should always be the first listed page.\n if os.path.splitext(relpath)[0] == 'index':\n pages.insert(0, relpath)\n else:\n pages.append(relpath)\n elif utils.is_css_file(filename):\n extra_css.append(relpath)\n elif utils.is_javascript_file(filename):\n extra_javascript.append(relpath)\n\n if config['pages'] is None:\n config['pages'] = pages\n\n if config['extra_css'] is None:\n config['extra_css'] = extra_css\n\n if config['extra_javascript'] is None:\n config['extra_javascript'] = extra_javascript\n\n package_dir = os.path.dirname(__file__)\n theme_dir = [os.path.join(package_dir, 'themes', config['theme'])]\n\n if config['theme_dir'] is not None:\n theme_dir.insert(0, config['theme_dir'])\n\n config['theme_dir'] = theme_dir\n\n if config['repo_url'] is not None and config['repo_name'] is None:\n repo_host = urlparse(config['repo_url']).netloc.lower()\n if repo_host == 'github.com':\n config['repo_name'] = 'GitHub'\n elif repo_host == 'bitbucket.com':\n config['repo_name'] = 'Bitbucket'\n else:\n config['repo_name'] = repo_host.split('.')[0].title()\n\n if config['include_next_prev'] is None:\n config['include_next_prev'] = len(config['pages']) > 1\n\n if config['include_nav'] is None:\n config['include_nav'] = len(config['pages']) > 1\n\n # To Do:\n\n # The docs dir must exist.\n # The theme dir must exist.\n # Ensure 'theme' is one of 'mkdocs', 'readthedocs', 'custom'\n # A homepage 'index' must exist.\n # The theme 'base.html' file must exist.\n # Cannot set repo_name without setting repo_url.\n # Cannot set 'include_next_prev: true' when only one page exists.\n # Cannot set 'include_nav: true' when only one page exists.\n # Error if any config keys provided that are not in the DEFAULT_CONFIG.\n\n return config\n", "path": "mkdocs/config.py"}], "after_files": [{"content": "# coding: utf-8\n\nfrom mkdocs import utils\nfrom mkdocs.compat import urlparse\nfrom mkdocs.exceptions import ConfigurationError\n\nimport os\nimport yaml\n\nDEFAULT_CONFIG = {\n 'site_name': None,\n 'pages': None,\n\n 'site_url': None,\n 'site_description': None,\n 'site_author': None,\n 'site_favicon': None,\n\n 'theme': 'mkdocs',\n 'docs_dir': 'docs',\n 'site_dir': 'site',\n 'theme_dir': None,\n\n 'copyright': None,\n 'google_analytics': None,\n\n # The address on which to serve the livereloading docs server.\n 'dev_addr': '127.0.0.1:8000',\n\n # If `True`, use `<page_name>/index.hmtl` style files with hyperlinks to the directory.\n # If `False`, use `<page_name>.html style file with hyperlinks to the file.\n # True generates nicer URLs, but False is useful if browsing the output on a filesystem.\n 'use_directory_urls': True,\n\n # Specify a link to the project source repo to be included\n # in the documentation pages.\n 'repo_url': None,\n\n # A name to use for the link to the project source repo.\n # Default: If repo_url is unset then None, otherwise\n # \"GitHub\" or \"Bitbucket\" for known url or Hostname for unknown urls.\n 'repo_name': None,\n\n # Specify which css or javascript files from the docs\n # directionary should be additionally included in the site.\n # Default: List of all .css and .js files in the docs dir.\n 'extra_css': None,\n 'extra_javascript': None,\n\n # Determine if the site should include the nav and next/prev elements.\n # Default: True if the site has more than one page, False otherwise.\n 'include_nav': None,\n 'include_next_prev': None,\n\n # PyMarkdown extension names.\n 'markdown_extensions': (),\n\n # Determine if the site should generate a json search index and include\n # search elements in the theme. - TODO\n 'include_search': False,\n\n # Determine if the site should include a 404.html page.\n # TODO: Implment this. Make this None, have it True if a 404.html\n # template exists in the theme or docs dir.\n 'include_404': False,\n\n # Determine if the site should include a sitemap.xml page.\n # TODO: Implement this. Make this None, have it True if a sitemap.xml\n # template exists in the theme or docs dir.\n 'include_sitemap': False,\n}\n\n\ndef load_config(filename='mkdocs.yml', options=None):\n options = options or {}\n if 'config' in options:\n filename = options.pop('config')\n if not os.path.exists(filename):\n raise ConfigurationError(\"Config file '%s' does not exist.\" % filename)\n with open(filename, 'r') as fp:\n user_config = yaml.load(fp)\n if not isinstance(user_config, dict):\n raise ConfigurationError(\"The mkdocs.yml file is invalid. See http://www.mkdocs.org/user-guide/configuration/ for more information.\")\n user_config.update(options)\n return validate_config(user_config)\n\n\ndef validate_config(user_config):\n config = DEFAULT_CONFIG.copy()\n config.update(user_config)\n\n if not config['site_name']:\n raise ConfigurationError(\"Config must contain 'site_name' setting.\")\n\n # If not specified, then the 'pages' config simply includes all\n # markdown files in the docs dir, without generating any header items\n # for them.\n pages = []\n extra_css = []\n extra_javascript = []\n for (dirpath, dirnames, filenames) in os.walk(config['docs_dir']):\n for filename in sorted(filenames):\n fullpath = os.path.join(dirpath, filename)\n relpath = os.path.relpath(fullpath, config['docs_dir'])\n\n if utils.is_markdown_file(filename):\n # index pages should always be the first listed page.\n if os.path.splitext(relpath)[0] == 'index':\n pages.insert(0, relpath)\n else:\n pages.append(relpath)\n elif utils.is_css_file(filename):\n extra_css.append(relpath)\n elif utils.is_javascript_file(filename):\n extra_javascript.append(relpath)\n\n if config['pages'] is None:\n config['pages'] = pages\n\n if config['extra_css'] is None:\n config['extra_css'] = extra_css\n\n if config['extra_javascript'] is None:\n config['extra_javascript'] = extra_javascript\n\n package_dir = os.path.dirname(__file__)\n theme_dir = [os.path.join(package_dir, 'themes', config['theme'])]\n\n if config['theme_dir'] is not None:\n theme_dir.insert(0, config['theme_dir'])\n\n config['theme_dir'] = theme_dir\n\n if config['repo_url'] is not None and config['repo_name'] is None:\n repo_host = urlparse(config['repo_url']).netloc.lower()\n if repo_host == 'github.com':\n config['repo_name'] = 'GitHub'\n elif repo_host == 'bitbucket.com':\n config['repo_name'] = 'Bitbucket'\n else:\n config['repo_name'] = repo_host.split('.')[0].title()\n\n if config['include_next_prev'] is None:\n config['include_next_prev'] = len(config['pages']) > 1\n\n if config['include_nav'] is None:\n config['include_nav'] = len(config['pages']) > 1\n\n # To Do:\n\n # The docs dir must exist.\n # The theme dir must exist.\n # Ensure 'theme' is one of 'mkdocs', 'readthedocs', 'custom'\n # A homepage 'index' must exist.\n # The theme 'base.html' file must exist.\n # Cannot set repo_name without setting repo_url.\n # Cannot set 'include_next_prev: true' when only one page exists.\n # Cannot set 'include_nav: true' when only one page exists.\n # Error if any config keys provided that are not in the DEFAULT_CONFIG.\n\n return config\n", "path": "mkdocs/config.py"}]} | 2,202 | 185 |
gh_patches_debug_5113 | rasdani/github-patches | git_diff | fedora-infra__bodhi-4102 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Bad characters in username
There's a bot with a bad username of `packagerbot/os-master01.phx2.fedoraproject.org` that makes CI tests failing.
https://bodhi.fedoraproject.org/users/packagerbot/os-master01.phx2.fedoraproject.org
I'm pushing a PR to safe check CI tests, but do we want to make Bodhi safe to bad usernames like this? Since usernames are from outside world, should we modify them in a safe way before storing in the database?
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `bodhi/server/services/user.py`
Content:
```
1 # Copyright 2014-2019 Red Hat, Inc. and others
2 #
3 # This file is part of Bodhi.
4 #
5 # This program is free software; you can redistribute it and/or
6 # modify it under the terms of the GNU General Public License
7 # as published by the Free Software Foundation; either version 2
8 # of the License, or (at your option) any later version.
9 #
10 # This program is distributed in the hope that it will be useful,
11 # but WITHOUT ANY WARRANTY; without even the implied warranty of
12 # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
13 # GNU General Public License for more details.
14 #
15 # You should have received a copy of the GNU General Public License
16 # along with this program; if not, write to the Free Software
17 # Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA.
18 """Defines API services that pertain to users."""
19 import math
20
21 from cornice import Service
22 from cornice.validators import colander_querystring_validator
23 from pyramid.exceptions import HTTPNotFound
24 from sqlalchemy import func, distinct
25 from sqlalchemy.sql import or_
26
27 from bodhi.server.models import Group, Update, User
28 from bodhi.server.validators import (validate_updates, validate_groups)
29 import bodhi.server.schemas
30 import bodhi.server.security
31 import bodhi.server.services.errors
32 import bodhi.server.services.updates
33
34
35 user = Service(name='user', path='/users/{name}',
36 description='Bodhi users',
37 # These we leave wide-open since these are only GETs
38 cors_origins=bodhi.server.security.cors_origins_ro)
39
40 users = Service(name='users', path='/users/',
41 description='Bodhi users',
42 # These we leave wide-open since these are only GETs
43 cors_origins=bodhi.server.security.cors_origins_ro)
44
45 users_rss = Service(name='users_rss', path='/rss/users/', description='Bodhi users RSS feed',
46 cors_origins=bodhi.server.security.cors_origins_ro)
47
48
49 @user.get(accept=("application/json", "text/json"), renderer="json",
50 error_handler=bodhi.server.services.errors.json_handler)
51 @user.get(accept=("application/javascript"), renderer="jsonp",
52 error_handler=bodhi.server.services.errors.json_handler)
53 @user.get(accept="text/html", renderer="user.html",
54 error_handler=bodhi.server.services.errors.html_handler)
55 def get_user(request):
56 """
57 Return a user given by username.
58
59 Args:
60 request (pyramid.request): The current request.
61 Returns:
62 dict: A dictionary with two keys. "user" maps to a dictionary representation of the User
63 object. "urls" maps to various URLs that describe various other objects related to the
64 user.
65 """
66 id = request.matchdict.get('name')
67 user = User.get(id)
68
69 if not user:
70 request.errors.add('body', 'name', 'No such user')
71 request.errors.status = HTTPNotFound.code
72 return
73
74 user = user.__json__(request)
75
76 # Throw some extra information in there
77 rurl = request.route_url # Just shorthand
78 urls = {
79 'comments_by': rurl('comments') + '?user=%s' % id,
80 'comments_on': rurl('comments') + '?update_owner=%s' % id,
81 'recent_updates': rurl('updates') + '?user=%s' % id,
82 'recent_overrides': rurl('overrides') + '?user=%s' % id,
83 'comments_by_rss': rurl('comments_rss') + '?user=%s' % id,
84 'comments_on_rss': rurl('comments_rss') + '?update_owner=%s' % id,
85 'recent_updates_rss': rurl('updates_rss') + '?user=%s' % id,
86 'recent_overrides_rss': rurl('overrides_rss') + '?user=%s' % id,
87 }
88
89 return dict(user=user, urls=urls)
90
91
92 validators = (
93 colander_querystring_validator,
94 validate_groups,
95 validate_updates,
96 )
97
98
99 @users.get(schema=bodhi.server.schemas.ListUserSchema,
100 accept=("application/json", "text/json"), renderer="json",
101 error_handler=bodhi.server.services.errors.json_handler,
102 validators=validators)
103 @users.get(schema=bodhi.server.schemas.ListUserSchema,
104 accept=("application/javascript"), renderer="jsonp",
105 error_handler=bodhi.server.services.errors.jsonp_handler,
106 validators=validators)
107 @users.get(schema=bodhi.server.schemas.ListUserSchema, renderer="rss",
108 accept=('application/atom+xml',),
109 error_handler=bodhi.server.services.errors.html_handler,
110 validators=validators)
111 @users_rss.get(schema=bodhi.server.schemas.ListUserSchema, renderer="rss",
112 error_handler=bodhi.server.services.errors.html_handler,
113 validators=validators)
114 def query_users(request):
115 """
116 Search for users by various criteria.
117
118 Args:
119 request (pyramid.request): The current web request.
120 Returns:
121 dict: A dictionary with the follow key mappings:
122 users: A list of users matching the search criteria.
123 page: The current page of results.
124 pages: The total number of pages available.
125 rows_per_page: The number of users on the page.
126 total: The total number of users matching the search criteria.
127 """
128 db = request.db
129 data = request.validated
130 query = db.query(User)
131
132 like = data.get('like')
133 if like is not None:
134 query = query.filter(or_(*[
135 User.name.like('%%%s%%' % like)
136 ]))
137
138 search = data.get('search')
139 if search is not None:
140 query = query.filter(User.name.ilike('%%%s%%' % search))
141
142 name = data.get('name')
143 if name is not None:
144 query = query.filter(User.name.like(name))
145
146 groups = data.get('groups')
147 if groups is not None:
148 query = query.join(User.groups)
149 query = query.filter(or_(*[Group.id == grp.id for grp in groups]))
150
151 updates = data.get('updates')
152 if updates is not None:
153 query = query.join(User.updates)
154 args = [Update.alias == update.alias for update in updates]
155 query = query.filter(or_(*args))
156
157 # We can't use ``query.count()`` here because it is naive with respect to
158 # all the joins that we're doing above.
159 count_query = query.with_labels().statement\
160 .with_only_columns([func.count(distinct(User.id))])\
161 .order_by(None)
162 total = request.db.execute(count_query).scalar()
163
164 page = data.get('page')
165 rows_per_page = data.get('rows_per_page')
166 pages = int(math.ceil(total / float(rows_per_page)))
167 query = query.offset(rows_per_page * (page - 1)).limit(rows_per_page)
168
169 return dict(
170 users=query.all(),
171 page=page,
172 pages=pages,
173 rows_per_page=rows_per_page,
174 total=total,
175 )
176
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/bodhi/server/services/user.py b/bodhi/server/services/user.py
--- a/bodhi/server/services/user.py
+++ b/bodhi/server/services/user.py
@@ -32,7 +32,7 @@
import bodhi.server.services.updates
-user = Service(name='user', path='/users/{name}',
+user = Service(name='user', path=r'/users/{name:\S+}',
description='Bodhi users',
# These we leave wide-open since these are only GETs
cors_origins=bodhi.server.security.cors_origins_ro)
| {"golden_diff": "diff --git a/bodhi/server/services/user.py b/bodhi/server/services/user.py\n--- a/bodhi/server/services/user.py\n+++ b/bodhi/server/services/user.py\n@@ -32,7 +32,7 @@\n import bodhi.server.services.updates\n \n \n-user = Service(name='user', path='/users/{name}',\n+user = Service(name='user', path=r'/users/{name:\\S+}',\n description='Bodhi users',\n # These we leave wide-open since these are only GETs\n cors_origins=bodhi.server.security.cors_origins_ro)\n", "issue": "Bad characters in username\nThere's a bot with a bad username of `packagerbot/os-master01.phx2.fedoraproject.org` that makes CI tests failing.\r\nhttps://bodhi.fedoraproject.org/users/packagerbot/os-master01.phx2.fedoraproject.org\r\n\r\nI'm pushing a PR to safe check CI tests, but do we want to make Bodhi safe to bad usernames like this? Since usernames are from outside world, should we modify them in a safe way before storing in the database?\n", "before_files": [{"content": "# Copyright 2014-2019 Red Hat, Inc. and others\n#\n# This file is part of Bodhi.\n#\n# This program is free software; you can redistribute it and/or\n# modify it under the terms of the GNU General Public License\n# as published by the Free Software Foundation; either version 2\n# of the License, or (at your option) any later version.\n#\n# This program is distributed in the hope that it will be useful,\n# but WITHOUT ANY WARRANTY; without even the implied warranty of\n# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the\n# GNU General Public License for more details.\n#\n# You should have received a copy of the GNU General Public License\n# along with this program; if not, write to the Free Software\n# Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA.\n\"\"\"Defines API services that pertain to users.\"\"\"\nimport math\n\nfrom cornice import Service\nfrom cornice.validators import colander_querystring_validator\nfrom pyramid.exceptions import HTTPNotFound\nfrom sqlalchemy import func, distinct\nfrom sqlalchemy.sql import or_\n\nfrom bodhi.server.models import Group, Update, User\nfrom bodhi.server.validators import (validate_updates, validate_groups)\nimport bodhi.server.schemas\nimport bodhi.server.security\nimport bodhi.server.services.errors\nimport bodhi.server.services.updates\n\n\nuser = Service(name='user', path='/users/{name}',\n description='Bodhi users',\n # These we leave wide-open since these are only GETs\n cors_origins=bodhi.server.security.cors_origins_ro)\n\nusers = Service(name='users', path='/users/',\n description='Bodhi users',\n # These we leave wide-open since these are only GETs\n cors_origins=bodhi.server.security.cors_origins_ro)\n\nusers_rss = Service(name='users_rss', path='/rss/users/', description='Bodhi users RSS feed',\n cors_origins=bodhi.server.security.cors_origins_ro)\n\n\[email protected](accept=(\"application/json\", \"text/json\"), renderer=\"json\",\n error_handler=bodhi.server.services.errors.json_handler)\[email protected](accept=(\"application/javascript\"), renderer=\"jsonp\",\n error_handler=bodhi.server.services.errors.json_handler)\[email protected](accept=\"text/html\", renderer=\"user.html\",\n error_handler=bodhi.server.services.errors.html_handler)\ndef get_user(request):\n \"\"\"\n Return a user given by username.\n\n Args:\n request (pyramid.request): The current request.\n Returns:\n dict: A dictionary with two keys. \"user\" maps to a dictionary representation of the User\n object. \"urls\" maps to various URLs that describe various other objects related to the\n user.\n \"\"\"\n id = request.matchdict.get('name')\n user = User.get(id)\n\n if not user:\n request.errors.add('body', 'name', 'No such user')\n request.errors.status = HTTPNotFound.code\n return\n\n user = user.__json__(request)\n\n # Throw some extra information in there\n rurl = request.route_url # Just shorthand\n urls = {\n 'comments_by': rurl('comments') + '?user=%s' % id,\n 'comments_on': rurl('comments') + '?update_owner=%s' % id,\n 'recent_updates': rurl('updates') + '?user=%s' % id,\n 'recent_overrides': rurl('overrides') + '?user=%s' % id,\n 'comments_by_rss': rurl('comments_rss') + '?user=%s' % id,\n 'comments_on_rss': rurl('comments_rss') + '?update_owner=%s' % id,\n 'recent_updates_rss': rurl('updates_rss') + '?user=%s' % id,\n 'recent_overrides_rss': rurl('overrides_rss') + '?user=%s' % id,\n }\n\n return dict(user=user, urls=urls)\n\n\nvalidators = (\n colander_querystring_validator,\n validate_groups,\n validate_updates,\n)\n\n\[email protected](schema=bodhi.server.schemas.ListUserSchema,\n accept=(\"application/json\", \"text/json\"), renderer=\"json\",\n error_handler=bodhi.server.services.errors.json_handler,\n validators=validators)\[email protected](schema=bodhi.server.schemas.ListUserSchema,\n accept=(\"application/javascript\"), renderer=\"jsonp\",\n error_handler=bodhi.server.services.errors.jsonp_handler,\n validators=validators)\[email protected](schema=bodhi.server.schemas.ListUserSchema, renderer=\"rss\",\n accept=('application/atom+xml',),\n error_handler=bodhi.server.services.errors.html_handler,\n validators=validators)\n@users_rss.get(schema=bodhi.server.schemas.ListUserSchema, renderer=\"rss\",\n error_handler=bodhi.server.services.errors.html_handler,\n validators=validators)\ndef query_users(request):\n \"\"\"\n Search for users by various criteria.\n\n Args:\n request (pyramid.request): The current web request.\n Returns:\n dict: A dictionary with the follow key mappings:\n users: A list of users matching the search criteria.\n page: The current page of results.\n pages: The total number of pages available.\n rows_per_page: The number of users on the page.\n total: The total number of users matching the search criteria.\n \"\"\"\n db = request.db\n data = request.validated\n query = db.query(User)\n\n like = data.get('like')\n if like is not None:\n query = query.filter(or_(*[\n User.name.like('%%%s%%' % like)\n ]))\n\n search = data.get('search')\n if search is not None:\n query = query.filter(User.name.ilike('%%%s%%' % search))\n\n name = data.get('name')\n if name is not None:\n query = query.filter(User.name.like(name))\n\n groups = data.get('groups')\n if groups is not None:\n query = query.join(User.groups)\n query = query.filter(or_(*[Group.id == grp.id for grp in groups]))\n\n updates = data.get('updates')\n if updates is not None:\n query = query.join(User.updates)\n args = [Update.alias == update.alias for update in updates]\n query = query.filter(or_(*args))\n\n # We can't use ``query.count()`` here because it is naive with respect to\n # all the joins that we're doing above.\n count_query = query.with_labels().statement\\\n .with_only_columns([func.count(distinct(User.id))])\\\n .order_by(None)\n total = request.db.execute(count_query).scalar()\n\n page = data.get('page')\n rows_per_page = data.get('rows_per_page')\n pages = int(math.ceil(total / float(rows_per_page)))\n query = query.offset(rows_per_page * (page - 1)).limit(rows_per_page)\n\n return dict(\n users=query.all(),\n page=page,\n pages=pages,\n rows_per_page=rows_per_page,\n total=total,\n )\n", "path": "bodhi/server/services/user.py"}], "after_files": [{"content": "# Copyright 2014-2019 Red Hat, Inc. and others\n#\n# This file is part of Bodhi.\n#\n# This program is free software; you can redistribute it and/or\n# modify it under the terms of the GNU General Public License\n# as published by the Free Software Foundation; either version 2\n# of the License, or (at your option) any later version.\n#\n# This program is distributed in the hope that it will be useful,\n# but WITHOUT ANY WARRANTY; without even the implied warranty of\n# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the\n# GNU General Public License for more details.\n#\n# You should have received a copy of the GNU General Public License\n# along with this program; if not, write to the Free Software\n# Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA.\n\"\"\"Defines API services that pertain to users.\"\"\"\nimport math\n\nfrom cornice import Service\nfrom cornice.validators import colander_querystring_validator\nfrom pyramid.exceptions import HTTPNotFound\nfrom sqlalchemy import func, distinct\nfrom sqlalchemy.sql import or_\n\nfrom bodhi.server.models import Group, Update, User\nfrom bodhi.server.validators import (validate_updates, validate_groups)\nimport bodhi.server.schemas\nimport bodhi.server.security\nimport bodhi.server.services.errors\nimport bodhi.server.services.updates\n\n\nuser = Service(name='user', path=r'/users/{name:\\S+}',\n description='Bodhi users',\n # These we leave wide-open since these are only GETs\n cors_origins=bodhi.server.security.cors_origins_ro)\n\nusers = Service(name='users', path='/users/',\n description='Bodhi users',\n # These we leave wide-open since these are only GETs\n cors_origins=bodhi.server.security.cors_origins_ro)\n\nusers_rss = Service(name='users_rss', path='/rss/users/', description='Bodhi users RSS feed',\n cors_origins=bodhi.server.security.cors_origins_ro)\n\n\[email protected](accept=(\"application/json\", \"text/json\"), renderer=\"json\",\n error_handler=bodhi.server.services.errors.json_handler)\[email protected](accept=(\"application/javascript\"), renderer=\"jsonp\",\n error_handler=bodhi.server.services.errors.json_handler)\[email protected](accept=\"text/html\", renderer=\"user.html\",\n error_handler=bodhi.server.services.errors.html_handler)\ndef get_user(request):\n \"\"\"\n Return a user given by username.\n\n Args:\n request (pyramid.request): The current request.\n Returns:\n dict: A dictionary with two keys. \"user\" maps to a dictionary representation of the User\n object. \"urls\" maps to various URLs that describe various other objects related to the\n user.\n \"\"\"\n id = request.matchdict.get('name')\n user = User.get(id)\n\n if not user:\n request.errors.add('body', 'name', 'No such user')\n request.errors.status = HTTPNotFound.code\n return\n\n user = user.__json__(request)\n\n # Throw some extra information in there\n rurl = request.route_url # Just shorthand\n urls = {\n 'comments_by': rurl('comments') + '?user=%s' % id,\n 'comments_on': rurl('comments') + '?update_owner=%s' % id,\n 'recent_updates': rurl('updates') + '?user=%s' % id,\n 'recent_overrides': rurl('overrides') + '?user=%s' % id,\n 'comments_by_rss': rurl('comments_rss') + '?user=%s' % id,\n 'comments_on_rss': rurl('comments_rss') + '?update_owner=%s' % id,\n 'recent_updates_rss': rurl('updates_rss') + '?user=%s' % id,\n 'recent_overrides_rss': rurl('overrides_rss') + '?user=%s' % id,\n }\n\n return dict(user=user, urls=urls)\n\n\nvalidators = (\n colander_querystring_validator,\n validate_groups,\n validate_updates,\n)\n\n\[email protected](schema=bodhi.server.schemas.ListUserSchema,\n accept=(\"application/json\", \"text/json\"), renderer=\"json\",\n error_handler=bodhi.server.services.errors.json_handler,\n validators=validators)\[email protected](schema=bodhi.server.schemas.ListUserSchema,\n accept=(\"application/javascript\"), renderer=\"jsonp\",\n error_handler=bodhi.server.services.errors.jsonp_handler,\n validators=validators)\[email protected](schema=bodhi.server.schemas.ListUserSchema, renderer=\"rss\",\n accept=('application/atom+xml',),\n error_handler=bodhi.server.services.errors.html_handler,\n validators=validators)\n@users_rss.get(schema=bodhi.server.schemas.ListUserSchema, renderer=\"rss\",\n error_handler=bodhi.server.services.errors.html_handler,\n validators=validators)\ndef query_users(request):\n \"\"\"\n Search for users by various criteria.\n\n Args:\n request (pyramid.request): The current web request.\n Returns:\n dict: A dictionary with the follow key mappings:\n users: A list of users matching the search criteria.\n page: The current page of results.\n pages: The total number of pages available.\n rows_per_page: The number of users on the page.\n total: The total number of users matching the search criteria.\n \"\"\"\n db = request.db\n data = request.validated\n query = db.query(User)\n\n like = data.get('like')\n if like is not None:\n query = query.filter(or_(*[\n User.name.like('%%%s%%' % like)\n ]))\n\n search = data.get('search')\n if search is not None:\n query = query.filter(User.name.ilike('%%%s%%' % search))\n\n name = data.get('name')\n if name is not None:\n query = query.filter(User.name.like(name))\n\n groups = data.get('groups')\n if groups is not None:\n query = query.join(User.groups)\n query = query.filter(or_(*[Group.id == grp.id for grp in groups]))\n\n updates = data.get('updates')\n if updates is not None:\n query = query.join(User.updates)\n args = [Update.alias == update.alias for update in updates]\n query = query.filter(or_(*args))\n\n # We can't use ``query.count()`` here because it is naive with respect to\n # all the joins that we're doing above.\n count_query = query.with_labels().statement\\\n .with_only_columns([func.count(distinct(User.id))])\\\n .order_by(None)\n total = request.db.execute(count_query).scalar()\n\n page = data.get('page')\n rows_per_page = data.get('rows_per_page')\n pages = int(math.ceil(total / float(rows_per_page)))\n query = query.offset(rows_per_page * (page - 1)).limit(rows_per_page)\n\n return dict(\n users=query.all(),\n page=page,\n pages=pages,\n rows_per_page=rows_per_page,\n total=total,\n )\n", "path": "bodhi/server/services/user.py"}]} | 2,340 | 129 |
gh_patches_debug_39224 | rasdani/github-patches | git_diff | spack__spack-12207 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Spack broken on Blue Waters
On the current develop, no Spack command works on Blue Waters.
### Steps to reproduce the issue
Any Spack command:
```console
$ spack help
```
### Error Message
```
Traceback (most recent call last):
File "/u/sciteam/stewart1/spack/bin/spack", line 48, in <module>
sys.exit(spack.main.main())
File "/mnt/a/u/sciteam/stewart1/spack/lib/spack/spack/main.py", line 704, in main
if spack.config.get('config:debug'):
File "/mnt/a/u/sciteam/stewart1/spack/lib/spack/spack/config.py", line 627, in get
return config.get(path, default, scope)
File "/mnt/a/u/sciteam/stewart1/spack/lib/spack/llnl/util/lang.py", line 558, in __getattr__
return getattr(self.instance, name)
File "/mnt/a/u/sciteam/stewart1/spack/lib/spack/llnl/util/lang.py", line 554, in instance
self._instance = self.factory()
File "/mnt/a/u/sciteam/stewart1/spack/lib/spack/spack/config.py", line 609, in _config
_add_platform_scope(cfg, ConfigScope, name, path)
File "/mnt/a/u/sciteam/stewart1/spack/lib/spack/spack/config.py", line 562, in _add_platform_scope
platform = spack.architecture.platform().name
File "/mnt/a/u/sciteam/stewart1/spack/lib/spack/llnl/util/lang.py", line 184, in _memoized_function
func.cache[args] = func(*args)
File "/mnt/a/u/sciteam/stewart1/spack/lib/spack/spack/architecture.py", line 388, in platform
return platform_cls()
File "/mnt/a/u/sciteam/stewart1/spack/lib/spack/spack/platforms/cray.py", line 76, in __init__
back_distro = Cnl()
File "/mnt/a/u/sciteam/stewart1/spack/lib/spack/spack/operating_systems/cnl.py", line 57, in __init__
version = self._detect_crayos_version()
File "/mnt/a/u/sciteam/stewart1/spack/lib/spack/spack/operating_systems/cnl.py", line 66, in _detect_crayos_version
release_attrs = read_cle_release_file()
File "/mnt/a/u/sciteam/stewart1/spack/lib/spack/spack/operating_systems/cnl.py", line 37, in read_cle_release_file
with open(_cle_release_file) as release_file:
IOError: [Errno 2] No such file or directory: '/etc/opt/cray/release/cle-release'
```
### Information on your system
```console
$ cat /etc/*-release
Cluster Manager v6.1
slave
LSB_VERSION="core-2.0-noarch:core-3.2-noarch:core-4.0-noarch:core-2.0-x86_64:core-3.2-x86_64:core-4.0-x86_64"
SUSE Linux Enterprise Server 11 (x86_64)
VERSION = 11
PATCHLEVEL = 3
$ uname -a
Linux h2ologin2 3.0.101-0.47.106.59-default #1 SMP Wed Jan 23 09:00:24 UTC 2019 (624897e) x86_64 x86_64 x86_64 GNU/Linux
```
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `lib/spack/spack/operating_systems/cnl.py`
Content:
```
1 # Copyright 2013-2019 Lawrence Livermore National Security, LLC and other
2 # Spack Project Developers. See the top-level COPYRIGHT file for details.
3 #
4 # SPDX-License-Identifier: (Apache-2.0 OR MIT)
5
6 import re
7
8 import llnl.util.tty as tty
9
10 import spack.version
11 from spack.architecture import OperatingSystem
12 from spack.util.module_cmd import module
13
14 #: Location of the Cray CLE release file, which we look at to get the CNL
15 #: OS version.
16 _cle_release_file = '/etc/opt/cray/release/cle-release'
17
18
19 def read_cle_release_file():
20 """Read the CLE release file and return a dict with its attributes.
21
22 The release file looks something like this::
23
24 RELEASE=6.0.UP07
25 BUILD=6.0.7424
26 ...
27
28 The dictionary we produce looks like this::
29
30 {
31 "RELEASE": "6.0.UP07",
32 "BUILD": "6.0.7424",
33 ...
34 }
35
36 """
37 with open(_cle_release_file) as release_file:
38 result = {}
39 for line in release_file:
40 # use partition instead of split() to ensure we only split on
41 # the first '=' in the line.
42 key, _, value = line.partition('=')
43 result[key] = value.strip()
44 return result
45
46
47 class Cnl(OperatingSystem):
48 """ Compute Node Linux (CNL) is the operating system used for the Cray XC
49 series super computers. It is a very stripped down version of GNU/Linux.
50 Any compilers found through this operating system will be used with
51 modules. If updated, user must make sure that version and name are
52 updated to indicate that OS has been upgraded (or downgraded)
53 """
54
55 def __init__(self):
56 name = 'cnl'
57 version = self._detect_crayos_version()
58 super(Cnl, self).__init__(name, version)
59 self.modulecmd = module
60
61 def __str__(self):
62 return self.name + str(self.version)
63
64 @classmethod
65 def _detect_crayos_version(cls):
66 release_attrs = read_cle_release_file()
67 v = spack.version.Version(release_attrs['RELEASE'])
68 return v[0]
69
70 def arguments_to_detect_version_fn(self, paths):
71 import spack.compilers
72
73 command_arguments = []
74 for compiler_name in spack.compilers.supported_compilers():
75 cmp_cls = spack.compilers.class_for_compiler_name(compiler_name)
76
77 # If the compiler doesn't have a corresponding
78 # Programming Environment, skip to the next
79 if cmp_cls.PrgEnv is None:
80 continue
81
82 if cmp_cls.PrgEnv_compiler is None:
83 tty.die('Must supply PrgEnv_compiler with PrgEnv')
84
85 compiler_id = spack.compilers.CompilerID(self, compiler_name, None)
86 detect_version_args = spack.compilers.DetectVersionArgs(
87 id=compiler_id, variation=(None, None),
88 language='cc', path='cc'
89 )
90 command_arguments.append(detect_version_args)
91 return command_arguments
92
93 def detect_version(self, detect_version_args):
94 import spack.compilers
95 modulecmd = self.modulecmd
96 compiler_name = detect_version_args.id.compiler_name
97 compiler_cls = spack.compilers.class_for_compiler_name(compiler_name)
98 output = modulecmd('avail', compiler_cls.PrgEnv_compiler)
99 version_regex = r'(%s)/([\d\.]+[\d])' % compiler_cls.PrgEnv_compiler
100 matches = re.findall(version_regex, output)
101 version = tuple(version for _, version in matches)
102 compiler_id = detect_version_args.id
103 value = detect_version_args._replace(
104 id=compiler_id._replace(version=version)
105 )
106 return value, None
107
108 def make_compilers(self, compiler_id, paths):
109 import spack.spec
110 name = compiler_id.compiler_name
111 cmp_cls = spack.compilers.class_for_compiler_name(name)
112 compilers = []
113 for v in compiler_id.version:
114 comp = cmp_cls(
115 spack.spec.CompilerSpec(name + '@' + v),
116 self, "any",
117 ['cc', 'CC', 'ftn'], [cmp_cls.PrgEnv, name + '/' + v])
118
119 compilers.append(comp)
120 return compilers
121
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/lib/spack/spack/operating_systems/cnl.py b/lib/spack/spack/operating_systems/cnl.py
--- a/lib/spack/spack/operating_systems/cnl.py
+++ b/lib/spack/spack/operating_systems/cnl.py
@@ -3,22 +3,27 @@
#
# SPDX-License-Identifier: (Apache-2.0 OR MIT)
+import os
import re
import llnl.util.tty as tty
+import spack.error
import spack.version
from spack.architecture import OperatingSystem
from spack.util.module_cmd import module
-#: Location of the Cray CLE release file, which we look at to get the CNL
-#: OS version.
+#: Possible locations of the Cray CLE release file,
+#: which we look at to get the CNL OS version.
_cle_release_file = '/etc/opt/cray/release/cle-release'
+_clerelease_file = '/etc/opt/cray/release/clerelease'
def read_cle_release_file():
"""Read the CLE release file and return a dict with its attributes.
+ This file is present on newer versions of Cray.
+
The release file looks something like this::
RELEASE=6.0.UP07
@@ -33,6 +38,8 @@
...
}
+ Returns:
+ dict: dictionary of release attributes
"""
with open(_cle_release_file) as release_file:
result = {}
@@ -44,8 +51,25 @@
return result
+def read_clerelease_file():
+ """Read the CLE release file and return the Cray OS version.
+
+ This file is present on older versions of Cray.
+
+ The release file looks something like this::
+
+ 5.2.UP04
+
+ Returns:
+ str: the Cray OS version
+ """
+ with open(_clerelease_file) as release_file:
+ for line in release_file:
+ return line.strip()
+
+
class Cnl(OperatingSystem):
- """ Compute Node Linux (CNL) is the operating system used for the Cray XC
+ """Compute Node Linux (CNL) is the operating system used for the Cray XC
series super computers. It is a very stripped down version of GNU/Linux.
Any compilers found through this operating system will be used with
modules. If updated, user must make sure that version and name are
@@ -63,9 +87,16 @@
@classmethod
def _detect_crayos_version(cls):
- release_attrs = read_cle_release_file()
- v = spack.version.Version(release_attrs['RELEASE'])
- return v[0]
+ if os.path.isfile(_cle_release_file):
+ release_attrs = read_cle_release_file()
+ v = spack.version.Version(release_attrs['RELEASE'])
+ return v[0]
+ elif os.path.isfile(_clerelease_file):
+ v = read_clerelease_file()
+ return spack.version.Version(v)[0]
+ else:
+ raise spack.error.UnsupportedPlatformError(
+ 'Unable to detect Cray OS version')
def arguments_to_detect_version_fn(self, paths):
import spack.compilers
| {"golden_diff": "diff --git a/lib/spack/spack/operating_systems/cnl.py b/lib/spack/spack/operating_systems/cnl.py\n--- a/lib/spack/spack/operating_systems/cnl.py\n+++ b/lib/spack/spack/operating_systems/cnl.py\n@@ -3,22 +3,27 @@\n #\n # SPDX-License-Identifier: (Apache-2.0 OR MIT)\n \n+import os\n import re\n \n import llnl.util.tty as tty\n \n+import spack.error\n import spack.version\n from spack.architecture import OperatingSystem\n from spack.util.module_cmd import module\n \n-#: Location of the Cray CLE release file, which we look at to get the CNL\n-#: OS version.\n+#: Possible locations of the Cray CLE release file,\n+#: which we look at to get the CNL OS version.\n _cle_release_file = '/etc/opt/cray/release/cle-release'\n+_clerelease_file = '/etc/opt/cray/release/clerelease'\n \n \n def read_cle_release_file():\n \"\"\"Read the CLE release file and return a dict with its attributes.\n \n+ This file is present on newer versions of Cray.\n+\n The release file looks something like this::\n \n RELEASE=6.0.UP07\n@@ -33,6 +38,8 @@\n ...\n }\n \n+ Returns:\n+ dict: dictionary of release attributes\n \"\"\"\n with open(_cle_release_file) as release_file:\n result = {}\n@@ -44,8 +51,25 @@\n return result\n \n \n+def read_clerelease_file():\n+ \"\"\"Read the CLE release file and return the Cray OS version.\n+\n+ This file is present on older versions of Cray.\n+\n+ The release file looks something like this::\n+\n+ 5.2.UP04\n+\n+ Returns:\n+ str: the Cray OS version\n+ \"\"\"\n+ with open(_clerelease_file) as release_file:\n+ for line in release_file:\n+ return line.strip()\n+\n+\n class Cnl(OperatingSystem):\n- \"\"\" Compute Node Linux (CNL) is the operating system used for the Cray XC\n+ \"\"\"Compute Node Linux (CNL) is the operating system used for the Cray XC\n series super computers. It is a very stripped down version of GNU/Linux.\n Any compilers found through this operating system will be used with\n modules. If updated, user must make sure that version and name are\n@@ -63,9 +87,16 @@\n \n @classmethod\n def _detect_crayos_version(cls):\n- release_attrs = read_cle_release_file()\n- v = spack.version.Version(release_attrs['RELEASE'])\n- return v[0]\n+ if os.path.isfile(_cle_release_file):\n+ release_attrs = read_cle_release_file()\n+ v = spack.version.Version(release_attrs['RELEASE'])\n+ return v[0]\n+ elif os.path.isfile(_clerelease_file):\n+ v = read_clerelease_file()\n+ return spack.version.Version(v)[0]\n+ else:\n+ raise spack.error.UnsupportedPlatformError(\n+ 'Unable to detect Cray OS version')\n \n def arguments_to_detect_version_fn(self, paths):\n import spack.compilers\n", "issue": "Spack broken on Blue Waters\nOn the current develop, no Spack command works on Blue Waters.\r\n\r\n### Steps to reproduce the issue\r\n\r\nAny Spack command:\r\n```console\r\n$ spack help\r\n```\r\n\r\n### Error Message\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"/u/sciteam/stewart1/spack/bin/spack\", line 48, in <module>\r\n sys.exit(spack.main.main())\r\n File \"/mnt/a/u/sciteam/stewart1/spack/lib/spack/spack/main.py\", line 704, in main\r\n if spack.config.get('config:debug'):\r\n File \"/mnt/a/u/sciteam/stewart1/spack/lib/spack/spack/config.py\", line 627, in get\r\n return config.get(path, default, scope)\r\n File \"/mnt/a/u/sciteam/stewart1/spack/lib/spack/llnl/util/lang.py\", line 558, in __getattr__\r\n return getattr(self.instance, name)\r\n File \"/mnt/a/u/sciteam/stewart1/spack/lib/spack/llnl/util/lang.py\", line 554, in instance\r\n self._instance = self.factory()\r\n File \"/mnt/a/u/sciteam/stewart1/spack/lib/spack/spack/config.py\", line 609, in _config\r\n _add_platform_scope(cfg, ConfigScope, name, path)\r\n File \"/mnt/a/u/sciteam/stewart1/spack/lib/spack/spack/config.py\", line 562, in _add_platform_scope\r\n platform = spack.architecture.platform().name\r\n File \"/mnt/a/u/sciteam/stewart1/spack/lib/spack/llnl/util/lang.py\", line 184, in _memoized_function\r\n func.cache[args] = func(*args)\r\n File \"/mnt/a/u/sciteam/stewart1/spack/lib/spack/spack/architecture.py\", line 388, in platform\r\n return platform_cls()\r\n File \"/mnt/a/u/sciteam/stewart1/spack/lib/spack/spack/platforms/cray.py\", line 76, in __init__\r\n back_distro = Cnl()\r\n File \"/mnt/a/u/sciteam/stewart1/spack/lib/spack/spack/operating_systems/cnl.py\", line 57, in __init__\r\n version = self._detect_crayos_version()\r\n File \"/mnt/a/u/sciteam/stewart1/spack/lib/spack/spack/operating_systems/cnl.py\", line 66, in _detect_crayos_version\r\n release_attrs = read_cle_release_file()\r\n File \"/mnt/a/u/sciteam/stewart1/spack/lib/spack/spack/operating_systems/cnl.py\", line 37, in read_cle_release_file\r\n with open(_cle_release_file) as release_file:\r\nIOError: [Errno 2] No such file or directory: '/etc/opt/cray/release/cle-release'\r\n```\r\n\r\n### Information on your system\r\n\r\n```console\r\n$ cat /etc/*-release\r\nCluster Manager v6.1\r\nslave\r\nLSB_VERSION=\"core-2.0-noarch:core-3.2-noarch:core-4.0-noarch:core-2.0-x86_64:core-3.2-x86_64:core-4.0-x86_64\"\r\nSUSE Linux Enterprise Server 11 (x86_64)\r\nVERSION = 11\r\nPATCHLEVEL = 3\r\n$ uname -a\r\nLinux h2ologin2 3.0.101-0.47.106.59-default #1 SMP Wed Jan 23 09:00:24 UTC 2019 (624897e) x86_64 x86_64 x86_64 GNU/Linux\r\n```\n", "before_files": [{"content": "# Copyright 2013-2019 Lawrence Livermore National Security, LLC and other\n# Spack Project Developers. See the top-level COPYRIGHT file for details.\n#\n# SPDX-License-Identifier: (Apache-2.0 OR MIT)\n\nimport re\n\nimport llnl.util.tty as tty\n\nimport spack.version\nfrom spack.architecture import OperatingSystem\nfrom spack.util.module_cmd import module\n\n#: Location of the Cray CLE release file, which we look at to get the CNL\n#: OS version.\n_cle_release_file = '/etc/opt/cray/release/cle-release'\n\n\ndef read_cle_release_file():\n \"\"\"Read the CLE release file and return a dict with its attributes.\n\n The release file looks something like this::\n\n RELEASE=6.0.UP07\n BUILD=6.0.7424\n ...\n\n The dictionary we produce looks like this::\n\n {\n \"RELEASE\": \"6.0.UP07\",\n \"BUILD\": \"6.0.7424\",\n ...\n }\n\n \"\"\"\n with open(_cle_release_file) as release_file:\n result = {}\n for line in release_file:\n # use partition instead of split() to ensure we only split on\n # the first '=' in the line.\n key, _, value = line.partition('=')\n result[key] = value.strip()\n return result\n\n\nclass Cnl(OperatingSystem):\n \"\"\" Compute Node Linux (CNL) is the operating system used for the Cray XC\n series super computers. It is a very stripped down version of GNU/Linux.\n Any compilers found through this operating system will be used with\n modules. If updated, user must make sure that version and name are\n updated to indicate that OS has been upgraded (or downgraded)\n \"\"\"\n\n def __init__(self):\n name = 'cnl'\n version = self._detect_crayos_version()\n super(Cnl, self).__init__(name, version)\n self.modulecmd = module\n\n def __str__(self):\n return self.name + str(self.version)\n\n @classmethod\n def _detect_crayos_version(cls):\n release_attrs = read_cle_release_file()\n v = spack.version.Version(release_attrs['RELEASE'])\n return v[0]\n\n def arguments_to_detect_version_fn(self, paths):\n import spack.compilers\n\n command_arguments = []\n for compiler_name in spack.compilers.supported_compilers():\n cmp_cls = spack.compilers.class_for_compiler_name(compiler_name)\n\n # If the compiler doesn't have a corresponding\n # Programming Environment, skip to the next\n if cmp_cls.PrgEnv is None:\n continue\n\n if cmp_cls.PrgEnv_compiler is None:\n tty.die('Must supply PrgEnv_compiler with PrgEnv')\n\n compiler_id = spack.compilers.CompilerID(self, compiler_name, None)\n detect_version_args = spack.compilers.DetectVersionArgs(\n id=compiler_id, variation=(None, None),\n language='cc', path='cc'\n )\n command_arguments.append(detect_version_args)\n return command_arguments\n\n def detect_version(self, detect_version_args):\n import spack.compilers\n modulecmd = self.modulecmd\n compiler_name = detect_version_args.id.compiler_name\n compiler_cls = spack.compilers.class_for_compiler_name(compiler_name)\n output = modulecmd('avail', compiler_cls.PrgEnv_compiler)\n version_regex = r'(%s)/([\\d\\.]+[\\d])' % compiler_cls.PrgEnv_compiler\n matches = re.findall(version_regex, output)\n version = tuple(version for _, version in matches)\n compiler_id = detect_version_args.id\n value = detect_version_args._replace(\n id=compiler_id._replace(version=version)\n )\n return value, None\n\n def make_compilers(self, compiler_id, paths):\n import spack.spec\n name = compiler_id.compiler_name\n cmp_cls = spack.compilers.class_for_compiler_name(name)\n compilers = []\n for v in compiler_id.version:\n comp = cmp_cls(\n spack.spec.CompilerSpec(name + '@' + v),\n self, \"any\",\n ['cc', 'CC', 'ftn'], [cmp_cls.PrgEnv, name + '/' + v])\n\n compilers.append(comp)\n return compilers\n", "path": "lib/spack/spack/operating_systems/cnl.py"}], "after_files": [{"content": "# Copyright 2013-2019 Lawrence Livermore National Security, LLC and other\n# Spack Project Developers. See the top-level COPYRIGHT file for details.\n#\n# SPDX-License-Identifier: (Apache-2.0 OR MIT)\n\nimport os\nimport re\n\nimport llnl.util.tty as tty\n\nimport spack.error\nimport spack.version\nfrom spack.architecture import OperatingSystem\nfrom spack.util.module_cmd import module\n\n#: Possible locations of the Cray CLE release file,\n#: which we look at to get the CNL OS version.\n_cle_release_file = '/etc/opt/cray/release/cle-release'\n_clerelease_file = '/etc/opt/cray/release/clerelease'\n\n\ndef read_cle_release_file():\n \"\"\"Read the CLE release file and return a dict with its attributes.\n\n This file is present on newer versions of Cray.\n\n The release file looks something like this::\n\n RELEASE=6.0.UP07\n BUILD=6.0.7424\n ...\n\n The dictionary we produce looks like this::\n\n {\n \"RELEASE\": \"6.0.UP07\",\n \"BUILD\": \"6.0.7424\",\n ...\n }\n\n Returns:\n dict: dictionary of release attributes\n \"\"\"\n with open(_cle_release_file) as release_file:\n result = {}\n for line in release_file:\n # use partition instead of split() to ensure we only split on\n # the first '=' in the line.\n key, _, value = line.partition('=')\n result[key] = value.strip()\n return result\n\n\ndef read_clerelease_file():\n \"\"\"Read the CLE release file and return the Cray OS version.\n\n This file is present on older versions of Cray.\n\n The release file looks something like this::\n\n 5.2.UP04\n\n Returns:\n str: the Cray OS version\n \"\"\"\n with open(_clerelease_file) as release_file:\n for line in release_file:\n return line.strip()\n\n\nclass Cnl(OperatingSystem):\n \"\"\"Compute Node Linux (CNL) is the operating system used for the Cray XC\n series super computers. It is a very stripped down version of GNU/Linux.\n Any compilers found through this operating system will be used with\n modules. If updated, user must make sure that version and name are\n updated to indicate that OS has been upgraded (or downgraded)\n \"\"\"\n\n def __init__(self):\n name = 'cnl'\n version = self._detect_crayos_version()\n super(Cnl, self).__init__(name, version)\n self.modulecmd = module\n\n def __str__(self):\n return self.name + str(self.version)\n\n @classmethod\n def _detect_crayos_version(cls):\n if os.path.isfile(_cle_release_file):\n release_attrs = read_cle_release_file()\n v = spack.version.Version(release_attrs['RELEASE'])\n return v[0]\n elif os.path.isfile(_clerelease_file):\n v = read_clerelease_file()\n return spack.version.Version(v)[0]\n else:\n raise spack.error.UnsupportedPlatformError(\n 'Unable to detect Cray OS version')\n\n def arguments_to_detect_version_fn(self, paths):\n import spack.compilers\n\n command_arguments = []\n for compiler_name in spack.compilers.supported_compilers():\n cmp_cls = spack.compilers.class_for_compiler_name(compiler_name)\n\n # If the compiler doesn't have a corresponding\n # Programming Environment, skip to the next\n if cmp_cls.PrgEnv is None:\n continue\n\n if cmp_cls.PrgEnv_compiler is None:\n tty.die('Must supply PrgEnv_compiler with PrgEnv')\n\n compiler_id = spack.compilers.CompilerID(self, compiler_name, None)\n detect_version_args = spack.compilers.DetectVersionArgs(\n id=compiler_id, variation=(None, None),\n language='cc', path='cc'\n )\n command_arguments.append(detect_version_args)\n return command_arguments\n\n def detect_version(self, detect_version_args):\n import spack.compilers\n modulecmd = self.modulecmd\n compiler_name = detect_version_args.id.compiler_name\n compiler_cls = spack.compilers.class_for_compiler_name(compiler_name)\n output = modulecmd('avail', compiler_cls.PrgEnv_compiler)\n version_regex = r'(%s)/([\\d\\.]+[\\d])' % compiler_cls.PrgEnv_compiler\n matches = re.findall(version_regex, output)\n version = tuple(version for _, version in matches)\n compiler_id = detect_version_args.id\n value = detect_version_args._replace(\n id=compiler_id._replace(version=version)\n )\n return value, None\n\n def make_compilers(self, compiler_id, paths):\n import spack.spec\n name = compiler_id.compiler_name\n cmp_cls = spack.compilers.class_for_compiler_name(name)\n compilers = []\n for v in compiler_id.version:\n comp = cmp_cls(\n spack.spec.CompilerSpec(name + '@' + v),\n self, \"any\",\n ['cc', 'CC', 'ftn'], [cmp_cls.PrgEnv, name + '/' + v])\n\n compilers.append(comp)\n return compilers\n", "path": "lib/spack/spack/operating_systems/cnl.py"}]} | 2,331 | 732 |
gh_patches_debug_24151 | rasdani/github-patches | git_diff | gammapy__gammapy-4924 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Remove deprecated features
This is a reminder issue to remove the features deprecated since 1.1 before the next release
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `gammapy/utils/table.py`
Content:
```
1 # Licensed under a 3-clause BSD style license - see LICENSE.rst
2 """Table helper utilities."""
3 import numpy as np
4 from astropy.table import Table
5 from astropy.units import Quantity
6 from .deprecation import deprecated
7 from .units import standardise_unit
8
9 __all__ = [
10 "hstack_columns",
11 "table_from_row_data",
12 "table_row_to_dict",
13 "table_standardise_units_copy",
14 "table_standardise_units_inplace",
15 ]
16
17
18 def hstack_columns(table, table_other):
19 """Stack the column data horizontally.
20
21 Parameters
22 ----------
23 table : `~astropy.table.Table`
24 Input table.
25 table_other : `~astropy.table.Table`
26 Other input table.
27
28 Returns
29 -------
30 stacked : `~astropy.table.Table`
31 Stacked table.
32 """
33 stacked = Table()
34
35 for column in table.colnames:
36 data = np.hstack([table[column].data[0], table_other[column].data[0]])
37 stacked[column] = data[np.newaxis, :]
38 return stacked
39
40
41 def table_standardise_units_copy(table):
42 """Standardise units for all columns in a table in a copy.
43
44 Calls `~gammapy.utils.units.standardise_unit`.
45
46 Parameters
47 ----------
48 table : `~astropy.table.Table`
49 Input table (won't be modified).
50
51 Returns
52 -------
53 table : `~astropy.table.Table`
54 Copy of the input table with standardised column units.
55 """
56 # Note: we could add an `inplace` option (or variant of this function)
57 # See https://github.com/astropy/astropy/issues/6098
58 table = Table(table)
59 return table_standardise_units_inplace(table)
60
61
62 def table_standardise_units_inplace(table):
63 """Standardise units for all columns in a table in place."""
64 for column in table.columns.values():
65 if column.unit:
66 column.unit = standardise_unit(column.unit)
67
68 return table
69
70
71 def table_row_to_dict(row, make_quantity=True):
72 """Make one source data dictionary.
73
74 Parameters
75 ----------
76 row : `~astropy.table.Row`
77 Row.
78 make_quantity : bool, optional
79 Make quantity values for columns with units.
80 Default is True.
81
82 Returns
83 -------
84 data : dict
85 Row data.
86 """
87 data = {}
88 for name, col in row.columns.items():
89 val = row[name]
90
91 if make_quantity and col.unit:
92 val = Quantity(val, unit=col.unit)
93 data[name] = val
94 return data
95
96
97 @deprecated("v1.1", alternative="astropy.table.Table")
98 def table_from_row_data(rows, **kwargs):
99 """Helper function to create table objects from row data.
100
101 Works with quantities.
102
103 Parameters
104 ----------
105 rows : list
106 List of row data (each row a dictionary).
107 """
108 table = Table(**kwargs)
109
110 if len(rows) == 0:
111 return table
112
113 colnames = list(rows[0].keys())
114
115 for name in colnames:
116 coldata = [_[name] for _ in rows]
117 if isinstance(rows[0][name], Quantity):
118 coldata = Quantity(coldata, unit=rows[0][name].unit)
119 table[name] = coldata
120
121 return table
122
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/gammapy/utils/table.py b/gammapy/utils/table.py
--- a/gammapy/utils/table.py
+++ b/gammapy/utils/table.py
@@ -3,12 +3,10 @@
import numpy as np
from astropy.table import Table
from astropy.units import Quantity
-from .deprecation import deprecated
from .units import standardise_unit
__all__ = [
"hstack_columns",
- "table_from_row_data",
"table_row_to_dict",
"table_standardise_units_copy",
"table_standardise_units_inplace",
@@ -92,30 +90,3 @@
val = Quantity(val, unit=col.unit)
data[name] = val
return data
-
-
-@deprecated("v1.1", alternative="astropy.table.Table")
-def table_from_row_data(rows, **kwargs):
- """Helper function to create table objects from row data.
-
- Works with quantities.
-
- Parameters
- ----------
- rows : list
- List of row data (each row a dictionary).
- """
- table = Table(**kwargs)
-
- if len(rows) == 0:
- return table
-
- colnames = list(rows[0].keys())
-
- for name in colnames:
- coldata = [_[name] for _ in rows]
- if isinstance(rows[0][name], Quantity):
- coldata = Quantity(coldata, unit=rows[0][name].unit)
- table[name] = coldata
-
- return table
| {"golden_diff": "diff --git a/gammapy/utils/table.py b/gammapy/utils/table.py\n--- a/gammapy/utils/table.py\n+++ b/gammapy/utils/table.py\n@@ -3,12 +3,10 @@\n import numpy as np\n from astropy.table import Table\n from astropy.units import Quantity\n-from .deprecation import deprecated\n from .units import standardise_unit\n \n __all__ = [\n \"hstack_columns\",\n- \"table_from_row_data\",\n \"table_row_to_dict\",\n \"table_standardise_units_copy\",\n \"table_standardise_units_inplace\",\n@@ -92,30 +90,3 @@\n val = Quantity(val, unit=col.unit)\n data[name] = val\n return data\n-\n-\n-@deprecated(\"v1.1\", alternative=\"astropy.table.Table\")\n-def table_from_row_data(rows, **kwargs):\n- \"\"\"Helper function to create table objects from row data.\n-\n- Works with quantities.\n-\n- Parameters\n- ----------\n- rows : list\n- List of row data (each row a dictionary).\n- \"\"\"\n- table = Table(**kwargs)\n-\n- if len(rows) == 0:\n- return table\n-\n- colnames = list(rows[0].keys())\n-\n- for name in colnames:\n- coldata = [_[name] for _ in rows]\n- if isinstance(rows[0][name], Quantity):\n- coldata = Quantity(coldata, unit=rows[0][name].unit)\n- table[name] = coldata\n-\n- return table\n", "issue": "Remove deprecated features\nThis is a reminder issue to remove the features deprecated since 1.1 before the next release\n", "before_files": [{"content": "# Licensed under a 3-clause BSD style license - see LICENSE.rst\n\"\"\"Table helper utilities.\"\"\"\nimport numpy as np\nfrom astropy.table import Table\nfrom astropy.units import Quantity\nfrom .deprecation import deprecated\nfrom .units import standardise_unit\n\n__all__ = [\n \"hstack_columns\",\n \"table_from_row_data\",\n \"table_row_to_dict\",\n \"table_standardise_units_copy\",\n \"table_standardise_units_inplace\",\n]\n\n\ndef hstack_columns(table, table_other):\n \"\"\"Stack the column data horizontally.\n\n Parameters\n ----------\n table : `~astropy.table.Table`\n Input table.\n table_other : `~astropy.table.Table`\n Other input table.\n\n Returns\n -------\n stacked : `~astropy.table.Table`\n Stacked table.\n \"\"\"\n stacked = Table()\n\n for column in table.colnames:\n data = np.hstack([table[column].data[0], table_other[column].data[0]])\n stacked[column] = data[np.newaxis, :]\n return stacked\n\n\ndef table_standardise_units_copy(table):\n \"\"\"Standardise units for all columns in a table in a copy.\n\n Calls `~gammapy.utils.units.standardise_unit`.\n\n Parameters\n ----------\n table : `~astropy.table.Table`\n Input table (won't be modified).\n\n Returns\n -------\n table : `~astropy.table.Table`\n Copy of the input table with standardised column units.\n \"\"\"\n # Note: we could add an `inplace` option (or variant of this function)\n # See https://github.com/astropy/astropy/issues/6098\n table = Table(table)\n return table_standardise_units_inplace(table)\n\n\ndef table_standardise_units_inplace(table):\n \"\"\"Standardise units for all columns in a table in place.\"\"\"\n for column in table.columns.values():\n if column.unit:\n column.unit = standardise_unit(column.unit)\n\n return table\n\n\ndef table_row_to_dict(row, make_quantity=True):\n \"\"\"Make one source data dictionary.\n\n Parameters\n ----------\n row : `~astropy.table.Row`\n Row.\n make_quantity : bool, optional\n Make quantity values for columns with units.\n Default is True.\n\n Returns\n -------\n data : dict\n Row data.\n \"\"\"\n data = {}\n for name, col in row.columns.items():\n val = row[name]\n\n if make_quantity and col.unit:\n val = Quantity(val, unit=col.unit)\n data[name] = val\n return data\n\n\n@deprecated(\"v1.1\", alternative=\"astropy.table.Table\")\ndef table_from_row_data(rows, **kwargs):\n \"\"\"Helper function to create table objects from row data.\n\n Works with quantities.\n\n Parameters\n ----------\n rows : list\n List of row data (each row a dictionary).\n \"\"\"\n table = Table(**kwargs)\n\n if len(rows) == 0:\n return table\n\n colnames = list(rows[0].keys())\n\n for name in colnames:\n coldata = [_[name] for _ in rows]\n if isinstance(rows[0][name], Quantity):\n coldata = Quantity(coldata, unit=rows[0][name].unit)\n table[name] = coldata\n\n return table\n", "path": "gammapy/utils/table.py"}], "after_files": [{"content": "# Licensed under a 3-clause BSD style license - see LICENSE.rst\n\"\"\"Table helper utilities.\"\"\"\nimport numpy as np\nfrom astropy.table import Table\nfrom astropy.units import Quantity\nfrom .units import standardise_unit\n\n__all__ = [\n \"hstack_columns\",\n \"table_row_to_dict\",\n \"table_standardise_units_copy\",\n \"table_standardise_units_inplace\",\n]\n\n\ndef hstack_columns(table, table_other):\n \"\"\"Stack the column data horizontally.\n\n Parameters\n ----------\n table : `~astropy.table.Table`\n Input table.\n table_other : `~astropy.table.Table`\n Other input table.\n\n Returns\n -------\n stacked : `~astropy.table.Table`\n Stacked table.\n \"\"\"\n stacked = Table()\n\n for column in table.colnames:\n data = np.hstack([table[column].data[0], table_other[column].data[0]])\n stacked[column] = data[np.newaxis, :]\n return stacked\n\n\ndef table_standardise_units_copy(table):\n \"\"\"Standardise units for all columns in a table in a copy.\n\n Calls `~gammapy.utils.units.standardise_unit`.\n\n Parameters\n ----------\n table : `~astropy.table.Table`\n Input table (won't be modified).\n\n Returns\n -------\n table : `~astropy.table.Table`\n Copy of the input table with standardised column units.\n \"\"\"\n # Note: we could add an `inplace` option (or variant of this function)\n # See https://github.com/astropy/astropy/issues/6098\n table = Table(table)\n return table_standardise_units_inplace(table)\n\n\ndef table_standardise_units_inplace(table):\n \"\"\"Standardise units for all columns in a table in place.\"\"\"\n for column in table.columns.values():\n if column.unit:\n column.unit = standardise_unit(column.unit)\n\n return table\n\n\ndef table_row_to_dict(row, make_quantity=True):\n \"\"\"Make one source data dictionary.\n\n Parameters\n ----------\n row : `~astropy.table.Row`\n Row.\n make_quantity : bool, optional\n Make quantity values for columns with units.\n Default is True.\n\n Returns\n -------\n data : dict\n Row data.\n \"\"\"\n data = {}\n for name, col in row.columns.items():\n val = row[name]\n\n if make_quantity and col.unit:\n val = Quantity(val, unit=col.unit)\n data[name] = val\n return data\n", "path": "gammapy/utils/table.py"}]} | 1,267 | 347 |
gh_patches_debug_476 | rasdani/github-patches | git_diff | rlworkgroup__garage-2133 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Unpin cloudpickle instead of pinning it to 1.3
Currently, #1879 pins cloudpickle to 1.3 because tensorflow-probability 0.11 does so. When tfp unpins cloudpickle, we should unpin it too.
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `setup.py`
Content:
```
1 """setuptools based setup module."""
2 import os
3
4 from setuptools import find_packages, setup
5
6 GARAGE_GH_TOKEN = os.environ.get('GARAGE_GH_TOKEN') or 'git'
7 GYM_VERSION = '0.17.2'
8
9 # Required dependencies
10 REQUIRED = [
11 # Please keep alphabetized
12 'akro',
13 'click>=2.0',
14 'cloudpickle==1.3',
15 'cma==2.7.0',
16 'dowel==0.0.3',
17 'numpy>=1.14.5',
18 'psutil',
19 'python-dateutil',
20 'ray',
21 'scikit-image',
22 'scipy',
23 'setproctitle>=1.0',
24 'tensorflow>=1.14',
25 'tensorflow-probability>=0.11.0',
26 'torch>=1.0.0,!=1.5.0',
27 'torchvision>=0.2.1',
28 ]
29
30 # Dependencies for optional features
31 EXTRAS = {}
32
33 EXTRAS['gym'] = [
34 f'gym[atari,box2d,classic_control]=={GYM_VERSION}',
35 ]
36
37 EXTRAS['mujoco'] = [
38 'mujoco-py>=2.0,<=2.0.2.8',
39 f'gym[all]=={GYM_VERSION}',
40 ]
41
42 EXTRAS['dm_control'] = [
43 # dm_control throws an error during install about not being able to
44 # find a build dependency (absl-py). Later pip executes the `install`
45 # command again and the install succeeds because absl-py has been
46 # installed. This is stupid, but harmless.
47 'dm_control',
48 ]
49
50 EXTRAS['bullet'] = ['mpi4py', 'pybullet>=2.8.7']
51
52 EXTRAS['all'] = list(set(sum(EXTRAS.values(), [])))
53
54 # Development dependencies (*not* included in 'all')
55 EXTRAS['dev'] = [
56 # Please keep alphabetized
57 'flake8',
58 'flake8-docstrings>=1.5.0',
59 'flake8-import-order',
60 f'metaworld @ https://{GARAGE_GH_TOKEN}@api.github.com/repos/rlworkgroup/metaworld/tarball/0875192baaa91c43523708f55866d98eaf3facaf', # noqa: E501
61 'isort>=4.3.21,<5.0.0',
62 'pep8-naming==0.7.0',
63 'pre-commit',
64 'pycodestyle>=2.5.0',
65 'pydocstyle>=4.0.0',
66 'pylint>=2.5.3',
67 'pytest>=4.5.0', # Required for strict-markers
68 'pytest-cov',
69 'pytest-rerunfailures',
70 'pytest-timeout',
71 'pytest-xdist',
72 'recommonmark',
73 'sphinx',
74 'sphinx-autoapi>=1.4.0',
75 'sphinx_rtd_theme',
76 'sphinxcontrib-bibtex',
77 'yapf==0.30.0',
78 ] # yapf: disable
79
80 with open('README.md') as f:
81 README = f.read()
82
83 # Get the package version dynamically
84 with open('VERSION') as v:
85 VERSION = v.read().strip()
86
87 setup(
88 name='garage',
89 version=VERSION,
90 author='Reinforcement Learning Working Group',
91 description='A toolkit for reproducible reinforcement learning research',
92 url='https://github.com/rlworkgroup/garage',
93 packages=find_packages(where='src'),
94 package_dir={'': 'src'},
95 scripts=['scripts/garage'],
96 python_requires='>=3.6',
97 install_requires=REQUIRED,
98 extras_require=EXTRAS,
99 license='MIT',
100 long_description=README,
101 long_description_content_type='text/markdown',
102 classifiers=[
103 'Development Status :: 4 - Beta',
104 'Intended Audience :: Developers',
105 'Intended Audience :: Education',
106 'Intended Audience :: Science/Research',
107 'License :: OSI Approved :: MIT License',
108 'Programming Language :: Python :: 3.6',
109 'Programming Language :: Python :: 3.7',
110 'Programming Language :: Python :: 3 :: Only',
111 'Topic :: Scientific/Engineering :: Artificial Intelligence',
112 'Topic :: Scientific/Engineering :: Mathematics',
113 'Topic :: Software Development :: Libraries',
114 ],
115 )
116
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/setup.py b/setup.py
--- a/setup.py
+++ b/setup.py
@@ -11,7 +11,7 @@
# Please keep alphabetized
'akro',
'click>=2.0',
- 'cloudpickle==1.3',
+ 'cloudpickle',
'cma==2.7.0',
'dowel==0.0.3',
'numpy>=1.14.5',
| {"golden_diff": "diff --git a/setup.py b/setup.py\n--- a/setup.py\n+++ b/setup.py\n@@ -11,7 +11,7 @@\n # Please keep alphabetized\n 'akro',\n 'click>=2.0',\n- 'cloudpickle==1.3',\n+ 'cloudpickle',\n 'cma==2.7.0',\n 'dowel==0.0.3',\n 'numpy>=1.14.5',\n", "issue": "Unpin cloudpickle instead of pinning it to 1.3\nCurrently, #1879 pins cloudpickle to 1.3 because tensorflow-probability 0.11 does so. When tfp unpins cloudpickle, we should unpin it too.\n", "before_files": [{"content": "\"\"\"setuptools based setup module.\"\"\"\nimport os\n\nfrom setuptools import find_packages, setup\n\nGARAGE_GH_TOKEN = os.environ.get('GARAGE_GH_TOKEN') or 'git'\nGYM_VERSION = '0.17.2'\n\n# Required dependencies\nREQUIRED = [\n # Please keep alphabetized\n 'akro',\n 'click>=2.0',\n 'cloudpickle==1.3',\n 'cma==2.7.0',\n 'dowel==0.0.3',\n 'numpy>=1.14.5',\n 'psutil',\n 'python-dateutil',\n 'ray',\n 'scikit-image',\n 'scipy',\n 'setproctitle>=1.0',\n 'tensorflow>=1.14',\n 'tensorflow-probability>=0.11.0',\n 'torch>=1.0.0,!=1.5.0',\n 'torchvision>=0.2.1',\n]\n\n# Dependencies for optional features\nEXTRAS = {}\n\nEXTRAS['gym'] = [\n f'gym[atari,box2d,classic_control]=={GYM_VERSION}',\n]\n\nEXTRAS['mujoco'] = [\n 'mujoco-py>=2.0,<=2.0.2.8',\n f'gym[all]=={GYM_VERSION}',\n]\n\nEXTRAS['dm_control'] = [\n # dm_control throws an error during install about not being able to\n # find a build dependency (absl-py). Later pip executes the `install`\n # command again and the install succeeds because absl-py has been\n # installed. This is stupid, but harmless.\n 'dm_control',\n]\n\nEXTRAS['bullet'] = ['mpi4py', 'pybullet>=2.8.7']\n\nEXTRAS['all'] = list(set(sum(EXTRAS.values(), [])))\n\n# Development dependencies (*not* included in 'all')\nEXTRAS['dev'] = [\n # Please keep alphabetized\n 'flake8',\n 'flake8-docstrings>=1.5.0',\n 'flake8-import-order',\n f'metaworld @ https://{GARAGE_GH_TOKEN}@api.github.com/repos/rlworkgroup/metaworld/tarball/0875192baaa91c43523708f55866d98eaf3facaf', # noqa: E501\n 'isort>=4.3.21,<5.0.0',\n 'pep8-naming==0.7.0',\n 'pre-commit',\n 'pycodestyle>=2.5.0',\n 'pydocstyle>=4.0.0',\n 'pylint>=2.5.3',\n 'pytest>=4.5.0', # Required for strict-markers\n 'pytest-cov',\n 'pytest-rerunfailures',\n 'pytest-timeout',\n 'pytest-xdist',\n 'recommonmark',\n 'sphinx',\n 'sphinx-autoapi>=1.4.0',\n 'sphinx_rtd_theme',\n 'sphinxcontrib-bibtex',\n 'yapf==0.30.0',\n] # yapf: disable\n\nwith open('README.md') as f:\n README = f.read()\n\n# Get the package version dynamically\nwith open('VERSION') as v:\n VERSION = v.read().strip()\n\nsetup(\n name='garage',\n version=VERSION,\n author='Reinforcement Learning Working Group',\n description='A toolkit for reproducible reinforcement learning research',\n url='https://github.com/rlworkgroup/garage',\n packages=find_packages(where='src'),\n package_dir={'': 'src'},\n scripts=['scripts/garage'],\n python_requires='>=3.6',\n install_requires=REQUIRED,\n extras_require=EXTRAS,\n license='MIT',\n long_description=README,\n long_description_content_type='text/markdown',\n classifiers=[\n 'Development Status :: 4 - Beta',\n 'Intended Audience :: Developers',\n 'Intended Audience :: Education',\n 'Intended Audience :: Science/Research',\n 'License :: OSI Approved :: MIT License',\n 'Programming Language :: Python :: 3.6',\n 'Programming Language :: Python :: 3.7',\n 'Programming Language :: Python :: 3 :: Only',\n 'Topic :: Scientific/Engineering :: Artificial Intelligence',\n 'Topic :: Scientific/Engineering :: Mathematics',\n 'Topic :: Software Development :: Libraries',\n ],\n)\n", "path": "setup.py"}], "after_files": [{"content": "\"\"\"setuptools based setup module.\"\"\"\nimport os\n\nfrom setuptools import find_packages, setup\n\nGARAGE_GH_TOKEN = os.environ.get('GARAGE_GH_TOKEN') or 'git'\nGYM_VERSION = '0.17.2'\n\n# Required dependencies\nREQUIRED = [\n # Please keep alphabetized\n 'akro',\n 'click>=2.0',\n 'cloudpickle',\n 'cma==2.7.0',\n 'dowel==0.0.3',\n 'numpy>=1.14.5',\n 'psutil',\n 'python-dateutil',\n 'ray',\n 'scikit-image',\n 'scipy',\n 'setproctitle>=1.0',\n 'tensorflow>=1.14',\n 'tensorflow-probability>=0.11.0',\n 'torch>=1.0.0,!=1.5.0',\n 'torchvision>=0.2.1',\n]\n\n# Dependencies for optional features\nEXTRAS = {}\n\nEXTRAS['gym'] = [\n f'gym[atari,box2d,classic_control]=={GYM_VERSION}',\n]\n\nEXTRAS['mujoco'] = [\n 'mujoco-py>=2.0,<=2.0.2.8',\n f'gym[all]=={GYM_VERSION}',\n]\n\nEXTRAS['dm_control'] = [\n # dm_control throws an error during install about not being able to\n # find a build dependency (absl-py). Later pip executes the `install`\n # command again and the install succeeds because absl-py has been\n # installed. This is stupid, but harmless.\n 'dm_control',\n]\n\nEXTRAS['bullet'] = ['mpi4py', 'pybullet>=2.8.7']\n\nEXTRAS['all'] = list(set(sum(EXTRAS.values(), [])))\n\n# Development dependencies (*not* included in 'all')\nEXTRAS['dev'] = [\n # Please keep alphabetized\n 'flake8',\n 'flake8-docstrings>=1.5.0',\n 'flake8-import-order',\n f'metaworld @ https://{GARAGE_GH_TOKEN}@api.github.com/repos/rlworkgroup/metaworld/tarball/0875192baaa91c43523708f55866d98eaf3facaf', # noqa: E501\n 'isort>=4.3.21,<5.0.0',\n 'pep8-naming==0.7.0',\n 'pre-commit',\n 'pycodestyle>=2.5.0',\n 'pydocstyle>=4.0.0',\n 'pylint>=2.5.3',\n 'pytest>=4.5.0', # Required for strict-markers\n 'pytest-cov',\n 'pytest-rerunfailures',\n 'pytest-timeout',\n 'pytest-xdist',\n 'recommonmark',\n 'sphinx',\n 'sphinx-autoapi>=1.4.0',\n 'sphinx_rtd_theme',\n 'sphinxcontrib-bibtex',\n 'yapf==0.30.0',\n] # yapf: disable\n\nwith open('README.md') as f:\n README = f.read()\n\n# Get the package version dynamically\nwith open('VERSION') as v:\n VERSION = v.read().strip()\n\nsetup(\n name='garage',\n version=VERSION,\n author='Reinforcement Learning Working Group',\n description='A toolkit for reproducible reinforcement learning research',\n url='https://github.com/rlworkgroup/garage',\n packages=find_packages(where='src'),\n package_dir={'': 'src'},\n scripts=['scripts/garage'],\n python_requires='>=3.6',\n install_requires=REQUIRED,\n extras_require=EXTRAS,\n license='MIT',\n long_description=README,\n long_description_content_type='text/markdown',\n classifiers=[\n 'Development Status :: 4 - Beta',\n 'Intended Audience :: Developers',\n 'Intended Audience :: Education',\n 'Intended Audience :: Science/Research',\n 'License :: OSI Approved :: MIT License',\n 'Programming Language :: Python :: 3.6',\n 'Programming Language :: Python :: 3.7',\n 'Programming Language :: Python :: 3 :: Only',\n 'Topic :: Scientific/Engineering :: Artificial Intelligence',\n 'Topic :: Scientific/Engineering :: Mathematics',\n 'Topic :: Software Development :: Libraries',\n ],\n)\n", "path": "setup.py"}]} | 1,555 | 101 |
gh_patches_debug_59565 | rasdani/github-patches | git_diff | saulpw__visidata-509 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
[html saver] Saving typed columns as html (int/vlen/bool) causes exception
I tried to copy (yank) a couple of rows from the frequency sheet and it provided me the following error. I believe this is due to the html parser expecting strings? A similar error also occurs in other sheets when using unexpected py types (e.g. bool).
FrequencySheet error
```
Traceback (most recent call last):
File "/Documents/pyv/py3/lib/python3.7/site-packages/visidata/threads.py", line 201, in _toplevelTryFunc
t.status = func(*args, **kwargs)
File "/Documents/pyv/py3/lib/python3.7/site-packages/visidata/loaders/html.py", line 124, in save_html
fp.write(html.escape(val))
File "/usr/local/Cellar/python/3.7.6_1/Frameworks/Python.framework/Versions/3.7/lib/python3.7/html/__init__.py", line 19, in escape
s = s.replace("&", "&") # Must be done first!
AttributeError: 'vlen' object has no attribute 'replace'
```
Sheet with a bool column error:
```
Traceback (most recent call last):
File "/Documents/pyv/py3/lib/python3.7/site-packages/visidata/threads.py", line 201, in _toplevelTryFunc
t.status = func(*args, **kwargs)
File "/Documents/pyv/py3/lib/python3.7/site-packages/visidata/loaders/html.py", line 124, in save_html
fp.write(html.escape(val))
File "/usr/local/Cellar/python/3.7.6_1/Frameworks/Python.framework/Versions/3.7/lib/python3.7/html/__init__.py", line 19, in escape
s = s.replace("&", "&") # Must be done first!
AttributeError: 'bool' object has no attribute 'replace'
```
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `visidata/loaders/html.py`
Content:
```
1 import html
2 from visidata import *
3
4
5 class HtmlTablesSheet(IndexSheet):
6 rowtype = 'sheets' # rowdef: HtmlTableSheet (sheet.html = lxml.html.HtmlElement)
7 columns = IndexSheet.columns + [
8 Column('tag', width=0, getter=lambda col,row: row.html.tag),
9 Column('id', getter=lambda col,row: row.html.attrib.get('id')),
10 Column('classes', getter=lambda col,row: row.html.attrib.get('class')),
11 ]
12 def iterload(self):
13 import lxml.html
14 from lxml import etree
15 utf8_parser = etree.HTMLParser(encoding='utf-8')
16 with self.source.open_text() as fp:
17 html = lxml.html.etree.parse(fp, parser=utf8_parser)
18 self.setKeys([self.column('name')])
19 self.column('keys').hide()
20 self.column('source').hide()
21
22 for i, e in enumerate(html.iter('table')):
23 if e.tag == 'table':
24 vs = HtmlTableSheet(e.attrib.get("id", "table_" + str(i)), source=e)
25 vs.reload()
26 vs.html = e
27 yield vs
28
29
30 def is_header(elem):
31 scope = elem.attrib.get('scope', '')
32
33 if elem.tag == 'th':
34 if not scope or scope == 'col':
35 return True
36
37 return False
38
39 class HtmlTableSheet(Sheet):
40 rowtype = 'rows' # list of strings
41 columns = []
42
43 def iterload(self):
44 headers = []
45
46 maxlinks = {} # [colnum] -> nlinks:int
47
48 for rownum, r in enumerate(self.source.iter('tr')):
49 row = []
50
51 colnum = 0
52 # get starting column, which might be different if there were rowspan>1 already
53 if rownum < len(headers):
54 while colnum < len(headers[rownum]):
55 if headers[rownum][colnum] is None:
56 break
57 colnum += 1
58
59 for cell in r.getchildren():
60 colspan = int(cell.attrib.get('colspan', 1))
61 rowspan = int(cell.attrib.get('rowspan', 1))
62 cellval = ' '.join(x.strip() for x in cell.itertext()) # text only without markup
63 links = [x.get('href') for x in cell.iter('a')]
64 maxlinks[colnum] = max(maxlinks.get(colnum, 0), len(links))
65
66 if is_header(cell):
67 for k in range(rownum, rownum+rowspan):
68 while k >= len(headers): # extend headers list with lists for all header rows
69 headers.append([])
70
71 for j in range(colnum, colnum+colspan):
72 while j >= len(headers[k]):
73 headers[k].append(None)
74 headers[k][j] = cellval
75 cellval = '' # use empty non-None value for subsequent rows in the rowspan
76 else:
77 while colnum >= len(row):
78 row.append(None)
79 row[colnum] = (cellval, links)
80
81 colnum += colspan
82
83 if any(row):
84 yield row
85
86 self.columns = []
87 if headers:
88 it = itertools.zip_longest(*headers, fillvalue='')
89 else:
90 it = [list(x) for x in self.rows[0]]
91 self.rows = self.rows[1:]
92
93 for colnum, names in enumerate(it):
94 name = '_'.join(str(x) for x in names if x)
95 self.addColumn(Column(name, getter=lambda c,r,i=colnum: r[i][0]))
96 for linknum in range(maxlinks.get(colnum, 0)):
97 self.addColumn(Column(name+'_link'+str(linknum), width=20, getter=lambda c,r,i=colnum,j=linknum: r[i][1][j]))
98
99
100 @VisiData.api
101 def save_html(vd, p, *vsheets):
102 'Save vsheets as HTML tables in a single file'
103
104 with open(p, 'w', encoding='ascii', errors='xmlcharrefreplace') as fp:
105 for sheet in vsheets:
106
107 fp.write('<h2 class="sheetname">%s</h2>\n'.format(sheetname=html.escape(sheet.name)))
108
109 fp.write('<table id="{sheetname}">\n'.format(sheetname=html.escape(sheet.name)))
110
111 # headers
112 fp.write('<tr>')
113 for col in sheet.visibleCols:
114 contents = html.escape(col.name)
115 fp.write('<th>{colname}</th>'.format(colname=contents))
116 fp.write('</tr>\n')
117
118 # rows
119 with Progress(gerund='saving'):
120 for typedvals in sheet.iterdispvals(format=False):
121 fp.write('<tr>')
122 for col, val in typedvals.items():
123 fp.write('<td>')
124 fp.write(html.escape(val))
125 fp.write('</td>')
126 fp.write('</tr>\n')
127
128 fp.write('</table>')
129 vd.status('%s save finished' % p)
130
131
132 VisiData.save_htm = VisiData.save_html
133
134
135 vd.filetype('html', HtmlTablesSheet)
136 vd.filetype('htm', HtmlTablesSheet)
137
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/visidata/loaders/html.py b/visidata/loaders/html.py
--- a/visidata/loaders/html.py
+++ b/visidata/loaders/html.py
@@ -121,7 +121,7 @@
fp.write('<tr>')
for col, val in typedvals.items():
fp.write('<td>')
- fp.write(html.escape(val))
+ fp.write(html.escape(str(val)))
fp.write('</td>')
fp.write('</tr>\n')
| {"golden_diff": "diff --git a/visidata/loaders/html.py b/visidata/loaders/html.py\n--- a/visidata/loaders/html.py\n+++ b/visidata/loaders/html.py\n@@ -121,7 +121,7 @@\n fp.write('<tr>')\n for col, val in typedvals.items():\n fp.write('<td>')\n- fp.write(html.escape(val))\n+ fp.write(html.escape(str(val)))\n fp.write('</td>')\n fp.write('</tr>\\n')\n", "issue": "[html saver] Saving typed columns as html (int/vlen/bool) causes exception\nI tried to copy (yank) a couple of rows from the frequency sheet and it provided me the following error. I believe this is due to the html parser expecting strings? A similar error also occurs in other sheets when using unexpected py types (e.g. bool).\r\n\r\nFrequencySheet error\r\n```\r\nTraceback (most recent call last):\r\n File \"/Documents/pyv/py3/lib/python3.7/site-packages/visidata/threads.py\", line 201, in _toplevelTryFunc\r\n t.status = func(*args, **kwargs)\r\n File \"/Documents/pyv/py3/lib/python3.7/site-packages/visidata/loaders/html.py\", line 124, in save_html\r\n fp.write(html.escape(val))\r\n File \"/usr/local/Cellar/python/3.7.6_1/Frameworks/Python.framework/Versions/3.7/lib/python3.7/html/__init__.py\", line 19, in escape\r\n s = s.replace(\"&\", \"&\") # Must be done first!\r\nAttributeError: 'vlen' object has no attribute 'replace'\r\n```\r\n\r\nSheet with a bool column error:\r\n```\r\nTraceback (most recent call last):\r\n File \"/Documents/pyv/py3/lib/python3.7/site-packages/visidata/threads.py\", line 201, in _toplevelTryFunc\r\n t.status = func(*args, **kwargs)\r\n File \"/Documents/pyv/py3/lib/python3.7/site-packages/visidata/loaders/html.py\", line 124, in save_html\r\n fp.write(html.escape(val))\r\n File \"/usr/local/Cellar/python/3.7.6_1/Frameworks/Python.framework/Versions/3.7/lib/python3.7/html/__init__.py\", line 19, in escape\r\n s = s.replace(\"&\", \"&\") # Must be done first!\r\nAttributeError: 'bool' object has no attribute 'replace'\r\n```\n", "before_files": [{"content": "import html\nfrom visidata import *\n\n\nclass HtmlTablesSheet(IndexSheet):\n rowtype = 'sheets' # rowdef: HtmlTableSheet (sheet.html = lxml.html.HtmlElement)\n columns = IndexSheet.columns + [\n Column('tag', width=0, getter=lambda col,row: row.html.tag),\n Column('id', getter=lambda col,row: row.html.attrib.get('id')),\n Column('classes', getter=lambda col,row: row.html.attrib.get('class')),\n ]\n def iterload(self):\n import lxml.html\n from lxml import etree\n utf8_parser = etree.HTMLParser(encoding='utf-8')\n with self.source.open_text() as fp:\n html = lxml.html.etree.parse(fp, parser=utf8_parser)\n self.setKeys([self.column('name')])\n self.column('keys').hide()\n self.column('source').hide()\n\n for i, e in enumerate(html.iter('table')):\n if e.tag == 'table':\n vs = HtmlTableSheet(e.attrib.get(\"id\", \"table_\" + str(i)), source=e)\n vs.reload()\n vs.html = e\n yield vs\n\n\ndef is_header(elem):\n scope = elem.attrib.get('scope', '')\n\n if elem.tag == 'th':\n if not scope or scope == 'col':\n return True\n\n return False\n\nclass HtmlTableSheet(Sheet):\n rowtype = 'rows' # list of strings\n columns = []\n\n def iterload(self):\n headers = []\n\n maxlinks = {} # [colnum] -> nlinks:int\n\n for rownum, r in enumerate(self.source.iter('tr')):\n row = []\n\n colnum = 0\n # get starting column, which might be different if there were rowspan>1 already\n if rownum < len(headers):\n while colnum < len(headers[rownum]):\n if headers[rownum][colnum] is None:\n break\n colnum += 1\n\n for cell in r.getchildren():\n colspan = int(cell.attrib.get('colspan', 1))\n rowspan = int(cell.attrib.get('rowspan', 1))\n cellval = ' '.join(x.strip() for x in cell.itertext()) # text only without markup\n links = [x.get('href') for x in cell.iter('a')]\n maxlinks[colnum] = max(maxlinks.get(colnum, 0), len(links))\n\n if is_header(cell):\n for k in range(rownum, rownum+rowspan):\n while k >= len(headers): # extend headers list with lists for all header rows\n headers.append([])\n\n for j in range(colnum, colnum+colspan):\n while j >= len(headers[k]):\n headers[k].append(None)\n headers[k][j] = cellval\n cellval = '' # use empty non-None value for subsequent rows in the rowspan\n else:\n while colnum >= len(row):\n row.append(None)\n row[colnum] = (cellval, links)\n\n colnum += colspan\n\n if any(row):\n yield row\n\n self.columns = []\n if headers:\n it = itertools.zip_longest(*headers, fillvalue='')\n else:\n it = [list(x) for x in self.rows[0]]\n self.rows = self.rows[1:]\n\n for colnum, names in enumerate(it):\n name = '_'.join(str(x) for x in names if x)\n self.addColumn(Column(name, getter=lambda c,r,i=colnum: r[i][0]))\n for linknum in range(maxlinks.get(colnum, 0)):\n self.addColumn(Column(name+'_link'+str(linknum), width=20, getter=lambda c,r,i=colnum,j=linknum: r[i][1][j]))\n\n\[email protected]\ndef save_html(vd, p, *vsheets):\n 'Save vsheets as HTML tables in a single file'\n\n with open(p, 'w', encoding='ascii', errors='xmlcharrefreplace') as fp:\n for sheet in vsheets:\n\n fp.write('<h2 class=\"sheetname\">%s</h2>\\n'.format(sheetname=html.escape(sheet.name)))\n\n fp.write('<table id=\"{sheetname}\">\\n'.format(sheetname=html.escape(sheet.name)))\n\n # headers\n fp.write('<tr>')\n for col in sheet.visibleCols:\n contents = html.escape(col.name)\n fp.write('<th>{colname}</th>'.format(colname=contents))\n fp.write('</tr>\\n')\n\n # rows\n with Progress(gerund='saving'):\n for typedvals in sheet.iterdispvals(format=False):\n fp.write('<tr>')\n for col, val in typedvals.items():\n fp.write('<td>')\n fp.write(html.escape(val))\n fp.write('</td>')\n fp.write('</tr>\\n')\n\n fp.write('</table>')\n vd.status('%s save finished' % p)\n\n\nVisiData.save_htm = VisiData.save_html\n\n\nvd.filetype('html', HtmlTablesSheet)\nvd.filetype('htm', HtmlTablesSheet)\n", "path": "visidata/loaders/html.py"}], "after_files": [{"content": "import html\nfrom visidata import *\n\n\nclass HtmlTablesSheet(IndexSheet):\n rowtype = 'sheets' # rowdef: HtmlTableSheet (sheet.html = lxml.html.HtmlElement)\n columns = IndexSheet.columns + [\n Column('tag', width=0, getter=lambda col,row: row.html.tag),\n Column('id', getter=lambda col,row: row.html.attrib.get('id')),\n Column('classes', getter=lambda col,row: row.html.attrib.get('class')),\n ]\n def iterload(self):\n import lxml.html\n from lxml import etree\n utf8_parser = etree.HTMLParser(encoding='utf-8')\n with self.source.open_text() as fp:\n html = lxml.html.etree.parse(fp, parser=utf8_parser)\n self.setKeys([self.column('name')])\n self.column('keys').hide()\n self.column('source').hide()\n\n for i, e in enumerate(html.iter('table')):\n if e.tag == 'table':\n vs = HtmlTableSheet(e.attrib.get(\"id\", \"table_\" + str(i)), source=e)\n vs.reload()\n vs.html = e\n yield vs\n\n\ndef is_header(elem):\n scope = elem.attrib.get('scope', '')\n\n if elem.tag == 'th':\n if not scope or scope == 'col':\n return True\n\n return False\n\nclass HtmlTableSheet(Sheet):\n rowtype = 'rows' # list of strings\n columns = []\n\n def iterload(self):\n headers = []\n\n maxlinks = {} # [colnum] -> nlinks:int\n\n for rownum, r in enumerate(self.source.iter('tr')):\n row = []\n\n colnum = 0\n # get starting column, which might be different if there were rowspan>1 already\n if rownum < len(headers):\n while colnum < len(headers[rownum]):\n if headers[rownum][colnum] is None:\n break\n colnum += 1\n\n for cell in r.getchildren():\n colspan = int(cell.attrib.get('colspan', 1))\n rowspan = int(cell.attrib.get('rowspan', 1))\n cellval = ' '.join(x.strip() for x in cell.itertext()) # text only without markup\n links = [x.get('href') for x in cell.iter('a')]\n maxlinks[colnum] = max(maxlinks.get(colnum, 0), len(links))\n\n if is_header(cell):\n for k in range(rownum, rownum+rowspan):\n while k >= len(headers): # extend headers list with lists for all header rows\n headers.append([])\n\n for j in range(colnum, colnum+colspan):\n while j >= len(headers[k]):\n headers[k].append(None)\n headers[k][j] = cellval\n cellval = '' # use empty non-None value for subsequent rows in the rowspan\n else:\n while colnum >= len(row):\n row.append(None)\n row[colnum] = (cellval, links)\n\n colnum += colspan\n\n if any(row):\n yield row\n\n self.columns = []\n if headers:\n it = itertools.zip_longest(*headers, fillvalue='')\n else:\n it = [list(x) for x in self.rows[0]]\n self.rows = self.rows[1:]\n\n for colnum, names in enumerate(it):\n name = '_'.join(str(x) for x in names if x)\n self.addColumn(Column(name, getter=lambda c,r,i=colnum: r[i][0]))\n for linknum in range(maxlinks.get(colnum, 0)):\n self.addColumn(Column(name+'_link'+str(linknum), width=20, getter=lambda c,r,i=colnum,j=linknum: r[i][1][j]))\n\n\[email protected]\ndef save_html(vd, p, *vsheets):\n 'Save vsheets as HTML tables in a single file'\n\n with open(p, 'w', encoding='ascii', errors='xmlcharrefreplace') as fp:\n for sheet in vsheets:\n\n fp.write('<h2 class=\"sheetname\">%s</h2>\\n'.format(sheetname=html.escape(sheet.name)))\n\n fp.write('<table id=\"{sheetname}\">\\n'.format(sheetname=html.escape(sheet.name)))\n\n # headers\n fp.write('<tr>')\n for col in sheet.visibleCols:\n contents = html.escape(col.name)\n fp.write('<th>{colname}</th>'.format(colname=contents))\n fp.write('</tr>\\n')\n\n # rows\n with Progress(gerund='saving'):\n for typedvals in sheet.iterdispvals(format=False):\n fp.write('<tr>')\n for col, val in typedvals.items():\n fp.write('<td>')\n fp.write(html.escape(str(val)))\n fp.write('</td>')\n fp.write('</tr>\\n')\n\n fp.write('</table>')\n vd.status('%s save finished' % p)\n\n\nVisiData.save_htm = VisiData.save_html\n\n\nvd.filetype('html', HtmlTablesSheet)\nvd.filetype('htm', HtmlTablesSheet)\n", "path": "visidata/loaders/html.py"}]} | 2,112 | 107 |
gh_patches_debug_26723 | rasdani/github-patches | git_diff | OpenCTI-Platform__connectors-51 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
[CVE] Download link to variable
## Description
Set the download CVE link to variable, because otherwise the tool can hardly be used offline. Offline we can host the CVEs on a link that is not : "https://nvd.nist.gov/feeds/json/cve/1.1/nvdcve-1.1-"
## Reproducible Steps
https://github.com/OpenCTI-Platform/connectors/blame/9d47ffdad1c2a7fbdd709565d5c3f670693b148f/cve/src/cve.py#L103
## Expected Output
Url as a variable in the .yml
## Actual Output
Permanent link : "https://nvd.nist.gov/feeds/json/cve/1.1/nvdcve-1.1-"
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `cve/src/cve.py`
Content:
```
1 # coding: utf-8
2
3 import os
4 import yaml
5 import time
6 import urllib.request
7 import gzip
8 import shutil
9
10 from datetime import datetime
11 from pycti import OpenCTIConnectorHelper, get_config_variable
12 from cvetostix2 import convert
13
14
15 class Cve:
16 def __init__(self):
17 # Instantiate the connector helper from config
18 config_file_path = os.path.dirname(os.path.abspath(__file__)) + "/config.yml"
19 config = (
20 yaml.load(open(config_file_path), Loader=yaml.FullLoader)
21 if os.path.isfile(config_file_path)
22 else {}
23 )
24 self.helper = OpenCTIConnectorHelper(config)
25 # Extra config
26 self.cve_import_history = get_config_variable(
27 "CVE_IMPORT_HISTORY", ["cve", "import_history"], config, False
28 )
29 self.cve_nvd_data_feed = get_config_variable(
30 "CVE_NVD_DATA_FEED", ["cve", "nvd_data_feed"], config
31 )
32 self.cve_interval = get_config_variable(
33 "CVE_INTERVAL", ["cve", "interval"], config, True
34 )
35 self.update_existing_data = get_config_variable(
36 "CONNECTOR_UPDATE_EXISTING_DATA",
37 ["connector", "update_existing_data"],
38 config,
39 )
40
41 def get_interval(self):
42 return int(self.cve_interval) * 60 * 60 * 24
43
44 def convert_and_send(self, url):
45 try:
46 # Downloading json.gz file
47 self.helper.log_info("Requesting the file " + url)
48 urllib.request.urlretrieve(
49 self.cve_nvd_data_feed,
50 os.path.dirname(os.path.abspath(__file__)) + "/data.json.gz",
51 )
52 # Unzipping the file
53 self.helper.log_info("Unzipping the file")
54 with gzip.open("data.json.gz", "rb") as f_in:
55 with open("data.json", "wb") as f_out:
56 shutil.copyfileobj(f_in, f_out)
57 # Converting the file to stix2
58 self.helper.log_info("Converting the file")
59 convert("data.json", "data-stix2.json")
60 with open("data-stix2.json") as stix_json:
61 contents = stix_json.read()
62 self.helper.send_stix2_bundle(
63 contents, self.helper.connect_scope, self.update_existing_data
64 )
65 # Remove files
66 os.remove("data.json")
67 os.remove("data.json.gz")
68 os.remove("data-stix2.json")
69 except Exception as e:
70 self.helper.log_error(str(e))
71 time.sleep(60)
72
73 def run(self):
74 self.helper.log_info("Fetching CVE knowledge...")
75 while True:
76 try:
77 # Get the current timestamp and check
78 timestamp = int(time.time())
79 current_state = self.helper.get_state()
80 if current_state is not None and "last_run" in current_state:
81 last_run = current_state["last_run"]
82 self.helper.log_info(
83 "Connector last run: "
84 + datetime.utcfromtimestamp(last_run).strftime(
85 "%Y-%m-%d %H:%M:%S"
86 )
87 )
88 else:
89 last_run = None
90 self.helper.log_info("Connector has never run")
91 # If the last_run is more than interval-1 day
92 if last_run is None or (
93 (timestamp - last_run)
94 > ((int(self.cve_interval) - 1) * 60 * 60 * 24)
95 ):
96 self.convert_and_send(self.cve_nvd_data_feed)
97 # If import history and never run
98 if last_run is None and self.cve_import_history:
99 now = datetime.now()
100 years = list(range(2002, now.year))
101 for year in years:
102 self.convert_and_send(
103 "https://nvd.nist.gov/feeds/json/cve/1.1/nvdcve-1.1-"
104 + str(year)
105 + ".json.gz"
106 )
107
108 # Store the current timestamp as a last run
109 self.helper.log_info(
110 "Connector successfully run, storing last_run as "
111 + str(timestamp)
112 )
113 self.helper.set_state({"last_run": timestamp})
114 self.helper.log_info(
115 "Last_run stored, next run in: "
116 + str(round(self.get_interval() / 60 / 60 / 24, 2))
117 + " days"
118 )
119 time.sleep(60)
120 else:
121 new_interval = self.get_interval() - (timestamp - last_run)
122 self.helper.log_info(
123 "Connector will not run, next run in: "
124 + str(round(new_interval / 60 / 60 / 24, 2))
125 + " days"
126 )
127 time.sleep(60)
128 except (KeyboardInterrupt, SystemExit):
129 self.helper.log_info("Connector stop")
130 exit(0)
131 except Exception as e:
132 self.helper.log_error(str(e))
133 time.sleep(60)
134
135
136 if __name__ == "__main__":
137 try:
138 cveConnector = Cve()
139 cveConnector.run()
140 except Exception as e:
141 print(e)
142 time.sleep(10)
143 exit(0)
144
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/cve/src/cve.py b/cve/src/cve.py
--- a/cve/src/cve.py
+++ b/cve/src/cve.py
@@ -29,6 +29,9 @@
self.cve_nvd_data_feed = get_config_variable(
"CVE_NVD_DATA_FEED", ["cve", "nvd_data_feed"], config
)
+ self.cve_history_data_feed = get_config_variable(
+ "CVE_HISTORY_DATA_FEED", ["cve", "history_data_feed"], config
+ )
self.cve_interval = get_config_variable(
"CVE_INTERVAL", ["cve", "interval"], config, True
)
@@ -97,12 +100,10 @@
# If import history and never run
if last_run is None and self.cve_import_history:
now = datetime.now()
- years = list(range(2002, now.year))
+ years = list(range(2002, now.year+1))
for year in years:
self.convert_and_send(
- "https://nvd.nist.gov/feeds/json/cve/1.1/nvdcve-1.1-"
- + str(year)
- + ".json.gz"
+ f"{self.cve_history_data_feed}nvdcve-1.1-{year}.json.gz"
)
# Store the current timestamp as a last run
| {"golden_diff": "diff --git a/cve/src/cve.py b/cve/src/cve.py\n--- a/cve/src/cve.py\n+++ b/cve/src/cve.py\n@@ -29,6 +29,9 @@\n self.cve_nvd_data_feed = get_config_variable(\n \"CVE_NVD_DATA_FEED\", [\"cve\", \"nvd_data_feed\"], config\n )\n+ self.cve_history_data_feed = get_config_variable(\n+ \"CVE_HISTORY_DATA_FEED\", [\"cve\", \"history_data_feed\"], config\n+ )\n self.cve_interval = get_config_variable(\n \"CVE_INTERVAL\", [\"cve\", \"interval\"], config, True\n )\n@@ -97,12 +100,10 @@\n # If import history and never run\n if last_run is None and self.cve_import_history:\n now = datetime.now()\n- years = list(range(2002, now.year))\n+ years = list(range(2002, now.year+1))\n for year in years:\n self.convert_and_send(\n- \"https://nvd.nist.gov/feeds/json/cve/1.1/nvdcve-1.1-\"\n- + str(year)\n- + \".json.gz\"\n+ f\"{self.cve_history_data_feed}nvdcve-1.1-{year}.json.gz\"\n )\n \n # Store the current timestamp as a last run\n", "issue": "[CVE] Download link to variable\n## Description\r\n\r\nSet the download CVE link to variable, because otherwise the tool can hardly be used offline. Offline we can host the CVEs on a link that is not : \"https://nvd.nist.gov/feeds/json/cve/1.1/nvdcve-1.1-\"\r\n\r\n## Reproducible Steps\r\n\r\nhttps://github.com/OpenCTI-Platform/connectors/blame/9d47ffdad1c2a7fbdd709565d5c3f670693b148f/cve/src/cve.py#L103\r\n\r\n## Expected Output\r\n\r\nUrl as a variable in the .yml\r\n\r\n## Actual Output\r\n\r\nPermanent link : \"https://nvd.nist.gov/feeds/json/cve/1.1/nvdcve-1.1-\"\r\n\n", "before_files": [{"content": "# coding: utf-8\n\nimport os\nimport yaml\nimport time\nimport urllib.request\nimport gzip\nimport shutil\n\nfrom datetime import datetime\nfrom pycti import OpenCTIConnectorHelper, get_config_variable\nfrom cvetostix2 import convert\n\n\nclass Cve:\n def __init__(self):\n # Instantiate the connector helper from config\n config_file_path = os.path.dirname(os.path.abspath(__file__)) + \"/config.yml\"\n config = (\n yaml.load(open(config_file_path), Loader=yaml.FullLoader)\n if os.path.isfile(config_file_path)\n else {}\n )\n self.helper = OpenCTIConnectorHelper(config)\n # Extra config\n self.cve_import_history = get_config_variable(\n \"CVE_IMPORT_HISTORY\", [\"cve\", \"import_history\"], config, False\n )\n self.cve_nvd_data_feed = get_config_variable(\n \"CVE_NVD_DATA_FEED\", [\"cve\", \"nvd_data_feed\"], config\n )\n self.cve_interval = get_config_variable(\n \"CVE_INTERVAL\", [\"cve\", \"interval\"], config, True\n )\n self.update_existing_data = get_config_variable(\n \"CONNECTOR_UPDATE_EXISTING_DATA\",\n [\"connector\", \"update_existing_data\"],\n config,\n )\n\n def get_interval(self):\n return int(self.cve_interval) * 60 * 60 * 24\n\n def convert_and_send(self, url):\n try:\n # Downloading json.gz file\n self.helper.log_info(\"Requesting the file \" + url)\n urllib.request.urlretrieve(\n self.cve_nvd_data_feed,\n os.path.dirname(os.path.abspath(__file__)) + \"/data.json.gz\",\n )\n # Unzipping the file\n self.helper.log_info(\"Unzipping the file\")\n with gzip.open(\"data.json.gz\", \"rb\") as f_in:\n with open(\"data.json\", \"wb\") as f_out:\n shutil.copyfileobj(f_in, f_out)\n # Converting the file to stix2\n self.helper.log_info(\"Converting the file\")\n convert(\"data.json\", \"data-stix2.json\")\n with open(\"data-stix2.json\") as stix_json:\n contents = stix_json.read()\n self.helper.send_stix2_bundle(\n contents, self.helper.connect_scope, self.update_existing_data\n )\n # Remove files\n os.remove(\"data.json\")\n os.remove(\"data.json.gz\")\n os.remove(\"data-stix2.json\")\n except Exception as e:\n self.helper.log_error(str(e))\n time.sleep(60)\n\n def run(self):\n self.helper.log_info(\"Fetching CVE knowledge...\")\n while True:\n try:\n # Get the current timestamp and check\n timestamp = int(time.time())\n current_state = self.helper.get_state()\n if current_state is not None and \"last_run\" in current_state:\n last_run = current_state[\"last_run\"]\n self.helper.log_info(\n \"Connector last run: \"\n + datetime.utcfromtimestamp(last_run).strftime(\n \"%Y-%m-%d %H:%M:%S\"\n )\n )\n else:\n last_run = None\n self.helper.log_info(\"Connector has never run\")\n # If the last_run is more than interval-1 day\n if last_run is None or (\n (timestamp - last_run)\n > ((int(self.cve_interval) - 1) * 60 * 60 * 24)\n ):\n self.convert_and_send(self.cve_nvd_data_feed)\n # If import history and never run\n if last_run is None and self.cve_import_history:\n now = datetime.now()\n years = list(range(2002, now.year))\n for year in years:\n self.convert_and_send(\n \"https://nvd.nist.gov/feeds/json/cve/1.1/nvdcve-1.1-\"\n + str(year)\n + \".json.gz\"\n )\n\n # Store the current timestamp as a last run\n self.helper.log_info(\n \"Connector successfully run, storing last_run as \"\n + str(timestamp)\n )\n self.helper.set_state({\"last_run\": timestamp})\n self.helper.log_info(\n \"Last_run stored, next run in: \"\n + str(round(self.get_interval() / 60 / 60 / 24, 2))\n + \" days\"\n )\n time.sleep(60)\n else:\n new_interval = self.get_interval() - (timestamp - last_run)\n self.helper.log_info(\n \"Connector will not run, next run in: \"\n + str(round(new_interval / 60 / 60 / 24, 2))\n + \" days\"\n )\n time.sleep(60)\n except (KeyboardInterrupt, SystemExit):\n self.helper.log_info(\"Connector stop\")\n exit(0)\n except Exception as e:\n self.helper.log_error(str(e))\n time.sleep(60)\n\n\nif __name__ == \"__main__\":\n try:\n cveConnector = Cve()\n cveConnector.run()\n except Exception as e:\n print(e)\n time.sleep(10)\n exit(0)\n", "path": "cve/src/cve.py"}], "after_files": [{"content": "# coding: utf-8\n\nimport os\nimport yaml\nimport time\nimport urllib.request\nimport gzip\nimport shutil\n\nfrom datetime import datetime\nfrom pycti import OpenCTIConnectorHelper, get_config_variable\nfrom cvetostix2 import convert\n\n\nclass Cve:\n def __init__(self):\n # Instantiate the connector helper from config\n config_file_path = os.path.dirname(os.path.abspath(__file__)) + \"/config.yml\"\n config = (\n yaml.load(open(config_file_path), Loader=yaml.FullLoader)\n if os.path.isfile(config_file_path)\n else {}\n )\n self.helper = OpenCTIConnectorHelper(config)\n # Extra config\n self.cve_import_history = get_config_variable(\n \"CVE_IMPORT_HISTORY\", [\"cve\", \"import_history\"], config, False\n )\n self.cve_nvd_data_feed = get_config_variable(\n \"CVE_NVD_DATA_FEED\", [\"cve\", \"nvd_data_feed\"], config\n )\n self.cve_history_data_feed = get_config_variable(\n \"CVE_HISTORY_DATA_FEED\", [\"cve\", \"history_data_feed\"], config\n )\n self.cve_interval = get_config_variable(\n \"CVE_INTERVAL\", [\"cve\", \"interval\"], config, True\n )\n self.update_existing_data = get_config_variable(\n \"CONNECTOR_UPDATE_EXISTING_DATA\",\n [\"connector\", \"update_existing_data\"],\n config,\n )\n\n def get_interval(self):\n return int(self.cve_interval) * 60 * 60 * 24\n\n def convert_and_send(self, url):\n try:\n # Downloading json.gz file\n self.helper.log_info(\"Requesting the file \" + url)\n urllib.request.urlretrieve(\n self.cve_nvd_data_feed,\n os.path.dirname(os.path.abspath(__file__)) + \"/data.json.gz\",\n )\n # Unzipping the file\n self.helper.log_info(\"Unzipping the file\")\n with gzip.open(\"data.json.gz\", \"rb\") as f_in:\n with open(\"data.json\", \"wb\") as f_out:\n shutil.copyfileobj(f_in, f_out)\n # Converting the file to stix2\n self.helper.log_info(\"Converting the file\")\n convert(\"data.json\", \"data-stix2.json\")\n with open(\"data-stix2.json\") as stix_json:\n contents = stix_json.read()\n self.helper.send_stix2_bundle(\n contents, self.helper.connect_scope, self.update_existing_data\n )\n # Remove files\n os.remove(\"data.json\")\n os.remove(\"data.json.gz\")\n os.remove(\"data-stix2.json\")\n except Exception as e:\n self.helper.log_error(str(e))\n time.sleep(60)\n\n def run(self):\n self.helper.log_info(\"Fetching CVE knowledge...\")\n while True:\n try:\n # Get the current timestamp and check\n timestamp = int(time.time())\n current_state = self.helper.get_state()\n if current_state is not None and \"last_run\" in current_state:\n last_run = current_state[\"last_run\"]\n self.helper.log_info(\n \"Connector last run: \"\n + datetime.utcfromtimestamp(last_run).strftime(\n \"%Y-%m-%d %H:%M:%S\"\n )\n )\n else:\n last_run = None\n self.helper.log_info(\"Connector has never run\")\n # If the last_run is more than interval-1 day\n if last_run is None or (\n (timestamp - last_run)\n > ((int(self.cve_interval) - 1) * 60 * 60 * 24)\n ):\n self.convert_and_send(self.cve_nvd_data_feed)\n # If import history and never run\n if last_run is None and self.cve_import_history:\n now = datetime.now()\n years = list(range(2002, now.year+1))\n for year in years:\n self.convert_and_send(\n f\"{self.cve_history_data_feed}nvdcve-1.1-{year}.json.gz\"\n )\n\n # Store the current timestamp as a last run\n self.helper.log_info(\n \"Connector successfully run, storing last_run as \"\n + str(timestamp)\n )\n self.helper.set_state({\"last_run\": timestamp})\n self.helper.log_info(\n \"Last_run stored, next run in: \"\n + str(round(self.get_interval() / 60 / 60 / 24, 2))\n + \" days\"\n )\n time.sleep(60)\n else:\n new_interval = self.get_interval() - (timestamp - last_run)\n self.helper.log_info(\n \"Connector will not run, next run in: \"\n + str(round(new_interval / 60 / 60 / 24, 2))\n + \" days\"\n )\n time.sleep(60)\n except (KeyboardInterrupt, SystemExit):\n self.helper.log_info(\"Connector stop\")\n exit(0)\n except Exception as e:\n self.helper.log_error(str(e))\n time.sleep(60)\n\n\nif __name__ == \"__main__\":\n try:\n cveConnector = Cve()\n cveConnector.run()\n except Exception as e:\n print(e)\n time.sleep(10)\n exit(0)\n", "path": "cve/src/cve.py"}]} | 1,902 | 313 |
gh_patches_debug_2261 | rasdani/github-patches | git_diff | mitmproxy__mitmproxy-4179 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
TypeError: Subscripted generics cannot be used with class and instance checks under python 3.9.0b1
#### Problem Description
Running mitmproxy 5.1.1 under python 3.9.0b1 fails with `TypeError: Subscripted generics cannot be used with class and instance checks`. The test suite fails as well with hundreds of ERROR and FAILED tests.
#### Steps to reproduce the behavior:
1. install mitmproxy 5.1.1 on Fedora rawhide
2. mitmproxy
3. pytest -v
There are:
```
=================== 303 failed, 994 passed, 2 xfailed, 115 warnings, 182 errors in 72.86s (0:01:12) ====================
```
Most of them throw a `TypeError: Subscripted generics cannot be used with class and instance checks` and have a stack trace similar to:
```
___________________________________ ERROR at setup of TestHTTPS.test_clientcert_dir ____________________________________
cls = <class 'test.mitmproxy.proxy.test_server.TestHTTPS'>
@classmethod
def setup_class(cls):
cls.server = pathod.test.Daemon(
ssl=cls.ssl,
ssloptions=cls.ssloptions)
cls.server2 = pathod.test.Daemon(
ssl=cls.ssl,
ssloptions=cls.ssloptions)
> cls.options = cls.get_options()
test/mitmproxy/tservers.py:146:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
test/mitmproxy/tservers.py:179: in get_options
return options.Options(
mitmproxy/options.py:50: in __init__
self.add_option(
mitmproxy/optmanager.py:109: in add_option
self._options[name] = _Option(name, typespec, default, help, choices)
mitmproxy/optmanager.py:34: in __init__
typecheck.check_option_type(name, default, typespec)
mitmproxy/utils/typecheck.py:73: in check_option_type
elif not isinstance(value, typeinfo):
/usr/lib64/python3.9/typing.py:649: in __instancecheck__
return self.__subclasscheck__(type(obj))
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
self = typing.Optional[str], cls = <class 'NoneType'>
def __subclasscheck__(self, cls):
> raise TypeError("Subscripted generics cannot be used with"
" class and instance checks")
E TypeError: Subscripted generics cannot be used with class and instance checks
/usr/lib64/python3.9/typing.py:652: TypeError
```
#### System Information
```
Traceback (most recent call last):
File "/builddir/build/BUILDROOT/mitmproxy-5.1.1-1.fc33.x86_64/usr/bin/./mitmproxy", line 11, in <module>
load_entry_point('mitmproxy==5.1.1', 'console_scripts', 'mitmproxy')()
File "/builddir/build/BUILDROOT/mitmproxy-5.1.1-1.fc33.x86_64/usr/lib/python3.9/site-packages/mitmproxy/tools/_main.py", line 147, in mitmproxy
run(console.master.ConsoleMaster, cmdline.mitmproxy, args)
File "/builddir/build/BUILDROOT/mitmproxy-5.1.1-1.fc33.x86_64/usr/lib/python3.9/site-packages/mitmproxy/tools/_main.py", line 71, in run
opts = options.Options()
File "/builddir/build/BUILDROOT/mitmproxy-5.1.1-1.fc33.x86_64/usr/lib/python3.9/site-packages/mitmproxy/options.py", line 50, in __init__
self.add_option(
File "/builddir/build/BUILDROOT/mitmproxy-5.1.1-1.fc33.x86_64/usr/lib/python3.9/site-packages/mitmproxy/optmanager.py", line 109, in add_option
self._options[name] = _Option(name, typespec, default, help, choices)
File "/builddir/build/BUILDROOT/mitmproxy-5.1.1-1.fc33.x86_64/usr/lib/python3.9/site-packages/mitmproxy/optmanager.py", line 34, in __init__
typecheck.check_option_type(name, default, typespec)
File "/builddir/build/BUILDROOT/mitmproxy-5.1.1-1.fc33.x86_64/usr/lib/python3.9/site-packages/mitmproxy/utils/typecheck.py", line 73, in check_option_type
elif not isinstance(value, typeinfo):
File "/usr/lib64/python3.9/typing.py", line 649, in __instancecheck__
return self.__subclasscheck__(type(obj))
File "/usr/lib64/python3.9/typing.py", line 652, in __subclasscheck__
raise TypeError("Subscripted generics cannot be used with"
TypeError: Subscripted generics cannot be used with class and instance checks
```
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `mitmproxy/utils/typecheck.py`
Content:
```
1 import typing
2
3 Type = typing.Union[
4 typing.Any # anything more elaborate really fails with mypy at the moment.
5 ]
6
7
8 def sequence_type(typeinfo: typing.Type[typing.List]) -> Type:
9 """Return the type of a sequence, e.g. typing.List"""
10 return typeinfo.__args__[0] # type: ignore
11
12
13 def tuple_types(typeinfo: typing.Type[typing.Tuple]) -> typing.Sequence[Type]:
14 """Return the types of a typing.Tuple"""
15 return typeinfo.__args__ # type: ignore
16
17
18 def union_types(typeinfo: typing.Type[typing.Tuple]) -> typing.Sequence[Type]:
19 """return the types of a typing.Union"""
20 return typeinfo.__args__ # type: ignore
21
22
23 def mapping_types(typeinfo: typing.Type[typing.Mapping]) -> typing.Tuple[Type, Type]:
24 """return the types of a mapping, e.g. typing.Dict"""
25 return typeinfo.__args__ # type: ignore
26
27
28 def check_option_type(name: str, value: typing.Any, typeinfo: Type) -> None:
29 """
30 Check if the provided value is an instance of typeinfo and raises a
31 TypeError otherwise. This function supports only those types required for
32 options.
33 """
34 e = TypeError("Expected {} for {}, but got {}.".format(
35 typeinfo,
36 name,
37 type(value)
38 ))
39
40 typename = str(typeinfo)
41
42 if typename.startswith("typing.Union"):
43 for T in union_types(typeinfo):
44 try:
45 check_option_type(name, value, T)
46 except TypeError:
47 pass
48 else:
49 return
50 raise e
51 elif typename.startswith("typing.Tuple"):
52 types = tuple_types(typeinfo)
53 if not isinstance(value, (tuple, list)):
54 raise e
55 if len(types) != len(value):
56 raise e
57 for i, (x, T) in enumerate(zip(value, types)):
58 check_option_type("{}[{}]".format(name, i), x, T)
59 return
60 elif typename.startswith("typing.Sequence"):
61 T = sequence_type(typeinfo)
62 if not isinstance(value, (tuple, list)):
63 raise e
64 for v in value:
65 check_option_type(name, v, T)
66 elif typename.startswith("typing.IO"):
67 if hasattr(value, "read"):
68 return
69 else:
70 raise e
71 elif typename.startswith("typing.Any"):
72 return
73 elif not isinstance(value, typeinfo):
74 if typeinfo is float and isinstance(value, int):
75 return
76 raise e
77
78
79 def typespec_to_str(typespec: typing.Any) -> str:
80 if typespec in (str, int, bool):
81 t = typespec.__name__
82 elif typespec == typing.Optional[str]:
83 t = 'optional str'
84 elif typespec == typing.Sequence[str]:
85 t = 'sequence of str'
86 elif typespec == typing.Optional[int]:
87 t = 'optional int'
88 else:
89 raise NotImplementedError
90 return t
91
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/mitmproxy/utils/typecheck.py b/mitmproxy/utils/typecheck.py
--- a/mitmproxy/utils/typecheck.py
+++ b/mitmproxy/utils/typecheck.py
@@ -39,7 +39,7 @@
typename = str(typeinfo)
- if typename.startswith("typing.Union"):
+ if typename.startswith("typing.Union") or typename.startswith("typing.Optional"):
for T in union_types(typeinfo):
try:
check_option_type(name, value, T)
| {"golden_diff": "diff --git a/mitmproxy/utils/typecheck.py b/mitmproxy/utils/typecheck.py\n--- a/mitmproxy/utils/typecheck.py\n+++ b/mitmproxy/utils/typecheck.py\n@@ -39,7 +39,7 @@\n \n typename = str(typeinfo)\n \n- if typename.startswith(\"typing.Union\"):\n+ if typename.startswith(\"typing.Union\") or typename.startswith(\"typing.Optional\"):\n for T in union_types(typeinfo):\n try:\n check_option_type(name, value, T)\n", "issue": "TypeError: Subscripted generics cannot be used with class and instance checks under python 3.9.0b1\n#### Problem Description\r\nRunning mitmproxy 5.1.1 under python 3.9.0b1 fails with `TypeError: Subscripted generics cannot be used with class and instance checks`. The test suite fails as well with hundreds of ERROR and FAILED tests.\r\n\r\n#### Steps to reproduce the behavior:\r\n1. install mitmproxy 5.1.1 on Fedora rawhide\r\n2. mitmproxy\r\n3. pytest -v\r\n\r\nThere are:\r\n```\r\n=================== 303 failed, 994 passed, 2 xfailed, 115 warnings, 182 errors in 72.86s (0:01:12) ====================\r\n```\r\nMost of them throw a `TypeError: Subscripted generics cannot be used with class and instance checks` and have a stack trace similar to:\r\n```\r\n___________________________________ ERROR at setup of TestHTTPS.test_clientcert_dir ____________________________________\r\n\r\ncls = <class 'test.mitmproxy.proxy.test_server.TestHTTPS'>\r\n\r\n @classmethod\r\n def setup_class(cls):\r\n cls.server = pathod.test.Daemon(\r\n ssl=cls.ssl,\r\n ssloptions=cls.ssloptions)\r\n cls.server2 = pathod.test.Daemon(\r\n ssl=cls.ssl,\r\n ssloptions=cls.ssloptions)\r\n \r\n> cls.options = cls.get_options()\r\n\r\ntest/mitmproxy/tservers.py:146: \r\n_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ \r\ntest/mitmproxy/tservers.py:179: in get_options\r\n return options.Options(\r\nmitmproxy/options.py:50: in __init__\r\n self.add_option(\r\nmitmproxy/optmanager.py:109: in add_option\r\n self._options[name] = _Option(name, typespec, default, help, choices)\r\nmitmproxy/optmanager.py:34: in __init__\r\n typecheck.check_option_type(name, default, typespec)\r\nmitmproxy/utils/typecheck.py:73: in check_option_type\r\n elif not isinstance(value, typeinfo):\r\n/usr/lib64/python3.9/typing.py:649: in __instancecheck__\r\n return self.__subclasscheck__(type(obj))\r\n_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ \r\n\r\nself = typing.Optional[str], cls = <class 'NoneType'>\r\n\r\n def __subclasscheck__(self, cls):\r\n> raise TypeError(\"Subscripted generics cannot be used with\"\r\n \" class and instance checks\")\r\nE TypeError: Subscripted generics cannot be used with class and instance checks\r\n\r\n/usr/lib64/python3.9/typing.py:652: TypeError\r\n```\r\n\r\n#### System Information\r\n```\r\nTraceback (most recent call last):\r\n File \"/builddir/build/BUILDROOT/mitmproxy-5.1.1-1.fc33.x86_64/usr/bin/./mitmproxy\", line 11, in <module>\r\n load_entry_point('mitmproxy==5.1.1', 'console_scripts', 'mitmproxy')()\r\n File \"/builddir/build/BUILDROOT/mitmproxy-5.1.1-1.fc33.x86_64/usr/lib/python3.9/site-packages/mitmproxy/tools/_main.py\", line 147, in mitmproxy\r\n run(console.master.ConsoleMaster, cmdline.mitmproxy, args)\r\n File \"/builddir/build/BUILDROOT/mitmproxy-5.1.1-1.fc33.x86_64/usr/lib/python3.9/site-packages/mitmproxy/tools/_main.py\", line 71, in run\r\n opts = options.Options()\r\n File \"/builddir/build/BUILDROOT/mitmproxy-5.1.1-1.fc33.x86_64/usr/lib/python3.9/site-packages/mitmproxy/options.py\", line 50, in __init__\r\n self.add_option(\r\n File \"/builddir/build/BUILDROOT/mitmproxy-5.1.1-1.fc33.x86_64/usr/lib/python3.9/site-packages/mitmproxy/optmanager.py\", line 109, in add_option\r\n self._options[name] = _Option(name, typespec, default, help, choices)\r\n File \"/builddir/build/BUILDROOT/mitmproxy-5.1.1-1.fc33.x86_64/usr/lib/python3.9/site-packages/mitmproxy/optmanager.py\", line 34, in __init__\r\n typecheck.check_option_type(name, default, typespec)\r\n File \"/builddir/build/BUILDROOT/mitmproxy-5.1.1-1.fc33.x86_64/usr/lib/python3.9/site-packages/mitmproxy/utils/typecheck.py\", line 73, in check_option_type\r\n elif not isinstance(value, typeinfo):\r\n File \"/usr/lib64/python3.9/typing.py\", line 649, in __instancecheck__\r\n return self.__subclasscheck__(type(obj))\r\n File \"/usr/lib64/python3.9/typing.py\", line 652, in __subclasscheck__\r\n raise TypeError(\"Subscripted generics cannot be used with\"\r\nTypeError: Subscripted generics cannot be used with class and instance checks\r\n```\n", "before_files": [{"content": "import typing\n\nType = typing.Union[\n typing.Any # anything more elaborate really fails with mypy at the moment.\n]\n\n\ndef sequence_type(typeinfo: typing.Type[typing.List]) -> Type:\n \"\"\"Return the type of a sequence, e.g. typing.List\"\"\"\n return typeinfo.__args__[0] # type: ignore\n\n\ndef tuple_types(typeinfo: typing.Type[typing.Tuple]) -> typing.Sequence[Type]:\n \"\"\"Return the types of a typing.Tuple\"\"\"\n return typeinfo.__args__ # type: ignore\n\n\ndef union_types(typeinfo: typing.Type[typing.Tuple]) -> typing.Sequence[Type]:\n \"\"\"return the types of a typing.Union\"\"\"\n return typeinfo.__args__ # type: ignore\n\n\ndef mapping_types(typeinfo: typing.Type[typing.Mapping]) -> typing.Tuple[Type, Type]:\n \"\"\"return the types of a mapping, e.g. typing.Dict\"\"\"\n return typeinfo.__args__ # type: ignore\n\n\ndef check_option_type(name: str, value: typing.Any, typeinfo: Type) -> None:\n \"\"\"\n Check if the provided value is an instance of typeinfo and raises a\n TypeError otherwise. This function supports only those types required for\n options.\n \"\"\"\n e = TypeError(\"Expected {} for {}, but got {}.\".format(\n typeinfo,\n name,\n type(value)\n ))\n\n typename = str(typeinfo)\n\n if typename.startswith(\"typing.Union\"):\n for T in union_types(typeinfo):\n try:\n check_option_type(name, value, T)\n except TypeError:\n pass\n else:\n return\n raise e\n elif typename.startswith(\"typing.Tuple\"):\n types = tuple_types(typeinfo)\n if not isinstance(value, (tuple, list)):\n raise e\n if len(types) != len(value):\n raise e\n for i, (x, T) in enumerate(zip(value, types)):\n check_option_type(\"{}[{}]\".format(name, i), x, T)\n return\n elif typename.startswith(\"typing.Sequence\"):\n T = sequence_type(typeinfo)\n if not isinstance(value, (tuple, list)):\n raise e\n for v in value:\n check_option_type(name, v, T)\n elif typename.startswith(\"typing.IO\"):\n if hasattr(value, \"read\"):\n return\n else:\n raise e\n elif typename.startswith(\"typing.Any\"):\n return\n elif not isinstance(value, typeinfo):\n if typeinfo is float and isinstance(value, int):\n return\n raise e\n\n\ndef typespec_to_str(typespec: typing.Any) -> str:\n if typespec in (str, int, bool):\n t = typespec.__name__\n elif typespec == typing.Optional[str]:\n t = 'optional str'\n elif typespec == typing.Sequence[str]:\n t = 'sequence of str'\n elif typespec == typing.Optional[int]:\n t = 'optional int'\n else:\n raise NotImplementedError\n return t\n", "path": "mitmproxy/utils/typecheck.py"}], "after_files": [{"content": "import typing\n\nType = typing.Union[\n typing.Any # anything more elaborate really fails with mypy at the moment.\n]\n\n\ndef sequence_type(typeinfo: typing.Type[typing.List]) -> Type:\n \"\"\"Return the type of a sequence, e.g. typing.List\"\"\"\n return typeinfo.__args__[0] # type: ignore\n\n\ndef tuple_types(typeinfo: typing.Type[typing.Tuple]) -> typing.Sequence[Type]:\n \"\"\"Return the types of a typing.Tuple\"\"\"\n return typeinfo.__args__ # type: ignore\n\n\ndef union_types(typeinfo: typing.Type[typing.Tuple]) -> typing.Sequence[Type]:\n \"\"\"return the types of a typing.Union\"\"\"\n return typeinfo.__args__ # type: ignore\n\n\ndef mapping_types(typeinfo: typing.Type[typing.Mapping]) -> typing.Tuple[Type, Type]:\n \"\"\"return the types of a mapping, e.g. typing.Dict\"\"\"\n return typeinfo.__args__ # type: ignore\n\n\ndef check_option_type(name: str, value: typing.Any, typeinfo: Type) -> None:\n \"\"\"\n Check if the provided value is an instance of typeinfo and raises a\n TypeError otherwise. This function supports only those types required for\n options.\n \"\"\"\n e = TypeError(\"Expected {} for {}, but got {}.\".format(\n typeinfo,\n name,\n type(value)\n ))\n\n typename = str(typeinfo)\n\n if typename.startswith(\"typing.Union\") or typename.startswith(\"typing.Optional\"):\n for T in union_types(typeinfo):\n try:\n check_option_type(name, value, T)\n except TypeError:\n pass\n else:\n return\n raise e\n elif typename.startswith(\"typing.Tuple\"):\n types = tuple_types(typeinfo)\n if not isinstance(value, (tuple, list)):\n raise e\n if len(types) != len(value):\n raise e\n for i, (x, T) in enumerate(zip(value, types)):\n check_option_type(\"{}[{}]\".format(name, i), x, T)\n return\n elif typename.startswith(\"typing.Sequence\"):\n T = sequence_type(typeinfo)\n if not isinstance(value, (tuple, list)):\n raise e\n for v in value:\n check_option_type(name, v, T)\n elif typename.startswith(\"typing.IO\"):\n if hasattr(value, \"read\"):\n return\n else:\n raise e\n elif typename.startswith(\"typing.Any\"):\n return\n elif not isinstance(value, typeinfo):\n if typeinfo is float and isinstance(value, int):\n return\n raise e\n\n\ndef typespec_to_str(typespec: typing.Any) -> str:\n if typespec in (str, int, bool):\n t = typespec.__name__\n elif typespec == typing.Optional[str]:\n t = 'optional str'\n elif typespec == typing.Sequence[str]:\n t = 'sequence of str'\n elif typespec == typing.Optional[int]:\n t = 'optional int'\n else:\n raise NotImplementedError\n return t\n", "path": "mitmproxy/utils/typecheck.py"}]} | 2,350 | 110 |
gh_patches_debug_16552 | rasdani/github-patches | git_diff | Kinto__kinto-1814 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Invalid account_create_principals key
The `account:create` check_permission code is looking at `account_account:create_principals` settings key rather than `account_create_principals`
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `kinto/core/authorization.py`
Content:
```
1 import functools
2 import logging
3
4 from pyramid.settings import aslist
5 from pyramid.security import IAuthorizationPolicy, Authenticated
6 from zope.interface import implementer
7
8 from kinto.core import utils
9 from kinto.core.storage import exceptions as storage_exceptions
10
11
12 logger = logging.getLogger(__name__)
13
14 # A permission is called "dynamic" when it's computed at request time.
15 DYNAMIC = "dynamic"
16
17 # When permission is set to "private", only the current user is allowed.
18 PRIVATE = "private"
19
20
21 def groupfinder(userid, request):
22 """Fetch principals from permission backend for the specified `userid`.
23
24 This is plugged by default using the ``multiauth.groupfinder`` setting.
25 """
26 backend = getattr(request.registry, "permission", None)
27 # Permission backend not configured. Ignore.
28 if not backend:
29 return []
30
31 # Safety check when Kinto-Core is used without pyramid_multiauth.
32 if request.prefixed_userid:
33 userid = request.prefixed_userid
34
35 # Query the permission backend only once per request (e.g. batch).
36 reify_key = userid + "_principals"
37 if reify_key not in request.bound_data:
38 principals = backend.get_user_principals(userid)
39 request.bound_data[reify_key] = principals
40
41 return request.bound_data[reify_key]
42
43
44 @implementer(IAuthorizationPolicy)
45 class AuthorizationPolicy:
46 """Default authorization class, that leverages the permission backend
47 for shareable resources.
48 """
49
50 get_bound_permissions = None
51 """Callable that takes an object id and a permission and returns
52 a list of tuples (<object id>, <permission>). Useful when objects
53 permission depend on others."""
54
55 def permits(self, context, principals, permission):
56 if permission == PRIVATE:
57 return Authenticated in principals
58
59 principals = context.get_prefixed_principals()
60
61 if permission == DYNAMIC:
62 permission = context.required_permission
63
64 create_permission = "{}:create".format(context.resource_name)
65 if permission == "create":
66 permission = create_permission
67
68 object_id = context.permission_object_id
69 bound_perms = self._get_bound_permissions(object_id, permission)
70
71 allowed = context.check_permission(principals, bound_perms)
72
73 # Here we consider that parent URI is one path level above.
74 parent_uri = "/".join(object_id.split("/")[:-1]) if object_id else None
75
76 # If not allowed to delete/patch, and target object is missing, and
77 # allowed to read the parent, then view is permitted (will raise 404
78 # later anyway). See Kinto/kinto#918
79 is_record_unknown = not context.on_collection and context.current_record is None
80 if context.required_permission == "write" and is_record_unknown:
81 bound_perms = self._get_bound_permissions(parent_uri, "read")
82 allowed = context.check_permission(principals, bound_perms)
83
84 # If not allowed on this collection, but some records are shared with
85 # the current user, then authorize.
86 # The ShareableResource class will take care of the filtering.
87 is_list_operation = context.on_collection and not permission.endswith("create")
88 if not allowed and is_list_operation:
89 allowed = bool(
90 context.fetch_shared_records(permission, principals, self.get_bound_permissions)
91 )
92 if not allowed:
93 # If allowed to create this kind of object on parent,
94 # then allow to obtain the list.
95 if len(bound_perms) > 0:
96 bound_perms = [(parent_uri, create_permission)]
97 else:
98 bound_perms = [("", "create")] # Root object.
99 allowed = context.check_permission(principals, bound_perms)
100
101 if not allowed:
102 logger.warn(
103 "Permission %r on %r not granted to %r.",
104 permission,
105 object_id,
106 principals[0],
107 extra=dict(userid=principals[0], uri=object_id, perm=permission),
108 )
109
110 return allowed
111
112 def _get_bound_permissions(self, object_id, permission):
113 if self.get_bound_permissions is None:
114 return [(object_id, permission)]
115 return self.get_bound_permissions(object_id, permission)
116
117 def principals_allowed_by_permission(self, context, permission):
118 raise NotImplementedError() # PRAGMA NOCOVER
119
120
121 class RouteFactory:
122 resource_name = None
123 on_collection = False
124 required_permission = None
125 permission_object_id = None
126 current_record = None
127 shared_ids = None
128
129 method_permissions = {
130 "head": "read",
131 "get": "read",
132 "post": "create",
133 "delete": "write",
134 "patch": "write",
135 }
136
137 def __init__(self, request):
138 # Store some shortcuts.
139 permission = request.registry.permission
140 self._check_permission = permission.check_permission
141 self._get_accessible_objects = permission.get_accessible_objects
142
143 self.get_prefixed_principals = functools.partial(utils.prefixed_principals, request)
144
145 # Store current resource and required permission.
146 service = utils.current_service(request)
147 is_on_resource = (
148 service is not None and hasattr(service, "viewset") and hasattr(service, "resource")
149 )
150 if is_on_resource:
151 self.resource_name = request.current_resource_name
152 self.on_collection = getattr(service, "type", None) == "collection"
153
154 # Try to fetch the target object. Its existence will affect permissions checking.
155 if not self.on_collection and request.method.lower() in ("put", "delete", "patch"):
156 resource = service.resource(request=request, context=self)
157 try:
158 # Save a reference, to avoid refetching from storage in resource.
159 self.current_record = resource.model.get_record(resource.record_id)
160 except storage_exceptions.RecordNotFoundError:
161 pass
162
163 self.permission_object_id, self.required_permission = self._find_required_permission(
164 request, service
165 )
166
167 # To obtain shared records on a collection endpoint, use a match:
168 self._object_id_match = self.get_permission_object_id(request, "*")
169
170 self._settings = request.registry.settings
171
172 def check_permission(self, principals, bound_perms):
173 """Read allowed principals from settings, if not any, query the permission
174 backend to check if view is allowed.
175 """
176 if not bound_perms:
177 bound_perms = [(self.resource_name, self.required_permission)]
178 for (_, permission) in bound_perms:
179 setting = "{}_{}_principals".format(self.resource_name, permission)
180 allowed_principals = aslist(self._settings.get(setting, ""))
181 if allowed_principals:
182 if bool(set(allowed_principals) & set(principals)):
183 return True
184 return self._check_permission(principals, bound_perms)
185
186 def fetch_shared_records(self, perm, principals, get_bound_permissions):
187 """Fetch records that are readable or writable for the current
188 principals.
189
190 See :meth:`kinto.core.authorization.AuthorizationPolicy.permits`
191
192 If no record is shared, it returns None.
193
194 .. warning::
195 This sets the ``shared_ids`` attribute to the context with the
196 return value. The attribute is then read by
197 :class:`kinto.core.resource.ShareableResource`
198 """
199 if get_bound_permissions:
200 bound_perms = get_bound_permissions(self._object_id_match, perm)
201 else:
202 bound_perms = [(self._object_id_match, perm)]
203 by_obj_id = self._get_accessible_objects(principals, bound_perms, with_children=False)
204 ids = by_obj_id.keys()
205 # Store for later use in ``ShareableResource``.
206 self.shared_ids = [self._extract_object_id(id_) for id_ in ids]
207 return self.shared_ids
208
209 def get_permission_object_id(self, request, object_id=None):
210 """Returns the permission object id for the current request.
211 In the nominal case, it is just the current URI without version prefix.
212 For collections, it is the related record URI using the specified
213 `object_id`.
214
215 See :meth:`kinto.core.resource.model.SharableModel` and
216 :meth:`kinto.core.authorization.RouteFactory.__init__`
217 """
218 object_uri = utils.strip_uri_prefix(request.path)
219
220 if self.on_collection and object_id is not None:
221 # With the current request on a collection, the record URI must
222 # be found out by inspecting the collection service and its sibling
223 # record service.
224 matchdict = {**request.matchdict, "id": object_id}
225 try:
226 object_uri = utils.instance_uri(request, self.resource_name, **matchdict)
227 object_uri = object_uri.replace("%2A", "*")
228 except KeyError:
229 # Maybe the resource has no single record endpoint.
230 # We consider that object URIs in permissions backend will
231 # be stored naively:
232 object_uri = "{}/{}".format(object_uri, object_id)
233
234 return object_uri
235
236 def _extract_object_id(self, object_uri):
237 # XXX: Rewrite using kinto.core.utils.view_lookup() and matchdict['id']
238 return object_uri.split("/")[-1]
239
240 def _find_required_permission(self, request, service):
241 """Find out what is the permission object id and the required
242 permission.
243
244 .. note::
245 This method saves an attribute ``self.current_record`` used
246 in :class:`kinto.core.resource.UserResource`.
247 """
248 # By default, it's a URI a and permission associated to the method.
249 permission_object_id = self.get_permission_object_id(request)
250 method = request.method.lower()
251 required_permission = self.method_permissions.get(method)
252
253 # For create permission, the object id is the plural endpoint.
254 collection_path = str(service.collection_path)
255 collection_path = collection_path.format_map(request.matchdict)
256
257 # In the case of a "PUT", check if the targetted record already
258 # exists, return "write" if it does, "create" otherwise.
259 if request.method.lower() == "put":
260 if self.current_record is None:
261 # The record does not exist, the permission to create on
262 # the related collection is required.
263 permission_object_id = collection_path
264 required_permission = "create"
265 else:
266 # For safe creations, the user needs a create permission.
267 # See Kinto/kinto#792
268 if request.headers.get("If-None-Match") == "*":
269 permission_object_id = collection_path
270 required_permission = "create"
271 else:
272 required_permission = "write"
273
274 return (permission_object_id, required_permission)
275
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/kinto/core/authorization.py b/kinto/core/authorization.py
--- a/kinto/core/authorization.py
+++ b/kinto/core/authorization.py
@@ -176,7 +176,11 @@
if not bound_perms:
bound_perms = [(self.resource_name, self.required_permission)]
for (_, permission) in bound_perms:
- setting = "{}_{}_principals".format(self.resource_name, permission)
+ # With Kinto inheritance tree, we can have: `permission = "record:create"`
+ if self.resource_name and permission.startswith(self.resource_name):
+ setting = "{}_principals".format(permission.replace(":", "_"))
+ else:
+ setting = "{}_{}_principals".format(self.resource_name, permission)
allowed_principals = aslist(self._settings.get(setting, ""))
if allowed_principals:
if bool(set(allowed_principals) & set(principals)):
| {"golden_diff": "diff --git a/kinto/core/authorization.py b/kinto/core/authorization.py\n--- a/kinto/core/authorization.py\n+++ b/kinto/core/authorization.py\n@@ -176,7 +176,11 @@\n if not bound_perms:\n bound_perms = [(self.resource_name, self.required_permission)]\n for (_, permission) in bound_perms:\n- setting = \"{}_{}_principals\".format(self.resource_name, permission)\n+ # With Kinto inheritance tree, we can have: `permission = \"record:create\"`\n+ if self.resource_name and permission.startswith(self.resource_name):\n+ setting = \"{}_principals\".format(permission.replace(\":\", \"_\"))\n+ else:\n+ setting = \"{}_{}_principals\".format(self.resource_name, permission)\n allowed_principals = aslist(self._settings.get(setting, \"\"))\n if allowed_principals:\n if bool(set(allowed_principals) & set(principals)):\n", "issue": "Invalid account_create_principals key\nThe `account:create` check_permission code is looking at `account_account:create_principals` settings key rather than `account_create_principals`\n", "before_files": [{"content": "import functools\nimport logging\n\nfrom pyramid.settings import aslist\nfrom pyramid.security import IAuthorizationPolicy, Authenticated\nfrom zope.interface import implementer\n\nfrom kinto.core import utils\nfrom kinto.core.storage import exceptions as storage_exceptions\n\n\nlogger = logging.getLogger(__name__)\n\n# A permission is called \"dynamic\" when it's computed at request time.\nDYNAMIC = \"dynamic\"\n\n# When permission is set to \"private\", only the current user is allowed.\nPRIVATE = \"private\"\n\n\ndef groupfinder(userid, request):\n \"\"\"Fetch principals from permission backend for the specified `userid`.\n\n This is plugged by default using the ``multiauth.groupfinder`` setting.\n \"\"\"\n backend = getattr(request.registry, \"permission\", None)\n # Permission backend not configured. Ignore.\n if not backend:\n return []\n\n # Safety check when Kinto-Core is used without pyramid_multiauth.\n if request.prefixed_userid:\n userid = request.prefixed_userid\n\n # Query the permission backend only once per request (e.g. batch).\n reify_key = userid + \"_principals\"\n if reify_key not in request.bound_data:\n principals = backend.get_user_principals(userid)\n request.bound_data[reify_key] = principals\n\n return request.bound_data[reify_key]\n\n\n@implementer(IAuthorizationPolicy)\nclass AuthorizationPolicy:\n \"\"\"Default authorization class, that leverages the permission backend\n for shareable resources.\n \"\"\"\n\n get_bound_permissions = None\n \"\"\"Callable that takes an object id and a permission and returns\n a list of tuples (<object id>, <permission>). Useful when objects\n permission depend on others.\"\"\"\n\n def permits(self, context, principals, permission):\n if permission == PRIVATE:\n return Authenticated in principals\n\n principals = context.get_prefixed_principals()\n\n if permission == DYNAMIC:\n permission = context.required_permission\n\n create_permission = \"{}:create\".format(context.resource_name)\n if permission == \"create\":\n permission = create_permission\n\n object_id = context.permission_object_id\n bound_perms = self._get_bound_permissions(object_id, permission)\n\n allowed = context.check_permission(principals, bound_perms)\n\n # Here we consider that parent URI is one path level above.\n parent_uri = \"/\".join(object_id.split(\"/\")[:-1]) if object_id else None\n\n # If not allowed to delete/patch, and target object is missing, and\n # allowed to read the parent, then view is permitted (will raise 404\n # later anyway). See Kinto/kinto#918\n is_record_unknown = not context.on_collection and context.current_record is None\n if context.required_permission == \"write\" and is_record_unknown:\n bound_perms = self._get_bound_permissions(parent_uri, \"read\")\n allowed = context.check_permission(principals, bound_perms)\n\n # If not allowed on this collection, but some records are shared with\n # the current user, then authorize.\n # The ShareableResource class will take care of the filtering.\n is_list_operation = context.on_collection and not permission.endswith(\"create\")\n if not allowed and is_list_operation:\n allowed = bool(\n context.fetch_shared_records(permission, principals, self.get_bound_permissions)\n )\n if not allowed:\n # If allowed to create this kind of object on parent,\n # then allow to obtain the list.\n if len(bound_perms) > 0:\n bound_perms = [(parent_uri, create_permission)]\n else:\n bound_perms = [(\"\", \"create\")] # Root object.\n allowed = context.check_permission(principals, bound_perms)\n\n if not allowed:\n logger.warn(\n \"Permission %r on %r not granted to %r.\",\n permission,\n object_id,\n principals[0],\n extra=dict(userid=principals[0], uri=object_id, perm=permission),\n )\n\n return allowed\n\n def _get_bound_permissions(self, object_id, permission):\n if self.get_bound_permissions is None:\n return [(object_id, permission)]\n return self.get_bound_permissions(object_id, permission)\n\n def principals_allowed_by_permission(self, context, permission):\n raise NotImplementedError() # PRAGMA NOCOVER\n\n\nclass RouteFactory:\n resource_name = None\n on_collection = False\n required_permission = None\n permission_object_id = None\n current_record = None\n shared_ids = None\n\n method_permissions = {\n \"head\": \"read\",\n \"get\": \"read\",\n \"post\": \"create\",\n \"delete\": \"write\",\n \"patch\": \"write\",\n }\n\n def __init__(self, request):\n # Store some shortcuts.\n permission = request.registry.permission\n self._check_permission = permission.check_permission\n self._get_accessible_objects = permission.get_accessible_objects\n\n self.get_prefixed_principals = functools.partial(utils.prefixed_principals, request)\n\n # Store current resource and required permission.\n service = utils.current_service(request)\n is_on_resource = (\n service is not None and hasattr(service, \"viewset\") and hasattr(service, \"resource\")\n )\n if is_on_resource:\n self.resource_name = request.current_resource_name\n self.on_collection = getattr(service, \"type\", None) == \"collection\"\n\n # Try to fetch the target object. Its existence will affect permissions checking.\n if not self.on_collection and request.method.lower() in (\"put\", \"delete\", \"patch\"):\n resource = service.resource(request=request, context=self)\n try:\n # Save a reference, to avoid refetching from storage in resource.\n self.current_record = resource.model.get_record(resource.record_id)\n except storage_exceptions.RecordNotFoundError:\n pass\n\n self.permission_object_id, self.required_permission = self._find_required_permission(\n request, service\n )\n\n # To obtain shared records on a collection endpoint, use a match:\n self._object_id_match = self.get_permission_object_id(request, \"*\")\n\n self._settings = request.registry.settings\n\n def check_permission(self, principals, bound_perms):\n \"\"\"Read allowed principals from settings, if not any, query the permission\n backend to check if view is allowed.\n \"\"\"\n if not bound_perms:\n bound_perms = [(self.resource_name, self.required_permission)]\n for (_, permission) in bound_perms:\n setting = \"{}_{}_principals\".format(self.resource_name, permission)\n allowed_principals = aslist(self._settings.get(setting, \"\"))\n if allowed_principals:\n if bool(set(allowed_principals) & set(principals)):\n return True\n return self._check_permission(principals, bound_perms)\n\n def fetch_shared_records(self, perm, principals, get_bound_permissions):\n \"\"\"Fetch records that are readable or writable for the current\n principals.\n\n See :meth:`kinto.core.authorization.AuthorizationPolicy.permits`\n\n If no record is shared, it returns None.\n\n .. warning::\n This sets the ``shared_ids`` attribute to the context with the\n return value. The attribute is then read by\n :class:`kinto.core.resource.ShareableResource`\n \"\"\"\n if get_bound_permissions:\n bound_perms = get_bound_permissions(self._object_id_match, perm)\n else:\n bound_perms = [(self._object_id_match, perm)]\n by_obj_id = self._get_accessible_objects(principals, bound_perms, with_children=False)\n ids = by_obj_id.keys()\n # Store for later use in ``ShareableResource``.\n self.shared_ids = [self._extract_object_id(id_) for id_ in ids]\n return self.shared_ids\n\n def get_permission_object_id(self, request, object_id=None):\n \"\"\"Returns the permission object id for the current request.\n In the nominal case, it is just the current URI without version prefix.\n For collections, it is the related record URI using the specified\n `object_id`.\n\n See :meth:`kinto.core.resource.model.SharableModel` and\n :meth:`kinto.core.authorization.RouteFactory.__init__`\n \"\"\"\n object_uri = utils.strip_uri_prefix(request.path)\n\n if self.on_collection and object_id is not None:\n # With the current request on a collection, the record URI must\n # be found out by inspecting the collection service and its sibling\n # record service.\n matchdict = {**request.matchdict, \"id\": object_id}\n try:\n object_uri = utils.instance_uri(request, self.resource_name, **matchdict)\n object_uri = object_uri.replace(\"%2A\", \"*\")\n except KeyError:\n # Maybe the resource has no single record endpoint.\n # We consider that object URIs in permissions backend will\n # be stored naively:\n object_uri = \"{}/{}\".format(object_uri, object_id)\n\n return object_uri\n\n def _extract_object_id(self, object_uri):\n # XXX: Rewrite using kinto.core.utils.view_lookup() and matchdict['id']\n return object_uri.split(\"/\")[-1]\n\n def _find_required_permission(self, request, service):\n \"\"\"Find out what is the permission object id and the required\n permission.\n\n .. note::\n This method saves an attribute ``self.current_record`` used\n in :class:`kinto.core.resource.UserResource`.\n \"\"\"\n # By default, it's a URI a and permission associated to the method.\n permission_object_id = self.get_permission_object_id(request)\n method = request.method.lower()\n required_permission = self.method_permissions.get(method)\n\n # For create permission, the object id is the plural endpoint.\n collection_path = str(service.collection_path)\n collection_path = collection_path.format_map(request.matchdict)\n\n # In the case of a \"PUT\", check if the targetted record already\n # exists, return \"write\" if it does, \"create\" otherwise.\n if request.method.lower() == \"put\":\n if self.current_record is None:\n # The record does not exist, the permission to create on\n # the related collection is required.\n permission_object_id = collection_path\n required_permission = \"create\"\n else:\n # For safe creations, the user needs a create permission.\n # See Kinto/kinto#792\n if request.headers.get(\"If-None-Match\") == \"*\":\n permission_object_id = collection_path\n required_permission = \"create\"\n else:\n required_permission = \"write\"\n\n return (permission_object_id, required_permission)\n", "path": "kinto/core/authorization.py"}], "after_files": [{"content": "import functools\nimport logging\n\nfrom pyramid.settings import aslist\nfrom pyramid.security import IAuthorizationPolicy, Authenticated\nfrom zope.interface import implementer\n\nfrom kinto.core import utils\nfrom kinto.core.storage import exceptions as storage_exceptions\n\n\nlogger = logging.getLogger(__name__)\n\n# A permission is called \"dynamic\" when it's computed at request time.\nDYNAMIC = \"dynamic\"\n\n# When permission is set to \"private\", only the current user is allowed.\nPRIVATE = \"private\"\n\n\ndef groupfinder(userid, request):\n \"\"\"Fetch principals from permission backend for the specified `userid`.\n\n This is plugged by default using the ``multiauth.groupfinder`` setting.\n \"\"\"\n backend = getattr(request.registry, \"permission\", None)\n # Permission backend not configured. Ignore.\n if not backend:\n return []\n\n # Safety check when Kinto-Core is used without pyramid_multiauth.\n if request.prefixed_userid:\n userid = request.prefixed_userid\n\n # Query the permission backend only once per request (e.g. batch).\n reify_key = userid + \"_principals\"\n if reify_key not in request.bound_data:\n principals = backend.get_user_principals(userid)\n request.bound_data[reify_key] = principals\n\n return request.bound_data[reify_key]\n\n\n@implementer(IAuthorizationPolicy)\nclass AuthorizationPolicy:\n \"\"\"Default authorization class, that leverages the permission backend\n for shareable resources.\n \"\"\"\n\n get_bound_permissions = None\n \"\"\"Callable that takes an object id and a permission and returns\n a list of tuples (<object id>, <permission>). Useful when objects\n permission depend on others.\"\"\"\n\n def permits(self, context, principals, permission):\n if permission == PRIVATE:\n return Authenticated in principals\n\n principals = context.get_prefixed_principals()\n\n if permission == DYNAMIC:\n permission = context.required_permission\n\n create_permission = \"{}:create\".format(context.resource_name)\n if permission == \"create\":\n permission = create_permission\n\n object_id = context.permission_object_id\n bound_perms = self._get_bound_permissions(object_id, permission)\n\n allowed = context.check_permission(principals, bound_perms)\n\n # Here we consider that parent URI is one path level above.\n parent_uri = \"/\".join(object_id.split(\"/\")[:-1]) if object_id else None\n\n # If not allowed to delete/patch, and target object is missing, and\n # allowed to read the parent, then view is permitted (will raise 404\n # later anyway). See Kinto/kinto#918\n is_record_unknown = not context.on_collection and context.current_record is None\n if context.required_permission == \"write\" and is_record_unknown:\n bound_perms = self._get_bound_permissions(parent_uri, \"read\")\n allowed = context.check_permission(principals, bound_perms)\n\n # If not allowed on this collection, but some records are shared with\n # the current user, then authorize.\n # The ShareableResource class will take care of the filtering.\n is_list_operation = context.on_collection and not permission.endswith(\"create\")\n if not allowed and is_list_operation:\n allowed = bool(\n context.fetch_shared_records(permission, principals, self.get_bound_permissions)\n )\n if not allowed:\n # If allowed to create this kind of object on parent,\n # then allow to obtain the list.\n if len(bound_perms) > 0:\n bound_perms = [(parent_uri, create_permission)]\n else:\n bound_perms = [(\"\", \"create\")] # Root object.\n allowed = context.check_permission(principals, bound_perms)\n\n if not allowed:\n logger.warn(\n \"Permission %r on %r not granted to %r.\",\n permission,\n object_id,\n principals[0],\n extra=dict(userid=principals[0], uri=object_id, perm=permission),\n )\n\n return allowed\n\n def _get_bound_permissions(self, object_id, permission):\n if self.get_bound_permissions is None:\n return [(object_id, permission)]\n return self.get_bound_permissions(object_id, permission)\n\n def principals_allowed_by_permission(self, context, permission):\n raise NotImplementedError() # PRAGMA NOCOVER\n\n\nclass RouteFactory:\n resource_name = None\n on_collection = False\n required_permission = None\n permission_object_id = None\n current_record = None\n shared_ids = None\n\n method_permissions = {\n \"head\": \"read\",\n \"get\": \"read\",\n \"post\": \"create\",\n \"delete\": \"write\",\n \"patch\": \"write\",\n }\n\n def __init__(self, request):\n # Store some shortcuts.\n permission = request.registry.permission\n self._check_permission = permission.check_permission\n self._get_accessible_objects = permission.get_accessible_objects\n\n self.get_prefixed_principals = functools.partial(utils.prefixed_principals, request)\n\n # Store current resource and required permission.\n service = utils.current_service(request)\n is_on_resource = (\n service is not None and hasattr(service, \"viewset\") and hasattr(service, \"resource\")\n )\n if is_on_resource:\n self.resource_name = request.current_resource_name\n self.on_collection = getattr(service, \"type\", None) == \"collection\"\n\n # Try to fetch the target object. Its existence will affect permissions checking.\n if not self.on_collection and request.method.lower() in (\"put\", \"delete\", \"patch\"):\n resource = service.resource(request=request, context=self)\n try:\n # Save a reference, to avoid refetching from storage in resource.\n self.current_record = resource.model.get_record(resource.record_id)\n except storage_exceptions.RecordNotFoundError:\n pass\n\n self.permission_object_id, self.required_permission = self._find_required_permission(\n request, service\n )\n\n # To obtain shared records on a collection endpoint, use a match:\n self._object_id_match = self.get_permission_object_id(request, \"*\")\n\n self._settings = request.registry.settings\n\n def check_permission(self, principals, bound_perms):\n \"\"\"Read allowed principals from settings, if not any, query the permission\n backend to check if view is allowed.\n \"\"\"\n if not bound_perms:\n bound_perms = [(self.resource_name, self.required_permission)]\n for (_, permission) in bound_perms:\n # With Kinto inheritance tree, we can have: `permission = \"record:create\"`\n if self.resource_name and permission.startswith(self.resource_name):\n setting = \"{}_principals\".format(permission.replace(\":\", \"_\"))\n else:\n setting = \"{}_{}_principals\".format(self.resource_name, permission)\n allowed_principals = aslist(self._settings.get(setting, \"\"))\n if allowed_principals:\n if bool(set(allowed_principals) & set(principals)):\n return True\n return self._check_permission(principals, bound_perms)\n\n def fetch_shared_records(self, perm, principals, get_bound_permissions):\n \"\"\"Fetch records that are readable or writable for the current\n principals.\n\n See :meth:`kinto.core.authorization.AuthorizationPolicy.permits`\n\n If no record is shared, it returns None.\n\n .. warning::\n This sets the ``shared_ids`` attribute to the context with the\n return value. The attribute is then read by\n :class:`kinto.core.resource.ShareableResource`\n \"\"\"\n if get_bound_permissions:\n bound_perms = get_bound_permissions(self._object_id_match, perm)\n else:\n bound_perms = [(self._object_id_match, perm)]\n by_obj_id = self._get_accessible_objects(principals, bound_perms, with_children=False)\n ids = by_obj_id.keys()\n # Store for later use in ``ShareableResource``.\n self.shared_ids = [self._extract_object_id(id_) for id_ in ids]\n return self.shared_ids\n\n def get_permission_object_id(self, request, object_id=None):\n \"\"\"Returns the permission object id for the current request.\n In the nominal case, it is just the current URI without version prefix.\n For collections, it is the related record URI using the specified\n `object_id`.\n\n See :meth:`kinto.core.resource.model.SharableModel` and\n :meth:`kinto.core.authorization.RouteFactory.__init__`\n \"\"\"\n object_uri = utils.strip_uri_prefix(request.path)\n\n if self.on_collection and object_id is not None:\n # With the current request on a collection, the record URI must\n # be found out by inspecting the collection service and its sibling\n # record service.\n matchdict = {**request.matchdict, \"id\": object_id}\n try:\n object_uri = utils.instance_uri(request, self.resource_name, **matchdict)\n object_uri = object_uri.replace(\"%2A\", \"*\")\n except KeyError:\n # Maybe the resource has no single record endpoint.\n # We consider that object URIs in permissions backend will\n # be stored naively:\n object_uri = \"{}/{}\".format(object_uri, object_id)\n\n return object_uri\n\n def _extract_object_id(self, object_uri):\n # XXX: Rewrite using kinto.core.utils.view_lookup() and matchdict['id']\n return object_uri.split(\"/\")[-1]\n\n def _find_required_permission(self, request, service):\n \"\"\"Find out what is the permission object id and the required\n permission.\n\n .. note::\n This method saves an attribute ``self.current_record`` used\n in :class:`kinto.core.resource.UserResource`.\n \"\"\"\n # By default, it's a URI a and permission associated to the method.\n permission_object_id = self.get_permission_object_id(request)\n method = request.method.lower()\n required_permission = self.method_permissions.get(method)\n\n # For create permission, the object id is the plural endpoint.\n collection_path = str(service.collection_path)\n collection_path = collection_path.format_map(request.matchdict)\n\n # In the case of a \"PUT\", check if the targetted record already\n # exists, return \"write\" if it does, \"create\" otherwise.\n if request.method.lower() == \"put\":\n if self.current_record is None:\n # The record does not exist, the permission to create on\n # the related collection is required.\n permission_object_id = collection_path\n required_permission = \"create\"\n else:\n # For safe creations, the user needs a create permission.\n # See Kinto/kinto#792\n if request.headers.get(\"If-None-Match\") == \"*\":\n permission_object_id = collection_path\n required_permission = \"create\"\n else:\n required_permission = \"write\"\n\n return (permission_object_id, required_permission)\n", "path": "kinto/core/authorization.py"}]} | 3,287 | 207 |
gh_patches_debug_20901 | rasdani/github-patches | git_diff | litestar-org__litestar-1426 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
StaticFilesConfig and virtual directories
I'm trying to write a ``FileSystemProtocol`` to load files from the package data using [importlib_resources](https://importlib-resources.readthedocs.io/en/latest/using.html#). But because ``directories`` is defined as ``DirectoryPath``, pydantic checks if the given directories exist in the local filesystem.
This is not generally true, especially in any kind of virtual filesystem (e.g. a zipped package). I think this condition should be relaxed to support virtual filesystems.
https://github.com/starlite-api/starlite/blob/9bb6dcd57c10a591377cf8e3a537e9292566d5b9/starlite/config/static_files.py#L32
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `starlite/contrib/sqlalchemy/init_plugin/config/common.py`
Content:
```
1 from __future__ import annotations
2
3 from dataclasses import dataclass, field
4 from typing import TYPE_CHECKING, Callable, Generic, TypeVar, cast
5
6 from starlite.constants import HTTP_DISCONNECT, HTTP_RESPONSE_START, WEBSOCKET_CLOSE, WEBSOCKET_DISCONNECT
7 from starlite.exceptions import ImproperlyConfiguredException
8 from starlite.types import Empty
9 from starlite.utils import get_starlite_scope_state, set_starlite_scope_state
10 from starlite.utils.dataclass import simple_asdict
11
12 from .engine import EngineConfig
13
14 if TYPE_CHECKING:
15 from typing import Any
16
17 from sqlalchemy import Connection, Engine
18 from sqlalchemy.ext.asyncio import AsyncConnection, AsyncEngine, AsyncSession, async_sessionmaker
19 from sqlalchemy.orm import Mapper, Query, Session, sessionmaker
20 from sqlalchemy.orm.session import JoinTransactionMode
21 from sqlalchemy.sql import TableClause
22
23 from starlite.datastructures.state import State
24 from starlite.types import BeforeMessageSendHookHandler, EmptyType, Scope
25
26 __all__ = (
27 "SESSION_SCOPE_KEY",
28 "SESSION_TERMINUS_ASGI_EVENTS",
29 "GenericSQLAlchemyConfig",
30 "GenericSessionConfig",
31 )
32
33 SESSION_SCOPE_KEY = "_sqlalchemy_db_session"
34 SESSION_TERMINUS_ASGI_EVENTS = {HTTP_RESPONSE_START, HTTP_DISCONNECT, WEBSOCKET_DISCONNECT, WEBSOCKET_CLOSE}
35
36 ConnectionT = TypeVar("ConnectionT", bound="Connection | AsyncConnection")
37 EngineT = TypeVar("EngineT", bound="Engine | AsyncEngine")
38 SessionT = TypeVar("SessionT", bound="Session | AsyncSession")
39 SessionMakerT = TypeVar("SessionMakerT", bound="sessionmaker | async_sessionmaker")
40
41
42 @dataclass
43 class GenericSessionConfig(Generic[ConnectionT, EngineT, SessionT]):
44 """SQLAlchemy async session config."""
45
46 autobegin: bool | EmptyType = Empty
47 """Automatically start transactions when database access is requested by an operation."""
48 autoflush: bool | EmptyType = Empty
49 """When ``True``, all query operations will issue a flush call to this :class:`Session <sqlalchemy.orm.Session>`
50 before proceeding"""
51 bind: EngineT | ConnectionT | None | EmptyType = Empty
52 """The :class:`Engine <sqlalchemy.engine.Engine>` or :class:`Connection <sqlalchemy.engine.Connection>` that new
53 :class:`Session <sqlalchemy.orm.Session>` objects will be bound to."""
54 binds: dict[type[Any] | Mapper[Any] | TableClause | str, EngineT | ConnectionT] | None | EmptyType = Empty
55 """A dictionary which may specify any number of :class:`Engine <sqlalchemy.engine.Engine>` or :class:`Connection
56 <sqlalchemy.engine.Connection>` objects as the source of connectivity for SQL operations on a per-entity basis. The
57 keys of the dictionary consist of any series of mapped classes, arbitrary Python classes that are bases for mapped
58 classes, :class:`Table <sqlalchemy.schema.Table>` objects and :class:`Mapper <sqlalchemy.orm.Mapper>` objects. The
59 values of the dictionary are then instances of :class:`Engine <sqlalchemy.engine.Engine>` or less commonly
60 :class:`Connection <sqlalchemy.engine.Connection>` objects."""
61 class_: type[SessionT] | EmptyType = Empty
62 """Class to use in order to create new :class:`Session <sqlalchemy.orm.Session>` objects."""
63 expire_on_commit: bool | EmptyType = Empty
64 """If ``True``, all instances will be expired after each commit."""
65 info: dict[str, Any] | None | EmptyType = Empty
66 """Optional dictionary of information that will be available via the
67 :attr:`Session.info <sqlalchemy.orm.Session.info>`"""
68 join_transaction_mode: JoinTransactionMode | EmptyType = Empty
69 """Describes the transactional behavior to take when a given bind is a Connection that has already begun a
70 transaction outside the scope of this Session; in other words the
71 :attr:`Connection.in_transaction() <sqlalchemy.Connection.in_transaction>` method returns True."""
72 query_cls: type[Query] | None | EmptyType = Empty
73 """Class which should be used to create new Query objects, as returned by the
74 :attr:`Session.query() <sqlalchemy.orm.Session.query>` method."""
75 twophase: bool | EmptyType = Empty
76 """When ``True``, all transactions will be started as a “two phase” transaction, i.e. using the “two phase”
77 semantics of the database in use along with an XID. During a :attr:`commit() <sqlalchemy.orm.Session.commit>`, after
78 :attr:`flush() <sqlalchemy.orm.Session.flush>` has been issued for all attached databases, the
79 :attr:`TwoPhaseTransaction.prepare() <sqlalchemy.engine.TwoPhaseTransaction.prepare>` method on each database`s
80 :class:`TwoPhaseTransaction <sqlalchemy.engine.TwoPhaseTransaction>` will be called. This allows each database to
81 roll back the entire transaction, before each transaction is committed."""
82
83
84 @dataclass
85 class GenericSQLAlchemyConfig(Generic[EngineT, SessionT, SessionMakerT]):
86 """Common SQLAlchemy Configuration."""
87
88 create_engine_callable: Callable[[str], EngineT]
89 """Callable that creates an :class:`AsyncEngine <sqlalchemy.ext.asyncio.AsyncEngine>` instance or instance of its
90 subclass.
91 """
92 session_config: GenericSessionConfig
93 """Configuration options for either the :class:`async_sessionmaker <sqlalchemy.ext.asyncio.async_sessionmaker>`
94 or :class:`sessionmaker <sqlalchemy.orm.sessionmaker>`.
95 """
96 session_maker_class: type[sessionmaker] | type[async_sessionmaker]
97 """Sessionmaker class to use."""
98 before_send_handler: BeforeMessageSendHookHandler
99 """Handler to call before the ASGI message is sent.
100
101 The handler should handle closing the session stored in the ASGI scope, if its still open, and committing and
102 uncommitted data.
103 """
104 connection_string: str | None = field(default=None)
105 """Database connection string in one of the formats supported by SQLAlchemy.
106
107 Notes:
108 - For async connections, the connection string must include the correct async prefix.
109 e.g. ``'postgresql+asyncpg://...'`` instead of ``'postgresql://'``, and for sync connections its the opposite.
110
111 """
112 engine_dependency_key: str = "db_engine"
113 """Key to use for the dependency injection of database engines."""
114 session_dependency_key: str = "db_session"
115 """Key to use for the dependency injection of database sessions."""
116 engine_app_state_key: str = "db_engine"
117 """Key under which to store the SQLAlchemy engine in the application :class:`State <.datastructures.State>`
118 instance.
119 """
120 engine_config: EngineConfig = field(default_factory=EngineConfig)
121 """Configuration for the SQLAlchemy engine.
122
123 The configuration options are documented in the SQLAlchemy documentation.
124 """
125 session_maker_app_state_key: str = "session_maker_class"
126 """Key under which to store the SQLAlchemy :class:`sessionmaker <sqlalchemy.orm.sessionmaker>` in the application
127 :class:`State <.datastructures.State>` instance.
128 """
129 session_maker: Callable[[], SessionT] | None = None
130 """Callable that returns a session.
131
132 If provided, the plugin will use this rather than instantiate a sessionmaker.
133 """
134 engine_instance: EngineT | None = None
135 """Optional engine to use.
136
137 If set, the plugin will use the provided instance rather than instantiate an engine.
138 """
139
140 def __post_init__(self) -> None:
141 if self.connection_string is not None and self.engine_instance is not None:
142 raise ImproperlyConfiguredException("Only one of 'connection_string' or 'engine_instance' can be provided.")
143
144 @property
145 def engine_config_dict(self) -> dict[str, Any]:
146 """Return the engine configuration as a dict.
147
148 Returns:
149 A string keyed dict of config kwargs for the SQLAlchemy :func:`create_engine <sqlalchemy.create_engine>`
150 function.
151 """
152 return simple_asdict(self.engine_config, exclude_empty=True)
153
154 @property
155 def session_config_dict(self) -> dict[str, Any]:
156 """Return the session configuration as a dict.
157
158 Returns:
159 A string keyed dict of config kwargs for the SQLAlchemy :class:`sessionmaker <sqlalchemy.orm.sessionmaker>`
160 class.
161 """
162 return simple_asdict(self.session_config, exclude_empty=True)
163
164 @property
165 def signature_namespace(self) -> dict[str, Any]:
166 """Return the plugin's signature namespace.
167
168 Returns:
169 A string keyed dict of names to be added to the namespace for signature forward reference resolution.
170 """
171 return {} # pragma: no cover
172
173 def create_engine(self) -> EngineT:
174 """Return an engine. If none exists yet, create one.
175
176 Returns:
177 Getter that returns the engine instance used by the plugin.
178 """
179 if self.engine_instance:
180 return self.engine_instance
181
182 if self.connection_string is None:
183 raise ImproperlyConfiguredException("One of 'connection_string' or 'engine_instance' must be provided.")
184
185 engine_config = self.engine_config_dict
186 try:
187 return self.create_engine_callable(self.connection_string, **engine_config)
188 except ValueError:
189 # likely due to a dialect that doesn't support json type
190 del engine_config["json_deserializer"]
191 del engine_config["json_serializer"]
192 return self.create_engine_callable(self.connection_string, **engine_config)
193
194 def create_session_maker(self) -> Callable[[], SessionT]:
195 """Get a session maker. If none exists yet, create one.
196
197 Returns:
198 Session factory used by the plugin.
199 """
200 if self.session_maker:
201 return self.session_maker
202
203 session_kws = self.session_config_dict
204 if session_kws.get("bind") is None:
205 session_kws["bind"] = self.create_engine()
206 return self.session_maker_class(**session_kws)
207
208 def provide_engine(self, state: State) -> EngineT:
209 """Create an engine instance.
210
211 Args:
212 state: The ``Starlite.state`` instance.
213
214 Returns:
215 An engine instance.
216 """
217 return cast("EngineT", state.get(self.engine_app_state_key))
218
219 def provide_session(self, state: State, scope: Scope) -> SessionT:
220 """Create a session instance.
221
222 Args:
223 state: The ``Starlite.state`` instance.
224 scope: The current connection's scope.
225
226 Returns:
227 A session instance.
228 """
229 session = cast("SessionT | None", get_starlite_scope_state(scope, SESSION_SCOPE_KEY))
230 if session is None:
231 session_maker = cast("Callable[[], SessionT]", state[self.session_maker_app_state_key])
232 session = session_maker()
233 set_starlite_scope_state(scope, SESSION_SCOPE_KEY, session)
234 return session
235
236 def create_app_state_items(self) -> dict[str, Any]:
237 """Key/value pairs to be stored in application state."""
238 return {
239 self.engine_app_state_key: self.create_engine(),
240 self.session_maker_app_state_key: self.create_session_maker(),
241 }
242
```
Path: `starlite/contrib/sqlalchemy/init_plugin/plugin.py`
Content:
```
1 from __future__ import annotations
2
3 from typing import TYPE_CHECKING
4
5 from starlite.di import Provide
6 from starlite.plugins import InitPluginProtocol
7
8 if TYPE_CHECKING:
9 from starlite.config.app import AppConfig
10
11 from .config import SQLAlchemyAsyncConfig, SQLAlchemySyncConfig
12
13 __all__ = ("SQLAlchemyInitPlugin",)
14
15
16 class SQLAlchemyInitPlugin(InitPluginProtocol):
17 """SQLAlchemy application lifecycle configuration."""
18
19 __slots__ = ("_config",)
20
21 def __init__(self, config: SQLAlchemyAsyncConfig | SQLAlchemySyncConfig) -> None:
22 """Initialize ``SQLAlchemyPlugin``.
23
24 Args:
25 config: configure DB connection and hook handlers and dependencies.
26 """
27 self._config = config
28
29 def on_app_init(self, app_config: AppConfig) -> AppConfig:
30 """Configure application for use with SQLAlchemy.
31
32 Args:
33 app_config: The :class:`AppConfig <.config.app.AppConfig>` instance.
34 """
35 app_config.dependencies.update(
36 {
37 self._config.engine_dependency_key: Provide(self._config.provide_engine),
38 self._config.session_dependency_key: Provide(self._config.provide_session),
39 }
40 )
41 app_config.before_send.append(self._config.before_send_handler)
42 app_config.on_shutdown.append(self._config.on_shutdown)
43 app_config.state.update(self._config.create_app_state_items())
44 app_config.signature_namespace.update(self._config.signature_namespace)
45 return app_config
46
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/starlite/contrib/sqlalchemy/init_plugin/config/common.py b/starlite/contrib/sqlalchemy/init_plugin/config/common.py
--- a/starlite/contrib/sqlalchemy/init_plugin/config/common.py
+++ b/starlite/contrib/sqlalchemy/init_plugin/config/common.py
@@ -239,3 +239,11 @@
self.engine_app_state_key: self.create_engine(),
self.session_maker_app_state_key: self.create_session_maker(),
}
+
+ def update_app_state(self, state: State) -> None:
+ """Set the app state with engine and session.
+
+ Args:
+ state: The ``Starlite.state`` instance.
+ """
+ state.update(self.create_app_state_items())
diff --git a/starlite/contrib/sqlalchemy/init_plugin/plugin.py b/starlite/contrib/sqlalchemy/init_plugin/plugin.py
--- a/starlite/contrib/sqlalchemy/init_plugin/plugin.py
+++ b/starlite/contrib/sqlalchemy/init_plugin/plugin.py
@@ -39,7 +39,7 @@
}
)
app_config.before_send.append(self._config.before_send_handler)
+ app_config.on_startup.append(self._config.update_app_state)
app_config.on_shutdown.append(self._config.on_shutdown)
- app_config.state.update(self._config.create_app_state_items())
app_config.signature_namespace.update(self._config.signature_namespace)
return app_config
| {"golden_diff": "diff --git a/starlite/contrib/sqlalchemy/init_plugin/config/common.py b/starlite/contrib/sqlalchemy/init_plugin/config/common.py\n--- a/starlite/contrib/sqlalchemy/init_plugin/config/common.py\n+++ b/starlite/contrib/sqlalchemy/init_plugin/config/common.py\n@@ -239,3 +239,11 @@\n self.engine_app_state_key: self.create_engine(),\n self.session_maker_app_state_key: self.create_session_maker(),\n }\n+\n+ def update_app_state(self, state: State) -> None:\n+ \"\"\"Set the app state with engine and session.\n+\n+ Args:\n+ state: The ``Starlite.state`` instance.\n+ \"\"\"\n+ state.update(self.create_app_state_items())\ndiff --git a/starlite/contrib/sqlalchemy/init_plugin/plugin.py b/starlite/contrib/sqlalchemy/init_plugin/plugin.py\n--- a/starlite/contrib/sqlalchemy/init_plugin/plugin.py\n+++ b/starlite/contrib/sqlalchemy/init_plugin/plugin.py\n@@ -39,7 +39,7 @@\n }\n )\n app_config.before_send.append(self._config.before_send_handler)\n+ app_config.on_startup.append(self._config.update_app_state)\n app_config.on_shutdown.append(self._config.on_shutdown)\n- app_config.state.update(self._config.create_app_state_items())\n app_config.signature_namespace.update(self._config.signature_namespace)\n return app_config\n", "issue": "StaticFilesConfig and virtual directories\nI'm trying to write a ``FileSystemProtocol`` to load files from the package data using [importlib_resources](https://importlib-resources.readthedocs.io/en/latest/using.html#). But because ``directories`` is defined as ``DirectoryPath``, pydantic checks if the given directories exist in the local filesystem. \r\n\r\nThis is not generally true, especially in any kind of virtual filesystem (e.g. a zipped package). I think this condition should be relaxed to support virtual filesystems.\r\n\r\nhttps://github.com/starlite-api/starlite/blob/9bb6dcd57c10a591377cf8e3a537e9292566d5b9/starlite/config/static_files.py#L32\n", "before_files": [{"content": "from __future__ import annotations\n\nfrom dataclasses import dataclass, field\nfrom typing import TYPE_CHECKING, Callable, Generic, TypeVar, cast\n\nfrom starlite.constants import HTTP_DISCONNECT, HTTP_RESPONSE_START, WEBSOCKET_CLOSE, WEBSOCKET_DISCONNECT\nfrom starlite.exceptions import ImproperlyConfiguredException\nfrom starlite.types import Empty\nfrom starlite.utils import get_starlite_scope_state, set_starlite_scope_state\nfrom starlite.utils.dataclass import simple_asdict\n\nfrom .engine import EngineConfig\n\nif TYPE_CHECKING:\n from typing import Any\n\n from sqlalchemy import Connection, Engine\n from sqlalchemy.ext.asyncio import AsyncConnection, AsyncEngine, AsyncSession, async_sessionmaker\n from sqlalchemy.orm import Mapper, Query, Session, sessionmaker\n from sqlalchemy.orm.session import JoinTransactionMode\n from sqlalchemy.sql import TableClause\n\n from starlite.datastructures.state import State\n from starlite.types import BeforeMessageSendHookHandler, EmptyType, Scope\n\n__all__ = (\n \"SESSION_SCOPE_KEY\",\n \"SESSION_TERMINUS_ASGI_EVENTS\",\n \"GenericSQLAlchemyConfig\",\n \"GenericSessionConfig\",\n)\n\nSESSION_SCOPE_KEY = \"_sqlalchemy_db_session\"\nSESSION_TERMINUS_ASGI_EVENTS = {HTTP_RESPONSE_START, HTTP_DISCONNECT, WEBSOCKET_DISCONNECT, WEBSOCKET_CLOSE}\n\nConnectionT = TypeVar(\"ConnectionT\", bound=\"Connection | AsyncConnection\")\nEngineT = TypeVar(\"EngineT\", bound=\"Engine | AsyncEngine\")\nSessionT = TypeVar(\"SessionT\", bound=\"Session | AsyncSession\")\nSessionMakerT = TypeVar(\"SessionMakerT\", bound=\"sessionmaker | async_sessionmaker\")\n\n\n@dataclass\nclass GenericSessionConfig(Generic[ConnectionT, EngineT, SessionT]):\n \"\"\"SQLAlchemy async session config.\"\"\"\n\n autobegin: bool | EmptyType = Empty\n \"\"\"Automatically start transactions when database access is requested by an operation.\"\"\"\n autoflush: bool | EmptyType = Empty\n \"\"\"When ``True``, all query operations will issue a flush call to this :class:`Session <sqlalchemy.orm.Session>`\n before proceeding\"\"\"\n bind: EngineT | ConnectionT | None | EmptyType = Empty\n \"\"\"The :class:`Engine <sqlalchemy.engine.Engine>` or :class:`Connection <sqlalchemy.engine.Connection>` that new\n :class:`Session <sqlalchemy.orm.Session>` objects will be bound to.\"\"\"\n binds: dict[type[Any] | Mapper[Any] | TableClause | str, EngineT | ConnectionT] | None | EmptyType = Empty\n \"\"\"A dictionary which may specify any number of :class:`Engine <sqlalchemy.engine.Engine>` or :class:`Connection\n <sqlalchemy.engine.Connection>` objects as the source of connectivity for SQL operations on a per-entity basis. The\n keys of the dictionary consist of any series of mapped classes, arbitrary Python classes that are bases for mapped\n classes, :class:`Table <sqlalchemy.schema.Table>` objects and :class:`Mapper <sqlalchemy.orm.Mapper>` objects. The\n values of the dictionary are then instances of :class:`Engine <sqlalchemy.engine.Engine>` or less commonly\n :class:`Connection <sqlalchemy.engine.Connection>` objects.\"\"\"\n class_: type[SessionT] | EmptyType = Empty\n \"\"\"Class to use in order to create new :class:`Session <sqlalchemy.orm.Session>` objects.\"\"\"\n expire_on_commit: bool | EmptyType = Empty\n \"\"\"If ``True``, all instances will be expired after each commit.\"\"\"\n info: dict[str, Any] | None | EmptyType = Empty\n \"\"\"Optional dictionary of information that will be available via the\n :attr:`Session.info <sqlalchemy.orm.Session.info>`\"\"\"\n join_transaction_mode: JoinTransactionMode | EmptyType = Empty\n \"\"\"Describes the transactional behavior to take when a given bind is a Connection that has already begun a\n transaction outside the scope of this Session; in other words the\n :attr:`Connection.in_transaction() <sqlalchemy.Connection.in_transaction>` method returns True.\"\"\"\n query_cls: type[Query] | None | EmptyType = Empty\n \"\"\"Class which should be used to create new Query objects, as returned by the\n :attr:`Session.query() <sqlalchemy.orm.Session.query>` method.\"\"\"\n twophase: bool | EmptyType = Empty\n \"\"\"When ``True``, all transactions will be started as a \u201ctwo phase\u201d transaction, i.e. using the \u201ctwo phase\u201d\n semantics of the database in use along with an XID. During a :attr:`commit() <sqlalchemy.orm.Session.commit>`, after\n :attr:`flush() <sqlalchemy.orm.Session.flush>` has been issued for all attached databases, the\n :attr:`TwoPhaseTransaction.prepare() <sqlalchemy.engine.TwoPhaseTransaction.prepare>` method on each database`s\n :class:`TwoPhaseTransaction <sqlalchemy.engine.TwoPhaseTransaction>` will be called. This allows each database to\n roll back the entire transaction, before each transaction is committed.\"\"\"\n\n\n@dataclass\nclass GenericSQLAlchemyConfig(Generic[EngineT, SessionT, SessionMakerT]):\n \"\"\"Common SQLAlchemy Configuration.\"\"\"\n\n create_engine_callable: Callable[[str], EngineT]\n \"\"\"Callable that creates an :class:`AsyncEngine <sqlalchemy.ext.asyncio.AsyncEngine>` instance or instance of its\n subclass.\n \"\"\"\n session_config: GenericSessionConfig\n \"\"\"Configuration options for either the :class:`async_sessionmaker <sqlalchemy.ext.asyncio.async_sessionmaker>`\n or :class:`sessionmaker <sqlalchemy.orm.sessionmaker>`.\n \"\"\"\n session_maker_class: type[sessionmaker] | type[async_sessionmaker]\n \"\"\"Sessionmaker class to use.\"\"\"\n before_send_handler: BeforeMessageSendHookHandler\n \"\"\"Handler to call before the ASGI message is sent.\n\n The handler should handle closing the session stored in the ASGI scope, if its still open, and committing and\n uncommitted data.\n \"\"\"\n connection_string: str | None = field(default=None)\n \"\"\"Database connection string in one of the formats supported by SQLAlchemy.\n\n Notes:\n - For async connections, the connection string must include the correct async prefix.\n e.g. ``'postgresql+asyncpg://...'`` instead of ``'postgresql://'``, and for sync connections its the opposite.\n\n \"\"\"\n engine_dependency_key: str = \"db_engine\"\n \"\"\"Key to use for the dependency injection of database engines.\"\"\"\n session_dependency_key: str = \"db_session\"\n \"\"\"Key to use for the dependency injection of database sessions.\"\"\"\n engine_app_state_key: str = \"db_engine\"\n \"\"\"Key under which to store the SQLAlchemy engine in the application :class:`State <.datastructures.State>`\n instance.\n \"\"\"\n engine_config: EngineConfig = field(default_factory=EngineConfig)\n \"\"\"Configuration for the SQLAlchemy engine.\n\n The configuration options are documented in the SQLAlchemy documentation.\n \"\"\"\n session_maker_app_state_key: str = \"session_maker_class\"\n \"\"\"Key under which to store the SQLAlchemy :class:`sessionmaker <sqlalchemy.orm.sessionmaker>` in the application\n :class:`State <.datastructures.State>` instance.\n \"\"\"\n session_maker: Callable[[], SessionT] | None = None\n \"\"\"Callable that returns a session.\n\n If provided, the plugin will use this rather than instantiate a sessionmaker.\n \"\"\"\n engine_instance: EngineT | None = None\n \"\"\"Optional engine to use.\n\n If set, the plugin will use the provided instance rather than instantiate an engine.\n \"\"\"\n\n def __post_init__(self) -> None:\n if self.connection_string is not None and self.engine_instance is not None:\n raise ImproperlyConfiguredException(\"Only one of 'connection_string' or 'engine_instance' can be provided.\")\n\n @property\n def engine_config_dict(self) -> dict[str, Any]:\n \"\"\"Return the engine configuration as a dict.\n\n Returns:\n A string keyed dict of config kwargs for the SQLAlchemy :func:`create_engine <sqlalchemy.create_engine>`\n function.\n \"\"\"\n return simple_asdict(self.engine_config, exclude_empty=True)\n\n @property\n def session_config_dict(self) -> dict[str, Any]:\n \"\"\"Return the session configuration as a dict.\n\n Returns:\n A string keyed dict of config kwargs for the SQLAlchemy :class:`sessionmaker <sqlalchemy.orm.sessionmaker>`\n class.\n \"\"\"\n return simple_asdict(self.session_config, exclude_empty=True)\n\n @property\n def signature_namespace(self) -> dict[str, Any]:\n \"\"\"Return the plugin's signature namespace.\n\n Returns:\n A string keyed dict of names to be added to the namespace for signature forward reference resolution.\n \"\"\"\n return {} # pragma: no cover\n\n def create_engine(self) -> EngineT:\n \"\"\"Return an engine. If none exists yet, create one.\n\n Returns:\n Getter that returns the engine instance used by the plugin.\n \"\"\"\n if self.engine_instance:\n return self.engine_instance\n\n if self.connection_string is None:\n raise ImproperlyConfiguredException(\"One of 'connection_string' or 'engine_instance' must be provided.\")\n\n engine_config = self.engine_config_dict\n try:\n return self.create_engine_callable(self.connection_string, **engine_config)\n except ValueError:\n # likely due to a dialect that doesn't support json type\n del engine_config[\"json_deserializer\"]\n del engine_config[\"json_serializer\"]\n return self.create_engine_callable(self.connection_string, **engine_config)\n\n def create_session_maker(self) -> Callable[[], SessionT]:\n \"\"\"Get a session maker. If none exists yet, create one.\n\n Returns:\n Session factory used by the plugin.\n \"\"\"\n if self.session_maker:\n return self.session_maker\n\n session_kws = self.session_config_dict\n if session_kws.get(\"bind\") is None:\n session_kws[\"bind\"] = self.create_engine()\n return self.session_maker_class(**session_kws)\n\n def provide_engine(self, state: State) -> EngineT:\n \"\"\"Create an engine instance.\n\n Args:\n state: The ``Starlite.state`` instance.\n\n Returns:\n An engine instance.\n \"\"\"\n return cast(\"EngineT\", state.get(self.engine_app_state_key))\n\n def provide_session(self, state: State, scope: Scope) -> SessionT:\n \"\"\"Create a session instance.\n\n Args:\n state: The ``Starlite.state`` instance.\n scope: The current connection's scope.\n\n Returns:\n A session instance.\n \"\"\"\n session = cast(\"SessionT | None\", get_starlite_scope_state(scope, SESSION_SCOPE_KEY))\n if session is None:\n session_maker = cast(\"Callable[[], SessionT]\", state[self.session_maker_app_state_key])\n session = session_maker()\n set_starlite_scope_state(scope, SESSION_SCOPE_KEY, session)\n return session\n\n def create_app_state_items(self) -> dict[str, Any]:\n \"\"\"Key/value pairs to be stored in application state.\"\"\"\n return {\n self.engine_app_state_key: self.create_engine(),\n self.session_maker_app_state_key: self.create_session_maker(),\n }\n", "path": "starlite/contrib/sqlalchemy/init_plugin/config/common.py"}, {"content": "from __future__ import annotations\n\nfrom typing import TYPE_CHECKING\n\nfrom starlite.di import Provide\nfrom starlite.plugins import InitPluginProtocol\n\nif TYPE_CHECKING:\n from starlite.config.app import AppConfig\n\n from .config import SQLAlchemyAsyncConfig, SQLAlchemySyncConfig\n\n__all__ = (\"SQLAlchemyInitPlugin\",)\n\n\nclass SQLAlchemyInitPlugin(InitPluginProtocol):\n \"\"\"SQLAlchemy application lifecycle configuration.\"\"\"\n\n __slots__ = (\"_config\",)\n\n def __init__(self, config: SQLAlchemyAsyncConfig | SQLAlchemySyncConfig) -> None:\n \"\"\"Initialize ``SQLAlchemyPlugin``.\n\n Args:\n config: configure DB connection and hook handlers and dependencies.\n \"\"\"\n self._config = config\n\n def on_app_init(self, app_config: AppConfig) -> AppConfig:\n \"\"\"Configure application for use with SQLAlchemy.\n\n Args:\n app_config: The :class:`AppConfig <.config.app.AppConfig>` instance.\n \"\"\"\n app_config.dependencies.update(\n {\n self._config.engine_dependency_key: Provide(self._config.provide_engine),\n self._config.session_dependency_key: Provide(self._config.provide_session),\n }\n )\n app_config.before_send.append(self._config.before_send_handler)\n app_config.on_shutdown.append(self._config.on_shutdown)\n app_config.state.update(self._config.create_app_state_items())\n app_config.signature_namespace.update(self._config.signature_namespace)\n return app_config\n", "path": "starlite/contrib/sqlalchemy/init_plugin/plugin.py"}], "after_files": [{"content": "from __future__ import annotations\n\nfrom dataclasses import dataclass, field\nfrom typing import TYPE_CHECKING, Callable, Generic, TypeVar, cast\n\nfrom starlite.constants import HTTP_DISCONNECT, HTTP_RESPONSE_START, WEBSOCKET_CLOSE, WEBSOCKET_DISCONNECT\nfrom starlite.exceptions import ImproperlyConfiguredException\nfrom starlite.types import Empty\nfrom starlite.utils import get_starlite_scope_state, set_starlite_scope_state\nfrom starlite.utils.dataclass import simple_asdict\n\nfrom .engine import EngineConfig\n\nif TYPE_CHECKING:\n from typing import Any\n\n from sqlalchemy import Connection, Engine\n from sqlalchemy.ext.asyncio import AsyncConnection, AsyncEngine, AsyncSession, async_sessionmaker\n from sqlalchemy.orm import Mapper, Query, Session, sessionmaker\n from sqlalchemy.orm.session import JoinTransactionMode\n from sqlalchemy.sql import TableClause\n\n from starlite.datastructures.state import State\n from starlite.types import BeforeMessageSendHookHandler, EmptyType, Scope\n\n__all__ = (\n \"SESSION_SCOPE_KEY\",\n \"SESSION_TERMINUS_ASGI_EVENTS\",\n \"GenericSQLAlchemyConfig\",\n \"GenericSessionConfig\",\n)\n\nSESSION_SCOPE_KEY = \"_sqlalchemy_db_session\"\nSESSION_TERMINUS_ASGI_EVENTS = {HTTP_RESPONSE_START, HTTP_DISCONNECT, WEBSOCKET_DISCONNECT, WEBSOCKET_CLOSE}\n\nConnectionT = TypeVar(\"ConnectionT\", bound=\"Connection | AsyncConnection\")\nEngineT = TypeVar(\"EngineT\", bound=\"Engine | AsyncEngine\")\nSessionT = TypeVar(\"SessionT\", bound=\"Session | AsyncSession\")\nSessionMakerT = TypeVar(\"SessionMakerT\", bound=\"sessionmaker | async_sessionmaker\")\n\n\n@dataclass\nclass GenericSessionConfig(Generic[ConnectionT, EngineT, SessionT]):\n \"\"\"SQLAlchemy async session config.\"\"\"\n\n autobegin: bool | EmptyType = Empty\n \"\"\"Automatically start transactions when database access is requested by an operation.\"\"\"\n autoflush: bool | EmptyType = Empty\n \"\"\"When ``True``, all query operations will issue a flush call to this :class:`Session <sqlalchemy.orm.Session>`\n before proceeding\"\"\"\n bind: EngineT | ConnectionT | None | EmptyType = Empty\n \"\"\"The :class:`Engine <sqlalchemy.engine.Engine>` or :class:`Connection <sqlalchemy.engine.Connection>` that new\n :class:`Session <sqlalchemy.orm.Session>` objects will be bound to.\"\"\"\n binds: dict[type[Any] | Mapper[Any] | TableClause | str, EngineT | ConnectionT] | None | EmptyType = Empty\n \"\"\"A dictionary which may specify any number of :class:`Engine <sqlalchemy.engine.Engine>` or :class:`Connection\n <sqlalchemy.engine.Connection>` objects as the source of connectivity for SQL operations on a per-entity basis. The\n keys of the dictionary consist of any series of mapped classes, arbitrary Python classes that are bases for mapped\n classes, :class:`Table <sqlalchemy.schema.Table>` objects and :class:`Mapper <sqlalchemy.orm.Mapper>` objects. The\n values of the dictionary are then instances of :class:`Engine <sqlalchemy.engine.Engine>` or less commonly\n :class:`Connection <sqlalchemy.engine.Connection>` objects.\"\"\"\n class_: type[SessionT] | EmptyType = Empty\n \"\"\"Class to use in order to create new :class:`Session <sqlalchemy.orm.Session>` objects.\"\"\"\n expire_on_commit: bool | EmptyType = Empty\n \"\"\"If ``True``, all instances will be expired after each commit.\"\"\"\n info: dict[str, Any] | None | EmptyType = Empty\n \"\"\"Optional dictionary of information that will be available via the\n :attr:`Session.info <sqlalchemy.orm.Session.info>`\"\"\"\n join_transaction_mode: JoinTransactionMode | EmptyType = Empty\n \"\"\"Describes the transactional behavior to take when a given bind is a Connection that has already begun a\n transaction outside the scope of this Session; in other words the\n :attr:`Connection.in_transaction() <sqlalchemy.Connection.in_transaction>` method returns True.\"\"\"\n query_cls: type[Query] | None | EmptyType = Empty\n \"\"\"Class which should be used to create new Query objects, as returned by the\n :attr:`Session.query() <sqlalchemy.orm.Session.query>` method.\"\"\"\n twophase: bool | EmptyType = Empty\n \"\"\"When ``True``, all transactions will be started as a \u201ctwo phase\u201d transaction, i.e. using the \u201ctwo phase\u201d\n semantics of the database in use along with an XID. During a :attr:`commit() <sqlalchemy.orm.Session.commit>`, after\n :attr:`flush() <sqlalchemy.orm.Session.flush>` has been issued for all attached databases, the\n :attr:`TwoPhaseTransaction.prepare() <sqlalchemy.engine.TwoPhaseTransaction.prepare>` method on each database`s\n :class:`TwoPhaseTransaction <sqlalchemy.engine.TwoPhaseTransaction>` will be called. This allows each database to\n roll back the entire transaction, before each transaction is committed.\"\"\"\n\n\n@dataclass\nclass GenericSQLAlchemyConfig(Generic[EngineT, SessionT, SessionMakerT]):\n \"\"\"Common SQLAlchemy Configuration.\"\"\"\n\n create_engine_callable: Callable[[str], EngineT]\n \"\"\"Callable that creates an :class:`AsyncEngine <sqlalchemy.ext.asyncio.AsyncEngine>` instance or instance of its\n subclass.\n \"\"\"\n session_config: GenericSessionConfig\n \"\"\"Configuration options for either the :class:`async_sessionmaker <sqlalchemy.ext.asyncio.async_sessionmaker>`\n or :class:`sessionmaker <sqlalchemy.orm.sessionmaker>`.\n \"\"\"\n session_maker_class: type[sessionmaker] | type[async_sessionmaker]\n \"\"\"Sessionmaker class to use.\"\"\"\n before_send_handler: BeforeMessageSendHookHandler\n \"\"\"Handler to call before the ASGI message is sent.\n\n The handler should handle closing the session stored in the ASGI scope, if its still open, and committing and\n uncommitted data.\n \"\"\"\n connection_string: str | None = field(default=None)\n \"\"\"Database connection string in one of the formats supported by SQLAlchemy.\n\n Notes:\n - For async connections, the connection string must include the correct async prefix.\n e.g. ``'postgresql+asyncpg://...'`` instead of ``'postgresql://'``, and for sync connections its the opposite.\n\n \"\"\"\n engine_dependency_key: str = \"db_engine\"\n \"\"\"Key to use for the dependency injection of database engines.\"\"\"\n session_dependency_key: str = \"db_session\"\n \"\"\"Key to use for the dependency injection of database sessions.\"\"\"\n engine_app_state_key: str = \"db_engine\"\n \"\"\"Key under which to store the SQLAlchemy engine in the application :class:`State <.datastructures.State>`\n instance.\n \"\"\"\n engine_config: EngineConfig = field(default_factory=EngineConfig)\n \"\"\"Configuration for the SQLAlchemy engine.\n\n The configuration options are documented in the SQLAlchemy documentation.\n \"\"\"\n session_maker_app_state_key: str = \"session_maker_class\"\n \"\"\"Key under which to store the SQLAlchemy :class:`sessionmaker <sqlalchemy.orm.sessionmaker>` in the application\n :class:`State <.datastructures.State>` instance.\n \"\"\"\n session_maker: Callable[[], SessionT] | None = None\n \"\"\"Callable that returns a session.\n\n If provided, the plugin will use this rather than instantiate a sessionmaker.\n \"\"\"\n engine_instance: EngineT | None = None\n \"\"\"Optional engine to use.\n\n If set, the plugin will use the provided instance rather than instantiate an engine.\n \"\"\"\n\n def __post_init__(self) -> None:\n if self.connection_string is not None and self.engine_instance is not None:\n raise ImproperlyConfiguredException(\"Only one of 'connection_string' or 'engine_instance' can be provided.\")\n\n @property\n def engine_config_dict(self) -> dict[str, Any]:\n \"\"\"Return the engine configuration as a dict.\n\n Returns:\n A string keyed dict of config kwargs for the SQLAlchemy :func:`create_engine <sqlalchemy.create_engine>`\n function.\n \"\"\"\n return simple_asdict(self.engine_config, exclude_empty=True)\n\n @property\n def session_config_dict(self) -> dict[str, Any]:\n \"\"\"Return the session configuration as a dict.\n\n Returns:\n A string keyed dict of config kwargs for the SQLAlchemy :class:`sessionmaker <sqlalchemy.orm.sessionmaker>`\n class.\n \"\"\"\n return simple_asdict(self.session_config, exclude_empty=True)\n\n @property\n def signature_namespace(self) -> dict[str, Any]:\n \"\"\"Return the plugin's signature namespace.\n\n Returns:\n A string keyed dict of names to be added to the namespace for signature forward reference resolution.\n \"\"\"\n return {} # pragma: no cover\n\n def create_engine(self) -> EngineT:\n \"\"\"Return an engine. If none exists yet, create one.\n\n Returns:\n Getter that returns the engine instance used by the plugin.\n \"\"\"\n if self.engine_instance:\n return self.engine_instance\n\n if self.connection_string is None:\n raise ImproperlyConfiguredException(\"One of 'connection_string' or 'engine_instance' must be provided.\")\n\n engine_config = self.engine_config_dict\n try:\n return self.create_engine_callable(self.connection_string, **engine_config)\n except ValueError:\n # likely due to a dialect that doesn't support json type\n del engine_config[\"json_deserializer\"]\n del engine_config[\"json_serializer\"]\n return self.create_engine_callable(self.connection_string, **engine_config)\n\n def create_session_maker(self) -> Callable[[], SessionT]:\n \"\"\"Get a session maker. If none exists yet, create one.\n\n Returns:\n Session factory used by the plugin.\n \"\"\"\n if self.session_maker:\n return self.session_maker\n\n session_kws = self.session_config_dict\n if session_kws.get(\"bind\") is None:\n session_kws[\"bind\"] = self.create_engine()\n return self.session_maker_class(**session_kws)\n\n def provide_engine(self, state: State) -> EngineT:\n \"\"\"Create an engine instance.\n\n Args:\n state: The ``Starlite.state`` instance.\n\n Returns:\n An engine instance.\n \"\"\"\n return cast(\"EngineT\", state.get(self.engine_app_state_key))\n\n def provide_session(self, state: State, scope: Scope) -> SessionT:\n \"\"\"Create a session instance.\n\n Args:\n state: The ``Starlite.state`` instance.\n scope: The current connection's scope.\n\n Returns:\n A session instance.\n \"\"\"\n session = cast(\"SessionT | None\", get_starlite_scope_state(scope, SESSION_SCOPE_KEY))\n if session is None:\n session_maker = cast(\"Callable[[], SessionT]\", state[self.session_maker_app_state_key])\n session = session_maker()\n set_starlite_scope_state(scope, SESSION_SCOPE_KEY, session)\n return session\n\n def create_app_state_items(self) -> dict[str, Any]:\n \"\"\"Key/value pairs to be stored in application state.\"\"\"\n return {\n self.engine_app_state_key: self.create_engine(),\n self.session_maker_app_state_key: self.create_session_maker(),\n }\n\n def update_app_state(self, state: State) -> None:\n \"\"\"Set the app state with engine and session.\n\n Args:\n state: The ``Starlite.state`` instance.\n \"\"\"\n state.update(self.create_app_state_items())\n", "path": "starlite/contrib/sqlalchemy/init_plugin/config/common.py"}, {"content": "from __future__ import annotations\n\nfrom typing import TYPE_CHECKING\n\nfrom starlite.di import Provide\nfrom starlite.plugins import InitPluginProtocol\n\nif TYPE_CHECKING:\n from starlite.config.app import AppConfig\n\n from .config import SQLAlchemyAsyncConfig, SQLAlchemySyncConfig\n\n__all__ = (\"SQLAlchemyInitPlugin\",)\n\n\nclass SQLAlchemyInitPlugin(InitPluginProtocol):\n \"\"\"SQLAlchemy application lifecycle configuration.\"\"\"\n\n __slots__ = (\"_config\",)\n\n def __init__(self, config: SQLAlchemyAsyncConfig | SQLAlchemySyncConfig) -> None:\n \"\"\"Initialize ``SQLAlchemyPlugin``.\n\n Args:\n config: configure DB connection and hook handlers and dependencies.\n \"\"\"\n self._config = config\n\n def on_app_init(self, app_config: AppConfig) -> AppConfig:\n \"\"\"Configure application for use with SQLAlchemy.\n\n Args:\n app_config: The :class:`AppConfig <.config.app.AppConfig>` instance.\n \"\"\"\n app_config.dependencies.update(\n {\n self._config.engine_dependency_key: Provide(self._config.provide_engine),\n self._config.session_dependency_key: Provide(self._config.provide_session),\n }\n )\n app_config.before_send.append(self._config.before_send_handler)\n app_config.on_startup.append(self._config.update_app_state)\n app_config.on_shutdown.append(self._config.on_shutdown)\n app_config.signature_namespace.update(self._config.signature_namespace)\n return app_config\n", "path": "starlite/contrib/sqlalchemy/init_plugin/plugin.py"}]} | 3,838 | 293 |
gh_patches_debug_18827 | rasdani/github-patches | git_diff | DataDog__dd-trace-py-1879 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Sanic resource names gets grouped incorrectly
Hi!
The Endpoints gets grouped incorrectly in the UI when using the Sanic integration because the path parameter values are used in the resource name instead of the parameter names and thus creating one unique Endpoint for every unique method + request path.
Example:

Is this by design? Other integrations (node express for example) groups them by the paramater names which imo seems to be the proper way to do it.
I have created a PR to solve this: #1879
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `ddtrace/contrib/sanic/patch.py`
Content:
```
1 import asyncio
2 import ddtrace
3 import sanic
4 from ddtrace import config
5 from ddtrace.constants import ANALYTICS_SAMPLE_RATE_KEY
6 from ddtrace.ext import SpanTypes
7 from ddtrace.propagation.http import HTTPPropagator
8 from ddtrace.utils.wrappers import unwrap as _u
9 from ddtrace.vendor import wrapt
10 from ddtrace.vendor.wrapt import wrap_function_wrapper as _w
11
12 from .. import trace_utils
13 from ...internal.logger import get_logger
14
15 log = get_logger(__name__)
16
17 config._add("sanic", dict(_default_service="sanic", distributed_tracing=True))
18
19
20 def _wrap_response_callback(span, callback):
21 # wrap response callbacks (either sync or async function) to set span tags
22 # based on response and finish span before returning response
23
24 def update_span(response):
25 if isinstance(response, sanic.response.BaseHTTPResponse):
26 status_code = response.status
27 response_headers = response.headers
28 else:
29 # invalid response causes ServerError exception which must be handled
30 status_code = 500
31 response_headers = None
32 trace_utils.set_http_meta(span, config.sanic, status_code=status_code, response_headers=response_headers)
33 span.finish()
34
35 @wrapt.function_wrapper
36 def wrap_sync(wrapped, instance, args, kwargs):
37 r = wrapped(*args, **kwargs)
38 response = args[0]
39 update_span(response)
40 return r
41
42 @wrapt.function_wrapper
43 async def wrap_async(wrapped, instance, args, kwargs):
44 r = await wrapped(*args, **kwargs)
45 response = args[0]
46 update_span(response)
47 return r
48
49 if asyncio.iscoroutinefunction(callback):
50 return wrap_async(callback)
51
52 return wrap_sync(callback)
53
54
55 def patch():
56 """Patch the instrumented methods."""
57 if getattr(sanic, "__datadog_patch", False):
58 return
59 setattr(sanic, "__datadog_patch", True)
60 _w("sanic", "Sanic.handle_request", patch_handle_request)
61
62
63 def unpatch():
64 """Unpatch the instrumented methods."""
65 _u(sanic.Sanic, "handle_request")
66 if not getattr(sanic, "__datadog_patch", False):
67 return
68 setattr(sanic, "__datadog_patch", False)
69
70
71 async def patch_handle_request(wrapped, instance, args, kwargs):
72 """Wrapper for Sanic.handle_request"""
73 request = kwargs.get("request", args[0])
74 write_callback = kwargs.get("write_callback", args[1])
75 stream_callback = kwargs.get("stream_callback", args[2])
76
77 if request.scheme not in ("http", "https"):
78 return await wrapped(request, write_callback, stream_callback, **kwargs)
79
80 resource = "{} {}".format(request.method, request.path)
81
82 headers = request.headers.copy()
83
84 if config.sanic.distributed_tracing:
85 propagator = HTTPPropagator()
86 context = propagator.extract(headers)
87 if context.trace_id:
88 ddtrace.tracer.context_provider.activate(context)
89
90 span = ddtrace.tracer.trace(
91 "sanic.request",
92 service=trace_utils.int_service(None, config.sanic),
93 resource=resource,
94 span_type=SpanTypes.WEB,
95 )
96 sample_rate = config.sanic.get_analytics_sample_rate(use_global_config=True)
97 if sample_rate is not None:
98 span.set_tag(ANALYTICS_SAMPLE_RATE_KEY, sample_rate)
99
100 method = request.method
101 url = "{scheme}://{host}{path}".format(scheme=request.scheme, host=request.host, path=request.path)
102 query_string = request.query_string
103 if isinstance(query_string, bytes):
104 query_string = query_string.decode()
105 trace_utils.set_http_meta(span, config.sanic, method=method, url=url, query=query_string, request_headers=headers)
106
107 if write_callback is not None:
108 write_callback = _wrap_response_callback(span, write_callback)
109 if stream_callback is not None:
110 stream_callback = _wrap_response_callback(span, stream_callback)
111
112 return await wrapped(request, write_callback, stream_callback, **kwargs)
113
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/ddtrace/contrib/sanic/patch.py b/ddtrace/contrib/sanic/patch.py
--- a/ddtrace/contrib/sanic/patch.py
+++ b/ddtrace/contrib/sanic/patch.py
@@ -52,6 +52,18 @@
return wrap_sync(callback)
+def _get_path(request):
+ """Get path and replace path parameter values with names if route exists."""
+ path = request.path
+ try:
+ match_info = request.match_info
+ except sanic.exceptions.SanicException:
+ return path
+ for key, value in match_info.items():
+ path = path.replace(value, f"<{key}>")
+ return path
+
+
def patch():
"""Patch the instrumented methods."""
if getattr(sanic, "__datadog_patch", False):
@@ -77,7 +89,7 @@
if request.scheme not in ("http", "https"):
return await wrapped(request, write_callback, stream_callback, **kwargs)
- resource = "{} {}".format(request.method, request.path)
+ resource = "{} {}".format(request.method, _get_path(request))
headers = request.headers.copy()
| {"golden_diff": "diff --git a/ddtrace/contrib/sanic/patch.py b/ddtrace/contrib/sanic/patch.py\n--- a/ddtrace/contrib/sanic/patch.py\n+++ b/ddtrace/contrib/sanic/patch.py\n@@ -52,6 +52,18 @@\n return wrap_sync(callback)\n \n \n+def _get_path(request):\n+ \"\"\"Get path and replace path parameter values with names if route exists.\"\"\"\n+ path = request.path\n+ try:\n+ match_info = request.match_info\n+ except sanic.exceptions.SanicException:\n+ return path\n+ for key, value in match_info.items():\n+ path = path.replace(value, f\"<{key}>\")\n+ return path\n+\n+\n def patch():\n \"\"\"Patch the instrumented methods.\"\"\"\n if getattr(sanic, \"__datadog_patch\", False):\n@@ -77,7 +89,7 @@\n if request.scheme not in (\"http\", \"https\"):\n return await wrapped(request, write_callback, stream_callback, **kwargs)\n \n- resource = \"{} {}\".format(request.method, request.path)\n+ resource = \"{} {}\".format(request.method, _get_path(request))\n \n headers = request.headers.copy()\n", "issue": "Sanic resource names gets grouped incorrectly\nHi!\r\n\r\nThe Endpoints gets grouped incorrectly in the UI when using the Sanic integration because the path parameter values are used in the resource name instead of the parameter names and thus creating one unique Endpoint for every unique method + request path.\r\n\r\nExample:\r\n\r\n\r\nIs this by design? Other integrations (node express for example) groups them by the paramater names which imo seems to be the proper way to do it.\r\n\r\nI have created a PR to solve this: #1879\n", "before_files": [{"content": "import asyncio\nimport ddtrace\nimport sanic\nfrom ddtrace import config\nfrom ddtrace.constants import ANALYTICS_SAMPLE_RATE_KEY\nfrom ddtrace.ext import SpanTypes\nfrom ddtrace.propagation.http import HTTPPropagator\nfrom ddtrace.utils.wrappers import unwrap as _u\nfrom ddtrace.vendor import wrapt\nfrom ddtrace.vendor.wrapt import wrap_function_wrapper as _w\n\nfrom .. import trace_utils\nfrom ...internal.logger import get_logger\n\nlog = get_logger(__name__)\n\nconfig._add(\"sanic\", dict(_default_service=\"sanic\", distributed_tracing=True))\n\n\ndef _wrap_response_callback(span, callback):\n # wrap response callbacks (either sync or async function) to set span tags\n # based on response and finish span before returning response\n\n def update_span(response):\n if isinstance(response, sanic.response.BaseHTTPResponse):\n status_code = response.status\n response_headers = response.headers\n else:\n # invalid response causes ServerError exception which must be handled\n status_code = 500\n response_headers = None\n trace_utils.set_http_meta(span, config.sanic, status_code=status_code, response_headers=response_headers)\n span.finish()\n\n @wrapt.function_wrapper\n def wrap_sync(wrapped, instance, args, kwargs):\n r = wrapped(*args, **kwargs)\n response = args[0]\n update_span(response)\n return r\n\n @wrapt.function_wrapper\n async def wrap_async(wrapped, instance, args, kwargs):\n r = await wrapped(*args, **kwargs)\n response = args[0]\n update_span(response)\n return r\n\n if asyncio.iscoroutinefunction(callback):\n return wrap_async(callback)\n\n return wrap_sync(callback)\n\n\ndef patch():\n \"\"\"Patch the instrumented methods.\"\"\"\n if getattr(sanic, \"__datadog_patch\", False):\n return\n setattr(sanic, \"__datadog_patch\", True)\n _w(\"sanic\", \"Sanic.handle_request\", patch_handle_request)\n\n\ndef unpatch():\n \"\"\"Unpatch the instrumented methods.\"\"\"\n _u(sanic.Sanic, \"handle_request\")\n if not getattr(sanic, \"__datadog_patch\", False):\n return\n setattr(sanic, \"__datadog_patch\", False)\n\n\nasync def patch_handle_request(wrapped, instance, args, kwargs):\n \"\"\"Wrapper for Sanic.handle_request\"\"\"\n request = kwargs.get(\"request\", args[0])\n write_callback = kwargs.get(\"write_callback\", args[1])\n stream_callback = kwargs.get(\"stream_callback\", args[2])\n\n if request.scheme not in (\"http\", \"https\"):\n return await wrapped(request, write_callback, stream_callback, **kwargs)\n\n resource = \"{} {}\".format(request.method, request.path)\n\n headers = request.headers.copy()\n\n if config.sanic.distributed_tracing:\n propagator = HTTPPropagator()\n context = propagator.extract(headers)\n if context.trace_id:\n ddtrace.tracer.context_provider.activate(context)\n\n span = ddtrace.tracer.trace(\n \"sanic.request\",\n service=trace_utils.int_service(None, config.sanic),\n resource=resource,\n span_type=SpanTypes.WEB,\n )\n sample_rate = config.sanic.get_analytics_sample_rate(use_global_config=True)\n if sample_rate is not None:\n span.set_tag(ANALYTICS_SAMPLE_RATE_KEY, sample_rate)\n\n method = request.method\n url = \"{scheme}://{host}{path}\".format(scheme=request.scheme, host=request.host, path=request.path)\n query_string = request.query_string\n if isinstance(query_string, bytes):\n query_string = query_string.decode()\n trace_utils.set_http_meta(span, config.sanic, method=method, url=url, query=query_string, request_headers=headers)\n\n if write_callback is not None:\n write_callback = _wrap_response_callback(span, write_callback)\n if stream_callback is not None:\n stream_callback = _wrap_response_callback(span, stream_callback)\n\n return await wrapped(request, write_callback, stream_callback, **kwargs)\n", "path": "ddtrace/contrib/sanic/patch.py"}], "after_files": [{"content": "import asyncio\nimport ddtrace\nimport sanic\nfrom ddtrace import config\nfrom ddtrace.constants import ANALYTICS_SAMPLE_RATE_KEY\nfrom ddtrace.ext import SpanTypes\nfrom ddtrace.propagation.http import HTTPPropagator\nfrom ddtrace.utils.wrappers import unwrap as _u\nfrom ddtrace.vendor import wrapt\nfrom ddtrace.vendor.wrapt import wrap_function_wrapper as _w\n\nfrom .. import trace_utils\nfrom ...internal.logger import get_logger\n\nlog = get_logger(__name__)\n\nconfig._add(\"sanic\", dict(_default_service=\"sanic\", distributed_tracing=True))\n\n\ndef _wrap_response_callback(span, callback):\n # wrap response callbacks (either sync or async function) to set span tags\n # based on response and finish span before returning response\n\n def update_span(response):\n if isinstance(response, sanic.response.BaseHTTPResponse):\n status_code = response.status\n response_headers = response.headers\n else:\n # invalid response causes ServerError exception which must be handled\n status_code = 500\n response_headers = None\n trace_utils.set_http_meta(span, config.sanic, status_code=status_code, response_headers=response_headers)\n span.finish()\n\n @wrapt.function_wrapper\n def wrap_sync(wrapped, instance, args, kwargs):\n r = wrapped(*args, **kwargs)\n response = args[0]\n update_span(response)\n return r\n\n @wrapt.function_wrapper\n async def wrap_async(wrapped, instance, args, kwargs):\n r = await wrapped(*args, **kwargs)\n response = args[0]\n update_span(response)\n return r\n\n if asyncio.iscoroutinefunction(callback):\n return wrap_async(callback)\n\n return wrap_sync(callback)\n\n\ndef _get_path(request):\n \"\"\"Get path and replace path parameter values with names if route exists.\"\"\"\n path = request.path\n try:\n match_info = request.match_info\n except sanic.exceptions.SanicException:\n return path\n for key, value in match_info.items():\n path = path.replace(value, f\"<{key}>\")\n return path\n\n\ndef patch():\n \"\"\"Patch the instrumented methods.\"\"\"\n if getattr(sanic, \"__datadog_patch\", False):\n return\n setattr(sanic, \"__datadog_patch\", True)\n _w(\"sanic\", \"Sanic.handle_request\", patch_handle_request)\n\n\ndef unpatch():\n \"\"\"Unpatch the instrumented methods.\"\"\"\n _u(sanic.Sanic, \"handle_request\")\n if not getattr(sanic, \"__datadog_patch\", False):\n return\n setattr(sanic, \"__datadog_patch\", False)\n\n\nasync def patch_handle_request(wrapped, instance, args, kwargs):\n \"\"\"Wrapper for Sanic.handle_request\"\"\"\n request = kwargs.get(\"request\", args[0])\n write_callback = kwargs.get(\"write_callback\", args[1])\n stream_callback = kwargs.get(\"stream_callback\", args[2])\n\n if request.scheme not in (\"http\", \"https\"):\n return await wrapped(request, write_callback, stream_callback, **kwargs)\n\n resource = \"{} {}\".format(request.method, _get_path(request))\n\n headers = request.headers.copy()\n\n if config.sanic.distributed_tracing:\n propagator = HTTPPropagator()\n context = propagator.extract(headers)\n if context.trace_id:\n ddtrace.tracer.context_provider.activate(context)\n\n span = ddtrace.tracer.trace(\n \"sanic.request\",\n service=trace_utils.int_service(None, config.sanic),\n resource=resource,\n span_type=SpanTypes.WEB,\n )\n sample_rate = config.sanic.get_analytics_sample_rate(use_global_config=True)\n if sample_rate is not None:\n span.set_tag(ANALYTICS_SAMPLE_RATE_KEY, sample_rate)\n\n method = request.method\n url = \"{scheme}://{host}{path}\".format(scheme=request.scheme, host=request.host, path=request.path)\n query_string = request.query_string\n if isinstance(query_string, bytes):\n query_string = query_string.decode()\n trace_utils.set_http_meta(span, config.sanic, method=method, url=url, query=query_string, request_headers=headers)\n\n if write_callback is not None:\n write_callback = _wrap_response_callback(span, write_callback)\n if stream_callback is not None:\n stream_callback = _wrap_response_callback(span, stream_callback)\n\n return await wrapped(request, write_callback, stream_callback, **kwargs)\n", "path": "ddtrace/contrib/sanic/patch.py"}]} | 1,540 | 259 |
gh_patches_debug_32673 | rasdani/github-patches | git_diff | coreproject-moe__CoreProject-Monorepo-19 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Add `djangorestframework-simplejwt` and add Django endpoints. ( Easiest part NGL )
Gonna leave it as is till i finish other stuff
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `backend/core/urls.py`
Content:
```
1 """core URL Configuration
2
3 The `urlpatterns` list routes URLs to views. For more information please see:
4 https://docs.djangoproject.com/en/3.2/topics/http/urls/
5 Examples:
6 Function views
7 1. Add an import: from my_app import views
8 2. Add a URL to urlpatterns: path('', views.home, name='home')
9 Class-based views
10 1. Add an import: from other_app.views import Home
11 2. Add a URL to urlpatterns: path('', Home.as_view(), name='home')
12 Including another URLconf
13 1. Import the include() function: from django.urls import include, path
14 2. Add a URL to urlpatterns: path('blog/', include('blog.urls'))
15 """
16 from django.contrib import admin
17 from django.urls import path
18 from django.urls import include
19 from django.conf.urls.static import static
20 from django.conf import settings
21
22 urlpatterns = [
23 path("admin/", admin.site.urls),
24 # Pages
25 path("user/", include("pages.users.urls")),
26 path("authentication/", include("pages.authentication.urls")),
27 # Api
28 path("api/v1/avatar/", include("api.v1.avatar.urls")),
29 # Rest endpoints
30 path("api/v1/users/", include("api.v1._user.urls")),
31 ]
32 if settings.DEBUG:
33 urlpatterns += static(settings.STATIC_URL, document_root=settings.STATIC_ROOT)
34 urlpatterns += static(settings.MEDIA_URL, document_root=settings.MEDIA_ROOT)
35
```
Path: `backend/core/settings.py`
Content:
```
1 """
2 Django settings for core project.
3
4 Generated by 'django-admin startproject' using Django 3.2.7.
5
6 For more information on this file, see
7 https://docs.djangoproject.com/en/3.2/topics/settings/
8
9 For the full list of settings and their values, see
10 https://docs.djangoproject.com/en/3.2/ref/settings/
11 """
12
13 from pathlib import Path
14 import os
15
16 # Build paths inside the project like this: BASE_DIR / 'subdir'.
17 BASE_DIR = Path(__file__).resolve().parent.parent
18
19
20 # Quick-start development settings - unsuitable for production
21 # See https://docs.djangoproject.com/en/3.2/howto/deployment/checklist/
22
23 # SECURITY WARNING: keep the secret key used in production secret!
24 SECRET_KEY = "django-insecure-mn19l@e%r^s&a^pa9%(bf173v-0c54^@3s(pb!ts_yuts0$+6p"
25
26 # SECURITY WARNING: don't run with debug turned on in production!
27 DEBUG = True
28
29 ALLOWED_HOSTS = []
30
31
32 # Application definition
33
34 INSTALLED_APPS = [
35 "django.contrib.admin",
36 "django.contrib.auth",
37 "django.contrib.contenttypes",
38 "django.contrib.sessions",
39 "django.contrib.messages",
40 "whitenoise.runserver_nostatic",
41 "django.contrib.staticfiles",
42 # Rest Framework
43 "rest_framework",
44 "rest_framework.authtoken",
45 "corsheaders",
46 # Custom Stuff
47 "custom.user",
48 # Pages
49 "pages.users",
50 "pages.authentication",
51 # Rest stuff
52 "api.v1.avatar",
53 "api.v1._user",
54 ]
55
56 MIDDLEWARE = [
57 "django.middleware.security.SecurityMiddleware",
58 "whitenoise.middleware.WhiteNoiseMiddleware",
59 "django.contrib.sessions.middleware.SessionMiddleware",
60 "corsheaders.middleware.CorsMiddleware",
61 "django.middleware.common.CommonMiddleware",
62 "django.middleware.csrf.CsrfViewMiddleware",
63 "django.contrib.auth.middleware.AuthenticationMiddleware",
64 "django.contrib.messages.middleware.MessageMiddleware",
65 "django.middleware.clickjacking.XFrameOptionsMiddleware",
66 ]
67
68 ROOT_URLCONF = "core.urls"
69
70 TEMPLATES = [
71 {
72 "BACKEND": "django.template.backends.django.DjangoTemplates",
73 "DIRS": [BASE_DIR / "templates"],
74 "APP_DIRS": True,
75 "OPTIONS": {
76 "context_processors": [
77 "django.template.context_processors.debug",
78 "django.template.context_processors.request",
79 "django.contrib.auth.context_processors.auth",
80 "django.contrib.messages.context_processors.messages",
81 ],
82 },
83 },
84 ]
85
86 WSGI_APPLICATION = "core.wsgi.application"
87
88
89 # Database
90 # https://docs.djangoproject.com/en/3.2/ref/settings/#databases
91
92 DATABASES = {
93 "default": {
94 "ENGINE": "django.db.backends.sqlite3",
95 "NAME": BASE_DIR / "db.sqlite3",
96 }
97 }
98
99
100 # Password validation
101 # https://docs.djangoproject.com/en/3.2/ref/settings/#auth-password-validators
102
103 AUTH_PASSWORD_VALIDATORS = [
104 {
105 "NAME": "django.contrib.auth.password_validation.UserAttributeSimilarityValidator",
106 },
107 {
108 "NAME": "django.contrib.auth.password_validation.MinimumLengthValidator",
109 },
110 {
111 "NAME": "django.contrib.auth.password_validation.CommonPasswordValidator",
112 },
113 {
114 "NAME": "django.contrib.auth.password_validation.NumericPasswordValidator",
115 },
116 ]
117
118 # Custom user model
119 # https://testdriven.io/blog/django-custom-user-model/
120
121 AUTH_USER_MODEL = "user.CustomUser"
122
123 # Password hashers
124 # https://docs.djangoproject.com/en/3.2/topics/auth/passwords/#using-argon2-with-django
125
126 PASSWORD_HASHERS = [
127 "django.contrib.auth.hashers.Argon2PasswordHasher",
128 "django.contrib.auth.hashers.PBKDF2PasswordHasher",
129 "django.contrib.auth.hashers.PBKDF2SHA1PasswordHasher",
130 "django.contrib.auth.hashers.BCryptSHA256PasswordHasher",
131 ]
132
133 # Internationalization
134 # https://docs.djangoproject.com/en/3.2/topics/i18n/
135
136 LANGUAGE_CODE = "en-us"
137
138 TIME_ZONE = "UTC"
139
140 USE_I18N = True
141
142 USE_L10N = True
143
144 USE_TZ = True
145
146
147 # Static files (CSS, JavaScript, Images)
148 # https://docs.djangoproject.com/en/3.2/howto/static-files/
149 STATIC_URL = "/static/"
150
151 STATICFILES_DIRS = [
152 os.path.join(BASE_DIR, "static"),
153 ]
154
155 STATIC_ROOT = os.path.join(BASE_DIR, "staticfiles")
156 STATICFILES_STORAGE = "whitenoise.storage.CompressedManifestStaticFilesStorage"
157
158 MEDIA_URL = "/media/"
159 MEDIA_ROOT = os.path.join(BASE_DIR, "media")
160 # Default primary key field type
161 # https://docs.djangoproject.com/en/3.2/ref/settings/#default-auto-field
162
163 DEFAULT_AUTO_FIELD = "django.db.models.BigAutoField"
164
165 # Rest framework auth
166 # https://django-rest-framework-simplejwt.readthedocs.io/en/latest/getting_started.html#installation
167 REST_FRAMEWORK = {
168 "DEFAULT_AUTHENTICATION_CLASSES": (
169 "rest_framework.authentication.BasicAuthentication",
170 "rest_framework.authentication.SessionAuthentication",
171 )
172 }
173
174 # Override the login url
175 # https://stackoverflow.com/questions/49532708/custom-login-url-in-django#49532812
176 LOGIN_URL = "login_page"
177
178
179 CORS_ALLOWED_ORIGINS = [
180 "http://127.0.0.1:3000",
181 "http://127.0.0.1:8000",
182 "http://localhost:3000",
183 "http://localhost:8000",
184 ]
185
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/backend/core/settings.py b/backend/core/settings.py
--- a/backend/core/settings.py
+++ b/backend/core/settings.py
@@ -41,7 +41,8 @@
"django.contrib.staticfiles",
# Rest Framework
"rest_framework",
- "rest_framework.authtoken",
+ "rest_framework_simplejwt.token_blacklist",
+ # 3rd party stuff
"corsheaders",
# Custom Stuff
"custom.user",
@@ -166,6 +167,7 @@
# https://django-rest-framework-simplejwt.readthedocs.io/en/latest/getting_started.html#installation
REST_FRAMEWORK = {
"DEFAULT_AUTHENTICATION_CLASSES": (
+ "rest_framework_simplejwt.authentication.JWTAuthentication",
"rest_framework.authentication.BasicAuthentication",
"rest_framework.authentication.SessionAuthentication",
)
diff --git a/backend/core/urls.py b/backend/core/urls.py
--- a/backend/core/urls.py
+++ b/backend/core/urls.py
@@ -19,6 +19,12 @@
from django.conf.urls.static import static
from django.conf import settings
+from rest_framework_simplejwt.views import (
+ TokenObtainPairView,
+ TokenRefreshView,
+ TokenBlacklistView,
+)
+
urlpatterns = [
path("admin/", admin.site.urls),
# Pages
@@ -26,6 +32,12 @@
path("authentication/", include("pages.authentication.urls")),
# Api
path("api/v1/avatar/", include("api.v1.avatar.urls")),
+ # https://django-rest-framework-simplejwt.readthedocs.io/en/latest/getting_started.html#installation
+ path("api/v1/token/", TokenObtainPairView.as_view(), name="token_obtain_pair"),
+ path("api/v1/token/refresh/", TokenRefreshView.as_view(), name="token_refresh"),
+ path(
+ "api/v1/token/blacklist/", TokenBlacklistView.as_view(), name="token_blacklist"
+ ),
# Rest endpoints
path("api/v1/users/", include("api.v1._user.urls")),
]
| {"golden_diff": "diff --git a/backend/core/settings.py b/backend/core/settings.py\n--- a/backend/core/settings.py\n+++ b/backend/core/settings.py\n@@ -41,7 +41,8 @@\n \"django.contrib.staticfiles\",\n # Rest Framework\n \"rest_framework\",\n- \"rest_framework.authtoken\",\n+ \"rest_framework_simplejwt.token_blacklist\",\n+ # 3rd party stuff\n \"corsheaders\",\n # Custom Stuff\n \"custom.user\",\n@@ -166,6 +167,7 @@\n # https://django-rest-framework-simplejwt.readthedocs.io/en/latest/getting_started.html#installation\n REST_FRAMEWORK = {\n \"DEFAULT_AUTHENTICATION_CLASSES\": (\n+ \"rest_framework_simplejwt.authentication.JWTAuthentication\",\n \"rest_framework.authentication.BasicAuthentication\",\n \"rest_framework.authentication.SessionAuthentication\",\n )\ndiff --git a/backend/core/urls.py b/backend/core/urls.py\n--- a/backend/core/urls.py\n+++ b/backend/core/urls.py\n@@ -19,6 +19,12 @@\n from django.conf.urls.static import static\n from django.conf import settings\n \n+from rest_framework_simplejwt.views import (\n+ TokenObtainPairView,\n+ TokenRefreshView,\n+ TokenBlacklistView,\n+)\n+\n urlpatterns = [\n path(\"admin/\", admin.site.urls),\n # Pages\n@@ -26,6 +32,12 @@\n path(\"authentication/\", include(\"pages.authentication.urls\")),\n # Api\n path(\"api/v1/avatar/\", include(\"api.v1.avatar.urls\")),\n+ # https://django-rest-framework-simplejwt.readthedocs.io/en/latest/getting_started.html#installation\n+ path(\"api/v1/token/\", TokenObtainPairView.as_view(), name=\"token_obtain_pair\"),\n+ path(\"api/v1/token/refresh/\", TokenRefreshView.as_view(), name=\"token_refresh\"),\n+ path(\n+ \"api/v1/token/blacklist/\", TokenBlacklistView.as_view(), name=\"token_blacklist\"\n+ ),\n # Rest endpoints\n path(\"api/v1/users/\", include(\"api.v1._user.urls\")),\n ]\n", "issue": "Add `djangorestframework-simplejwt` and add Django endpoints. ( Easiest part NGL )\nGonna leave it as is till i finish other stuff\r\n\r\n\n", "before_files": [{"content": "\"\"\"core URL Configuration\n\nThe `urlpatterns` list routes URLs to views. For more information please see:\n https://docs.djangoproject.com/en/3.2/topics/http/urls/\nExamples:\nFunction views\n 1. Add an import: from my_app import views\n 2. Add a URL to urlpatterns: path('', views.home, name='home')\nClass-based views\n 1. Add an import: from other_app.views import Home\n 2. Add a URL to urlpatterns: path('', Home.as_view(), name='home')\nIncluding another URLconf\n 1. Import the include() function: from django.urls import include, path\n 2. Add a URL to urlpatterns: path('blog/', include('blog.urls'))\n\"\"\"\nfrom django.contrib import admin\nfrom django.urls import path\nfrom django.urls import include\nfrom django.conf.urls.static import static\nfrom django.conf import settings\n\nurlpatterns = [\n path(\"admin/\", admin.site.urls),\n # Pages\n path(\"user/\", include(\"pages.users.urls\")),\n path(\"authentication/\", include(\"pages.authentication.urls\")),\n # Api\n path(\"api/v1/avatar/\", include(\"api.v1.avatar.urls\")),\n # Rest endpoints\n path(\"api/v1/users/\", include(\"api.v1._user.urls\")),\n]\nif settings.DEBUG:\n urlpatterns += static(settings.STATIC_URL, document_root=settings.STATIC_ROOT)\n urlpatterns += static(settings.MEDIA_URL, document_root=settings.MEDIA_ROOT)\n", "path": "backend/core/urls.py"}, {"content": "\"\"\"\nDjango settings for core project.\n\nGenerated by 'django-admin startproject' using Django 3.2.7.\n\nFor more information on this file, see\nhttps://docs.djangoproject.com/en/3.2/topics/settings/\n\nFor the full list of settings and their values, see\nhttps://docs.djangoproject.com/en/3.2/ref/settings/\n\"\"\"\n\nfrom pathlib import Path\nimport os\n\n# Build paths inside the project like this: BASE_DIR / 'subdir'.\nBASE_DIR = Path(__file__).resolve().parent.parent\n\n\n# Quick-start development settings - unsuitable for production\n# See https://docs.djangoproject.com/en/3.2/howto/deployment/checklist/\n\n# SECURITY WARNING: keep the secret key used in production secret!\nSECRET_KEY = \"django-insecure-mn19l@e%r^s&a^pa9%(bf173v-0c54^@3s(pb!ts_yuts0$+6p\"\n\n# SECURITY WARNING: don't run with debug turned on in production!\nDEBUG = True\n\nALLOWED_HOSTS = []\n\n\n# Application definition\n\nINSTALLED_APPS = [\n \"django.contrib.admin\",\n \"django.contrib.auth\",\n \"django.contrib.contenttypes\",\n \"django.contrib.sessions\",\n \"django.contrib.messages\",\n \"whitenoise.runserver_nostatic\",\n \"django.contrib.staticfiles\",\n # Rest Framework\n \"rest_framework\",\n \"rest_framework.authtoken\",\n \"corsheaders\",\n # Custom Stuff\n \"custom.user\",\n # Pages\n \"pages.users\",\n \"pages.authentication\",\n # Rest stuff\n \"api.v1.avatar\",\n \"api.v1._user\",\n]\n\nMIDDLEWARE = [\n \"django.middleware.security.SecurityMiddleware\",\n \"whitenoise.middleware.WhiteNoiseMiddleware\",\n \"django.contrib.sessions.middleware.SessionMiddleware\",\n \"corsheaders.middleware.CorsMiddleware\",\n \"django.middleware.common.CommonMiddleware\",\n \"django.middleware.csrf.CsrfViewMiddleware\",\n \"django.contrib.auth.middleware.AuthenticationMiddleware\",\n \"django.contrib.messages.middleware.MessageMiddleware\",\n \"django.middleware.clickjacking.XFrameOptionsMiddleware\",\n]\n\nROOT_URLCONF = \"core.urls\"\n\nTEMPLATES = [\n {\n \"BACKEND\": \"django.template.backends.django.DjangoTemplates\",\n \"DIRS\": [BASE_DIR / \"templates\"],\n \"APP_DIRS\": True,\n \"OPTIONS\": {\n \"context_processors\": [\n \"django.template.context_processors.debug\",\n \"django.template.context_processors.request\",\n \"django.contrib.auth.context_processors.auth\",\n \"django.contrib.messages.context_processors.messages\",\n ],\n },\n },\n]\n\nWSGI_APPLICATION = \"core.wsgi.application\"\n\n\n# Database\n# https://docs.djangoproject.com/en/3.2/ref/settings/#databases\n\nDATABASES = {\n \"default\": {\n \"ENGINE\": \"django.db.backends.sqlite3\",\n \"NAME\": BASE_DIR / \"db.sqlite3\",\n }\n}\n\n\n# Password validation\n# https://docs.djangoproject.com/en/3.2/ref/settings/#auth-password-validators\n\nAUTH_PASSWORD_VALIDATORS = [\n {\n \"NAME\": \"django.contrib.auth.password_validation.UserAttributeSimilarityValidator\",\n },\n {\n \"NAME\": \"django.contrib.auth.password_validation.MinimumLengthValidator\",\n },\n {\n \"NAME\": \"django.contrib.auth.password_validation.CommonPasswordValidator\",\n },\n {\n \"NAME\": \"django.contrib.auth.password_validation.NumericPasswordValidator\",\n },\n]\n\n# Custom user model\n# https://testdriven.io/blog/django-custom-user-model/\n\nAUTH_USER_MODEL = \"user.CustomUser\"\n\n# Password hashers\n# https://docs.djangoproject.com/en/3.2/topics/auth/passwords/#using-argon2-with-django\n\nPASSWORD_HASHERS = [\n \"django.contrib.auth.hashers.Argon2PasswordHasher\",\n \"django.contrib.auth.hashers.PBKDF2PasswordHasher\",\n \"django.contrib.auth.hashers.PBKDF2SHA1PasswordHasher\",\n \"django.contrib.auth.hashers.BCryptSHA256PasswordHasher\",\n]\n\n# Internationalization\n# https://docs.djangoproject.com/en/3.2/topics/i18n/\n\nLANGUAGE_CODE = \"en-us\"\n\nTIME_ZONE = \"UTC\"\n\nUSE_I18N = True\n\nUSE_L10N = True\n\nUSE_TZ = True\n\n\n# Static files (CSS, JavaScript, Images)\n# https://docs.djangoproject.com/en/3.2/howto/static-files/\nSTATIC_URL = \"/static/\"\n\nSTATICFILES_DIRS = [\n os.path.join(BASE_DIR, \"static\"),\n]\n\nSTATIC_ROOT = os.path.join(BASE_DIR, \"staticfiles\")\nSTATICFILES_STORAGE = \"whitenoise.storage.CompressedManifestStaticFilesStorage\"\n\nMEDIA_URL = \"/media/\"\nMEDIA_ROOT = os.path.join(BASE_DIR, \"media\")\n# Default primary key field type\n# https://docs.djangoproject.com/en/3.2/ref/settings/#default-auto-field\n\nDEFAULT_AUTO_FIELD = \"django.db.models.BigAutoField\"\n\n# Rest framework auth\n# https://django-rest-framework-simplejwt.readthedocs.io/en/latest/getting_started.html#installation\nREST_FRAMEWORK = {\n \"DEFAULT_AUTHENTICATION_CLASSES\": (\n \"rest_framework.authentication.BasicAuthentication\",\n \"rest_framework.authentication.SessionAuthentication\",\n )\n}\n\n# Override the login url\n# https://stackoverflow.com/questions/49532708/custom-login-url-in-django#49532812\nLOGIN_URL = \"login_page\"\n\n\nCORS_ALLOWED_ORIGINS = [\n \"http://127.0.0.1:3000\",\n \"http://127.0.0.1:8000\",\n \"http://localhost:3000\",\n \"http://localhost:8000\",\n]\n", "path": "backend/core/settings.py"}], "after_files": [{"content": "\"\"\"core URL Configuration\n\nThe `urlpatterns` list routes URLs to views. For more information please see:\n https://docs.djangoproject.com/en/3.2/topics/http/urls/\nExamples:\nFunction views\n 1. Add an import: from my_app import views\n 2. Add a URL to urlpatterns: path('', views.home, name='home')\nClass-based views\n 1. Add an import: from other_app.views import Home\n 2. Add a URL to urlpatterns: path('', Home.as_view(), name='home')\nIncluding another URLconf\n 1. Import the include() function: from django.urls import include, path\n 2. Add a URL to urlpatterns: path('blog/', include('blog.urls'))\n\"\"\"\nfrom django.contrib import admin\nfrom django.urls import path\nfrom django.urls import include\nfrom django.conf.urls.static import static\nfrom django.conf import settings\n\nfrom rest_framework_simplejwt.views import (\n TokenObtainPairView,\n TokenRefreshView,\n TokenBlacklistView,\n)\n\nurlpatterns = [\n path(\"admin/\", admin.site.urls),\n # Pages\n path(\"user/\", include(\"pages.users.urls\")),\n path(\"authentication/\", include(\"pages.authentication.urls\")),\n # Api\n path(\"api/v1/avatar/\", include(\"api.v1.avatar.urls\")),\n # https://django-rest-framework-simplejwt.readthedocs.io/en/latest/getting_started.html#installation\n path(\"api/v1/token/\", TokenObtainPairView.as_view(), name=\"token_obtain_pair\"),\n path(\"api/v1/token/refresh/\", TokenRefreshView.as_view(), name=\"token_refresh\"),\n path(\n \"api/v1/token/blacklist/\", TokenBlacklistView.as_view(), name=\"token_blacklist\"\n ),\n # Rest endpoints\n path(\"api/v1/users/\", include(\"api.v1._user.urls\")),\n]\nif settings.DEBUG:\n urlpatterns += static(settings.STATIC_URL, document_root=settings.STATIC_ROOT)\n urlpatterns += static(settings.MEDIA_URL, document_root=settings.MEDIA_ROOT)\n", "path": "backend/core/urls.py"}, {"content": "\"\"\"\nDjango settings for core project.\n\nGenerated by 'django-admin startproject' using Django 3.2.7.\n\nFor more information on this file, see\nhttps://docs.djangoproject.com/en/3.2/topics/settings/\n\nFor the full list of settings and their values, see\nhttps://docs.djangoproject.com/en/3.2/ref/settings/\n\"\"\"\n\nfrom pathlib import Path\nimport os\n\n# Build paths inside the project like this: BASE_DIR / 'subdir'.\nBASE_DIR = Path(__file__).resolve().parent.parent\n\n\n# Quick-start development settings - unsuitable for production\n# See https://docs.djangoproject.com/en/3.2/howto/deployment/checklist/\n\n# SECURITY WARNING: keep the secret key used in production secret!\nSECRET_KEY = \"django-insecure-mn19l@e%r^s&a^pa9%(bf173v-0c54^@3s(pb!ts_yuts0$+6p\"\n\n# SECURITY WARNING: don't run with debug turned on in production!\nDEBUG = True\n\nALLOWED_HOSTS = []\n\n\n# Application definition\n\nINSTALLED_APPS = [\n \"django.contrib.admin\",\n \"django.contrib.auth\",\n \"django.contrib.contenttypes\",\n \"django.contrib.sessions\",\n \"django.contrib.messages\",\n \"whitenoise.runserver_nostatic\",\n \"django.contrib.staticfiles\",\n # Rest Framework\n \"rest_framework\",\n \"rest_framework_simplejwt.token_blacklist\",\n # 3rd party stuff\n \"corsheaders\",\n # Custom Stuff\n \"custom.user\",\n # Pages\n \"pages.users\",\n \"pages.authentication\",\n # Rest stuff\n \"api.v1.avatar\",\n \"api.v1._user\",\n]\n\nMIDDLEWARE = [\n \"django.middleware.security.SecurityMiddleware\",\n \"whitenoise.middleware.WhiteNoiseMiddleware\",\n \"django.contrib.sessions.middleware.SessionMiddleware\",\n \"corsheaders.middleware.CorsMiddleware\",\n \"django.middleware.common.CommonMiddleware\",\n \"django.middleware.csrf.CsrfViewMiddleware\",\n \"django.contrib.auth.middleware.AuthenticationMiddleware\",\n \"django.contrib.messages.middleware.MessageMiddleware\",\n \"django.middleware.clickjacking.XFrameOptionsMiddleware\",\n]\n\nROOT_URLCONF = \"core.urls\"\n\nTEMPLATES = [\n {\n \"BACKEND\": \"django.template.backends.django.DjangoTemplates\",\n \"DIRS\": [BASE_DIR / \"templates\"],\n \"APP_DIRS\": True,\n \"OPTIONS\": {\n \"context_processors\": [\n \"django.template.context_processors.debug\",\n \"django.template.context_processors.request\",\n \"django.contrib.auth.context_processors.auth\",\n \"django.contrib.messages.context_processors.messages\",\n ],\n },\n },\n]\n\nWSGI_APPLICATION = \"core.wsgi.application\"\n\n\n# Database\n# https://docs.djangoproject.com/en/3.2/ref/settings/#databases\n\nDATABASES = {\n \"default\": {\n \"ENGINE\": \"django.db.backends.sqlite3\",\n \"NAME\": BASE_DIR / \"db.sqlite3\",\n }\n}\n\n\n# Password validation\n# https://docs.djangoproject.com/en/3.2/ref/settings/#auth-password-validators\n\nAUTH_PASSWORD_VALIDATORS = [\n {\n \"NAME\": \"django.contrib.auth.password_validation.UserAttributeSimilarityValidator\",\n },\n {\n \"NAME\": \"django.contrib.auth.password_validation.MinimumLengthValidator\",\n },\n {\n \"NAME\": \"django.contrib.auth.password_validation.CommonPasswordValidator\",\n },\n {\n \"NAME\": \"django.contrib.auth.password_validation.NumericPasswordValidator\",\n },\n]\n\n# Custom user model\n# https://testdriven.io/blog/django-custom-user-model/\n\nAUTH_USER_MODEL = \"user.CustomUser\"\n\n# Password hashers\n# https://docs.djangoproject.com/en/3.2/topics/auth/passwords/#using-argon2-with-django\n\nPASSWORD_HASHERS = [\n \"django.contrib.auth.hashers.Argon2PasswordHasher\",\n \"django.contrib.auth.hashers.PBKDF2PasswordHasher\",\n \"django.contrib.auth.hashers.PBKDF2SHA1PasswordHasher\",\n \"django.contrib.auth.hashers.BCryptSHA256PasswordHasher\",\n]\n\n# Internationalization\n# https://docs.djangoproject.com/en/3.2/topics/i18n/\n\nLANGUAGE_CODE = \"en-us\"\n\nTIME_ZONE = \"UTC\"\n\nUSE_I18N = True\n\nUSE_L10N = True\n\nUSE_TZ = True\n\n\n# Static files (CSS, JavaScript, Images)\n# https://docs.djangoproject.com/en/3.2/howto/static-files/\nSTATIC_URL = \"/static/\"\n\nSTATICFILES_DIRS = [\n os.path.join(BASE_DIR, \"static\"),\n]\n\nSTATIC_ROOT = os.path.join(BASE_DIR, \"staticfiles\")\nSTATICFILES_STORAGE = \"whitenoise.storage.CompressedManifestStaticFilesStorage\"\n\nMEDIA_URL = \"/media/\"\nMEDIA_ROOT = os.path.join(BASE_DIR, \"media\")\n# Default primary key field type\n# https://docs.djangoproject.com/en/3.2/ref/settings/#default-auto-field\n\nDEFAULT_AUTO_FIELD = \"django.db.models.BigAutoField\"\n\n# Rest framework auth\n# https://django-rest-framework-simplejwt.readthedocs.io/en/latest/getting_started.html#installation\nREST_FRAMEWORK = {\n \"DEFAULT_AUTHENTICATION_CLASSES\": (\n \"rest_framework_simplejwt.authentication.JWTAuthentication\",\n \"rest_framework.authentication.BasicAuthentication\",\n \"rest_framework.authentication.SessionAuthentication\",\n )\n}\n\n# Override the login url\n# https://stackoverflow.com/questions/49532708/custom-login-url-in-django#49532812\nLOGIN_URL = \"login_page\"\n\n\nCORS_ALLOWED_ORIGINS = [\n \"http://127.0.0.1:3000\",\n \"http://127.0.0.1:8000\",\n \"http://localhost:3000\",\n \"http://localhost:8000\",\n]\n", "path": "backend/core/settings.py"}]} | 2,372 | 452 |
gh_patches_debug_28758 | rasdani/github-patches | git_diff | microsoft__botbuilder-python-1402 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
[PORT] Add EndOfConversationCodes to EndOfConversation activity from Skill
> Port this change from botbuilder-dotnet/master branch:
https://github.com/microsoft/botbuilder-dotnet/pull/4235
Fixes https://github.com/microsoft/botframework-sdk/issues/5852
# Changed projects
* Microsoft.Bot.Builder.Dialogs
* Microsoft.Bot.Builder.Dialogs.Tests
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `libraries/botbuilder-dialogs/botbuilder/dialogs/dialog_extensions.py`
Content:
```
1 # Copyright (c) Microsoft Corporation. All rights reserved.
2 # Licensed under the MIT License.
3
4 from botbuilder.core import BotAdapter, StatePropertyAccessor, TurnContext
5 from botbuilder.core.skills import SkillHandler, SkillConversationReference
6
7 from botbuilder.dialogs import (
8 Dialog,
9 DialogEvents,
10 DialogSet,
11 DialogTurnStatus,
12 )
13 from botbuilder.schema import Activity, ActivityTypes
14 from botframework.connector.auth import (
15 ClaimsIdentity,
16 SkillValidation,
17 AuthenticationConstants,
18 GovernmentConstants,
19 )
20
21
22 class DialogExtensions:
23 @staticmethod
24 async def run_dialog(
25 dialog: Dialog, turn_context: TurnContext, accessor: StatePropertyAccessor
26 ):
27 """
28 Creates a dialog stack and starts a dialog, pushing it onto the stack.
29 """
30
31 dialog_set = DialogSet(accessor)
32 dialog_set.add(dialog)
33
34 dialog_context = await dialog_set.create_context(turn_context)
35
36 # Handle EoC and Reprompt event from a parent bot (can be root bot to skill or skill to skill)
37 if DialogExtensions.__is_from_parent_to_skill(turn_context):
38 # Handle remote cancellation request from parent.
39 if turn_context.activity.type == ActivityTypes.end_of_conversation:
40 if not dialog_context.stack:
41 # No dialogs to cancel, just return.
42 return
43
44 remote_cancel_text = "Skill was canceled through an EndOfConversation activity from the parent."
45 await turn_context.send_trace_activity(
46 f"Extension {Dialog.__name__}.run_dialog", label=remote_cancel_text,
47 )
48
49 # Send cancellation message to the dialog to ensure all the parents are canceled
50 # in the right order.
51 await dialog_context.cancel_all_dialogs()
52 return
53
54 # Handle a reprompt event sent from the parent.
55 if (
56 turn_context.activity.type == ActivityTypes.event
57 and turn_context.activity.name == DialogEvents.reprompt_dialog
58 ):
59 if not dialog_context.stack:
60 # No dialogs to reprompt, just return.
61 return
62
63 await dialog_context.reprompt_dialog()
64 return
65
66 # Continue or start the dialog.
67 result = await dialog_context.continue_dialog()
68 if result.status == DialogTurnStatus.Empty:
69 result = await dialog_context.begin_dialog(dialog.id)
70
71 # Skills should send EoC when the dialog completes.
72 if (
73 result.status == DialogTurnStatus.Complete
74 or result.status == DialogTurnStatus.Cancelled
75 ):
76 if DialogExtensions.__send_eoc_to_parent(turn_context):
77 end_message_text = (
78 f"Dialog {dialog.id} has **completed**. Sending EndOfConversation."
79 )
80 await turn_context.send_trace_activity(
81 f"Extension {Dialog.__name__}.run_dialog",
82 label=end_message_text,
83 value=result.result,
84 )
85
86 activity = Activity(
87 type=ActivityTypes.end_of_conversation,
88 value=result.result,
89 locale=turn_context.activity.locale,
90 )
91 await turn_context.send_activity(activity)
92
93 @staticmethod
94 def __is_from_parent_to_skill(turn_context: TurnContext) -> bool:
95 if turn_context.turn_state.get(SkillHandler.SKILL_CONVERSATION_REFERENCE_KEY):
96 return False
97
98 claims_identity = turn_context.turn_state.get(BotAdapter.BOT_IDENTITY_KEY)
99 return isinstance(
100 claims_identity, ClaimsIdentity
101 ) and SkillValidation.is_skill_claim(claims_identity.claims)
102
103 @staticmethod
104 def __send_eoc_to_parent(turn_context: TurnContext) -> bool:
105 claims_identity = turn_context.turn_state.get(BotAdapter.BOT_IDENTITY_KEY)
106 if isinstance(
107 claims_identity, ClaimsIdentity
108 ) and SkillValidation.is_skill_claim(claims_identity.claims):
109 # EoC Activities returned by skills are bounced back to the bot by SkillHandler.
110 # In those cases we will have a SkillConversationReference instance in state.
111 skill_conversation_reference: SkillConversationReference = turn_context.turn_state.get(
112 SkillHandler.SKILL_CONVERSATION_REFERENCE_KEY
113 )
114 if skill_conversation_reference:
115 # If the skillConversationReference.OAuthScope is for one of the supported channels,
116 # we are at the root and we should not send an EoC.
117 return (
118 skill_conversation_reference.oauth_scope
119 != AuthenticationConstants.TO_CHANNEL_FROM_BOT_OAUTH_SCOPE
120 and skill_conversation_reference.oauth_scope
121 != GovernmentConstants.TO_CHANNEL_FROM_BOT_OAUTH_SCOPE
122 )
123 return True
124
125 return False
126
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/libraries/botbuilder-dialogs/botbuilder/dialogs/dialog_extensions.py b/libraries/botbuilder-dialogs/botbuilder/dialogs/dialog_extensions.py
--- a/libraries/botbuilder-dialogs/botbuilder/dialogs/dialog_extensions.py
+++ b/libraries/botbuilder-dialogs/botbuilder/dialogs/dialog_extensions.py
@@ -1,22 +1,21 @@
# Copyright (c) Microsoft Corporation. All rights reserved.
# Licensed under the MIT License.
+from botframework.connector.auth import (
+ ClaimsIdentity,
+ SkillValidation,
+ AuthenticationConstants,
+ GovernmentConstants,
+)
from botbuilder.core import BotAdapter, StatePropertyAccessor, TurnContext
from botbuilder.core.skills import SkillHandler, SkillConversationReference
-
from botbuilder.dialogs import (
Dialog,
DialogEvents,
DialogSet,
DialogTurnStatus,
)
-from botbuilder.schema import Activity, ActivityTypes
-from botframework.connector.auth import (
- ClaimsIdentity,
- SkillValidation,
- AuthenticationConstants,
- GovernmentConstants,
-)
+from botbuilder.schema import Activity, ActivityTypes, EndOfConversationCodes
class DialogExtensions:
@@ -87,6 +86,9 @@
type=ActivityTypes.end_of_conversation,
value=result.result,
locale=turn_context.activity.locale,
+ code=EndOfConversationCodes.completed_successfully
+ if result.status == DialogTurnStatus.Complete
+ else EndOfConversationCodes.user_cancelled,
)
await turn_context.send_activity(activity)
| {"golden_diff": "diff --git a/libraries/botbuilder-dialogs/botbuilder/dialogs/dialog_extensions.py b/libraries/botbuilder-dialogs/botbuilder/dialogs/dialog_extensions.py\n--- a/libraries/botbuilder-dialogs/botbuilder/dialogs/dialog_extensions.py\n+++ b/libraries/botbuilder-dialogs/botbuilder/dialogs/dialog_extensions.py\n@@ -1,22 +1,21 @@\n # Copyright (c) Microsoft Corporation. All rights reserved.\n # Licensed under the MIT License.\n \n+from botframework.connector.auth import (\n+ ClaimsIdentity,\n+ SkillValidation,\n+ AuthenticationConstants,\n+ GovernmentConstants,\n+)\n from botbuilder.core import BotAdapter, StatePropertyAccessor, TurnContext\n from botbuilder.core.skills import SkillHandler, SkillConversationReference\n-\n from botbuilder.dialogs import (\n Dialog,\n DialogEvents,\n DialogSet,\n DialogTurnStatus,\n )\n-from botbuilder.schema import Activity, ActivityTypes\n-from botframework.connector.auth import (\n- ClaimsIdentity,\n- SkillValidation,\n- AuthenticationConstants,\n- GovernmentConstants,\n-)\n+from botbuilder.schema import Activity, ActivityTypes, EndOfConversationCodes\n \n \n class DialogExtensions:\n@@ -87,6 +86,9 @@\n type=ActivityTypes.end_of_conversation,\n value=result.result,\n locale=turn_context.activity.locale,\n+ code=EndOfConversationCodes.completed_successfully\n+ if result.status == DialogTurnStatus.Complete\n+ else EndOfConversationCodes.user_cancelled,\n )\n await turn_context.send_activity(activity)\n", "issue": "[PORT] Add EndOfConversationCodes to EndOfConversation activity from Skill\n> Port this change from botbuilder-dotnet/master branch:\nhttps://github.com/microsoft/botbuilder-dotnet/pull/4235\n\nFixes https://github.com/microsoft/botframework-sdk/issues/5852\n\n\r\n# Changed projects\r\n* Microsoft.Bot.Builder.Dialogs\r\n* Microsoft.Bot.Builder.Dialogs.Tests\r\n\r\n\r\n\n\n", "before_files": [{"content": "# Copyright (c) Microsoft Corporation. All rights reserved.\n# Licensed under the MIT License.\n\nfrom botbuilder.core import BotAdapter, StatePropertyAccessor, TurnContext\nfrom botbuilder.core.skills import SkillHandler, SkillConversationReference\n\nfrom botbuilder.dialogs import (\n Dialog,\n DialogEvents,\n DialogSet,\n DialogTurnStatus,\n)\nfrom botbuilder.schema import Activity, ActivityTypes\nfrom botframework.connector.auth import (\n ClaimsIdentity,\n SkillValidation,\n AuthenticationConstants,\n GovernmentConstants,\n)\n\n\nclass DialogExtensions:\n @staticmethod\n async def run_dialog(\n dialog: Dialog, turn_context: TurnContext, accessor: StatePropertyAccessor\n ):\n \"\"\"\n Creates a dialog stack and starts a dialog, pushing it onto the stack.\n \"\"\"\n\n dialog_set = DialogSet(accessor)\n dialog_set.add(dialog)\n\n dialog_context = await dialog_set.create_context(turn_context)\n\n # Handle EoC and Reprompt event from a parent bot (can be root bot to skill or skill to skill)\n if DialogExtensions.__is_from_parent_to_skill(turn_context):\n # Handle remote cancellation request from parent.\n if turn_context.activity.type == ActivityTypes.end_of_conversation:\n if not dialog_context.stack:\n # No dialogs to cancel, just return.\n return\n\n remote_cancel_text = \"Skill was canceled through an EndOfConversation activity from the parent.\"\n await turn_context.send_trace_activity(\n f\"Extension {Dialog.__name__}.run_dialog\", label=remote_cancel_text,\n )\n\n # Send cancellation message to the dialog to ensure all the parents are canceled\n # in the right order.\n await dialog_context.cancel_all_dialogs()\n return\n\n # Handle a reprompt event sent from the parent.\n if (\n turn_context.activity.type == ActivityTypes.event\n and turn_context.activity.name == DialogEvents.reprompt_dialog\n ):\n if not dialog_context.stack:\n # No dialogs to reprompt, just return.\n return\n\n await dialog_context.reprompt_dialog()\n return\n\n # Continue or start the dialog.\n result = await dialog_context.continue_dialog()\n if result.status == DialogTurnStatus.Empty:\n result = await dialog_context.begin_dialog(dialog.id)\n\n # Skills should send EoC when the dialog completes.\n if (\n result.status == DialogTurnStatus.Complete\n or result.status == DialogTurnStatus.Cancelled\n ):\n if DialogExtensions.__send_eoc_to_parent(turn_context):\n end_message_text = (\n f\"Dialog {dialog.id} has **completed**. Sending EndOfConversation.\"\n )\n await turn_context.send_trace_activity(\n f\"Extension {Dialog.__name__}.run_dialog\",\n label=end_message_text,\n value=result.result,\n )\n\n activity = Activity(\n type=ActivityTypes.end_of_conversation,\n value=result.result,\n locale=turn_context.activity.locale,\n )\n await turn_context.send_activity(activity)\n\n @staticmethod\n def __is_from_parent_to_skill(turn_context: TurnContext) -> bool:\n if turn_context.turn_state.get(SkillHandler.SKILL_CONVERSATION_REFERENCE_KEY):\n return False\n\n claims_identity = turn_context.turn_state.get(BotAdapter.BOT_IDENTITY_KEY)\n return isinstance(\n claims_identity, ClaimsIdentity\n ) and SkillValidation.is_skill_claim(claims_identity.claims)\n\n @staticmethod\n def __send_eoc_to_parent(turn_context: TurnContext) -> bool:\n claims_identity = turn_context.turn_state.get(BotAdapter.BOT_IDENTITY_KEY)\n if isinstance(\n claims_identity, ClaimsIdentity\n ) and SkillValidation.is_skill_claim(claims_identity.claims):\n # EoC Activities returned by skills are bounced back to the bot by SkillHandler.\n # In those cases we will have a SkillConversationReference instance in state.\n skill_conversation_reference: SkillConversationReference = turn_context.turn_state.get(\n SkillHandler.SKILL_CONVERSATION_REFERENCE_KEY\n )\n if skill_conversation_reference:\n # If the skillConversationReference.OAuthScope is for one of the supported channels,\n # we are at the root and we should not send an EoC.\n return (\n skill_conversation_reference.oauth_scope\n != AuthenticationConstants.TO_CHANNEL_FROM_BOT_OAUTH_SCOPE\n and skill_conversation_reference.oauth_scope\n != GovernmentConstants.TO_CHANNEL_FROM_BOT_OAUTH_SCOPE\n )\n return True\n\n return False\n", "path": "libraries/botbuilder-dialogs/botbuilder/dialogs/dialog_extensions.py"}], "after_files": [{"content": "# Copyright (c) Microsoft Corporation. All rights reserved.\n# Licensed under the MIT License.\n\nfrom botframework.connector.auth import (\n ClaimsIdentity,\n SkillValidation,\n AuthenticationConstants,\n GovernmentConstants,\n)\nfrom botbuilder.core import BotAdapter, StatePropertyAccessor, TurnContext\nfrom botbuilder.core.skills import SkillHandler, SkillConversationReference\nfrom botbuilder.dialogs import (\n Dialog,\n DialogEvents,\n DialogSet,\n DialogTurnStatus,\n)\nfrom botbuilder.schema import Activity, ActivityTypes, EndOfConversationCodes\n\n\nclass DialogExtensions:\n @staticmethod\n async def run_dialog(\n dialog: Dialog, turn_context: TurnContext, accessor: StatePropertyAccessor\n ):\n \"\"\"\n Creates a dialog stack and starts a dialog, pushing it onto the stack.\n \"\"\"\n\n dialog_set = DialogSet(accessor)\n dialog_set.add(dialog)\n\n dialog_context = await dialog_set.create_context(turn_context)\n\n # Handle EoC and Reprompt event from a parent bot (can be root bot to skill or skill to skill)\n if DialogExtensions.__is_from_parent_to_skill(turn_context):\n # Handle remote cancellation request from parent.\n if turn_context.activity.type == ActivityTypes.end_of_conversation:\n if not dialog_context.stack:\n # No dialogs to cancel, just return.\n return\n\n remote_cancel_text = \"Skill was canceled through an EndOfConversation activity from the parent.\"\n await turn_context.send_trace_activity(\n f\"Extension {Dialog.__name__}.run_dialog\", label=remote_cancel_text,\n )\n\n # Send cancellation message to the dialog to ensure all the parents are canceled\n # in the right order.\n await dialog_context.cancel_all_dialogs()\n return\n\n # Handle a reprompt event sent from the parent.\n if (\n turn_context.activity.type == ActivityTypes.event\n and turn_context.activity.name == DialogEvents.reprompt_dialog\n ):\n if not dialog_context.stack:\n # No dialogs to reprompt, just return.\n return\n\n await dialog_context.reprompt_dialog()\n return\n\n # Continue or start the dialog.\n result = await dialog_context.continue_dialog()\n if result.status == DialogTurnStatus.Empty:\n result = await dialog_context.begin_dialog(dialog.id)\n\n # Skills should send EoC when the dialog completes.\n if (\n result.status == DialogTurnStatus.Complete\n or result.status == DialogTurnStatus.Cancelled\n ):\n if DialogExtensions.__send_eoc_to_parent(turn_context):\n end_message_text = (\n f\"Dialog {dialog.id} has **completed**. Sending EndOfConversation.\"\n )\n await turn_context.send_trace_activity(\n f\"Extension {Dialog.__name__}.run_dialog\",\n label=end_message_text,\n value=result.result,\n )\n\n activity = Activity(\n type=ActivityTypes.end_of_conversation,\n value=result.result,\n locale=turn_context.activity.locale,\n code=EndOfConversationCodes.completed_successfully\n if result.status == DialogTurnStatus.Complete\n else EndOfConversationCodes.user_cancelled,\n )\n await turn_context.send_activity(activity)\n\n @staticmethod\n def __is_from_parent_to_skill(turn_context: TurnContext) -> bool:\n if turn_context.turn_state.get(SkillHandler.SKILL_CONVERSATION_REFERENCE_KEY):\n return False\n\n claims_identity = turn_context.turn_state.get(BotAdapter.BOT_IDENTITY_KEY)\n return isinstance(\n claims_identity, ClaimsIdentity\n ) and SkillValidation.is_skill_claim(claims_identity.claims)\n\n @staticmethod\n def __send_eoc_to_parent(turn_context: TurnContext) -> bool:\n claims_identity = turn_context.turn_state.get(BotAdapter.BOT_IDENTITY_KEY)\n if isinstance(\n claims_identity, ClaimsIdentity\n ) and SkillValidation.is_skill_claim(claims_identity.claims):\n # EoC Activities returned by skills are bounced back to the bot by SkillHandler.\n # In those cases we will have a SkillConversationReference instance in state.\n skill_conversation_reference: SkillConversationReference = turn_context.turn_state.get(\n SkillHandler.SKILL_CONVERSATION_REFERENCE_KEY\n )\n if skill_conversation_reference:\n # If the skillConversationReference.OAuthScope is for one of the supported channels,\n # we are at the root and we should not send an EoC.\n return (\n skill_conversation_reference.oauth_scope\n != AuthenticationConstants.TO_CHANNEL_FROM_BOT_OAUTH_SCOPE\n and skill_conversation_reference.oauth_scope\n != GovernmentConstants.TO_CHANNEL_FROM_BOT_OAUTH_SCOPE\n )\n return True\n\n return False\n", "path": "libraries/botbuilder-dialogs/botbuilder/dialogs/dialog_extensions.py"}]} | 1,566 | 326 |
gh_patches_debug_9886 | rasdani/github-patches | git_diff | yt-dlp__yt-dlp-8144 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
[media.ccc.de:lists] playlist_id should be case-sensitive
### DO NOT REMOVE OR SKIP THE ISSUE TEMPLATE
- [X] I understand that I will be **blocked** if I *intentionally* remove or skip any mandatory\* field
### Checklist
- [X] I'm reporting that yt-dlp is broken on a **supported** site
- [X] I've verified that I'm running yt-dlp version **2023.07.06** ([update instructions](https://github.com/yt-dlp/yt-dlp#update)) or later (specify commit)
- [X] I've checked that all provided URLs are playable in a browser with the same IP and same login details
- [X] I've checked that all URLs and arguments with special characters are [properly quoted or escaped](https://github.com/yt-dlp/yt-dlp/wiki/FAQ#video-url-contains-an-ampersand--and-im-getting-some-strange-output-1-2839-or-v-is-not-recognized-as-an-internal-or-external-command)
- [X] I've searched [known issues](https://github.com/yt-dlp/yt-dlp/issues/3766) and the [bugtracker](https://github.com/yt-dlp/yt-dlp/issues?q=) for similar issues **including closed ones**. DO NOT post duplicates
- [X] I've read the [guidelines for opening an issue](https://github.com/yt-dlp/yt-dlp/blob/master/CONTRIBUTING.md#opening-an-issue)
- [ ] I've read about [sharing account credentials](https://github.com/yt-dlp/yt-dlp/blob/master/CONTRIBUTING.md#are-you-willing-to-share-account-details-if-needed) and I'm willing to share it if required
### Region
_No response_
### Provide a description that is worded well enough to be understood
Some playlists use uppercase `playlist_id`s, like `https://media.ccc.de/c/DS2023` → `https://media.ccc.de/public/conferences/DS2022` or `https://media.ccc.de/c/MCH2022` → `https://media.ccc.de/public/conferences/MCH2022`. So I guess removing `.lower()` in https://github.com/yt-dlp/yt-dlp/blob/master/yt_dlp/extractor/ccc.py#L96 should resolve this.
### Provide verbose output that clearly demonstrates the problem
- [X] Run **your** yt-dlp command with **-vU** flag added (`yt-dlp -vU <your command line>`)
- [ ] If using API, add `'verbose': True` to `YoutubeDL` params instead
- [X] Copy the WHOLE output (starting with `[debug] Command-line config`) and insert it below
### Complete Verbose Output
```shell
$ python -m yt_dlp --verbose --ignore-config https://media.ccc.de/c/DS2023
[debug] Command-line config: ['--verbose', '--ignore-config', 'https://media.ccc.de/c/DS2023']
[debug] Encodings: locale UTF-8, fs utf-8, pref UTF-8, out utf-8, error utf-8, screen utf-8
[debug] yt-dlp version [email protected] [b532a3481] (source)
[debug] Lazy loading extractors is disabled
[debug] Git HEAD: 30ba233d4
[debug] Python 3.11.5 (CPython x86_64 64bit) - Linux-6.5.0-5-generic-x86_64-with-glibc2.38 (OpenSSL 3.0.10 1 Aug 2023, glibc 2.38)
[debug] exe versions: ffmpeg 6.0 (setts), ffprobe 6.0, rtmpdump 2.4
[debug] Optional libraries: Cryptodome-3.11.0, brotli-1.0.9, certifi-2022.09.24, mutagen-1.46.0, pyxattr-0.8.1, secretstorage-3.3.3, sqlite3-2.6.0, websockets-10.4
[debug] Proxy map: {}
[debug] Loaded 1866 extractors
[media.ccc.de:lists] Extracting URL: https://media.ccc.de/c/DS2023
[media.ccc.de:lists] ds2023: Downloading JSON metadata
ERROR: [media.ccc.de:lists] DS2023: Unable to download JSON metadata: HTTP Error 404: Not Found (caused by <HTTPError 404: Not Found>); please report this issue on https://github.com/yt-dlp/yt-dlp/issues?q= , filling out the appropriate issue template. Confirm you are on the latest version using yt-dlp -U
[…]
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/tmp/yt-dlp/yt_dlp/extractor/common.py", line 847, in _request_webpage
return self._downloader.urlopen(self._create_request(url_or_request, data, headers, query))
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/tmp/yt-dlp/yt_dlp/YoutubeDL.py", line 4078, in urlopen
raise _CompatHTTPError(e) from e
yt_dlp.networking.exceptions._CompatHTTPError: HTTP Error 404: Not Found
```
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `yt_dlp/extractor/ccc.py`
Content:
```
1 from .common import InfoExtractor
2 from ..utils import (
3 int_or_none,
4 parse_iso8601,
5 try_get,
6 url_or_none,
7 )
8
9
10 class CCCIE(InfoExtractor):
11 IE_NAME = 'media.ccc.de'
12 _VALID_URL = r'https?://(?:www\.)?media\.ccc\.de/v/(?P<id>[^/?#&]+)'
13
14 _TESTS = [{
15 'url': 'https://media.ccc.de/v/30C3_-_5443_-_en_-_saal_g_-_201312281830_-_introduction_to_processor_design_-_byterazor#video',
16 'md5': '3a1eda8f3a29515d27f5adb967d7e740',
17 'info_dict': {
18 'id': '1839',
19 'ext': 'mp4',
20 'title': 'Introduction to Processor Design',
21 'creator': 'byterazor',
22 'description': 'md5:df55f6d073d4ceae55aae6f2fd98a0ac',
23 'thumbnail': r're:^https?://.*\.jpg$',
24 'upload_date': '20131228',
25 'timestamp': 1388188800,
26 'duration': 3710,
27 'tags': list,
28 }
29 }, {
30 'url': 'https://media.ccc.de/v/32c3-7368-shopshifting#download',
31 'only_matching': True,
32 }]
33
34 def _real_extract(self, url):
35 display_id = self._match_id(url)
36 webpage = self._download_webpage(url, display_id)
37 event_id = self._search_regex(r"data-id='(\d+)'", webpage, 'event id')
38 event_data = self._download_json('https://media.ccc.de/public/events/%s' % event_id, event_id)
39
40 formats = []
41 for recording in event_data.get('recordings', []):
42 recording_url = recording.get('recording_url')
43 if not recording_url:
44 continue
45 language = recording.get('language')
46 folder = recording.get('folder')
47 format_id = None
48 if language:
49 format_id = language
50 if folder:
51 if language:
52 format_id += '-' + folder
53 else:
54 format_id = folder
55 vcodec = 'h264' if 'h264' in folder else (
56 'none' if folder in ('mp3', 'opus') else None
57 )
58 formats.append({
59 'format_id': format_id,
60 'url': recording_url,
61 'width': int_or_none(recording.get('width')),
62 'height': int_or_none(recording.get('height')),
63 'filesize': int_or_none(recording.get('size'), invscale=1024 * 1024),
64 'language': language,
65 'vcodec': vcodec,
66 })
67
68 return {
69 'id': event_id,
70 'display_id': display_id,
71 'title': event_data['title'],
72 'creator': try_get(event_data, lambda x: ', '.join(x['persons'])),
73 'description': event_data.get('description'),
74 'thumbnail': event_data.get('thumb_url'),
75 'timestamp': parse_iso8601(event_data.get('date')),
76 'duration': int_or_none(event_data.get('length')),
77 'view_count': int_or_none(event_data.get('view_count')),
78 'tags': event_data.get('tags'),
79 'formats': formats,
80 }
81
82
83 class CCCPlaylistIE(InfoExtractor):
84 IE_NAME = 'media.ccc.de:lists'
85 _VALID_URL = r'https?://(?:www\.)?media\.ccc\.de/c/(?P<id>[^/?#&]+)'
86 _TESTS = [{
87 'url': 'https://media.ccc.de/c/30c3',
88 'info_dict': {
89 'title': '30C3',
90 'id': '30c3',
91 },
92 'playlist_count': 135,
93 }]
94
95 def _real_extract(self, url):
96 playlist_id = self._match_id(url).lower()
97
98 conf = self._download_json(
99 'https://media.ccc.de/public/conferences/' + playlist_id,
100 playlist_id)
101
102 entries = []
103 for e in conf['events']:
104 event_url = url_or_none(e.get('frontend_link'))
105 if event_url:
106 entries.append(self.url_result(event_url, ie=CCCIE.ie_key()))
107
108 return self.playlist_result(entries, playlist_id, conf.get('title'))
109
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/yt_dlp/extractor/ccc.py b/yt_dlp/extractor/ccc.py
--- a/yt_dlp/extractor/ccc.py
+++ b/yt_dlp/extractor/ccc.py
@@ -90,10 +90,17 @@
'id': '30c3',
},
'playlist_count': 135,
+ }, {
+ 'url': 'https://media.ccc.de/c/DS2023',
+ 'info_dict': {
+ 'title': 'Datenspuren 2023',
+ 'id': 'DS2023',
+ },
+ 'playlist_count': 37
}]
def _real_extract(self, url):
- playlist_id = self._match_id(url).lower()
+ playlist_id = self._match_id(url)
conf = self._download_json(
'https://media.ccc.de/public/conferences/' + playlist_id,
| {"golden_diff": "diff --git a/yt_dlp/extractor/ccc.py b/yt_dlp/extractor/ccc.py\n--- a/yt_dlp/extractor/ccc.py\n+++ b/yt_dlp/extractor/ccc.py\n@@ -90,10 +90,17 @@\n 'id': '30c3',\n },\n 'playlist_count': 135,\n+ }, {\n+ 'url': 'https://media.ccc.de/c/DS2023',\n+ 'info_dict': {\n+ 'title': 'Datenspuren 2023',\n+ 'id': 'DS2023',\n+ },\n+ 'playlist_count': 37\n }]\n \n def _real_extract(self, url):\n- playlist_id = self._match_id(url).lower()\n+ playlist_id = self._match_id(url)\n \n conf = self._download_json(\n 'https://media.ccc.de/public/conferences/' + playlist_id,\n", "issue": "[media.ccc.de:lists] playlist_id should be case-sensitive\n### DO NOT REMOVE OR SKIP THE ISSUE TEMPLATE\n\n- [X] I understand that I will be **blocked** if I *intentionally* remove or skip any mandatory\\* field\n\n### Checklist\n\n- [X] I'm reporting that yt-dlp is broken on a **supported** site\n- [X] I've verified that I'm running yt-dlp version **2023.07.06** ([update instructions](https://github.com/yt-dlp/yt-dlp#update)) or later (specify commit)\n- [X] I've checked that all provided URLs are playable in a browser with the same IP and same login details\n- [X] I've checked that all URLs and arguments with special characters are [properly quoted or escaped](https://github.com/yt-dlp/yt-dlp/wiki/FAQ#video-url-contains-an-ampersand--and-im-getting-some-strange-output-1-2839-or-v-is-not-recognized-as-an-internal-or-external-command)\n- [X] I've searched [known issues](https://github.com/yt-dlp/yt-dlp/issues/3766) and the [bugtracker](https://github.com/yt-dlp/yt-dlp/issues?q=) for similar issues **including closed ones**. DO NOT post duplicates\n- [X] I've read the [guidelines for opening an issue](https://github.com/yt-dlp/yt-dlp/blob/master/CONTRIBUTING.md#opening-an-issue)\n- [ ] I've read about [sharing account credentials](https://github.com/yt-dlp/yt-dlp/blob/master/CONTRIBUTING.md#are-you-willing-to-share-account-details-if-needed) and I'm willing to share it if required\n\n### Region\n\n_No response_\n\n### Provide a description that is worded well enough to be understood\n\nSome playlists use uppercase `playlist_id`s, like `https://media.ccc.de/c/DS2023` \u2192 `https://media.ccc.de/public/conferences/DS2022` or `https://media.ccc.de/c/MCH2022` \u2192 `https://media.ccc.de/public/conferences/MCH2022`. So I guess removing `.lower()` in https://github.com/yt-dlp/yt-dlp/blob/master/yt_dlp/extractor/ccc.py#L96 should resolve this.\n\n### Provide verbose output that clearly demonstrates the problem\n\n- [X] Run **your** yt-dlp command with **-vU** flag added (`yt-dlp -vU <your command line>`)\n- [ ] If using API, add `'verbose': True` to `YoutubeDL` params instead\n- [X] Copy the WHOLE output (starting with `[debug] Command-line config`) and insert it below\n\n### Complete Verbose Output\n\n```shell\n$ python -m yt_dlp --verbose --ignore-config https://media.ccc.de/c/DS2023\r\n[debug] Command-line config: ['--verbose', '--ignore-config', 'https://media.ccc.de/c/DS2023']\r\n[debug] Encodings: locale UTF-8, fs utf-8, pref UTF-8, out utf-8, error utf-8, screen utf-8\r\n[debug] yt-dlp version [email protected] [b532a3481] (source)\r\n[debug] Lazy loading extractors is disabled\r\n[debug] Git HEAD: 30ba233d4\r\n[debug] Python 3.11.5 (CPython x86_64 64bit) - Linux-6.5.0-5-generic-x86_64-with-glibc2.38 (OpenSSL 3.0.10 1 Aug 2023, glibc 2.38)\r\n[debug] exe versions: ffmpeg 6.0 (setts), ffprobe 6.0, rtmpdump 2.4\r\n[debug] Optional libraries: Cryptodome-3.11.0, brotli-1.0.9, certifi-2022.09.24, mutagen-1.46.0, pyxattr-0.8.1, secretstorage-3.3.3, sqlite3-2.6.0, websockets-10.4\r\n[debug] Proxy map: {}\r\n[debug] Loaded 1866 extractors\r\n[media.ccc.de:lists] Extracting URL: https://media.ccc.de/c/DS2023\r\n[media.ccc.de:lists] ds2023: Downloading JSON metadata\r\nERROR: [media.ccc.de:lists] DS2023: Unable to download JSON metadata: HTTP Error 404: Not Found (caused by <HTTPError 404: Not Found>); please report this issue on https://github.com/yt-dlp/yt-dlp/issues?q= , filling out the appropriate issue template. Confirm you are on the latest version using yt-dlp -U\r\n[\u2026]\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n File \"/tmp/yt-dlp/yt_dlp/extractor/common.py\", line 847, in _request_webpage\r\n return self._downloader.urlopen(self._create_request(url_or_request, data, headers, query))\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File \"/tmp/yt-dlp/yt_dlp/YoutubeDL.py\", line 4078, in urlopen\r\n raise _CompatHTTPError(e) from e\r\nyt_dlp.networking.exceptions._CompatHTTPError: HTTP Error 404: Not Found\n```\n\n", "before_files": [{"content": "from .common import InfoExtractor\nfrom ..utils import (\n int_or_none,\n parse_iso8601,\n try_get,\n url_or_none,\n)\n\n\nclass CCCIE(InfoExtractor):\n IE_NAME = 'media.ccc.de'\n _VALID_URL = r'https?://(?:www\\.)?media\\.ccc\\.de/v/(?P<id>[^/?#&]+)'\n\n _TESTS = [{\n 'url': 'https://media.ccc.de/v/30C3_-_5443_-_en_-_saal_g_-_201312281830_-_introduction_to_processor_design_-_byterazor#video',\n 'md5': '3a1eda8f3a29515d27f5adb967d7e740',\n 'info_dict': {\n 'id': '1839',\n 'ext': 'mp4',\n 'title': 'Introduction to Processor Design',\n 'creator': 'byterazor',\n 'description': 'md5:df55f6d073d4ceae55aae6f2fd98a0ac',\n 'thumbnail': r're:^https?://.*\\.jpg$',\n 'upload_date': '20131228',\n 'timestamp': 1388188800,\n 'duration': 3710,\n 'tags': list,\n }\n }, {\n 'url': 'https://media.ccc.de/v/32c3-7368-shopshifting#download',\n 'only_matching': True,\n }]\n\n def _real_extract(self, url):\n display_id = self._match_id(url)\n webpage = self._download_webpage(url, display_id)\n event_id = self._search_regex(r\"data-id='(\\d+)'\", webpage, 'event id')\n event_data = self._download_json('https://media.ccc.de/public/events/%s' % event_id, event_id)\n\n formats = []\n for recording in event_data.get('recordings', []):\n recording_url = recording.get('recording_url')\n if not recording_url:\n continue\n language = recording.get('language')\n folder = recording.get('folder')\n format_id = None\n if language:\n format_id = language\n if folder:\n if language:\n format_id += '-' + folder\n else:\n format_id = folder\n vcodec = 'h264' if 'h264' in folder else (\n 'none' if folder in ('mp3', 'opus') else None\n )\n formats.append({\n 'format_id': format_id,\n 'url': recording_url,\n 'width': int_or_none(recording.get('width')),\n 'height': int_or_none(recording.get('height')),\n 'filesize': int_or_none(recording.get('size'), invscale=1024 * 1024),\n 'language': language,\n 'vcodec': vcodec,\n })\n\n return {\n 'id': event_id,\n 'display_id': display_id,\n 'title': event_data['title'],\n 'creator': try_get(event_data, lambda x: ', '.join(x['persons'])),\n 'description': event_data.get('description'),\n 'thumbnail': event_data.get('thumb_url'),\n 'timestamp': parse_iso8601(event_data.get('date')),\n 'duration': int_or_none(event_data.get('length')),\n 'view_count': int_or_none(event_data.get('view_count')),\n 'tags': event_data.get('tags'),\n 'formats': formats,\n }\n\n\nclass CCCPlaylistIE(InfoExtractor):\n IE_NAME = 'media.ccc.de:lists'\n _VALID_URL = r'https?://(?:www\\.)?media\\.ccc\\.de/c/(?P<id>[^/?#&]+)'\n _TESTS = [{\n 'url': 'https://media.ccc.de/c/30c3',\n 'info_dict': {\n 'title': '30C3',\n 'id': '30c3',\n },\n 'playlist_count': 135,\n }]\n\n def _real_extract(self, url):\n playlist_id = self._match_id(url).lower()\n\n conf = self._download_json(\n 'https://media.ccc.de/public/conferences/' + playlist_id,\n playlist_id)\n\n entries = []\n for e in conf['events']:\n event_url = url_or_none(e.get('frontend_link'))\n if event_url:\n entries.append(self.url_result(event_url, ie=CCCIE.ie_key()))\n\n return self.playlist_result(entries, playlist_id, conf.get('title'))\n", "path": "yt_dlp/extractor/ccc.py"}], "after_files": [{"content": "from .common import InfoExtractor\nfrom ..utils import (\n int_or_none,\n parse_iso8601,\n try_get,\n url_or_none,\n)\n\n\nclass CCCIE(InfoExtractor):\n IE_NAME = 'media.ccc.de'\n _VALID_URL = r'https?://(?:www\\.)?media\\.ccc\\.de/v/(?P<id>[^/?#&]+)'\n\n _TESTS = [{\n 'url': 'https://media.ccc.de/v/30C3_-_5443_-_en_-_saal_g_-_201312281830_-_introduction_to_processor_design_-_byterazor#video',\n 'md5': '3a1eda8f3a29515d27f5adb967d7e740',\n 'info_dict': {\n 'id': '1839',\n 'ext': 'mp4',\n 'title': 'Introduction to Processor Design',\n 'creator': 'byterazor',\n 'description': 'md5:df55f6d073d4ceae55aae6f2fd98a0ac',\n 'thumbnail': r're:^https?://.*\\.jpg$',\n 'upload_date': '20131228',\n 'timestamp': 1388188800,\n 'duration': 3710,\n 'tags': list,\n }\n }, {\n 'url': 'https://media.ccc.de/v/32c3-7368-shopshifting#download',\n 'only_matching': True,\n }]\n\n def _real_extract(self, url):\n display_id = self._match_id(url)\n webpage = self._download_webpage(url, display_id)\n event_id = self._search_regex(r\"data-id='(\\d+)'\", webpage, 'event id')\n event_data = self._download_json('https://media.ccc.de/public/events/%s' % event_id, event_id)\n\n formats = []\n for recording in event_data.get('recordings', []):\n recording_url = recording.get('recording_url')\n if not recording_url:\n continue\n language = recording.get('language')\n folder = recording.get('folder')\n format_id = None\n if language:\n format_id = language\n if folder:\n if language:\n format_id += '-' + folder\n else:\n format_id = folder\n vcodec = 'h264' if 'h264' in folder else (\n 'none' if folder in ('mp3', 'opus') else None\n )\n formats.append({\n 'format_id': format_id,\n 'url': recording_url,\n 'width': int_or_none(recording.get('width')),\n 'height': int_or_none(recording.get('height')),\n 'filesize': int_or_none(recording.get('size'), invscale=1024 * 1024),\n 'language': language,\n 'vcodec': vcodec,\n })\n\n return {\n 'id': event_id,\n 'display_id': display_id,\n 'title': event_data['title'],\n 'creator': try_get(event_data, lambda x: ', '.join(x['persons'])),\n 'description': event_data.get('description'),\n 'thumbnail': event_data.get('thumb_url'),\n 'timestamp': parse_iso8601(event_data.get('date')),\n 'duration': int_or_none(event_data.get('length')),\n 'view_count': int_or_none(event_data.get('view_count')),\n 'tags': event_data.get('tags'),\n 'formats': formats,\n }\n\n\nclass CCCPlaylistIE(InfoExtractor):\n IE_NAME = 'media.ccc.de:lists'\n _VALID_URL = r'https?://(?:www\\.)?media\\.ccc\\.de/c/(?P<id>[^/?#&]+)'\n _TESTS = [{\n 'url': 'https://media.ccc.de/c/30c3',\n 'info_dict': {\n 'title': '30C3',\n 'id': '30c3',\n },\n 'playlist_count': 135,\n }, {\n 'url': 'https://media.ccc.de/c/DS2023',\n 'info_dict': {\n 'title': 'Datenspuren 2023',\n 'id': 'DS2023',\n },\n 'playlist_count': 37\n }]\n\n def _real_extract(self, url):\n playlist_id = self._match_id(url)\n\n conf = self._download_json(\n 'https://media.ccc.de/public/conferences/' + playlist_id,\n playlist_id)\n\n entries = []\n for e in conf['events']:\n event_url = url_or_none(e.get('frontend_link'))\n if event_url:\n entries.append(self.url_result(event_url, ie=CCCIE.ie_key()))\n\n return self.playlist_result(entries, playlist_id, conf.get('title'))\n", "path": "yt_dlp/extractor/ccc.py"}]} | 2,800 | 221 |
gh_patches_debug_40887 | rasdani/github-patches | git_diff | mozilla__pontoon-2853 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Pretranslate access keys using the algorithm to extract candidate keys
Fluent Rich editor has a special UI for messages with access keys, which lists access key candidates.
We should use the same logic when pretranslating accesskeys and use the first candidate as the translation.
We should also take into account #2717.
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `pontoon/pretranslation/transformer.py`
Content:
```
1 from copy import deepcopy
2 from typing import Callable, Optional, cast
3
4 from fluent.syntax import ast as FTL
5 from fluent.syntax.serializer import serialize_expression
6 from fluent.syntax.visitor import Transformer
7
8 from pontoon.base.fluent import is_plural_expression
9 from pontoon.base.models import Locale
10
11
12 def flatten_select_expressions(pattern: FTL.Pattern):
13 """
14 If the pattern contains any select expressions,
15 flatten it to only contain select expressions.
16 Leading and trailing elements are copied into each variant,
17 and any single leading or trailing spaces are lifted out of the select expressions.
18 """
19
20 def isSelExp(el: FTL.PatternElement):
21 return isinstance(el, FTL.Placeable) and isinstance(
22 el.expression, FTL.SelectExpression
23 )
24
25 def patternStartsWithSpace(pat: list[FTL.PatternElement]):
26 return isinstance(pat[0], FTL.TextElement) and pat[0].value.startswith(" ")
27
28 def patternEndsWithSpace(pat: list[FTL.PatternElement]):
29 return isinstance(pat[-1], FTL.TextElement) and pat[-1].value.endswith(" ")
30
31 prev = -1
32 select = None
33 for idx, placeable in filter(lambda x: isSelExp(x[1]), enumerate(pattern.elements)):
34 before = pattern.elements[prev + 1 : idx]
35 if before:
36 select = cast(FTL.SelectExpression, placeable.expression)
37 for variant in select.variants:
38 variant.value.elements[0:0] = deepcopy(before)
39 prev = idx
40 if select:
41 after = pattern.elements[prev + 1 :]
42 if after:
43 for variant in select.variants:
44 variant.value.elements += deepcopy(after)
45
46 res: list[FTL.PatternElement] = []
47 for placeable in filter(isSelExp, pattern.elements):
48 patterns = tuple(
49 map(lambda var: var.value.elements, placeable.expression.variants)
50 )
51
52 # Collect leading spaces
53 if all(map(patternStartsWithSpace, patterns)):
54 res.append(FTL.Placeable(FTL.StringLiteral(" ")))
55 for pat in patterns:
56 pat[0].value = pat[0].value[1:]
57
58 res.append(placeable)
59
60 # Collect trailing spaces
61 if all(map(patternEndsWithSpace, patterns)):
62 res.append(FTL.Placeable(FTL.StringLiteral(" ")))
63 for pat in patterns:
64 pat[-1].value = pat[-1].value[:-1]
65 pattern.elements = res
66
67
68 def create_locale_plural_variants(node: FTL.SelectExpression, locale: Locale):
69 variants: list[FTL.Variant] = []
70 source_plurals: dict[str, FTL.Variant] = {}
71 default = cast(FTL.Variant, None)
72
73 for variant in node.variants:
74 key = variant.key
75 if isinstance(key, FTL.NumberLiteral):
76 variants.append(variant)
77 else:
78 source_plurals[key.name] = variant
79 if variant.default:
80 default = variant
81
82 for plural in locale.cldr_plurals_list():
83 if plural in source_plurals.keys():
84 variant = source_plurals[plural]
85 else:
86 variant = deepcopy(default)
87 variant.key.name = plural
88 variant.default = False
89 variants.append(variant)
90
91 variants[-1].default = True
92
93 node.variants = variants
94
95
96 class PreparePretranslation(Transformer):
97 """
98 Flattens the given Pattern, uplifting selectors to the highest possible level and
99 duplicating shared parts in the variants. Transforms plural variants to match the
100 locale.
101 """
102
103 def __init__(self, locale: Locale):
104 self.locale = locale
105
106 def visit_Attribute(self, node: FTL.Attribute):
107 flatten_select_expressions(node.value)
108 return self.generic_visit(node)
109
110 def visit_Message(self, node: FTL.Message):
111 if node.value:
112 flatten_select_expressions(node.value)
113 return self.generic_visit(node)
114
115 def visit_SelectExpression(self, node: FTL.SelectExpression):
116 if is_plural_expression(node):
117 create_locale_plural_variants(node, self.locale)
118 return self.generic_visit(node)
119
120
121 class ApplyPretranslation(Transformer):
122 """
123 During `visit()`, calls `callback(source, locale) -> (translation, service)` for each pattern.
124 """
125
126 def __init__(
127 self,
128 locale: Locale,
129 entry: FTL.EntryType,
130 callback: Callable[[str, str], tuple[Optional[str], str]],
131 ):
132 prep = PreparePretranslation(locale)
133 prep.visit(entry)
134 self.callback = callback
135 self.locale = locale
136 self.services: list[str] = []
137
138 def visit_Attribute(self, node):
139 if (
140 node.id.name.endswith("accesskey")
141 and not self.locale.accesskey_localization
142 ):
143 return node
144 return self.generic_visit(node)
145
146 def visit_Pattern(self, node: FTL.Pattern):
147 has_selects = False
148 source = ""
149 for el in node.elements:
150 if isinstance(el, FTL.TextElement):
151 source += el.value
152 elif isinstance(el.expression, FTL.SelectExpression):
153 self.generic_visit(el.expression)
154 has_selects = True
155 else:
156 source += serialize_expression(el)
157 if not has_selects and source != "":
158 # Machine translation treats each line as a separate sentence,
159 # hence we replace newline characters with spaces.
160 source = source.replace("\n", " ")
161
162 translation, service = self.callback(source, self.locale)
163 if translation is None:
164 raise ValueError(
165 f"Pretranslation for `{source}` to {self.locale.code} not available."
166 )
167 node.elements = [FTL.TextElement(translation)]
168 self.services.append(service)
169 return node
170
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/pontoon/pretranslation/transformer.py b/pontoon/pretranslation/transformer.py
--- a/pontoon/pretranslation/transformer.py
+++ b/pontoon/pretranslation/transformer.py
@@ -1,3 +1,5 @@
+import re
+
from copy import deepcopy
from typing import Callable, Optional, cast
@@ -93,6 +95,51 @@
node.variants = variants
+def extract_accesskey_candidates(message: FTL.Message, label: str, variant_name=None):
+ def get_source(names):
+ for attribute in message.attributes:
+ if attribute.id.name in names:
+ element = attribute.value.elements[0]
+
+ if isinstance(element, FTL.TextElement):
+ return element.value
+ elif isinstance(element.expression, FTL.SelectExpression):
+ variants = element.expression.variants
+ variant = next(
+ (v for v in variants if v.key.name == variant_name), variants[0]
+ )
+ variant_element = variant.value.elements[0]
+
+ if isinstance(variant_element, FTL.TextElement):
+ return variant_element.value
+
+ return None
+
+ prefix_end = label.index("accesskey")
+ prefix = label[0:prefix_end]
+
+ # Generate access key candidates:
+ if prefix:
+ # From a prefixed "label" attribute
+ name = f"{prefix}label"
+ source = get_source([name])
+ else:
+ # From a pre-defined list of attribute names
+ source = get_source(["label", "value", "aria-label"])
+ # From a message value
+ if not source and message.value:
+ source = message.value.elements[0].value
+
+ if not source:
+ return []
+
+ # Exclude placeables (message is flat). See bug 1447103 for details.
+ keys = re.sub(r"(?s){.*?}|[\W_]", "", source)
+
+ # Extract unique candidates
+ return list(dict.fromkeys(keys))
+
+
class PreparePretranslation(Transformer):
"""
Flattens the given Pattern, uplifting selectors to the highest possible level and
@@ -132,15 +179,43 @@
prep = PreparePretranslation(locale)
prep.visit(entry)
self.callback = callback
+ self.entry = entry
self.locale = locale
self.services: list[str] = []
- def visit_Attribute(self, node):
- if (
- node.id.name.endswith("accesskey")
- and not self.locale.accesskey_localization
- ):
- return node
+ def visit_Attribute(self, node: FTL.Pattern):
+ name = node.id.name
+
+ def set_accesskey(element, variant_name=None):
+ if isinstance(element, FTL.TextElement) and len(element.value) <= 1:
+ candidates = extract_accesskey_candidates(
+ self.entry, name, variant_name
+ )
+ if candidates:
+ element.value = candidates[0]
+ return True
+
+ if name.endswith("accesskey"):
+ if self.locale.accesskey_localization:
+ element = node.value.elements[0]
+
+ if set_accesskey(element):
+ return node
+ elif isinstance(element, FTL.Placeable) and isinstance(
+ element.expression, FTL.SelectExpression
+ ):
+ variants = element.expression.variants
+ processed_variants = 0
+ for variant in variants:
+ variant_element = variant.value.elements[0]
+ if set_accesskey(variant_element, variant.key.name):
+ processed_variants += 1
+ if processed_variants == len(variants):
+ return node
+
+ else:
+ return node
+
return self.generic_visit(node)
def visit_Pattern(self, node: FTL.Pattern):
| {"golden_diff": "diff --git a/pontoon/pretranslation/transformer.py b/pontoon/pretranslation/transformer.py\n--- a/pontoon/pretranslation/transformer.py\n+++ b/pontoon/pretranslation/transformer.py\n@@ -1,3 +1,5 @@\n+import re\n+\n from copy import deepcopy\n from typing import Callable, Optional, cast\n \n@@ -93,6 +95,51 @@\n node.variants = variants\n \n \n+def extract_accesskey_candidates(message: FTL.Message, label: str, variant_name=None):\n+ def get_source(names):\n+ for attribute in message.attributes:\n+ if attribute.id.name in names:\n+ element = attribute.value.elements[0]\n+\n+ if isinstance(element, FTL.TextElement):\n+ return element.value\n+ elif isinstance(element.expression, FTL.SelectExpression):\n+ variants = element.expression.variants\n+ variant = next(\n+ (v for v in variants if v.key.name == variant_name), variants[0]\n+ )\n+ variant_element = variant.value.elements[0]\n+\n+ if isinstance(variant_element, FTL.TextElement):\n+ return variant_element.value\n+\n+ return None\n+\n+ prefix_end = label.index(\"accesskey\")\n+ prefix = label[0:prefix_end]\n+\n+ # Generate access key candidates:\n+ if prefix:\n+ # From a prefixed \"label\" attribute\n+ name = f\"{prefix}label\"\n+ source = get_source([name])\n+ else:\n+ # From a pre-defined list of attribute names\n+ source = get_source([\"label\", \"value\", \"aria-label\"])\n+ # From a message value\n+ if not source and message.value:\n+ source = message.value.elements[0].value\n+\n+ if not source:\n+ return []\n+\n+ # Exclude placeables (message is flat). See bug 1447103 for details.\n+ keys = re.sub(r\"(?s){.*?}|[\\W_]\", \"\", source)\n+\n+ # Extract unique candidates\n+ return list(dict.fromkeys(keys))\n+\n+\n class PreparePretranslation(Transformer):\n \"\"\"\n Flattens the given Pattern, uplifting selectors to the highest possible level and\n@@ -132,15 +179,43 @@\n prep = PreparePretranslation(locale)\n prep.visit(entry)\n self.callback = callback\n+ self.entry = entry\n self.locale = locale\n self.services: list[str] = []\n \n- def visit_Attribute(self, node):\n- if (\n- node.id.name.endswith(\"accesskey\")\n- and not self.locale.accesskey_localization\n- ):\n- return node\n+ def visit_Attribute(self, node: FTL.Pattern):\n+ name = node.id.name\n+\n+ def set_accesskey(element, variant_name=None):\n+ if isinstance(element, FTL.TextElement) and len(element.value) <= 1:\n+ candidates = extract_accesskey_candidates(\n+ self.entry, name, variant_name\n+ )\n+ if candidates:\n+ element.value = candidates[0]\n+ return True\n+\n+ if name.endswith(\"accesskey\"):\n+ if self.locale.accesskey_localization:\n+ element = node.value.elements[0]\n+\n+ if set_accesskey(element):\n+ return node\n+ elif isinstance(element, FTL.Placeable) and isinstance(\n+ element.expression, FTL.SelectExpression\n+ ):\n+ variants = element.expression.variants\n+ processed_variants = 0\n+ for variant in variants:\n+ variant_element = variant.value.elements[0]\n+ if set_accesskey(variant_element, variant.key.name):\n+ processed_variants += 1\n+ if processed_variants == len(variants):\n+ return node\n+\n+ else:\n+ return node\n+\n return self.generic_visit(node)\n \n def visit_Pattern(self, node: FTL.Pattern):\n", "issue": "Pretranslate access keys using the algorithm to extract candidate keys\nFluent Rich editor has a special UI for messages with access keys, which lists access key candidates.\r\n\r\nWe should use the same logic when pretranslating accesskeys and use the first candidate as the translation.\r\n\r\nWe should also take into account #2717.\n", "before_files": [{"content": "from copy import deepcopy\nfrom typing import Callable, Optional, cast\n\nfrom fluent.syntax import ast as FTL\nfrom fluent.syntax.serializer import serialize_expression\nfrom fluent.syntax.visitor import Transformer\n\nfrom pontoon.base.fluent import is_plural_expression\nfrom pontoon.base.models import Locale\n\n\ndef flatten_select_expressions(pattern: FTL.Pattern):\n \"\"\"\n If the pattern contains any select expressions,\n flatten it to only contain select expressions.\n Leading and trailing elements are copied into each variant,\n and any single leading or trailing spaces are lifted out of the select expressions.\n \"\"\"\n\n def isSelExp(el: FTL.PatternElement):\n return isinstance(el, FTL.Placeable) and isinstance(\n el.expression, FTL.SelectExpression\n )\n\n def patternStartsWithSpace(pat: list[FTL.PatternElement]):\n return isinstance(pat[0], FTL.TextElement) and pat[0].value.startswith(\" \")\n\n def patternEndsWithSpace(pat: list[FTL.PatternElement]):\n return isinstance(pat[-1], FTL.TextElement) and pat[-1].value.endswith(\" \")\n\n prev = -1\n select = None\n for idx, placeable in filter(lambda x: isSelExp(x[1]), enumerate(pattern.elements)):\n before = pattern.elements[prev + 1 : idx]\n if before:\n select = cast(FTL.SelectExpression, placeable.expression)\n for variant in select.variants:\n variant.value.elements[0:0] = deepcopy(before)\n prev = idx\n if select:\n after = pattern.elements[prev + 1 :]\n if after:\n for variant in select.variants:\n variant.value.elements += deepcopy(after)\n\n res: list[FTL.PatternElement] = []\n for placeable in filter(isSelExp, pattern.elements):\n patterns = tuple(\n map(lambda var: var.value.elements, placeable.expression.variants)\n )\n\n # Collect leading spaces\n if all(map(patternStartsWithSpace, patterns)):\n res.append(FTL.Placeable(FTL.StringLiteral(\" \")))\n for pat in patterns:\n pat[0].value = pat[0].value[1:]\n\n res.append(placeable)\n\n # Collect trailing spaces\n if all(map(patternEndsWithSpace, patterns)):\n res.append(FTL.Placeable(FTL.StringLiteral(\" \")))\n for pat in patterns:\n pat[-1].value = pat[-1].value[:-1]\n pattern.elements = res\n\n\ndef create_locale_plural_variants(node: FTL.SelectExpression, locale: Locale):\n variants: list[FTL.Variant] = []\n source_plurals: dict[str, FTL.Variant] = {}\n default = cast(FTL.Variant, None)\n\n for variant in node.variants:\n key = variant.key\n if isinstance(key, FTL.NumberLiteral):\n variants.append(variant)\n else:\n source_plurals[key.name] = variant\n if variant.default:\n default = variant\n\n for plural in locale.cldr_plurals_list():\n if plural in source_plurals.keys():\n variant = source_plurals[plural]\n else:\n variant = deepcopy(default)\n variant.key.name = plural\n variant.default = False\n variants.append(variant)\n\n variants[-1].default = True\n\n node.variants = variants\n\n\nclass PreparePretranslation(Transformer):\n \"\"\"\n Flattens the given Pattern, uplifting selectors to the highest possible level and\n duplicating shared parts in the variants. Transforms plural variants to match the\n locale.\n \"\"\"\n\n def __init__(self, locale: Locale):\n self.locale = locale\n\n def visit_Attribute(self, node: FTL.Attribute):\n flatten_select_expressions(node.value)\n return self.generic_visit(node)\n\n def visit_Message(self, node: FTL.Message):\n if node.value:\n flatten_select_expressions(node.value)\n return self.generic_visit(node)\n\n def visit_SelectExpression(self, node: FTL.SelectExpression):\n if is_plural_expression(node):\n create_locale_plural_variants(node, self.locale)\n return self.generic_visit(node)\n\n\nclass ApplyPretranslation(Transformer):\n \"\"\"\n During `visit()`, calls `callback(source, locale) -> (translation, service)` for each pattern.\n \"\"\"\n\n def __init__(\n self,\n locale: Locale,\n entry: FTL.EntryType,\n callback: Callable[[str, str], tuple[Optional[str], str]],\n ):\n prep = PreparePretranslation(locale)\n prep.visit(entry)\n self.callback = callback\n self.locale = locale\n self.services: list[str] = []\n\n def visit_Attribute(self, node):\n if (\n node.id.name.endswith(\"accesskey\")\n and not self.locale.accesskey_localization\n ):\n return node\n return self.generic_visit(node)\n\n def visit_Pattern(self, node: FTL.Pattern):\n has_selects = False\n source = \"\"\n for el in node.elements:\n if isinstance(el, FTL.TextElement):\n source += el.value\n elif isinstance(el.expression, FTL.SelectExpression):\n self.generic_visit(el.expression)\n has_selects = True\n else:\n source += serialize_expression(el)\n if not has_selects and source != \"\":\n # Machine translation treats each line as a separate sentence,\n # hence we replace newline characters with spaces.\n source = source.replace(\"\\n\", \" \")\n\n translation, service = self.callback(source, self.locale)\n if translation is None:\n raise ValueError(\n f\"Pretranslation for `{source}` to {self.locale.code} not available.\"\n )\n node.elements = [FTL.TextElement(translation)]\n self.services.append(service)\n return node\n", "path": "pontoon/pretranslation/transformer.py"}], "after_files": [{"content": "import re\n\nfrom copy import deepcopy\nfrom typing import Callable, Optional, cast\n\nfrom fluent.syntax import ast as FTL\nfrom fluent.syntax.serializer import serialize_expression\nfrom fluent.syntax.visitor import Transformer\n\nfrom pontoon.base.fluent import is_plural_expression\nfrom pontoon.base.models import Locale\n\n\ndef flatten_select_expressions(pattern: FTL.Pattern):\n \"\"\"\n If the pattern contains any select expressions,\n flatten it to only contain select expressions.\n Leading and trailing elements are copied into each variant,\n and any single leading or trailing spaces are lifted out of the select expressions.\n \"\"\"\n\n def isSelExp(el: FTL.PatternElement):\n return isinstance(el, FTL.Placeable) and isinstance(\n el.expression, FTL.SelectExpression\n )\n\n def patternStartsWithSpace(pat: list[FTL.PatternElement]):\n return isinstance(pat[0], FTL.TextElement) and pat[0].value.startswith(\" \")\n\n def patternEndsWithSpace(pat: list[FTL.PatternElement]):\n return isinstance(pat[-1], FTL.TextElement) and pat[-1].value.endswith(\" \")\n\n prev = -1\n select = None\n for idx, placeable in filter(lambda x: isSelExp(x[1]), enumerate(pattern.elements)):\n before = pattern.elements[prev + 1 : idx]\n if before:\n select = cast(FTL.SelectExpression, placeable.expression)\n for variant in select.variants:\n variant.value.elements[0:0] = deepcopy(before)\n prev = idx\n if select:\n after = pattern.elements[prev + 1 :]\n if after:\n for variant in select.variants:\n variant.value.elements += deepcopy(after)\n\n res: list[FTL.PatternElement] = []\n for placeable in filter(isSelExp, pattern.elements):\n patterns = tuple(\n map(lambda var: var.value.elements, placeable.expression.variants)\n )\n\n # Collect leading spaces\n if all(map(patternStartsWithSpace, patterns)):\n res.append(FTL.Placeable(FTL.StringLiteral(\" \")))\n for pat in patterns:\n pat[0].value = pat[0].value[1:]\n\n res.append(placeable)\n\n # Collect trailing spaces\n if all(map(patternEndsWithSpace, patterns)):\n res.append(FTL.Placeable(FTL.StringLiteral(\" \")))\n for pat in patterns:\n pat[-1].value = pat[-1].value[:-1]\n pattern.elements = res\n\n\ndef create_locale_plural_variants(node: FTL.SelectExpression, locale: Locale):\n variants: list[FTL.Variant] = []\n source_plurals: dict[str, FTL.Variant] = {}\n default = cast(FTL.Variant, None)\n\n for variant in node.variants:\n key = variant.key\n if isinstance(key, FTL.NumberLiteral):\n variants.append(variant)\n else:\n source_plurals[key.name] = variant\n if variant.default:\n default = variant\n\n for plural in locale.cldr_plurals_list():\n if plural in source_plurals.keys():\n variant = source_plurals[plural]\n else:\n variant = deepcopy(default)\n variant.key.name = plural\n variant.default = False\n variants.append(variant)\n\n variants[-1].default = True\n\n node.variants = variants\n\n\ndef extract_accesskey_candidates(message: FTL.Message, label: str, variant_name=None):\n def get_source(names):\n for attribute in message.attributes:\n if attribute.id.name in names:\n element = attribute.value.elements[0]\n\n if isinstance(element, FTL.TextElement):\n return element.value\n elif isinstance(element.expression, FTL.SelectExpression):\n variants = element.expression.variants\n variant = next(\n (v for v in variants if v.key.name == variant_name), variants[0]\n )\n variant_element = variant.value.elements[0]\n\n if isinstance(variant_element, FTL.TextElement):\n return variant_element.value\n\n return None\n\n prefix_end = label.index(\"accesskey\")\n prefix = label[0:prefix_end]\n\n # Generate access key candidates:\n if prefix:\n # From a prefixed \"label\" attribute\n name = f\"{prefix}label\"\n source = get_source([name])\n else:\n # From a pre-defined list of attribute names\n source = get_source([\"label\", \"value\", \"aria-label\"])\n # From a message value\n if not source and message.value:\n source = message.value.elements[0].value\n\n if not source:\n return []\n\n # Exclude placeables (message is flat). See bug 1447103 for details.\n keys = re.sub(r\"(?s){.*?}|[\\W_]\", \"\", source)\n\n # Extract unique candidates\n return list(dict.fromkeys(keys))\n\n\nclass PreparePretranslation(Transformer):\n \"\"\"\n Flattens the given Pattern, uplifting selectors to the highest possible level and\n duplicating shared parts in the variants. Transforms plural variants to match the\n locale.\n \"\"\"\n\n def __init__(self, locale: Locale):\n self.locale = locale\n\n def visit_Attribute(self, node: FTL.Attribute):\n flatten_select_expressions(node.value)\n return self.generic_visit(node)\n\n def visit_Message(self, node: FTL.Message):\n if node.value:\n flatten_select_expressions(node.value)\n return self.generic_visit(node)\n\n def visit_SelectExpression(self, node: FTL.SelectExpression):\n if is_plural_expression(node):\n create_locale_plural_variants(node, self.locale)\n return self.generic_visit(node)\n\n\nclass ApplyPretranslation(Transformer):\n \"\"\"\n During `visit()`, calls `callback(source, locale) -> (translation, service)` for each pattern.\n \"\"\"\n\n def __init__(\n self,\n locale: Locale,\n entry: FTL.EntryType,\n callback: Callable[[str, str], tuple[Optional[str], str]],\n ):\n prep = PreparePretranslation(locale)\n prep.visit(entry)\n self.callback = callback\n self.entry = entry\n self.locale = locale\n self.services: list[str] = []\n\n def visit_Attribute(self, node: FTL.Pattern):\n name = node.id.name\n\n def set_accesskey(element, variant_name=None):\n if isinstance(element, FTL.TextElement) and len(element.value) <= 1:\n candidates = extract_accesskey_candidates(\n self.entry, name, variant_name\n )\n if candidates:\n element.value = candidates[0]\n return True\n\n if name.endswith(\"accesskey\"):\n if self.locale.accesskey_localization:\n element = node.value.elements[0]\n\n if set_accesskey(element):\n return node\n elif isinstance(element, FTL.Placeable) and isinstance(\n element.expression, FTL.SelectExpression\n ):\n variants = element.expression.variants\n processed_variants = 0\n for variant in variants:\n variant_element = variant.value.elements[0]\n if set_accesskey(variant_element, variant.key.name):\n processed_variants += 1\n if processed_variants == len(variants):\n return node\n\n else:\n return node\n\n return self.generic_visit(node)\n\n def visit_Pattern(self, node: FTL.Pattern):\n has_selects = False\n source = \"\"\n for el in node.elements:\n if isinstance(el, FTL.TextElement):\n source += el.value\n elif isinstance(el.expression, FTL.SelectExpression):\n self.generic_visit(el.expression)\n has_selects = True\n else:\n source += serialize_expression(el)\n if not has_selects and source != \"\":\n # Machine translation treats each line as a separate sentence,\n # hence we replace newline characters with spaces.\n source = source.replace(\"\\n\", \" \")\n\n translation, service = self.callback(source, self.locale)\n if translation is None:\n raise ValueError(\n f\"Pretranslation for `{source}` to {self.locale.code} not available.\"\n )\n node.elements = [FTL.TextElement(translation)]\n self.services.append(service)\n return node\n", "path": "pontoon/pretranslation/transformer.py"}]} | 1,969 | 863 |
gh_patches_debug_39375 | rasdani/github-patches | git_diff | opendatacube__datacube-core-694 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
write_geotiff helper function fails if CRS is string, not object
### Expected behaviour
The write_geotiff helper function assumes that you will pass the function a datacube crs object. If you are writing out a geotiff from data that was not loaded using `dc.load`, this crs object is not present.
E.g. I read a Geotiff file produced by datacube-stats into a Notebook, ran some analysis on it, and wanted to write it back out to Geotiff. I have a crs string from the attributes of my original Geotiff, but no datacube crs object, so the write_geotiff function fails.
### Error
```
AttributeError Traceback (most recent call last)
<ipython-input-41-736bab55bae5> in <module>()
3 Differenceds.attrs['crs'] = (GeotiffData.crs)
4
----> 5 write_geotiff(PercentileConfidence, Differenceds)
/g/data/v10/public/modules/dea/20180515/lib/python3.6/site-packages/datacube/helpers.py in write_geotiff(filename, dataset, profile_override, time_index)
44 profile = DEFAULT_PROFILE.copy()
45 profile.update({
---> 46 'width': dataset.dims[dataset.crs.dimensions[1]],
47 'height': dataset.dims[dataset.crs.dimensions[0]],
48 'transform': dataset.affine,
AttributeError: 'str' object has no attribute 'dimensions'
```
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `datacube/helpers.py`
Content:
```
1 """
2 Useful functions for Datacube users
3
4 Not used internally, those should go in `utils.py`
5 """
6
7 import numpy as np
8 import rasterio
9
10 DEFAULT_PROFILE = {
11 'blockxsize': 256,
12 'blockysize': 256,
13 'compress': 'lzw',
14 'driver': 'GTiff',
15 'interleave': 'band',
16 'nodata': 0.0,
17 'tiled': True}
18
19
20 def write_geotiff(filename, dataset, profile_override=None, time_index=None):
21 """
22 Write an ODC style xarray.Dataset to a GeoTIFF file.
23
24 :param filename: Output filename
25 :param dataset: xarray dataset containing one or more bands to write to a file.
26 :param profile_override: option dict, overrides rasterio file creation options.
27 :param time_index: DEPRECATED
28 """
29 profile_override = profile_override or {}
30
31 if time_index is not None:
32 raise ValueError('''The write_geotiff function no longer supports passing in `time_index`.
33 The same function can be achieved by calling `dataset.isel(time=<time_index>)` before passing
34 in your dataset. It was removed because it made the function much less useful for more advanced cases.''')
35
36 try:
37 dtypes = {val.dtype for val in dataset.data_vars.values()}
38 assert len(dtypes) == 1 # Check for multiple dtypes
39 except AttributeError:
40 dtypes = [dataset.dtype]
41
42 profile = DEFAULT_PROFILE.copy()
43 profile.update({
44 'width': dataset.dims[dataset.crs.dimensions[1]],
45 'height': dataset.dims[dataset.crs.dimensions[0]],
46 'transform': dataset.affine,
47 'crs': dataset.crs.crs_str,
48 'count': len(dataset.data_vars),
49 'dtype': str(dtypes.pop())
50 })
51 profile.update(profile_override)
52
53 _calculate_blocksize(profile)
54
55 with rasterio.open(str(filename), 'w', **profile) as dest:
56 if hasattr(dataset, 'data_vars'):
57 for bandnum, data in enumerate(dataset.data_vars.values(), start=1):
58 dest.write(data.data, bandnum)
59
60
61 def _calculate_blocksize(profile):
62 # Block size must be smaller than the image size, and for geotiffs must be divisible by 16
63 # Fix for small images.
64 if profile['blockxsize'] > profile['width']:
65 if profile['width'] % 16 == 0 or profile['width'] < 16:
66 profile['blockxsize'] = profile['width']
67 else:
68 profile['blockxsize'] = 16
69
70 if profile['blockysize'] > profile['height']:
71 if profile['height'] % 16 == 0 or profile['height'] < 16:
72 profile['blockysize'] = profile['height']
73 else:
74 profile['blockysize'] = 16
75
76
77 def ga_pq_fuser(dest, src):
78 """
79 Fuse two Geoscience Australia Pixel Quality ndarrays
80
81 To be used as a `fuse_func` when loaded `grouped` data, for example when grouping
82 by solar day to avoid duplicate data from scene overlaps.
83 """
84 valid_bit = 8
85 valid_val = (1 << valid_bit)
86
87 no_data_dest_mask = ~(dest & valid_val).astype(bool)
88 np.copyto(dest, src, where=no_data_dest_mask)
89
90 both_data_mask = (valid_val & dest & src).astype(bool)
91 np.copyto(dest, src & dest, where=both_data_mask)
92
```
Path: `datacube/utils/xarray_geoextensions.py`
Content:
```
1 """
2 Add geometric extensions to :class:`xarray.Dataset` and :class:`xarray.DataArray` for use
3 with Data Cube by Monkey Patching those classes.
4
5 This extension is reliant on an `xarray` object having a `.crs` property of type
6 :class:`datacube.utils.geometry.CRS`. This is used to inspect the spatial dimensions of the
7 :class:`Dataset` or :class:`DataArray`, and provide new attributes for accessing a
8 :class:`datacube.utils.geometry.GeoBox`, affine transform and extent for the dataset as
9 `.geobox`, `.affine` and `.extent` respectively.
10
11 """
12
13 import xarray
14 from affine import Affine
15
16 from datacube.utils import data_resolution_and_offset, geometry
17
18
19 def _xarray_affine(obj):
20 dims = obj.crs.dimensions
21 xres, xoff = data_resolution_and_offset(obj[dims[1]].values)
22 yres, yoff = data_resolution_and_offset(obj[dims[0]].values)
23 return Affine.translation(xoff, yoff) * Affine.scale(xres, yres)
24
25
26 def _xarray_extent(obj):
27 return obj.geobox.extent
28
29
30 def _xarray_geobox(obj):
31 dims = obj.crs.dimensions
32 return geometry.GeoBox(obj[dims[1]].size, obj[dims[0]].size, obj.affine, obj.crs)
33
34
35 xarray.Dataset.geobox = property(_xarray_geobox)
36 xarray.Dataset.affine = property(_xarray_affine)
37 xarray.Dataset.extent = property(_xarray_extent)
38 xarray.DataArray.geobox = property(_xarray_geobox)
39 xarray.DataArray.affine = property(_xarray_affine)
40 xarray.DataArray.extent = property(_xarray_extent)
41
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/datacube/helpers.py b/datacube/helpers.py
--- a/datacube/helpers.py
+++ b/datacube/helpers.py
@@ -26,12 +26,17 @@
:param profile_override: option dict, overrides rasterio file creation options.
:param time_index: DEPRECATED
"""
- profile_override = profile_override or {}
-
if time_index is not None:
raise ValueError('''The write_geotiff function no longer supports passing in `time_index`.
The same function can be achieved by calling `dataset.isel(time=<time_index>)` before passing
- in your dataset. It was removed because it made the function much less useful for more advanced cases.''')
+ in your dataset. It was removed because it made the function much less useful for more advanced cases.''')
+
+ profile_override = profile_override or {}
+
+ geobox = getattr(dataset, 'geobox', None)
+
+ if geobox is None:
+ raise ValueError('Can only write datasets with specified `crs` attribute')
try:
dtypes = {val.dtype for val in dataset.data_vars.values()}
@@ -40,11 +45,13 @@
dtypes = [dataset.dtype]
profile = DEFAULT_PROFILE.copy()
+ height, width = geobox.shape
+
profile.update({
- 'width': dataset.dims[dataset.crs.dimensions[1]],
- 'height': dataset.dims[dataset.crs.dimensions[0]],
- 'transform': dataset.affine,
- 'crs': dataset.crs.crs_str,
+ 'width': width,
+ 'height': height,
+ 'transform': geobox.affine,
+ 'crs': geobox.crs.crs_str,
'count': len(dataset.data_vars),
'dtype': str(dtypes.pop())
})
diff --git a/datacube/utils/xarray_geoextensions.py b/datacube/utils/xarray_geoextensions.py
--- a/datacube/utils/xarray_geoextensions.py
+++ b/datacube/utils/xarray_geoextensions.py
@@ -16,20 +16,38 @@
from datacube.utils import data_resolution_and_offset, geometry
+def _norm_crs(crs):
+ if crs is None or isinstance(crs, geometry.CRS):
+ return crs
+ elif isinstance(crs, str):
+ return geometry.CRS(crs)
+ else:
+ raise ValueError('Can not interpret {} as CRS'.format(type(crs)))
+
+
def _xarray_affine(obj):
- dims = obj.crs.dimensions
+ crs = _norm_crs(obj.crs)
+ if crs is None:
+ return None
+
+ dims = crs.dimensions
xres, xoff = data_resolution_and_offset(obj[dims[1]].values)
yres, yoff = data_resolution_and_offset(obj[dims[0]].values)
return Affine.translation(xoff, yoff) * Affine.scale(xres, yres)
def _xarray_extent(obj):
- return obj.geobox.extent
+ geobox = obj.geobox
+ return None if geobox is None else geobox.extent
def _xarray_geobox(obj):
- dims = obj.crs.dimensions
- return geometry.GeoBox(obj[dims[1]].size, obj[dims[0]].size, obj.affine, obj.crs)
+ crs = _norm_crs(obj.crs)
+ if crs is None:
+ return None
+
+ dims = crs.dimensions
+ return geometry.GeoBox(obj[dims[1]].size, obj[dims[0]].size, obj.affine, crs)
xarray.Dataset.geobox = property(_xarray_geobox)
| {"golden_diff": "diff --git a/datacube/helpers.py b/datacube/helpers.py\n--- a/datacube/helpers.py\n+++ b/datacube/helpers.py\n@@ -26,12 +26,17 @@\n :param profile_override: option dict, overrides rasterio file creation options.\n :param time_index: DEPRECATED\n \"\"\"\n- profile_override = profile_override or {}\n-\n if time_index is not None:\n raise ValueError('''The write_geotiff function no longer supports passing in `time_index`.\n The same function can be achieved by calling `dataset.isel(time=<time_index>)` before passing\n- in your dataset. It was removed because it made the function much less useful for more advanced cases.''')\n+ in your dataset. It was removed because it made the function much less useful for more advanced cases.''')\n+\n+ profile_override = profile_override or {}\n+\n+ geobox = getattr(dataset, 'geobox', None)\n+\n+ if geobox is None:\n+ raise ValueError('Can only write datasets with specified `crs` attribute')\n \n try:\n dtypes = {val.dtype for val in dataset.data_vars.values()}\n@@ -40,11 +45,13 @@\n dtypes = [dataset.dtype]\n \n profile = DEFAULT_PROFILE.copy()\n+ height, width = geobox.shape\n+\n profile.update({\n- 'width': dataset.dims[dataset.crs.dimensions[1]],\n- 'height': dataset.dims[dataset.crs.dimensions[0]],\n- 'transform': dataset.affine,\n- 'crs': dataset.crs.crs_str,\n+ 'width': width,\n+ 'height': height,\n+ 'transform': geobox.affine,\n+ 'crs': geobox.crs.crs_str,\n 'count': len(dataset.data_vars),\n 'dtype': str(dtypes.pop())\n })\ndiff --git a/datacube/utils/xarray_geoextensions.py b/datacube/utils/xarray_geoextensions.py\n--- a/datacube/utils/xarray_geoextensions.py\n+++ b/datacube/utils/xarray_geoextensions.py\n@@ -16,20 +16,38 @@\n from datacube.utils import data_resolution_and_offset, geometry\n \n \n+def _norm_crs(crs):\n+ if crs is None or isinstance(crs, geometry.CRS):\n+ return crs\n+ elif isinstance(crs, str):\n+ return geometry.CRS(crs)\n+ else:\n+ raise ValueError('Can not interpret {} as CRS'.format(type(crs)))\n+\n+\n def _xarray_affine(obj):\n- dims = obj.crs.dimensions\n+ crs = _norm_crs(obj.crs)\n+ if crs is None:\n+ return None\n+\n+ dims = crs.dimensions\n xres, xoff = data_resolution_and_offset(obj[dims[1]].values)\n yres, yoff = data_resolution_and_offset(obj[dims[0]].values)\n return Affine.translation(xoff, yoff) * Affine.scale(xres, yres)\n \n \n def _xarray_extent(obj):\n- return obj.geobox.extent\n+ geobox = obj.geobox\n+ return None if geobox is None else geobox.extent\n \n \n def _xarray_geobox(obj):\n- dims = obj.crs.dimensions\n- return geometry.GeoBox(obj[dims[1]].size, obj[dims[0]].size, obj.affine, obj.crs)\n+ crs = _norm_crs(obj.crs)\n+ if crs is None:\n+ return None\n+\n+ dims = crs.dimensions\n+ return geometry.GeoBox(obj[dims[1]].size, obj[dims[0]].size, obj.affine, crs)\n \n \n xarray.Dataset.geobox = property(_xarray_geobox)\n", "issue": "write_geotiff helper function fails if CRS is string, not object\n### Expected behaviour\r\nThe write_geotiff helper function assumes that you will pass the function a datacube crs object. If you are writing out a geotiff from data that was not loaded using `dc.load`, this crs object is not present. \r\nE.g. I read a Geotiff file produced by datacube-stats into a Notebook, ran some analysis on it, and wanted to write it back out to Geotiff. I have a crs string from the attributes of my original Geotiff, but no datacube crs object, so the write_geotiff function fails.\r\n\r\n### Error\r\n```\r\nAttributeError Traceback (most recent call last)\r\n<ipython-input-41-736bab55bae5> in <module>()\r\n 3 Differenceds.attrs['crs'] = (GeotiffData.crs)\r\n 4 \r\n----> 5 write_geotiff(PercentileConfidence, Differenceds)\r\n\r\n/g/data/v10/public/modules/dea/20180515/lib/python3.6/site-packages/datacube/helpers.py in write_geotiff(filename, dataset, profile_override, time_index)\r\n 44 profile = DEFAULT_PROFILE.copy()\r\n 45 profile.update({\r\n---> 46 'width': dataset.dims[dataset.crs.dimensions[1]],\r\n 47 'height': dataset.dims[dataset.crs.dimensions[0]],\r\n 48 'transform': dataset.affine,\r\n\r\nAttributeError: 'str' object has no attribute 'dimensions'\r\n```\n", "before_files": [{"content": "\"\"\"\nUseful functions for Datacube users\n\nNot used internally, those should go in `utils.py`\n\"\"\"\n\nimport numpy as np\nimport rasterio\n\nDEFAULT_PROFILE = {\n 'blockxsize': 256,\n 'blockysize': 256,\n 'compress': 'lzw',\n 'driver': 'GTiff',\n 'interleave': 'band',\n 'nodata': 0.0,\n 'tiled': True}\n\n\ndef write_geotiff(filename, dataset, profile_override=None, time_index=None):\n \"\"\"\n Write an ODC style xarray.Dataset to a GeoTIFF file.\n\n :param filename: Output filename\n :param dataset: xarray dataset containing one or more bands to write to a file.\n :param profile_override: option dict, overrides rasterio file creation options.\n :param time_index: DEPRECATED\n \"\"\"\n profile_override = profile_override or {}\n\n if time_index is not None:\n raise ValueError('''The write_geotiff function no longer supports passing in `time_index`.\n The same function can be achieved by calling `dataset.isel(time=<time_index>)` before passing\n in your dataset. It was removed because it made the function much less useful for more advanced cases.''')\n\n try:\n dtypes = {val.dtype for val in dataset.data_vars.values()}\n assert len(dtypes) == 1 # Check for multiple dtypes\n except AttributeError:\n dtypes = [dataset.dtype]\n\n profile = DEFAULT_PROFILE.copy()\n profile.update({\n 'width': dataset.dims[dataset.crs.dimensions[1]],\n 'height': dataset.dims[dataset.crs.dimensions[0]],\n 'transform': dataset.affine,\n 'crs': dataset.crs.crs_str,\n 'count': len(dataset.data_vars),\n 'dtype': str(dtypes.pop())\n })\n profile.update(profile_override)\n\n _calculate_blocksize(profile)\n\n with rasterio.open(str(filename), 'w', **profile) as dest:\n if hasattr(dataset, 'data_vars'):\n for bandnum, data in enumerate(dataset.data_vars.values(), start=1):\n dest.write(data.data, bandnum)\n\n\ndef _calculate_blocksize(profile):\n # Block size must be smaller than the image size, and for geotiffs must be divisible by 16\n # Fix for small images.\n if profile['blockxsize'] > profile['width']:\n if profile['width'] % 16 == 0 or profile['width'] < 16:\n profile['blockxsize'] = profile['width']\n else:\n profile['blockxsize'] = 16\n\n if profile['blockysize'] > profile['height']:\n if profile['height'] % 16 == 0 or profile['height'] < 16:\n profile['blockysize'] = profile['height']\n else:\n profile['blockysize'] = 16\n\n\ndef ga_pq_fuser(dest, src):\n \"\"\"\n Fuse two Geoscience Australia Pixel Quality ndarrays\n\n To be used as a `fuse_func` when loaded `grouped` data, for example when grouping\n by solar day to avoid duplicate data from scene overlaps.\n \"\"\"\n valid_bit = 8\n valid_val = (1 << valid_bit)\n\n no_data_dest_mask = ~(dest & valid_val).astype(bool)\n np.copyto(dest, src, where=no_data_dest_mask)\n\n both_data_mask = (valid_val & dest & src).astype(bool)\n np.copyto(dest, src & dest, where=both_data_mask)\n", "path": "datacube/helpers.py"}, {"content": "\"\"\"\nAdd geometric extensions to :class:`xarray.Dataset` and :class:`xarray.DataArray` for use\nwith Data Cube by Monkey Patching those classes.\n\nThis extension is reliant on an `xarray` object having a `.crs` property of type\n:class:`datacube.utils.geometry.CRS`. This is used to inspect the spatial dimensions of the\n:class:`Dataset` or :class:`DataArray`, and provide new attributes for accessing a\n:class:`datacube.utils.geometry.GeoBox`, affine transform and extent for the dataset as\n`.geobox`, `.affine` and `.extent` respectively.\n\n\"\"\"\n\nimport xarray\nfrom affine import Affine\n\nfrom datacube.utils import data_resolution_and_offset, geometry\n\n\ndef _xarray_affine(obj):\n dims = obj.crs.dimensions\n xres, xoff = data_resolution_and_offset(obj[dims[1]].values)\n yres, yoff = data_resolution_and_offset(obj[dims[0]].values)\n return Affine.translation(xoff, yoff) * Affine.scale(xres, yres)\n\n\ndef _xarray_extent(obj):\n return obj.geobox.extent\n\n\ndef _xarray_geobox(obj):\n dims = obj.crs.dimensions\n return geometry.GeoBox(obj[dims[1]].size, obj[dims[0]].size, obj.affine, obj.crs)\n\n\nxarray.Dataset.geobox = property(_xarray_geobox)\nxarray.Dataset.affine = property(_xarray_affine)\nxarray.Dataset.extent = property(_xarray_extent)\nxarray.DataArray.geobox = property(_xarray_geobox)\nxarray.DataArray.affine = property(_xarray_affine)\nxarray.DataArray.extent = property(_xarray_extent)\n", "path": "datacube/utils/xarray_geoextensions.py"}], "after_files": [{"content": "\"\"\"\nUseful functions for Datacube users\n\nNot used internally, those should go in `utils.py`\n\"\"\"\n\nimport numpy as np\nimport rasterio\n\nDEFAULT_PROFILE = {\n 'blockxsize': 256,\n 'blockysize': 256,\n 'compress': 'lzw',\n 'driver': 'GTiff',\n 'interleave': 'band',\n 'nodata': 0.0,\n 'tiled': True}\n\n\ndef write_geotiff(filename, dataset, profile_override=None, time_index=None):\n \"\"\"\n Write an ODC style xarray.Dataset to a GeoTIFF file.\n\n :param filename: Output filename\n :param dataset: xarray dataset containing one or more bands to write to a file.\n :param profile_override: option dict, overrides rasterio file creation options.\n :param time_index: DEPRECATED\n \"\"\"\n if time_index is not None:\n raise ValueError('''The write_geotiff function no longer supports passing in `time_index`.\n The same function can be achieved by calling `dataset.isel(time=<time_index>)` before passing\n in your dataset. It was removed because it made the function much less useful for more advanced cases.''')\n\n profile_override = profile_override or {}\n\n geobox = getattr(dataset, 'geobox', None)\n\n if geobox is None:\n raise ValueError('Can only write datasets with specified `crs` attribute')\n\n try:\n dtypes = {val.dtype for val in dataset.data_vars.values()}\n assert len(dtypes) == 1 # Check for multiple dtypes\n except AttributeError:\n dtypes = [dataset.dtype]\n\n profile = DEFAULT_PROFILE.copy()\n height, width = geobox.shape\n\n profile.update({\n 'width': width,\n 'height': height,\n 'transform': geobox.affine,\n 'crs': geobox.crs.crs_str,\n 'count': len(dataset.data_vars),\n 'dtype': str(dtypes.pop())\n })\n profile.update(profile_override)\n\n _calculate_blocksize(profile)\n\n with rasterio.open(str(filename), 'w', **profile) as dest:\n if hasattr(dataset, 'data_vars'):\n for bandnum, data in enumerate(dataset.data_vars.values(), start=1):\n dest.write(data.data, bandnum)\n\n\ndef _calculate_blocksize(profile):\n # Block size must be smaller than the image size, and for geotiffs must be divisible by 16\n # Fix for small images.\n if profile['blockxsize'] > profile['width']:\n if profile['width'] % 16 == 0 or profile['width'] < 16:\n profile['blockxsize'] = profile['width']\n else:\n profile['blockxsize'] = 16\n\n if profile['blockysize'] > profile['height']:\n if profile['height'] % 16 == 0 or profile['height'] < 16:\n profile['blockysize'] = profile['height']\n else:\n profile['blockysize'] = 16\n\n\ndef ga_pq_fuser(dest, src):\n \"\"\"\n Fuse two Geoscience Australia Pixel Quality ndarrays\n\n To be used as a `fuse_func` when loaded `grouped` data, for example when grouping\n by solar day to avoid duplicate data from scene overlaps.\n \"\"\"\n valid_bit = 8\n valid_val = (1 << valid_bit)\n\n no_data_dest_mask = ~(dest & valid_val).astype(bool)\n np.copyto(dest, src, where=no_data_dest_mask)\n\n both_data_mask = (valid_val & dest & src).astype(bool)\n np.copyto(dest, src & dest, where=both_data_mask)\n", "path": "datacube/helpers.py"}, {"content": "\"\"\"\nAdd geometric extensions to :class:`xarray.Dataset` and :class:`xarray.DataArray` for use\nwith Data Cube by Monkey Patching those classes.\n\nThis extension is reliant on an `xarray` object having a `.crs` property of type\n:class:`datacube.utils.geometry.CRS`. This is used to inspect the spatial dimensions of the\n:class:`Dataset` or :class:`DataArray`, and provide new attributes for accessing a\n:class:`datacube.utils.geometry.GeoBox`, affine transform and extent for the dataset as\n`.geobox`, `.affine` and `.extent` respectively.\n\n\"\"\"\n\nimport xarray\nfrom affine import Affine\n\nfrom datacube.utils import data_resolution_and_offset, geometry\n\n\ndef _norm_crs(crs):\n if crs is None or isinstance(crs, geometry.CRS):\n return crs\n elif isinstance(crs, str):\n return geometry.CRS(crs)\n else:\n raise ValueError('Can not interpret {} as CRS'.format(type(crs)))\n\n\ndef _xarray_affine(obj):\n crs = _norm_crs(obj.crs)\n if crs is None:\n return None\n\n dims = crs.dimensions\n xres, xoff = data_resolution_and_offset(obj[dims[1]].values)\n yres, yoff = data_resolution_and_offset(obj[dims[0]].values)\n return Affine.translation(xoff, yoff) * Affine.scale(xres, yres)\n\n\ndef _xarray_extent(obj):\n geobox = obj.geobox\n return None if geobox is None else geobox.extent\n\n\ndef _xarray_geobox(obj):\n crs = _norm_crs(obj.crs)\n if crs is None:\n return None\n\n dims = crs.dimensions\n return geometry.GeoBox(obj[dims[1]].size, obj[dims[0]].size, obj.affine, crs)\n\n\nxarray.Dataset.geobox = property(_xarray_geobox)\nxarray.Dataset.affine = property(_xarray_affine)\nxarray.Dataset.extent = property(_xarray_extent)\nxarray.DataArray.geobox = property(_xarray_geobox)\nxarray.DataArray.affine = property(_xarray_affine)\nxarray.DataArray.extent = property(_xarray_extent)\n", "path": "datacube/utils/xarray_geoextensions.py"}]} | 2,037 | 825 |
gh_patches_debug_12402 | rasdani/github-patches | git_diff | deepset-ai__haystack-2973 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Code snippet wrongly formatted in docs
**Describe the bug**
The code snippet at:
https://github.com/deepset-ai/haystack/blob/master/haystack/nodes/ranker/sentence_transformers.py#L32
is rendered as plain text in the corresponding documentation page
https://haystack.deepset.ai/reference/ranker
**Expected behavior**
The snipped should be formatted as Python code
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `haystack/nodes/ranker/sentence_transformers.py`
Content:
```
1 from typing import List, Optional, Union, Tuple, Iterator, Any
2 import logging
3 from pathlib import Path
4
5 import torch
6 from torch.nn import DataParallel
7 from transformers import AutoModelForSequenceClassification, AutoTokenizer
8
9 from haystack.errors import HaystackError
10 from haystack.schema import Document
11 from haystack.nodes.ranker.base import BaseRanker
12 from haystack.modeling.utils import initialize_device_settings
13
14 logger = logging.getLogger(__name__)
15
16
17 class SentenceTransformersRanker(BaseRanker):
18 """
19 Sentence Transformer based pre-trained Cross-Encoder model for Document Re-ranking (https://huggingface.co/cross-encoder).
20 Re-Ranking can be used on top of a retriever to boost the performance for document search. This is particularly useful if the retriever has a high recall but is bad in sorting the documents by relevance.
21
22 SentenceTransformerRanker handles Cross-Encoder models
23 - use a single logit as similarity score e.g. cross-encoder/ms-marco-MiniLM-L-12-v2
24 - use two output logits (no_answer, has_answer) e.g. deepset/gbert-base-germandpr-reranking
25 https://www.sbert.net/docs/pretrained-models/ce-msmarco.html#usage-with-transformers
26
27 | With a SentenceTransformersRanker, you can:
28 - directly get predictions via predict()
29
30 Usage example:
31
32 ```python
33 | retriever = BM25Retriever(document_store=document_store)
34 | ranker = SentenceTransformersRanker(model_name_or_path="cross-encoder/ms-marco-MiniLM-L-12-v2")
35 | p = Pipeline()
36 | p.add_node(component=retriever, name="ESRetriever", inputs=["Query"])
37 | p.add_node(component=ranker, name="Ranker", inputs=["ESRetriever"])
38 ```
39 """
40
41 def __init__(
42 self,
43 model_name_or_path: Union[str, Path],
44 model_version: Optional[str] = None,
45 top_k: int = 10,
46 use_gpu: bool = True,
47 devices: Optional[List[Union[str, torch.device]]] = None,
48 batch_size: int = 16,
49 scale_score: bool = True,
50 ):
51 """
52 :param model_name_or_path: Directory of a saved model or the name of a public model e.g.
53 'cross-encoder/ms-marco-MiniLM-L-12-v2'.
54 See https://huggingface.co/cross-encoder for full list of available models
55 :param model_version: The version of model to use from the HuggingFace model hub. Can be tag name, branch name, or commit hash.
56 :param top_k: The maximum number of documents to return
57 :param use_gpu: Whether to use all available GPUs or the CPU. Falls back on CPU if no GPU is available.
58 :param devices: List of GPU (or CPU) devices, to limit inference to certain GPUs and not use all available ones
59 The strings will be converted into pytorch devices, so use the string notation described here:
60 https://pytorch.org/docs/stable/tensor_attributes.html?highlight=torch%20device#torch.torch.device
61 (e.g. ["cuda:0"]).
62 :param batch_size: Number of documents to process at a time.
63 :param scale_score: The raw predictions will be transformed using a Sigmoid activation function in case the model
64 only predicts a single label. For multi-label predictions, no scaling is applied. Set this
65 to False if you do not want any scaling of the raw predictions.
66 """
67 super().__init__()
68
69 self.top_k = top_k
70
71 if devices is not None:
72 self.devices = [torch.device(device) for device in devices]
73 else:
74 self.devices, _ = initialize_device_settings(use_cuda=use_gpu, multi_gpu=True)
75
76 self.transformer_model = AutoModelForSequenceClassification.from_pretrained(
77 pretrained_model_name_or_path=model_name_or_path, revision=model_version
78 )
79 self.transformer_model.to(str(self.devices[0]))
80 self.transformer_tokenizer = AutoTokenizer.from_pretrained(
81 pretrained_model_name_or_path=model_name_or_path, revision=model_version
82 )
83 self.transformer_model.eval()
84
85 # we use sigmoid activation function to scale the score in case there is only a single label
86 # we do not apply any scaling when scale_score is set to False
87 num_labels = self.transformer_model.num_labels
88 self.activation_function: torch.nn.Module
89 if num_labels == 1 and scale_score:
90 self.activation_function = torch.nn.Sigmoid()
91 else:
92 self.activation_function = torch.nn.Identity()
93
94 if len(self.devices) > 1:
95 self.model = DataParallel(self.transformer_model, device_ids=self.devices)
96
97 self.batch_size = batch_size
98
99 def predict(self, query: str, documents: List[Document], top_k: Optional[int] = None) -> List[Document]:
100 """
101 Use loaded ranker model to re-rank the supplied list of Document.
102
103 Returns list of Document sorted by (desc.) similarity with the query.
104
105 :param query: Query string
106 :param documents: List of Document to be re-ranked
107 :param top_k: The maximum number of documents to return
108 :return: List of Document
109 """
110 if top_k is None:
111 top_k = self.top_k
112
113 features = self.transformer_tokenizer(
114 [query for doc in documents],
115 [doc.content for doc in documents],
116 padding=True,
117 truncation=True,
118 return_tensors="pt",
119 ).to(self.devices[0])
120
121 # SentenceTransformerRanker uses:
122 # 1. the logit as similarity score/answerable classification
123 # 2. the logits as answerable classification (no_answer / has_answer)
124 # https://www.sbert.net/docs/pretrained-models/ce-msmarco.html#usage-with-transformers
125 with torch.no_grad():
126 similarity_scores = self.transformer_model(**features).logits
127
128 logits_dim = similarity_scores.shape[1] # [batch_size, logits_dim]
129 sorted_scores_and_documents = sorted(
130 zip(similarity_scores, documents),
131 key=lambda similarity_document_tuple:
132 # assume the last element in logits represents the `has_answer` label
133 similarity_document_tuple[0][-1] if logits_dim >= 2 else similarity_document_tuple[0],
134 reverse=True,
135 )
136
137 # add normalized scores to documents
138 sorted_documents = self._add_scores_to_documents(sorted_scores_and_documents[:top_k], logits_dim)
139
140 return sorted_documents
141
142 def _add_scores_to_documents(
143 self, sorted_scores_and_documents: List[Tuple[Any, Document]], logits_dim: int
144 ) -> List[Document]:
145 """
146 Normalize and add scores to retrieved result documents.
147
148 :param sorted_scores_and_documents: List of score, Document Tuples.
149 :param logits_dim: Dimensionality of the returned scores.
150 """
151 sorted_documents = []
152 for raw_score, doc in sorted_scores_and_documents:
153 if logits_dim >= 2:
154 score = self.activation_function(raw_score)[-1]
155 else:
156 score = self.activation_function(raw_score)[0]
157
158 doc.score = score.detach().cpu().numpy().tolist()
159 sorted_documents.append(doc)
160
161 return sorted_documents
162
163 def predict_batch(
164 self,
165 queries: List[str],
166 documents: Union[List[Document], List[List[Document]]],
167 top_k: Optional[int] = None,
168 batch_size: Optional[int] = None,
169 ) -> Union[List[Document], List[List[Document]]]:
170 """
171 Use loaded ranker model to re-rank the supplied lists of Documents.
172
173 Returns lists of Documents sorted by (desc.) similarity with the corresponding queries.
174
175
176 - If you provide a list containing a single query...
177
178 - ... and a single list of Documents, the single list of Documents will be re-ranked based on the
179 supplied query.
180 - ... and a list of lists of Documents, each list of Documents will be re-ranked individually based on the
181 supplied query.
182
183
184 - If you provide a list of multiple queries...
185
186 - ... you need to provide a list of lists of Documents. Each list of Documents will be re-ranked based on
187 its corresponding query.
188
189 :param queries: Single query string or list of queries
190 :param documents: Single list of Documents or list of lists of Documents to be reranked.
191 :param top_k: The maximum number of documents to return per Document list.
192 :param batch_size: Number of Documents to process at a time.
193 """
194 if top_k is None:
195 top_k = self.top_k
196
197 if batch_size is None:
198 batch_size = self.batch_size
199
200 number_of_docs, all_queries, all_docs, single_list_of_docs = self._preprocess_batch_queries_and_docs(
201 queries=queries, documents=documents
202 )
203
204 batches = self._get_batches(all_queries=all_queries, all_docs=all_docs, batch_size=batch_size)
205 preds = []
206 for cur_queries, cur_docs in batches:
207 features = self.transformer_tokenizer(
208 cur_queries, [doc.content for doc in cur_docs], padding=True, truncation=True, return_tensors="pt"
209 ).to(self.devices[0])
210
211 with torch.no_grad():
212 similarity_scores = self.transformer_model(**features).logits
213 preds.extend(similarity_scores)
214
215 logits_dim = similarity_scores.shape[1] # [batch_size, logits_dim]
216 if single_list_of_docs:
217 sorted_scores_and_documents = sorted(
218 zip(similarity_scores, documents),
219 key=lambda similarity_document_tuple:
220 # assume the last element in logits represents the `has_answer` label
221 similarity_document_tuple[0][-1] if logits_dim >= 2 else similarity_document_tuple[0],
222 reverse=True,
223 )
224
225 # is this step needed?
226 sorted_documents = [(score, doc) for score, doc in sorted_scores_and_documents if isinstance(doc, Document)]
227 sorted_documents_with_scores = self._add_scores_to_documents(sorted_documents[:top_k], logits_dim)
228
229 return sorted_documents_with_scores
230 else:
231 # Group predictions together
232 grouped_predictions = []
233 left_idx = 0
234 right_idx = 0
235 for number in number_of_docs:
236 right_idx = left_idx + number
237 grouped_predictions.append(similarity_scores[left_idx:right_idx])
238 left_idx = right_idx
239
240 result = []
241 for pred_group, doc_group in zip(grouped_predictions, documents):
242 sorted_scores_and_documents = sorted(
243 zip(pred_group, doc_group), # type: ignore
244 key=lambda similarity_document_tuple:
245 # assume the last element in logits represents the `has_answer` label
246 similarity_document_tuple[0][-1] if logits_dim >= 2 else similarity_document_tuple[0],
247 reverse=True,
248 )
249
250 # rank documents according to scores
251 sorted_documents = [
252 (score, doc) for score, doc in sorted_scores_and_documents if isinstance(doc, Document)
253 ]
254 sorted_documents_with_scores = self._add_scores_to_documents(sorted_documents[:top_k], logits_dim)
255
256 result.append(sorted_documents_with_scores)
257
258 return result
259
260 def _preprocess_batch_queries_and_docs(
261 self, queries: List[str], documents: Union[List[Document], List[List[Document]]]
262 ) -> Tuple[List[int], List[str], List[Document], bool]:
263 number_of_docs = []
264 all_queries = []
265 all_docs: List[Document] = []
266 single_list_of_docs = False
267
268 # Docs case 1: single list of Documents -> rerank single list of Documents based on single query
269 if len(documents) > 0 and isinstance(documents[0], Document):
270 if len(queries) != 1:
271 raise HaystackError("Number of queries must be 1 if a single list of Documents is provided.")
272 query = queries[0]
273 number_of_docs = [len(documents)]
274 all_queries = [query] * len(documents)
275 all_docs = documents # type: ignore
276 single_list_of_docs = True
277
278 # Docs case 2: list of lists of Documents -> rerank each list of Documents based on corresponding query
279 # If queries contains a single query, apply it to each list of Documents
280 if len(documents) > 0 and isinstance(documents[0], list):
281 if len(queries) == 1:
282 queries = queries * len(documents)
283 if len(queries) != len(documents):
284 raise HaystackError("Number of queries must be equal to number of provided Document lists.")
285 for query, cur_docs in zip(queries, documents):
286 if not isinstance(cur_docs, list):
287 raise HaystackError(f"cur_docs was of type {type(cur_docs)}, but expected a list of Documents.")
288 number_of_docs.append(len(cur_docs))
289 all_queries.extend([query] * len(cur_docs))
290 all_docs.extend(cur_docs)
291
292 return number_of_docs, all_queries, all_docs, single_list_of_docs
293
294 @staticmethod
295 def _get_batches(
296 all_queries: List[str], all_docs: List[Document], batch_size: Optional[int]
297 ) -> Iterator[Tuple[List[str], List[Document]]]:
298 if batch_size is None:
299 yield all_queries, all_docs
300 return
301 else:
302 for index in range(0, len(all_queries), batch_size):
303 yield all_queries[index : index + batch_size], all_docs[index : index + batch_size]
304
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/haystack/nodes/ranker/sentence_transformers.py b/haystack/nodes/ranker/sentence_transformers.py
--- a/haystack/nodes/ranker/sentence_transformers.py
+++ b/haystack/nodes/ranker/sentence_transformers.py
@@ -30,11 +30,11 @@
Usage example:
```python
- | retriever = BM25Retriever(document_store=document_store)
- | ranker = SentenceTransformersRanker(model_name_or_path="cross-encoder/ms-marco-MiniLM-L-12-v2")
- | p = Pipeline()
- | p.add_node(component=retriever, name="ESRetriever", inputs=["Query"])
- | p.add_node(component=ranker, name="Ranker", inputs=["ESRetriever"])
+ retriever = BM25Retriever(document_store=document_store)
+ ranker = SentenceTransformersRanker(model_name_or_path="cross-encoder/ms-marco-MiniLM-L-12-v2")
+ p = Pipeline()
+ p.add_node(component=retriever, name="ESRetriever", inputs=["Query"])
+ p.add_node(component=ranker, name="Ranker", inputs=["ESRetriever"])
```
"""
| {"golden_diff": "diff --git a/haystack/nodes/ranker/sentence_transformers.py b/haystack/nodes/ranker/sentence_transformers.py\n--- a/haystack/nodes/ranker/sentence_transformers.py\n+++ b/haystack/nodes/ranker/sentence_transformers.py\n@@ -30,11 +30,11 @@\n Usage example:\n \n ```python\n- | retriever = BM25Retriever(document_store=document_store)\n- | ranker = SentenceTransformersRanker(model_name_or_path=\"cross-encoder/ms-marco-MiniLM-L-12-v2\")\n- | p = Pipeline()\n- | p.add_node(component=retriever, name=\"ESRetriever\", inputs=[\"Query\"])\n- | p.add_node(component=ranker, name=\"Ranker\", inputs=[\"ESRetriever\"])\n+ retriever = BM25Retriever(document_store=document_store)\n+ ranker = SentenceTransformersRanker(model_name_or_path=\"cross-encoder/ms-marco-MiniLM-L-12-v2\")\n+ p = Pipeline()\n+ p.add_node(component=retriever, name=\"ESRetriever\", inputs=[\"Query\"])\n+ p.add_node(component=ranker, name=\"Ranker\", inputs=[\"ESRetriever\"])\n ```\n \"\"\"\n", "issue": "Code snippet wrongly formatted in docs\n**Describe the bug**\r\nThe code snippet at:\r\n\r\nhttps://github.com/deepset-ai/haystack/blob/master/haystack/nodes/ranker/sentence_transformers.py#L32\r\n\r\nis rendered as plain text in the corresponding documentation page\r\n\r\nhttps://haystack.deepset.ai/reference/ranker\r\n\r\n\r\n**Expected behavior**\r\nThe snipped should be formatted as Python code\r\n\n", "before_files": [{"content": "from typing import List, Optional, Union, Tuple, Iterator, Any\nimport logging\nfrom pathlib import Path\n\nimport torch\nfrom torch.nn import DataParallel\nfrom transformers import AutoModelForSequenceClassification, AutoTokenizer\n\nfrom haystack.errors import HaystackError\nfrom haystack.schema import Document\nfrom haystack.nodes.ranker.base import BaseRanker\nfrom haystack.modeling.utils import initialize_device_settings\n\nlogger = logging.getLogger(__name__)\n\n\nclass SentenceTransformersRanker(BaseRanker):\n \"\"\"\n Sentence Transformer based pre-trained Cross-Encoder model for Document Re-ranking (https://huggingface.co/cross-encoder).\n Re-Ranking can be used on top of a retriever to boost the performance for document search. This is particularly useful if the retriever has a high recall but is bad in sorting the documents by relevance.\n\n SentenceTransformerRanker handles Cross-Encoder models\n - use a single logit as similarity score e.g. cross-encoder/ms-marco-MiniLM-L-12-v2\n - use two output logits (no_answer, has_answer) e.g. deepset/gbert-base-germandpr-reranking\n https://www.sbert.net/docs/pretrained-models/ce-msmarco.html#usage-with-transformers\n\n | With a SentenceTransformersRanker, you can:\n - directly get predictions via predict()\n\n Usage example:\n\n ```python\n | retriever = BM25Retriever(document_store=document_store)\n | ranker = SentenceTransformersRanker(model_name_or_path=\"cross-encoder/ms-marco-MiniLM-L-12-v2\")\n | p = Pipeline()\n | p.add_node(component=retriever, name=\"ESRetriever\", inputs=[\"Query\"])\n | p.add_node(component=ranker, name=\"Ranker\", inputs=[\"ESRetriever\"])\n ```\n \"\"\"\n\n def __init__(\n self,\n model_name_or_path: Union[str, Path],\n model_version: Optional[str] = None,\n top_k: int = 10,\n use_gpu: bool = True,\n devices: Optional[List[Union[str, torch.device]]] = None,\n batch_size: int = 16,\n scale_score: bool = True,\n ):\n \"\"\"\n :param model_name_or_path: Directory of a saved model or the name of a public model e.g.\n 'cross-encoder/ms-marco-MiniLM-L-12-v2'.\n See https://huggingface.co/cross-encoder for full list of available models\n :param model_version: The version of model to use from the HuggingFace model hub. Can be tag name, branch name, or commit hash.\n :param top_k: The maximum number of documents to return\n :param use_gpu: Whether to use all available GPUs or the CPU. Falls back on CPU if no GPU is available.\n :param devices: List of GPU (or CPU) devices, to limit inference to certain GPUs and not use all available ones\n The strings will be converted into pytorch devices, so use the string notation described here:\n https://pytorch.org/docs/stable/tensor_attributes.html?highlight=torch%20device#torch.torch.device\n (e.g. [\"cuda:0\"]).\n :param batch_size: Number of documents to process at a time.\n :param scale_score: The raw predictions will be transformed using a Sigmoid activation function in case the model\n only predicts a single label. For multi-label predictions, no scaling is applied. Set this\n to False if you do not want any scaling of the raw predictions.\n \"\"\"\n super().__init__()\n\n self.top_k = top_k\n\n if devices is not None:\n self.devices = [torch.device(device) for device in devices]\n else:\n self.devices, _ = initialize_device_settings(use_cuda=use_gpu, multi_gpu=True)\n\n self.transformer_model = AutoModelForSequenceClassification.from_pretrained(\n pretrained_model_name_or_path=model_name_or_path, revision=model_version\n )\n self.transformer_model.to(str(self.devices[0]))\n self.transformer_tokenizer = AutoTokenizer.from_pretrained(\n pretrained_model_name_or_path=model_name_or_path, revision=model_version\n )\n self.transformer_model.eval()\n\n # we use sigmoid activation function to scale the score in case there is only a single label\n # we do not apply any scaling when scale_score is set to False\n num_labels = self.transformer_model.num_labels\n self.activation_function: torch.nn.Module\n if num_labels == 1 and scale_score:\n self.activation_function = torch.nn.Sigmoid()\n else:\n self.activation_function = torch.nn.Identity()\n\n if len(self.devices) > 1:\n self.model = DataParallel(self.transformer_model, device_ids=self.devices)\n\n self.batch_size = batch_size\n\n def predict(self, query: str, documents: List[Document], top_k: Optional[int] = None) -> List[Document]:\n \"\"\"\n Use loaded ranker model to re-rank the supplied list of Document.\n\n Returns list of Document sorted by (desc.) similarity with the query.\n\n :param query: Query string\n :param documents: List of Document to be re-ranked\n :param top_k: The maximum number of documents to return\n :return: List of Document\n \"\"\"\n if top_k is None:\n top_k = self.top_k\n\n features = self.transformer_tokenizer(\n [query for doc in documents],\n [doc.content for doc in documents],\n padding=True,\n truncation=True,\n return_tensors=\"pt\",\n ).to(self.devices[0])\n\n # SentenceTransformerRanker uses:\n # 1. the logit as similarity score/answerable classification\n # 2. the logits as answerable classification (no_answer / has_answer)\n # https://www.sbert.net/docs/pretrained-models/ce-msmarco.html#usage-with-transformers\n with torch.no_grad():\n similarity_scores = self.transformer_model(**features).logits\n\n logits_dim = similarity_scores.shape[1] # [batch_size, logits_dim]\n sorted_scores_and_documents = sorted(\n zip(similarity_scores, documents),\n key=lambda similarity_document_tuple:\n # assume the last element in logits represents the `has_answer` label\n similarity_document_tuple[0][-1] if logits_dim >= 2 else similarity_document_tuple[0],\n reverse=True,\n )\n\n # add normalized scores to documents\n sorted_documents = self._add_scores_to_documents(sorted_scores_and_documents[:top_k], logits_dim)\n\n return sorted_documents\n\n def _add_scores_to_documents(\n self, sorted_scores_and_documents: List[Tuple[Any, Document]], logits_dim: int\n ) -> List[Document]:\n \"\"\"\n Normalize and add scores to retrieved result documents.\n\n :param sorted_scores_and_documents: List of score, Document Tuples.\n :param logits_dim: Dimensionality of the returned scores.\n \"\"\"\n sorted_documents = []\n for raw_score, doc in sorted_scores_and_documents:\n if logits_dim >= 2:\n score = self.activation_function(raw_score)[-1]\n else:\n score = self.activation_function(raw_score)[0]\n\n doc.score = score.detach().cpu().numpy().tolist()\n sorted_documents.append(doc)\n\n return sorted_documents\n\n def predict_batch(\n self,\n queries: List[str],\n documents: Union[List[Document], List[List[Document]]],\n top_k: Optional[int] = None,\n batch_size: Optional[int] = None,\n ) -> Union[List[Document], List[List[Document]]]:\n \"\"\"\n Use loaded ranker model to re-rank the supplied lists of Documents.\n\n Returns lists of Documents sorted by (desc.) similarity with the corresponding queries.\n\n\n - If you provide a list containing a single query...\n\n - ... and a single list of Documents, the single list of Documents will be re-ranked based on the\n supplied query.\n - ... and a list of lists of Documents, each list of Documents will be re-ranked individually based on the\n supplied query.\n\n\n - If you provide a list of multiple queries...\n\n - ... you need to provide a list of lists of Documents. Each list of Documents will be re-ranked based on\n its corresponding query.\n\n :param queries: Single query string or list of queries\n :param documents: Single list of Documents or list of lists of Documents to be reranked.\n :param top_k: The maximum number of documents to return per Document list.\n :param batch_size: Number of Documents to process at a time.\n \"\"\"\n if top_k is None:\n top_k = self.top_k\n\n if batch_size is None:\n batch_size = self.batch_size\n\n number_of_docs, all_queries, all_docs, single_list_of_docs = self._preprocess_batch_queries_and_docs(\n queries=queries, documents=documents\n )\n\n batches = self._get_batches(all_queries=all_queries, all_docs=all_docs, batch_size=batch_size)\n preds = []\n for cur_queries, cur_docs in batches:\n features = self.transformer_tokenizer(\n cur_queries, [doc.content for doc in cur_docs], padding=True, truncation=True, return_tensors=\"pt\"\n ).to(self.devices[0])\n\n with torch.no_grad():\n similarity_scores = self.transformer_model(**features).logits\n preds.extend(similarity_scores)\n\n logits_dim = similarity_scores.shape[1] # [batch_size, logits_dim]\n if single_list_of_docs:\n sorted_scores_and_documents = sorted(\n zip(similarity_scores, documents),\n key=lambda similarity_document_tuple:\n # assume the last element in logits represents the `has_answer` label\n similarity_document_tuple[0][-1] if logits_dim >= 2 else similarity_document_tuple[0],\n reverse=True,\n )\n\n # is this step needed?\n sorted_documents = [(score, doc) for score, doc in sorted_scores_and_documents if isinstance(doc, Document)]\n sorted_documents_with_scores = self._add_scores_to_documents(sorted_documents[:top_k], logits_dim)\n\n return sorted_documents_with_scores\n else:\n # Group predictions together\n grouped_predictions = []\n left_idx = 0\n right_idx = 0\n for number in number_of_docs:\n right_idx = left_idx + number\n grouped_predictions.append(similarity_scores[left_idx:right_idx])\n left_idx = right_idx\n\n result = []\n for pred_group, doc_group in zip(grouped_predictions, documents):\n sorted_scores_and_documents = sorted(\n zip(pred_group, doc_group), # type: ignore\n key=lambda similarity_document_tuple:\n # assume the last element in logits represents the `has_answer` label\n similarity_document_tuple[0][-1] if logits_dim >= 2 else similarity_document_tuple[0],\n reverse=True,\n )\n\n # rank documents according to scores\n sorted_documents = [\n (score, doc) for score, doc in sorted_scores_and_documents if isinstance(doc, Document)\n ]\n sorted_documents_with_scores = self._add_scores_to_documents(sorted_documents[:top_k], logits_dim)\n\n result.append(sorted_documents_with_scores)\n\n return result\n\n def _preprocess_batch_queries_and_docs(\n self, queries: List[str], documents: Union[List[Document], List[List[Document]]]\n ) -> Tuple[List[int], List[str], List[Document], bool]:\n number_of_docs = []\n all_queries = []\n all_docs: List[Document] = []\n single_list_of_docs = False\n\n # Docs case 1: single list of Documents -> rerank single list of Documents based on single query\n if len(documents) > 0 and isinstance(documents[0], Document):\n if len(queries) != 1:\n raise HaystackError(\"Number of queries must be 1 if a single list of Documents is provided.\")\n query = queries[0]\n number_of_docs = [len(documents)]\n all_queries = [query] * len(documents)\n all_docs = documents # type: ignore\n single_list_of_docs = True\n\n # Docs case 2: list of lists of Documents -> rerank each list of Documents based on corresponding query\n # If queries contains a single query, apply it to each list of Documents\n if len(documents) > 0 and isinstance(documents[0], list):\n if len(queries) == 1:\n queries = queries * len(documents)\n if len(queries) != len(documents):\n raise HaystackError(\"Number of queries must be equal to number of provided Document lists.\")\n for query, cur_docs in zip(queries, documents):\n if not isinstance(cur_docs, list):\n raise HaystackError(f\"cur_docs was of type {type(cur_docs)}, but expected a list of Documents.\")\n number_of_docs.append(len(cur_docs))\n all_queries.extend([query] * len(cur_docs))\n all_docs.extend(cur_docs)\n\n return number_of_docs, all_queries, all_docs, single_list_of_docs\n\n @staticmethod\n def _get_batches(\n all_queries: List[str], all_docs: List[Document], batch_size: Optional[int]\n ) -> Iterator[Tuple[List[str], List[Document]]]:\n if batch_size is None:\n yield all_queries, all_docs\n return\n else:\n for index in range(0, len(all_queries), batch_size):\n yield all_queries[index : index + batch_size], all_docs[index : index + batch_size]\n", "path": "haystack/nodes/ranker/sentence_transformers.py"}], "after_files": [{"content": "from typing import List, Optional, Union, Tuple, Iterator, Any\nimport logging\nfrom pathlib import Path\n\nimport torch\nfrom torch.nn import DataParallel\nfrom transformers import AutoModelForSequenceClassification, AutoTokenizer\n\nfrom haystack.errors import HaystackError\nfrom haystack.schema import Document\nfrom haystack.nodes.ranker.base import BaseRanker\nfrom haystack.modeling.utils import initialize_device_settings\n\nlogger = logging.getLogger(__name__)\n\n\nclass SentenceTransformersRanker(BaseRanker):\n \"\"\"\n Sentence Transformer based pre-trained Cross-Encoder model for Document Re-ranking (https://huggingface.co/cross-encoder).\n Re-Ranking can be used on top of a retriever to boost the performance for document search. This is particularly useful if the retriever has a high recall but is bad in sorting the documents by relevance.\n\n SentenceTransformerRanker handles Cross-Encoder models\n - use a single logit as similarity score e.g. cross-encoder/ms-marco-MiniLM-L-12-v2\n - use two output logits (no_answer, has_answer) e.g. deepset/gbert-base-germandpr-reranking\n https://www.sbert.net/docs/pretrained-models/ce-msmarco.html#usage-with-transformers\n\n | With a SentenceTransformersRanker, you can:\n - directly get predictions via predict()\n\n Usage example:\n\n ```python\n retriever = BM25Retriever(document_store=document_store)\n ranker = SentenceTransformersRanker(model_name_or_path=\"cross-encoder/ms-marco-MiniLM-L-12-v2\")\n p = Pipeline()\n p.add_node(component=retriever, name=\"ESRetriever\", inputs=[\"Query\"])\n p.add_node(component=ranker, name=\"Ranker\", inputs=[\"ESRetriever\"])\n ```\n \"\"\"\n\n def __init__(\n self,\n model_name_or_path: Union[str, Path],\n model_version: Optional[str] = None,\n top_k: int = 10,\n use_gpu: bool = True,\n devices: Optional[List[Union[str, torch.device]]] = None,\n batch_size: int = 16,\n scale_score: bool = True,\n ):\n \"\"\"\n :param model_name_or_path: Directory of a saved model or the name of a public model e.g.\n 'cross-encoder/ms-marco-MiniLM-L-12-v2'.\n See https://huggingface.co/cross-encoder for full list of available models\n :param model_version: The version of model to use from the HuggingFace model hub. Can be tag name, branch name, or commit hash.\n :param top_k: The maximum number of documents to return\n :param use_gpu: Whether to use all available GPUs or the CPU. Falls back on CPU if no GPU is available.\n :param devices: List of GPU (or CPU) devices, to limit inference to certain GPUs and not use all available ones\n The strings will be converted into pytorch devices, so use the string notation described here:\n https://pytorch.org/docs/stable/tensor_attributes.html?highlight=torch%20device#torch.torch.device\n (e.g. [\"cuda:0\"]).\n :param batch_size: Number of documents to process at a time.\n :param scale_score: The raw predictions will be transformed using a Sigmoid activation function in case the model\n only predicts a single label. For multi-label predictions, no scaling is applied. Set this\n to False if you do not want any scaling of the raw predictions.\n \"\"\"\n super().__init__()\n\n self.top_k = top_k\n\n if devices is not None:\n self.devices = [torch.device(device) for device in devices]\n else:\n self.devices, _ = initialize_device_settings(use_cuda=use_gpu, multi_gpu=True)\n\n self.transformer_model = AutoModelForSequenceClassification.from_pretrained(\n pretrained_model_name_or_path=model_name_or_path, revision=model_version\n )\n self.transformer_model.to(str(self.devices[0]))\n self.transformer_tokenizer = AutoTokenizer.from_pretrained(\n pretrained_model_name_or_path=model_name_or_path, revision=model_version\n )\n self.transformer_model.eval()\n\n # we use sigmoid activation function to scale the score in case there is only a single label\n # we do not apply any scaling when scale_score is set to False\n num_labels = self.transformer_model.num_labels\n self.activation_function: torch.nn.Module\n if num_labels == 1 and scale_score:\n self.activation_function = torch.nn.Sigmoid()\n else:\n self.activation_function = torch.nn.Identity()\n\n if len(self.devices) > 1:\n self.model = DataParallel(self.transformer_model, device_ids=self.devices)\n\n self.batch_size = batch_size\n\n def predict(self, query: str, documents: List[Document], top_k: Optional[int] = None) -> List[Document]:\n \"\"\"\n Use loaded ranker model to re-rank the supplied list of Document.\n\n Returns list of Document sorted by (desc.) similarity with the query.\n\n :param query: Query string\n :param documents: List of Document to be re-ranked\n :param top_k: The maximum number of documents to return\n :return: List of Document\n \"\"\"\n if top_k is None:\n top_k = self.top_k\n\n features = self.transformer_tokenizer(\n [query for doc in documents],\n [doc.content for doc in documents],\n padding=True,\n truncation=True,\n return_tensors=\"pt\",\n ).to(self.devices[0])\n\n # SentenceTransformerRanker uses:\n # 1. the logit as similarity score/answerable classification\n # 2. the logits as answerable classification (no_answer / has_answer)\n # https://www.sbert.net/docs/pretrained-models/ce-msmarco.html#usage-with-transformers\n with torch.no_grad():\n similarity_scores = self.transformer_model(**features).logits\n\n logits_dim = similarity_scores.shape[1] # [batch_size, logits_dim]\n sorted_scores_and_documents = sorted(\n zip(similarity_scores, documents),\n key=lambda similarity_document_tuple:\n # assume the last element in logits represents the `has_answer` label\n similarity_document_tuple[0][-1] if logits_dim >= 2 else similarity_document_tuple[0],\n reverse=True,\n )\n\n # add normalized scores to documents\n sorted_documents = self._add_scores_to_documents(sorted_scores_and_documents[:top_k], logits_dim)\n\n return sorted_documents\n\n def _add_scores_to_documents(\n self, sorted_scores_and_documents: List[Tuple[Any, Document]], logits_dim: int\n ) -> List[Document]:\n \"\"\"\n Normalize and add scores to retrieved result documents.\n\n :param sorted_scores_and_documents: List of score, Document Tuples.\n :param logits_dim: Dimensionality of the returned scores.\n \"\"\"\n sorted_documents = []\n for raw_score, doc in sorted_scores_and_documents:\n if logits_dim >= 2:\n score = self.activation_function(raw_score)[-1]\n else:\n score = self.activation_function(raw_score)[0]\n\n doc.score = score.detach().cpu().numpy().tolist()\n sorted_documents.append(doc)\n\n return sorted_documents\n\n def predict_batch(\n self,\n queries: List[str],\n documents: Union[List[Document], List[List[Document]]],\n top_k: Optional[int] = None,\n batch_size: Optional[int] = None,\n ) -> Union[List[Document], List[List[Document]]]:\n \"\"\"\n Use loaded ranker model to re-rank the supplied lists of Documents.\n\n Returns lists of Documents sorted by (desc.) similarity with the corresponding queries.\n\n\n - If you provide a list containing a single query...\n\n - ... and a single list of Documents, the single list of Documents will be re-ranked based on the\n supplied query.\n - ... and a list of lists of Documents, each list of Documents will be re-ranked individually based on the\n supplied query.\n\n\n - If you provide a list of multiple queries...\n\n - ... you need to provide a list of lists of Documents. Each list of Documents will be re-ranked based on\n its corresponding query.\n\n :param queries: Single query string or list of queries\n :param documents: Single list of Documents or list of lists of Documents to be reranked.\n :param top_k: The maximum number of documents to return per Document list.\n :param batch_size: Number of Documents to process at a time.\n \"\"\"\n if top_k is None:\n top_k = self.top_k\n\n if batch_size is None:\n batch_size = self.batch_size\n\n number_of_docs, all_queries, all_docs, single_list_of_docs = self._preprocess_batch_queries_and_docs(\n queries=queries, documents=documents\n )\n\n batches = self._get_batches(all_queries=all_queries, all_docs=all_docs, batch_size=batch_size)\n preds = []\n for cur_queries, cur_docs in batches:\n features = self.transformer_tokenizer(\n cur_queries, [doc.content for doc in cur_docs], padding=True, truncation=True, return_tensors=\"pt\"\n ).to(self.devices[0])\n\n with torch.no_grad():\n similarity_scores = self.transformer_model(**features).logits\n preds.extend(similarity_scores)\n\n logits_dim = similarity_scores.shape[1] # [batch_size, logits_dim]\n if single_list_of_docs:\n sorted_scores_and_documents = sorted(\n zip(similarity_scores, documents),\n key=lambda similarity_document_tuple:\n # assume the last element in logits represents the `has_answer` label\n similarity_document_tuple[0][-1] if logits_dim >= 2 else similarity_document_tuple[0],\n reverse=True,\n )\n\n # is this step needed?\n sorted_documents = [(score, doc) for score, doc in sorted_scores_and_documents if isinstance(doc, Document)]\n sorted_documents_with_scores = self._add_scores_to_documents(sorted_documents[:top_k], logits_dim)\n\n return sorted_documents_with_scores\n else:\n # Group predictions together\n grouped_predictions = []\n left_idx = 0\n right_idx = 0\n for number in number_of_docs:\n right_idx = left_idx + number\n grouped_predictions.append(similarity_scores[left_idx:right_idx])\n left_idx = right_idx\n\n result = []\n for pred_group, doc_group in zip(grouped_predictions, documents):\n sorted_scores_and_documents = sorted(\n zip(pred_group, doc_group), # type: ignore\n key=lambda similarity_document_tuple:\n # assume the last element in logits represents the `has_answer` label\n similarity_document_tuple[0][-1] if logits_dim >= 2 else similarity_document_tuple[0],\n reverse=True,\n )\n\n # rank documents according to scores\n sorted_documents = [\n (score, doc) for score, doc in sorted_scores_and_documents if isinstance(doc, Document)\n ]\n sorted_documents_with_scores = self._add_scores_to_documents(sorted_documents[:top_k], logits_dim)\n\n result.append(sorted_documents_with_scores)\n\n return result\n\n def _preprocess_batch_queries_and_docs(\n self, queries: List[str], documents: Union[List[Document], List[List[Document]]]\n ) -> Tuple[List[int], List[str], List[Document], bool]:\n number_of_docs = []\n all_queries = []\n all_docs: List[Document] = []\n single_list_of_docs = False\n\n # Docs case 1: single list of Documents -> rerank single list of Documents based on single query\n if len(documents) > 0 and isinstance(documents[0], Document):\n if len(queries) != 1:\n raise HaystackError(\"Number of queries must be 1 if a single list of Documents is provided.\")\n query = queries[0]\n number_of_docs = [len(documents)]\n all_queries = [query] * len(documents)\n all_docs = documents # type: ignore\n single_list_of_docs = True\n\n # Docs case 2: list of lists of Documents -> rerank each list of Documents based on corresponding query\n # If queries contains a single query, apply it to each list of Documents\n if len(documents) > 0 and isinstance(documents[0], list):\n if len(queries) == 1:\n queries = queries * len(documents)\n if len(queries) != len(documents):\n raise HaystackError(\"Number of queries must be equal to number of provided Document lists.\")\n for query, cur_docs in zip(queries, documents):\n if not isinstance(cur_docs, list):\n raise HaystackError(f\"cur_docs was of type {type(cur_docs)}, but expected a list of Documents.\")\n number_of_docs.append(len(cur_docs))\n all_queries.extend([query] * len(cur_docs))\n all_docs.extend(cur_docs)\n\n return number_of_docs, all_queries, all_docs, single_list_of_docs\n\n @staticmethod\n def _get_batches(\n all_queries: List[str], all_docs: List[Document], batch_size: Optional[int]\n ) -> Iterator[Tuple[List[str], List[Document]]]:\n if batch_size is None:\n yield all_queries, all_docs\n return\n else:\n for index in range(0, len(all_queries), batch_size):\n yield all_queries[index : index + batch_size], all_docs[index : index + batch_size]\n", "path": "haystack/nodes/ranker/sentence_transformers.py"}]} | 4,085 | 289 |
gh_patches_debug_39493 | rasdani/github-patches | git_diff | ethereum__web3.py-3077 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
cannot import name 'DEFAULT_WEBSOCKET_TIMEOUT'
ImportError: cannot import name 'DEFAULT_WEBSOCKET_TIMEOUT' from 'web3.providers.websocket' (C:\Python310\lib\site-packages\web3\providers\websocket\__init__.py)
any one know what happened?
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `web3/manager.py`
Content:
```
1 import asyncio
2 import json
3 import logging
4 from typing import (
5 TYPE_CHECKING,
6 Any,
7 AsyncGenerator,
8 Callable,
9 List,
10 Optional,
11 Sequence,
12 Tuple,
13 Union,
14 cast,
15 )
16
17 from eth_utils.toolz import (
18 pipe,
19 )
20 from hexbytes import (
21 HexBytes,
22 )
23 from websockets.exceptions import (
24 ConnectionClosedOK,
25 )
26
27 from web3._utils.caching import (
28 generate_cache_key,
29 )
30 from web3.datastructures import (
31 NamedElementOnion,
32 )
33 from web3.exceptions import (
34 BadResponseFormat,
35 MethodUnavailable,
36 )
37 from web3.middleware import (
38 abi_middleware,
39 async_attrdict_middleware,
40 async_buffered_gas_estimate_middleware,
41 async_gas_price_strategy_middleware,
42 async_name_to_address_middleware,
43 async_validation_middleware,
44 attrdict_middleware,
45 buffered_gas_estimate_middleware,
46 gas_price_strategy_middleware,
47 name_to_address_middleware,
48 validation_middleware,
49 )
50 from web3.module import (
51 apply_result_formatters,
52 )
53 from web3.providers import (
54 AutoProvider,
55 PersistentConnectionProvider,
56 )
57 from web3.types import (
58 AsyncMiddleware,
59 AsyncMiddlewareOnion,
60 Middleware,
61 MiddlewareOnion,
62 RPCEndpoint,
63 RPCResponse,
64 )
65
66 if TYPE_CHECKING:
67 from web3 import ( # noqa: F401
68 AsyncWeb3,
69 Web3,
70 )
71 from web3.providers import ( # noqa: F401
72 AsyncBaseProvider,
73 BaseProvider,
74 )
75
76
77 NULL_RESPONSES = [None, HexBytes("0x"), "0x"]
78
79
80 def apply_error_formatters(
81 error_formatters: Callable[..., Any],
82 response: RPCResponse,
83 ) -> RPCResponse:
84 if error_formatters:
85 formatted_resp = pipe(response, error_formatters)
86 return formatted_resp
87 else:
88 return response
89
90
91 def apply_null_result_formatters(
92 null_result_formatters: Callable[..., Any],
93 response: RPCResponse,
94 params: Optional[Any] = None,
95 ) -> RPCResponse:
96 if null_result_formatters:
97 formatted_resp = pipe(params, null_result_formatters)
98 return formatted_resp
99 else:
100 return response
101
102
103 class RequestManager:
104 logger = logging.getLogger("web3.RequestManager")
105
106 middleware_onion: Union[
107 MiddlewareOnion, AsyncMiddlewareOnion, NamedElementOnion[None, None]
108 ]
109
110 def __init__(
111 self,
112 w3: Union["AsyncWeb3", "Web3"],
113 provider: Optional[Union["BaseProvider", "AsyncBaseProvider"]] = None,
114 middlewares: Optional[
115 Union[
116 Sequence[Tuple[Middleware, str]], Sequence[Tuple[AsyncMiddleware, str]]
117 ]
118 ] = None,
119 ) -> None:
120 self.w3 = w3
121
122 if provider is None:
123 self.provider = AutoProvider()
124 else:
125 self.provider = provider
126
127 if middlewares is None:
128 middlewares = (
129 self.async_default_middlewares()
130 if self.provider.is_async
131 else self.default_middlewares(cast("Web3", w3))
132 )
133
134 self.middleware_onion = NamedElementOnion(middlewares)
135
136 w3: Union["AsyncWeb3", "Web3"] = None
137 _provider = None
138
139 @property
140 def provider(self) -> Union["BaseProvider", "AsyncBaseProvider"]:
141 return self._provider
142
143 @provider.setter
144 def provider(self, provider: Union["BaseProvider", "AsyncBaseProvider"]) -> None:
145 self._provider = provider
146
147 @staticmethod
148 def default_middlewares(w3: "Web3") -> List[Tuple[Middleware, str]]:
149 """
150 List the default middlewares for the request manager.
151 Leaving w3 unspecified will prevent the middleware from resolving names.
152 Documentation should remain in sync with these defaults.
153 """
154 return [
155 (gas_price_strategy_middleware, "gas_price_strategy"),
156 (name_to_address_middleware(w3), "name_to_address"),
157 (attrdict_middleware, "attrdict"),
158 (validation_middleware, "validation"),
159 (abi_middleware, "abi"),
160 (buffered_gas_estimate_middleware, "gas_estimate"),
161 ]
162
163 @staticmethod
164 def async_default_middlewares() -> List[Tuple[AsyncMiddleware, str]]:
165 """
166 List the default async middlewares for the request manager.
167 Documentation should remain in sync with these defaults.
168 """
169 return [
170 (async_gas_price_strategy_middleware, "gas_price_strategy"),
171 (async_name_to_address_middleware, "name_to_address"),
172 (async_attrdict_middleware, "attrdict"),
173 (async_validation_middleware, "validation"),
174 (async_buffered_gas_estimate_middleware, "gas_estimate"),
175 ]
176
177 #
178 # Provider requests and response
179 #
180 def _make_request(
181 self, method: Union[RPCEndpoint, Callable[..., RPCEndpoint]], params: Any
182 ) -> RPCResponse:
183 provider = cast("BaseProvider", self.provider)
184 request_func = provider.request_func(
185 cast("Web3", self.w3), cast(MiddlewareOnion, self.middleware_onion)
186 )
187 self.logger.debug(f"Making request. Method: {method}")
188 return request_func(method, params)
189
190 async def _coro_make_request(
191 self, method: Union[RPCEndpoint, Callable[..., RPCEndpoint]], params: Any
192 ) -> RPCResponse:
193 provider = cast("AsyncBaseProvider", self.provider)
194 request_func = await provider.request_func(
195 cast("AsyncWeb3", self.w3),
196 cast(AsyncMiddlewareOnion, self.middleware_onion),
197 )
198 self.logger.debug(f"Making request. Method: {method}")
199 return await request_func(method, params)
200
201 @staticmethod
202 def formatted_response(
203 response: RPCResponse,
204 params: Any,
205 error_formatters: Optional[Callable[..., Any]] = None,
206 null_result_formatters: Optional[Callable[..., Any]] = None,
207 ) -> Any:
208 if "error" in response:
209 apply_error_formatters(error_formatters, response)
210
211 # guard against eth-tester case - eth-tester returns a string
212 # with no code, so can't parse what the error is.
213 if isinstance(response["error"], dict):
214 resp_code = response["error"].get("code")
215 if resp_code == -32601:
216 raise MethodUnavailable(response["error"])
217 raise ValueError(response["error"])
218 # NULL_RESPONSES includes None, so return False here as the default
219 # so we don't apply the null_result_formatters if there is no 'result' key
220 elif response.get("result", False) in NULL_RESPONSES:
221 # null_result_formatters raise either a BlockNotFound
222 # or a TransactionNotFound error, depending on the method called
223 apply_null_result_formatters(null_result_formatters, response, params)
224 return response["result"]
225 elif response.get("result") is not None:
226 return response["result"]
227 elif (
228 # eth_subscribe case
229 response.get("params") is not None
230 and response["params"].get("result") is not None
231 ):
232 return response["params"]["result"]
233 else:
234 raise BadResponseFormat(
235 "The response was in an unexpected format and unable to be parsed. "
236 f"The raw response is: {response}"
237 )
238
239 def request_blocking(
240 self,
241 method: Union[RPCEndpoint, Callable[..., RPCEndpoint]],
242 params: Any,
243 error_formatters: Optional[Callable[..., Any]] = None,
244 null_result_formatters: Optional[Callable[..., Any]] = None,
245 ) -> Any:
246 """
247 Make a synchronous request using the provider
248 """
249 response = self._make_request(method, params)
250 return self.formatted_response(
251 response, params, error_formatters, null_result_formatters
252 )
253
254 async def coro_request(
255 self,
256 method: Union[RPCEndpoint, Callable[..., RPCEndpoint]],
257 params: Any,
258 error_formatters: Optional[Callable[..., Any]] = None,
259 null_result_formatters: Optional[Callable[..., Any]] = None,
260 ) -> Any:
261 """
262 Coroutine for making a request using the provider
263 """
264 response = await self._coro_make_request(method, params)
265 return self.formatted_response(
266 response, params, error_formatters, null_result_formatters
267 )
268
269 # persistent connection
270 async def ws_send(
271 self,
272 method: Union[RPCEndpoint, Callable[..., RPCEndpoint]],
273 params: Any,
274 ) -> RPCResponse:
275 provider = cast(PersistentConnectionProvider, self._provider)
276 request_func = await provider.request_func(
277 cast("AsyncWeb3", self.w3),
278 cast(AsyncMiddlewareOnion, self.middleware_onion),
279 )
280 self.logger.debug(
281 "Making request to open websocket connection - "
282 f"uri: {provider.endpoint_uri}, method: {method}"
283 )
284 await request_func(method, params)
285 return await asyncio.wait_for(
286 self.ws_recv(),
287 timeout=provider.call_timeout,
288 )
289
290 async def ws_recv(self) -> Any:
291 return await self._ws_recv_stream().__anext__()
292
293 def persistent_recv_stream(self) -> "_AsyncPersistentRecvStream":
294 return _AsyncPersistentRecvStream(self)
295
296 async def _ws_recv_stream(self) -> AsyncGenerator[RPCResponse, None]:
297 if not isinstance(self._provider, PersistentConnectionProvider):
298 raise TypeError(
299 "Only websocket providers that maintain an open, persistent connection "
300 "can listen to websocket recv streams."
301 )
302
303 response = json.loads(
304 await asyncio.wait_for(
305 self._provider.ws.recv(),
306 timeout=self._provider.call_timeout,
307 )
308 )
309 request_info = self._provider._get_request_information_for_response(response)
310
311 if request_info is None:
312 self.logger.debug("No cache key found for response, returning raw response")
313 yield response
314
315 if request_info.method == "eth_subscribe" and "result" in response.keys():
316 # if response for the initial eth_subscribe request, which returns the
317 # subscription id
318 subscription_id = response["result"]
319 cache_key = generate_cache_key(subscription_id)
320 if cache_key not in self._provider._async_response_processing_cache:
321 # cache by subscription id in order to process each response for the
322 # subscription as it comes in
323 self._provider.logger.debug(
324 f"Caching eth_subscription info:\n cache_key={cache_key},\n "
325 f"request_info={request_info.__dict__}"
326 )
327 self._provider._async_response_processing_cache.cache(
328 cache_key, request_info
329 )
330
331 # pipe response back through middleware response processors
332 if len(request_info.middleware_response_processors) > 0:
333 response = pipe(response, *request_info.middleware_response_processors)
334
335 (
336 result_formatters,
337 error_formatters,
338 null_formatters,
339 ) = request_info.response_formatters
340 partly_formatted_response = self.formatted_response(
341 response,
342 request_info.params,
343 error_formatters,
344 null_formatters,
345 )
346 yield apply_result_formatters(result_formatters, partly_formatted_response)
347
348
349 class _AsyncPersistentRecvStream:
350 """
351 Async generator for receiving responses from a persistent connection. This
352 abstraction is necessary to define the `__aiter__()` method required for
353 use with "async for" loops.
354 """
355
356 def __init__(self, manager: RequestManager, *args: Any, **kwargs: Any) -> None:
357 self.manager = manager
358 super().__init__(*args, **kwargs)
359
360 def __aiter__(self) -> AsyncGenerator[RPCResponse, None]:
361 while True:
362 try:
363 return self.manager._ws_recv_stream()
364 except ConnectionClosedOK:
365 pass
366
```
Path: `web3/providers/websocket/__init__.py`
Content:
```
1 from .websocket import ( # noqa: F401
2 WebsocketProvider,
3 )
4 from .websocket_v2 import ( # noqa: F401
5 WebsocketProviderV2,
6 )
7
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/web3/manager.py b/web3/manager.py
--- a/web3/manager.py
+++ b/web3/manager.py
@@ -312,38 +312,40 @@
self.logger.debug("No cache key found for response, returning raw response")
yield response
- if request_info.method == "eth_subscribe" and "result" in response.keys():
- # if response for the initial eth_subscribe request, which returns the
- # subscription id
- subscription_id = response["result"]
- cache_key = generate_cache_key(subscription_id)
- if cache_key not in self._provider._async_response_processing_cache:
- # cache by subscription id in order to process each response for the
- # subscription as it comes in
- self._provider.logger.debug(
- f"Caching eth_subscription info:\n cache_key={cache_key},\n "
- f"request_info={request_info.__dict__}"
- )
- self._provider._async_response_processing_cache.cache(
- cache_key, request_info
- )
-
- # pipe response back through middleware response processors
- if len(request_info.middleware_response_processors) > 0:
- response = pipe(response, *request_info.middleware_response_processors)
-
- (
- result_formatters,
- error_formatters,
- null_formatters,
- ) = request_info.response_formatters
- partly_formatted_response = self.formatted_response(
- response,
- request_info.params,
- error_formatters,
- null_formatters,
- )
- yield apply_result_formatters(result_formatters, partly_formatted_response)
+ else:
+ if request_info.method == "eth_subscribe" and "result" in response.keys():
+ # if response for the initial eth_subscribe request, which returns the
+ # subscription id
+ subscription_id = response["result"]
+ cache_key = generate_cache_key(subscription_id)
+ if cache_key not in self._provider._async_response_processing_cache:
+ # cache by subscription id in order to process each response for the
+ # subscription as it comes in
+ self._provider.logger.debug(
+ f"Caching eth_subscription info:\n "
+ f"cache_key={cache_key},\n "
+ f"request_info={request_info.__dict__}"
+ )
+ self._provider._async_response_processing_cache.cache(
+ cache_key, request_info
+ )
+
+ # pipe response back through middleware response processors
+ if len(request_info.middleware_response_processors) > 0:
+ response = pipe(response, *request_info.middleware_response_processors)
+
+ (
+ result_formatters,
+ error_formatters,
+ null_formatters,
+ ) = request_info.response_formatters
+ partly_formatted_response = self.formatted_response(
+ response,
+ request_info.params,
+ error_formatters,
+ null_formatters,
+ )
+ yield apply_result_formatters(result_formatters, partly_formatted_response)
class _AsyncPersistentRecvStream:
diff --git a/web3/providers/websocket/__init__.py b/web3/providers/websocket/__init__.py
--- a/web3/providers/websocket/__init__.py
+++ b/web3/providers/websocket/__init__.py
@@ -1,4 +1,6 @@
from .websocket import ( # noqa: F401
+ DEFAULT_WEBSOCKET_TIMEOUT,
+ RESTRICTED_WEBSOCKET_KWARGS,
WebsocketProvider,
)
from .websocket_v2 import ( # noqa: F401
| {"golden_diff": "diff --git a/web3/manager.py b/web3/manager.py\n--- a/web3/manager.py\n+++ b/web3/manager.py\n@@ -312,38 +312,40 @@\n self.logger.debug(\"No cache key found for response, returning raw response\")\n yield response\n \n- if request_info.method == \"eth_subscribe\" and \"result\" in response.keys():\n- # if response for the initial eth_subscribe request, which returns the\n- # subscription id\n- subscription_id = response[\"result\"]\n- cache_key = generate_cache_key(subscription_id)\n- if cache_key not in self._provider._async_response_processing_cache:\n- # cache by subscription id in order to process each response for the\n- # subscription as it comes in\n- self._provider.logger.debug(\n- f\"Caching eth_subscription info:\\n cache_key={cache_key},\\n \"\n- f\"request_info={request_info.__dict__}\"\n- )\n- self._provider._async_response_processing_cache.cache(\n- cache_key, request_info\n- )\n-\n- # pipe response back through middleware response processors\n- if len(request_info.middleware_response_processors) > 0:\n- response = pipe(response, *request_info.middleware_response_processors)\n-\n- (\n- result_formatters,\n- error_formatters,\n- null_formatters,\n- ) = request_info.response_formatters\n- partly_formatted_response = self.formatted_response(\n- response,\n- request_info.params,\n- error_formatters,\n- null_formatters,\n- )\n- yield apply_result_formatters(result_formatters, partly_formatted_response)\n+ else:\n+ if request_info.method == \"eth_subscribe\" and \"result\" in response.keys():\n+ # if response for the initial eth_subscribe request, which returns the\n+ # subscription id\n+ subscription_id = response[\"result\"]\n+ cache_key = generate_cache_key(subscription_id)\n+ if cache_key not in self._provider._async_response_processing_cache:\n+ # cache by subscription id in order to process each response for the\n+ # subscription as it comes in\n+ self._provider.logger.debug(\n+ f\"Caching eth_subscription info:\\n \"\n+ f\"cache_key={cache_key},\\n \"\n+ f\"request_info={request_info.__dict__}\"\n+ )\n+ self._provider._async_response_processing_cache.cache(\n+ cache_key, request_info\n+ )\n+\n+ # pipe response back through middleware response processors\n+ if len(request_info.middleware_response_processors) > 0:\n+ response = pipe(response, *request_info.middleware_response_processors)\n+\n+ (\n+ result_formatters,\n+ error_formatters,\n+ null_formatters,\n+ ) = request_info.response_formatters\n+ partly_formatted_response = self.formatted_response(\n+ response,\n+ request_info.params,\n+ error_formatters,\n+ null_formatters,\n+ )\n+ yield apply_result_formatters(result_formatters, partly_formatted_response)\n \n \n class _AsyncPersistentRecvStream:\ndiff --git a/web3/providers/websocket/__init__.py b/web3/providers/websocket/__init__.py\n--- a/web3/providers/websocket/__init__.py\n+++ b/web3/providers/websocket/__init__.py\n@@ -1,4 +1,6 @@\n from .websocket import ( # noqa: F401\n+ DEFAULT_WEBSOCKET_TIMEOUT,\n+ RESTRICTED_WEBSOCKET_KWARGS,\n WebsocketProvider,\n )\n from .websocket_v2 import ( # noqa: F401\n", "issue": " cannot import name 'DEFAULT_WEBSOCKET_TIMEOUT'\n\r\nImportError: cannot import name 'DEFAULT_WEBSOCKET_TIMEOUT' from 'web3.providers.websocket' (C:\\Python310\\lib\\site-packages\\web3\\providers\\websocket\\__init__.py)\r\n\r\nany one know what happened?\n", "before_files": [{"content": "import asyncio\nimport json\nimport logging\nfrom typing import (\n TYPE_CHECKING,\n Any,\n AsyncGenerator,\n Callable,\n List,\n Optional,\n Sequence,\n Tuple,\n Union,\n cast,\n)\n\nfrom eth_utils.toolz import (\n pipe,\n)\nfrom hexbytes import (\n HexBytes,\n)\nfrom websockets.exceptions import (\n ConnectionClosedOK,\n)\n\nfrom web3._utils.caching import (\n generate_cache_key,\n)\nfrom web3.datastructures import (\n NamedElementOnion,\n)\nfrom web3.exceptions import (\n BadResponseFormat,\n MethodUnavailable,\n)\nfrom web3.middleware import (\n abi_middleware,\n async_attrdict_middleware,\n async_buffered_gas_estimate_middleware,\n async_gas_price_strategy_middleware,\n async_name_to_address_middleware,\n async_validation_middleware,\n attrdict_middleware,\n buffered_gas_estimate_middleware,\n gas_price_strategy_middleware,\n name_to_address_middleware,\n validation_middleware,\n)\nfrom web3.module import (\n apply_result_formatters,\n)\nfrom web3.providers import (\n AutoProvider,\n PersistentConnectionProvider,\n)\nfrom web3.types import (\n AsyncMiddleware,\n AsyncMiddlewareOnion,\n Middleware,\n MiddlewareOnion,\n RPCEndpoint,\n RPCResponse,\n)\n\nif TYPE_CHECKING:\n from web3 import ( # noqa: F401\n AsyncWeb3,\n Web3,\n )\n from web3.providers import ( # noqa: F401\n AsyncBaseProvider,\n BaseProvider,\n )\n\n\nNULL_RESPONSES = [None, HexBytes(\"0x\"), \"0x\"]\n\n\ndef apply_error_formatters(\n error_formatters: Callable[..., Any],\n response: RPCResponse,\n) -> RPCResponse:\n if error_formatters:\n formatted_resp = pipe(response, error_formatters)\n return formatted_resp\n else:\n return response\n\n\ndef apply_null_result_formatters(\n null_result_formatters: Callable[..., Any],\n response: RPCResponse,\n params: Optional[Any] = None,\n) -> RPCResponse:\n if null_result_formatters:\n formatted_resp = pipe(params, null_result_formatters)\n return formatted_resp\n else:\n return response\n\n\nclass RequestManager:\n logger = logging.getLogger(\"web3.RequestManager\")\n\n middleware_onion: Union[\n MiddlewareOnion, AsyncMiddlewareOnion, NamedElementOnion[None, None]\n ]\n\n def __init__(\n self,\n w3: Union[\"AsyncWeb3\", \"Web3\"],\n provider: Optional[Union[\"BaseProvider\", \"AsyncBaseProvider\"]] = None,\n middlewares: Optional[\n Union[\n Sequence[Tuple[Middleware, str]], Sequence[Tuple[AsyncMiddleware, str]]\n ]\n ] = None,\n ) -> None:\n self.w3 = w3\n\n if provider is None:\n self.provider = AutoProvider()\n else:\n self.provider = provider\n\n if middlewares is None:\n middlewares = (\n self.async_default_middlewares()\n if self.provider.is_async\n else self.default_middlewares(cast(\"Web3\", w3))\n )\n\n self.middleware_onion = NamedElementOnion(middlewares)\n\n w3: Union[\"AsyncWeb3\", \"Web3\"] = None\n _provider = None\n\n @property\n def provider(self) -> Union[\"BaseProvider\", \"AsyncBaseProvider\"]:\n return self._provider\n\n @provider.setter\n def provider(self, provider: Union[\"BaseProvider\", \"AsyncBaseProvider\"]) -> None:\n self._provider = provider\n\n @staticmethod\n def default_middlewares(w3: \"Web3\") -> List[Tuple[Middleware, str]]:\n \"\"\"\n List the default middlewares for the request manager.\n Leaving w3 unspecified will prevent the middleware from resolving names.\n Documentation should remain in sync with these defaults.\n \"\"\"\n return [\n (gas_price_strategy_middleware, \"gas_price_strategy\"),\n (name_to_address_middleware(w3), \"name_to_address\"),\n (attrdict_middleware, \"attrdict\"),\n (validation_middleware, \"validation\"),\n (abi_middleware, \"abi\"),\n (buffered_gas_estimate_middleware, \"gas_estimate\"),\n ]\n\n @staticmethod\n def async_default_middlewares() -> List[Tuple[AsyncMiddleware, str]]:\n \"\"\"\n List the default async middlewares for the request manager.\n Documentation should remain in sync with these defaults.\n \"\"\"\n return [\n (async_gas_price_strategy_middleware, \"gas_price_strategy\"),\n (async_name_to_address_middleware, \"name_to_address\"),\n (async_attrdict_middleware, \"attrdict\"),\n (async_validation_middleware, \"validation\"),\n (async_buffered_gas_estimate_middleware, \"gas_estimate\"),\n ]\n\n #\n # Provider requests and response\n #\n def _make_request(\n self, method: Union[RPCEndpoint, Callable[..., RPCEndpoint]], params: Any\n ) -> RPCResponse:\n provider = cast(\"BaseProvider\", self.provider)\n request_func = provider.request_func(\n cast(\"Web3\", self.w3), cast(MiddlewareOnion, self.middleware_onion)\n )\n self.logger.debug(f\"Making request. Method: {method}\")\n return request_func(method, params)\n\n async def _coro_make_request(\n self, method: Union[RPCEndpoint, Callable[..., RPCEndpoint]], params: Any\n ) -> RPCResponse:\n provider = cast(\"AsyncBaseProvider\", self.provider)\n request_func = await provider.request_func(\n cast(\"AsyncWeb3\", self.w3),\n cast(AsyncMiddlewareOnion, self.middleware_onion),\n )\n self.logger.debug(f\"Making request. Method: {method}\")\n return await request_func(method, params)\n\n @staticmethod\n def formatted_response(\n response: RPCResponse,\n params: Any,\n error_formatters: Optional[Callable[..., Any]] = None,\n null_result_formatters: Optional[Callable[..., Any]] = None,\n ) -> Any:\n if \"error\" in response:\n apply_error_formatters(error_formatters, response)\n\n # guard against eth-tester case - eth-tester returns a string\n # with no code, so can't parse what the error is.\n if isinstance(response[\"error\"], dict):\n resp_code = response[\"error\"].get(\"code\")\n if resp_code == -32601:\n raise MethodUnavailable(response[\"error\"])\n raise ValueError(response[\"error\"])\n # NULL_RESPONSES includes None, so return False here as the default\n # so we don't apply the null_result_formatters if there is no 'result' key\n elif response.get(\"result\", False) in NULL_RESPONSES:\n # null_result_formatters raise either a BlockNotFound\n # or a TransactionNotFound error, depending on the method called\n apply_null_result_formatters(null_result_formatters, response, params)\n return response[\"result\"]\n elif response.get(\"result\") is not None:\n return response[\"result\"]\n elif (\n # eth_subscribe case\n response.get(\"params\") is not None\n and response[\"params\"].get(\"result\") is not None\n ):\n return response[\"params\"][\"result\"]\n else:\n raise BadResponseFormat(\n \"The response was in an unexpected format and unable to be parsed. \"\n f\"The raw response is: {response}\"\n )\n\n def request_blocking(\n self,\n method: Union[RPCEndpoint, Callable[..., RPCEndpoint]],\n params: Any,\n error_formatters: Optional[Callable[..., Any]] = None,\n null_result_formatters: Optional[Callable[..., Any]] = None,\n ) -> Any:\n \"\"\"\n Make a synchronous request using the provider\n \"\"\"\n response = self._make_request(method, params)\n return self.formatted_response(\n response, params, error_formatters, null_result_formatters\n )\n\n async def coro_request(\n self,\n method: Union[RPCEndpoint, Callable[..., RPCEndpoint]],\n params: Any,\n error_formatters: Optional[Callable[..., Any]] = None,\n null_result_formatters: Optional[Callable[..., Any]] = None,\n ) -> Any:\n \"\"\"\n Coroutine for making a request using the provider\n \"\"\"\n response = await self._coro_make_request(method, params)\n return self.formatted_response(\n response, params, error_formatters, null_result_formatters\n )\n\n # persistent connection\n async def ws_send(\n self,\n method: Union[RPCEndpoint, Callable[..., RPCEndpoint]],\n params: Any,\n ) -> RPCResponse:\n provider = cast(PersistentConnectionProvider, self._provider)\n request_func = await provider.request_func(\n cast(\"AsyncWeb3\", self.w3),\n cast(AsyncMiddlewareOnion, self.middleware_onion),\n )\n self.logger.debug(\n \"Making request to open websocket connection - \"\n f\"uri: {provider.endpoint_uri}, method: {method}\"\n )\n await request_func(method, params)\n return await asyncio.wait_for(\n self.ws_recv(),\n timeout=provider.call_timeout,\n )\n\n async def ws_recv(self) -> Any:\n return await self._ws_recv_stream().__anext__()\n\n def persistent_recv_stream(self) -> \"_AsyncPersistentRecvStream\":\n return _AsyncPersistentRecvStream(self)\n\n async def _ws_recv_stream(self) -> AsyncGenerator[RPCResponse, None]:\n if not isinstance(self._provider, PersistentConnectionProvider):\n raise TypeError(\n \"Only websocket providers that maintain an open, persistent connection \"\n \"can listen to websocket recv streams.\"\n )\n\n response = json.loads(\n await asyncio.wait_for(\n self._provider.ws.recv(),\n timeout=self._provider.call_timeout,\n )\n )\n request_info = self._provider._get_request_information_for_response(response)\n\n if request_info is None:\n self.logger.debug(\"No cache key found for response, returning raw response\")\n yield response\n\n if request_info.method == \"eth_subscribe\" and \"result\" in response.keys():\n # if response for the initial eth_subscribe request, which returns the\n # subscription id\n subscription_id = response[\"result\"]\n cache_key = generate_cache_key(subscription_id)\n if cache_key not in self._provider._async_response_processing_cache:\n # cache by subscription id in order to process each response for the\n # subscription as it comes in\n self._provider.logger.debug(\n f\"Caching eth_subscription info:\\n cache_key={cache_key},\\n \"\n f\"request_info={request_info.__dict__}\"\n )\n self._provider._async_response_processing_cache.cache(\n cache_key, request_info\n )\n\n # pipe response back through middleware response processors\n if len(request_info.middleware_response_processors) > 0:\n response = pipe(response, *request_info.middleware_response_processors)\n\n (\n result_formatters,\n error_formatters,\n null_formatters,\n ) = request_info.response_formatters\n partly_formatted_response = self.formatted_response(\n response,\n request_info.params,\n error_formatters,\n null_formatters,\n )\n yield apply_result_formatters(result_formatters, partly_formatted_response)\n\n\nclass _AsyncPersistentRecvStream:\n \"\"\"\n Async generator for receiving responses from a persistent connection. This\n abstraction is necessary to define the `__aiter__()` method required for\n use with \"async for\" loops.\n \"\"\"\n\n def __init__(self, manager: RequestManager, *args: Any, **kwargs: Any) -> None:\n self.manager = manager\n super().__init__(*args, **kwargs)\n\n def __aiter__(self) -> AsyncGenerator[RPCResponse, None]:\n while True:\n try:\n return self.manager._ws_recv_stream()\n except ConnectionClosedOK:\n pass\n", "path": "web3/manager.py"}, {"content": "from .websocket import ( # noqa: F401\n WebsocketProvider,\n)\nfrom .websocket_v2 import ( # noqa: F401\n WebsocketProviderV2,\n)\n", "path": "web3/providers/websocket/__init__.py"}], "after_files": [{"content": "import asyncio\nimport json\nimport logging\nfrom typing import (\n TYPE_CHECKING,\n Any,\n AsyncGenerator,\n Callable,\n List,\n Optional,\n Sequence,\n Tuple,\n Union,\n cast,\n)\n\nfrom eth_utils.toolz import (\n pipe,\n)\nfrom hexbytes import (\n HexBytes,\n)\nfrom websockets.exceptions import (\n ConnectionClosedOK,\n)\n\nfrom web3._utils.caching import (\n generate_cache_key,\n)\nfrom web3.datastructures import (\n NamedElementOnion,\n)\nfrom web3.exceptions import (\n BadResponseFormat,\n MethodUnavailable,\n)\nfrom web3.middleware import (\n abi_middleware,\n async_attrdict_middleware,\n async_buffered_gas_estimate_middleware,\n async_gas_price_strategy_middleware,\n async_name_to_address_middleware,\n async_validation_middleware,\n attrdict_middleware,\n buffered_gas_estimate_middleware,\n gas_price_strategy_middleware,\n name_to_address_middleware,\n validation_middleware,\n)\nfrom web3.module import (\n apply_result_formatters,\n)\nfrom web3.providers import (\n AutoProvider,\n PersistentConnectionProvider,\n)\nfrom web3.types import (\n AsyncMiddleware,\n AsyncMiddlewareOnion,\n Middleware,\n MiddlewareOnion,\n RPCEndpoint,\n RPCResponse,\n)\n\nif TYPE_CHECKING:\n from web3 import ( # noqa: F401\n AsyncWeb3,\n Web3,\n )\n from web3.providers import ( # noqa: F401\n AsyncBaseProvider,\n BaseProvider,\n )\n\n\nNULL_RESPONSES = [None, HexBytes(\"0x\"), \"0x\"]\n\n\ndef apply_error_formatters(\n error_formatters: Callable[..., Any],\n response: RPCResponse,\n) -> RPCResponse:\n if error_formatters:\n formatted_resp = pipe(response, error_formatters)\n return formatted_resp\n else:\n return response\n\n\ndef apply_null_result_formatters(\n null_result_formatters: Callable[..., Any],\n response: RPCResponse,\n params: Optional[Any] = None,\n) -> RPCResponse:\n if null_result_formatters:\n formatted_resp = pipe(params, null_result_formatters)\n return formatted_resp\n else:\n return response\n\n\nclass RequestManager:\n logger = logging.getLogger(\"web3.RequestManager\")\n\n middleware_onion: Union[\n MiddlewareOnion, AsyncMiddlewareOnion, NamedElementOnion[None, None]\n ]\n\n def __init__(\n self,\n w3: Union[\"AsyncWeb3\", \"Web3\"],\n provider: Optional[Union[\"BaseProvider\", \"AsyncBaseProvider\"]] = None,\n middlewares: Optional[\n Union[\n Sequence[Tuple[Middleware, str]], Sequence[Tuple[AsyncMiddleware, str]]\n ]\n ] = None,\n ) -> None:\n self.w3 = w3\n\n if provider is None:\n self.provider = AutoProvider()\n else:\n self.provider = provider\n\n if middlewares is None:\n middlewares = (\n self.async_default_middlewares()\n if self.provider.is_async\n else self.default_middlewares(cast(\"Web3\", w3))\n )\n\n self.middleware_onion = NamedElementOnion(middlewares)\n\n w3: Union[\"AsyncWeb3\", \"Web3\"] = None\n _provider = None\n\n @property\n def provider(self) -> Union[\"BaseProvider\", \"AsyncBaseProvider\"]:\n return self._provider\n\n @provider.setter\n def provider(self, provider: Union[\"BaseProvider\", \"AsyncBaseProvider\"]) -> None:\n self._provider = provider\n\n @staticmethod\n def default_middlewares(w3: \"Web3\") -> List[Tuple[Middleware, str]]:\n \"\"\"\n List the default middlewares for the request manager.\n Leaving w3 unspecified will prevent the middleware from resolving names.\n Documentation should remain in sync with these defaults.\n \"\"\"\n return [\n (gas_price_strategy_middleware, \"gas_price_strategy\"),\n (name_to_address_middleware(w3), \"name_to_address\"),\n (attrdict_middleware, \"attrdict\"),\n (validation_middleware, \"validation\"),\n (abi_middleware, \"abi\"),\n (buffered_gas_estimate_middleware, \"gas_estimate\"),\n ]\n\n @staticmethod\n def async_default_middlewares() -> List[Tuple[AsyncMiddleware, str]]:\n \"\"\"\n List the default async middlewares for the request manager.\n Documentation should remain in sync with these defaults.\n \"\"\"\n return [\n (async_gas_price_strategy_middleware, \"gas_price_strategy\"),\n (async_name_to_address_middleware, \"name_to_address\"),\n (async_attrdict_middleware, \"attrdict\"),\n (async_validation_middleware, \"validation\"),\n (async_buffered_gas_estimate_middleware, \"gas_estimate\"),\n ]\n\n #\n # Provider requests and response\n #\n def _make_request(\n self, method: Union[RPCEndpoint, Callable[..., RPCEndpoint]], params: Any\n ) -> RPCResponse:\n provider = cast(\"BaseProvider\", self.provider)\n request_func = provider.request_func(\n cast(\"Web3\", self.w3), cast(MiddlewareOnion, self.middleware_onion)\n )\n self.logger.debug(f\"Making request. Method: {method}\")\n return request_func(method, params)\n\n async def _coro_make_request(\n self, method: Union[RPCEndpoint, Callable[..., RPCEndpoint]], params: Any\n ) -> RPCResponse:\n provider = cast(\"AsyncBaseProvider\", self.provider)\n request_func = await provider.request_func(\n cast(\"AsyncWeb3\", self.w3),\n cast(AsyncMiddlewareOnion, self.middleware_onion),\n )\n self.logger.debug(f\"Making request. Method: {method}\")\n return await request_func(method, params)\n\n @staticmethod\n def formatted_response(\n response: RPCResponse,\n params: Any,\n error_formatters: Optional[Callable[..., Any]] = None,\n null_result_formatters: Optional[Callable[..., Any]] = None,\n ) -> Any:\n if \"error\" in response:\n apply_error_formatters(error_formatters, response)\n\n # guard against eth-tester case - eth-tester returns a string\n # with no code, so can't parse what the error is.\n if isinstance(response[\"error\"], dict):\n resp_code = response[\"error\"].get(\"code\")\n if resp_code == -32601:\n raise MethodUnavailable(response[\"error\"])\n raise ValueError(response[\"error\"])\n # NULL_RESPONSES includes None, so return False here as the default\n # so we don't apply the null_result_formatters if there is no 'result' key\n elif response.get(\"result\", False) in NULL_RESPONSES:\n # null_result_formatters raise either a BlockNotFound\n # or a TransactionNotFound error, depending on the method called\n apply_null_result_formatters(null_result_formatters, response, params)\n return response[\"result\"]\n elif response.get(\"result\") is not None:\n return response[\"result\"]\n elif (\n # eth_subscribe case\n response.get(\"params\") is not None\n and response[\"params\"].get(\"result\") is not None\n ):\n return response[\"params\"][\"result\"]\n else:\n raise BadResponseFormat(\n \"The response was in an unexpected format and unable to be parsed. \"\n f\"The raw response is: {response}\"\n )\n\n def request_blocking(\n self,\n method: Union[RPCEndpoint, Callable[..., RPCEndpoint]],\n params: Any,\n error_formatters: Optional[Callable[..., Any]] = None,\n null_result_formatters: Optional[Callable[..., Any]] = None,\n ) -> Any:\n \"\"\"\n Make a synchronous request using the provider\n \"\"\"\n response = self._make_request(method, params)\n return self.formatted_response(\n response, params, error_formatters, null_result_formatters\n )\n\n async def coro_request(\n self,\n method: Union[RPCEndpoint, Callable[..., RPCEndpoint]],\n params: Any,\n error_formatters: Optional[Callable[..., Any]] = None,\n null_result_formatters: Optional[Callable[..., Any]] = None,\n ) -> Any:\n \"\"\"\n Coroutine for making a request using the provider\n \"\"\"\n response = await self._coro_make_request(method, params)\n return self.formatted_response(\n response, params, error_formatters, null_result_formatters\n )\n\n # persistent connection\n async def ws_send(\n self,\n method: Union[RPCEndpoint, Callable[..., RPCEndpoint]],\n params: Any,\n ) -> RPCResponse:\n provider = cast(PersistentConnectionProvider, self._provider)\n request_func = await provider.request_func(\n cast(\"AsyncWeb3\", self.w3),\n cast(AsyncMiddlewareOnion, self.middleware_onion),\n )\n self.logger.debug(\n \"Making request to open websocket connection - \"\n f\"uri: {provider.endpoint_uri}, method: {method}\"\n )\n await request_func(method, params)\n return await asyncio.wait_for(\n self.ws_recv(),\n timeout=provider.call_timeout,\n )\n\n async def ws_recv(self) -> Any:\n return await self._ws_recv_stream().__anext__()\n\n def persistent_recv_stream(self) -> \"_AsyncPersistentRecvStream\":\n return _AsyncPersistentRecvStream(self)\n\n async def _ws_recv_stream(self) -> AsyncGenerator[RPCResponse, None]:\n if not isinstance(self._provider, PersistentConnectionProvider):\n raise TypeError(\n \"Only websocket providers that maintain an open, persistent connection \"\n \"can listen to websocket recv streams.\"\n )\n\n response = json.loads(\n await asyncio.wait_for(\n self._provider.ws.recv(),\n timeout=self._provider.call_timeout,\n )\n )\n request_info = self._provider._get_request_information_for_response(response)\n\n if request_info is None:\n self.logger.debug(\"No cache key found for response, returning raw response\")\n yield response\n\n else:\n if request_info.method == \"eth_subscribe\" and \"result\" in response.keys():\n # if response for the initial eth_subscribe request, which returns the\n # subscription id\n subscription_id = response[\"result\"]\n cache_key = generate_cache_key(subscription_id)\n if cache_key not in self._provider._async_response_processing_cache:\n # cache by subscription id in order to process each response for the\n # subscription as it comes in\n self._provider.logger.debug(\n f\"Caching eth_subscription info:\\n \"\n f\"cache_key={cache_key},\\n \"\n f\"request_info={request_info.__dict__}\"\n )\n self._provider._async_response_processing_cache.cache(\n cache_key, request_info\n )\n\n # pipe response back through middleware response processors\n if len(request_info.middleware_response_processors) > 0:\n response = pipe(response, *request_info.middleware_response_processors)\n\n (\n result_formatters,\n error_formatters,\n null_formatters,\n ) = request_info.response_formatters\n partly_formatted_response = self.formatted_response(\n response,\n request_info.params,\n error_formatters,\n null_formatters,\n )\n yield apply_result_formatters(result_formatters, partly_formatted_response)\n\n\nclass _AsyncPersistentRecvStream:\n \"\"\"\n Async generator for receiving responses from a persistent connection. This\n abstraction is necessary to define the `__aiter__()` method required for\n use with \"async for\" loops.\n \"\"\"\n\n def __init__(self, manager: RequestManager, *args: Any, **kwargs: Any) -> None:\n self.manager = manager\n super().__init__(*args, **kwargs)\n\n def __aiter__(self) -> AsyncGenerator[RPCResponse, None]:\n while True:\n try:\n return self.manager._ws_recv_stream()\n except ConnectionClosedOK:\n pass\n", "path": "web3/manager.py"}, {"content": "from .websocket import ( # noqa: F401\n DEFAULT_WEBSOCKET_TIMEOUT,\n RESTRICTED_WEBSOCKET_KWARGS,\n WebsocketProvider,\n)\nfrom .websocket_v2 import ( # noqa: F401\n WebsocketProviderV2,\n)\n", "path": "web3/providers/websocket/__init__.py"}]} | 3,949 | 787 |
gh_patches_debug_41475 | rasdani/github-patches | git_diff | automl__auto-sklearn-1407 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
run_wrapper() got an unexpected keyword argument 'pure'
While running a fitting on a classifier (*output[0]* with data *output[1]* to *output[4]*), I get the following error; I think it's packages issue;
> output[0].fit(output[1], output[2], output[3], output[4])
> File "/usr/local/lib/python3.8/dist-packages/autosklearn/estimators.py", line 1045, in fit
super().fit(
> File "/usr/local/lib/python3.8/dist-packages/autosklearn/estimators.py", line 375, in fit
self.automl_.fit(load_models=self.load_models, **kwargs)
> File "/usr/local/lib/python3.8/dist-packages/autosklearn/automl.py", line 2056, in fit
return super().fit(
> File "/usr/local/lib/python3.8/dist-packages/autosklearn/automl.py", line 931, in fit
_proc_smac.run_smbo()
> File "/usr/local/lib/python3.8/dist-packages/autosklearn/smbo.py", line 498, in run_smbo
smac.optimize()
> File "/usr/local/lib/python3.8/dist-packages/smac/facade/smac_ac_facade.py", line 720, in optimize
incumbent = self.solver.run()
> File "/usr/local/lib/python3.8/dist-packages/smac/optimizer/smbo.py", line 287, in run
self.tae_runner.submit_run(run_info=run_info)
> File "/usr/local/lib/python3.8/dist-packages/smac/tae/dask_runner.py", line 166, in submit_run
self.client.submit(
> File "/usr/local/lib/python3.8/dist-packages/autosklearn/util/single_thread_client.py", line 59, in submit
return DummyFuture(func(*args, **kwargs))
> **TypeError: run_wrapper() got an unexpected keyword argument 'pure'**
It seems like it's an error with Dask. Here are the installed packages on ubuntu 18.04
pandas==1.3.0
scikit-learn==0.24
dask==2021.12.0
auto-sklearn==0.14.5 #AutoML
tensorflow==2.8.0
I've tried all versions of dask from 2021.12.0 to 2022.02.0 (Current) and nothing seems to work. Downgrading to auto-sklearn 0.14.4 and lower didn't solve the problem.
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `autosklearn/util/single_thread_client.py`
Content:
```
1 import typing
2 from pathlib import Path
3
4 import dask.distributed
5
6
7 class DummyFuture(dask.distributed.Future):
8 """
9 A class that mimics a distributed Future, the outcome of
10 performing submit on a distributed client.
11 """
12 def __init__(self, result: typing.Any) -> None:
13 self._result = result # type: typing.Any
14
15 def result(self, timeout: typing.Optional[int] = None) -> typing.Any:
16 return self._result
17
18 def cancel(self) -> None:
19 pass
20
21 def done(self) -> bool:
22 return True
23
24 def __repr__(self) -> str:
25 return "DummyFuture: {}".format(self._result)
26
27 def __del__(self) -> None:
28 pass
29
30
31 class SingleThreadedClient(dask.distributed.Client):
32 """
33 A class to Mock the Distributed Client class, in case
34 Auto-Sklearn is meant to run in the current Thread.
35 """
36 def __init__(self) -> None:
37
38 # Raise a not implemented error if using a method from Client
39 implemented_methods = ['submit', 'close', 'shutdown', 'write_scheduler_file',
40 '_get_scheduler_info', 'nthreads']
41 method_list = [func for func in dir(dask.distributed.Client) if callable(
42 getattr(dask.distributed.Client, func)) and not func.startswith('__')]
43 for method in method_list:
44 if method in implemented_methods:
45 continue
46 setattr(self, method, self._unsupported_method)
47 pass
48
49 def _unsupported_method(self) -> None:
50 raise NotImplementedError()
51
52 def submit(
53 self,
54 func: typing.Callable,
55 *args: typing.List,
56 priority: int = 0,
57 **kwargs: typing.Dict,
58 ) -> typing.Any:
59 return DummyFuture(func(*args, **kwargs))
60
61 def close(self) -> None:
62 pass
63
64 def shutdown(self) -> None:
65 pass
66
67 def write_scheduler_file(self, scheduler_file: str) -> None:
68 Path(scheduler_file).touch()
69 return
70
71 def _get_scheduler_info(self) -> typing.Dict:
72 return {
73 'workers': ['127.0.0.1'],
74 'type': 'Scheduler',
75 }
76
77 def nthreads(self) -> typing.Dict:
78 return {
79 '127.0.0.1': 1,
80 }
81
82 def __repr__(self) -> str:
83 return 'SingleThreadedClient()'
84
85 def __del__(self) -> None:
86 pass
87
```
Path: `autosklearn/__version__.py`
Content:
```
1 """Version information."""
2
3 # The following line *must* be the last in the module, exactly as formatted:
4 __version__ = "0.14.4"
5
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/autosklearn/__version__.py b/autosklearn/__version__.py
--- a/autosklearn/__version__.py
+++ b/autosklearn/__version__.py
@@ -1,4 +1,4 @@
"""Version information."""
# The following line *must* be the last in the module, exactly as formatted:
-__version__ = "0.14.4"
+__version__ = "0.14.6"
diff --git a/autosklearn/util/single_thread_client.py b/autosklearn/util/single_thread_client.py
--- a/autosklearn/util/single_thread_client.py
+++ b/autosklearn/util/single_thread_client.py
@@ -1,5 +1,6 @@
import typing
from pathlib import Path
+from typing import Any
import dask.distributed
@@ -9,6 +10,7 @@
A class that mimics a distributed Future, the outcome of
performing submit on a distributed client.
"""
+
def __init__(self, result: typing.Any) -> None:
self._result = result # type: typing.Any
@@ -33,13 +35,24 @@
A class to Mock the Distributed Client class, in case
Auto-Sklearn is meant to run in the current Thread.
"""
+
def __init__(self) -> None:
# Raise a not implemented error if using a method from Client
- implemented_methods = ['submit', 'close', 'shutdown', 'write_scheduler_file',
- '_get_scheduler_info', 'nthreads']
- method_list = [func for func in dir(dask.distributed.Client) if callable(
- getattr(dask.distributed.Client, func)) and not func.startswith('__')]
+ implemented_methods = [
+ "submit",
+ "close",
+ "shutdown",
+ "write_scheduler_file",
+ "_get_scheduler_info",
+ "nthreads",
+ ]
+ method_list = [
+ func
+ for func in dir(dask.distributed.Client)
+ if callable(getattr(dask.distributed.Client, func))
+ and not func.startswith("__")
+ ]
for method in method_list:
if method in implemented_methods:
continue
@@ -54,8 +67,24 @@
func: typing.Callable,
*args: typing.List,
priority: int = 0,
- **kwargs: typing.Dict,
+ key: Any = None,
+ workers: Any = None,
+ resources: Any = None,
+ retries: Any = None,
+ fifo_timeout: Any = "100 ms",
+ allow_other_workers: Any = False,
+ actor: Any = False,
+ actors: Any = False,
+ pure: Any = None,
+ **kwargs: Any,
) -> typing.Any:
+ """
+ Note
+ ----
+ The keyword arguments caught in `dask.distributed.Client` need to
+ be specified here so they don't get passed in as ``**kwargs`` to the
+ ``func``.
+ """
return DummyFuture(func(*args, **kwargs))
def close(self) -> None:
@@ -70,17 +99,17 @@
def _get_scheduler_info(self) -> typing.Dict:
return {
- 'workers': ['127.0.0.1'],
- 'type': 'Scheduler',
+ "workers": ["127.0.0.1"],
+ "type": "Scheduler",
}
def nthreads(self) -> typing.Dict:
return {
- '127.0.0.1': 1,
+ "127.0.0.1": 1,
}
def __repr__(self) -> str:
- return 'SingleThreadedClient()'
+ return "SingleThreadedClient()"
def __del__(self) -> None:
pass
| {"golden_diff": "diff --git a/autosklearn/__version__.py b/autosklearn/__version__.py\n--- a/autosklearn/__version__.py\n+++ b/autosklearn/__version__.py\n@@ -1,4 +1,4 @@\n \"\"\"Version information.\"\"\"\n \n # The following line *must* be the last in the module, exactly as formatted:\n-__version__ = \"0.14.4\"\n+__version__ = \"0.14.6\"\ndiff --git a/autosklearn/util/single_thread_client.py b/autosklearn/util/single_thread_client.py\n--- a/autosklearn/util/single_thread_client.py\n+++ b/autosklearn/util/single_thread_client.py\n@@ -1,5 +1,6 @@\n import typing\n from pathlib import Path\n+from typing import Any\n \n import dask.distributed\n \n@@ -9,6 +10,7 @@\n A class that mimics a distributed Future, the outcome of\n performing submit on a distributed client.\n \"\"\"\n+\n def __init__(self, result: typing.Any) -> None:\n self._result = result # type: typing.Any\n \n@@ -33,13 +35,24 @@\n A class to Mock the Distributed Client class, in case\n Auto-Sklearn is meant to run in the current Thread.\n \"\"\"\n+\n def __init__(self) -> None:\n \n # Raise a not implemented error if using a method from Client\n- implemented_methods = ['submit', 'close', 'shutdown', 'write_scheduler_file',\n- '_get_scheduler_info', 'nthreads']\n- method_list = [func for func in dir(dask.distributed.Client) if callable(\n- getattr(dask.distributed.Client, func)) and not func.startswith('__')]\n+ implemented_methods = [\n+ \"submit\",\n+ \"close\",\n+ \"shutdown\",\n+ \"write_scheduler_file\",\n+ \"_get_scheduler_info\",\n+ \"nthreads\",\n+ ]\n+ method_list = [\n+ func\n+ for func in dir(dask.distributed.Client)\n+ if callable(getattr(dask.distributed.Client, func))\n+ and not func.startswith(\"__\")\n+ ]\n for method in method_list:\n if method in implemented_methods:\n continue\n@@ -54,8 +67,24 @@\n func: typing.Callable,\n *args: typing.List,\n priority: int = 0,\n- **kwargs: typing.Dict,\n+ key: Any = None,\n+ workers: Any = None,\n+ resources: Any = None,\n+ retries: Any = None,\n+ fifo_timeout: Any = \"100 ms\",\n+ allow_other_workers: Any = False,\n+ actor: Any = False,\n+ actors: Any = False,\n+ pure: Any = None,\n+ **kwargs: Any,\n ) -> typing.Any:\n+ \"\"\"\n+ Note\n+ ----\n+ The keyword arguments caught in `dask.distributed.Client` need to\n+ be specified here so they don't get passed in as ``**kwargs`` to the\n+ ``func``.\n+ \"\"\"\n return DummyFuture(func(*args, **kwargs))\n \n def close(self) -> None:\n@@ -70,17 +99,17 @@\n \n def _get_scheduler_info(self) -> typing.Dict:\n return {\n- 'workers': ['127.0.0.1'],\n- 'type': 'Scheduler',\n+ \"workers\": [\"127.0.0.1\"],\n+ \"type\": \"Scheduler\",\n }\n \n def nthreads(self) -> typing.Dict:\n return {\n- '127.0.0.1': 1,\n+ \"127.0.0.1\": 1,\n }\n \n def __repr__(self) -> str:\n- return 'SingleThreadedClient()'\n+ return \"SingleThreadedClient()\"\n \n def __del__(self) -> None:\n pass\n", "issue": "run_wrapper() got an unexpected keyword argument 'pure'\nWhile running a fitting on a classifier (*output[0]* with data *output[1]* to *output[4]*), I get the following error; I think it's packages issue;\r\n\r\n > output[0].fit(output[1], output[2], output[3], output[4])\r\n > File \"/usr/local/lib/python3.8/dist-packages/autosklearn/estimators.py\", line 1045, in fit\r\n super().fit(\r\n > File \"/usr/local/lib/python3.8/dist-packages/autosklearn/estimators.py\", line 375, in fit\r\n self.automl_.fit(load_models=self.load_models, **kwargs)\r\n > File \"/usr/local/lib/python3.8/dist-packages/autosklearn/automl.py\", line 2056, in fit\r\n return super().fit(\r\n > File \"/usr/local/lib/python3.8/dist-packages/autosklearn/automl.py\", line 931, in fit\r\n _proc_smac.run_smbo()\r\n > File \"/usr/local/lib/python3.8/dist-packages/autosklearn/smbo.py\", line 498, in run_smbo\r\n smac.optimize()\r\n > File \"/usr/local/lib/python3.8/dist-packages/smac/facade/smac_ac_facade.py\", line 720, in optimize\r\n incumbent = self.solver.run()\r\n > File \"/usr/local/lib/python3.8/dist-packages/smac/optimizer/smbo.py\", line 287, in run\r\n self.tae_runner.submit_run(run_info=run_info)\r\n > File \"/usr/local/lib/python3.8/dist-packages/smac/tae/dask_runner.py\", line 166, in submit_run\r\n self.client.submit(\r\n > File \"/usr/local/lib/python3.8/dist-packages/autosklearn/util/single_thread_client.py\", line 59, in submit\r\n return DummyFuture(func(*args, **kwargs))\r\n > **TypeError: run_wrapper() got an unexpected keyword argument 'pure'**\r\n\r\nIt seems like it's an error with Dask. Here are the installed packages on ubuntu 18.04\r\n\r\npandas==1.3.0\r\nscikit-learn==0.24\r\ndask==2021.12.0\r\nauto-sklearn==0.14.5 #AutoML\r\ntensorflow==2.8.0\r\n\r\nI've tried all versions of dask from 2021.12.0 to 2022.02.0 (Current) and nothing seems to work. Downgrading to auto-sklearn 0.14.4 and lower didn't solve the problem.\n", "before_files": [{"content": "import typing\nfrom pathlib import Path\n\nimport dask.distributed\n\n\nclass DummyFuture(dask.distributed.Future):\n \"\"\"\n A class that mimics a distributed Future, the outcome of\n performing submit on a distributed client.\n \"\"\"\n def __init__(self, result: typing.Any) -> None:\n self._result = result # type: typing.Any\n\n def result(self, timeout: typing.Optional[int] = None) -> typing.Any:\n return self._result\n\n def cancel(self) -> None:\n pass\n\n def done(self) -> bool:\n return True\n\n def __repr__(self) -> str:\n return \"DummyFuture: {}\".format(self._result)\n\n def __del__(self) -> None:\n pass\n\n\nclass SingleThreadedClient(dask.distributed.Client):\n \"\"\"\n A class to Mock the Distributed Client class, in case\n Auto-Sklearn is meant to run in the current Thread.\n \"\"\"\n def __init__(self) -> None:\n\n # Raise a not implemented error if using a method from Client\n implemented_methods = ['submit', 'close', 'shutdown', 'write_scheduler_file',\n '_get_scheduler_info', 'nthreads']\n method_list = [func for func in dir(dask.distributed.Client) if callable(\n getattr(dask.distributed.Client, func)) and not func.startswith('__')]\n for method in method_list:\n if method in implemented_methods:\n continue\n setattr(self, method, self._unsupported_method)\n pass\n\n def _unsupported_method(self) -> None:\n raise NotImplementedError()\n\n def submit(\n self,\n func: typing.Callable,\n *args: typing.List,\n priority: int = 0,\n **kwargs: typing.Dict,\n ) -> typing.Any:\n return DummyFuture(func(*args, **kwargs))\n\n def close(self) -> None:\n pass\n\n def shutdown(self) -> None:\n pass\n\n def write_scheduler_file(self, scheduler_file: str) -> None:\n Path(scheduler_file).touch()\n return\n\n def _get_scheduler_info(self) -> typing.Dict:\n return {\n 'workers': ['127.0.0.1'],\n 'type': 'Scheduler',\n }\n\n def nthreads(self) -> typing.Dict:\n return {\n '127.0.0.1': 1,\n }\n\n def __repr__(self) -> str:\n return 'SingleThreadedClient()'\n\n def __del__(self) -> None:\n pass\n", "path": "autosklearn/util/single_thread_client.py"}, {"content": "\"\"\"Version information.\"\"\"\n\n# The following line *must* be the last in the module, exactly as formatted:\n__version__ = \"0.14.4\"\n", "path": "autosklearn/__version__.py"}], "after_files": [{"content": "import typing\nfrom pathlib import Path\nfrom typing import Any\n\nimport dask.distributed\n\n\nclass DummyFuture(dask.distributed.Future):\n \"\"\"\n A class that mimics a distributed Future, the outcome of\n performing submit on a distributed client.\n \"\"\"\n\n def __init__(self, result: typing.Any) -> None:\n self._result = result # type: typing.Any\n\n def result(self, timeout: typing.Optional[int] = None) -> typing.Any:\n return self._result\n\n def cancel(self) -> None:\n pass\n\n def done(self) -> bool:\n return True\n\n def __repr__(self) -> str:\n return \"DummyFuture: {}\".format(self._result)\n\n def __del__(self) -> None:\n pass\n\n\nclass SingleThreadedClient(dask.distributed.Client):\n \"\"\"\n A class to Mock the Distributed Client class, in case\n Auto-Sklearn is meant to run in the current Thread.\n \"\"\"\n\n def __init__(self) -> None:\n\n # Raise a not implemented error if using a method from Client\n implemented_methods = [\n \"submit\",\n \"close\",\n \"shutdown\",\n \"write_scheduler_file\",\n \"_get_scheduler_info\",\n \"nthreads\",\n ]\n method_list = [\n func\n for func in dir(dask.distributed.Client)\n if callable(getattr(dask.distributed.Client, func))\n and not func.startswith(\"__\")\n ]\n for method in method_list:\n if method in implemented_methods:\n continue\n setattr(self, method, self._unsupported_method)\n pass\n\n def _unsupported_method(self) -> None:\n raise NotImplementedError()\n\n def submit(\n self,\n func: typing.Callable,\n *args: typing.List,\n priority: int = 0,\n key: Any = None,\n workers: Any = None,\n resources: Any = None,\n retries: Any = None,\n fifo_timeout: Any = \"100 ms\",\n allow_other_workers: Any = False,\n actor: Any = False,\n actors: Any = False,\n pure: Any = None,\n **kwargs: Any,\n ) -> typing.Any:\n \"\"\"\n Note\n ----\n The keyword arguments caught in `dask.distributed.Client` need to\n be specified here so they don't get passed in as ``**kwargs`` to the\n ``func``.\n \"\"\"\n return DummyFuture(func(*args, **kwargs))\n\n def close(self) -> None:\n pass\n\n def shutdown(self) -> None:\n pass\n\n def write_scheduler_file(self, scheduler_file: str) -> None:\n Path(scheduler_file).touch()\n return\n\n def _get_scheduler_info(self) -> typing.Dict:\n return {\n \"workers\": [\"127.0.0.1\"],\n \"type\": \"Scheduler\",\n }\n\n def nthreads(self) -> typing.Dict:\n return {\n \"127.0.0.1\": 1,\n }\n\n def __repr__(self) -> str:\n return \"SingleThreadedClient()\"\n\n def __del__(self) -> None:\n pass\n", "path": "autosklearn/util/single_thread_client.py"}, {"content": "\"\"\"Version information.\"\"\"\n\n# The following line *must* be the last in the module, exactly as formatted:\n__version__ = \"0.14.6\"\n", "path": "autosklearn/__version__.py"}]} | 1,634 | 873 |
gh_patches_debug_12360 | rasdani/github-patches | git_diff | akvo__akvo-rsr-2517 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
UnicodeEncodeError with ProjectCustomField
```python
File "django/core/handlers/base.py", line 111, in get_response
response = wrapped_callback(request, *callback_args, **callback_kwargs)
File "django/views/decorators/csrf.py", line 57, in wrapped_view
return view_func(*args, **kwargs)
File "django/views/generic/base.py", line 69, in view
return self.dispatch(request, *args, **kwargs)
File "rest_framework/views.py", line 466, in dispatch
response = self.handle_exception(exc)
File "rest_framework/views.py", line 463, in dispatch
response = handler(request, *args, **kwargs)
File "rest_framework/decorators.py", line 53, in handler
return func(*args, **kwargs)
File "akvo/rest/views/project_editor.py", line 576, in project_editor
'changes': log_changes(changes, user, project),
File "akvo/rest/views/project_editor.py", line 92, in log_changes
object_repr=obj.__unicode__(),
File "akvo/rsr/models/custom_field.py", line 77, in __unicode__
return u'%s' % str(self.value)
```
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `akvo/rsr/models/custom_field.py`
Content:
```
1 # -*- coding: utf-8 -*-
2
3 # Akvo RSR is covered by the GNU Affero General Public License.
4 # See more details in the license.txt file located at the root folder of the Akvo RSR module.
5 # For additional details on the GNU license please see < http://www.gnu.org/licenses/agpl.html >.
6
7
8 from django.db import models
9 from django.utils.translation import ugettext_lazy as _
10
11 from ..fields import ValidXMLCharField, ValidXMLTextField
12
13
14 class ProjectCustomField(models.Model):
15 """
16 Custom fields make it possible for partner to specify additional fields. When specified for a
17 project, the fields will appear in the admin (under the specified section) and can then be
18 filled in.
19
20 Custom fields for a project, linking the project to its' custom fields.
21
22 Name: name of the custom field (label in the admin)
23 Section: the section in the admin where the field should be added
24 Maxlength: the maximum number of characters of the field
25 Help text: the help text belonging to the field
26 Value: the value which can be filled in the project admin.
27 """
28 SECTIONS = (
29 (1, _(u'01 - General information')),
30 (2, _(u'02 - Contact information')),
31 (3, _(u'03 - Project partners')),
32 (4, _(u'04 - Project descriptions')),
33 (5, _(u'05 - Results and indicators')),
34 (6, _(u'06 - Finance')),
35 (7, _(u'07 - Project locations')),
36 (8, _(u'08 - Project focus')),
37 (9, _(u'09 - Links and documents')),
38 (10, _(u'10 - Project comments')),
39 )
40
41 TYPES = (
42 ('text', _(u'Text')),
43 ('boolean', _(u'Checkbox')),
44 )
45
46 project = models.ForeignKey('Project', verbose_name=_(u'project'), related_name='custom_fields')
47 name = ValidXMLCharField(_(u'name'), max_length=255, help_text=_(u'(max 255 characters)'))
48 section = models.IntegerField(
49 _(u'admin section'), choices=SECTIONS,
50 help_text=_(u'Select the section of the admin where the custom field should be displayed')
51 )
52 max_characters = models.IntegerField(
53 _(u'maximum characters'), blank=True, null=True,
54 help_text=_(u'Set the maximum amount of characters that the user is allowed to fill in. '
55 u'Leave empty or fill in 0 if there is no character limit.')
56 )
57 help_text = ValidXMLTextField(
58 _(u'help text'), max_length=1000, blank=True,
59 help_text=_(u'The help text to be displayed with the field in the admin. Leave empty if '
60 u'there is no need for a help text. (max 1000 characters)')
61 )
62 value = ValidXMLTextField(_(u'value'), blank=True)
63 mandatory = models.BooleanField(_(u'mandatory'), default=False,
64 help_text=_(u'Indicate whether this field is mandatory or not'))
65 order = models.PositiveSmallIntegerField(
66 _(u'order'), help_text=_(u'The order of the fields as they will be displayed in the '
67 u'project editor. Must be a positive number, and the lowest '
68 u'number will be shown on top.')
69 )
70 type = ValidXMLCharField(
71 _(u'type'), max_length=20, choices=TYPES, default='text',
72 help_text=_(u'Select the type of custom field. Text will show a text area in the project '
73 u'editor, and checkbox will show a checkbox.')
74 )
75
76 def __unicode__(self):
77 return u'%s' % str(self.value)
78
79
80 class OrganisationCustomField(models.Model):
81 """
82 Custom fields make it possible for partner to specify additional fields. When specified for a
83 project, the fields will appear in the admin (under the specified section) and can then be
84 filled in.
85
86 Custom fields for an organisation, linking the organisation to its' custom fields.
87
88 These custom fields will be used for the projects whenever a user of the organisation
89 creates a new project.
90
91 Name: name of the custom field (label in the admin)
92 Section: the section in the admin where the field should be added
93 Maxlength: the maximum number of characters of the field
94 Help text: the help text belonging to the field
95 """
96 SECTIONS = (
97 (1, _(u'01 - General information')),
98 (2, _(u'02 - Contact information')),
99 (3, _(u'03 - Project partners')),
100 (4, _(u'04 - Project descriptions')),
101 (5, _(u'05 - Results and indicators')),
102 (6, _(u'06 - Finance')),
103 (7, _(u'07 - Project locations')),
104 (8, _(u'08 - Project focus')),
105 (9, _(u'09 - Links and documents')),
106 (10, _(u'10 - Project comments')),
107 )
108
109 TYPES = (
110 ('text', _(u'Text')),
111 ('boolean', _(u'Checkbox')),
112 )
113
114 organisation = models.ForeignKey(
115 'Organisation', verbose_name=_(u'organisation'), related_name='custom_fields'
116 )
117 name = ValidXMLCharField(_(u'name'), max_length=255, help_text=_(u'(max 255 characters)'))
118 section = models.IntegerField(
119 _(u'admin section'), choices=SECTIONS,
120 help_text=_(u'Select the section of the admin where the custom field should be displayed')
121 )
122 max_characters = models.IntegerField(
123 _(u'maximum characters'), blank=True, null=True,
124 help_text=_(u'Set the maximum amount of characters that the user is allowed to fill in. '
125 u'Leave empty or fill in 0 if there is no character limit.')
126 )
127 help_text = ValidXMLTextField(
128 _(u'help text'), max_length=1000, blank=True,
129 help_text=_(u'The help text to be displayed with the field in the admin. Leave empty if '
130 u'there is no need for a help text. (max 1000 characters)')
131 )
132 mandatory = models.BooleanField(_(u'mandatory'), default=False,
133 help_text=_(u'Indicate whether this field is mandatory or not'))
134 order = models.PositiveSmallIntegerField(
135 _(u'order'), help_text=_(u'The order of the fields as they will be displayed in the '
136 u'project editor. Must be a positive number, and the lowest '
137 u'number will be shown on top.')
138 )
139 type = ValidXMLCharField(
140 _(u'type'), max_length=20, choices=TYPES, default='text',
141 help_text=_(u'Select the type of custom field. Text will show a text area in the project '
142 u'editor, and checkbox will show a checkbox.')
143 )
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/akvo/rsr/models/custom_field.py b/akvo/rsr/models/custom_field.py
--- a/akvo/rsr/models/custom_field.py
+++ b/akvo/rsr/models/custom_field.py
@@ -74,7 +74,7 @@
)
def __unicode__(self):
- return u'%s' % str(self.value)
+ return u'%s' % self.value
class OrganisationCustomField(models.Model):
@@ -140,4 +140,4 @@
_(u'type'), max_length=20, choices=TYPES, default='text',
help_text=_(u'Select the type of custom field. Text will show a text area in the project '
u'editor, and checkbox will show a checkbox.')
- )
\ No newline at end of file
+ )
| {"golden_diff": "diff --git a/akvo/rsr/models/custom_field.py b/akvo/rsr/models/custom_field.py\n--- a/akvo/rsr/models/custom_field.py\n+++ b/akvo/rsr/models/custom_field.py\n@@ -74,7 +74,7 @@\n )\n \n def __unicode__(self):\n- return u'%s' % str(self.value)\n+ return u'%s' % self.value\n \n \n class OrganisationCustomField(models.Model):\n@@ -140,4 +140,4 @@\n _(u'type'), max_length=20, choices=TYPES, default='text',\n help_text=_(u'Select the type of custom field. Text will show a text area in the project '\n u'editor, and checkbox will show a checkbox.')\n- )\n\\ No newline at end of file\n+ )\n", "issue": "UnicodeEncodeError with ProjectCustomField\n```python\r\n\r\n File \"django/core/handlers/base.py\", line 111, in get_response\r\n response = wrapped_callback(request, *callback_args, **callback_kwargs)\r\n File \"django/views/decorators/csrf.py\", line 57, in wrapped_view\r\n return view_func(*args, **kwargs)\r\n File \"django/views/generic/base.py\", line 69, in view\r\n return self.dispatch(request, *args, **kwargs)\r\n File \"rest_framework/views.py\", line 466, in dispatch\r\n response = self.handle_exception(exc)\r\n File \"rest_framework/views.py\", line 463, in dispatch\r\n response = handler(request, *args, **kwargs)\r\n File \"rest_framework/decorators.py\", line 53, in handler\r\n return func(*args, **kwargs)\r\n File \"akvo/rest/views/project_editor.py\", line 576, in project_editor\r\n 'changes': log_changes(changes, user, project),\r\n File \"akvo/rest/views/project_editor.py\", line 92, in log_changes\r\n object_repr=obj.__unicode__(),\r\n File \"akvo/rsr/models/custom_field.py\", line 77, in __unicode__\r\n return u'%s' % str(self.value)\r\n```\n", "before_files": [{"content": "# -*- coding: utf-8 -*-\n\n# Akvo RSR is covered by the GNU Affero General Public License.\n# See more details in the license.txt file located at the root folder of the Akvo RSR module.\n# For additional details on the GNU license please see < http://www.gnu.org/licenses/agpl.html >.\n\n\nfrom django.db import models\nfrom django.utils.translation import ugettext_lazy as _\n\nfrom ..fields import ValidXMLCharField, ValidXMLTextField\n\n\nclass ProjectCustomField(models.Model):\n \"\"\"\n Custom fields make it possible for partner to specify additional fields. When specified for a\n project, the fields will appear in the admin (under the specified section) and can then be\n filled in.\n\n Custom fields for a project, linking the project to its' custom fields.\n\n Name: name of the custom field (label in the admin)\n Section: the section in the admin where the field should be added\n Maxlength: the maximum number of characters of the field\n Help text: the help text belonging to the field\n Value: the value which can be filled in the project admin.\n \"\"\"\n SECTIONS = (\n (1, _(u'01 - General information')),\n (2, _(u'02 - Contact information')),\n (3, _(u'03 - Project partners')),\n (4, _(u'04 - Project descriptions')),\n (5, _(u'05 - Results and indicators')),\n (6, _(u'06 - Finance')),\n (7, _(u'07 - Project locations')),\n (8, _(u'08 - Project focus')),\n (9, _(u'09 - Links and documents')),\n (10, _(u'10 - Project comments')),\n )\n\n TYPES = (\n ('text', _(u'Text')),\n ('boolean', _(u'Checkbox')),\n )\n\n project = models.ForeignKey('Project', verbose_name=_(u'project'), related_name='custom_fields')\n name = ValidXMLCharField(_(u'name'), max_length=255, help_text=_(u'(max 255 characters)'))\n section = models.IntegerField(\n _(u'admin section'), choices=SECTIONS,\n help_text=_(u'Select the section of the admin where the custom field should be displayed')\n )\n max_characters = models.IntegerField(\n _(u'maximum characters'), blank=True, null=True,\n help_text=_(u'Set the maximum amount of characters that the user is allowed to fill in. '\n u'Leave empty or fill in 0 if there is no character limit.')\n )\n help_text = ValidXMLTextField(\n _(u'help text'), max_length=1000, blank=True,\n help_text=_(u'The help text to be displayed with the field in the admin. Leave empty if '\n u'there is no need for a help text. (max 1000 characters)')\n )\n value = ValidXMLTextField(_(u'value'), blank=True)\n mandatory = models.BooleanField(_(u'mandatory'), default=False,\n help_text=_(u'Indicate whether this field is mandatory or not'))\n order = models.PositiveSmallIntegerField(\n _(u'order'), help_text=_(u'The order of the fields as they will be displayed in the '\n u'project editor. Must be a positive number, and the lowest '\n u'number will be shown on top.')\n )\n type = ValidXMLCharField(\n _(u'type'), max_length=20, choices=TYPES, default='text',\n help_text=_(u'Select the type of custom field. Text will show a text area in the project '\n u'editor, and checkbox will show a checkbox.')\n )\n\n def __unicode__(self):\n return u'%s' % str(self.value)\n\n\nclass OrganisationCustomField(models.Model):\n \"\"\"\n Custom fields make it possible for partner to specify additional fields. When specified for a\n project, the fields will appear in the admin (under the specified section) and can then be\n filled in.\n\n Custom fields for an organisation, linking the organisation to its' custom fields.\n\n These custom fields will be used for the projects whenever a user of the organisation\n creates a new project.\n\n Name: name of the custom field (label in the admin)\n Section: the section in the admin where the field should be added\n Maxlength: the maximum number of characters of the field\n Help text: the help text belonging to the field\n \"\"\"\n SECTIONS = (\n (1, _(u'01 - General information')),\n (2, _(u'02 - Contact information')),\n (3, _(u'03 - Project partners')),\n (4, _(u'04 - Project descriptions')),\n (5, _(u'05 - Results and indicators')),\n (6, _(u'06 - Finance')),\n (7, _(u'07 - Project locations')),\n (8, _(u'08 - Project focus')),\n (9, _(u'09 - Links and documents')),\n (10, _(u'10 - Project comments')),\n )\n\n TYPES = (\n ('text', _(u'Text')),\n ('boolean', _(u'Checkbox')),\n )\n\n organisation = models.ForeignKey(\n 'Organisation', verbose_name=_(u'organisation'), related_name='custom_fields'\n )\n name = ValidXMLCharField(_(u'name'), max_length=255, help_text=_(u'(max 255 characters)'))\n section = models.IntegerField(\n _(u'admin section'), choices=SECTIONS,\n help_text=_(u'Select the section of the admin where the custom field should be displayed')\n )\n max_characters = models.IntegerField(\n _(u'maximum characters'), blank=True, null=True,\n help_text=_(u'Set the maximum amount of characters that the user is allowed to fill in. '\n u'Leave empty or fill in 0 if there is no character limit.')\n )\n help_text = ValidXMLTextField(\n _(u'help text'), max_length=1000, blank=True,\n help_text=_(u'The help text to be displayed with the field in the admin. Leave empty if '\n u'there is no need for a help text. (max 1000 characters)')\n )\n mandatory = models.BooleanField(_(u'mandatory'), default=False,\n help_text=_(u'Indicate whether this field is mandatory or not'))\n order = models.PositiveSmallIntegerField(\n _(u'order'), help_text=_(u'The order of the fields as they will be displayed in the '\n u'project editor. Must be a positive number, and the lowest '\n u'number will be shown on top.')\n )\n type = ValidXMLCharField(\n _(u'type'), max_length=20, choices=TYPES, default='text',\n help_text=_(u'Select the type of custom field. Text will show a text area in the project '\n u'editor, and checkbox will show a checkbox.')\n )", "path": "akvo/rsr/models/custom_field.py"}], "after_files": [{"content": "# -*- coding: utf-8 -*-\n\n# Akvo RSR is covered by the GNU Affero General Public License.\n# See more details in the license.txt file located at the root folder of the Akvo RSR module.\n# For additional details on the GNU license please see < http://www.gnu.org/licenses/agpl.html >.\n\n\nfrom django.db import models\nfrom django.utils.translation import ugettext_lazy as _\n\nfrom ..fields import ValidXMLCharField, ValidXMLTextField\n\n\nclass ProjectCustomField(models.Model):\n \"\"\"\n Custom fields make it possible for partner to specify additional fields. When specified for a\n project, the fields will appear in the admin (under the specified section) and can then be\n filled in.\n\n Custom fields for a project, linking the project to its' custom fields.\n\n Name: name of the custom field (label in the admin)\n Section: the section in the admin where the field should be added\n Maxlength: the maximum number of characters of the field\n Help text: the help text belonging to the field\n Value: the value which can be filled in the project admin.\n \"\"\"\n SECTIONS = (\n (1, _(u'01 - General information')),\n (2, _(u'02 - Contact information')),\n (3, _(u'03 - Project partners')),\n (4, _(u'04 - Project descriptions')),\n (5, _(u'05 - Results and indicators')),\n (6, _(u'06 - Finance')),\n (7, _(u'07 - Project locations')),\n (8, _(u'08 - Project focus')),\n (9, _(u'09 - Links and documents')),\n (10, _(u'10 - Project comments')),\n )\n\n TYPES = (\n ('text', _(u'Text')),\n ('boolean', _(u'Checkbox')),\n )\n\n project = models.ForeignKey('Project', verbose_name=_(u'project'), related_name='custom_fields')\n name = ValidXMLCharField(_(u'name'), max_length=255, help_text=_(u'(max 255 characters)'))\n section = models.IntegerField(\n _(u'admin section'), choices=SECTIONS,\n help_text=_(u'Select the section of the admin where the custom field should be displayed')\n )\n max_characters = models.IntegerField(\n _(u'maximum characters'), blank=True, null=True,\n help_text=_(u'Set the maximum amount of characters that the user is allowed to fill in. '\n u'Leave empty or fill in 0 if there is no character limit.')\n )\n help_text = ValidXMLTextField(\n _(u'help text'), max_length=1000, blank=True,\n help_text=_(u'The help text to be displayed with the field in the admin. Leave empty if '\n u'there is no need for a help text. (max 1000 characters)')\n )\n value = ValidXMLTextField(_(u'value'), blank=True)\n mandatory = models.BooleanField(_(u'mandatory'), default=False,\n help_text=_(u'Indicate whether this field is mandatory or not'))\n order = models.PositiveSmallIntegerField(\n _(u'order'), help_text=_(u'The order of the fields as they will be displayed in the '\n u'project editor. Must be a positive number, and the lowest '\n u'number will be shown on top.')\n )\n type = ValidXMLCharField(\n _(u'type'), max_length=20, choices=TYPES, default='text',\n help_text=_(u'Select the type of custom field. Text will show a text area in the project '\n u'editor, and checkbox will show a checkbox.')\n )\n\n def __unicode__(self):\n return u'%s' % self.value\n\n\nclass OrganisationCustomField(models.Model):\n \"\"\"\n Custom fields make it possible for partner to specify additional fields. When specified for a\n project, the fields will appear in the admin (under the specified section) and can then be\n filled in.\n\n Custom fields for an organisation, linking the organisation to its' custom fields.\n\n These custom fields will be used for the projects whenever a user of the organisation\n creates a new project.\n\n Name: name of the custom field (label in the admin)\n Section: the section in the admin where the field should be added\n Maxlength: the maximum number of characters of the field\n Help text: the help text belonging to the field\n \"\"\"\n SECTIONS = (\n (1, _(u'01 - General information')),\n (2, _(u'02 - Contact information')),\n (3, _(u'03 - Project partners')),\n (4, _(u'04 - Project descriptions')),\n (5, _(u'05 - Results and indicators')),\n (6, _(u'06 - Finance')),\n (7, _(u'07 - Project locations')),\n (8, _(u'08 - Project focus')),\n (9, _(u'09 - Links and documents')),\n (10, _(u'10 - Project comments')),\n )\n\n TYPES = (\n ('text', _(u'Text')),\n ('boolean', _(u'Checkbox')),\n )\n\n organisation = models.ForeignKey(\n 'Organisation', verbose_name=_(u'organisation'), related_name='custom_fields'\n )\n name = ValidXMLCharField(_(u'name'), max_length=255, help_text=_(u'(max 255 characters)'))\n section = models.IntegerField(\n _(u'admin section'), choices=SECTIONS,\n help_text=_(u'Select the section of the admin where the custom field should be displayed')\n )\n max_characters = models.IntegerField(\n _(u'maximum characters'), blank=True, null=True,\n help_text=_(u'Set the maximum amount of characters that the user is allowed to fill in. '\n u'Leave empty or fill in 0 if there is no character limit.')\n )\n help_text = ValidXMLTextField(\n _(u'help text'), max_length=1000, blank=True,\n help_text=_(u'The help text to be displayed with the field in the admin. Leave empty if '\n u'there is no need for a help text. (max 1000 characters)')\n )\n mandatory = models.BooleanField(_(u'mandatory'), default=False,\n help_text=_(u'Indicate whether this field is mandatory or not'))\n order = models.PositiveSmallIntegerField(\n _(u'order'), help_text=_(u'The order of the fields as they will be displayed in the '\n u'project editor. Must be a positive number, and the lowest '\n u'number will be shown on top.')\n )\n type = ValidXMLCharField(\n _(u'type'), max_length=20, choices=TYPES, default='text',\n help_text=_(u'Select the type of custom field. Text will show a text area in the project '\n u'editor, and checkbox will show a checkbox.')\n )\n", "path": "akvo/rsr/models/custom_field.py"}]} | 2,406 | 185 |
gh_patches_debug_16694 | rasdani/github-patches | git_diff | mitmproxy__mitmproxy-2069 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Entering Palette options crashes mitmproxy
##### Steps to reproduce the problem:
1. Press 'O' for options
2. Select 'Palette'
3. mitmproxy will crash
##### Any other comments? What have you tried so far?
```
Traceback (most recent call last):
File "/usr/local/Cellar/mitmproxy/2.0.0/libexec/lib/python3.6/site-packages/mitmproxy/tools/console/master.py", line 281, in run
self.loop.run()
File "/usr/local/Cellar/mitmproxy/2.0.0/libexec/lib/python3.6/site-packages/urwid/main_loop.py", line 278, in run
self._run()
File "/usr/local/Cellar/mitmproxy/2.0.0/libexec/lib/python3.6/site-packages/urwid/main_loop.py", line 376, in _run
self.event_loop.run()
File "/usr/local/Cellar/mitmproxy/2.0.0/libexec/lib/python3.6/site-packages/urwid/main_loop.py", line 682, in run
self._loop()
File "/usr/local/Cellar/mitmproxy/2.0.0/libexec/lib/python3.6/site-packages/urwid/main_loop.py", line 719, in _loop
self._watch_files[fd]()
File "/usr/local/Cellar/mitmproxy/2.0.0/libexec/lib/python3.6/site-packages/urwid/raw_display.py", line 393, in <lambda>
event_loop, callback, self.get_available_raw_input())
File "/usr/local/Cellar/mitmproxy/2.0.0/libexec/lib/python3.6/site-packages/urwid/raw_display.py", line 493, in parse_input
callback(processed, processed_codes)
File "/usr/local/Cellar/mitmproxy/2.0.0/libexec/lib/python3.6/site-packages/urwid/main_loop.py", line 403, in _update
self.process_input(keys)
File "/usr/local/Cellar/mitmproxy/2.0.0/libexec/lib/python3.6/site-packages/urwid/main_loop.py", line 503, in process_input
k = self._topmost_widget.keypress(self.screen_size, k)
File "/usr/local/Cellar/mitmproxy/2.0.0/libexec/lib/python3.6/site-packages/mitmproxy/tools/console/window.py", line 84, in keypress
k = super().keypress(size, k)
File "/usr/local/Cellar/mitmproxy/2.0.0/libexec/lib/python3.6/site-packages/urwid/container.py", line 1128, in keypress
return self.body.keypress( (maxcol, remaining), key )
File "/usr/local/Cellar/mitmproxy/2.0.0/libexec/lib/python3.6/site-packages/urwid/container.py", line 1128, in keypress
return self.body.keypress( (maxcol, remaining), key )
File "/usr/local/Cellar/mitmproxy/2.0.0/libexec/lib/python3.6/site-packages/mitmproxy/tools/console/select.py", line 114, in keypress
self.get_focus()[0].option.activate()
File "/usr/local/Cellar/mitmproxy/2.0.0/libexec/lib/python3.6/site-packages/mitmproxy/tools/console/palettepicker.py", line 46, in <lambda>
lambda: setattr(self.master.options, "palette", name)
File "/usr/local/Cellar/mitmproxy/2.0.0/libexec/lib/python3.6/site-packages/mitmproxy/optmanager.py", line 114, in __setattr__
self.update(**{attr: value})
File "/usr/local/Cellar/mitmproxy/2.0.0/libexec/lib/python3.6/site-packages/mitmproxy/optmanager.py", line 141, in update
raise KeyError("No such option: %s" % k)
KeyError: 'No such option: palette'
```
The option names in mitmproxy/options.py were prefixed with 'console_', but line 46 and line 62 of mitmproxy/tools/console/palettepicker.py were not updated to include this prefix.
This appears to have been broken by commit [35aff3b](https://github.com/mitmproxy/mitmproxy/commit/35aff3b7838f8df718cc574d2643f1355849fa8e)
##### System information
Mitmproxy version: 2.0.0 (release version)
Python version: 3.6.0
Platform: Darwin-16.4.0-x86_64-i386-64bit
SSL version: OpenSSL 1.1.0e 16 Feb 2017
Mac version: 10.12.3 ('', '', '') x86_64
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `mitmproxy/tools/console/palettepicker.py`
Content:
```
1 import urwid
2
3 from mitmproxy.tools.console import common
4 from mitmproxy.tools.console import palettes
5 from mitmproxy.tools.console import select
6
7 footer = [
8 ('heading_key', "enter/space"), ":select",
9 ]
10
11
12 def _mkhelp():
13 text = []
14 keys = [
15 ("enter/space", "select"),
16 ]
17 text.extend(common.format_keyvals(keys, key="key", val="text", indent=4))
18 return text
19
20
21 help_context = _mkhelp()
22
23
24 class PalettePicker(urwid.WidgetWrap):
25
26 def __init__(self, master):
27 self.master = master
28 low, high = [], []
29 for k, v in palettes.palettes.items():
30 if v.high:
31 high.append(k)
32 else:
33 low.append(k)
34 high.sort()
35 low.sort()
36
37 options = [
38 select.Heading("High Colour")
39 ]
40
41 def mkopt(name):
42 return select.Option(
43 i,
44 None,
45 lambda: self.master.options.console_palette == name,
46 lambda: setattr(self.master.options, "palette", name)
47 )
48
49 for i in high:
50 options.append(mkopt(i))
51 options.append(select.Heading("Low Colour"))
52 for i in low:
53 options.append(mkopt(i))
54
55 options.extend(
56 [
57 select.Heading("Options"),
58 select.Option(
59 "Transparent",
60 "T",
61 lambda: master.options.console_palette_transparent,
62 master.options.toggler("palette_transparent")
63 )
64 ]
65 )
66
67 self.lb = select.Select(options)
68 title = urwid.Text("Palettes")
69 title = urwid.Padding(title, align="left", width=("relative", 100))
70 title = urwid.AttrWrap(title, "heading")
71 self._w = urwid.Frame(
72 self.lb,
73 header = title
74 )
75 master.options.changed.connect(self.sig_options_changed)
76
77 def sig_options_changed(self, options, updated):
78 self.lb.walker._modified()
79
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/mitmproxy/tools/console/palettepicker.py b/mitmproxy/tools/console/palettepicker.py
--- a/mitmproxy/tools/console/palettepicker.py
+++ b/mitmproxy/tools/console/palettepicker.py
@@ -43,7 +43,7 @@
i,
None,
lambda: self.master.options.console_palette == name,
- lambda: setattr(self.master.options, "palette", name)
+ lambda: setattr(self.master.options, "console_palette", name)
)
for i in high:
@@ -59,7 +59,7 @@
"Transparent",
"T",
lambda: master.options.console_palette_transparent,
- master.options.toggler("palette_transparent")
+ master.options.toggler("console_palette_transparent")
)
]
)
| {"golden_diff": "diff --git a/mitmproxy/tools/console/palettepicker.py b/mitmproxy/tools/console/palettepicker.py\n--- a/mitmproxy/tools/console/palettepicker.py\n+++ b/mitmproxy/tools/console/palettepicker.py\n@@ -43,7 +43,7 @@\n i,\n None,\n lambda: self.master.options.console_palette == name,\n- lambda: setattr(self.master.options, \"palette\", name)\n+ lambda: setattr(self.master.options, \"console_palette\", name)\n )\n \n for i in high:\n@@ -59,7 +59,7 @@\n \"Transparent\",\n \"T\",\n lambda: master.options.console_palette_transparent,\n- master.options.toggler(\"palette_transparent\")\n+ master.options.toggler(\"console_palette_transparent\")\n )\n ]\n )\n", "issue": "Entering Palette options crashes mitmproxy\n##### Steps to reproduce the problem:\r\n\r\n1. Press 'O' for options\r\n2. Select 'Palette'\r\n3. mitmproxy will crash\r\n\r\n\r\n##### Any other comments? What have you tried so far?\r\n```\r\nTraceback (most recent call last):\r\n File \"/usr/local/Cellar/mitmproxy/2.0.0/libexec/lib/python3.6/site-packages/mitmproxy/tools/console/master.py\", line 281, in run\r\n self.loop.run()\r\n File \"/usr/local/Cellar/mitmproxy/2.0.0/libexec/lib/python3.6/site-packages/urwid/main_loop.py\", line 278, in run\r\n self._run()\r\n File \"/usr/local/Cellar/mitmproxy/2.0.0/libexec/lib/python3.6/site-packages/urwid/main_loop.py\", line 376, in _run\r\n self.event_loop.run()\r\n File \"/usr/local/Cellar/mitmproxy/2.0.0/libexec/lib/python3.6/site-packages/urwid/main_loop.py\", line 682, in run\r\n self._loop()\r\n File \"/usr/local/Cellar/mitmproxy/2.0.0/libexec/lib/python3.6/site-packages/urwid/main_loop.py\", line 719, in _loop\r\n self._watch_files[fd]()\r\n File \"/usr/local/Cellar/mitmproxy/2.0.0/libexec/lib/python3.6/site-packages/urwid/raw_display.py\", line 393, in <lambda>\r\n event_loop, callback, self.get_available_raw_input())\r\n File \"/usr/local/Cellar/mitmproxy/2.0.0/libexec/lib/python3.6/site-packages/urwid/raw_display.py\", line 493, in parse_input\r\n callback(processed, processed_codes)\r\n File \"/usr/local/Cellar/mitmproxy/2.0.0/libexec/lib/python3.6/site-packages/urwid/main_loop.py\", line 403, in _update\r\n self.process_input(keys)\r\n File \"/usr/local/Cellar/mitmproxy/2.0.0/libexec/lib/python3.6/site-packages/urwid/main_loop.py\", line 503, in process_input\r\n k = self._topmost_widget.keypress(self.screen_size, k)\r\n File \"/usr/local/Cellar/mitmproxy/2.0.0/libexec/lib/python3.6/site-packages/mitmproxy/tools/console/window.py\", line 84, in keypress\r\n k = super().keypress(size, k)\r\n File \"/usr/local/Cellar/mitmproxy/2.0.0/libexec/lib/python3.6/site-packages/urwid/container.py\", line 1128, in keypress\r\n return self.body.keypress( (maxcol, remaining), key )\r\n File \"/usr/local/Cellar/mitmproxy/2.0.0/libexec/lib/python3.6/site-packages/urwid/container.py\", line 1128, in keypress\r\n return self.body.keypress( (maxcol, remaining), key )\r\n File \"/usr/local/Cellar/mitmproxy/2.0.0/libexec/lib/python3.6/site-packages/mitmproxy/tools/console/select.py\", line 114, in keypress\r\n self.get_focus()[0].option.activate()\r\n File \"/usr/local/Cellar/mitmproxy/2.0.0/libexec/lib/python3.6/site-packages/mitmproxy/tools/console/palettepicker.py\", line 46, in <lambda>\r\n lambda: setattr(self.master.options, \"palette\", name)\r\n File \"/usr/local/Cellar/mitmproxy/2.0.0/libexec/lib/python3.6/site-packages/mitmproxy/optmanager.py\", line 114, in __setattr__\r\n self.update(**{attr: value})\r\n File \"/usr/local/Cellar/mitmproxy/2.0.0/libexec/lib/python3.6/site-packages/mitmproxy/optmanager.py\", line 141, in update\r\n raise KeyError(\"No such option: %s\" % k)\r\nKeyError: 'No such option: palette'\r\n\r\n```\r\nThe option names in mitmproxy/options.py were prefixed with 'console_', but line 46 and line 62 of mitmproxy/tools/console/palettepicker.py were not updated to include this prefix. \r\n\r\nThis appears to have been broken by commit [35aff3b](https://github.com/mitmproxy/mitmproxy/commit/35aff3b7838f8df718cc574d2643f1355849fa8e)\r\n\r\n##### System information\r\n\r\nMitmproxy version: 2.0.0 (release version) \r\nPython version: 3.6.0\r\nPlatform: Darwin-16.4.0-x86_64-i386-64bit\r\nSSL version: OpenSSL 1.1.0e 16 Feb 2017\r\nMac version: 10.12.3 ('', '', '') x86_64\r\n\n", "before_files": [{"content": "import urwid\n\nfrom mitmproxy.tools.console import common\nfrom mitmproxy.tools.console import palettes\nfrom mitmproxy.tools.console import select\n\nfooter = [\n ('heading_key', \"enter/space\"), \":select\",\n]\n\n\ndef _mkhelp():\n text = []\n keys = [\n (\"enter/space\", \"select\"),\n ]\n text.extend(common.format_keyvals(keys, key=\"key\", val=\"text\", indent=4))\n return text\n\n\nhelp_context = _mkhelp()\n\n\nclass PalettePicker(urwid.WidgetWrap):\n\n def __init__(self, master):\n self.master = master\n low, high = [], []\n for k, v in palettes.palettes.items():\n if v.high:\n high.append(k)\n else:\n low.append(k)\n high.sort()\n low.sort()\n\n options = [\n select.Heading(\"High Colour\")\n ]\n\n def mkopt(name):\n return select.Option(\n i,\n None,\n lambda: self.master.options.console_palette == name,\n lambda: setattr(self.master.options, \"palette\", name)\n )\n\n for i in high:\n options.append(mkopt(i))\n options.append(select.Heading(\"Low Colour\"))\n for i in low:\n options.append(mkopt(i))\n\n options.extend(\n [\n select.Heading(\"Options\"),\n select.Option(\n \"Transparent\",\n \"T\",\n lambda: master.options.console_palette_transparent,\n master.options.toggler(\"palette_transparent\")\n )\n ]\n )\n\n self.lb = select.Select(options)\n title = urwid.Text(\"Palettes\")\n title = urwid.Padding(title, align=\"left\", width=(\"relative\", 100))\n title = urwid.AttrWrap(title, \"heading\")\n self._w = urwid.Frame(\n self.lb,\n header = title\n )\n master.options.changed.connect(self.sig_options_changed)\n\n def sig_options_changed(self, options, updated):\n self.lb.walker._modified()\n", "path": "mitmproxy/tools/console/palettepicker.py"}], "after_files": [{"content": "import urwid\n\nfrom mitmproxy.tools.console import common\nfrom mitmproxy.tools.console import palettes\nfrom mitmproxy.tools.console import select\n\nfooter = [\n ('heading_key', \"enter/space\"), \":select\",\n]\n\n\ndef _mkhelp():\n text = []\n keys = [\n (\"enter/space\", \"select\"),\n ]\n text.extend(common.format_keyvals(keys, key=\"key\", val=\"text\", indent=4))\n return text\n\n\nhelp_context = _mkhelp()\n\n\nclass PalettePicker(urwid.WidgetWrap):\n\n def __init__(self, master):\n self.master = master\n low, high = [], []\n for k, v in palettes.palettes.items():\n if v.high:\n high.append(k)\n else:\n low.append(k)\n high.sort()\n low.sort()\n\n options = [\n select.Heading(\"High Colour\")\n ]\n\n def mkopt(name):\n return select.Option(\n i,\n None,\n lambda: self.master.options.console_palette == name,\n lambda: setattr(self.master.options, \"console_palette\", name)\n )\n\n for i in high:\n options.append(mkopt(i))\n options.append(select.Heading(\"Low Colour\"))\n for i in low:\n options.append(mkopt(i))\n\n options.extend(\n [\n select.Heading(\"Options\"),\n select.Option(\n \"Transparent\",\n \"T\",\n lambda: master.options.console_palette_transparent,\n master.options.toggler(\"console_palette_transparent\")\n )\n ]\n )\n\n self.lb = select.Select(options)\n title = urwid.Text(\"Palettes\")\n title = urwid.Padding(title, align=\"left\", width=(\"relative\", 100))\n title = urwid.AttrWrap(title, \"heading\")\n self._w = urwid.Frame(\n self.lb,\n header = title\n )\n master.options.changed.connect(self.sig_options_changed)\n\n def sig_options_changed(self, options, updated):\n self.lb.walker._modified()\n", "path": "mitmproxy/tools/console/palettepicker.py"}]} | 1,974 | 174 |
gh_patches_debug_10877 | rasdani/github-patches | git_diff | ivy-llc__ivy-20857 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
fftshift
This subtask is to add the `fftshift` function to the Jax frontend
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `ivy/functional/frontends/jax/numpy/fft.py`
Content:
```
1 # local
2 import ivy
3 from ivy.functional.frontends.jax.func_wrapper import to_ivy_arrays_and_back
4
5
6 @to_ivy_arrays_and_back
7 def fft(a, n=None, axis=-1, norm=None):
8 if norm is None:
9 norm = "backward"
10 return ivy.fft(a, axis, norm=norm, n=n)
11
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/ivy/functional/frontends/jax/numpy/fft.py b/ivy/functional/frontends/jax/numpy/fft.py
--- a/ivy/functional/frontends/jax/numpy/fft.py
+++ b/ivy/functional/frontends/jax/numpy/fft.py
@@ -1,6 +1,25 @@
# local
import ivy
from ivy.functional.frontends.jax.func_wrapper import to_ivy_arrays_and_back
+from ivy.func_wrapper import with_unsupported_dtypes
+
+
+@to_ivy_arrays_and_back
+@with_unsupported_dtypes({"2.4.2 and below": ("float16", "bfloat16")}, "paddle")
+def fftshift(x, axes=None, name=None):
+ shape = x.shape
+
+ if axes is None:
+ axes = tuple(range(x.ndim))
+ shifts = [(dim // 2) for dim in shape]
+ elif isinstance(axes, int):
+ shifts = shape[axes] // 2
+ else:
+ shifts = [shape[ax] // 2 for ax in axes]
+
+ roll = ivy.roll(x, shifts, axis=axes)
+
+ return roll
@to_ivy_arrays_and_back
| {"golden_diff": "diff --git a/ivy/functional/frontends/jax/numpy/fft.py b/ivy/functional/frontends/jax/numpy/fft.py\n--- a/ivy/functional/frontends/jax/numpy/fft.py\n+++ b/ivy/functional/frontends/jax/numpy/fft.py\n@@ -1,6 +1,25 @@\n # local\n import ivy\n from ivy.functional.frontends.jax.func_wrapper import to_ivy_arrays_and_back\n+from ivy.func_wrapper import with_unsupported_dtypes\n+\n+\n+@to_ivy_arrays_and_back\n+@with_unsupported_dtypes({\"2.4.2 and below\": (\"float16\", \"bfloat16\")}, \"paddle\")\n+def fftshift(x, axes=None, name=None):\n+ shape = x.shape\n+\n+ if axes is None:\n+ axes = tuple(range(x.ndim))\n+ shifts = [(dim // 2) for dim in shape]\n+ elif isinstance(axes, int):\n+ shifts = shape[axes] // 2\n+ else:\n+ shifts = [shape[ax] // 2 for ax in axes]\n+\n+ roll = ivy.roll(x, shifts, axis=axes)\n+\n+ return roll\n \n \n @to_ivy_arrays_and_back\n", "issue": "fftshift\nThis subtask is to add the `fftshift` function to the Jax frontend\n", "before_files": [{"content": "# local\nimport ivy\nfrom ivy.functional.frontends.jax.func_wrapper import to_ivy_arrays_and_back\n\n\n@to_ivy_arrays_and_back\ndef fft(a, n=None, axis=-1, norm=None):\n if norm is None:\n norm = \"backward\"\n return ivy.fft(a, axis, norm=norm, n=n)\n", "path": "ivy/functional/frontends/jax/numpy/fft.py"}], "after_files": [{"content": "# local\nimport ivy\nfrom ivy.functional.frontends.jax.func_wrapper import to_ivy_arrays_and_back\nfrom ivy.func_wrapper import with_unsupported_dtypes\n\n\n@to_ivy_arrays_and_back\n@with_unsupported_dtypes({\"2.4.2 and below\": (\"float16\", \"bfloat16\")}, \"paddle\")\ndef fftshift(x, axes=None, name=None):\n shape = x.shape\n\n if axes is None:\n axes = tuple(range(x.ndim))\n shifts = [(dim // 2) for dim in shape]\n elif isinstance(axes, int):\n shifts = shape[axes] // 2\n else:\n shifts = [shape[ax] // 2 for ax in axes]\n\n roll = ivy.roll(x, shifts, axis=axes)\n\n return roll\n\n\n@to_ivy_arrays_and_back\ndef fft(a, n=None, axis=-1, norm=None):\n if norm is None:\n norm = \"backward\"\n return ivy.fft(a, axis, norm=norm, n=n)\n", "path": "ivy/functional/frontends/jax/numpy/fft.py"}]} | 377 | 280 |
gh_patches_debug_963 | rasdani/github-patches | git_diff | pyjanitor-devs__pyjanitor-289 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Utilize autosummary Sphinx directive in API Reference
## Proposal
A consolidated list of functionality would go a long way in [our API Reference](https://pyjanitor.readthedocs.io/api.html) section.
Other libraries have leveraged the [autosummary](http://www.sphinx-doc.org/en/master/usage/extensions/autosummary.html#directive-autosummary) Sphinx directive to achieve this to great effect. For instance:
* Pandas: [Docs](https://pandas.pydata.org/pandas-docs/stable/reference/indexing.html), [Raw](https://raw.githubusercontent.com/pandas-dev/pandas/master/doc/source/reference/indexing.rst)
* Matplotlib: [Docs](https://matplotlib.org/api/axes_api.html), [Raw](https://matplotlib.org/_sources/api/axes_api.rst.txt)
## Implementation Details
Apart from rolling `sphinx.ext.autosummary` into the `conf.py` this would also involve going through and enumerating the different functions in the `api.rst` documentation.
A concern here, though-- this would mean that all future feature introductions would have to get appended to the lists in these files, **which necessitates adding this step to the PR checklist**... Until someone figures out a more programmatic way to do this, anyhow 😉
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `docs/conf.py`
Content:
```
1 # -*- coding: utf-8 -*-
2 #
3 # Configuration file for the Sphinx documentation builder.
4 #
5 # This file does only contain a selection of the most common options. For a
6 # full list see the documentation:
7 # http://www.sphinx-doc.org/en/stable/config
8
9 # -- Path setup --------------------------------------------------------------
10
11 # If extensions (or modules to document with autodoc) are in another directory,
12 # add these directories to sys.path here. If the directory is relative to the
13 # documentation root, use os.path.abspath to make it absolute, like shown here.
14 #
15 import os
16 import sys
17 from pathlib import Path
18
19 sys.path.insert(0, os.path.abspath("."))
20 sys.path.insert(0, os.path.abspath("../examples"))
21
22 # Make a symlink in our sphinx source directory to the top-level
23 # examples/notebooks directory so we can include notebooks in the doc
24 notebooks = Path("./notebooks")
25 if not notebooks.exists():
26 print("Making symlink to ../examples/notebooks")
27 notebooks.symlink_to("../examples/notebooks")
28
29
30 # -- Project information -----------------------------------------------------
31
32 project = "pyjanitor"
33 copyright = "2018, Eric J. Ma"
34 author = "Eric J. Ma"
35
36 # The short X.Y version
37 version = "0.1.0"
38 # The full version, including alpha/beta/rc tags
39 release = ""
40
41
42 # -- General configuration ---------------------------------------------------
43
44 # If your documentation needs a minimal Sphinx version, state it here.
45 #
46 # needs_sphinx = '1.0'
47
48 # Add any Sphinx extension module names here, as strings. They can be
49 # extensions coming with Sphinx (named 'sphinx.ext.*') or your custom
50 # ones.
51 extensions = [
52 "sphinx.ext.autodoc",
53 "sphinx.ext.doctest",
54 "sphinx.ext.intersphinx",
55 "sphinx.ext.todo",
56 "sphinx.ext.coverage",
57 "sphinx.ext.viewcode",
58 "sphinx.ext.githubpages",
59 "sphinxcontrib.fulltoc",
60 "nbsphinx",
61 ]
62
63 # Add any paths that contain templates here, relative to this directory.
64 templates_path = ["_templates"]
65
66 # The suffix(es) of source filenames.
67 # You can specify multiple suffix as a list of string:
68 #
69 # source_suffix = ['.rst', '.md']
70 source_suffix = [".md", ".rst", ".ipynb"]
71
72 # The master toctree document.
73 master_doc = "index"
74
75 # The language for content autogenerated by Sphinx. Refer to documentation
76 # for a list of supported languages.
77 #
78 # This is also used if you do content translation via gettext catalogs.
79 # Usually you set "language" from the command line for these cases.
80 language = None
81
82 # List of patterns, relative to source directory, that match files and
83 # directories to ignore when looking for source files.
84 # This pattern also affects html_static_path and html_extra_path .
85 exclude_patterns = ["_build", "Thumbs.db", ".DS_Store", "**.ipynb_checkpoints"]
86
87 # The name of the Pygments (syntax highlighting) style to use.
88 pygments_style = "sphinx"
89
90
91 # -- Options for HTML output -------------------------------------------------
92
93 # The theme to use for HTML and HTML Help pages. See the documentation for
94 # a list of builtin themes.
95 #
96 html_theme = "alabaster"
97
98 # Theme options are theme-specific and customize the look and feel of a theme
99 # further. For a list of options available for each theme, see the
100 # documentation.
101 #
102 html_theme_options = {"logo": "logo_title.svg"}
103
104 # Add any paths that contain custom static files (such as style sheets) here,
105 # relative to this directory. They are copied after the builtin static files,
106 # so a file named "default.css" will overwrite the builtin "default.css".
107 html_static_path = ["_static"]
108
109 # Custom sidebar templates, must be a dictionary that maps document names
110 # to template names.
111 #
112 # The default sidebars (for documents that don't match any pattern) are
113 # defined by theme itself. Builtin themes are using these templates by
114 # default: ``['localtoc.html', 'relations.html', 'sourcelink.html',
115 # 'searchbox.html']``.
116 #
117 html_sidebars = {
118 "**": ["about.html", "navigation.html", "relations.html", "searchbox.html"]
119 }
120
121
122 # -- Options for HTMLHelp output ---------------------------------------------
123
124 # Output file base name for HTML help builder.
125 htmlhelp_basename = "pyjanitordoc"
126
127
128 # -- Options for LaTeX output ------------------------------------------------
129
130 latex_elements = {
131 # The paper size ('letterpaper' or 'a4paper').
132 #
133 # 'papersize': 'letterpaper',
134 # The font size ('10pt', '11pt' or '12pt').
135 #
136 # 'pointsize': '10pt',
137 # Additional stuff for the LaTeX preamble.
138 #
139 # 'preamble': '',
140 # Latex figure (float) alignment
141 #
142 # 'figure_align': 'htbp',
143 }
144
145 # Grouping the document tree into LaTeX files. List of tuples
146 # (source start file, target name, title,
147 # author, documentclass [howto, manual, or own class]).
148 latex_documents = [
149 (
150 master_doc,
151 "pyjanitor.tex",
152 "pyjanitor Documentation",
153 "Eric J. Ma",
154 "manual",
155 )
156 ]
157
158
159 # -- Options for manual page output ------------------------------------------
160
161 # One entry per manual page. List of tuples
162 # (source start file, name, description, authors, manual section).
163 man_pages = [(master_doc, "pyjanitor", "pyjanitor Documentation", [author], 1)]
164
165
166 # -- Options for Texinfo output ----------------------------------------------
167
168 # Grouping the document tree into Texinfo files. List of tuples
169 # (source start file, target name, title, author,
170 # dir menu entry, description, category)
171 texinfo_documents = [
172 (
173 master_doc,
174 "pyjanitor",
175 "pyjanitor Documentation",
176 author,
177 "pyjanitor",
178 "One line description of project.",
179 "Miscellaneous",
180 )
181 ]
182
183
184 # -- Extension configuration -------------------------------------------------
185
186 # -- Options for intersphinx extension ---------------------------------------
187
188 # Example configuration for intersphinx: refer to the Python standard library.
189 intersphinx_mapping = {
190 "https://docs.python.org/": None,
191 "https://pandas.pydata.org/pandas-docs/stable": None,
192 }
193
194 # -- Options for todo extension ----------------------------------------------
195
196 # If true, `todo` and `todoList` produce output, else they produce nothing.
197 todo_include_todos = True
198
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/docs/conf.py b/docs/conf.py
--- a/docs/conf.py
+++ b/docs/conf.py
@@ -58,6 +58,7 @@
"sphinx.ext.githubpages",
"sphinxcontrib.fulltoc",
"nbsphinx",
+ "sphinx.ext.autosummary",
]
# Add any paths that contain templates here, relative to this directory.
| {"golden_diff": "diff --git a/docs/conf.py b/docs/conf.py\n--- a/docs/conf.py\n+++ b/docs/conf.py\n@@ -58,6 +58,7 @@\n \"sphinx.ext.githubpages\",\n \"sphinxcontrib.fulltoc\",\n \"nbsphinx\",\n+ \"sphinx.ext.autosummary\",\n ]\n \n # Add any paths that contain templates here, relative to this directory.\n", "issue": "Utilize autosummary Sphinx directive in API Reference\n## Proposal\r\n\r\nA consolidated list of functionality would go a long way in [our API Reference](https://pyjanitor.readthedocs.io/api.html) section.\r\n\r\nOther libraries have leveraged the [autosummary](http://www.sphinx-doc.org/en/master/usage/extensions/autosummary.html#directive-autosummary) Sphinx directive to achieve this to great effect. For instance:\r\n\r\n* Pandas: [Docs](https://pandas.pydata.org/pandas-docs/stable/reference/indexing.html), [Raw](https://raw.githubusercontent.com/pandas-dev/pandas/master/doc/source/reference/indexing.rst)\r\n* Matplotlib: [Docs](https://matplotlib.org/api/axes_api.html), [Raw](https://matplotlib.org/_sources/api/axes_api.rst.txt)\r\n\r\n## Implementation Details\r\n\r\nApart from rolling `sphinx.ext.autosummary` into the `conf.py` this would also involve going through and enumerating the different functions in the `api.rst` documentation.\r\n\r\nA concern here, though-- this would mean that all future feature introductions would have to get appended to the lists in these files, **which necessitates adding this step to the PR checklist**... Until someone figures out a more programmatic way to do this, anyhow \ud83d\ude09 \n", "before_files": [{"content": "# -*- coding: utf-8 -*-\n#\n# Configuration file for the Sphinx documentation builder.\n#\n# This file does only contain a selection of the most common options. For a\n# full list see the documentation:\n# http://www.sphinx-doc.org/en/stable/config\n\n# -- Path setup --------------------------------------------------------------\n\n# If extensions (or modules to document with autodoc) are in another directory,\n# add these directories to sys.path here. If the directory is relative to the\n# documentation root, use os.path.abspath to make it absolute, like shown here.\n#\nimport os\nimport sys\nfrom pathlib import Path\n\nsys.path.insert(0, os.path.abspath(\".\"))\nsys.path.insert(0, os.path.abspath(\"../examples\"))\n\n# Make a symlink in our sphinx source directory to the top-level\n# examples/notebooks directory so we can include notebooks in the doc\nnotebooks = Path(\"./notebooks\")\nif not notebooks.exists():\n print(\"Making symlink to ../examples/notebooks\")\n notebooks.symlink_to(\"../examples/notebooks\")\n\n\n# -- Project information -----------------------------------------------------\n\nproject = \"pyjanitor\"\ncopyright = \"2018, Eric J. Ma\"\nauthor = \"Eric J. Ma\"\n\n# The short X.Y version\nversion = \"0.1.0\"\n# The full version, including alpha/beta/rc tags\nrelease = \"\"\n\n\n# -- General configuration ---------------------------------------------------\n\n# If your documentation needs a minimal Sphinx version, state it here.\n#\n# needs_sphinx = '1.0'\n\n# Add any Sphinx extension module names here, as strings. They can be\n# extensions coming with Sphinx (named 'sphinx.ext.*') or your custom\n# ones.\nextensions = [\n \"sphinx.ext.autodoc\",\n \"sphinx.ext.doctest\",\n \"sphinx.ext.intersphinx\",\n \"sphinx.ext.todo\",\n \"sphinx.ext.coverage\",\n \"sphinx.ext.viewcode\",\n \"sphinx.ext.githubpages\",\n \"sphinxcontrib.fulltoc\",\n \"nbsphinx\",\n]\n\n# Add any paths that contain templates here, relative to this directory.\ntemplates_path = [\"_templates\"]\n\n# The suffix(es) of source filenames.\n# You can specify multiple suffix as a list of string:\n#\n# source_suffix = ['.rst', '.md']\nsource_suffix = [\".md\", \".rst\", \".ipynb\"]\n\n# The master toctree document.\nmaster_doc = \"index\"\n\n# The language for content autogenerated by Sphinx. Refer to documentation\n# for a list of supported languages.\n#\n# This is also used if you do content translation via gettext catalogs.\n# Usually you set \"language\" from the command line for these cases.\nlanguage = None\n\n# List of patterns, relative to source directory, that match files and\n# directories to ignore when looking for source files.\n# This pattern also affects html_static_path and html_extra_path .\nexclude_patterns = [\"_build\", \"Thumbs.db\", \".DS_Store\", \"**.ipynb_checkpoints\"]\n\n# The name of the Pygments (syntax highlighting) style to use.\npygments_style = \"sphinx\"\n\n\n# -- Options for HTML output -------------------------------------------------\n\n# The theme to use for HTML and HTML Help pages. See the documentation for\n# a list of builtin themes.\n#\nhtml_theme = \"alabaster\"\n\n# Theme options are theme-specific and customize the look and feel of a theme\n# further. For a list of options available for each theme, see the\n# documentation.\n#\nhtml_theme_options = {\"logo\": \"logo_title.svg\"}\n\n# Add any paths that contain custom static files (such as style sheets) here,\n# relative to this directory. They are copied after the builtin static files,\n# so a file named \"default.css\" will overwrite the builtin \"default.css\".\nhtml_static_path = [\"_static\"]\n\n# Custom sidebar templates, must be a dictionary that maps document names\n# to template names.\n#\n# The default sidebars (for documents that don't match any pattern) are\n# defined by theme itself. Builtin themes are using these templates by\n# default: ``['localtoc.html', 'relations.html', 'sourcelink.html',\n# 'searchbox.html']``.\n#\nhtml_sidebars = {\n \"**\": [\"about.html\", \"navigation.html\", \"relations.html\", \"searchbox.html\"]\n}\n\n\n# -- Options for HTMLHelp output ---------------------------------------------\n\n# Output file base name for HTML help builder.\nhtmlhelp_basename = \"pyjanitordoc\"\n\n\n# -- Options for LaTeX output ------------------------------------------------\n\nlatex_elements = {\n # The paper size ('letterpaper' or 'a4paper').\n #\n # 'papersize': 'letterpaper',\n # The font size ('10pt', '11pt' or '12pt').\n #\n # 'pointsize': '10pt',\n # Additional stuff for the LaTeX preamble.\n #\n # 'preamble': '',\n # Latex figure (float) alignment\n #\n # 'figure_align': 'htbp',\n}\n\n# Grouping the document tree into LaTeX files. List of tuples\n# (source start file, target name, title,\n# author, documentclass [howto, manual, or own class]).\nlatex_documents = [\n (\n master_doc,\n \"pyjanitor.tex\",\n \"pyjanitor Documentation\",\n \"Eric J. Ma\",\n \"manual\",\n )\n]\n\n\n# -- Options for manual page output ------------------------------------------\n\n# One entry per manual page. List of tuples\n# (source start file, name, description, authors, manual section).\nman_pages = [(master_doc, \"pyjanitor\", \"pyjanitor Documentation\", [author], 1)]\n\n\n# -- Options for Texinfo output ----------------------------------------------\n\n# Grouping the document tree into Texinfo files. List of tuples\n# (source start file, target name, title, author,\n# dir menu entry, description, category)\ntexinfo_documents = [\n (\n master_doc,\n \"pyjanitor\",\n \"pyjanitor Documentation\",\n author,\n \"pyjanitor\",\n \"One line description of project.\",\n \"Miscellaneous\",\n )\n]\n\n\n# -- Extension configuration -------------------------------------------------\n\n# -- Options for intersphinx extension ---------------------------------------\n\n# Example configuration for intersphinx: refer to the Python standard library.\nintersphinx_mapping = {\n \"https://docs.python.org/\": None,\n \"https://pandas.pydata.org/pandas-docs/stable\": None,\n}\n\n# -- Options for todo extension ----------------------------------------------\n\n# If true, `todo` and `todoList` produce output, else they produce nothing.\ntodo_include_todos = True\n", "path": "docs/conf.py"}], "after_files": [{"content": "# -*- coding: utf-8 -*-\n#\n# Configuration file for the Sphinx documentation builder.\n#\n# This file does only contain a selection of the most common options. For a\n# full list see the documentation:\n# http://www.sphinx-doc.org/en/stable/config\n\n# -- Path setup --------------------------------------------------------------\n\n# If extensions (or modules to document with autodoc) are in another directory,\n# add these directories to sys.path here. If the directory is relative to the\n# documentation root, use os.path.abspath to make it absolute, like shown here.\n#\nimport os\nimport sys\nfrom pathlib import Path\n\nsys.path.insert(0, os.path.abspath(\".\"))\nsys.path.insert(0, os.path.abspath(\"../examples\"))\n\n# Make a symlink in our sphinx source directory to the top-level\n# examples/notebooks directory so we can include notebooks in the doc\nnotebooks = Path(\"./notebooks\")\nif not notebooks.exists():\n print(\"Making symlink to ../examples/notebooks\")\n notebooks.symlink_to(\"../examples/notebooks\")\n\n\n# -- Project information -----------------------------------------------------\n\nproject = \"pyjanitor\"\ncopyright = \"2018, Eric J. Ma\"\nauthor = \"Eric J. Ma\"\n\n# The short X.Y version\nversion = \"0.1.0\"\n# The full version, including alpha/beta/rc tags\nrelease = \"\"\n\n\n# -- General configuration ---------------------------------------------------\n\n# If your documentation needs a minimal Sphinx version, state it here.\n#\n# needs_sphinx = '1.0'\n\n# Add any Sphinx extension module names here, as strings. They can be\n# extensions coming with Sphinx (named 'sphinx.ext.*') or your custom\n# ones.\nextensions = [\n \"sphinx.ext.autodoc\",\n \"sphinx.ext.doctest\",\n \"sphinx.ext.intersphinx\",\n \"sphinx.ext.todo\",\n \"sphinx.ext.coverage\",\n \"sphinx.ext.viewcode\",\n \"sphinx.ext.githubpages\",\n \"sphinxcontrib.fulltoc\",\n \"nbsphinx\",\n \"sphinx.ext.autosummary\",\n]\n\n# Add any paths that contain templates here, relative to this directory.\ntemplates_path = [\"_templates\"]\n\n# The suffix(es) of source filenames.\n# You can specify multiple suffix as a list of string:\n#\n# source_suffix = ['.rst', '.md']\nsource_suffix = [\".md\", \".rst\", \".ipynb\"]\n\n# The master toctree document.\nmaster_doc = \"index\"\n\n# The language for content autogenerated by Sphinx. Refer to documentation\n# for a list of supported languages.\n#\n# This is also used if you do content translation via gettext catalogs.\n# Usually you set \"language\" from the command line for these cases.\nlanguage = None\n\n# List of patterns, relative to source directory, that match files and\n# directories to ignore when looking for source files.\n# This pattern also affects html_static_path and html_extra_path .\nexclude_patterns = [\"_build\", \"Thumbs.db\", \".DS_Store\", \"**.ipynb_checkpoints\"]\n\n# The name of the Pygments (syntax highlighting) style to use.\npygments_style = \"sphinx\"\n\n\n# -- Options for HTML output -------------------------------------------------\n\n# The theme to use for HTML and HTML Help pages. See the documentation for\n# a list of builtin themes.\n#\nhtml_theme = \"alabaster\"\n\n# Theme options are theme-specific and customize the look and feel of a theme\n# further. For a list of options available for each theme, see the\n# documentation.\n#\nhtml_theme_options = {\"logo\": \"logo_title.svg\"}\n\n# Add any paths that contain custom static files (such as style sheets) here,\n# relative to this directory. They are copied after the builtin static files,\n# so a file named \"default.css\" will overwrite the builtin \"default.css\".\nhtml_static_path = [\"_static\"]\n\n# Custom sidebar templates, must be a dictionary that maps document names\n# to template names.\n#\n# The default sidebars (for documents that don't match any pattern) are\n# defined by theme itself. Builtin themes are using these templates by\n# default: ``['localtoc.html', 'relations.html', 'sourcelink.html',\n# 'searchbox.html']``.\n#\nhtml_sidebars = {\n \"**\": [\"about.html\", \"navigation.html\", \"relations.html\", \"searchbox.html\"]\n}\n\n\n# -- Options for HTMLHelp output ---------------------------------------------\n\n# Output file base name for HTML help builder.\nhtmlhelp_basename = \"pyjanitordoc\"\n\n\n# -- Options for LaTeX output ------------------------------------------------\n\nlatex_elements = {\n # The paper size ('letterpaper' or 'a4paper').\n #\n # 'papersize': 'letterpaper',\n # The font size ('10pt', '11pt' or '12pt').\n #\n # 'pointsize': '10pt',\n # Additional stuff for the LaTeX preamble.\n #\n # 'preamble': '',\n # Latex figure (float) alignment\n #\n # 'figure_align': 'htbp',\n}\n\n# Grouping the document tree into LaTeX files. List of tuples\n# (source start file, target name, title,\n# author, documentclass [howto, manual, or own class]).\nlatex_documents = [\n (\n master_doc,\n \"pyjanitor.tex\",\n \"pyjanitor Documentation\",\n \"Eric J. Ma\",\n \"manual\",\n )\n]\n\n\n# -- Options for manual page output ------------------------------------------\n\n# One entry per manual page. List of tuples\n# (source start file, name, description, authors, manual section).\nman_pages = [(master_doc, \"pyjanitor\", \"pyjanitor Documentation\", [author], 1)]\n\n\n# -- Options for Texinfo output ----------------------------------------------\n\n# Grouping the document tree into Texinfo files. List of tuples\n# (source start file, target name, title, author,\n# dir menu entry, description, category)\ntexinfo_documents = [\n (\n master_doc,\n \"pyjanitor\",\n \"pyjanitor Documentation\",\n author,\n \"pyjanitor\",\n \"One line description of project.\",\n \"Miscellaneous\",\n )\n]\n\n\n# -- Extension configuration -------------------------------------------------\n\n# -- Options for intersphinx extension ---------------------------------------\n\n# Example configuration for intersphinx: refer to the Python standard library.\nintersphinx_mapping = {\n \"https://docs.python.org/\": None,\n \"https://pandas.pydata.org/pandas-docs/stable\": None,\n}\n\n# -- Options for todo extension ----------------------------------------------\n\n# If true, `todo` and `todoList` produce output, else they produce nothing.\ntodo_include_todos = True\n", "path": "docs/conf.py"}]} | 2,433 | 85 |
gh_patches_debug_21747 | rasdani/github-patches | git_diff | huggingface__dataset-viewer-2770 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
FineWeb: Unexpected end of stream: Page was smaller (1862094) than expected (2055611)
The config-parquet-metadata job succeeds but the split-first-rows job fails when using `compute_first_rows_from_parquet_response`.
In the meantime I set the error code in the config-parquet-metadata response as `CachedResponseNotFound` to make the split-first-rows succeed
This workaround causes `ResponseNotFound` when opening page 2 in the viewer unfortunately (can't do random access in the parquet data without a valid config-parquet-metadata response)
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `services/worker/src/worker/job_runners/config/parquet_metadata.py`
Content:
```
1 # SPDX-License-Identifier: Apache-2.0
2 # Copyright 2022 The HuggingFace Authors.
3
4 import functools
5 import logging
6 from typing import Optional
7
8 from fsspec.implementations.http import HTTPFileSystem
9 from libcommon.dtos import JobInfo, SplitHubFile
10 from libcommon.exceptions import (
11 FileSystemError,
12 ParquetResponseEmptyError,
13 PreviousStepFormatError,
14 )
15 from libcommon.simple_cache import get_previous_step_or_raise
16 from libcommon.storage import StrPath
17 from libcommon.viewer_utils.parquet_metadata import create_parquet_metadata_file
18 from tqdm.contrib.concurrent import thread_map
19
20 from worker.config import AppConfig
21 from worker.dtos import (
22 CompleteJobResult,
23 ConfigParquetMetadataResponse,
24 ParquetFileMetadataItem,
25 )
26 from worker.job_runners.config.config_job_runner import ConfigJobRunner
27 from worker.utils import get_parquet_file
28
29
30 def create_parquet_metadata_file_from_remote_parquet(
31 parquet_file_item: SplitHubFile, fs: HTTPFileSystem, hf_token: Optional[str], parquet_metadata_directory: StrPath
32 ) -> ParquetFileMetadataItem:
33 try:
34 parquet_file = get_parquet_file(url=parquet_file_item["url"], fs=fs, hf_token=hf_token)
35 except Exception as e:
36 raise FileSystemError(f"Could not read the parquet files: {e}") from e
37 parquet_metadata_subpath = create_parquet_metadata_file(
38 dataset=parquet_file_item["dataset"],
39 config=parquet_file_item["config"],
40 split=parquet_file_item["split"],
41 parquet_file_metadata=parquet_file.metadata,
42 filename=parquet_file_item["filename"],
43 parquet_metadata_directory=parquet_metadata_directory,
44 )
45 return ParquetFileMetadataItem(
46 dataset=parquet_file_item["dataset"],
47 config=parquet_file_item["config"],
48 split=parquet_file_item["split"],
49 url=parquet_file_item["url"],
50 filename=parquet_file_item["filename"],
51 size=parquet_file_item["size"],
52 num_rows=parquet_file.metadata.num_rows,
53 parquet_metadata_subpath=parquet_metadata_subpath,
54 )
55
56
57 def compute_parquet_metadata_response(
58 dataset: str, config: str, hf_token: Optional[str], parquet_metadata_directory: StrPath
59 ) -> ConfigParquetMetadataResponse:
60 """
61 Get the response of 'config-parquet-metadata' for one specific dataset and config on huggingface.co.
62 Store the config's parquet metadata on the disk and return the list of local metadata files.
63
64 Args:
65 dataset (`str`):
66 A namespace (user or an organization) and a repo name separated
67 by a `/`.
68 config (`str`):
69 A configuration name.
70 hf_token (`str`, *optional*):
71 An authentication token (See https://huggingface.co/settings/token)
72 parquet_metadata_directory (`str` or `pathlib.Path`):
73 The directory where the parquet metadata files are stored.
74
75 Raises:
76 [~`libcommon.simple_cache.CachedArtifactError`]:
77 If the previous step gave an error.
78 [~`libcommon.exceptions.PreviousStepFormatError`]:
79 If the content of the previous step has not the expected format
80 [~`libcommon.exceptions.ParquetResponseEmptyError`]:
81 If the previous step provided an empty list of parquet files.
82 [~`libcommon.exceptions.FileSystemError`]:
83 If the HfFileSystem couldn't access the parquet files.
84
85 Returns:
86 `ConfigParquetMetadataResponse`: An object with the list of parquet metadata files.
87 """
88 logging.info(f"compute 'config-parquet-metadata' for {dataset=} {config=}")
89
90 config_parquet_response = get_previous_step_or_raise(kind="config-parquet", dataset=dataset, config=config)
91 try:
92 parquet_files_content = config_parquet_response["content"]["parquet_files"]
93 parquet_file_items: list[SplitHubFile] = [
94 parquet_file_item for parquet_file_item in parquet_files_content if parquet_file_item["config"] == config
95 ]
96 if not parquet_file_items:
97 raise ParquetResponseEmptyError("No parquet files found.")
98 content = config_parquet_response["content"]
99 if "features" in content and isinstance(content["features"], dict):
100 features = content["features"] # config-parquet version<6 didn't have features
101 else:
102 # (July 23) we can remove this later and raise an error instead (can be None for backward compatibility)
103 features = None
104 partial = config_parquet_response["content"]["partial"]
105 except Exception as e:
106 raise PreviousStepFormatError("Previous step did not return the expected content.") from e
107
108 fs = HTTPFileSystem()
109 desc = f"{dataset}/{config}"
110 parquet_files_metadata: list[ParquetFileMetadataItem] = thread_map(
111 functools.partial(
112 create_parquet_metadata_file_from_remote_parquet,
113 fs=fs,
114 hf_token=hf_token,
115 parquet_metadata_directory=parquet_metadata_directory,
116 ),
117 parquet_file_items,
118 desc=desc,
119 unit="pq",
120 disable=True,
121 )
122 return ConfigParquetMetadataResponse(
123 parquet_files_metadata=parquet_files_metadata, features=features, partial=partial
124 )
125
126
127 class ConfigParquetMetadataJobRunner(ConfigJobRunner):
128 parquet_metadata_directory: StrPath
129
130 @staticmethod
131 def get_job_type() -> str:
132 return "config-parquet-metadata"
133
134 def __init__(
135 self,
136 job_info: JobInfo,
137 app_config: AppConfig,
138 parquet_metadata_directory: StrPath,
139 ) -> None:
140 super().__init__(
141 job_info=job_info,
142 app_config=app_config,
143 )
144 self.parquet_metadata_directory = parquet_metadata_directory
145
146 def compute(self) -> CompleteJobResult:
147 return CompleteJobResult(
148 compute_parquet_metadata_response(
149 dataset=self.dataset,
150 config=self.config,
151 hf_token=self.app_config.common.hf_token,
152 parquet_metadata_directory=self.parquet_metadata_directory,
153 )
154 )
155
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/services/worker/src/worker/job_runners/config/parquet_metadata.py b/services/worker/src/worker/job_runners/config/parquet_metadata.py
--- a/services/worker/src/worker/job_runners/config/parquet_metadata.py
+++ b/services/worker/src/worker/job_runners/config/parquet_metadata.py
@@ -34,10 +34,14 @@
parquet_file = get_parquet_file(url=parquet_file_item["url"], fs=fs, hf_token=hf_token)
except Exception as e:
raise FileSystemError(f"Could not read the parquet files: {e}") from e
+ split = parquet_file_item["url"].split("/")[-2]
+ # ^ https://github.com/huggingface/dataset-viewer/issues/2768
+ # to support more than 10k parquet files, in which case, instead of "train" for example,
+ # the subdirectories are "train-part0", "train-part1", "train-part2", etc.
parquet_metadata_subpath = create_parquet_metadata_file(
dataset=parquet_file_item["dataset"],
config=parquet_file_item["config"],
- split=parquet_file_item["split"],
+ split=split,
parquet_file_metadata=parquet_file.metadata,
filename=parquet_file_item["filename"],
parquet_metadata_directory=parquet_metadata_directory,
| {"golden_diff": "diff --git a/services/worker/src/worker/job_runners/config/parquet_metadata.py b/services/worker/src/worker/job_runners/config/parquet_metadata.py\n--- a/services/worker/src/worker/job_runners/config/parquet_metadata.py\n+++ b/services/worker/src/worker/job_runners/config/parquet_metadata.py\n@@ -34,10 +34,14 @@\n parquet_file = get_parquet_file(url=parquet_file_item[\"url\"], fs=fs, hf_token=hf_token)\n except Exception as e:\n raise FileSystemError(f\"Could not read the parquet files: {e}\") from e\n+ split = parquet_file_item[\"url\"].split(\"/\")[-2]\n+ # ^ https://github.com/huggingface/dataset-viewer/issues/2768\n+ # to support more than 10k parquet files, in which case, instead of \"train\" for example,\n+ # the subdirectories are \"train-part0\", \"train-part1\", \"train-part2\", etc.\n parquet_metadata_subpath = create_parquet_metadata_file(\n dataset=parquet_file_item[\"dataset\"],\n config=parquet_file_item[\"config\"],\n- split=parquet_file_item[\"split\"],\n+ split=split,\n parquet_file_metadata=parquet_file.metadata,\n filename=parquet_file_item[\"filename\"],\n parquet_metadata_directory=parquet_metadata_directory,\n", "issue": "FineWeb: Unexpected end of stream: Page was smaller (1862094) than expected (2055611)\nThe config-parquet-metadata job succeeds but the split-first-rows job fails when using `compute_first_rows_from_parquet_response`.\r\n\r\nIn the meantime I set the error code in the config-parquet-metadata response as `CachedResponseNotFound` to make the split-first-rows succeed\r\n\r\nThis workaround causes `ResponseNotFound` when opening page 2 in the viewer unfortunately (can't do random access in the parquet data without a valid config-parquet-metadata response)\n", "before_files": [{"content": "# SPDX-License-Identifier: Apache-2.0\n# Copyright 2022 The HuggingFace Authors.\n\nimport functools\nimport logging\nfrom typing import Optional\n\nfrom fsspec.implementations.http import HTTPFileSystem\nfrom libcommon.dtos import JobInfo, SplitHubFile\nfrom libcommon.exceptions import (\n FileSystemError,\n ParquetResponseEmptyError,\n PreviousStepFormatError,\n)\nfrom libcommon.simple_cache import get_previous_step_or_raise\nfrom libcommon.storage import StrPath\nfrom libcommon.viewer_utils.parquet_metadata import create_parquet_metadata_file\nfrom tqdm.contrib.concurrent import thread_map\n\nfrom worker.config import AppConfig\nfrom worker.dtos import (\n CompleteJobResult,\n ConfigParquetMetadataResponse,\n ParquetFileMetadataItem,\n)\nfrom worker.job_runners.config.config_job_runner import ConfigJobRunner\nfrom worker.utils import get_parquet_file\n\n\ndef create_parquet_metadata_file_from_remote_parquet(\n parquet_file_item: SplitHubFile, fs: HTTPFileSystem, hf_token: Optional[str], parquet_metadata_directory: StrPath\n) -> ParquetFileMetadataItem:\n try:\n parquet_file = get_parquet_file(url=parquet_file_item[\"url\"], fs=fs, hf_token=hf_token)\n except Exception as e:\n raise FileSystemError(f\"Could not read the parquet files: {e}\") from e\n parquet_metadata_subpath = create_parquet_metadata_file(\n dataset=parquet_file_item[\"dataset\"],\n config=parquet_file_item[\"config\"],\n split=parquet_file_item[\"split\"],\n parquet_file_metadata=parquet_file.metadata,\n filename=parquet_file_item[\"filename\"],\n parquet_metadata_directory=parquet_metadata_directory,\n )\n return ParquetFileMetadataItem(\n dataset=parquet_file_item[\"dataset\"],\n config=parquet_file_item[\"config\"],\n split=parquet_file_item[\"split\"],\n url=parquet_file_item[\"url\"],\n filename=parquet_file_item[\"filename\"],\n size=parquet_file_item[\"size\"],\n num_rows=parquet_file.metadata.num_rows,\n parquet_metadata_subpath=parquet_metadata_subpath,\n )\n\n\ndef compute_parquet_metadata_response(\n dataset: str, config: str, hf_token: Optional[str], parquet_metadata_directory: StrPath\n) -> ConfigParquetMetadataResponse:\n \"\"\"\n Get the response of 'config-parquet-metadata' for one specific dataset and config on huggingface.co.\n Store the config's parquet metadata on the disk and return the list of local metadata files.\n\n Args:\n dataset (`str`):\n A namespace (user or an organization) and a repo name separated\n by a `/`.\n config (`str`):\n A configuration name.\n hf_token (`str`, *optional*):\n An authentication token (See https://huggingface.co/settings/token)\n parquet_metadata_directory (`str` or `pathlib.Path`):\n The directory where the parquet metadata files are stored.\n\n Raises:\n [~`libcommon.simple_cache.CachedArtifactError`]:\n If the previous step gave an error.\n [~`libcommon.exceptions.PreviousStepFormatError`]:\n If the content of the previous step has not the expected format\n [~`libcommon.exceptions.ParquetResponseEmptyError`]:\n If the previous step provided an empty list of parquet files.\n [~`libcommon.exceptions.FileSystemError`]:\n If the HfFileSystem couldn't access the parquet files.\n\n Returns:\n `ConfigParquetMetadataResponse`: An object with the list of parquet metadata files.\n \"\"\"\n logging.info(f\"compute 'config-parquet-metadata' for {dataset=} {config=}\")\n\n config_parquet_response = get_previous_step_or_raise(kind=\"config-parquet\", dataset=dataset, config=config)\n try:\n parquet_files_content = config_parquet_response[\"content\"][\"parquet_files\"]\n parquet_file_items: list[SplitHubFile] = [\n parquet_file_item for parquet_file_item in parquet_files_content if parquet_file_item[\"config\"] == config\n ]\n if not parquet_file_items:\n raise ParquetResponseEmptyError(\"No parquet files found.\")\n content = config_parquet_response[\"content\"]\n if \"features\" in content and isinstance(content[\"features\"], dict):\n features = content[\"features\"] # config-parquet version<6 didn't have features\n else:\n # (July 23) we can remove this later and raise an error instead (can be None for backward compatibility)\n features = None\n partial = config_parquet_response[\"content\"][\"partial\"]\n except Exception as e:\n raise PreviousStepFormatError(\"Previous step did not return the expected content.\") from e\n\n fs = HTTPFileSystem()\n desc = f\"{dataset}/{config}\"\n parquet_files_metadata: list[ParquetFileMetadataItem] = thread_map(\n functools.partial(\n create_parquet_metadata_file_from_remote_parquet,\n fs=fs,\n hf_token=hf_token,\n parquet_metadata_directory=parquet_metadata_directory,\n ),\n parquet_file_items,\n desc=desc,\n unit=\"pq\",\n disable=True,\n )\n return ConfigParquetMetadataResponse(\n parquet_files_metadata=parquet_files_metadata, features=features, partial=partial\n )\n\n\nclass ConfigParquetMetadataJobRunner(ConfigJobRunner):\n parquet_metadata_directory: StrPath\n\n @staticmethod\n def get_job_type() -> str:\n return \"config-parquet-metadata\"\n\n def __init__(\n self,\n job_info: JobInfo,\n app_config: AppConfig,\n parquet_metadata_directory: StrPath,\n ) -> None:\n super().__init__(\n job_info=job_info,\n app_config=app_config,\n )\n self.parquet_metadata_directory = parquet_metadata_directory\n\n def compute(self) -> CompleteJobResult:\n return CompleteJobResult(\n compute_parquet_metadata_response(\n dataset=self.dataset,\n config=self.config,\n hf_token=self.app_config.common.hf_token,\n parquet_metadata_directory=self.parquet_metadata_directory,\n )\n )\n", "path": "services/worker/src/worker/job_runners/config/parquet_metadata.py"}], "after_files": [{"content": "# SPDX-License-Identifier: Apache-2.0\n# Copyright 2022 The HuggingFace Authors.\n\nimport functools\nimport logging\nfrom typing import Optional\n\nfrom fsspec.implementations.http import HTTPFileSystem\nfrom libcommon.dtos import JobInfo, SplitHubFile\nfrom libcommon.exceptions import (\n FileSystemError,\n ParquetResponseEmptyError,\n PreviousStepFormatError,\n)\nfrom libcommon.simple_cache import get_previous_step_or_raise\nfrom libcommon.storage import StrPath\nfrom libcommon.viewer_utils.parquet_metadata import create_parquet_metadata_file\nfrom tqdm.contrib.concurrent import thread_map\n\nfrom worker.config import AppConfig\nfrom worker.dtos import (\n CompleteJobResult,\n ConfigParquetMetadataResponse,\n ParquetFileMetadataItem,\n)\nfrom worker.job_runners.config.config_job_runner import ConfigJobRunner\nfrom worker.utils import get_parquet_file\n\n\ndef create_parquet_metadata_file_from_remote_parquet(\n parquet_file_item: SplitHubFile, fs: HTTPFileSystem, hf_token: Optional[str], parquet_metadata_directory: StrPath\n) -> ParquetFileMetadataItem:\n try:\n parquet_file = get_parquet_file(url=parquet_file_item[\"url\"], fs=fs, hf_token=hf_token)\n except Exception as e:\n raise FileSystemError(f\"Could not read the parquet files: {e}\") from e\n split = parquet_file_item[\"url\"].split(\"/\")[-2]\n # ^ https://github.com/huggingface/dataset-viewer/issues/2768\n # to support more than 10k parquet files, in which case, instead of \"train\" for example,\n # the subdirectories are \"train-part0\", \"train-part1\", \"train-part2\", etc.\n parquet_metadata_subpath = create_parquet_metadata_file(\n dataset=parquet_file_item[\"dataset\"],\n config=parquet_file_item[\"config\"],\n split=split,\n parquet_file_metadata=parquet_file.metadata,\n filename=parquet_file_item[\"filename\"],\n parquet_metadata_directory=parquet_metadata_directory,\n )\n return ParquetFileMetadataItem(\n dataset=parquet_file_item[\"dataset\"],\n config=parquet_file_item[\"config\"],\n split=parquet_file_item[\"split\"],\n url=parquet_file_item[\"url\"],\n filename=parquet_file_item[\"filename\"],\n size=parquet_file_item[\"size\"],\n num_rows=parquet_file.metadata.num_rows,\n parquet_metadata_subpath=parquet_metadata_subpath,\n )\n\n\ndef compute_parquet_metadata_response(\n dataset: str, config: str, hf_token: Optional[str], parquet_metadata_directory: StrPath\n) -> ConfigParquetMetadataResponse:\n \"\"\"\n Get the response of 'config-parquet-metadata' for one specific dataset and config on huggingface.co.\n Store the config's parquet metadata on the disk and return the list of local metadata files.\n\n Args:\n dataset (`str`):\n A namespace (user or an organization) and a repo name separated\n by a `/`.\n config (`str`):\n A configuration name.\n hf_token (`str`, *optional*):\n An authentication token (See https://huggingface.co/settings/token)\n parquet_metadata_directory (`str` or `pathlib.Path`):\n The directory where the parquet metadata files are stored.\n\n Raises:\n [~`libcommon.simple_cache.CachedArtifactError`]:\n If the previous step gave an error.\n [~`libcommon.exceptions.PreviousStepFormatError`]:\n If the content of the previous step has not the expected format\n [~`libcommon.exceptions.ParquetResponseEmptyError`]:\n If the previous step provided an empty list of parquet files.\n [~`libcommon.exceptions.FileSystemError`]:\n If the HfFileSystem couldn't access the parquet files.\n\n Returns:\n `ConfigParquetMetadataResponse`: An object with the list of parquet metadata files.\n \"\"\"\n logging.info(f\"compute 'config-parquet-metadata' for {dataset=} {config=}\")\n\n config_parquet_response = get_previous_step_or_raise(kind=\"config-parquet\", dataset=dataset, config=config)\n try:\n parquet_files_content = config_parquet_response[\"content\"][\"parquet_files\"]\n parquet_file_items: list[SplitHubFile] = [\n parquet_file_item for parquet_file_item in parquet_files_content if parquet_file_item[\"config\"] == config\n ]\n if not parquet_file_items:\n raise ParquetResponseEmptyError(\"No parquet files found.\")\n content = config_parquet_response[\"content\"]\n if \"features\" in content and isinstance(content[\"features\"], dict):\n features = content[\"features\"] # config-parquet version<6 didn't have features\n else:\n # (July 23) we can remove this later and raise an error instead (can be None for backward compatibility)\n features = None\n partial = config_parquet_response[\"content\"][\"partial\"]\n except Exception as e:\n raise PreviousStepFormatError(\"Previous step did not return the expected content.\") from e\n\n fs = HTTPFileSystem()\n desc = f\"{dataset}/{config}\"\n parquet_files_metadata: list[ParquetFileMetadataItem] = thread_map(\n functools.partial(\n create_parquet_metadata_file_from_remote_parquet,\n fs=fs,\n hf_token=hf_token,\n parquet_metadata_directory=parquet_metadata_directory,\n ),\n parquet_file_items,\n desc=desc,\n unit=\"pq\",\n disable=True,\n )\n return ConfigParquetMetadataResponse(\n parquet_files_metadata=parquet_files_metadata, features=features, partial=partial\n )\n\n\nclass ConfigParquetMetadataJobRunner(ConfigJobRunner):\n parquet_metadata_directory: StrPath\n\n @staticmethod\n def get_job_type() -> str:\n return \"config-parquet-metadata\"\n\n def __init__(\n self,\n job_info: JobInfo,\n app_config: AppConfig,\n parquet_metadata_directory: StrPath,\n ) -> None:\n super().__init__(\n job_info=job_info,\n app_config=app_config,\n )\n self.parquet_metadata_directory = parquet_metadata_directory\n\n def compute(self) -> CompleteJobResult:\n return CompleteJobResult(\n compute_parquet_metadata_response(\n dataset=self.dataset,\n config=self.config,\n hf_token=self.app_config.common.hf_token,\n parquet_metadata_directory=self.parquet_metadata_directory,\n )\n )\n", "path": "services/worker/src/worker/job_runners/config/parquet_metadata.py"}]} | 2,067 | 311 |
gh_patches_debug_20482 | rasdani/github-patches | git_diff | crytic__slither-546 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
function-id not zero-padding function signature
```
ERC20:
+---------------------------------------+------------+
| Name | ID |
+---------------------------------------+------------+
| totalSupply() | 0x18160ddd |
| balanceOf(address) | 0x70a08231 |
| allowance(address,address) | 0xdd62ed3e |
| transfer(address,uint256) | 0xa9059cbb |
| transferFrom(address,address,uint256) | 0x23b872dd |
| approve(address,uint256) | 0x95ea7b3 |
+---------------------------------------+------------+
```
It's a minor annoyance, but for approve it outputs `0x95ea7b3` instead of `0x095ea7b3`. It is the same numerically, of course, but the function signature is more of an opaque 4-byte identifier than something numerically quantified.
function-id not zero-padding function signature
```
ERC20:
+---------------------------------------+------------+
| Name | ID |
+---------------------------------------+------------+
| totalSupply() | 0x18160ddd |
| balanceOf(address) | 0x70a08231 |
| allowance(address,address) | 0xdd62ed3e |
| transfer(address,uint256) | 0xa9059cbb |
| transferFrom(address,address,uint256) | 0x23b872dd |
| approve(address,uint256) | 0x95ea7b3 |
+---------------------------------------+------------+
```
It's a minor annoyance, but for approve it outputs `0x95ea7b3` instead of `0x095ea7b3`. It is the same numerically, of course, but the function signature is more of an opaque 4-byte identifier than something numerically quantified.
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `slither/printers/summary/function_ids.py`
Content:
```
1 """
2 Module printing summary of the contract
3 """
4 from slither.printers.abstract_printer import AbstractPrinter
5 from slither.utils.function import get_function_id
6 from slither.utils.myprettytable import MyPrettyTable
7
8
9 class FunctionIds(AbstractPrinter):
10
11 ARGUMENT = 'function-id'
12 HELP = 'Print the keccack256 signature of the functions'
13
14 WIKI = 'https://github.com/trailofbits/slither/wiki/Printer-documentation#function-id'
15
16 def output(self, _filename):
17 """
18 _filename is not used
19 Args:
20 _filename(string)
21 """
22
23 txt = ''
24 all_tables = []
25 for contract in self.slither.contracts_derived:
26 txt += '\n{}:\n'.format(contract.name)
27 table = MyPrettyTable(['Name', 'ID'])
28 for function in contract.functions:
29 if function.visibility in ['public', 'external']:
30 table.add_row([function.solidity_signature, hex(get_function_id(function.solidity_signature))])
31 for variable in contract.state_variables:
32 if variable.visibility in ['public']:
33 sig = variable.function_name
34 table.add_row([sig, hex(get_function_id(sig))])
35 txt += str(table) + '\n'
36 all_tables.append((contract.name, table))
37
38 self.info(txt)
39
40 res = self.generate_output(txt)
41 for name, table in all_tables:
42 res.add_pretty_table(table, name)
43
44 return res
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/slither/printers/summary/function_ids.py b/slither/printers/summary/function_ids.py
--- a/slither/printers/summary/function_ids.py
+++ b/slither/printers/summary/function_ids.py
@@ -27,11 +27,13 @@
table = MyPrettyTable(['Name', 'ID'])
for function in contract.functions:
if function.visibility in ['public', 'external']:
- table.add_row([function.solidity_signature, hex(get_function_id(function.solidity_signature))])
+ function_id = get_function_id(function.solidity_signature)
+ table.add_row([function.solidity_signature, f"{function_id:#0{10}x}"])
for variable in contract.state_variables:
if variable.visibility in ['public']:
sig = variable.function_name
- table.add_row([sig, hex(get_function_id(sig))])
+ function_id = get_function_id(sig)
+ table.add_row([sig, f"{function_id:#0{10}x}"])
txt += str(table) + '\n'
all_tables.append((contract.name, table))
| {"golden_diff": "diff --git a/slither/printers/summary/function_ids.py b/slither/printers/summary/function_ids.py\n--- a/slither/printers/summary/function_ids.py\n+++ b/slither/printers/summary/function_ids.py\n@@ -27,11 +27,13 @@\n table = MyPrettyTable(['Name', 'ID'])\n for function in contract.functions:\n if function.visibility in ['public', 'external']:\n- table.add_row([function.solidity_signature, hex(get_function_id(function.solidity_signature))])\n+ function_id = get_function_id(function.solidity_signature)\n+ table.add_row([function.solidity_signature, f\"{function_id:#0{10}x}\"])\n for variable in contract.state_variables:\n if variable.visibility in ['public']:\n sig = variable.function_name\n- table.add_row([sig, hex(get_function_id(sig))])\n+ function_id = get_function_id(sig)\n+ table.add_row([sig, f\"{function_id:#0{10}x}\"])\n txt += str(table) + '\\n'\n all_tables.append((contract.name, table))\n", "issue": "function-id not zero-padding function signature \n```\r\nERC20:\r\n+---------------------------------------+------------+\r\n| Name | ID |\r\n+---------------------------------------+------------+\r\n| totalSupply() | 0x18160ddd |\r\n| balanceOf(address) | 0x70a08231 |\r\n| allowance(address,address) | 0xdd62ed3e |\r\n| transfer(address,uint256) | 0xa9059cbb |\r\n| transferFrom(address,address,uint256) | 0x23b872dd |\r\n| approve(address,uint256) | 0x95ea7b3 |\r\n+---------------------------------------+------------+\r\n\r\n```\r\n\r\nIt's a minor annoyance, but for approve it outputs `0x95ea7b3` instead of `0x095ea7b3`. It is the same numerically, of course, but the function signature is more of an opaque 4-byte identifier than something numerically quantified.\r\n\r\n\nfunction-id not zero-padding function signature \n```\r\nERC20:\r\n+---------------------------------------+------------+\r\n| Name | ID |\r\n+---------------------------------------+------------+\r\n| totalSupply() | 0x18160ddd |\r\n| balanceOf(address) | 0x70a08231 |\r\n| allowance(address,address) | 0xdd62ed3e |\r\n| transfer(address,uint256) | 0xa9059cbb |\r\n| transferFrom(address,address,uint256) | 0x23b872dd |\r\n| approve(address,uint256) | 0x95ea7b3 |\r\n+---------------------------------------+------------+\r\n\r\n```\r\n\r\nIt's a minor annoyance, but for approve it outputs `0x95ea7b3` instead of `0x095ea7b3`. It is the same numerically, of course, but the function signature is more of an opaque 4-byte identifier than something numerically quantified.\r\n\r\n\n", "before_files": [{"content": "\"\"\"\n Module printing summary of the contract\n\"\"\"\nfrom slither.printers.abstract_printer import AbstractPrinter\nfrom slither.utils.function import get_function_id\nfrom slither.utils.myprettytable import MyPrettyTable\n\n\nclass FunctionIds(AbstractPrinter):\n\n ARGUMENT = 'function-id'\n HELP = 'Print the keccack256 signature of the functions'\n\n WIKI = 'https://github.com/trailofbits/slither/wiki/Printer-documentation#function-id'\n\n def output(self, _filename):\n \"\"\"\n _filename is not used\n Args:\n _filename(string)\n \"\"\"\n\n txt = ''\n all_tables = []\n for contract in self.slither.contracts_derived:\n txt += '\\n{}:\\n'.format(contract.name)\n table = MyPrettyTable(['Name', 'ID'])\n for function in contract.functions:\n if function.visibility in ['public', 'external']:\n table.add_row([function.solidity_signature, hex(get_function_id(function.solidity_signature))])\n for variable in contract.state_variables:\n if variable.visibility in ['public']:\n sig = variable.function_name\n table.add_row([sig, hex(get_function_id(sig))])\n txt += str(table) + '\\n'\n all_tables.append((contract.name, table))\n\n self.info(txt)\n\n res = self.generate_output(txt)\n for name, table in all_tables:\n res.add_pretty_table(table, name)\n\n return res", "path": "slither/printers/summary/function_ids.py"}], "after_files": [{"content": "\"\"\"\n Module printing summary of the contract\n\"\"\"\nfrom slither.printers.abstract_printer import AbstractPrinter\nfrom slither.utils.function import get_function_id\nfrom slither.utils.myprettytable import MyPrettyTable\n\n\nclass FunctionIds(AbstractPrinter):\n\n ARGUMENT = 'function-id'\n HELP = 'Print the keccack256 signature of the functions'\n\n WIKI = 'https://github.com/trailofbits/slither/wiki/Printer-documentation#function-id'\n\n def output(self, _filename):\n \"\"\"\n _filename is not used\n Args:\n _filename(string)\n \"\"\"\n\n txt = ''\n all_tables = []\n for contract in self.slither.contracts_derived:\n txt += '\\n{}:\\n'.format(contract.name)\n table = MyPrettyTable(['Name', 'ID'])\n for function in contract.functions:\n if function.visibility in ['public', 'external']:\n function_id = get_function_id(function.solidity_signature)\n table.add_row([function.solidity_signature, f\"{function_id:#0{10}x}\"])\n for variable in contract.state_variables:\n if variable.visibility in ['public']:\n sig = variable.function_name\n function_id = get_function_id(sig)\n table.add_row([sig, f\"{function_id:#0{10}x}\"])\n txt += str(table) + '\\n'\n all_tables.append((contract.name, table))\n\n self.info(txt)\n\n res = self.generate_output(txt)\n for name, table in all_tables:\n res.add_pretty_table(table, name)\n\n return res", "path": "slither/printers/summary/function_ids.py"}]} | 1,117 | 243 |
gh_patches_debug_34752 | rasdani/github-patches | git_diff | litestar-org__litestar-288 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Partial doesn't work with inherited fields
```python
from starlite import Partial, get
from pydantic import BaseModel
class Parent(BaseModel):
foo: int
class Child(Parent):
bar: int
@get("/test")
def example(obj: Partial[Child]) -> None:
print(obj)
```
In the above example, `Partial[Child]` only accepts the field `bar: Optional[int]` and ignores all fields from the superclass. I couldn't find this behaviour documented anywhere so I assume this isn't intended?
```python
Python 3.10.5 (main, Jun 23 2022, 17:14:57) [Clang 13.1.6 (clang-1316.0.21.2.5)] on darwin
Type "help", "copyright", "credits" or "license" for more information.
>>> from starlite import Partial
>>> from pydantic import BaseModel
>>> class Parent(BaseModel):
... foo: int
...
>>> class Child(Parent):
... bar: int
...
>>> PartialChild = Partial[Child]
>>> PartialChild.__annotations__
{'bar': typing.Optional[int]}
>>>
```
This behaviour can also be seen above
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `starlite/types.py`
Content:
```
1 from typing import (
2 TYPE_CHECKING,
3 Any,
4 Awaitable,
5 Callable,
6 Dict,
7 Generic,
8 Optional,
9 Tuple,
10 Type,
11 TypeVar,
12 Union,
13 cast,
14 )
15
16 from openapi_schema_pydantic.v3.v3_1_0.header import Header
17 from pydantic import BaseModel, create_model
18 from pydantic.typing import AnyCallable
19 from starlette.exceptions import HTTPException as StarletteHTTPException
20 from starlette.middleware import Middleware as StarletteMiddleware
21 from starlette.middleware.base import BaseHTTPMiddleware
22 from starlette.requests import HTTPConnection
23 from starlette.responses import Response as StarletteResponse
24 from typing_extensions import Literal, Protocol, runtime_checkable
25
26 from starlite.exceptions import HTTPException
27 from starlite.response import Response
28
29 try:
30 # python 3.9 changed these variable
31 from typing import _UnionGenericAlias as GenericAlias # type: ignore
32 except ImportError: # pragma: no cover
33 from typing import _GenericAlias as GenericAlias # type: ignore
34
35 if TYPE_CHECKING:
36 from starlette.types import ASGIApp, Receive, Scope, Send
37
38 from starlite.connection import Request # noqa: TC004
39 from starlite.controller import Controller # noqa: TC004
40 from starlite.datastructures import State # noqa: TC004
41 from starlite.handlers import BaseRouteHandler # noqa: TC004
42 from starlite.router import Router # noqa: TC004
43 else:
44 Request = Any
45 WebSocket = Any
46 BaseRouteHandler = Any
47 Controller = Any
48 Router = Any
49 State = Any
50
51 T = TypeVar("T", bound=BaseModel)
52 H = TypeVar("H", bound=HTTPConnection)
53
54 ExceptionHandler = Callable[
55 [Request, Union[Exception, HTTPException, StarletteHTTPException]], Union[Response, StarletteResponse]
56 ]
57 LifeCycleHandler = Union[
58 Callable[[], Any],
59 Callable[[State], Any],
60 Callable[[], Awaitable[Any]],
61 Callable[[State], Awaitable[Any]],
62 ]
63 Guard = Union[Callable[[H, BaseRouteHandler], Awaitable[None]], Callable[[H, BaseRouteHandler], None]]
64 Method = Union[Literal["GET"], Literal["POST"], Literal["DELETE"], Literal["PATCH"], Literal["PUT"], Literal["HEAD"]]
65 ReservedKwargs = Union[
66 Literal["request"],
67 Literal["socket"],
68 Literal["headers"],
69 Literal["query"],
70 Literal["cookies"],
71 Literal["state"],
72 Literal["data"],
73 ]
74 ControllerRouterHandler = Union[Type[Controller], BaseRouteHandler, Router, AnyCallable]
75
76 # connection-lifecycle hook handlers
77 BeforeRequestHandler = Union[Callable[[Request], Any], Callable[[Request], Awaitable[Any]]]
78 AfterRequestHandler = Union[
79 Callable[[Response], Response],
80 Callable[[Response], Awaitable[Response]],
81 Callable[[StarletteResponse], StarletteResponse],
82 Callable[[StarletteResponse], Awaitable[StarletteResponse]],
83 ]
84 AfterResponseHandler = Union[Callable[[Request], None], Callable[[Request], Awaitable[None]]]
85
86 AsyncAnyCallable = Callable[..., Awaitable[Any]]
87 CacheKeyBuilder = Callable[[Request], str]
88
89
90 @runtime_checkable
91 class MiddlewareProtocol(Protocol):
92 def __init__(self, app: "ASGIApp"): # pragma: no cover
93 ...
94
95 async def __call__(self, scope: "Scope", receive: "Receive", send: "Send") -> None: # pragma: no cover
96 ...
97
98
99 class Partial(Generic[T]):
100 _models: Dict[Type[T], Any] = {}
101
102 def __class_getitem__(cls, item: Type[T]) -> Type[T]:
103 """
104 Modifies a given T subclass of BaseModel to be all optional
105 """
106 if not cls._models.get(item):
107 field_definitions: Dict[str, Tuple[Any, None]] = {}
108 for field_name, field_type in item.__annotations__.items():
109 # we modify the field annotations to make it optional
110 if not isinstance(field_type, GenericAlias) or type(None) not in field_type.__args__:
111 field_definitions[field_name] = (Optional[field_type], None)
112 else:
113 field_definitions[field_name] = (field_type, None)
114 cls._models[item] = create_model("Partial" + item.__name__, **field_definitions) # type: ignore
115 return cast(Type[T], cls._models.get(item))
116
117
118 class ResponseHeader(Header):
119 value: Any = ...
120
121
122 Middleware = Union[StarletteMiddleware, Type[BaseHTTPMiddleware], Type[MiddlewareProtocol]]
123
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/starlite/types.py b/starlite/types.py
--- a/starlite/types.py
+++ b/starlite/types.py
@@ -11,6 +11,7 @@
TypeVar,
Union,
cast,
+ get_type_hints,
)
from openapi_schema_pydantic.v3.v3_1_0.header import Header
@@ -23,7 +24,7 @@
from starlette.responses import Response as StarletteResponse
from typing_extensions import Literal, Protocol, runtime_checkable
-from starlite.exceptions import HTTPException
+from starlite.exceptions import HTTPException, ImproperlyConfiguredException
from starlite.response import Response
try:
@@ -103,15 +104,23 @@
"""
Modifies a given T subclass of BaseModel to be all optional
"""
+ if not issubclass(item, BaseModel):
+ raise ImproperlyConfiguredException(f"Partial[{item}] must be a subclass of BaseModel")
if not cls._models.get(item):
field_definitions: Dict[str, Tuple[Any, None]] = {}
- for field_name, field_type in item.__annotations__.items():
- # we modify the field annotations to make it optional
- if not isinstance(field_type, GenericAlias) or type(None) not in field_type.__args__:
- field_definitions[field_name] = (Optional[field_type], None)
+ # traverse the object's mro and get all annotations
+ # until we find a BaseModel.
+ for obj in item.mro():
+ if issubclass(obj, BaseModel):
+ for field_name, field_type in get_type_hints(obj).items():
+ # we modify the field annotations to make it optional
+ if not isinstance(field_type, GenericAlias) or type(None) not in field_type.__args__:
+ field_definitions[field_name] = (Optional[field_type], None)
+ else:
+ field_definitions[field_name] = (field_type, None)
else:
- field_definitions[field_name] = (field_type, None)
- cls._models[item] = create_model("Partial" + item.__name__, **field_definitions) # type: ignore
+ break
+ cls._models[item] = create_model(f"Partial{item.__name__}", **field_definitions) # type: ignore
return cast(Type[T], cls._models.get(item))
| {"golden_diff": "diff --git a/starlite/types.py b/starlite/types.py\n--- a/starlite/types.py\n+++ b/starlite/types.py\n@@ -11,6 +11,7 @@\n TypeVar,\n Union,\n cast,\n+ get_type_hints,\n )\n \n from openapi_schema_pydantic.v3.v3_1_0.header import Header\n@@ -23,7 +24,7 @@\n from starlette.responses import Response as StarletteResponse\n from typing_extensions import Literal, Protocol, runtime_checkable\n \n-from starlite.exceptions import HTTPException\n+from starlite.exceptions import HTTPException, ImproperlyConfiguredException\n from starlite.response import Response\n \n try:\n@@ -103,15 +104,23 @@\n \"\"\"\n Modifies a given T subclass of BaseModel to be all optional\n \"\"\"\n+ if not issubclass(item, BaseModel):\n+ raise ImproperlyConfiguredException(f\"Partial[{item}] must be a subclass of BaseModel\")\n if not cls._models.get(item):\n field_definitions: Dict[str, Tuple[Any, None]] = {}\n- for field_name, field_type in item.__annotations__.items():\n- # we modify the field annotations to make it optional\n- if not isinstance(field_type, GenericAlias) or type(None) not in field_type.__args__:\n- field_definitions[field_name] = (Optional[field_type], None)\n+ # traverse the object's mro and get all annotations\n+ # until we find a BaseModel.\n+ for obj in item.mro():\n+ if issubclass(obj, BaseModel):\n+ for field_name, field_type in get_type_hints(obj).items():\n+ # we modify the field annotations to make it optional\n+ if not isinstance(field_type, GenericAlias) or type(None) not in field_type.__args__:\n+ field_definitions[field_name] = (Optional[field_type], None)\n+ else:\n+ field_definitions[field_name] = (field_type, None)\n else:\n- field_definitions[field_name] = (field_type, None)\n- cls._models[item] = create_model(\"Partial\" + item.__name__, **field_definitions) # type: ignore\n+ break\n+ cls._models[item] = create_model(f\"Partial{item.__name__}\", **field_definitions) # type: ignore\n return cast(Type[T], cls._models.get(item))\n", "issue": "Partial doesn't work with inherited fields\n```python\r\nfrom starlite import Partial, get\r\nfrom pydantic import BaseModel\r\n\r\nclass Parent(BaseModel):\r\n foo: int\r\n\r\nclass Child(Parent):\r\n bar: int\r\n\r\n@get(\"/test\")\r\ndef example(obj: Partial[Child]) -> None:\r\n print(obj)\r\n```\r\n\r\nIn the above example, `Partial[Child]` only accepts the field `bar: Optional[int]` and ignores all fields from the superclass. I couldn't find this behaviour documented anywhere so I assume this isn't intended?\r\n\r\n```python\r\nPython 3.10.5 (main, Jun 23 2022, 17:14:57) [Clang 13.1.6 (clang-1316.0.21.2.5)] on darwin\r\nType \"help\", \"copyright\", \"credits\" or \"license\" for more information.\r\n>>> from starlite import Partial\r\n>>> from pydantic import BaseModel\r\n>>> class Parent(BaseModel):\r\n... foo: int\r\n...\r\n>>> class Child(Parent):\r\n... bar: int\r\n...\r\n>>> PartialChild = Partial[Child]\r\n>>> PartialChild.__annotations__\r\n{'bar': typing.Optional[int]}\r\n>>>\r\n```\r\n\r\nThis behaviour can also be seen above\r\n\n", "before_files": [{"content": "from typing import (\n TYPE_CHECKING,\n Any,\n Awaitable,\n Callable,\n Dict,\n Generic,\n Optional,\n Tuple,\n Type,\n TypeVar,\n Union,\n cast,\n)\n\nfrom openapi_schema_pydantic.v3.v3_1_0.header import Header\nfrom pydantic import BaseModel, create_model\nfrom pydantic.typing import AnyCallable\nfrom starlette.exceptions import HTTPException as StarletteHTTPException\nfrom starlette.middleware import Middleware as StarletteMiddleware\nfrom starlette.middleware.base import BaseHTTPMiddleware\nfrom starlette.requests import HTTPConnection\nfrom starlette.responses import Response as StarletteResponse\nfrom typing_extensions import Literal, Protocol, runtime_checkable\n\nfrom starlite.exceptions import HTTPException\nfrom starlite.response import Response\n\ntry:\n # python 3.9 changed these variable\n from typing import _UnionGenericAlias as GenericAlias # type: ignore\nexcept ImportError: # pragma: no cover\n from typing import _GenericAlias as GenericAlias # type: ignore\n\nif TYPE_CHECKING:\n from starlette.types import ASGIApp, Receive, Scope, Send\n\n from starlite.connection import Request # noqa: TC004\n from starlite.controller import Controller # noqa: TC004\n from starlite.datastructures import State # noqa: TC004\n from starlite.handlers import BaseRouteHandler # noqa: TC004\n from starlite.router import Router # noqa: TC004\nelse:\n Request = Any\n WebSocket = Any\n BaseRouteHandler = Any\n Controller = Any\n Router = Any\n State = Any\n\nT = TypeVar(\"T\", bound=BaseModel)\nH = TypeVar(\"H\", bound=HTTPConnection)\n\nExceptionHandler = Callable[\n [Request, Union[Exception, HTTPException, StarletteHTTPException]], Union[Response, StarletteResponse]\n]\nLifeCycleHandler = Union[\n Callable[[], Any],\n Callable[[State], Any],\n Callable[[], Awaitable[Any]],\n Callable[[State], Awaitable[Any]],\n]\nGuard = Union[Callable[[H, BaseRouteHandler], Awaitable[None]], Callable[[H, BaseRouteHandler], None]]\nMethod = Union[Literal[\"GET\"], Literal[\"POST\"], Literal[\"DELETE\"], Literal[\"PATCH\"], Literal[\"PUT\"], Literal[\"HEAD\"]]\nReservedKwargs = Union[\n Literal[\"request\"],\n Literal[\"socket\"],\n Literal[\"headers\"],\n Literal[\"query\"],\n Literal[\"cookies\"],\n Literal[\"state\"],\n Literal[\"data\"],\n]\nControllerRouterHandler = Union[Type[Controller], BaseRouteHandler, Router, AnyCallable]\n\n# connection-lifecycle hook handlers\nBeforeRequestHandler = Union[Callable[[Request], Any], Callable[[Request], Awaitable[Any]]]\nAfterRequestHandler = Union[\n Callable[[Response], Response],\n Callable[[Response], Awaitable[Response]],\n Callable[[StarletteResponse], StarletteResponse],\n Callable[[StarletteResponse], Awaitable[StarletteResponse]],\n]\nAfterResponseHandler = Union[Callable[[Request], None], Callable[[Request], Awaitable[None]]]\n\nAsyncAnyCallable = Callable[..., Awaitable[Any]]\nCacheKeyBuilder = Callable[[Request], str]\n\n\n@runtime_checkable\nclass MiddlewareProtocol(Protocol):\n def __init__(self, app: \"ASGIApp\"): # pragma: no cover\n ...\n\n async def __call__(self, scope: \"Scope\", receive: \"Receive\", send: \"Send\") -> None: # pragma: no cover\n ...\n\n\nclass Partial(Generic[T]):\n _models: Dict[Type[T], Any] = {}\n\n def __class_getitem__(cls, item: Type[T]) -> Type[T]:\n \"\"\"\n Modifies a given T subclass of BaseModel to be all optional\n \"\"\"\n if not cls._models.get(item):\n field_definitions: Dict[str, Tuple[Any, None]] = {}\n for field_name, field_type in item.__annotations__.items():\n # we modify the field annotations to make it optional\n if not isinstance(field_type, GenericAlias) or type(None) not in field_type.__args__:\n field_definitions[field_name] = (Optional[field_type], None)\n else:\n field_definitions[field_name] = (field_type, None)\n cls._models[item] = create_model(\"Partial\" + item.__name__, **field_definitions) # type: ignore\n return cast(Type[T], cls._models.get(item))\n\n\nclass ResponseHeader(Header):\n value: Any = ...\n\n\nMiddleware = Union[StarletteMiddleware, Type[BaseHTTPMiddleware], Type[MiddlewareProtocol]]\n", "path": "starlite/types.py"}], "after_files": [{"content": "from typing import (\n TYPE_CHECKING,\n Any,\n Awaitable,\n Callable,\n Dict,\n Generic,\n Optional,\n Tuple,\n Type,\n TypeVar,\n Union,\n cast,\n get_type_hints,\n)\n\nfrom openapi_schema_pydantic.v3.v3_1_0.header import Header\nfrom pydantic import BaseModel, create_model\nfrom pydantic.typing import AnyCallable\nfrom starlette.exceptions import HTTPException as StarletteHTTPException\nfrom starlette.middleware import Middleware as StarletteMiddleware\nfrom starlette.middleware.base import BaseHTTPMiddleware\nfrom starlette.requests import HTTPConnection\nfrom starlette.responses import Response as StarletteResponse\nfrom typing_extensions import Literal, Protocol, runtime_checkable\n\nfrom starlite.exceptions import HTTPException, ImproperlyConfiguredException\nfrom starlite.response import Response\n\ntry:\n # python 3.9 changed these variable\n from typing import _UnionGenericAlias as GenericAlias # type: ignore\nexcept ImportError: # pragma: no cover\n from typing import _GenericAlias as GenericAlias # type: ignore\n\nif TYPE_CHECKING:\n from starlette.types import ASGIApp, Receive, Scope, Send\n\n from starlite.connection import Request # noqa: TC004\n from starlite.controller import Controller # noqa: TC004\n from starlite.datastructures import State # noqa: TC004\n from starlite.handlers import BaseRouteHandler # noqa: TC004\n from starlite.router import Router # noqa: TC004\nelse:\n Request = Any\n WebSocket = Any\n BaseRouteHandler = Any\n Controller = Any\n Router = Any\n State = Any\n\nT = TypeVar(\"T\", bound=BaseModel)\nH = TypeVar(\"H\", bound=HTTPConnection)\n\nExceptionHandler = Callable[\n [Request, Union[Exception, HTTPException, StarletteHTTPException]], Union[Response, StarletteResponse]\n]\nLifeCycleHandler = Union[\n Callable[[], Any],\n Callable[[State], Any],\n Callable[[], Awaitable[Any]],\n Callable[[State], Awaitable[Any]],\n]\nGuard = Union[Callable[[H, BaseRouteHandler], Awaitable[None]], Callable[[H, BaseRouteHandler], None]]\nMethod = Union[Literal[\"GET\"], Literal[\"POST\"], Literal[\"DELETE\"], Literal[\"PATCH\"], Literal[\"PUT\"], Literal[\"HEAD\"]]\nReservedKwargs = Union[\n Literal[\"request\"],\n Literal[\"socket\"],\n Literal[\"headers\"],\n Literal[\"query\"],\n Literal[\"cookies\"],\n Literal[\"state\"],\n Literal[\"data\"],\n]\nControllerRouterHandler = Union[Type[Controller], BaseRouteHandler, Router, AnyCallable]\n\n# connection-lifecycle hook handlers\nBeforeRequestHandler = Union[Callable[[Request], Any], Callable[[Request], Awaitable[Any]]]\nAfterRequestHandler = Union[\n Callable[[Response], Response],\n Callable[[Response], Awaitable[Response]],\n Callable[[StarletteResponse], StarletteResponse],\n Callable[[StarletteResponse], Awaitable[StarletteResponse]],\n]\nAfterResponseHandler = Union[Callable[[Request], None], Callable[[Request], Awaitable[None]]]\n\nAsyncAnyCallable = Callable[..., Awaitable[Any]]\nCacheKeyBuilder = Callable[[Request], str]\n\n\n@runtime_checkable\nclass MiddlewareProtocol(Protocol):\n def __init__(self, app: \"ASGIApp\"): # pragma: no cover\n ...\n\n async def __call__(self, scope: \"Scope\", receive: \"Receive\", send: \"Send\") -> None: # pragma: no cover\n ...\n\n\nclass Partial(Generic[T]):\n _models: Dict[Type[T], Any] = {}\n\n def __class_getitem__(cls, item: Type[T]) -> Type[T]:\n \"\"\"\n Modifies a given T subclass of BaseModel to be all optional\n \"\"\"\n if not issubclass(item, BaseModel):\n raise ImproperlyConfiguredException(f\"Partial[{item}] must be a subclass of BaseModel\")\n if not cls._models.get(item):\n field_definitions: Dict[str, Tuple[Any, None]] = {}\n # traverse the object's mro and get all annotations\n # until we find a BaseModel.\n for obj in item.mro():\n if issubclass(obj, BaseModel):\n for field_name, field_type in get_type_hints(obj).items():\n # we modify the field annotations to make it optional\n if not isinstance(field_type, GenericAlias) or type(None) not in field_type.__args__:\n field_definitions[field_name] = (Optional[field_type], None)\n else:\n field_definitions[field_name] = (field_type, None)\n else:\n break\n cls._models[item] = create_model(f\"Partial{item.__name__}\", **field_definitions) # type: ignore\n return cast(Type[T], cls._models.get(item))\n\n\nclass ResponseHeader(Header):\n value: Any = ...\n\n\nMiddleware = Union[StarletteMiddleware, Type[BaseHTTPMiddleware], Type[MiddlewareProtocol]]\n", "path": "starlite/types.py"}]} | 1,791 | 518 |
gh_patches_debug_64987 | rasdani/github-patches | git_diff | googleapis__google-auth-library-python-937 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
new cachetools version (5.0.0+) can't work with google-auth
`cachetools` has released a new version (5.0.0) which conflicts with google-auth requirements of it being <5, this prevents updates to the `cachetools` package and pose a potential security concern (as updates are no longer possible to it)
```
The conflict is caused by:
The user requested cachetools==5.0.0
google-auth 2.3.3 depends on cachetools<5.0 and >=2.0.0
```
issue seems in https://github.com/googleapis/google-auth-library-python/blob/3c3fbf40b07e090f2be7fac5b304dbf438b5cd6c/setup.py#L23
#### Environment details
- OS: alpine3.11
- Python version: python:3.8.6
- pip version: 20.3.3
- `google-auth` version: 2.3.3 (latest at time of writing)
#### Steps to reproduce
1. try pip install using latest `cachetools` with latest `google-auth`
2. pip fails
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `setup.py`
Content:
```
1 # Copyright 2014 Google Inc.
2 #
3 # Licensed under the Apache License, Version 2.0 (the "License");
4 # you may not use this file except in compliance with the License.
5 # You may obtain a copy of the License at
6 #
7 # http://www.apache.org/licenses/LICENSE-2.0
8 #
9 # Unless required by applicable law or agreed to in writing, software
10 # distributed under the License is distributed on an "AS IS" BASIS,
11 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12 # See the License for the specific language governing permissions and
13 # limitations under the License.
14
15 import io
16 import os
17
18 from setuptools import find_packages
19 from setuptools import setup
20
21
22 DEPENDENCIES = (
23 "cachetools>=2.0.0,<5.0",
24 "pyasn1-modules>=0.2.1",
25 # rsa==4.5 is the last version to support 2.7
26 # https://github.com/sybrenstuvel/python-rsa/issues/152#issuecomment-643470233
27 'rsa<4.6; python_version < "3.6"',
28 'rsa>=3.1.4,<5; python_version >= "3.6"',
29 # install enum34 to support 2.7. enum34 only works up to python version 3.3.
30 'enum34>=1.1.10; python_version < "3.4"',
31 "six>=1.9.0",
32 )
33
34 extras = {
35 "aiohttp": [
36 "aiohttp >= 3.6.2, < 4.0.0dev; python_version>='3.6'",
37 "requests >= 2.20.0, < 3.0.0dev",
38 ],
39 "pyopenssl": "pyopenssl>=20.0.0",
40 "reauth": "pyu2f>=0.1.5",
41 }
42
43 with io.open("README.rst", "r") as fh:
44 long_description = fh.read()
45
46 package_root = os.path.abspath(os.path.dirname(__file__))
47
48 version = {}
49 with open(os.path.join(package_root, "google/auth/version.py")) as fp:
50 exec(fp.read(), version)
51 version = version["__version__"]
52
53 setup(
54 name="google-auth",
55 version=version,
56 author="Google Cloud Platform",
57 author_email="[email protected]",
58 description="Google Authentication Library",
59 long_description=long_description,
60 url="https://github.com/googleapis/google-auth-library-python",
61 packages=find_packages(exclude=("tests*", "system_tests*")),
62 namespace_packages=("google",),
63 install_requires=DEPENDENCIES,
64 extras_require=extras,
65 python_requires=">=2.7,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*,!=3.5.*",
66 license="Apache 2.0",
67 keywords="google auth oauth client",
68 classifiers=[
69 "Programming Language :: Python :: 3",
70 "Programming Language :: Python :: 3.6",
71 "Programming Language :: Python :: 3.7",
72 "Programming Language :: Python :: 3.8",
73 "Programming Language :: Python :: 3.9",
74 "Programming Language :: Python :: 3.10",
75 "Development Status :: 5 - Production/Stable",
76 "Intended Audience :: Developers",
77 "License :: OSI Approved :: Apache Software License",
78 "Operating System :: POSIX",
79 "Operating System :: Microsoft :: Windows",
80 "Operating System :: MacOS :: MacOS X",
81 "Operating System :: OS Independent",
82 "Topic :: Internet :: WWW/HTTP",
83 ],
84 )
85
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/setup.py b/setup.py
--- a/setup.py
+++ b/setup.py
@@ -20,7 +20,7 @@
DEPENDENCIES = (
- "cachetools>=2.0.0,<5.0",
+ "cachetools>=2.0.0,<6.0",
"pyasn1-modules>=0.2.1",
# rsa==4.5 is the last version to support 2.7
# https://github.com/sybrenstuvel/python-rsa/issues/152#issuecomment-643470233
| {"golden_diff": "diff --git a/setup.py b/setup.py\n--- a/setup.py\n+++ b/setup.py\n@@ -20,7 +20,7 @@\n \n \n DEPENDENCIES = (\n- \"cachetools>=2.0.0,<5.0\",\n+ \"cachetools>=2.0.0,<6.0\",\n \"pyasn1-modules>=0.2.1\",\n # rsa==4.5 is the last version to support 2.7\n # https://github.com/sybrenstuvel/python-rsa/issues/152#issuecomment-643470233\n", "issue": "new cachetools version (5.0.0+) can't work with google-auth\n`cachetools` has released a new version (5.0.0) which conflicts with google-auth requirements of it being <5, this prevents updates to the `cachetools` package and pose a potential security concern (as updates are no longer possible to it)\r\n\r\n```\r\nThe conflict is caused by:\r\n The user requested cachetools==5.0.0\r\n google-auth 2.3.3 depends on cachetools<5.0 and >=2.0.0\r\n```\r\n\r\nissue seems in https://github.com/googleapis/google-auth-library-python/blob/3c3fbf40b07e090f2be7fac5b304dbf438b5cd6c/setup.py#L23 \r\n\r\n#### Environment details\r\n\r\n - OS: alpine3.11\r\n - Python version: python:3.8.6\r\n - pip version: 20.3.3\r\n - `google-auth` version: 2.3.3 (latest at time of writing)\r\n\r\n#### Steps to reproduce\r\n\r\n 1. try pip install using latest `cachetools` with latest `google-auth`\r\n 2. pip fails\r\n\n", "before_files": [{"content": "# Copyright 2014 Google Inc.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nimport io\nimport os\n\nfrom setuptools import find_packages\nfrom setuptools import setup\n\n\nDEPENDENCIES = (\n \"cachetools>=2.0.0,<5.0\",\n \"pyasn1-modules>=0.2.1\",\n # rsa==4.5 is the last version to support 2.7\n # https://github.com/sybrenstuvel/python-rsa/issues/152#issuecomment-643470233\n 'rsa<4.6; python_version < \"3.6\"',\n 'rsa>=3.1.4,<5; python_version >= \"3.6\"',\n # install enum34 to support 2.7. enum34 only works up to python version 3.3.\n 'enum34>=1.1.10; python_version < \"3.4\"',\n \"six>=1.9.0\",\n)\n\nextras = {\n \"aiohttp\": [\n \"aiohttp >= 3.6.2, < 4.0.0dev; python_version>='3.6'\",\n \"requests >= 2.20.0, < 3.0.0dev\",\n ],\n \"pyopenssl\": \"pyopenssl>=20.0.0\",\n \"reauth\": \"pyu2f>=0.1.5\",\n}\n\nwith io.open(\"README.rst\", \"r\") as fh:\n long_description = fh.read()\n\npackage_root = os.path.abspath(os.path.dirname(__file__))\n\nversion = {}\nwith open(os.path.join(package_root, \"google/auth/version.py\")) as fp:\n exec(fp.read(), version)\nversion = version[\"__version__\"]\n\nsetup(\n name=\"google-auth\",\n version=version,\n author=\"Google Cloud Platform\",\n author_email=\"[email protected]\",\n description=\"Google Authentication Library\",\n long_description=long_description,\n url=\"https://github.com/googleapis/google-auth-library-python\",\n packages=find_packages(exclude=(\"tests*\", \"system_tests*\")),\n namespace_packages=(\"google\",),\n install_requires=DEPENDENCIES,\n extras_require=extras,\n python_requires=\">=2.7,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*,!=3.5.*\",\n license=\"Apache 2.0\",\n keywords=\"google auth oauth client\",\n classifiers=[\n \"Programming Language :: Python :: 3\",\n \"Programming Language :: Python :: 3.6\",\n \"Programming Language :: Python :: 3.7\",\n \"Programming Language :: Python :: 3.8\",\n \"Programming Language :: Python :: 3.9\",\n \"Programming Language :: Python :: 3.10\",\n \"Development Status :: 5 - Production/Stable\",\n \"Intended Audience :: Developers\",\n \"License :: OSI Approved :: Apache Software License\",\n \"Operating System :: POSIX\",\n \"Operating System :: Microsoft :: Windows\",\n \"Operating System :: MacOS :: MacOS X\",\n \"Operating System :: OS Independent\",\n \"Topic :: Internet :: WWW/HTTP\",\n ],\n)\n", "path": "setup.py"}], "after_files": [{"content": "# Copyright 2014 Google Inc.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nimport io\nimport os\n\nfrom setuptools import find_packages\nfrom setuptools import setup\n\n\nDEPENDENCIES = (\n \"cachetools>=2.0.0,<6.0\",\n \"pyasn1-modules>=0.2.1\",\n # rsa==4.5 is the last version to support 2.7\n # https://github.com/sybrenstuvel/python-rsa/issues/152#issuecomment-643470233\n 'rsa<4.6; python_version < \"3.6\"',\n 'rsa>=3.1.4,<5; python_version >= \"3.6\"',\n # install enum34 to support 2.7. enum34 only works up to python version 3.3.\n 'enum34>=1.1.10; python_version < \"3.4\"',\n \"six>=1.9.0\",\n)\n\nextras = {\n \"aiohttp\": [\n \"aiohttp >= 3.6.2, < 4.0.0dev; python_version>='3.6'\",\n \"requests >= 2.20.0, < 3.0.0dev\",\n ],\n \"pyopenssl\": \"pyopenssl>=20.0.0\",\n \"reauth\": \"pyu2f>=0.1.5\",\n}\n\nwith io.open(\"README.rst\", \"r\") as fh:\n long_description = fh.read()\n\npackage_root = os.path.abspath(os.path.dirname(__file__))\n\nversion = {}\nwith open(os.path.join(package_root, \"google/auth/version.py\")) as fp:\n exec(fp.read(), version)\nversion = version[\"__version__\"]\n\nsetup(\n name=\"google-auth\",\n version=version,\n author=\"Google Cloud Platform\",\n author_email=\"[email protected]\",\n description=\"Google Authentication Library\",\n long_description=long_description,\n url=\"https://github.com/googleapis/google-auth-library-python\",\n packages=find_packages(exclude=(\"tests*\", \"system_tests*\")),\n namespace_packages=(\"google\",),\n install_requires=DEPENDENCIES,\n extras_require=extras,\n python_requires=\">=2.7,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*,!=3.5.*\",\n license=\"Apache 2.0\",\n keywords=\"google auth oauth client\",\n classifiers=[\n \"Programming Language :: Python :: 3\",\n \"Programming Language :: Python :: 3.6\",\n \"Programming Language :: Python :: 3.7\",\n \"Programming Language :: Python :: 3.8\",\n \"Programming Language :: Python :: 3.9\",\n \"Programming Language :: Python :: 3.10\",\n \"Development Status :: 5 - Production/Stable\",\n \"Intended Audience :: Developers\",\n \"License :: OSI Approved :: Apache Software License\",\n \"Operating System :: POSIX\",\n \"Operating System :: Microsoft :: Windows\",\n \"Operating System :: MacOS :: MacOS X\",\n \"Operating System :: OS Independent\",\n \"Topic :: Internet :: WWW/HTTP\",\n ],\n)\n", "path": "setup.py"}]} | 1,500 | 138 |
gh_patches_debug_20415 | rasdani/github-patches | git_diff | ansible__awx-12803 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Token and Session Expiration never run after the first time
### Please confirm the following
- [X] I agree to follow this project's [code of conduct](https://docs.ansible.com/ansible/latest/community/code_of_conduct.html).
- [X] I have checked the [current issues](https://github.com/ansible/awx/issues) for duplicates.
- [X] I understand that AWX is open source software provided for free and that I might not receive a timely response.
### Bug Summary
Looks like when we implemented token and session cleanup way back in https://github.com/ansible/awx/pull/3856
We populated the recurrence rule incorrectly:
https://github.com/ansible/awx/blob/8a06ffbe15c9f8e68b1da86e5ca7daf5ecfd6da4/awx/main/migrations/_create_system_jobs.py#L39
This schedule will only ever run once due to `COUNT=1`.... we should omit that so that it will periodically run.
### AWX version
latest
### Select the relevant components
- [ ] UI
- [X] API
- [ ] Docs
- [ ] Collection
- [ ] CLI
- [ ] Other
### Installation method
minishift
### Modifications
no
### Ansible version
_No response_
### Operating system
_No response_
### Web browser
_No response_
### Steps to reproduce
install awx
### Expected results
periodic running of these management jobs
### Actual results
the jobs only run once
### Additional information
_No response_
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `awx/main/migrations/_create_system_jobs.py`
Content:
```
1 import logging
2
3 from django.utils.timezone import now
4
5 logger = logging.getLogger('awx.main.migrations')
6
7 __all__ = ['create_collection_jt', 'create_clearsessions_jt', 'create_cleartokens_jt']
8
9 '''
10 These methods are called by migrations to create various system job templates
11
12 Create default system job templates if not present. Create default schedules
13 only if new system job templates were created (i.e. new database).
14 '''
15
16
17 def create_clearsessions_jt(apps, schema_editor):
18
19 SystemJobTemplate = apps.get_model('main', 'SystemJobTemplate')
20 Schedule = apps.get_model('main', 'Schedule')
21 ContentType = apps.get_model('contenttypes', 'ContentType')
22 sjt_ct = ContentType.objects.get_for_model(SystemJobTemplate)
23 now_dt = now()
24 schedule_time = now_dt.strftime('%Y%m%dT%H%M%SZ')
25
26 sjt, created = SystemJobTemplate.objects.get_or_create(
27 job_type='cleanup_sessions',
28 defaults=dict(
29 name='Cleanup Expired Sessions',
30 description='Cleans out expired browser sessions',
31 polymorphic_ctype=sjt_ct,
32 created=now_dt,
33 modified=now_dt,
34 ),
35 )
36 if created:
37 sched = Schedule(
38 name='Cleanup Expired Sessions',
39 rrule='DTSTART:%s RRULE:FREQ=WEEKLY;INTERVAL=1;COUNT=1' % schedule_time,
40 description='Cleans out expired browser sessions',
41 enabled=True,
42 created=now_dt,
43 modified=now_dt,
44 extra_data={},
45 )
46 sched.unified_job_template = sjt
47 sched.save()
48
49
50 def create_cleartokens_jt(apps, schema_editor):
51
52 SystemJobTemplate = apps.get_model('main', 'SystemJobTemplate')
53 Schedule = apps.get_model('main', 'Schedule')
54 ContentType = apps.get_model('contenttypes', 'ContentType')
55 sjt_ct = ContentType.objects.get_for_model(SystemJobTemplate)
56 now_dt = now()
57 schedule_time = now_dt.strftime('%Y%m%dT%H%M%SZ')
58
59 sjt, created = SystemJobTemplate.objects.get_or_create(
60 job_type='cleanup_tokens',
61 defaults=dict(
62 name='Cleanup Expired OAuth 2 Tokens',
63 description='Cleanup expired OAuth 2 access and refresh tokens',
64 polymorphic_ctype=sjt_ct,
65 created=now_dt,
66 modified=now_dt,
67 ),
68 )
69 if created:
70 sched = Schedule(
71 name='Cleanup Expired OAuth 2 Tokens',
72 rrule='DTSTART:%s RRULE:FREQ=WEEKLY;INTERVAL=1;COUNT=1' % schedule_time,
73 description='Removes expired OAuth 2 access and refresh tokens',
74 enabled=True,
75 created=now_dt,
76 modified=now_dt,
77 extra_data={},
78 )
79 sched.unified_job_template = sjt
80 sched.save()
81
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/awx/main/migrations/_create_system_jobs.py b/awx/main/migrations/_create_system_jobs.py
--- a/awx/main/migrations/_create_system_jobs.py
+++ b/awx/main/migrations/_create_system_jobs.py
@@ -36,7 +36,7 @@
if created:
sched = Schedule(
name='Cleanup Expired Sessions',
- rrule='DTSTART:%s RRULE:FREQ=WEEKLY;INTERVAL=1;COUNT=1' % schedule_time,
+ rrule='DTSTART:%s RRULE:FREQ=WEEKLY;INTERVAL=1' % schedule_time,
description='Cleans out expired browser sessions',
enabled=True,
created=now_dt,
@@ -69,7 +69,7 @@
if created:
sched = Schedule(
name='Cleanup Expired OAuth 2 Tokens',
- rrule='DTSTART:%s RRULE:FREQ=WEEKLY;INTERVAL=1;COUNT=1' % schedule_time,
+ rrule='DTSTART:%s RRULE:FREQ=WEEKLY;INTERVAL=1' % schedule_time,
description='Removes expired OAuth 2 access and refresh tokens',
enabled=True,
created=now_dt,
| {"golden_diff": "diff --git a/awx/main/migrations/_create_system_jobs.py b/awx/main/migrations/_create_system_jobs.py\n--- a/awx/main/migrations/_create_system_jobs.py\n+++ b/awx/main/migrations/_create_system_jobs.py\n@@ -36,7 +36,7 @@\n if created:\n sched = Schedule(\n name='Cleanup Expired Sessions',\n- rrule='DTSTART:%s RRULE:FREQ=WEEKLY;INTERVAL=1;COUNT=1' % schedule_time,\n+ rrule='DTSTART:%s RRULE:FREQ=WEEKLY;INTERVAL=1' % schedule_time,\n description='Cleans out expired browser sessions',\n enabled=True,\n created=now_dt,\n@@ -69,7 +69,7 @@\n if created:\n sched = Schedule(\n name='Cleanup Expired OAuth 2 Tokens',\n- rrule='DTSTART:%s RRULE:FREQ=WEEKLY;INTERVAL=1;COUNT=1' % schedule_time,\n+ rrule='DTSTART:%s RRULE:FREQ=WEEKLY;INTERVAL=1' % schedule_time,\n description='Removes expired OAuth 2 access and refresh tokens',\n enabled=True,\n created=now_dt,\n", "issue": "Token and Session Expiration never run after the first time\n### Please confirm the following\n\n- [X] I agree to follow this project's [code of conduct](https://docs.ansible.com/ansible/latest/community/code_of_conduct.html).\n- [X] I have checked the [current issues](https://github.com/ansible/awx/issues) for duplicates.\n- [X] I understand that AWX is open source software provided for free and that I might not receive a timely response.\n\n### Bug Summary\n\nLooks like when we implemented token and session cleanup way back in https://github.com/ansible/awx/pull/3856\r\n\r\nWe populated the recurrence rule incorrectly:\r\n\r\nhttps://github.com/ansible/awx/blob/8a06ffbe15c9f8e68b1da86e5ca7daf5ecfd6da4/awx/main/migrations/_create_system_jobs.py#L39\r\n\r\nThis schedule will only ever run once due to `COUNT=1`.... we should omit that so that it will periodically run.\n\n### AWX version\n\nlatest\n\n### Select the relevant components\n\n- [ ] UI\n- [X] API\n- [ ] Docs\n- [ ] Collection\n- [ ] CLI\n- [ ] Other\n\n### Installation method\n\nminishift\n\n### Modifications\n\nno\n\n### Ansible version\n\n_No response_\n\n### Operating system\n\n_No response_\n\n### Web browser\n\n_No response_\n\n### Steps to reproduce\n\ninstall awx\n\n### Expected results\n\nperiodic running of these management jobs\n\n### Actual results\n\nthe jobs only run once\n\n### Additional information\n\n_No response_\n", "before_files": [{"content": "import logging\n\nfrom django.utils.timezone import now\n\nlogger = logging.getLogger('awx.main.migrations')\n\n__all__ = ['create_collection_jt', 'create_clearsessions_jt', 'create_cleartokens_jt']\n\n'''\nThese methods are called by migrations to create various system job templates\n\nCreate default system job templates if not present. Create default schedules\nonly if new system job templates were created (i.e. new database).\n'''\n\n\ndef create_clearsessions_jt(apps, schema_editor):\n\n SystemJobTemplate = apps.get_model('main', 'SystemJobTemplate')\n Schedule = apps.get_model('main', 'Schedule')\n ContentType = apps.get_model('contenttypes', 'ContentType')\n sjt_ct = ContentType.objects.get_for_model(SystemJobTemplate)\n now_dt = now()\n schedule_time = now_dt.strftime('%Y%m%dT%H%M%SZ')\n\n sjt, created = SystemJobTemplate.objects.get_or_create(\n job_type='cleanup_sessions',\n defaults=dict(\n name='Cleanup Expired Sessions',\n description='Cleans out expired browser sessions',\n polymorphic_ctype=sjt_ct,\n created=now_dt,\n modified=now_dt,\n ),\n )\n if created:\n sched = Schedule(\n name='Cleanup Expired Sessions',\n rrule='DTSTART:%s RRULE:FREQ=WEEKLY;INTERVAL=1;COUNT=1' % schedule_time,\n description='Cleans out expired browser sessions',\n enabled=True,\n created=now_dt,\n modified=now_dt,\n extra_data={},\n )\n sched.unified_job_template = sjt\n sched.save()\n\n\ndef create_cleartokens_jt(apps, schema_editor):\n\n SystemJobTemplate = apps.get_model('main', 'SystemJobTemplate')\n Schedule = apps.get_model('main', 'Schedule')\n ContentType = apps.get_model('contenttypes', 'ContentType')\n sjt_ct = ContentType.objects.get_for_model(SystemJobTemplate)\n now_dt = now()\n schedule_time = now_dt.strftime('%Y%m%dT%H%M%SZ')\n\n sjt, created = SystemJobTemplate.objects.get_or_create(\n job_type='cleanup_tokens',\n defaults=dict(\n name='Cleanup Expired OAuth 2 Tokens',\n description='Cleanup expired OAuth 2 access and refresh tokens',\n polymorphic_ctype=sjt_ct,\n created=now_dt,\n modified=now_dt,\n ),\n )\n if created:\n sched = Schedule(\n name='Cleanup Expired OAuth 2 Tokens',\n rrule='DTSTART:%s RRULE:FREQ=WEEKLY;INTERVAL=1;COUNT=1' % schedule_time,\n description='Removes expired OAuth 2 access and refresh tokens',\n enabled=True,\n created=now_dt,\n modified=now_dt,\n extra_data={},\n )\n sched.unified_job_template = sjt\n sched.save()\n", "path": "awx/main/migrations/_create_system_jobs.py"}], "after_files": [{"content": "import logging\n\nfrom django.utils.timezone import now\n\nlogger = logging.getLogger('awx.main.migrations')\n\n__all__ = ['create_collection_jt', 'create_clearsessions_jt', 'create_cleartokens_jt']\n\n'''\nThese methods are called by migrations to create various system job templates\n\nCreate default system job templates if not present. Create default schedules\nonly if new system job templates were created (i.e. new database).\n'''\n\n\ndef create_clearsessions_jt(apps, schema_editor):\n\n SystemJobTemplate = apps.get_model('main', 'SystemJobTemplate')\n Schedule = apps.get_model('main', 'Schedule')\n ContentType = apps.get_model('contenttypes', 'ContentType')\n sjt_ct = ContentType.objects.get_for_model(SystemJobTemplate)\n now_dt = now()\n schedule_time = now_dt.strftime('%Y%m%dT%H%M%SZ')\n\n sjt, created = SystemJobTemplate.objects.get_or_create(\n job_type='cleanup_sessions',\n defaults=dict(\n name='Cleanup Expired Sessions',\n description='Cleans out expired browser sessions',\n polymorphic_ctype=sjt_ct,\n created=now_dt,\n modified=now_dt,\n ),\n )\n if created:\n sched = Schedule(\n name='Cleanup Expired Sessions',\n rrule='DTSTART:%s RRULE:FREQ=WEEKLY;INTERVAL=1' % schedule_time,\n description='Cleans out expired browser sessions',\n enabled=True,\n created=now_dt,\n modified=now_dt,\n extra_data={},\n )\n sched.unified_job_template = sjt\n sched.save()\n\n\ndef create_cleartokens_jt(apps, schema_editor):\n\n SystemJobTemplate = apps.get_model('main', 'SystemJobTemplate')\n Schedule = apps.get_model('main', 'Schedule')\n ContentType = apps.get_model('contenttypes', 'ContentType')\n sjt_ct = ContentType.objects.get_for_model(SystemJobTemplate)\n now_dt = now()\n schedule_time = now_dt.strftime('%Y%m%dT%H%M%SZ')\n\n sjt, created = SystemJobTemplate.objects.get_or_create(\n job_type='cleanup_tokens',\n defaults=dict(\n name='Cleanup Expired OAuth 2 Tokens',\n description='Cleanup expired OAuth 2 access and refresh tokens',\n polymorphic_ctype=sjt_ct,\n created=now_dt,\n modified=now_dt,\n ),\n )\n if created:\n sched = Schedule(\n name='Cleanup Expired OAuth 2 Tokens',\n rrule='DTSTART:%s RRULE:FREQ=WEEKLY;INTERVAL=1' % schedule_time,\n description='Removes expired OAuth 2 access and refresh tokens',\n enabled=True,\n created=now_dt,\n modified=now_dt,\n extra_data={},\n )\n sched.unified_job_template = sjt\n sched.save()\n", "path": "awx/main/migrations/_create_system_jobs.py"}]} | 1,380 | 273 |
gh_patches_debug_8777 | rasdani/github-patches | git_diff | searx__searx-2385 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Startpage not returning any results
**Version:**
I tried it with `v0.17.0` and `820b468bfe96f693d60ce06f1e78af51f00deefc`
**Installation-Method:**
Manually according to wiki (with uwsgi)
**What happened?**
The startpage engine is not returning any results
**How To Reproduce**
Execute a query with `!sp test`
**Expected behavior**
Results from startpage
**Additional context**
I added some log lines to the startpage engine file and it seems like it gets some
response back from startpage.
Maybe startpage changed their formatting?
I didn't have a closer look on the way results are parsed yet.
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `searx/engines/startpage.py`
Content:
```
1 # Startpage (Web)
2 #
3 # @website https://startpage.com
4 # @provide-api no (nothing found)
5 #
6 # @using-api no
7 # @results HTML
8 # @stable no (HTML can change)
9 # @parse url, title, content
10 #
11 # @todo paging
12
13 from lxml import html
14 from dateutil import parser
15 from datetime import datetime, timedelta
16 import re
17 from unicodedata import normalize, combining
18 from babel import Locale
19 from babel.localedata import locale_identifiers
20 from searx.utils import extract_text, eval_xpath, match_language
21
22 # engine dependent config
23 categories = ['general']
24 # there is a mechanism to block "bot" search
25 # (probably the parameter qid), require
26 # storing of qid's between mulitble search-calls
27
28 paging = True
29 language_support = True
30 supported_languages_url = 'https://www.startpage.com/do/settings'
31
32 # search-url
33 base_url = 'https://startpage.com/'
34 search_url = base_url + 'do/search'
35
36 # specific xpath variables
37 # ads xpath //div[@id="results"]/div[@id="sponsored"]//div[@class="result"]
38 # not ads: div[@class="result"] are the direct childs of div[@id="results"]
39 results_xpath = '//div[@class="w-gl__result"]'
40 link_xpath = './/a[@class="w-gl__result-title"]'
41 content_xpath = './/p[@class="w-gl__description"]'
42
43
44 # do search-request
45 def request(query, params):
46
47 params['url'] = search_url
48 params['method'] = 'POST'
49 params['data'] = {
50 'query': query,
51 'page': params['pageno'],
52 'cat': 'web',
53 'cmd': 'process_search',
54 'engine0': 'v1all',
55 }
56
57 # set language if specified
58 if params['language'] != 'all':
59 lang_code = match_language(params['language'], supported_languages, fallback=None)
60 if lang_code:
61 language_name = supported_languages[lang_code]['alias']
62 params['data']['language'] = language_name
63 params['data']['lui'] = language_name
64
65 return params
66
67
68 # get response from search-request
69 def response(resp):
70 results = []
71
72 dom = html.fromstring(resp.text)
73
74 # parse results
75 for result in eval_xpath(dom, results_xpath):
76 links = eval_xpath(result, link_xpath)
77 if not links:
78 continue
79 link = links[0]
80 url = link.attrib.get('href')
81
82 # block google-ad url's
83 if re.match(r"^http(s|)://(www\.)?google\.[a-z]+/aclk.*$", url):
84 continue
85
86 # block startpage search url's
87 if re.match(r"^http(s|)://(www\.)?startpage\.com/do/search\?.*$", url):
88 continue
89
90 title = extract_text(link)
91
92 if eval_xpath(result, content_xpath):
93 content = extract_text(eval_xpath(result, content_xpath))
94 else:
95 content = ''
96
97 published_date = None
98
99 # check if search result starts with something like: "2 Sep 2014 ... "
100 if re.match(r"^([1-9]|[1-2][0-9]|3[0-1]) [A-Z][a-z]{2} [0-9]{4} \.\.\. ", content):
101 date_pos = content.find('...') + 4
102 date_string = content[0:date_pos - 5]
103 # fix content string
104 content = content[date_pos:]
105
106 try:
107 published_date = parser.parse(date_string, dayfirst=True)
108 except ValueError:
109 pass
110
111 # check if search result starts with something like: "5 days ago ... "
112 elif re.match(r"^[0-9]+ days? ago \.\.\. ", content):
113 date_pos = content.find('...') + 4
114 date_string = content[0:date_pos - 5]
115
116 # calculate datetime
117 published_date = datetime.now() - timedelta(days=int(re.match(r'\d+', date_string).group()))
118
119 # fix content string
120 content = content[date_pos:]
121
122 if published_date:
123 # append result
124 results.append({'url': url,
125 'title': title,
126 'content': content,
127 'publishedDate': published_date})
128 else:
129 # append result
130 results.append({'url': url,
131 'title': title,
132 'content': content})
133
134 # return results
135 return results
136
137
138 # get supported languages from their site
139 def _fetch_supported_languages(resp):
140 # startpage's language selector is a mess
141 # each option has a displayed name and a value, either of which may represent the language name
142 # in the native script, the language name in English, an English transliteration of the native name,
143 # the English name of the writing script used by the language, or occasionally something else entirely.
144
145 # this cases are so special they need to be hardcoded, a couple of them are mispellings
146 language_names = {
147 'english_uk': 'en-GB',
148 'fantizhengwen': ['zh-TW', 'zh-HK'],
149 'hangul': 'ko',
150 'malayam': 'ml',
151 'norsk': 'nb',
152 'sinhalese': 'si',
153 'sudanese': 'su'
154 }
155
156 # get the English name of every language known by babel
157 language_names.update({name.lower(): lang_code for lang_code, name in Locale('en')._data['languages'].items()})
158
159 # get the native name of every language known by babel
160 for lang_code in filter(lambda lang_code: lang_code.find('_') == -1, locale_identifiers()):
161 native_name = Locale(lang_code).get_language_name().lower()
162 # add native name exactly as it is
163 language_names[native_name] = lang_code
164
165 # add "normalized" language name (i.e. français becomes francais and español becomes espanol)
166 unaccented_name = ''.join(filter(lambda c: not combining(c), normalize('NFKD', native_name)))
167 if len(unaccented_name) == len(unaccented_name.encode()):
168 # add only if result is ascii (otherwise "normalization" didn't work)
169 language_names[unaccented_name] = lang_code
170
171 dom = html.fromstring(resp.text)
172 sp_lang_names = []
173 for option in dom.xpath('//form[@id="settings-form"]//select[@name="language"]/option'):
174 sp_lang_names.append((option.get('value'), extract_text(option).lower()))
175
176 supported_languages = {}
177 for sp_option_value, sp_option_text in sp_lang_names:
178 lang_code = language_names.get(sp_option_value) or language_names.get(sp_option_text)
179 if isinstance(lang_code, str):
180 supported_languages[lang_code] = {'alias': sp_option_value}
181 elif isinstance(lang_code, list):
182 for lc in lang_code:
183 supported_languages[lc] = {'alias': sp_option_value}
184 else:
185 print('Unknown language option in Startpage: {} ({})'.format(sp_option_value, sp_option_text))
186
187 return supported_languages
188
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/searx/engines/startpage.py b/searx/engines/startpage.py
--- a/searx/engines/startpage.py
+++ b/searx/engines/startpage.py
@@ -36,8 +36,8 @@
# specific xpath variables
# ads xpath //div[@id="results"]/div[@id="sponsored"]//div[@class="result"]
# not ads: div[@class="result"] are the direct childs of div[@id="results"]
-results_xpath = '//div[@class="w-gl__result"]'
-link_xpath = './/a[@class="w-gl__result-title"]'
+results_xpath = '//div[@class="w-gl__result__main"]'
+link_xpath = './/a[@class="w-gl__result-url result-link"]'
content_xpath = './/p[@class="w-gl__description"]'
| {"golden_diff": "diff --git a/searx/engines/startpage.py b/searx/engines/startpage.py\n--- a/searx/engines/startpage.py\n+++ b/searx/engines/startpage.py\n@@ -36,8 +36,8 @@\n # specific xpath variables\n # ads xpath //div[@id=\"results\"]/div[@id=\"sponsored\"]//div[@class=\"result\"]\n # not ads: div[@class=\"result\"] are the direct childs of div[@id=\"results\"]\n-results_xpath = '//div[@class=\"w-gl__result\"]'\n-link_xpath = './/a[@class=\"w-gl__result-title\"]'\n+results_xpath = '//div[@class=\"w-gl__result__main\"]'\n+link_xpath = './/a[@class=\"w-gl__result-url result-link\"]'\n content_xpath = './/p[@class=\"w-gl__description\"]'\n", "issue": "Startpage not returning any results\n**Version:**\r\nI tried it with `v0.17.0` and `820b468bfe96f693d60ce06f1e78af51f00deefc`\r\n\r\n**Installation-Method:** \r\nManually according to wiki (with uwsgi)\r\n\r\n**What happened?**\r\nThe startpage engine is not returning any results\r\n\r\n**How To Reproduce**\r\nExecute a query with `!sp test`\r\n\r\n**Expected behavior**\r\nResults from startpage\r\n\r\n**Additional context**\r\nI added some log lines to the startpage engine file and it seems like it gets some\r\nresponse back from startpage. \r\n\r\nMaybe startpage changed their formatting? \r\nI didn't have a closer look on the way results are parsed yet. \n", "before_files": [{"content": "# Startpage (Web)\n#\n# @website https://startpage.com\n# @provide-api no (nothing found)\n#\n# @using-api no\n# @results HTML\n# @stable no (HTML can change)\n# @parse url, title, content\n#\n# @todo paging\n\nfrom lxml import html\nfrom dateutil import parser\nfrom datetime import datetime, timedelta\nimport re\nfrom unicodedata import normalize, combining\nfrom babel import Locale\nfrom babel.localedata import locale_identifiers\nfrom searx.utils import extract_text, eval_xpath, match_language\n\n# engine dependent config\ncategories = ['general']\n# there is a mechanism to block \"bot\" search\n# (probably the parameter qid), require\n# storing of qid's between mulitble search-calls\n\npaging = True\nlanguage_support = True\nsupported_languages_url = 'https://www.startpage.com/do/settings'\n\n# search-url\nbase_url = 'https://startpage.com/'\nsearch_url = base_url + 'do/search'\n\n# specific xpath variables\n# ads xpath //div[@id=\"results\"]/div[@id=\"sponsored\"]//div[@class=\"result\"]\n# not ads: div[@class=\"result\"] are the direct childs of div[@id=\"results\"]\nresults_xpath = '//div[@class=\"w-gl__result\"]'\nlink_xpath = './/a[@class=\"w-gl__result-title\"]'\ncontent_xpath = './/p[@class=\"w-gl__description\"]'\n\n\n# do search-request\ndef request(query, params):\n\n params['url'] = search_url\n params['method'] = 'POST'\n params['data'] = {\n 'query': query,\n 'page': params['pageno'],\n 'cat': 'web',\n 'cmd': 'process_search',\n 'engine0': 'v1all',\n }\n\n # set language if specified\n if params['language'] != 'all':\n lang_code = match_language(params['language'], supported_languages, fallback=None)\n if lang_code:\n language_name = supported_languages[lang_code]['alias']\n params['data']['language'] = language_name\n params['data']['lui'] = language_name\n\n return params\n\n\n# get response from search-request\ndef response(resp):\n results = []\n\n dom = html.fromstring(resp.text)\n\n # parse results\n for result in eval_xpath(dom, results_xpath):\n links = eval_xpath(result, link_xpath)\n if not links:\n continue\n link = links[0]\n url = link.attrib.get('href')\n\n # block google-ad url's\n if re.match(r\"^http(s|)://(www\\.)?google\\.[a-z]+/aclk.*$\", url):\n continue\n\n # block startpage search url's\n if re.match(r\"^http(s|)://(www\\.)?startpage\\.com/do/search\\?.*$\", url):\n continue\n\n title = extract_text(link)\n\n if eval_xpath(result, content_xpath):\n content = extract_text(eval_xpath(result, content_xpath))\n else:\n content = ''\n\n published_date = None\n\n # check if search result starts with something like: \"2 Sep 2014 ... \"\n if re.match(r\"^([1-9]|[1-2][0-9]|3[0-1]) [A-Z][a-z]{2} [0-9]{4} \\.\\.\\. \", content):\n date_pos = content.find('...') + 4\n date_string = content[0:date_pos - 5]\n # fix content string\n content = content[date_pos:]\n\n try:\n published_date = parser.parse(date_string, dayfirst=True)\n except ValueError:\n pass\n\n # check if search result starts with something like: \"5 days ago ... \"\n elif re.match(r\"^[0-9]+ days? ago \\.\\.\\. \", content):\n date_pos = content.find('...') + 4\n date_string = content[0:date_pos - 5]\n\n # calculate datetime\n published_date = datetime.now() - timedelta(days=int(re.match(r'\\d+', date_string).group()))\n\n # fix content string\n content = content[date_pos:]\n\n if published_date:\n # append result\n results.append({'url': url,\n 'title': title,\n 'content': content,\n 'publishedDate': published_date})\n else:\n # append result\n results.append({'url': url,\n 'title': title,\n 'content': content})\n\n # return results\n return results\n\n\n# get supported languages from their site\ndef _fetch_supported_languages(resp):\n # startpage's language selector is a mess\n # each option has a displayed name and a value, either of which may represent the language name\n # in the native script, the language name in English, an English transliteration of the native name,\n # the English name of the writing script used by the language, or occasionally something else entirely.\n\n # this cases are so special they need to be hardcoded, a couple of them are mispellings\n language_names = {\n 'english_uk': 'en-GB',\n 'fantizhengwen': ['zh-TW', 'zh-HK'],\n 'hangul': 'ko',\n 'malayam': 'ml',\n 'norsk': 'nb',\n 'sinhalese': 'si',\n 'sudanese': 'su'\n }\n\n # get the English name of every language known by babel\n language_names.update({name.lower(): lang_code for lang_code, name in Locale('en')._data['languages'].items()})\n\n # get the native name of every language known by babel\n for lang_code in filter(lambda lang_code: lang_code.find('_') == -1, locale_identifiers()):\n native_name = Locale(lang_code).get_language_name().lower()\n # add native name exactly as it is\n language_names[native_name] = lang_code\n\n # add \"normalized\" language name (i.e. fran\u00e7ais becomes francais and espa\u00f1ol becomes espanol)\n unaccented_name = ''.join(filter(lambda c: not combining(c), normalize('NFKD', native_name)))\n if len(unaccented_name) == len(unaccented_name.encode()):\n # add only if result is ascii (otherwise \"normalization\" didn't work)\n language_names[unaccented_name] = lang_code\n\n dom = html.fromstring(resp.text)\n sp_lang_names = []\n for option in dom.xpath('//form[@id=\"settings-form\"]//select[@name=\"language\"]/option'):\n sp_lang_names.append((option.get('value'), extract_text(option).lower()))\n\n supported_languages = {}\n for sp_option_value, sp_option_text in sp_lang_names:\n lang_code = language_names.get(sp_option_value) or language_names.get(sp_option_text)\n if isinstance(lang_code, str):\n supported_languages[lang_code] = {'alias': sp_option_value}\n elif isinstance(lang_code, list):\n for lc in lang_code:\n supported_languages[lc] = {'alias': sp_option_value}\n else:\n print('Unknown language option in Startpage: {} ({})'.format(sp_option_value, sp_option_text))\n\n return supported_languages\n", "path": "searx/engines/startpage.py"}], "after_files": [{"content": "# Startpage (Web)\n#\n# @website https://startpage.com\n# @provide-api no (nothing found)\n#\n# @using-api no\n# @results HTML\n# @stable no (HTML can change)\n# @parse url, title, content\n#\n# @todo paging\n\nfrom lxml import html\nfrom dateutil import parser\nfrom datetime import datetime, timedelta\nimport re\nfrom unicodedata import normalize, combining\nfrom babel import Locale\nfrom babel.localedata import locale_identifiers\nfrom searx.utils import extract_text, eval_xpath, match_language\n\n# engine dependent config\ncategories = ['general']\n# there is a mechanism to block \"bot\" search\n# (probably the parameter qid), require\n# storing of qid's between mulitble search-calls\n\npaging = True\nlanguage_support = True\nsupported_languages_url = 'https://www.startpage.com/do/settings'\n\n# search-url\nbase_url = 'https://startpage.com/'\nsearch_url = base_url + 'do/search'\n\n# specific xpath variables\n# ads xpath //div[@id=\"results\"]/div[@id=\"sponsored\"]//div[@class=\"result\"]\n# not ads: div[@class=\"result\"] are the direct childs of div[@id=\"results\"]\nresults_xpath = '//div[@class=\"w-gl__result__main\"]'\nlink_xpath = './/a[@class=\"w-gl__result-url result-link\"]'\ncontent_xpath = './/p[@class=\"w-gl__description\"]'\n\n\n# do search-request\ndef request(query, params):\n\n params['url'] = search_url\n params['method'] = 'POST'\n params['data'] = {\n 'query': query,\n 'page': params['pageno'],\n 'cat': 'web',\n 'cmd': 'process_search',\n 'engine0': 'v1all',\n }\n\n # set language if specified\n if params['language'] != 'all':\n lang_code = match_language(params['language'], supported_languages, fallback=None)\n if lang_code:\n language_name = supported_languages[lang_code]['alias']\n params['data']['language'] = language_name\n params['data']['lui'] = language_name\n\n return params\n\n\n# get response from search-request\ndef response(resp):\n results = []\n\n dom = html.fromstring(resp.text)\n\n # parse results\n for result in eval_xpath(dom, results_xpath):\n links = eval_xpath(result, link_xpath)\n if not links:\n continue\n link = links[0]\n url = link.attrib.get('href')\n\n # block google-ad url's\n if re.match(r\"^http(s|)://(www\\.)?google\\.[a-z]+/aclk.*$\", url):\n continue\n\n # block startpage search url's\n if re.match(r\"^http(s|)://(www\\.)?startpage\\.com/do/search\\?.*$\", url):\n continue\n\n title = extract_text(link)\n\n if eval_xpath(result, content_xpath):\n content = extract_text(eval_xpath(result, content_xpath))\n else:\n content = ''\n\n published_date = None\n\n # check if search result starts with something like: \"2 Sep 2014 ... \"\n if re.match(r\"^([1-9]|[1-2][0-9]|3[0-1]) [A-Z][a-z]{2} [0-9]{4} \\.\\.\\. \", content):\n date_pos = content.find('...') + 4\n date_string = content[0:date_pos - 5]\n # fix content string\n content = content[date_pos:]\n\n try:\n published_date = parser.parse(date_string, dayfirst=True)\n except ValueError:\n pass\n\n # check if search result starts with something like: \"5 days ago ... \"\n elif re.match(r\"^[0-9]+ days? ago \\.\\.\\. \", content):\n date_pos = content.find('...') + 4\n date_string = content[0:date_pos - 5]\n\n # calculate datetime\n published_date = datetime.now() - timedelta(days=int(re.match(r'\\d+', date_string).group()))\n\n # fix content string\n content = content[date_pos:]\n\n if published_date:\n # append result\n results.append({'url': url,\n 'title': title,\n 'content': content,\n 'publishedDate': published_date})\n else:\n # append result\n results.append({'url': url,\n 'title': title,\n 'content': content})\n\n # return results\n return results\n\n\n# get supported languages from their site\ndef _fetch_supported_languages(resp):\n # startpage's language selector is a mess\n # each option has a displayed name and a value, either of which may represent the language name\n # in the native script, the language name in English, an English transliteration of the native name,\n # the English name of the writing script used by the language, or occasionally something else entirely.\n\n # this cases are so special they need to be hardcoded, a couple of them are mispellings\n language_names = {\n 'english_uk': 'en-GB',\n 'fantizhengwen': ['zh-TW', 'zh-HK'],\n 'hangul': 'ko',\n 'malayam': 'ml',\n 'norsk': 'nb',\n 'sinhalese': 'si',\n 'sudanese': 'su'\n }\n\n # get the English name of every language known by babel\n language_names.update({name.lower(): lang_code for lang_code, name in Locale('en')._data['languages'].items()})\n\n # get the native name of every language known by babel\n for lang_code in filter(lambda lang_code: lang_code.find('_') == -1, locale_identifiers()):\n native_name = Locale(lang_code).get_language_name().lower()\n # add native name exactly as it is\n language_names[native_name] = lang_code\n\n # add \"normalized\" language name (i.e. fran\u00e7ais becomes francais and espa\u00f1ol becomes espanol)\n unaccented_name = ''.join(filter(lambda c: not combining(c), normalize('NFKD', native_name)))\n if len(unaccented_name) == len(unaccented_name.encode()):\n # add only if result is ascii (otherwise \"normalization\" didn't work)\n language_names[unaccented_name] = lang_code\n\n dom = html.fromstring(resp.text)\n sp_lang_names = []\n for option in dom.xpath('//form[@id=\"settings-form\"]//select[@name=\"language\"]/option'):\n sp_lang_names.append((option.get('value'), extract_text(option).lower()))\n\n supported_languages = {}\n for sp_option_value, sp_option_text in sp_lang_names:\n lang_code = language_names.get(sp_option_value) or language_names.get(sp_option_text)\n if isinstance(lang_code, str):\n supported_languages[lang_code] = {'alias': sp_option_value}\n elif isinstance(lang_code, list):\n for lc in lang_code:\n supported_languages[lc] = {'alias': sp_option_value}\n else:\n print('Unknown language option in Startpage: {} ({})'.format(sp_option_value, sp_option_text))\n\n return supported_languages\n", "path": "searx/engines/startpage.py"}]} | 2,474 | 191 |
gh_patches_debug_20553 | rasdani/github-patches | git_diff | paperless-ngx__paperless-ngx-1605 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
[BUG] permission error if no consume share is mapped
### Description
starting with Paperless-ngx 1.8.0, Paperless on docker does not start when no consume share is mapped:
`
SystemCheckError: System check identified some issues:
ERRORS:
?: PAPERLESS_CONSUMPTION_DIR is not writeable
HINT: Set the permissions of
drwxr-xr-x /usr/src/paperless/consume
to be writeable by the user running the Paperless services
`
I've some containers running without a mapped share, as I don't need a consumption folder. Was no issue in previous versions; I assume the permissions of the folder in the docker container has changed.
### Steps to reproduce
1. create docker-compose without mapped consumption folder
2. look at the error message
### Webserver logs
```bash
SystemCheckError: System check identified some issues:
ERRORS:
?: PAPERLESS_CONSUMPTION_DIR is not writeable
HINT: Set the permissions of
drwxr-xr-x /usr/src/paperless/consume
to be writeable by the user running the Paperless services
```
### Paperless-ngx version
1.8,0
### Host OS
docker
### Installation method
Docker - official image
### Browser
_No response_
### Configuration changes
_No response_
### Other
_No response_
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `src/paperless/checks.py`
Content:
```
1 import os
2 import shutil
3 import stat
4
5 from django.conf import settings
6 from django.core.checks import Error
7 from django.core.checks import register
8 from django.core.checks import Warning
9
10 exists_message = "{} is set but doesn't exist."
11 exists_hint = "Create a directory at {}"
12 writeable_message = "{} is not writeable"
13 writeable_hint = (
14 "Set the permissions of {} to be writeable by the user running the "
15 "Paperless services"
16 )
17
18
19 def path_check(var, directory):
20 messages = []
21 if directory:
22 if not os.path.isdir(directory):
23 messages.append(
24 Error(exists_message.format(var), exists_hint.format(directory)),
25 )
26 else:
27 test_file = os.path.join(
28 directory,
29 f"__paperless_write_test_{os.getpid()}__",
30 )
31 try:
32 with open(test_file, "w"):
33 pass
34 except PermissionError:
35 messages.append(
36 Error(
37 writeable_message.format(var),
38 writeable_hint.format(
39 f"\n{stat.filemode(os.stat(directory).st_mode)} "
40 f"{directory}\n",
41 ),
42 ),
43 )
44 finally:
45 if os.path.isfile(test_file):
46 os.remove(test_file)
47
48 return messages
49
50
51 @register()
52 def paths_check(app_configs, **kwargs):
53 """
54 Check the various paths for existence, readability and writeability
55 """
56
57 return (
58 path_check("PAPERLESS_DATA_DIR", settings.DATA_DIR)
59 + path_check("PAPERLESS_TRASH_DIR", settings.TRASH_DIR)
60 + path_check("PAPERLESS_MEDIA_ROOT", settings.MEDIA_ROOT)
61 + path_check("PAPERLESS_CONSUMPTION_DIR", settings.CONSUMPTION_DIR)
62 )
63
64
65 @register()
66 def binaries_check(app_configs, **kwargs):
67 """
68 Paperless requires the existence of a few binaries, so we do some checks
69 for those here.
70 """
71
72 error = "Paperless can't find {}. Without it, consumption is impossible."
73 hint = "Either it's not in your ${PATH} or it's not installed."
74
75 binaries = (settings.CONVERT_BINARY, "tesseract")
76
77 check_messages = []
78 for binary in binaries:
79 if shutil.which(binary) is None:
80 check_messages.append(Warning(error.format(binary), hint))
81
82 return check_messages
83
84
85 @register()
86 def debug_mode_check(app_configs, **kwargs):
87 if settings.DEBUG:
88 return [
89 Warning(
90 "DEBUG mode is enabled. Disable Debug mode. This is a serious "
91 "security issue, since it puts security overides in place which "
92 "are meant to be only used during development. This "
93 "also means that paperless will tell anyone various "
94 "debugging information when something goes wrong.",
95 ),
96 ]
97 else:
98 return []
99
100
101 @register()
102 def settings_values_check(app_configs, **kwargs):
103 """
104 Validates at least some of the user provided settings
105 """
106
107 def _ocrmypdf_settings_check():
108 """
109 Validates some of the arguments which will be provided to ocrmypdf
110 against the valid options. Use "ocrmypdf --help" to see the valid
111 inputs
112 """
113 msgs = []
114 if settings.OCR_OUTPUT_TYPE not in {
115 "pdfa",
116 "pdf",
117 "pdfa-1",
118 "pdfa-2",
119 "pdfa-3",
120 }:
121 msgs.append(
122 Error(f'OCR output type "{settings.OCR_OUTPUT_TYPE}" is not valid'),
123 )
124
125 if settings.OCR_MODE not in {"force", "skip", "redo_ocr"}:
126 msgs.append(Error(f'OCR output mode "{settings.OCR_MODE}" is not valid'))
127
128 if settings.OCR_CLEAN not in {"clean", "clean_final"}:
129 msgs.append(Error(f'OCR clean mode "{settings.OCR_CLEAN}" is not valid'))
130 return msgs
131
132 def _timezone_validate():
133 """
134 Validates the user provided timezone is a valid timezone
135 """
136 try:
137 import zoneinfo
138 except ImportError: # pragma: nocover
139 import backports.zoneinfo as zoneinfo
140 msgs = []
141 if settings.TIME_ZONE not in zoneinfo.available_timezones():
142 msgs.append(
143 Error(f'Timezone "{settings.TIME_ZONE}" is not a valid timezone'),
144 )
145 return msgs
146
147 return _ocrmypdf_settings_check() + _timezone_validate()
148
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/src/paperless/checks.py b/src/paperless/checks.py
--- a/src/paperless/checks.py
+++ b/src/paperless/checks.py
@@ -1,4 +1,6 @@
+import grp
import os
+import pwd
import shutil
import stat
@@ -32,12 +34,15 @@
with open(test_file, "w"):
pass
except PermissionError:
+ dir_stat = os.stat(directory)
+ dir_mode = stat.filemode(dir_stat.st_mode)
+ dir_owner = pwd.getpwuid(dir_stat.st_uid).pw_name
+ dir_group = grp.getgrgid(dir_stat.st_gid).gr_name
messages.append(
Error(
writeable_message.format(var),
writeable_hint.format(
- f"\n{stat.filemode(os.stat(directory).st_mode)} "
- f"{directory}\n",
+ f"\n{dir_mode} {dir_owner} {dir_group} " f"{directory}\n",
),
),
)
| {"golden_diff": "diff --git a/src/paperless/checks.py b/src/paperless/checks.py\n--- a/src/paperless/checks.py\n+++ b/src/paperless/checks.py\n@@ -1,4 +1,6 @@\n+import grp\n import os\n+import pwd\n import shutil\n import stat\n \n@@ -32,12 +34,15 @@\n with open(test_file, \"w\"):\n pass\n except PermissionError:\n+ dir_stat = os.stat(directory)\n+ dir_mode = stat.filemode(dir_stat.st_mode)\n+ dir_owner = pwd.getpwuid(dir_stat.st_uid).pw_name\n+ dir_group = grp.getgrgid(dir_stat.st_gid).gr_name\n messages.append(\n Error(\n writeable_message.format(var),\n writeable_hint.format(\n- f\"\\n{stat.filemode(os.stat(directory).st_mode)} \"\n- f\"{directory}\\n\",\n+ f\"\\n{dir_mode} {dir_owner} {dir_group} \" f\"{directory}\\n\",\n ),\n ),\n )\n", "issue": "[BUG] permission error if no consume share is mapped\n### Description\r\n\r\nstarting with Paperless-ngx 1.8.0, Paperless on docker does not start when no consume share is mapped:\r\n\r\n`\r\n SystemCheckError: System check identified some issues:\r\n ERRORS:\r\n ?: PAPERLESS_CONSUMPTION_DIR is not writeable\r\n\tHINT: Set the permissions of \r\n drwxr-xr-x /usr/src/paperless/consume\r\n to be writeable by the user running the Paperless services\r\n\r\n`\r\n\r\nI've some containers running without a mapped share, as I don't need a consumption folder. Was no issue in previous versions; I assume the permissions of the folder in the docker container has changed.\r\n\r\n### Steps to reproduce\r\n\r\n1. create docker-compose without mapped consumption folder\r\n2. look at the error message\r\n\r\n### Webserver logs\r\n\r\n```bash\r\nSystemCheckError: System check identified some issues:\r\nERRORS:\r\n?: PAPERLESS_CONSUMPTION_DIR is not writeable\r\n\tHINT: Set the permissions of \r\ndrwxr-xr-x /usr/src/paperless/consume\r\n to be writeable by the user running the Paperless services\r\n```\r\n\r\n\r\n### Paperless-ngx version\r\n\r\n1.8,0\r\n\r\n### Host OS\r\n\r\ndocker\r\n\r\n### Installation method\r\n\r\nDocker - official image\r\n\r\n### Browser\r\n\r\n_No response_\r\n\r\n### Configuration changes\r\n\r\n_No response_\r\n\r\n### Other\r\n\r\n_No response_\n", "before_files": [{"content": "import os\nimport shutil\nimport stat\n\nfrom django.conf import settings\nfrom django.core.checks import Error\nfrom django.core.checks import register\nfrom django.core.checks import Warning\n\nexists_message = \"{} is set but doesn't exist.\"\nexists_hint = \"Create a directory at {}\"\nwriteable_message = \"{} is not writeable\"\nwriteable_hint = (\n \"Set the permissions of {} to be writeable by the user running the \"\n \"Paperless services\"\n)\n\n\ndef path_check(var, directory):\n messages = []\n if directory:\n if not os.path.isdir(directory):\n messages.append(\n Error(exists_message.format(var), exists_hint.format(directory)),\n )\n else:\n test_file = os.path.join(\n directory,\n f\"__paperless_write_test_{os.getpid()}__\",\n )\n try:\n with open(test_file, \"w\"):\n pass\n except PermissionError:\n messages.append(\n Error(\n writeable_message.format(var),\n writeable_hint.format(\n f\"\\n{stat.filemode(os.stat(directory).st_mode)} \"\n f\"{directory}\\n\",\n ),\n ),\n )\n finally:\n if os.path.isfile(test_file):\n os.remove(test_file)\n\n return messages\n\n\n@register()\ndef paths_check(app_configs, **kwargs):\n \"\"\"\n Check the various paths for existence, readability and writeability\n \"\"\"\n\n return (\n path_check(\"PAPERLESS_DATA_DIR\", settings.DATA_DIR)\n + path_check(\"PAPERLESS_TRASH_DIR\", settings.TRASH_DIR)\n + path_check(\"PAPERLESS_MEDIA_ROOT\", settings.MEDIA_ROOT)\n + path_check(\"PAPERLESS_CONSUMPTION_DIR\", settings.CONSUMPTION_DIR)\n )\n\n\n@register()\ndef binaries_check(app_configs, **kwargs):\n \"\"\"\n Paperless requires the existence of a few binaries, so we do some checks\n for those here.\n \"\"\"\n\n error = \"Paperless can't find {}. Without it, consumption is impossible.\"\n hint = \"Either it's not in your ${PATH} or it's not installed.\"\n\n binaries = (settings.CONVERT_BINARY, \"tesseract\")\n\n check_messages = []\n for binary in binaries:\n if shutil.which(binary) is None:\n check_messages.append(Warning(error.format(binary), hint))\n\n return check_messages\n\n\n@register()\ndef debug_mode_check(app_configs, **kwargs):\n if settings.DEBUG:\n return [\n Warning(\n \"DEBUG mode is enabled. Disable Debug mode. This is a serious \"\n \"security issue, since it puts security overides in place which \"\n \"are meant to be only used during development. This \"\n \"also means that paperless will tell anyone various \"\n \"debugging information when something goes wrong.\",\n ),\n ]\n else:\n return []\n\n\n@register()\ndef settings_values_check(app_configs, **kwargs):\n \"\"\"\n Validates at least some of the user provided settings\n \"\"\"\n\n def _ocrmypdf_settings_check():\n \"\"\"\n Validates some of the arguments which will be provided to ocrmypdf\n against the valid options. Use \"ocrmypdf --help\" to see the valid\n inputs\n \"\"\"\n msgs = []\n if settings.OCR_OUTPUT_TYPE not in {\n \"pdfa\",\n \"pdf\",\n \"pdfa-1\",\n \"pdfa-2\",\n \"pdfa-3\",\n }:\n msgs.append(\n Error(f'OCR output type \"{settings.OCR_OUTPUT_TYPE}\" is not valid'),\n )\n\n if settings.OCR_MODE not in {\"force\", \"skip\", \"redo_ocr\"}:\n msgs.append(Error(f'OCR output mode \"{settings.OCR_MODE}\" is not valid'))\n\n if settings.OCR_CLEAN not in {\"clean\", \"clean_final\"}:\n msgs.append(Error(f'OCR clean mode \"{settings.OCR_CLEAN}\" is not valid'))\n return msgs\n\n def _timezone_validate():\n \"\"\"\n Validates the user provided timezone is a valid timezone\n \"\"\"\n try:\n import zoneinfo\n except ImportError: # pragma: nocover\n import backports.zoneinfo as zoneinfo\n msgs = []\n if settings.TIME_ZONE not in zoneinfo.available_timezones():\n msgs.append(\n Error(f'Timezone \"{settings.TIME_ZONE}\" is not a valid timezone'),\n )\n return msgs\n\n return _ocrmypdf_settings_check() + _timezone_validate()\n", "path": "src/paperless/checks.py"}], "after_files": [{"content": "import grp\nimport os\nimport pwd\nimport shutil\nimport stat\n\nfrom django.conf import settings\nfrom django.core.checks import Error\nfrom django.core.checks import register\nfrom django.core.checks import Warning\n\nexists_message = \"{} is set but doesn't exist.\"\nexists_hint = \"Create a directory at {}\"\nwriteable_message = \"{} is not writeable\"\nwriteable_hint = (\n \"Set the permissions of {} to be writeable by the user running the \"\n \"Paperless services\"\n)\n\n\ndef path_check(var, directory):\n messages = []\n if directory:\n if not os.path.isdir(directory):\n messages.append(\n Error(exists_message.format(var), exists_hint.format(directory)),\n )\n else:\n test_file = os.path.join(\n directory,\n f\"__paperless_write_test_{os.getpid()}__\",\n )\n try:\n with open(test_file, \"w\"):\n pass\n except PermissionError:\n dir_stat = os.stat(directory)\n dir_mode = stat.filemode(dir_stat.st_mode)\n dir_owner = pwd.getpwuid(dir_stat.st_uid).pw_name\n dir_group = grp.getgrgid(dir_stat.st_gid).gr_name\n messages.append(\n Error(\n writeable_message.format(var),\n writeable_hint.format(\n f\"\\n{dir_mode} {dir_owner} {dir_group} \" f\"{directory}\\n\",\n ),\n ),\n )\n finally:\n if os.path.isfile(test_file):\n os.remove(test_file)\n\n return messages\n\n\n@register()\ndef paths_check(app_configs, **kwargs):\n \"\"\"\n Check the various paths for existence, readability and writeability\n \"\"\"\n\n return (\n path_check(\"PAPERLESS_DATA_DIR\", settings.DATA_DIR)\n + path_check(\"PAPERLESS_TRASH_DIR\", settings.TRASH_DIR)\n + path_check(\"PAPERLESS_MEDIA_ROOT\", settings.MEDIA_ROOT)\n + path_check(\"PAPERLESS_CONSUMPTION_DIR\", settings.CONSUMPTION_DIR)\n )\n\n\n@register()\ndef binaries_check(app_configs, **kwargs):\n \"\"\"\n Paperless requires the existence of a few binaries, so we do some checks\n for those here.\n \"\"\"\n\n error = \"Paperless can't find {}. Without it, consumption is impossible.\"\n hint = \"Either it's not in your ${PATH} or it's not installed.\"\n\n binaries = (settings.CONVERT_BINARY, \"tesseract\")\n\n check_messages = []\n for binary in binaries:\n if shutil.which(binary) is None:\n check_messages.append(Warning(error.format(binary), hint))\n\n return check_messages\n\n\n@register()\ndef debug_mode_check(app_configs, **kwargs):\n if settings.DEBUG:\n return [\n Warning(\n \"DEBUG mode is enabled. Disable Debug mode. This is a serious \"\n \"security issue, since it puts security overides in place which \"\n \"are meant to be only used during development. This \"\n \"also means that paperless will tell anyone various \"\n \"debugging information when something goes wrong.\",\n ),\n ]\n else:\n return []\n\n\n@register()\ndef settings_values_check(app_configs, **kwargs):\n \"\"\"\n Validates at least some of the user provided settings\n \"\"\"\n\n def _ocrmypdf_settings_check():\n \"\"\"\n Validates some of the arguments which will be provided to ocrmypdf\n against the valid options. Use \"ocrmypdf --help\" to see the valid\n inputs\n \"\"\"\n msgs = []\n if settings.OCR_OUTPUT_TYPE not in {\n \"pdfa\",\n \"pdf\",\n \"pdfa-1\",\n \"pdfa-2\",\n \"pdfa-3\",\n }:\n msgs.append(\n Error(f'OCR output type \"{settings.OCR_OUTPUT_TYPE}\" is not valid'),\n )\n\n if settings.OCR_MODE not in {\"force\", \"skip\", \"redo_ocr\"}:\n msgs.append(Error(f'OCR output mode \"{settings.OCR_MODE}\" is not valid'))\n\n if settings.OCR_CLEAN not in {\"clean\", \"clean_final\"}:\n msgs.append(Error(f'OCR clean mode \"{settings.OCR_CLEAN}\" is not valid'))\n return msgs\n\n def _timezone_validate():\n \"\"\"\n Validates the user provided timezone is a valid timezone\n \"\"\"\n try:\n import zoneinfo\n except ImportError: # pragma: nocover\n import backports.zoneinfo as zoneinfo\n msgs = []\n if settings.TIME_ZONE not in zoneinfo.available_timezones():\n msgs.append(\n Error(f'Timezone \"{settings.TIME_ZONE}\" is not a valid timezone'),\n )\n return msgs\n\n return _ocrmypdf_settings_check() + _timezone_validate()\n", "path": "src/paperless/checks.py"}]} | 1,837 | 229 |
gh_patches_debug_3959 | rasdani/github-patches | git_diff | great-expectations__great_expectations-5468 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Use cleaner solution for non-truncating division in python 2
Prefer `from __future__ import division` to `1.*x/y`
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `contrib/capitalone_dataprofiler_expectations/capitalone_dataprofiler_expectations/expectations/__init__.py`
Content:
```
1 # Make sure to include any Expectations your want exported below!
2
3 from .expect_column_values_confidence_for_data_label_to_be_greater_than_or_equal_to_threshold import (
4 ExpectColumnValuesConfidenceForDataLabelToBeGreaterThanOrEqualToThreshold,
5 )
6 from .expect_column_values_confidence_for_data_label_to_be_less_than_or_equal_to_threshold import (
7 ExpectColumnValuesConfidenceForDataLabelToBeLessThanOrEqualToThreshold,
8 )
9 from .expect_column_values_to_be_equal_to_or_greater_than_profile_min import (
10 ExpectColumnValuesToBeEqualToOrGreaterThanProfileMin,
11 )
12 from .expect_column_values_to_be_equal_to_or_less_than_profile_max import (
13 ExpectColumnValuesToBeEqualToOrLessThanProfileMax,
14 )
15 from .expect_column_values_to_be_probabilistically_greater_than_or_equal_to_threshold import (
16 ExpectColumnValuesToBeProbabilisticallyGreaterThanOrEqualToThreshold,
17 )
18 from .expect_profile_numeric_columns_diff_between_threshold_range import (
19 ExpectProfileNumericColumnsDiffBetweenThresholdRange
20 )
21
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/contrib/capitalone_dataprofiler_expectations/capitalone_dataprofiler_expectations/expectations/__init__.py b/contrib/capitalone_dataprofiler_expectations/capitalone_dataprofiler_expectations/expectations/__init__.py
--- a/contrib/capitalone_dataprofiler_expectations/capitalone_dataprofiler_expectations/expectations/__init__.py
+++ b/contrib/capitalone_dataprofiler_expectations/capitalone_dataprofiler_expectations/expectations/__init__.py
@@ -16,5 +16,5 @@
ExpectColumnValuesToBeProbabilisticallyGreaterThanOrEqualToThreshold,
)
from .expect_profile_numeric_columns_diff_between_threshold_range import (
- ExpectProfileNumericColumnsDiffBetweenThresholdRange
+ ExpectProfileNumericColumnsDiffBetweenThresholdRange,
)
| {"golden_diff": "diff --git a/contrib/capitalone_dataprofiler_expectations/capitalone_dataprofiler_expectations/expectations/__init__.py b/contrib/capitalone_dataprofiler_expectations/capitalone_dataprofiler_expectations/expectations/__init__.py\n--- a/contrib/capitalone_dataprofiler_expectations/capitalone_dataprofiler_expectations/expectations/__init__.py\n+++ b/contrib/capitalone_dataprofiler_expectations/capitalone_dataprofiler_expectations/expectations/__init__.py\n@@ -16,5 +16,5 @@\n ExpectColumnValuesToBeProbabilisticallyGreaterThanOrEqualToThreshold,\n )\n from .expect_profile_numeric_columns_diff_between_threshold_range import (\n- ExpectProfileNumericColumnsDiffBetweenThresholdRange\n+ ExpectProfileNumericColumnsDiffBetweenThresholdRange,\n )\n", "issue": "Use cleaner solution for non-truncating division in python 2\nPrefer `from __future__ import division` to `1.*x/y`\n", "before_files": [{"content": "# Make sure to include any Expectations your want exported below!\n\nfrom .expect_column_values_confidence_for_data_label_to_be_greater_than_or_equal_to_threshold import (\n ExpectColumnValuesConfidenceForDataLabelToBeGreaterThanOrEqualToThreshold,\n)\nfrom .expect_column_values_confidence_for_data_label_to_be_less_than_or_equal_to_threshold import (\n ExpectColumnValuesConfidenceForDataLabelToBeLessThanOrEqualToThreshold,\n)\nfrom .expect_column_values_to_be_equal_to_or_greater_than_profile_min import (\n ExpectColumnValuesToBeEqualToOrGreaterThanProfileMin,\n)\nfrom .expect_column_values_to_be_equal_to_or_less_than_profile_max import (\n ExpectColumnValuesToBeEqualToOrLessThanProfileMax,\n)\nfrom .expect_column_values_to_be_probabilistically_greater_than_or_equal_to_threshold import (\n ExpectColumnValuesToBeProbabilisticallyGreaterThanOrEqualToThreshold,\n)\nfrom .expect_profile_numeric_columns_diff_between_threshold_range import (\n ExpectProfileNumericColumnsDiffBetweenThresholdRange\n)\n", "path": "contrib/capitalone_dataprofiler_expectations/capitalone_dataprofiler_expectations/expectations/__init__.py"}], "after_files": [{"content": "# Make sure to include any Expectations your want exported below!\n\nfrom .expect_column_values_confidence_for_data_label_to_be_greater_than_or_equal_to_threshold import (\n ExpectColumnValuesConfidenceForDataLabelToBeGreaterThanOrEqualToThreshold,\n)\nfrom .expect_column_values_confidence_for_data_label_to_be_less_than_or_equal_to_threshold import (\n ExpectColumnValuesConfidenceForDataLabelToBeLessThanOrEqualToThreshold,\n)\nfrom .expect_column_values_to_be_equal_to_or_greater_than_profile_min import (\n ExpectColumnValuesToBeEqualToOrGreaterThanProfileMin,\n)\nfrom .expect_column_values_to_be_equal_to_or_less_than_profile_max import (\n ExpectColumnValuesToBeEqualToOrLessThanProfileMax,\n)\nfrom .expect_column_values_to_be_probabilistically_greater_than_or_equal_to_threshold import (\n ExpectColumnValuesToBeProbabilisticallyGreaterThanOrEqualToThreshold,\n)\nfrom .expect_profile_numeric_columns_diff_between_threshold_range import (\n ExpectProfileNumericColumnsDiffBetweenThresholdRange,\n)\n", "path": "contrib/capitalone_dataprofiler_expectations/capitalone_dataprofiler_expectations/expectations/__init__.py"}]} | 544 | 181 |
gh_patches_debug_25951 | rasdani/github-patches | git_diff | napari__napari-6475 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Weird docs build error on vortex.py example
### 🐛 Bug Report
It looks like we are missing the gallery example for [vortex.py](https://github.com/napari/napari/blob/1b510bab020aae485565f000ddab842ab17ef608/examples/vortex.py). That file runs fine locally and *I think* ~~it ran fine in the PR~~ (**Edit:** nope, it was [broken](https://output.circle-artifacts.com/output/job/1b750fb4-4df5-462b-84ca-cdddeede41ff/artifacts/0/napari-docs/docs/_build/gallery.html#gallery), but gallery errors aren't errors. I don't know what the right answer is there but either we should turn them into errors or we should turn them into errors *when the contribution is a gallery example*?) But the error suggests some form of race condition during addition of the layer, which makes me think this is an async issue. Here's the error message from [this build](https://github.com/napari/docs/actions/runs/6658042739/job/18094063805#step:9:238):
```pytb
Downloading file 'data/pivchallenge-B-B001_1.tif' from 'https://gitlab.com/scikit-image/data/-/raw/2cdc5ce89b334d28f06a58c9f0ca21aa6992a5ba/pivchallenge/B/B001_1.tif' to '/home/runner/.cache/scikit-image/0.22.0'.
Downloading file 'data/pivchallenge-B-B001_2.tif' from 'https://gitlab.com/scikit-image/data/-/raw/2cdc5ce89b334d28f06a58c9f0ca21aa6992a5ba/pivchallenge/B/B001_2.tif' to '/home/runner/.cache/scikit-image/0.22.0'.
WARNING: /home/runner/work/docs/docs/docs/examples/vortex.py failed to execute correctly: Traceback (most recent call last):
File "/home/runner/work/docs/docs/docs/examples/vortex.py", line 59, in <module>
flow_layer = viewer.add_vectors(
File "/opt/hostedtoolcache/Python/3.10.13/x64/lib/python3.10/site-packages/napari/components/viewer_model.py", line 5, in add_vectors
import os
File "/opt/hostedtoolcache/Python/3.10.13/x64/lib/python3.10/_collections_abc.py", line 1128, in append
self.insert(len(self), value)
File "/opt/hostedtoolcache/Python/3.10.13/x64/lib/python3.10/site-packages/napari/components/layerlist.py", line 194, in insert
super().insert(index, new_layer)
File "/opt/hostedtoolcache/Python/3.10.13/x64/lib/python3.10/site-packages/napari/utils/events/containers/_selectable_list.py", line 71, in insert
self.selection.active = value
File "/opt/hostedtoolcache/Python/3.10.13/x64/lib/python3.10/site-packages/napari/utils/events/containers/_selection.py", line 108, in active
self.events.active(value=value)
File "/opt/hostedtoolcache/Python/3.10.13/x64/lib/python3.10/site-packages/napari/utils/events/event.py", line 771, in __call__
self._invoke_callback(cb, event if pass_event else None)
File "/opt/hostedtoolcache/Python/3.10.13/x64/lib/python3.10/site-packages/napari/utils/events/event.py", line 809, in _invoke_callback
_handle_exception(
File "/opt/hostedtoolcache/Python/3.10.13/x64/lib/python3.10/site-packages/napari/utils/events/event.py", line 796, in _invoke_callback
cb(event)
File "/opt/hostedtoolcache/Python/3.10.13/x64/lib/python3.10/site-packages/napari/_qt/layer_controls/qt_layer_controls_container.py", line 130, in _display
controls = self.widgets[layer]
KeyError: <Vectors layer 'optical flow' at 0x7f6f55be4be0>
```
### 💡 Steps to Reproduce
I don't have a working docs build at the moment so I don't know whether this reproduces in local docs builds, but the example runs fine locally. So this is either a problem with sphinx gallery or with sphinx gallery on CI.
### 💡 Expected Behavior
Example should run fine on GHA.
### 🌎 Environment
napari main on CI 😬
(Note: should we echo `napari --info` on all our CI?)
### 💡 Additional Context
_No response_
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `examples/vortex.py`
Content:
```
1 """Visualizing optical flow in napari.
2
3 Adapted from the scikit-image gallery [1]_.
4
5 In napari, we can show the flowing vortex as an additional dimension in the
6 image, visible by moving the slider.
7
8 .. tags:: visualization-advanced, layers
9
10 .. [1] https://scikit-image.org/docs/stable/auto_examples/registration/plot_opticalflow.html
11 """
12 import numpy as np
13 from skimage.data import vortex
14 from skimage.registration import optical_flow_ilk
15
16 import napari
17
18 #######################################################################
19 # First, we load the vortex image as a 3D array. (time, row, column)
20
21 vortex_im = np.asarray(vortex())
22
23 #######################################################################
24 # We compute the optical flow using scikit-image. (Note: as of
25 # scikit-image 0.21, there seems to be a transposition of the image in
26 # the output, which we account for later.)
27
28 u, v = optical_flow_ilk(vortex_im[0], vortex_im[1], radius=15)
29
30 #######################################################################
31 # Compute the flow magnitude, for visualization.
32
33 magnitude = np.sqrt(u ** 2 + v ** 2)
34
35 #######################################################################
36 # Create a viewer, add the vortex frames, and overlay the flow
37 # magnitude.
38
39 viewer, vortex_layer = napari.imshow(vortex_im)
40 mag_layer = viewer.add_image(magnitude, colormap='magma', opacity=0.3)
41
42 #######################################################################
43 # Finally, we subsample the vector field to display it — it's too
44 # messy otherwise! And we transpose the rows/columns axes to match the
45 # current scikit-image output.
46
47 nvec = 21
48 nr, nc = magnitude.shape
49 step = max(nr//nvec, nc//nvec)
50 offset = step // 2
51 usub = u[offset::step, offset::step]
52 vsub = v[offset::step, offset::step]
53
54 vectors_field = np.transpose( # transpose required — skimage bug?
55 np.stack([usub, vsub], axis=-1),
56 (1, 0, 2),
57 )
58
59 flow_layer = viewer.add_vectors(
60 vectors_field,
61 name='optical flow',
62 scale=[step, step],
63 translate=[offset, offset],
64 edge_width=0.3,
65 length=0.3,
66 )
67
68 if __name__ == '__main__':
69 napari.run()
70
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/examples/vortex.py b/examples/vortex.py
--- a/examples/vortex.py
+++ b/examples/vortex.py
@@ -1,4 +1,6 @@
-"""Visualizing optical flow in napari.
+"""
+Visualizing optical flow in napari
+==================================
Adapted from the scikit-image gallery [1]_.
@@ -33,14 +35,7 @@
magnitude = np.sqrt(u ** 2 + v ** 2)
#######################################################################
-# Create a viewer, add the vortex frames, and overlay the flow
-# magnitude.
-
-viewer, vortex_layer = napari.imshow(vortex_im)
-mag_layer = viewer.add_image(magnitude, colormap='magma', opacity=0.3)
-
-#######################################################################
-# Finally, we subsample the vector field to display it — it's too
+# We subsample the vector field to display it — it's too
# messy otherwise! And we transpose the rows/columns axes to match the
# current scikit-image output.
@@ -56,6 +51,12 @@
(1, 0, 2),
)
+#######################################################################
+# Finally, we create a viewer, and add the vortex frames, the flow
+# magnitude, and the vector field.
+
+viewer, vortex_layer = napari.imshow(vortex_im)
+mag_layer = viewer.add_image(magnitude, colormap='magma', opacity=0.3)
flow_layer = viewer.add_vectors(
vectors_field,
name='optical flow',
| {"golden_diff": "diff --git a/examples/vortex.py b/examples/vortex.py\n--- a/examples/vortex.py\n+++ b/examples/vortex.py\n@@ -1,4 +1,6 @@\n-\"\"\"Visualizing optical flow in napari.\n+\"\"\"\n+Visualizing optical flow in napari\n+==================================\n \n Adapted from the scikit-image gallery [1]_.\n \n@@ -33,14 +35,7 @@\n magnitude = np.sqrt(u ** 2 + v ** 2)\n \n #######################################################################\n-# Create a viewer, add the vortex frames, and overlay the flow\n-# magnitude.\n-\n-viewer, vortex_layer = napari.imshow(vortex_im)\n-mag_layer = viewer.add_image(magnitude, colormap='magma', opacity=0.3)\n-\n-#######################################################################\n-# Finally, we subsample the vector field to display it \u2014 it's too\n+# We subsample the vector field to display it \u2014 it's too\n # messy otherwise! And we transpose the rows/columns axes to match the\n # current scikit-image output.\n \n@@ -56,6 +51,12 @@\n (1, 0, 2),\n )\n \n+#######################################################################\n+# Finally, we create a viewer, and add the vortex frames, the flow\n+# magnitude, and the vector field.\n+\n+viewer, vortex_layer = napari.imshow(vortex_im)\n+mag_layer = viewer.add_image(magnitude, colormap='magma', opacity=0.3)\n flow_layer = viewer.add_vectors(\n vectors_field,\n name='optical flow',\n", "issue": "Weird docs build error on vortex.py example\n### \ud83d\udc1b Bug Report\n\nIt looks like we are missing the gallery example for [vortex.py](https://github.com/napari/napari/blob/1b510bab020aae485565f000ddab842ab17ef608/examples/vortex.py). That file runs fine locally and *I think* ~~it ran fine in the PR~~ (**Edit:** nope, it was [broken](https://output.circle-artifacts.com/output/job/1b750fb4-4df5-462b-84ca-cdddeede41ff/artifacts/0/napari-docs/docs/_build/gallery.html#gallery), but gallery errors aren't errors. I don't know what the right answer is there but either we should turn them into errors or we should turn them into errors *when the contribution is a gallery example*?) But the error suggests some form of race condition during addition of the layer, which makes me think this is an async issue. Here's the error message from [this build](https://github.com/napari/docs/actions/runs/6658042739/job/18094063805#step:9:238):\r\n\r\n```pytb\r\nDownloading file 'data/pivchallenge-B-B001_1.tif' from 'https://gitlab.com/scikit-image/data/-/raw/2cdc5ce89b334d28f06a58c9f0ca21aa6992a5ba/pivchallenge/B/B001_1.tif' to '/home/runner/.cache/scikit-image/0.22.0'.\r\nDownloading file 'data/pivchallenge-B-B001_2.tif' from 'https://gitlab.com/scikit-image/data/-/raw/2cdc5ce89b334d28f06a58c9f0ca21aa6992a5ba/pivchallenge/B/B001_2.tif' to '/home/runner/.cache/scikit-image/0.22.0'.\r\nWARNING: /home/runner/work/docs/docs/docs/examples/vortex.py failed to execute correctly: Traceback (most recent call last):\r\n File \"/home/runner/work/docs/docs/docs/examples/vortex.py\", line 59, in <module>\r\n flow_layer = viewer.add_vectors(\r\n File \"/opt/hostedtoolcache/Python/3.10.13/x64/lib/python3.10/site-packages/napari/components/viewer_model.py\", line 5, in add_vectors\r\n import os\r\n File \"/opt/hostedtoolcache/Python/3.10.13/x64/lib/python3.10/_collections_abc.py\", line 1128, in append\r\n self.insert(len(self), value)\r\n File \"/opt/hostedtoolcache/Python/3.10.13/x64/lib/python3.10/site-packages/napari/components/layerlist.py\", line 194, in insert\r\n super().insert(index, new_layer)\r\n File \"/opt/hostedtoolcache/Python/3.10.13/x64/lib/python3.10/site-packages/napari/utils/events/containers/_selectable_list.py\", line 71, in insert\r\n self.selection.active = value\r\n File \"/opt/hostedtoolcache/Python/3.10.13/x64/lib/python3.10/site-packages/napari/utils/events/containers/_selection.py\", line 108, in active\r\n self.events.active(value=value)\r\n File \"/opt/hostedtoolcache/Python/3.10.13/x64/lib/python3.10/site-packages/napari/utils/events/event.py\", line 771, in __call__\r\n self._invoke_callback(cb, event if pass_event else None)\r\n File \"/opt/hostedtoolcache/Python/3.10.13/x64/lib/python3.10/site-packages/napari/utils/events/event.py\", line 809, in _invoke_callback\r\n _handle_exception(\r\n File \"/opt/hostedtoolcache/Python/3.10.13/x64/lib/python3.10/site-packages/napari/utils/events/event.py\", line 796, in _invoke_callback\r\n cb(event)\r\n File \"/opt/hostedtoolcache/Python/3.10.13/x64/lib/python3.10/site-packages/napari/_qt/layer_controls/qt_layer_controls_container.py\", line 130, in _display\r\n controls = self.widgets[layer]\r\nKeyError: <Vectors layer 'optical flow' at 0x7f6f55be4be0>\r\n```\n\n### \ud83d\udca1 Steps to Reproduce\n\nI don't have a working docs build at the moment so I don't know whether this reproduces in local docs builds, but the example runs fine locally. So this is either a problem with sphinx gallery or with sphinx gallery on CI.\n\n### \ud83d\udca1 Expected Behavior\n\nExample should run fine on GHA.\n\n### \ud83c\udf0e Environment\n\nnapari main on CI \ud83d\ude2c \r\n\r\n(Note: should we echo `napari --info` on all our CI?)\n\n### \ud83d\udca1 Additional Context\n\n_No response_\n", "before_files": [{"content": "\"\"\"Visualizing optical flow in napari.\n\nAdapted from the scikit-image gallery [1]_.\n\nIn napari, we can show the flowing vortex as an additional dimension in the\nimage, visible by moving the slider.\n\n.. tags:: visualization-advanced, layers\n\n.. [1] https://scikit-image.org/docs/stable/auto_examples/registration/plot_opticalflow.html\n\"\"\"\nimport numpy as np\nfrom skimage.data import vortex\nfrom skimage.registration import optical_flow_ilk\n\nimport napari\n\n#######################################################################\n# First, we load the vortex image as a 3D array. (time, row, column)\n\nvortex_im = np.asarray(vortex())\n\n#######################################################################\n# We compute the optical flow using scikit-image. (Note: as of\n# scikit-image 0.21, there seems to be a transposition of the image in\n# the output, which we account for later.)\n\nu, v = optical_flow_ilk(vortex_im[0], vortex_im[1], radius=15)\n\n#######################################################################\n# Compute the flow magnitude, for visualization.\n\nmagnitude = np.sqrt(u ** 2 + v ** 2)\n\n#######################################################################\n# Create a viewer, add the vortex frames, and overlay the flow\n# magnitude.\n\nviewer, vortex_layer = napari.imshow(vortex_im)\nmag_layer = viewer.add_image(magnitude, colormap='magma', opacity=0.3)\n\n#######################################################################\n# Finally, we subsample the vector field to display it \u2014 it's too\n# messy otherwise! And we transpose the rows/columns axes to match the\n# current scikit-image output.\n\nnvec = 21\nnr, nc = magnitude.shape\nstep = max(nr//nvec, nc//nvec)\noffset = step // 2\nusub = u[offset::step, offset::step]\nvsub = v[offset::step, offset::step]\n\nvectors_field = np.transpose( # transpose required \u2014 skimage bug?\n np.stack([usub, vsub], axis=-1),\n (1, 0, 2),\n )\n\nflow_layer = viewer.add_vectors(\n vectors_field,\n name='optical flow',\n scale=[step, step],\n translate=[offset, offset],\n edge_width=0.3,\n length=0.3,\n )\n\nif __name__ == '__main__':\n napari.run()\n", "path": "examples/vortex.py"}], "after_files": [{"content": "\"\"\"\nVisualizing optical flow in napari\n==================================\n\nAdapted from the scikit-image gallery [1]_.\n\nIn napari, we can show the flowing vortex as an additional dimension in the\nimage, visible by moving the slider.\n\n.. tags:: visualization-advanced, layers\n\n.. [1] https://scikit-image.org/docs/stable/auto_examples/registration/plot_opticalflow.html\n\"\"\"\nimport numpy as np\nfrom skimage.data import vortex\nfrom skimage.registration import optical_flow_ilk\n\nimport napari\n\n#######################################################################\n# First, we load the vortex image as a 3D array. (time, row, column)\n\nvortex_im = np.asarray(vortex())\n\n#######################################################################\n# We compute the optical flow using scikit-image. (Note: as of\n# scikit-image 0.21, there seems to be a transposition of the image in\n# the output, which we account for later.)\n\nu, v = optical_flow_ilk(vortex_im[0], vortex_im[1], radius=15)\n\n#######################################################################\n# Compute the flow magnitude, for visualization.\n\nmagnitude = np.sqrt(u ** 2 + v ** 2)\n\n#######################################################################\n# We subsample the vector field to display it \u2014 it's too\n# messy otherwise! And we transpose the rows/columns axes to match the\n# current scikit-image output.\n\nnvec = 21\nnr, nc = magnitude.shape\nstep = max(nr//nvec, nc//nvec)\noffset = step // 2\nusub = u[offset::step, offset::step]\nvsub = v[offset::step, offset::step]\n\nvectors_field = np.transpose( # transpose required \u2014 skimage bug?\n np.stack([usub, vsub], axis=-1),\n (1, 0, 2),\n )\n\n#######################################################################\n# Finally, we create a viewer, and add the vortex frames, the flow\n# magnitude, and the vector field.\n\nviewer, vortex_layer = napari.imshow(vortex_im)\nmag_layer = viewer.add_image(magnitude, colormap='magma', opacity=0.3)\nflow_layer = viewer.add_vectors(\n vectors_field,\n name='optical flow',\n scale=[step, step],\n translate=[offset, offset],\n edge_width=0.3,\n length=0.3,\n )\n\nif __name__ == '__main__':\n napari.run()\n", "path": "examples/vortex.py"}]} | 2,074 | 330 |
gh_patches_debug_29254 | rasdani/github-patches | git_diff | web2py__web2py-1907 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Streamer.py handles IOError exception in non-python3 compatible way
Under python 3 and the latest repository, browsing to `http://127.0.0.1:8000/welcome/favicon.ico` causes `streamer.py` to crash because it treats the `IOError` exception in a non-python3 compatible way. The IOError exception occurs because `favicon.ico` is not found at `/ `. No error ticket is generated.
except IOError as e:
if e[0] == errno.EISDIR:
raise HTTP(403, error_message, web2py_error='file is a directory')
elif e[0] == errno.EACCES:
raise HTTP(403, error_message, web2py_error='inaccessible file')
else:
raise HTTP(404, error_message, web2py_error='invalid file')
This works in python 2, but `e[0]` should be accesed as `e.errno` under python 3
Partial stack trace:
Traceback (most recent call last):
File "C:\web2py\gluon\main.py", line 329, in wsgibase
response.stream(static_file, request=request)
File "C:\web2py\gluon\globals.py", line 617, in stream
status=self.status)
File "C:\web2py\gluon\streamer.py", line 66, in stream_file_or_304_or_206
if e[0] == errno.EISDIR:
TypeError: 'FileNotFoundError' object is not subscriptable
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `gluon/streamer.py`
Content:
```
1 #!/usr/bin/env python
2 # -*- coding: utf-8 -*-
3
4 """
5 | This file is part of the web2py Web Framework
6 | Copyrighted by Massimo Di Pierro <[email protected]>
7 | License: LGPLv3 (http://www.gnu.org/licenses/lgpl.html)
8
9 Facilities to handle file streaming
10 ------------------------------------
11 """
12
13 import os
14 import stat
15 import time
16 import re
17 import errno
18 from gluon.http import HTTP
19 from gluon.contenttype import contenttype
20 from gluon._compat import PY2
21
22
23 regex_start_range = re.compile('\d+(?=\-)')
24 regex_stop_range = re.compile('(?<=\-)\d+')
25
26 DEFAULT_CHUNK_SIZE = 64 * 1024
27
28 def streamer(stream, chunk_size=DEFAULT_CHUNK_SIZE, bytes=None, callback=None):
29 try:
30 offset = 0
31 while bytes is None or offset < bytes:
32 if not bytes is None and bytes - offset < chunk_size:
33 chunk_size = bytes - offset
34 data = stream.read(chunk_size)
35 length = len(data)
36 if not length:
37 break
38 else:
39 yield data
40 if length < chunk_size:
41 break
42 offset += length
43 finally:
44 stream.close()
45 if callback:
46 callback()
47
48 def stream_file_or_304_or_206(
49 static_file,
50 chunk_size=DEFAULT_CHUNK_SIZE,
51 request=None,
52 headers={},
53 status=200,
54 error_message=None
55 ):
56 # FIX THIS
57 # if error_message is None:
58 # error_message = rewrite.THREAD_LOCAL.routes.error_message % 'invalid request'
59 try:
60 if PY2:
61 open_f = file # this makes no sense but without it GAE cannot open files
62 else:
63 open_f = open
64 fp = open_f(static_file,'rb')
65 except IOError as e:
66 if e[0] == errno.EISDIR:
67 raise HTTP(403, error_message, web2py_error='file is a directory')
68 elif e[0] == errno.EACCES:
69 raise HTTP(403, error_message, web2py_error='inaccessible file')
70 else:
71 raise HTTP(404, error_message, web2py_error='invalid file')
72 else:
73 fp.close()
74 stat_file = os.stat(static_file)
75 fsize = stat_file[stat.ST_SIZE]
76 modified = stat_file[stat.ST_MTIME]
77 mtime = time.strftime('%a, %d %b %Y %H:%M:%S GMT', time.gmtime(modified))
78 headers.setdefault('Content-Type', contenttype(static_file))
79 headers.setdefault('Last-Modified', mtime)
80 headers.setdefault('Pragma', 'cache')
81 headers.setdefault('Cache-Control', 'private')
82
83 # if this is a normal response and not a respnse to an error page
84 if status == 200:
85 if request and request.env.http_if_modified_since == mtime:
86 raise HTTP(304, **{'Content-Type': headers['Content-Type']})
87
88 elif request and request.env.http_range:
89 start_items = regex_start_range.findall(request.env.http_range)
90 if not start_items:
91 start_items = [0]
92 stop_items = regex_stop_range.findall(request.env.http_range)
93 if not stop_items or int(stop_items[0]) > fsize - 1:
94 stop_items = [fsize - 1]
95 part = (int(start_items[0]), int(stop_items[0]), fsize)
96 bytes = part[1] - part[0] + 1
97 try:
98 stream = open(static_file, 'rb')
99 except IOError as e:
100 if e[0] in (errno.EISDIR, errno.EACCES):
101 raise HTTP(403)
102 else:
103 raise HTTP(404)
104 stream.seek(part[0])
105 headers['Content-Range'] = 'bytes %i-%i/%i' % part
106 headers['Content-Length'] = '%i' % bytes
107 status = 206
108 # in all the other cases (not 304, not 206, but 200 or error page)
109 if status != 206:
110 enc = request.env.http_accept_encoding
111 if enc and 'gzip' in enc and not 'Content-Encoding' in headers:
112 gzipped = static_file + '.gz'
113 if os.path.isfile(gzipped) and os.path.getmtime(gzipped) >= modified:
114 static_file = gzipped
115 fsize = os.path.getsize(gzipped)
116 headers['Content-Encoding'] = 'gzip'
117 headers['Vary'] = 'Accept-Encoding'
118 try:
119 stream = open(static_file, 'rb')
120 except IOError as e:
121 # this better does not happer when returning an error page ;-)
122 if e[0] in (errno.EISDIR, errno.EACCES):
123 raise HTTP(403)
124 else:
125 raise HTTP(404)
126 headers['Content-Length'] = fsize
127 bytes = None
128 if request and request.env.web2py_use_wsgi_file_wrapper:
129 wrapped = request.env.wsgi_file_wrapper(stream, chunk_size)
130 else:
131 wrapped = streamer(stream, chunk_size=chunk_size, bytes=bytes)
132 raise HTTP(status, wrapped, **headers)
133
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/gluon/streamer.py b/gluon/streamer.py
--- a/gluon/streamer.py
+++ b/gluon/streamer.py
@@ -63,9 +63,9 @@
open_f = open
fp = open_f(static_file,'rb')
except IOError as e:
- if e[0] == errno.EISDIR:
+ if e.errno == errno.EISDIR:
raise HTTP(403, error_message, web2py_error='file is a directory')
- elif e[0] == errno.EACCES:
+ elif e.errno == errno.EACCES:
raise HTTP(403, error_message, web2py_error='inaccessible file')
else:
raise HTTP(404, error_message, web2py_error='invalid file')
@@ -97,7 +97,7 @@
try:
stream = open(static_file, 'rb')
except IOError as e:
- if e[0] in (errno.EISDIR, errno.EACCES):
+ if e.errno in (errno.EISDIR, errno.EACCES):
raise HTTP(403)
else:
raise HTTP(404)
@@ -118,8 +118,8 @@
try:
stream = open(static_file, 'rb')
except IOError as e:
- # this better does not happer when returning an error page ;-)
- if e[0] in (errno.EISDIR, errno.EACCES):
+ # this better not happen when returning an error page ;-)
+ if e.errno in (errno.EISDIR, errno.EACCES):
raise HTTP(403)
else:
raise HTTP(404)
| {"golden_diff": "diff --git a/gluon/streamer.py b/gluon/streamer.py\n--- a/gluon/streamer.py\n+++ b/gluon/streamer.py\n@@ -63,9 +63,9 @@\n open_f = open\n fp = open_f(static_file,'rb')\n except IOError as e:\n- if e[0] == errno.EISDIR:\n+ if e.errno == errno.EISDIR:\n raise HTTP(403, error_message, web2py_error='file is a directory')\n- elif e[0] == errno.EACCES:\n+ elif e.errno == errno.EACCES:\n raise HTTP(403, error_message, web2py_error='inaccessible file')\n else:\n raise HTTP(404, error_message, web2py_error='invalid file')\n@@ -97,7 +97,7 @@\n try:\n stream = open(static_file, 'rb')\n except IOError as e:\n- if e[0] in (errno.EISDIR, errno.EACCES):\n+ if e.errno in (errno.EISDIR, errno.EACCES):\n raise HTTP(403)\n else:\n raise HTTP(404)\n@@ -118,8 +118,8 @@\n try:\n stream = open(static_file, 'rb')\n except IOError as e:\n- # this better does not happer when returning an error page ;-)\n- if e[0] in (errno.EISDIR, errno.EACCES):\n+ # this better not happen when returning an error page ;-)\n+ if e.errno in (errno.EISDIR, errno.EACCES):\n raise HTTP(403)\n else:\n raise HTTP(404)\n", "issue": "Streamer.py handles IOError exception in non-python3 compatible way\nUnder python 3 and the latest repository, browsing to `http://127.0.0.1:8000/welcome/favicon.ico` causes `streamer.py` to crash because it treats the `IOError` exception in a non-python3 compatible way. The IOError exception occurs because `favicon.ico` is not found at `/ `. No error ticket is generated.\r\n\r\n except IOError as e:\r\n if e[0] == errno.EISDIR:\r\n raise HTTP(403, error_message, web2py_error='file is a directory')\r\n elif e[0] == errno.EACCES:\r\n raise HTTP(403, error_message, web2py_error='inaccessible file')\r\n else:\r\n raise HTTP(404, error_message, web2py_error='invalid file')\r\n\r\nThis works in python 2, but `e[0]` should be accesed as `e.errno` under python 3\r\n\r\nPartial stack trace:\r\n\r\n Traceback (most recent call last):\r\n File \"C:\\web2py\\gluon\\main.py\", line 329, in wsgibase\r\n response.stream(static_file, request=request)\r\n File \"C:\\web2py\\gluon\\globals.py\", line 617, in stream\r\n status=self.status)\r\n File \"C:\\web2py\\gluon\\streamer.py\", line 66, in stream_file_or_304_or_206\r\n if e[0] == errno.EISDIR:\r\n TypeError: 'FileNotFoundError' object is not subscriptable\r\n\n", "before_files": [{"content": "#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\n\"\"\"\n| This file is part of the web2py Web Framework\n| Copyrighted by Massimo Di Pierro <[email protected]>\n| License: LGPLv3 (http://www.gnu.org/licenses/lgpl.html)\n\nFacilities to handle file streaming\n------------------------------------\n\"\"\"\n\nimport os\nimport stat\nimport time\nimport re\nimport errno\nfrom gluon.http import HTTP\nfrom gluon.contenttype import contenttype\nfrom gluon._compat import PY2\n\n\nregex_start_range = re.compile('\\d+(?=\\-)')\nregex_stop_range = re.compile('(?<=\\-)\\d+')\n\nDEFAULT_CHUNK_SIZE = 64 * 1024\n\ndef streamer(stream, chunk_size=DEFAULT_CHUNK_SIZE, bytes=None, callback=None):\n try:\n offset = 0\n while bytes is None or offset < bytes:\n if not bytes is None and bytes - offset < chunk_size:\n chunk_size = bytes - offset\n data = stream.read(chunk_size)\n length = len(data)\n if not length:\n break\n else:\n yield data\n if length < chunk_size:\n break\n offset += length\n finally:\n stream.close()\n if callback:\n callback()\n\ndef stream_file_or_304_or_206(\n static_file,\n chunk_size=DEFAULT_CHUNK_SIZE,\n request=None,\n headers={},\n status=200,\n error_message=None\n ):\n # FIX THIS\n # if error_message is None:\n # error_message = rewrite.THREAD_LOCAL.routes.error_message % 'invalid request'\n try:\n if PY2:\n open_f = file # this makes no sense but without it GAE cannot open files\n else:\n open_f = open\n fp = open_f(static_file,'rb')\n except IOError as e:\n if e[0] == errno.EISDIR:\n raise HTTP(403, error_message, web2py_error='file is a directory')\n elif e[0] == errno.EACCES:\n raise HTTP(403, error_message, web2py_error='inaccessible file')\n else:\n raise HTTP(404, error_message, web2py_error='invalid file')\n else:\n fp.close()\n stat_file = os.stat(static_file)\n fsize = stat_file[stat.ST_SIZE]\n modified = stat_file[stat.ST_MTIME]\n mtime = time.strftime('%a, %d %b %Y %H:%M:%S GMT', time.gmtime(modified))\n headers.setdefault('Content-Type', contenttype(static_file))\n headers.setdefault('Last-Modified', mtime)\n headers.setdefault('Pragma', 'cache')\n headers.setdefault('Cache-Control', 'private')\n\n # if this is a normal response and not a respnse to an error page\n if status == 200:\n if request and request.env.http_if_modified_since == mtime:\n raise HTTP(304, **{'Content-Type': headers['Content-Type']})\n\n elif request and request.env.http_range:\n start_items = regex_start_range.findall(request.env.http_range)\n if not start_items:\n start_items = [0]\n stop_items = regex_stop_range.findall(request.env.http_range)\n if not stop_items or int(stop_items[0]) > fsize - 1:\n stop_items = [fsize - 1]\n part = (int(start_items[0]), int(stop_items[0]), fsize)\n bytes = part[1] - part[0] + 1\n try:\n stream = open(static_file, 'rb')\n except IOError as e:\n if e[0] in (errno.EISDIR, errno.EACCES):\n raise HTTP(403)\n else:\n raise HTTP(404)\n stream.seek(part[0])\n headers['Content-Range'] = 'bytes %i-%i/%i' % part\n headers['Content-Length'] = '%i' % bytes\n status = 206\n # in all the other cases (not 304, not 206, but 200 or error page)\n if status != 206:\n enc = request.env.http_accept_encoding\n if enc and 'gzip' in enc and not 'Content-Encoding' in headers:\n gzipped = static_file + '.gz'\n if os.path.isfile(gzipped) and os.path.getmtime(gzipped) >= modified:\n static_file = gzipped\n fsize = os.path.getsize(gzipped)\n headers['Content-Encoding'] = 'gzip'\n headers['Vary'] = 'Accept-Encoding'\n try:\n stream = open(static_file, 'rb')\n except IOError as e:\n # this better does not happer when returning an error page ;-)\n if e[0] in (errno.EISDIR, errno.EACCES):\n raise HTTP(403)\n else:\n raise HTTP(404)\n headers['Content-Length'] = fsize\n bytes = None\n if request and request.env.web2py_use_wsgi_file_wrapper:\n wrapped = request.env.wsgi_file_wrapper(stream, chunk_size)\n else:\n wrapped = streamer(stream, chunk_size=chunk_size, bytes=bytes)\n raise HTTP(status, wrapped, **headers)\n", "path": "gluon/streamer.py"}], "after_files": [{"content": "#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\n\"\"\"\n| This file is part of the web2py Web Framework\n| Copyrighted by Massimo Di Pierro <[email protected]>\n| License: LGPLv3 (http://www.gnu.org/licenses/lgpl.html)\n\nFacilities to handle file streaming\n------------------------------------\n\"\"\"\n\nimport os\nimport stat\nimport time\nimport re\nimport errno\nfrom gluon.http import HTTP\nfrom gluon.contenttype import contenttype\nfrom gluon._compat import PY2\n\n\nregex_start_range = re.compile('\\d+(?=\\-)')\nregex_stop_range = re.compile('(?<=\\-)\\d+')\n\nDEFAULT_CHUNK_SIZE = 64 * 1024\n\ndef streamer(stream, chunk_size=DEFAULT_CHUNK_SIZE, bytes=None, callback=None):\n try:\n offset = 0\n while bytes is None or offset < bytes:\n if not bytes is None and bytes - offset < chunk_size:\n chunk_size = bytes - offset\n data = stream.read(chunk_size)\n length = len(data)\n if not length:\n break\n else:\n yield data\n if length < chunk_size:\n break\n offset += length\n finally:\n stream.close()\n if callback:\n callback()\n\ndef stream_file_or_304_or_206(\n static_file,\n chunk_size=DEFAULT_CHUNK_SIZE,\n request=None,\n headers={},\n status=200,\n error_message=None\n ):\n # FIX THIS\n # if error_message is None:\n # error_message = rewrite.THREAD_LOCAL.routes.error_message % 'invalid request'\n try:\n if PY2:\n open_f = file # this makes no sense but without it GAE cannot open files\n else:\n open_f = open\n fp = open_f(static_file,'rb')\n except IOError as e:\n if e.errno == errno.EISDIR:\n raise HTTP(403, error_message, web2py_error='file is a directory')\n elif e.errno == errno.EACCES:\n raise HTTP(403, error_message, web2py_error='inaccessible file')\n else:\n raise HTTP(404, error_message, web2py_error='invalid file')\n else:\n fp.close()\n stat_file = os.stat(static_file)\n fsize = stat_file[stat.ST_SIZE]\n modified = stat_file[stat.ST_MTIME]\n mtime = time.strftime('%a, %d %b %Y %H:%M:%S GMT', time.gmtime(modified))\n headers.setdefault('Content-Type', contenttype(static_file))\n headers.setdefault('Last-Modified', mtime)\n headers.setdefault('Pragma', 'cache')\n headers.setdefault('Cache-Control', 'private')\n\n # if this is a normal response and not a respnse to an error page\n if status == 200:\n if request and request.env.http_if_modified_since == mtime:\n raise HTTP(304, **{'Content-Type': headers['Content-Type']})\n\n elif request and request.env.http_range:\n start_items = regex_start_range.findall(request.env.http_range)\n if not start_items:\n start_items = [0]\n stop_items = regex_stop_range.findall(request.env.http_range)\n if not stop_items or int(stop_items[0]) > fsize - 1:\n stop_items = [fsize - 1]\n part = (int(start_items[0]), int(stop_items[0]), fsize)\n bytes = part[1] - part[0] + 1\n try:\n stream = open(static_file, 'rb')\n except IOError as e:\n if e.errno in (errno.EISDIR, errno.EACCES):\n raise HTTP(403)\n else:\n raise HTTP(404)\n stream.seek(part[0])\n headers['Content-Range'] = 'bytes %i-%i/%i' % part\n headers['Content-Length'] = '%i' % bytes\n status = 206\n # in all the other cases (not 304, not 206, but 200 or error page)\n if status != 206:\n enc = request.env.http_accept_encoding\n if enc and 'gzip' in enc and not 'Content-Encoding' in headers:\n gzipped = static_file + '.gz'\n if os.path.isfile(gzipped) and os.path.getmtime(gzipped) >= modified:\n static_file = gzipped\n fsize = os.path.getsize(gzipped)\n headers['Content-Encoding'] = 'gzip'\n headers['Vary'] = 'Accept-Encoding'\n try:\n stream = open(static_file, 'rb')\n except IOError as e:\n # this better not happen when returning an error page ;-)\n if e.errno in (errno.EISDIR, errno.EACCES):\n raise HTTP(403)\n else:\n raise HTTP(404)\n headers['Content-Length'] = fsize\n bytes = None\n if request and request.env.web2py_use_wsgi_file_wrapper:\n wrapped = request.env.wsgi_file_wrapper(stream, chunk_size)\n else:\n wrapped = streamer(stream, chunk_size=chunk_size, bytes=bytes)\n raise HTTP(status, wrapped, **headers)\n", "path": "gluon/streamer.py"}]} | 2,072 | 389 |
gh_patches_debug_5751 | rasdani/github-patches | git_diff | ansible__ansible-lint-1128 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
[701] No 'galaxy_info' found results in meta/requirements.yml file
### Summary
ansible-lint reporting `[701] No 'galaxy_info' found` in my `meta/requirements.yml`, a file that unlike `meta/main.yml` does not (to my knowledge) support a `galaxy_info` field.
##### Issue Type
- Bug Report
##### Ansible and Ansible Lint details
<!--- Paste verbatim output between tripple backticks -->
```console (paste below)
$ ansible --version
ansible 2.10.1
$ ansible-lint --version
ansible-lint 4.3.5
```
- ansible installation method: pipenv (pip)
- ansible-lint installation method: pipenv (pip)
##### OS / ENVIRONMENT
MacOS 10.15.7 (Catalina Latest)
##### STEPS TO REPRODUCE
Using this `meta/requirements.yml`
```bash
---
# insert third party deps here. download with:
# ansible-galaxy install -r requirements.yml
# https://docs.ansible.com/ansible/galaxy.html
- name: singleplatform-eng.users
version: v1.2.6
- name: weareinteractive.sudo
version: 1.14.1
- name: geerlingguy.fluentd
version: 1.1.0
```
Note that `meta/main.yml` does include `galaxy_info`, but does not include as dependencies, the roles listed in requirements.yml. This is purposeful because I'm choosing `meta/requirements.yml` instead of `meta/main.yml` because I prefer the workflow and do not want the roles running first, as they do when in `meta/main.yml`. I'm following the previously linked user-guide on this topic.
To reproduce, I simply run ansible-lint directly or via molecule.
##### Desired Behaviour
I would expect ansible-lint not to flag these as issues... unless I'm completely misunderstanding the finding and misreading the documentation associated with this rule.
##### Actual Behaviour
Below are the ansible-lint results when run on my role.
```bash
$ ansible-lint
[701] No 'galaxy_info' found
meta/requirements.yml:7
{'meta/main.yml': {'name': 'singleplatform-eng.users', 'version': 'v1.2.6', '__line__': 7, '__file__': '/Users/tmichael/orgs/tmb/ansible_roles/base/meta/requirements.yml', 'skipped_rules': []}}
[701] No 'galaxy_info' found
meta/requirements.yml:10
{'meta/main.yml': {'name': 'weareinteractive.sudo', 'version': '1.14.1', '__line__': 10, '__file__': '/Users/tmichael/orgs/tmb/ansible_roles/base/meta/requirements.yml'}}
[701] No 'galaxy_info' found
meta/requirements.yml:13
{'meta/main.yml': {'name': 'geerlingguy.fluentd', 'version': '1.1.0', '__line__': 13, '__file__': '/Users/tmichael/orgs/tmb/ansible_roles/base/meta/requirements.yml'}}
```
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `lib/ansiblelint/rules/MetaMainHasInfoRule.py`
Content:
```
1 # Copyright (c) 2016, Will Thames and contributors
2 # Copyright (c) 2018, Ansible Project
3
4 from ansiblelint.rules import AnsibleLintRule
5
6 META_STR_INFO = (
7 'author',
8 'description'
9 )
10 META_INFO = tuple(list(META_STR_INFO) + [
11 'license',
12 'min_ansible_version',
13 'platforms',
14 ])
15
16
17 def _platform_info_errors_itr(platforms):
18 if not isinstance(platforms, list):
19 yield 'Platforms should be a list of dictionaries'
20 return
21
22 for platform in platforms:
23 if not isinstance(platform, dict):
24 yield 'Platforms should be a list of dictionaries'
25 elif 'name' not in platform:
26 yield 'Platform should contain name'
27
28
29 def _galaxy_info_errors_itr(galaxy_info,
30 info_list=META_INFO,
31 str_info_list=META_STR_INFO):
32 for info in info_list:
33 ginfo = galaxy_info.get(info, False)
34 if ginfo:
35 if info in str_info_list and not isinstance(ginfo, str):
36 yield '{info} should be a string'.format(info=info)
37 elif info == 'platforms':
38 for err in _platform_info_errors_itr(ginfo):
39 yield err
40 else:
41 yield 'Role info should contain {info}'.format(info=info)
42
43
44 class MetaMainHasInfoRule(AnsibleLintRule):
45 id = '701'
46 shortdesc = 'meta/main.yml should contain relevant info'
47 str_info = META_STR_INFO
48 info = META_INFO
49 description = (
50 'meta/main.yml should contain: ``{}``'.format(', '.join(info))
51 )
52 severity = 'HIGH'
53 tags = ['metadata']
54 version_added = 'v4.0.0'
55
56 def matchplay(self, file, data):
57 if file['type'] != 'meta':
58 return False
59
60 meta = {'meta/main.yml': data}
61 galaxy_info = data.get('galaxy_info', False)
62 if galaxy_info:
63 return [(meta, err) for err
64 in _galaxy_info_errors_itr(galaxy_info)]
65
66 return [(meta, "No 'galaxy_info' found")]
67
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/lib/ansiblelint/rules/MetaMainHasInfoRule.py b/lib/ansiblelint/rules/MetaMainHasInfoRule.py
--- a/lib/ansiblelint/rules/MetaMainHasInfoRule.py
+++ b/lib/ansiblelint/rules/MetaMainHasInfoRule.py
@@ -57,6 +57,11 @@
if file['type'] != 'meta':
return False
+ # since Ansible 2.10 we can add a meta/requirements.yml but
+ # we only want to match on meta/main.yml
+ if not file['path'].endswith('/main.yml'):
+ return False
+
meta = {'meta/main.yml': data}
galaxy_info = data.get('galaxy_info', False)
if galaxy_info:
| {"golden_diff": "diff --git a/lib/ansiblelint/rules/MetaMainHasInfoRule.py b/lib/ansiblelint/rules/MetaMainHasInfoRule.py\n--- a/lib/ansiblelint/rules/MetaMainHasInfoRule.py\n+++ b/lib/ansiblelint/rules/MetaMainHasInfoRule.py\n@@ -57,6 +57,11 @@\n if file['type'] != 'meta':\n return False\n \n+ # since Ansible 2.10 we can add a meta/requirements.yml but\n+ # we only want to match on meta/main.yml\n+ if not file['path'].endswith('/main.yml'):\n+ return False\n+\n meta = {'meta/main.yml': data}\n galaxy_info = data.get('galaxy_info', False)\n if galaxy_info:\n", "issue": "[701] No 'galaxy_info' found results in meta/requirements.yml file\n### Summary\r\n\r\nansible-lint reporting `[701] No 'galaxy_info' found` in my `meta/requirements.yml`, a file that unlike `meta/main.yml` does not (to my knowledge) support a `galaxy_info` field.\r\n\r\n##### Issue Type\r\n\r\n- Bug Report\r\n\r\n##### Ansible and Ansible Lint details\r\n<!--- Paste verbatim output between tripple backticks -->\r\n```console (paste below)\r\n$ ansible --version\r\nansible 2.10.1\r\n\r\n$ ansible-lint --version\r\nansible-lint 4.3.5\r\n```\r\n\r\n- ansible installation method: pipenv (pip)\r\n- ansible-lint installation method: pipenv (pip)\r\n\r\n##### OS / ENVIRONMENT\r\nMacOS 10.15.7 (Catalina Latest)\r\n\r\n##### STEPS TO REPRODUCE\r\n\r\nUsing this `meta/requirements.yml`\r\n```bash\r\n---\r\n\r\n# insert third party deps here. download with:\r\n# ansible-galaxy install -r requirements.yml\r\n# https://docs.ansible.com/ansible/galaxy.html\r\n\r\n- name: singleplatform-eng.users\r\n version: v1.2.6\r\n\r\n- name: weareinteractive.sudo\r\n version: 1.14.1\r\n\r\n- name: geerlingguy.fluentd\r\n version: 1.1.0\r\n```\r\n\r\nNote that `meta/main.yml` does include `galaxy_info`, but does not include as dependencies, the roles listed in requirements.yml. This is purposeful because I'm choosing `meta/requirements.yml` instead of `meta/main.yml` because I prefer the workflow and do not want the roles running first, as they do when in `meta/main.yml`. I'm following the previously linked user-guide on this topic.\r\n\r\nTo reproduce, I simply run ansible-lint directly or via molecule.\r\n\r\n##### Desired Behaviour\r\n\r\nI would expect ansible-lint not to flag these as issues... unless I'm completely misunderstanding the finding and misreading the documentation associated with this rule.\r\n\r\n##### Actual Behaviour\r\n\r\n\r\nBelow are the ansible-lint results when run on my role.\r\n```bash\r\n$ ansible-lint\r\n[701] No 'galaxy_info' found\r\nmeta/requirements.yml:7\r\n{'meta/main.yml': {'name': 'singleplatform-eng.users', 'version': 'v1.2.6', '__line__': 7, '__file__': '/Users/tmichael/orgs/tmb/ansible_roles/base/meta/requirements.yml', 'skipped_rules': []}}\r\n\r\n[701] No 'galaxy_info' found\r\nmeta/requirements.yml:10\r\n{'meta/main.yml': {'name': 'weareinteractive.sudo', 'version': '1.14.1', '__line__': 10, '__file__': '/Users/tmichael/orgs/tmb/ansible_roles/base/meta/requirements.yml'}}\r\n\r\n[701] No 'galaxy_info' found\r\nmeta/requirements.yml:13\r\n{'meta/main.yml': {'name': 'geerlingguy.fluentd', 'version': '1.1.0', '__line__': 13, '__file__': '/Users/tmichael/orgs/tmb/ansible_roles/base/meta/requirements.yml'}}\r\n```\n", "before_files": [{"content": "# Copyright (c) 2016, Will Thames and contributors\n# Copyright (c) 2018, Ansible Project\n\nfrom ansiblelint.rules import AnsibleLintRule\n\nMETA_STR_INFO = (\n 'author',\n 'description'\n)\nMETA_INFO = tuple(list(META_STR_INFO) + [\n 'license',\n 'min_ansible_version',\n 'platforms',\n])\n\n\ndef _platform_info_errors_itr(platforms):\n if not isinstance(platforms, list):\n yield 'Platforms should be a list of dictionaries'\n return\n\n for platform in platforms:\n if not isinstance(platform, dict):\n yield 'Platforms should be a list of dictionaries'\n elif 'name' not in platform:\n yield 'Platform should contain name'\n\n\ndef _galaxy_info_errors_itr(galaxy_info,\n info_list=META_INFO,\n str_info_list=META_STR_INFO):\n for info in info_list:\n ginfo = galaxy_info.get(info, False)\n if ginfo:\n if info in str_info_list and not isinstance(ginfo, str):\n yield '{info} should be a string'.format(info=info)\n elif info == 'platforms':\n for err in _platform_info_errors_itr(ginfo):\n yield err\n else:\n yield 'Role info should contain {info}'.format(info=info)\n\n\nclass MetaMainHasInfoRule(AnsibleLintRule):\n id = '701'\n shortdesc = 'meta/main.yml should contain relevant info'\n str_info = META_STR_INFO\n info = META_INFO\n description = (\n 'meta/main.yml should contain: ``{}``'.format(', '.join(info))\n )\n severity = 'HIGH'\n tags = ['metadata']\n version_added = 'v4.0.0'\n\n def matchplay(self, file, data):\n if file['type'] != 'meta':\n return False\n\n meta = {'meta/main.yml': data}\n galaxy_info = data.get('galaxy_info', False)\n if galaxy_info:\n return [(meta, err) for err\n in _galaxy_info_errors_itr(galaxy_info)]\n\n return [(meta, \"No 'galaxy_info' found\")]\n", "path": "lib/ansiblelint/rules/MetaMainHasInfoRule.py"}], "after_files": [{"content": "# Copyright (c) 2016, Will Thames and contributors\n# Copyright (c) 2018, Ansible Project\n\nfrom ansiblelint.rules import AnsibleLintRule\n\nMETA_STR_INFO = (\n 'author',\n 'description'\n)\nMETA_INFO = tuple(list(META_STR_INFO) + [\n 'license',\n 'min_ansible_version',\n 'platforms',\n])\n\n\ndef _platform_info_errors_itr(platforms):\n if not isinstance(platforms, list):\n yield 'Platforms should be a list of dictionaries'\n return\n\n for platform in platforms:\n if not isinstance(platform, dict):\n yield 'Platforms should be a list of dictionaries'\n elif 'name' not in platform:\n yield 'Platform should contain name'\n\n\ndef _galaxy_info_errors_itr(galaxy_info,\n info_list=META_INFO,\n str_info_list=META_STR_INFO):\n for info in info_list:\n ginfo = galaxy_info.get(info, False)\n if ginfo:\n if info in str_info_list and not isinstance(ginfo, str):\n yield '{info} should be a string'.format(info=info)\n elif info == 'platforms':\n for err in _platform_info_errors_itr(ginfo):\n yield err\n else:\n yield 'Role info should contain {info}'.format(info=info)\n\n\nclass MetaMainHasInfoRule(AnsibleLintRule):\n id = '701'\n shortdesc = 'meta/main.yml should contain relevant info'\n str_info = META_STR_INFO\n info = META_INFO\n description = (\n 'meta/main.yml should contain: ``{}``'.format(', '.join(info))\n )\n severity = 'HIGH'\n tags = ['metadata']\n version_added = 'v4.0.0'\n\n def matchplay(self, file, data):\n if file['type'] != 'meta':\n return False\n\n # since Ansible 2.10 we can add a meta/requirements.yml but\n # we only want to match on meta/main.yml\n if not file['path'].endswith('/main.yml'):\n return False\n\n meta = {'meta/main.yml': data}\n galaxy_info = data.get('galaxy_info', False)\n if galaxy_info:\n return [(meta, err) for err\n in _galaxy_info_errors_itr(galaxy_info)]\n\n return [(meta, \"No 'galaxy_info' found\")]\n", "path": "lib/ansiblelint/rules/MetaMainHasInfoRule.py"}]} | 1,576 | 169 |
gh_patches_debug_32180 | rasdani/github-patches | git_diff | mathesar-foundation__mathesar-3459 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Demo mode results in 5XX
## Description
<!-- A clear and concise description of what the bug is. -->
Mathesar is broken (as of 0.1.4) for Demo Mode. It doesn't load, and just says "Server Error (500)" instead.
## Expected behavior
<!-- A clear and concise description of what you expected to happen. -->
Demo mode should work.
## To Reproduce
<!-- How can we recreate this bug? Please try to provide a Minimal, Complete, and Verifiable (http://stackoverflow.com/help/mcve) example if code-related. -->
Change the `.env` file according to the comment to use the demo mode settings, try to build and start mathesar (dev environment is fine)
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `mathesar/install.py`
Content:
```
1 """
2 This script installs functions and types for Mathesar onto the configured DB.
3 """
4 import getopt
5 import os
6 import sys
7
8 import django
9 from django.core import management
10 from decouple import config as decouple_config
11 from django.conf import settings
12 from django.db.utils import IntegrityError
13 from sqlalchemy.exc import OperationalError
14 from db import install
15
16
17 def main(skip_static_collection=False):
18 # skip_confirm is temporarily enabled by default as we don't have any use
19 # for interactive prompts with docker only deployments
20 skip_confirm = True
21 (opts, _) = getopt.getopt(sys.argv[1:], ":s", ["skip-confirm"])
22 for (opt, value) in opts:
23 if (opt == "-s") or (opt == "--skip-confirm"):
24 skip_confirm = True
25 os.environ.setdefault("DJANGO_SETTINGS_MODULE", "config.settings.production")
26 django.setup()
27 management.call_command('migrate')
28 debug_mode = decouple_config('DEBUG', default=False, cast=bool)
29 #
30 if not debug_mode and not skip_static_collection:
31 management.call_command('collectstatic', '--noinput', '--clear')
32 print("------------Setting up User Databases------------")
33 django_db_key = decouple_config('DJANGO_DATABASE_KEY', default="default")
34 user_databases = [key for key in settings.DATABASES if key != django_db_key]
35 for database_key in user_databases:
36 try:
37 install_on_db_with_key(database_key, skip_confirm)
38 except IntegrityError:
39 continue
40
41
42 def install_on_db_with_key(database_key, skip_confirm):
43 from mathesar.models.base import Database
44 db_model = Database.create_from_settings_key(database_key)
45 db_model.save()
46 try:
47 install.install_mathesar(
48 database_name=db_model.db_name,
49 hostname=db_model.host,
50 username=db_model.username,
51 password=db_model.password,
52 port=db_model.port,
53 skip_confirm=skip_confirm
54 )
55 except OperationalError as e:
56 db_model.delete()
57 raise e
58
59
60 if __name__ == "__main__":
61 main()
62
```
Path: `demo/settings.py`
Content:
```
1 from config.settings.production import * # noqa
2 from config.settings import * # noqa
3 from decouple import config as decouple_config
4
5 INSTALLED_APPS += [ # noqa
6 "demo"
7 ]
8
9 MIDDLEWARE += [ # noqa
10 "demo.middleware.LiveDemoModeMiddleware",
11 ]
12
13 MATHESAR_LIVE_DEMO = True
14 MATHESAR_LIVE_DEMO_USERNAME = decouple_config('MATHESAR_LIVE_DEMO_USERNAME', default=None)
15 MATHESAR_LIVE_DEMO_PASSWORD = decouple_config('MATHESAR_LIVE_DEMO_PASSWORD', default=None)
16
17 MATHESAR_DEMO_TEMPLATE = 'mathesar_demo_template'
18 MATHESAR_DEMO_ARXIV_LOG_PATH = decouple_config(
19 'MATHESAR_DEMO_ARXIV_LOG_PATH',
20 default='/var/lib/mathesar/demo/arxiv_db_schema_log'
21 )
22 BASE_TEMPLATE_ADDITIONAL_SCRIPT_TEMPLATES += ['demo/analytics.html'] # noqa
23 ROOT_URLCONF = "demo.urls"
24
```
Path: `demo/management/commands/setup_demo_template_db.py`
Content:
```
1 from sqlalchemy import text
2 from sqlalchemy.exc import OperationalError
3 from django.conf import settings
4 from django.core.management import BaseCommand
5
6 from db.install import install_mathesar
7 from demo.install.datasets import load_datasets
8 from mathesar.database.base import create_mathesar_engine
9 from mathesar.models.base import Database
10
11
12 class Command(BaseCommand):
13 help = 'Initialize the demo template database.'
14
15 def handle(self, *args, **options):
16 _setup_demo_template_db()
17
18
19 def _setup_demo_template_db():
20 print("Initializing demo template database...")
21
22 template_db_name = settings.MATHESAR_DEMO_TEMPLATE
23 django_model = Database.current_objects.get(name=settings.DATABASES["default"]["NAME"])
24 root_engine = create_mathesar_engine(django_model)
25 with root_engine.connect() as conn:
26 conn.execution_options(isolation_level="AUTOCOMMIT")
27 conn.execute(text(f"DROP DATABASE IF EXISTS {template_db_name} WITH (FORCE)"))
28 root_engine.dispose()
29 db_model, _ = Database.current_objects.get_or_create(
30 name=template_db_name,
31 defaults={
32 'db_name': template_db_name,
33 'username': django_model.username,
34 'password': django_model.password,
35 'host': django_model.host,
36 'port': django_model.port
37 }
38 )
39 try:
40 install_mathesar(
41 database_name=template_db_name,
42 hostname=db_model.host,
43 username=db_model.username,
44 password=db_model.password,
45 port=db_model.port,
46 skip_confirm=True
47 )
48 except OperationalError as e:
49 db_model.delete()
50 raise e
51 user_engine = create_mathesar_engine(db_model)
52 load_datasets(user_engine)
53 user_engine.dispose()
54
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/demo/management/commands/setup_demo_template_db.py b/demo/management/commands/setup_demo_template_db.py
--- a/demo/management/commands/setup_demo_template_db.py
+++ b/demo/management/commands/setup_demo_template_db.py
@@ -20,7 +20,7 @@
print("Initializing demo template database...")
template_db_name = settings.MATHESAR_DEMO_TEMPLATE
- django_model = Database.current_objects.get(name=settings.DATABASES["default"]["NAME"])
+ django_model = Database.create_from_settings_key("default")
root_engine = create_mathesar_engine(django_model)
with root_engine.connect() as conn:
conn.execution_options(isolation_level="AUTOCOMMIT")
diff --git a/demo/settings.py b/demo/settings.py
--- a/demo/settings.py
+++ b/demo/settings.py
@@ -1,5 +1,4 @@
-from config.settings.production import * # noqa
-from config.settings import * # noqa
+from config.settings.common_settings import * # noqa
from decouple import config as decouple_config
INSTALLED_APPS += [ # noqa
@@ -10,6 +9,7 @@
"demo.middleware.LiveDemoModeMiddleware",
]
+MATHESAR_MODE = 'PRODUCTION'
MATHESAR_LIVE_DEMO = True
MATHESAR_LIVE_DEMO_USERNAME = decouple_config('MATHESAR_LIVE_DEMO_USERNAME', default=None)
MATHESAR_LIVE_DEMO_PASSWORD = decouple_config('MATHESAR_LIVE_DEMO_PASSWORD', default=None)
diff --git a/mathesar/install.py b/mathesar/install.py
--- a/mathesar/install.py
+++ b/mathesar/install.py
@@ -37,6 +37,14 @@
install_on_db_with_key(database_key, skip_confirm)
except IntegrityError:
continue
+ if getattr(settings, 'MATHESAR_LIVE_DEMO', False) is True:
+ management.call_command(
+ 'createsuperuser',
+ '--no-input',
+ '--username', 'demo',
+ '--email', '[email protected]',
+ )
+ management.call_command('setup_demo_template_db')
def install_on_db_with_key(database_key, skip_confirm):
| {"golden_diff": "diff --git a/demo/management/commands/setup_demo_template_db.py b/demo/management/commands/setup_demo_template_db.py\n--- a/demo/management/commands/setup_demo_template_db.py\n+++ b/demo/management/commands/setup_demo_template_db.py\n@@ -20,7 +20,7 @@\n print(\"Initializing demo template database...\")\n \n template_db_name = settings.MATHESAR_DEMO_TEMPLATE\n- django_model = Database.current_objects.get(name=settings.DATABASES[\"default\"][\"NAME\"])\n+ django_model = Database.create_from_settings_key(\"default\")\n root_engine = create_mathesar_engine(django_model)\n with root_engine.connect() as conn:\n conn.execution_options(isolation_level=\"AUTOCOMMIT\")\ndiff --git a/demo/settings.py b/demo/settings.py\n--- a/demo/settings.py\n+++ b/demo/settings.py\n@@ -1,5 +1,4 @@\n-from config.settings.production import * # noqa\n-from config.settings import * # noqa\n+from config.settings.common_settings import * # noqa\n from decouple import config as decouple_config\n \n INSTALLED_APPS += [ # noqa\n@@ -10,6 +9,7 @@\n \"demo.middleware.LiveDemoModeMiddleware\",\n ]\n \n+MATHESAR_MODE = 'PRODUCTION'\n MATHESAR_LIVE_DEMO = True\n MATHESAR_LIVE_DEMO_USERNAME = decouple_config('MATHESAR_LIVE_DEMO_USERNAME', default=None)\n MATHESAR_LIVE_DEMO_PASSWORD = decouple_config('MATHESAR_LIVE_DEMO_PASSWORD', default=None)\ndiff --git a/mathesar/install.py b/mathesar/install.py\n--- a/mathesar/install.py\n+++ b/mathesar/install.py\n@@ -37,6 +37,14 @@\n install_on_db_with_key(database_key, skip_confirm)\n except IntegrityError:\n continue\n+ if getattr(settings, 'MATHESAR_LIVE_DEMO', False) is True:\n+ management.call_command(\n+ 'createsuperuser',\n+ '--no-input',\n+ '--username', 'demo',\n+ '--email', '[email protected]',\n+ )\n+ management.call_command('setup_demo_template_db')\n \n \n def install_on_db_with_key(database_key, skip_confirm):\n", "issue": "Demo mode results in 5XX\n## Description\r\n<!-- A clear and concise description of what the bug is. -->\r\n\r\nMathesar is broken (as of 0.1.4) for Demo Mode. It doesn't load, and just says \"Server Error (500)\" instead.\r\n\r\n## Expected behavior\r\n<!-- A clear and concise description of what you expected to happen. -->\r\n\r\nDemo mode should work.\r\n\r\n## To Reproduce\r\n<!-- How can we recreate this bug? Please try to provide a Minimal, Complete, and Verifiable (http://stackoverflow.com/help/mcve) example if code-related. -->\r\n\r\nChange the `.env` file according to the comment to use the demo mode settings, try to build and start mathesar (dev environment is fine)\n", "before_files": [{"content": "\"\"\"\nThis script installs functions and types for Mathesar onto the configured DB.\n\"\"\"\nimport getopt\nimport os\nimport sys\n\nimport django\nfrom django.core import management\nfrom decouple import config as decouple_config\nfrom django.conf import settings\nfrom django.db.utils import IntegrityError\nfrom sqlalchemy.exc import OperationalError\nfrom db import install\n\n\ndef main(skip_static_collection=False):\n # skip_confirm is temporarily enabled by default as we don't have any use\n # for interactive prompts with docker only deployments\n skip_confirm = True\n (opts, _) = getopt.getopt(sys.argv[1:], \":s\", [\"skip-confirm\"])\n for (opt, value) in opts:\n if (opt == \"-s\") or (opt == \"--skip-confirm\"):\n skip_confirm = True\n os.environ.setdefault(\"DJANGO_SETTINGS_MODULE\", \"config.settings.production\")\n django.setup()\n management.call_command('migrate')\n debug_mode = decouple_config('DEBUG', default=False, cast=bool)\n #\n if not debug_mode and not skip_static_collection:\n management.call_command('collectstatic', '--noinput', '--clear')\n print(\"------------Setting up User Databases------------\")\n django_db_key = decouple_config('DJANGO_DATABASE_KEY', default=\"default\")\n user_databases = [key for key in settings.DATABASES if key != django_db_key]\n for database_key in user_databases:\n try:\n install_on_db_with_key(database_key, skip_confirm)\n except IntegrityError:\n continue\n\n\ndef install_on_db_with_key(database_key, skip_confirm):\n from mathesar.models.base import Database\n db_model = Database.create_from_settings_key(database_key)\n db_model.save()\n try:\n install.install_mathesar(\n database_name=db_model.db_name,\n hostname=db_model.host,\n username=db_model.username,\n password=db_model.password,\n port=db_model.port,\n skip_confirm=skip_confirm\n )\n except OperationalError as e:\n db_model.delete()\n raise e\n\n\nif __name__ == \"__main__\":\n main()\n", "path": "mathesar/install.py"}, {"content": "from config.settings.production import * # noqa\nfrom config.settings import * # noqa\nfrom decouple import config as decouple_config\n\nINSTALLED_APPS += [ # noqa\n \"demo\"\n]\n\nMIDDLEWARE += [ # noqa\n \"demo.middleware.LiveDemoModeMiddleware\",\n]\n\nMATHESAR_LIVE_DEMO = True\nMATHESAR_LIVE_DEMO_USERNAME = decouple_config('MATHESAR_LIVE_DEMO_USERNAME', default=None)\nMATHESAR_LIVE_DEMO_PASSWORD = decouple_config('MATHESAR_LIVE_DEMO_PASSWORD', default=None)\n\nMATHESAR_DEMO_TEMPLATE = 'mathesar_demo_template'\nMATHESAR_DEMO_ARXIV_LOG_PATH = decouple_config(\n 'MATHESAR_DEMO_ARXIV_LOG_PATH',\n default='/var/lib/mathesar/demo/arxiv_db_schema_log'\n)\nBASE_TEMPLATE_ADDITIONAL_SCRIPT_TEMPLATES += ['demo/analytics.html'] # noqa\nROOT_URLCONF = \"demo.urls\"\n", "path": "demo/settings.py"}, {"content": "from sqlalchemy import text\nfrom sqlalchemy.exc import OperationalError\nfrom django.conf import settings\nfrom django.core.management import BaseCommand\n\nfrom db.install import install_mathesar\nfrom demo.install.datasets import load_datasets\nfrom mathesar.database.base import create_mathesar_engine\nfrom mathesar.models.base import Database\n\n\nclass Command(BaseCommand):\n help = 'Initialize the demo template database.'\n\n def handle(self, *args, **options):\n _setup_demo_template_db()\n\n\ndef _setup_demo_template_db():\n print(\"Initializing demo template database...\")\n\n template_db_name = settings.MATHESAR_DEMO_TEMPLATE\n django_model = Database.current_objects.get(name=settings.DATABASES[\"default\"][\"NAME\"])\n root_engine = create_mathesar_engine(django_model)\n with root_engine.connect() as conn:\n conn.execution_options(isolation_level=\"AUTOCOMMIT\")\n conn.execute(text(f\"DROP DATABASE IF EXISTS {template_db_name} WITH (FORCE)\"))\n root_engine.dispose()\n db_model, _ = Database.current_objects.get_or_create(\n name=template_db_name,\n defaults={\n 'db_name': template_db_name,\n 'username': django_model.username,\n 'password': django_model.password,\n 'host': django_model.host,\n 'port': django_model.port\n }\n )\n try:\n install_mathesar(\n database_name=template_db_name,\n hostname=db_model.host,\n username=db_model.username,\n password=db_model.password,\n port=db_model.port,\n skip_confirm=True\n )\n except OperationalError as e:\n db_model.delete()\n raise e\n user_engine = create_mathesar_engine(db_model)\n load_datasets(user_engine)\n user_engine.dispose()\n", "path": "demo/management/commands/setup_demo_template_db.py"}], "after_files": [{"content": "\"\"\"\nThis script installs functions and types for Mathesar onto the configured DB.\n\"\"\"\nimport getopt\nimport os\nimport sys\n\nimport django\nfrom django.core import management\nfrom decouple import config as decouple_config\nfrom django.conf import settings\nfrom django.db.utils import IntegrityError\nfrom sqlalchemy.exc import OperationalError\nfrom db import install\n\n\ndef main(skip_static_collection=False):\n # skip_confirm is temporarily enabled by default as we don't have any use\n # for interactive prompts with docker only deployments\n skip_confirm = True\n (opts, _) = getopt.getopt(sys.argv[1:], \":s\", [\"skip-confirm\"])\n for (opt, value) in opts:\n if (opt == \"-s\") or (opt == \"--skip-confirm\"):\n skip_confirm = True\n os.environ.setdefault(\"DJANGO_SETTINGS_MODULE\", \"config.settings.production\")\n django.setup()\n management.call_command('migrate')\n debug_mode = decouple_config('DEBUG', default=False, cast=bool)\n #\n if not debug_mode and not skip_static_collection:\n management.call_command('collectstatic', '--noinput', '--clear')\n print(\"------------Setting up User Databases------------\")\n django_db_key = decouple_config('DJANGO_DATABASE_KEY', default=\"default\")\n user_databases = [key for key in settings.DATABASES if key != django_db_key]\n for database_key in user_databases:\n try:\n install_on_db_with_key(database_key, skip_confirm)\n except IntegrityError:\n continue\n if getattr(settings, 'MATHESAR_LIVE_DEMO', False) is True:\n management.call_command(\n 'createsuperuser',\n '--no-input',\n '--username', 'demo',\n '--email', '[email protected]',\n )\n management.call_command('setup_demo_template_db')\n\n\ndef install_on_db_with_key(database_key, skip_confirm):\n from mathesar.models.base import Database\n db_model = Database.create_from_settings_key(database_key)\n db_model.save()\n try:\n install.install_mathesar(\n database_name=db_model.db_name,\n hostname=db_model.host,\n username=db_model.username,\n password=db_model.password,\n port=db_model.port,\n skip_confirm=skip_confirm\n )\n except OperationalError as e:\n db_model.delete()\n raise e\n\n\nif __name__ == \"__main__\":\n main()\n", "path": "mathesar/install.py"}, {"content": "from config.settings.common_settings import * # noqa\nfrom decouple import config as decouple_config\n\nINSTALLED_APPS += [ # noqa\n \"demo\"\n]\n\nMIDDLEWARE += [ # noqa\n \"demo.middleware.LiveDemoModeMiddleware\",\n]\n\nMATHESAR_MODE = 'PRODUCTION'\nMATHESAR_LIVE_DEMO = True\nMATHESAR_LIVE_DEMO_USERNAME = decouple_config('MATHESAR_LIVE_DEMO_USERNAME', default=None)\nMATHESAR_LIVE_DEMO_PASSWORD = decouple_config('MATHESAR_LIVE_DEMO_PASSWORD', default=None)\n\nMATHESAR_DEMO_TEMPLATE = 'mathesar_demo_template'\nMATHESAR_DEMO_ARXIV_LOG_PATH = decouple_config(\n 'MATHESAR_DEMO_ARXIV_LOG_PATH',\n default='/var/lib/mathesar/demo/arxiv_db_schema_log'\n)\nBASE_TEMPLATE_ADDITIONAL_SCRIPT_TEMPLATES += ['demo/analytics.html'] # noqa\nROOT_URLCONF = \"demo.urls\"\n", "path": "demo/settings.py"}, {"content": "from sqlalchemy import text\nfrom sqlalchemy.exc import OperationalError\nfrom django.conf import settings\nfrom django.core.management import BaseCommand\n\nfrom db.install import install_mathesar\nfrom demo.install.datasets import load_datasets\nfrom mathesar.database.base import create_mathesar_engine\nfrom mathesar.models.base import Database\n\n\nclass Command(BaseCommand):\n help = 'Initialize the demo template database.'\n\n def handle(self, *args, **options):\n _setup_demo_template_db()\n\n\ndef _setup_demo_template_db():\n print(\"Initializing demo template database...\")\n\n template_db_name = settings.MATHESAR_DEMO_TEMPLATE\n django_model = Database.create_from_settings_key(\"default\")\n root_engine = create_mathesar_engine(django_model)\n with root_engine.connect() as conn:\n conn.execution_options(isolation_level=\"AUTOCOMMIT\")\n conn.execute(text(f\"DROP DATABASE IF EXISTS {template_db_name} WITH (FORCE)\"))\n root_engine.dispose()\n db_model, _ = Database.current_objects.get_or_create(\n name=template_db_name,\n defaults={\n 'db_name': template_db_name,\n 'username': django_model.username,\n 'password': django_model.password,\n 'host': django_model.host,\n 'port': django_model.port\n }\n )\n try:\n install_mathesar(\n database_name=template_db_name,\n hostname=db_model.host,\n username=db_model.username,\n password=db_model.password,\n port=db_model.port,\n skip_confirm=True\n )\n except OperationalError as e:\n db_model.delete()\n raise e\n user_engine = create_mathesar_engine(db_model)\n load_datasets(user_engine)\n user_engine.dispose()\n", "path": "demo/management/commands/setup_demo_template_db.py"}]} | 1,714 | 481 |
gh_patches_debug_35751 | rasdani/github-patches | git_diff | litestar-org__litestar-2810 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Bug: Exception in event listener breaks listener
### Description
In a separate production code base, we encountered what I believe to be an issue/behavior that could be fixed/improved and did not find anything other references of it when searching the issues here.
Basically, if an exception occurs inside of an event listener you will then see
```log
INFO: ASGI 'lifespan' protocol appears unsupported.
```
and the listener stream is closed which then prevents any additional events from being picked up. Any time `request.app.emit` is then called you will get `ClosedResourceError`.
I don't think a single exception inside an event execution should prevent all future events from getting picked up or at least the error should be a little more obvious as to what is going on.
I've created an example repo that recreates the issue consistently.
### URL to code causing the issue
https://github.com/bnjmn/litestar-emit-fail-example
### MCVE
See [full example here](https://github.com/bnjmn/litestar-emit-fail-example/tree/main)
```python
from litestar import Litestar, get, Request
from litestar.events import listener
import logging
logger = logging.getLogger(__name__)
@listener("raise_exception")
async def raise_exception_if_odd(value) -> None:
"""Raise an exception to test Emit error."""
if value is not None and value % 2 != 0:
raise ValueError(f"{value} is odd")
else:
return "The value is even. No exception raised."
@get("/")
async def index() -> str:
return "Hello, world!"
@get("/check-value/{value:int}")
async def check_value(request: Request, value: int) -> str:
try:
request.app.emit("raise_exception", value)
return f"Checked {value}: No exception raised."
except ValueError as e:
return str(e)
app = Litestar([index, check_value], listeners=[raise_exception_if_odd])
```
### Steps to reproduce
1. Follow the steps outlined in the [README](https://github.com/bnjmn/litestar-emit-fail-example/blob/main/README.md#run-it)
```bash
pipenv install
litestar run -d
```
- Go to http://localhost:8000/check-value/2 to see a successful request
- Go to http://localhost:8000/check-value/3 to see an exception raised in the listener
- Attempt to go to http://localhost:8000/check-value/2 again, but you should get ClosedResourceError
### Litestar Version
Recreated on both 2.3.2 and 2.4.1
### Platform
- [ ] Linux
- [X] Mac
- [ ] Windows
- [x] Other (Please specify in the description above)
<!-- POLAR PLEDGE BADGE START -->
---
> [!NOTE]
> While we are open for sponsoring on [GitHub Sponsors](https://github.com/sponsors/litestar-org/) and
> [OpenCollective](https://opencollective.com/litestar), we also utilize [Polar.sh](https://polar.sh/) to engage in pledge-based sponsorship.
>
> Check out all issues funded or available for funding [on our Polar.sh Litestar dashboard](https://polar.sh/litestar-org)
> * If you would like to see an issue prioritized, make a pledge towards it!
> * We receive the pledge once the issue is completed & verified
> * This, along with engagement in the community, helps us know which features are a priority to our users.
<a href="https://polar.sh/litestar-org/litestar/issues/2809">
<picture>
<source media="(prefers-color-scheme: dark)" srcset="https://polar.sh/api/github/litestar-org/litestar/issues/2809/pledge.svg?darkmode=1">
<img alt="Fund with Polar" src="https://polar.sh/api/github/litestar-org/litestar/issues/2809/pledge.svg">
</picture>
</a>
<!-- POLAR PLEDGE BADGE END -->
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `litestar/events/emitter.py`
Content:
```
1 from __future__ import annotations
2
3 import math
4 import sys
5 from abc import ABC, abstractmethod
6 from collections import defaultdict
7 from contextlib import AsyncExitStack
8 from functools import partial
9 from typing import TYPE_CHECKING, Any, Sequence
10
11 if sys.version_info < (3, 9):
12 from typing import AsyncContextManager
13 else:
14 from contextlib import AbstractAsyncContextManager as AsyncContextManager
15
16 import anyio
17
18 from litestar.exceptions import ImproperlyConfiguredException
19
20 __all__ = ("BaseEventEmitterBackend", "SimpleEventEmitter")
21
22
23 if TYPE_CHECKING:
24 from types import TracebackType
25
26 from anyio.streams.memory import MemoryObjectReceiveStream, MemoryObjectSendStream
27
28 from litestar.events.listener import EventListener
29
30
31 class BaseEventEmitterBackend(AsyncContextManager["BaseEventEmitterBackend"], ABC):
32 """Abstract class used to define event emitter backends."""
33
34 __slots__ = ("listeners",)
35
36 listeners: defaultdict[str, set[EventListener]]
37
38 def __init__(self, listeners: Sequence[EventListener]) -> None:
39 """Create an event emitter instance.
40
41 Args:
42 listeners: A list of listeners.
43 """
44 self.listeners = defaultdict(set)
45 for listener in listeners:
46 for event_id in listener.event_ids:
47 self.listeners[event_id].add(listener)
48
49 @abstractmethod
50 def emit(self, event_id: str, *args: Any, **kwargs: Any) -> None:
51 """Emit an event to all attached listeners.
52
53 Args:
54 event_id: The ID of the event to emit, e.g 'my_event'.
55 *args: args to pass to the listener(s).
56 **kwargs: kwargs to pass to the listener(s)
57
58 Returns:
59 None
60 """
61 raise NotImplementedError("not implemented")
62
63
64 class SimpleEventEmitter(BaseEventEmitterBackend):
65 """Event emitter the works only in the current process"""
66
67 __slots__ = ("_queue", "_exit_stack", "_receive_stream", "_send_stream")
68
69 def __init__(self, listeners: Sequence[EventListener]) -> None:
70 """Create an event emitter instance.
71
72 Args:
73 listeners: A list of listeners.
74 """
75 super().__init__(listeners=listeners)
76 self._receive_stream: MemoryObjectReceiveStream | None = None
77 self._send_stream: MemoryObjectSendStream | None = None
78 self._exit_stack: AsyncExitStack | None = None
79
80 @staticmethod
81 async def _worker(receive_stream: MemoryObjectReceiveStream) -> None:
82 """Run items from ``receive_stream`` in a task group.
83
84 Returns:
85 None
86 """
87 async with receive_stream, anyio.create_task_group() as task_group:
88 async for item in receive_stream:
89 fn, args, kwargs = item
90 if kwargs:
91 fn = partial(fn, **kwargs)
92 task_group.start_soon(fn, *args)
93
94 async def __aenter__(self) -> SimpleEventEmitter:
95 self._exit_stack = AsyncExitStack()
96 send_stream, receive_stream = anyio.create_memory_object_stream(math.inf) # type: ignore[var-annotated]
97 self._send_stream = send_stream
98 task_group = anyio.create_task_group()
99
100 await self._exit_stack.enter_async_context(task_group)
101 await self._exit_stack.enter_async_context(send_stream)
102 task_group.start_soon(self._worker, receive_stream)
103
104 return self
105
106 async def __aexit__(
107 self,
108 exc_type: type[BaseException] | None,
109 exc_val: BaseException | None,
110 exc_tb: TracebackType | None,
111 ) -> None:
112 if self._exit_stack:
113 await self._exit_stack.__aexit__(exc_type, exc_val, exc_tb)
114
115 self._exit_stack = None
116 self._send_stream = None
117
118 def emit(self, event_id: str, *args: Any, **kwargs: Any) -> None:
119 """Emit an event to all attached listeners.
120
121 Args:
122 event_id: The ID of the event to emit, e.g 'my_event'.
123 *args: args to pass to the listener(s).
124 **kwargs: kwargs to pass to the listener(s)
125
126 Returns:
127 None
128 """
129 if not (self._send_stream and self._exit_stack):
130 raise RuntimeError("Emitter not initialized")
131
132 if listeners := self.listeners.get(event_id):
133 for listener in listeners:
134 self._send_stream.send_nowait((listener.fn, args, kwargs))
135 return
136 raise ImproperlyConfiguredException(f"no event listeners are registered for event ID: {event_id}")
137
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/litestar/events/emitter.py b/litestar/events/emitter.py
--- a/litestar/events/emitter.py
+++ b/litestar/events/emitter.py
@@ -1,5 +1,6 @@
from __future__ import annotations
+import logging
import math
import sys
from abc import ABC, abstractmethod
@@ -17,9 +18,6 @@
from litestar.exceptions import ImproperlyConfiguredException
-__all__ = ("BaseEventEmitterBackend", "SimpleEventEmitter")
-
-
if TYPE_CHECKING:
from types import TracebackType
@@ -27,6 +25,10 @@
from litestar.events.listener import EventListener
+__all__ = ("BaseEventEmitterBackend", "SimpleEventEmitter")
+
+logger = logging.getLogger(__name__)
+
class BaseEventEmitterBackend(AsyncContextManager["BaseEventEmitterBackend"], ABC):
"""Abstract class used to define event emitter backends."""
@@ -77,19 +79,25 @@
self._send_stream: MemoryObjectSendStream | None = None
self._exit_stack: AsyncExitStack | None = None
- @staticmethod
- async def _worker(receive_stream: MemoryObjectReceiveStream) -> None:
+ async def _worker(self, receive_stream: MemoryObjectReceiveStream) -> None:
"""Run items from ``receive_stream`` in a task group.
Returns:
None
"""
- async with receive_stream, anyio.create_task_group() as task_group:
+ async with receive_stream:
async for item in receive_stream:
- fn, args, kwargs = item
+ await self._run_listener_in_task_group(*item)
+
+ @staticmethod
+ async def _run_listener_in_task_group(fn: Any, args: tuple[Any], kwargs: dict[str, Any]) -> None:
+ try:
+ async with anyio.create_task_group() as task_group:
if kwargs:
fn = partial(fn, **kwargs)
task_group.start_soon(fn, *args)
+ except Exception as exc:
+ logger.exception("Error in event listener: %s", exc)
async def __aenter__(self) -> SimpleEventEmitter:
self._exit_stack = AsyncExitStack()
| {"golden_diff": "diff --git a/litestar/events/emitter.py b/litestar/events/emitter.py\n--- a/litestar/events/emitter.py\n+++ b/litestar/events/emitter.py\n@@ -1,5 +1,6 @@\n from __future__ import annotations\n \n+import logging\n import math\n import sys\n from abc import ABC, abstractmethod\n@@ -17,9 +18,6 @@\n \n from litestar.exceptions import ImproperlyConfiguredException\n \n-__all__ = (\"BaseEventEmitterBackend\", \"SimpleEventEmitter\")\n-\n-\n if TYPE_CHECKING:\n from types import TracebackType\n \n@@ -27,6 +25,10 @@\n \n from litestar.events.listener import EventListener\n \n+__all__ = (\"BaseEventEmitterBackend\", \"SimpleEventEmitter\")\n+\n+logger = logging.getLogger(__name__)\n+\n \n class BaseEventEmitterBackend(AsyncContextManager[\"BaseEventEmitterBackend\"], ABC):\n \"\"\"Abstract class used to define event emitter backends.\"\"\"\n@@ -77,19 +79,25 @@\n self._send_stream: MemoryObjectSendStream | None = None\n self._exit_stack: AsyncExitStack | None = None\n \n- @staticmethod\n- async def _worker(receive_stream: MemoryObjectReceiveStream) -> None:\n+ async def _worker(self, receive_stream: MemoryObjectReceiveStream) -> None:\n \"\"\"Run items from ``receive_stream`` in a task group.\n \n Returns:\n None\n \"\"\"\n- async with receive_stream, anyio.create_task_group() as task_group:\n+ async with receive_stream:\n async for item in receive_stream:\n- fn, args, kwargs = item\n+ await self._run_listener_in_task_group(*item)\n+\n+ @staticmethod\n+ async def _run_listener_in_task_group(fn: Any, args: tuple[Any], kwargs: dict[str, Any]) -> None:\n+ try:\n+ async with anyio.create_task_group() as task_group:\n if kwargs:\n fn = partial(fn, **kwargs)\n task_group.start_soon(fn, *args)\n+ except Exception as exc:\n+ logger.exception(\"Error in event listener: %s\", exc)\n \n async def __aenter__(self) -> SimpleEventEmitter:\n self._exit_stack = AsyncExitStack()\n", "issue": "Bug: Exception in event listener breaks listener\n### Description\r\n\r\nIn a separate production code base, we encountered what I believe to be an issue/behavior that could be fixed/improved and did not find anything other references of it when searching the issues here. \r\n\r\nBasically, if an exception occurs inside of an event listener you will then see\r\n```log\r\nINFO: ASGI 'lifespan' protocol appears unsupported.\r\n```\r\n and the listener stream is closed which then prevents any additional events from being picked up. Any time `request.app.emit` is then called you will get `ClosedResourceError`.\r\n\r\nI don't think a single exception inside an event execution should prevent all future events from getting picked up or at least the error should be a little more obvious as to what is going on.\r\n\r\nI've created an example repo that recreates the issue consistently.\r\n\r\n### URL to code causing the issue\r\n\r\nhttps://github.com/bnjmn/litestar-emit-fail-example\r\n\r\n### MCVE\r\n\r\nSee [full example here](https://github.com/bnjmn/litestar-emit-fail-example/tree/main)\r\n```python\r\nfrom litestar import Litestar, get, Request\r\nfrom litestar.events import listener\r\n\r\nimport logging\r\n\r\nlogger = logging.getLogger(__name__)\r\n\r\n@listener(\"raise_exception\")\r\nasync def raise_exception_if_odd(value) -> None:\r\n \"\"\"Raise an exception to test Emit error.\"\"\"\r\n if value is not None and value % 2 != 0:\r\n raise ValueError(f\"{value} is odd\")\r\n else:\r\n return \"The value is even. No exception raised.\"\r\n\r\n@get(\"/\")\r\nasync def index() -> str:\r\n return \"Hello, world!\"\r\n\r\n@get(\"/check-value/{value:int}\")\r\nasync def check_value(request: Request, value: int) -> str:\r\n try:\r\n request.app.emit(\"raise_exception\", value)\r\n return f\"Checked {value}: No exception raised.\"\r\n except ValueError as e:\r\n return str(e)\r\n\r\napp = Litestar([index, check_value], listeners=[raise_exception_if_odd])\r\n```\r\n\r\n\r\n### Steps to reproduce\r\n\r\n1. Follow the steps outlined in the [README](https://github.com/bnjmn/litestar-emit-fail-example/blob/main/README.md#run-it)\r\n```bash\r\npipenv install\r\nlitestar run -d\r\n```\r\n- Go to http://localhost:8000/check-value/2 to see a successful request\r\n- Go to http://localhost:8000/check-value/3 to see an exception raised in the listener\r\n- Attempt to go to http://localhost:8000/check-value/2 again, but you should get ClosedResourceError\r\n\r\n### Litestar Version\r\n\r\nRecreated on both 2.3.2 and 2.4.1\r\n\r\n### Platform\r\n\r\n- [ ] Linux\r\n- [X] Mac\r\n- [ ] Windows\r\n- [x] Other (Please specify in the description above)\r\n\r\n<!-- POLAR PLEDGE BADGE START -->\r\n---\r\n> [!NOTE] \r\n> While we are open for sponsoring on [GitHub Sponsors](https://github.com/sponsors/litestar-org/) and \r\n> [OpenCollective](https://opencollective.com/litestar), we also utilize [Polar.sh](https://polar.sh/) to engage in pledge-based sponsorship.\r\n>\r\n> Check out all issues funded or available for funding [on our Polar.sh Litestar dashboard](https://polar.sh/litestar-org)\r\n> * If you would like to see an issue prioritized, make a pledge towards it!\r\n> * We receive the pledge once the issue is completed & verified\r\n> * This, along with engagement in the community, helps us know which features are a priority to our users.\r\n\r\n<a href=\"https://polar.sh/litestar-org/litestar/issues/2809\">\r\n<picture>\r\n <source media=\"(prefers-color-scheme: dark)\" srcset=\"https://polar.sh/api/github/litestar-org/litestar/issues/2809/pledge.svg?darkmode=1\">\r\n <img alt=\"Fund with Polar\" src=\"https://polar.sh/api/github/litestar-org/litestar/issues/2809/pledge.svg\">\r\n</picture>\r\n</a>\r\n<!-- POLAR PLEDGE BADGE END -->\r\n\n", "before_files": [{"content": "from __future__ import annotations\n\nimport math\nimport sys\nfrom abc import ABC, abstractmethod\nfrom collections import defaultdict\nfrom contextlib import AsyncExitStack\nfrom functools import partial\nfrom typing import TYPE_CHECKING, Any, Sequence\n\nif sys.version_info < (3, 9):\n from typing import AsyncContextManager\nelse:\n from contextlib import AbstractAsyncContextManager as AsyncContextManager\n\nimport anyio\n\nfrom litestar.exceptions import ImproperlyConfiguredException\n\n__all__ = (\"BaseEventEmitterBackend\", \"SimpleEventEmitter\")\n\n\nif TYPE_CHECKING:\n from types import TracebackType\n\n from anyio.streams.memory import MemoryObjectReceiveStream, MemoryObjectSendStream\n\n from litestar.events.listener import EventListener\n\n\nclass BaseEventEmitterBackend(AsyncContextManager[\"BaseEventEmitterBackend\"], ABC):\n \"\"\"Abstract class used to define event emitter backends.\"\"\"\n\n __slots__ = (\"listeners\",)\n\n listeners: defaultdict[str, set[EventListener]]\n\n def __init__(self, listeners: Sequence[EventListener]) -> None:\n \"\"\"Create an event emitter instance.\n\n Args:\n listeners: A list of listeners.\n \"\"\"\n self.listeners = defaultdict(set)\n for listener in listeners:\n for event_id in listener.event_ids:\n self.listeners[event_id].add(listener)\n\n @abstractmethod\n def emit(self, event_id: str, *args: Any, **kwargs: Any) -> None:\n \"\"\"Emit an event to all attached listeners.\n\n Args:\n event_id: The ID of the event to emit, e.g 'my_event'.\n *args: args to pass to the listener(s).\n **kwargs: kwargs to pass to the listener(s)\n\n Returns:\n None\n \"\"\"\n raise NotImplementedError(\"not implemented\")\n\n\nclass SimpleEventEmitter(BaseEventEmitterBackend):\n \"\"\"Event emitter the works only in the current process\"\"\"\n\n __slots__ = (\"_queue\", \"_exit_stack\", \"_receive_stream\", \"_send_stream\")\n\n def __init__(self, listeners: Sequence[EventListener]) -> None:\n \"\"\"Create an event emitter instance.\n\n Args:\n listeners: A list of listeners.\n \"\"\"\n super().__init__(listeners=listeners)\n self._receive_stream: MemoryObjectReceiveStream | None = None\n self._send_stream: MemoryObjectSendStream | None = None\n self._exit_stack: AsyncExitStack | None = None\n\n @staticmethod\n async def _worker(receive_stream: MemoryObjectReceiveStream) -> None:\n \"\"\"Run items from ``receive_stream`` in a task group.\n\n Returns:\n None\n \"\"\"\n async with receive_stream, anyio.create_task_group() as task_group:\n async for item in receive_stream:\n fn, args, kwargs = item\n if kwargs:\n fn = partial(fn, **kwargs)\n task_group.start_soon(fn, *args)\n\n async def __aenter__(self) -> SimpleEventEmitter:\n self._exit_stack = AsyncExitStack()\n send_stream, receive_stream = anyio.create_memory_object_stream(math.inf) # type: ignore[var-annotated]\n self._send_stream = send_stream\n task_group = anyio.create_task_group()\n\n await self._exit_stack.enter_async_context(task_group)\n await self._exit_stack.enter_async_context(send_stream)\n task_group.start_soon(self._worker, receive_stream)\n\n return self\n\n async def __aexit__(\n self,\n exc_type: type[BaseException] | None,\n exc_val: BaseException | None,\n exc_tb: TracebackType | None,\n ) -> None:\n if self._exit_stack:\n await self._exit_stack.__aexit__(exc_type, exc_val, exc_tb)\n\n self._exit_stack = None\n self._send_stream = None\n\n def emit(self, event_id: str, *args: Any, **kwargs: Any) -> None:\n \"\"\"Emit an event to all attached listeners.\n\n Args:\n event_id: The ID of the event to emit, e.g 'my_event'.\n *args: args to pass to the listener(s).\n **kwargs: kwargs to pass to the listener(s)\n\n Returns:\n None\n \"\"\"\n if not (self._send_stream and self._exit_stack):\n raise RuntimeError(\"Emitter not initialized\")\n\n if listeners := self.listeners.get(event_id):\n for listener in listeners:\n self._send_stream.send_nowait((listener.fn, args, kwargs))\n return\n raise ImproperlyConfiguredException(f\"no event listeners are registered for event ID: {event_id}\")\n", "path": "litestar/events/emitter.py"}], "after_files": [{"content": "from __future__ import annotations\n\nimport logging\nimport math\nimport sys\nfrom abc import ABC, abstractmethod\nfrom collections import defaultdict\nfrom contextlib import AsyncExitStack\nfrom functools import partial\nfrom typing import TYPE_CHECKING, Any, Sequence\n\nif sys.version_info < (3, 9):\n from typing import AsyncContextManager\nelse:\n from contextlib import AbstractAsyncContextManager as AsyncContextManager\n\nimport anyio\n\nfrom litestar.exceptions import ImproperlyConfiguredException\n\nif TYPE_CHECKING:\n from types import TracebackType\n\n from anyio.streams.memory import MemoryObjectReceiveStream, MemoryObjectSendStream\n\n from litestar.events.listener import EventListener\n\n__all__ = (\"BaseEventEmitterBackend\", \"SimpleEventEmitter\")\n\nlogger = logging.getLogger(__name__)\n\n\nclass BaseEventEmitterBackend(AsyncContextManager[\"BaseEventEmitterBackend\"], ABC):\n \"\"\"Abstract class used to define event emitter backends.\"\"\"\n\n __slots__ = (\"listeners\",)\n\n listeners: defaultdict[str, set[EventListener]]\n\n def __init__(self, listeners: Sequence[EventListener]) -> None:\n \"\"\"Create an event emitter instance.\n\n Args:\n listeners: A list of listeners.\n \"\"\"\n self.listeners = defaultdict(set)\n for listener in listeners:\n for event_id in listener.event_ids:\n self.listeners[event_id].add(listener)\n\n @abstractmethod\n def emit(self, event_id: str, *args: Any, **kwargs: Any) -> None:\n \"\"\"Emit an event to all attached listeners.\n\n Args:\n event_id: The ID of the event to emit, e.g 'my_event'.\n *args: args to pass to the listener(s).\n **kwargs: kwargs to pass to the listener(s)\n\n Returns:\n None\n \"\"\"\n raise NotImplementedError(\"not implemented\")\n\n\nclass SimpleEventEmitter(BaseEventEmitterBackend):\n \"\"\"Event emitter the works only in the current process\"\"\"\n\n __slots__ = (\"_queue\", \"_exit_stack\", \"_receive_stream\", \"_send_stream\")\n\n def __init__(self, listeners: Sequence[EventListener]) -> None:\n \"\"\"Create an event emitter instance.\n\n Args:\n listeners: A list of listeners.\n \"\"\"\n super().__init__(listeners=listeners)\n self._receive_stream: MemoryObjectReceiveStream | None = None\n self._send_stream: MemoryObjectSendStream | None = None\n self._exit_stack: AsyncExitStack | None = None\n\n async def _worker(self, receive_stream: MemoryObjectReceiveStream) -> None:\n \"\"\"Run items from ``receive_stream`` in a task group.\n\n Returns:\n None\n \"\"\"\n async with receive_stream:\n async for item in receive_stream:\n await self._run_listener_in_task_group(*item)\n\n @staticmethod\n async def _run_listener_in_task_group(fn: Any, args: tuple[Any], kwargs: dict[str, Any]) -> None:\n try:\n async with anyio.create_task_group() as task_group:\n if kwargs:\n fn = partial(fn, **kwargs)\n task_group.start_soon(fn, *args)\n except Exception as exc:\n logger.exception(\"Error in event listener: %s\", exc)\n\n async def __aenter__(self) -> SimpleEventEmitter:\n self._exit_stack = AsyncExitStack()\n send_stream, receive_stream = anyio.create_memory_object_stream(math.inf) # type: ignore[var-annotated]\n self._send_stream = send_stream\n task_group = anyio.create_task_group()\n\n await self._exit_stack.enter_async_context(task_group)\n await self._exit_stack.enter_async_context(send_stream)\n task_group.start_soon(self._worker, receive_stream)\n\n return self\n\n async def __aexit__(\n self,\n exc_type: type[BaseException] | None,\n exc_val: BaseException | None,\n exc_tb: TracebackType | None,\n ) -> None:\n if self._exit_stack:\n await self._exit_stack.__aexit__(exc_type, exc_val, exc_tb)\n\n self._exit_stack = None\n self._send_stream = None\n\n def emit(self, event_id: str, *args: Any, **kwargs: Any) -> None:\n \"\"\"Emit an event to all attached listeners.\n\n Args:\n event_id: The ID of the event to emit, e.g 'my_event'.\n *args: args to pass to the listener(s).\n **kwargs: kwargs to pass to the listener(s)\n\n Returns:\n None\n \"\"\"\n if not (self._send_stream and self._exit_stack):\n raise RuntimeError(\"Emitter not initialized\")\n\n if listeners := self.listeners.get(event_id):\n for listener in listeners:\n self._send_stream.send_nowait((listener.fn, args, kwargs))\n return\n raise ImproperlyConfiguredException(f\"no event listeners are registered for event ID: {event_id}\")\n", "path": "litestar/events/emitter.py"}]} | 2,464 | 498 |
gh_patches_debug_38209 | rasdani/github-patches | git_diff | digitalfabrik__integreat-cms-445 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Improve documentation of cms rules
Explain the rules module and how it interacts with our permission management. Add docstrings of the following format:
```
"""
[Summary]
:param [ParamName]: [ParamDescription], defaults to [DefaultParamVal]
:type [ParamName]: [ParamType](, optional)
...
:raises [ErrorType]: [ErrorDescription]
...
:return: [ReturnDescription]
:rtype: [ReturnType]
"""
```
Improve documentation of cms rules
Explain the rules module and how it interacts with our permission management. Add docstrings of the following format:
```
"""
[Summary]
:param [ParamName]: [ParamDescription], defaults to [DefaultParamVal]
:type [ParamName]: [ParamType](, optional)
...
:raises [ErrorType]: [ErrorDescription]
...
:return: [ReturnDescription]
:rtype: [ReturnType]
"""
```
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `src/cms/rules.py`
Content:
```
1 from rules import add_perm, predicate
2
3
4 # Predicates
5
6 @predicate
7 def is_page_editor(user, page):
8 if not page:
9 return False
10 return user in page.editors.all()
11
12 @predicate
13 def is_page_publisher(user, page):
14 if not page:
15 return False
16 return user in page.publishers.all()
17
18 @predicate
19 # pylint: disable=unused-argument
20 def can_edit_all_pages(user, page):
21 return user.has_perm('cms.edit_pages')
22
23 @predicate
24 # pylint: disable=unused-argument
25 def can_publish_all_pages(user, page):
26 return user.has_perm('cms.publish_pages')
27
28
29 # Permissions
30
31 add_perm('cms.edit_page', can_edit_all_pages | is_page_editor | can_publish_all_pages | is_page_publisher)
32 add_perm('cms.publish_page', can_publish_all_pages | is_page_publisher)
33
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/src/cms/rules.py b/src/cms/rules.py
--- a/src/cms/rules.py
+++ b/src/cms/rules.py
@@ -1,3 +1,22 @@
+"""
+We use `django-rules <https://pypi.org/project/rules/>`_ to add custom permissions for specific pages.
+
+For a given user and page, the following permissions are added:
+
+* ``cms.edit_page`` if one of the following predicates return true:
+
+ * :func:`~cms.rules.can_edit_all_pages`
+ * :func:`~cms.rules.is_page_editor`
+ * :func:`~cms.rules.can_publish_all_pages`
+ * :func:`~cms.rules.is_page_publisher`
+
+* ``cms.publish_page`` if one of the following predicates return true:
+
+ * :func:`~cms.rules.can_publish_all_pages`
+ * :func:`~cms.rules.is_page_publisher`
+
+See the project's `README <https://github.com/dfunckt/django-rules/blob/master/README.rst>`_ to learn more.
+"""
from rules import add_perm, predicate
@@ -5,12 +24,36 @@
@predicate
def is_page_editor(user, page):
+ """
+ This predicate checks whether the given user is one of the editors of the given page.
+
+ :param user: The user who's permission should be checked
+ :type user: ~django.contrib.auth.models.User
+
+ :param page: The requested page
+ :type page: ~cms.models.pages.page.Page
+
+ :return: Whether or not ``user`` is an editor of ``page``
+ :rtype: bool
+ """
if not page:
return False
return user in page.editors.all()
@predicate
def is_page_publisher(user, page):
+ """
+ This predicate checks whether the given user is one of the publishers of the given page.
+
+ :param user: The user who's permission should be checked
+ :type user: ~django.contrib.auth.models.User
+
+ :param page: The requested page
+ :type page: ~cms.models.pages.page.Page
+
+ :return: Whether or not ``user`` is a publisher of ``page``
+ :rtype: bool
+ """
if not page:
return False
return user in page.publishers.all()
@@ -18,11 +61,35 @@
@predicate
# pylint: disable=unused-argument
def can_edit_all_pages(user, page):
+ """
+ This predicate checks whether the given user can edit all pages.
+
+ :param user: The user who's permission should be checked
+ :type user: ~django.contrib.auth.models.User
+
+ :param page: Unused page parameter (the function signature must match the other predicates)
+ :type page: ~cms.models.pages.page.Page
+
+ :return: Whether or not ``user`` can edit all pages
+ :rtype: bool
+ """
return user.has_perm('cms.edit_pages')
@predicate
# pylint: disable=unused-argument
def can_publish_all_pages(user, page):
+ """
+ This predicate checks whether the given user can publish all pages.
+
+ :param user: The user who's permission should be checked
+ :type user: ~django.contrib.auth.models.User
+
+ :param page: Unused page parameter (the function signature must match the other predicates)
+ :type page: ~cms.models.pages.page.Page
+
+ :return: Whether or not ``user`` can publish all pages
+ :rtype: bool
+ """
return user.has_perm('cms.publish_pages')
| {"golden_diff": "diff --git a/src/cms/rules.py b/src/cms/rules.py\n--- a/src/cms/rules.py\n+++ b/src/cms/rules.py\n@@ -1,3 +1,22 @@\n+\"\"\"\n+We use `django-rules <https://pypi.org/project/rules/>`_ to add custom permissions for specific pages.\n+\n+For a given user and page, the following permissions are added:\n+\n+* ``cms.edit_page`` if one of the following predicates return true:\n+\n+ * :func:`~cms.rules.can_edit_all_pages`\n+ * :func:`~cms.rules.is_page_editor`\n+ * :func:`~cms.rules.can_publish_all_pages`\n+ * :func:`~cms.rules.is_page_publisher`\n+\n+* ``cms.publish_page`` if one of the following predicates return true:\n+\n+ * :func:`~cms.rules.can_publish_all_pages`\n+ * :func:`~cms.rules.is_page_publisher`\n+\n+See the project's `README <https://github.com/dfunckt/django-rules/blob/master/README.rst>`_ to learn more.\n+\"\"\"\n from rules import add_perm, predicate\n \n \n@@ -5,12 +24,36 @@\n \n @predicate\n def is_page_editor(user, page):\n+ \"\"\"\n+ This predicate checks whether the given user is one of the editors of the given page.\n+\n+ :param user: The user who's permission should be checked\n+ :type user: ~django.contrib.auth.models.User\n+\n+ :param page: The requested page\n+ :type page: ~cms.models.pages.page.Page\n+\n+ :return: Whether or not ``user`` is an editor of ``page``\n+ :rtype: bool\n+ \"\"\"\n if not page:\n return False\n return user in page.editors.all()\n \n @predicate\n def is_page_publisher(user, page):\n+ \"\"\"\n+ This predicate checks whether the given user is one of the publishers of the given page.\n+\n+ :param user: The user who's permission should be checked\n+ :type user: ~django.contrib.auth.models.User\n+\n+ :param page: The requested page\n+ :type page: ~cms.models.pages.page.Page\n+\n+ :return: Whether or not ``user`` is a publisher of ``page``\n+ :rtype: bool\n+ \"\"\"\n if not page:\n return False\n return user in page.publishers.all()\n@@ -18,11 +61,35 @@\n @predicate\n # pylint: disable=unused-argument\n def can_edit_all_pages(user, page):\n+ \"\"\"\n+ This predicate checks whether the given user can edit all pages.\n+\n+ :param user: The user who's permission should be checked\n+ :type user: ~django.contrib.auth.models.User\n+\n+ :param page: Unused page parameter (the function signature must match the other predicates)\n+ :type page: ~cms.models.pages.page.Page\n+\n+ :return: Whether or not ``user`` can edit all pages\n+ :rtype: bool\n+ \"\"\"\n return user.has_perm('cms.edit_pages')\n \n @predicate\n # pylint: disable=unused-argument\n def can_publish_all_pages(user, page):\n+ \"\"\"\n+ This predicate checks whether the given user can publish all pages.\n+\n+ :param user: The user who's permission should be checked\n+ :type user: ~django.contrib.auth.models.User\n+\n+ :param page: Unused page parameter (the function signature must match the other predicates)\n+ :type page: ~cms.models.pages.page.Page\n+\n+ :return: Whether or not ``user`` can publish all pages\n+ :rtype: bool\n+ \"\"\"\n return user.has_perm('cms.publish_pages')\n", "issue": "Improve documentation of cms rules\nExplain the rules module and how it interacts with our permission management. Add docstrings of the following format:\r\n```\r\n\"\"\"\r\n[Summary]\r\n\r\n:param [ParamName]: [ParamDescription], defaults to [DefaultParamVal]\r\n:type [ParamName]: [ParamType](, optional)\r\n...\r\n:raises [ErrorType]: [ErrorDescription]\r\n...\r\n:return: [ReturnDescription]\r\n:rtype: [ReturnType]\r\n\"\"\"\r\n```\nImprove documentation of cms rules\nExplain the rules module and how it interacts with our permission management. Add docstrings of the following format:\r\n```\r\n\"\"\"\r\n[Summary]\r\n\r\n:param [ParamName]: [ParamDescription], defaults to [DefaultParamVal]\r\n:type [ParamName]: [ParamType](, optional)\r\n...\r\n:raises [ErrorType]: [ErrorDescription]\r\n...\r\n:return: [ReturnDescription]\r\n:rtype: [ReturnType]\r\n\"\"\"\r\n```\n", "before_files": [{"content": "from rules import add_perm, predicate\n\n\n# Predicates\n\n@predicate\ndef is_page_editor(user, page):\n if not page:\n return False\n return user in page.editors.all()\n\n@predicate\ndef is_page_publisher(user, page):\n if not page:\n return False\n return user in page.publishers.all()\n\n@predicate\n# pylint: disable=unused-argument\ndef can_edit_all_pages(user, page):\n return user.has_perm('cms.edit_pages')\n\n@predicate\n# pylint: disable=unused-argument\ndef can_publish_all_pages(user, page):\n return user.has_perm('cms.publish_pages')\n\n\n# Permissions\n\nadd_perm('cms.edit_page', can_edit_all_pages | is_page_editor | can_publish_all_pages | is_page_publisher)\nadd_perm('cms.publish_page', can_publish_all_pages | is_page_publisher)\n", "path": "src/cms/rules.py"}], "after_files": [{"content": "\"\"\"\nWe use `django-rules <https://pypi.org/project/rules/>`_ to add custom permissions for specific pages.\n\nFor a given user and page, the following permissions are added:\n\n* ``cms.edit_page`` if one of the following predicates return true:\n\n * :func:`~cms.rules.can_edit_all_pages`\n * :func:`~cms.rules.is_page_editor`\n * :func:`~cms.rules.can_publish_all_pages`\n * :func:`~cms.rules.is_page_publisher`\n\n* ``cms.publish_page`` if one of the following predicates return true:\n\n * :func:`~cms.rules.can_publish_all_pages`\n * :func:`~cms.rules.is_page_publisher`\n\nSee the project's `README <https://github.com/dfunckt/django-rules/blob/master/README.rst>`_ to learn more.\n\"\"\"\nfrom rules import add_perm, predicate\n\n\n# Predicates\n\n@predicate\ndef is_page_editor(user, page):\n \"\"\"\n This predicate checks whether the given user is one of the editors of the given page.\n\n :param user: The user who's permission should be checked\n :type user: ~django.contrib.auth.models.User\n\n :param page: The requested page\n :type page: ~cms.models.pages.page.Page\n\n :return: Whether or not ``user`` is an editor of ``page``\n :rtype: bool\n \"\"\"\n if not page:\n return False\n return user in page.editors.all()\n\n@predicate\ndef is_page_publisher(user, page):\n \"\"\"\n This predicate checks whether the given user is one of the publishers of the given page.\n\n :param user: The user who's permission should be checked\n :type user: ~django.contrib.auth.models.User\n\n :param page: The requested page\n :type page: ~cms.models.pages.page.Page\n\n :return: Whether or not ``user`` is a publisher of ``page``\n :rtype: bool\n \"\"\"\n if not page:\n return False\n return user in page.publishers.all()\n\n@predicate\n# pylint: disable=unused-argument\ndef can_edit_all_pages(user, page):\n \"\"\"\n This predicate checks whether the given user can edit all pages.\n\n :param user: The user who's permission should be checked\n :type user: ~django.contrib.auth.models.User\n\n :param page: Unused page parameter (the function signature must match the other predicates)\n :type page: ~cms.models.pages.page.Page\n\n :return: Whether or not ``user`` can edit all pages\n :rtype: bool\n \"\"\"\n return user.has_perm('cms.edit_pages')\n\n@predicate\n# pylint: disable=unused-argument\ndef can_publish_all_pages(user, page):\n \"\"\"\n This predicate checks whether the given user can publish all pages.\n\n :param user: The user who's permission should be checked\n :type user: ~django.contrib.auth.models.User\n\n :param page: Unused page parameter (the function signature must match the other predicates)\n :type page: ~cms.models.pages.page.Page\n\n :return: Whether or not ``user`` can publish all pages\n :rtype: bool\n \"\"\"\n return user.has_perm('cms.publish_pages')\n\n\n# Permissions\n\nadd_perm('cms.edit_page', can_edit_all_pages | is_page_editor | can_publish_all_pages | is_page_publisher)\nadd_perm('cms.publish_page', can_publish_all_pages | is_page_publisher)\n", "path": "src/cms/rules.py"}]} | 680 | 807 |
gh_patches_debug_28434 | rasdani/github-patches | git_diff | openstates__openstates-scrapers-2400 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
NE failing since at least 2018-06-28
NE has been failing since 2018-06-28
Based on automated runs it appears that NE has not run successfully in 2 days (2018-06-28).
```
02:30:54 INFO pupa: save membership 5b4c2cb2-7b6e-11e8-9e19-02e29baaa692 membership in ~{"classification": "legislature"} as membership_5b4c30ea-7b6e-11e8-9e19-02e29baaa692.json
02:30:54 INFO pupa: save membership 5b4c2cb2-7b6e-11e8-9e19-02e29baaa692 membership in ~{"classification": "party", "name": "Nonpartisan"} as membership_5b4c3400-7b6e-11e8-9e19-02e29baaa692.json
02:30:54 INFO scrapelib: GET - http://news.legislature.ne.gov/dist44
02:30:55 INFO pupa: save person Dan Hughes as person_5bea39ca-7b6e-11e8-9e19-02e29baaa692.json
02:30:55 INFO pupa: save membership 5bea39ca-7b6e-11e8-9e19-02e29baaa692 membership in ~{"classification": "legislature"} as membership_5bea3dda-7b6e-11e8-9e19-02e29baaa692.json
02:30:55 INFO pupa: save membership 5bea39ca-7b6e-11e8-9e19-02e29baaa692 membership in ~{"classification": "party", "name": "Nonpartisan"} as membership_5bea4028-7b6e-11e8-9e19-02e29baaa692.json
02:30:55 INFO scrapelib: GET - http://news.legislature.ne.gov/dist45
02:30:56 INFO pupa: save person Sue Crawford as person_5c938c46-7b6e-11e8-9e19-02e29baaa692.json
02:30:56 INFO pupa: save membership 5c938c46-7b6e-11e8-9e19-02e29baaa692 membership in ~{"classification": "legislature"} as membership_5c93909c-7b6e-11e8-9e19-02e29baaa692.json
02:30:56 INFO pupa: save membership 5c938c46-7b6e-11e8-9e19-02e29baaa692 membership in ~{"classification": "party", "name": "Nonpartisan"} as membership_5c93939e-7b6e-11e8-9e19-02e29baaa692.json
02:30:56 INFO scrapelib: GET - http://news.legislature.ne.gov/dist46
02:30:57 INFO pupa: save person Adam Morfeld as person_5d0a16ea-7b6e-11e8-9e19-02e29baaa692.json
02:30:57 INFO pupa: save membership 5d0a16ea-7b6e-11e8-9e19-02e29baaa692 membership in ~{"classification": "legislature"} as membership_5d0a1ae6-7b6e-11e8-9e19-02e29baaa692.json
02:30:57 INFO pupa: save membership 5d0a16ea-7b6e-11e8-9e19-02e29baaa692 membership in ~{"classification": "party", "name": "Nonpartisan"} as membership_5d0a1d2a-7b6e-11e8-9e19-02e29baaa692.json
02:30:57 INFO scrapelib: GET - http://news.legislature.ne.gov/dist47
02:30:58 INFO pupa: save person Steve Erdman as person_5dc322ac-7b6e-11e8-9e19-02e29baaa692.json
02:30:58 INFO pupa: save membership 5dc322ac-7b6e-11e8-9e19-02e29baaa692 membership in ~{"classification": "legislature"} as membership_5dc32694-7b6e-11e8-9e19-02e29baaa692.json
02:30:58 INFO pupa: save membership 5dc322ac-7b6e-11e8-9e19-02e29baaa692 membership in ~{"classification": "party", "name": "Nonpartisan"} as membership_5dc32afe-7b6e-11e8-9e19-02e29baaa692.json
02:30:58 INFO scrapelib: GET - http://news.legislature.ne.gov/dist48
02:30:59 INFO pupa: save person John Stinner as person_5e89a800-7b6e-11e8-9e19-02e29baaa692.json
02:30:59 INFO pupa: save membership 5e89a800-7b6e-11e8-9e19-02e29baaa692 membership in ~{"classification": "legislature"} as membership_5e89ac4c-7b6e-11e8-9e19-02e29baaa692.json
02:30:59 INFO pupa: save membership 5e89a800-7b6e-11e8-9e19-02e29baaa692 membership in ~{"classification": "party", "name": "Nonpartisan"} as membership_5e89aed6-7b6e-11e8-9e19-02e29baaa692.json
02:30:59 INFO scrapelib: GET - http://news.legislature.ne.gov/dist49
02:31:00 INFO pupa: save person John Murante as person_5ed7988a-7b6e-11e8-9e19-02e29baaa692.json
02:31:00 INFO pupa: save membership 5ed7988a-7b6e-11e8-9e19-02e29baaa692 membership in ~{"classification": "legislature"} as membership_5ed79c90-7b6e-11e8-9e19-02e29baaa692.json
02:31:00 INFO pupa: save membership 5ed7988a-7b6e-11e8-9e19-02e29baaa692 membership in ~{"classification": "party", "name": "Nonpartisan"} as membership_5ed79ede-7b6e-11e8-9e19-02e29baaa692.json
02:31:00 INFO scrapelib: GET - http://www.nebraskalegislature.gov/committees/standing-committees.php
02:31:00 INFO scrapelib: GET - http://www.nebraskalegislature.gov/committees/select-committees.php
02:31:01 WARNING pupa: No members found in Select committees committee.
02:31:01 WARNING pupa: No members found in Special committees committee.
Traceback (most recent call last):
File "/opt/openstates/venv-pupa//bin/pupa", line 11, in <module>
load_entry_point('pupa', 'console_scripts', 'pupa')()
File "/opt/openstates/venv-pupa/src/pupa/pupa/cli/__main__.py", line 68, in main
subcommands[args.subcommand].handle(args, other)
File "/opt/openstates/venv-pupa/src/pupa/pupa/cli/commands/update.py", line 260, in handle
return self.do_handle(args, other, juris)
File "/opt/openstates/venv-pupa/src/pupa/pupa/cli/commands/update.py", line 305, in do_handle
report['scrape'] = self.do_scrape(juris, args, scrapers)
File "/opt/openstates/venv-pupa/src/pupa/pupa/cli/commands/update.py", line 173, in do_scrape
report[scraper_name] = scraper.do_scrape(**scrape_args)
File "/opt/openstates/venv-pupa/src/pupa/pupa/scrape/base.py", line 120, in do_scrape
raise ScrapeError('no objects returned from {} scrape'.format(self.__class__.__name__))
pupa.exceptions.ScrapeError: no objects returned from NECommitteeScraper scrape
loaded Open States pupa settings...
ne (scrape, import)
bills: {}
votes: {}
people: {}
committees: {}
```
Visit http://bobsled.openstates.org for more info.
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `openstates/ne/committees.py`
Content:
```
1 import re
2
3 from pupa.scrape import Scraper, Organization
4
5 from openstates.utils import LXMLMixin
6
7
8 class NECommitteeScraper(Scraper, LXMLMixin):
9 def _scrape_standing_committees(self):
10 """Scrapes the Standing Committees page of the Nebraska state
11 legislature."""
12 main_url = 'http://www.nebraskalegislature.gov/committees/standing-committees.php'
13 page = self.lxmlize(main_url)
14
15 committee_nodes = self.get_nodes(
16 page,
17 '//div[@class="main-content"]/div[@class="panel panel-leg"][1]/'
18 'div[@class="list-group"]/a[@class="list-group-item"]')
19
20 for committee_node in committee_nodes:
21 committee_page_url = committee_node.attrib['href']
22 committee_page = self.lxmlize(committee_page_url)
23
24 name_text = self.get_node(
25 committee_page,
26 '//div[@class="container view-front"]/div[@class="row"]/'
27 'div[@class="col-sm-6 col-md-7"]/h1/text()[normalize-space()]')
28 name = name_text.split()[0:-1]
29
30 committee_name = ''
31 for x in range(len(name)):
32 committee_name += name[x] + ' '
33 committee_name = committee_name[0: -1]
34
35 org = Organization(name=committee_name, chamber='legislature',
36 classification='committee')
37
38 members = self.get_nodes(
39 committee_page,
40 '//div[@class="col-sm-4 col-md-3 ltc-col-right"][1]/'
41 'div[@class="block-box"][1]/ul[@class="list-unstyled '
42 'feature-content"]/li/a/text()[normalize-space()]')
43
44 for member in members:
45 member_name = re.sub(r'\Sen\.\s+', '', member)
46 member_name = re.sub(r', Chairperson', '', member_name).strip()
47 if 'Chairperson' in member:
48 member_role = 'Chairperson'
49 else:
50 member_role = 'member'
51 org.add_member(member_name, member_role)
52
53 org.add_source(main_url)
54 org.add_source(committee_page_url)
55
56 yield org
57
58 def _scrape_select_special_committees(self):
59 """Scrapes the Select and Special Committees page of the
60 Nebraska state legislature."""
61 main_url = 'http://www.nebraskalegislature.gov/committees/select-committees.php'
62 page = self.lxmlize(main_url)
63
64 committee_nodes = self.get_nodes(
65 page,
66 '//div[@class="main-content"]/div[@class="panel panel-leg"]')
67
68 for committee_node in committee_nodes:
69 committee_name = self.get_node(
70 committee_node,
71 './/h2[@class="panel-title"]/text()[normalize-space()]')
72
73 if committee_name is None:
74 committee_name = self.get_node(
75 committee_node,
76 './/h2[@class="panel-title"]/a/text()[normalize-space()]')
77
78 org = Organization(name=committee_name, chamber='legislature',
79 classification='committee')
80 org.add_source(main_url)
81
82 members = self.get_nodes(
83 committee_node,
84 './div[@class="list-group"]/a[@class="list-group-item"]/'
85 'text()[normalize-space()]')
86
87 for member in members:
88 member_name = re.sub(r'\Sen\.\s+', '', member)
89 member_name = re.sub(r', Chairperson', '', member_name).strip()
90 if 'Chairperson' in member:
91 member_role = 'Chairperson'
92 else:
93 member_role = 'member'
94 org.add_member(member_name, member_role)
95
96 if not org._related:
97 self.warning('No members found in {} committee.'.format(
98 org.name))
99 else:
100 yield org
101
102 def scrape(self):
103 yield from self._scrape_standing_committees()
104 yield from self._scrape_select_special_committees()
105
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/openstates/ne/committees.py b/openstates/ne/committees.py
--- a/openstates/ne/committees.py
+++ b/openstates/ne/committees.py
@@ -14,8 +14,9 @@
committee_nodes = self.get_nodes(
page,
- '//div[@class="main-content"]/div[@class="panel panel-leg"][1]/'
- 'div[@class="list-group"]/a[@class="list-group-item"]')
+ '//a[@class="accordion-switch"][contains(text(), "Standing Committees")]'
+ '/ancestor::div[@class="panel panel-leg"]//div[@class="list-group"]'
+ '/a[@class="list-group-item"]')
for committee_node in committee_nodes:
committee_page_url = committee_node.attrib['href']
@@ -63,7 +64,8 @@
committee_nodes = self.get_nodes(
page,
- '//div[@class="main-content"]/div[@class="panel panel-leg"]')
+ '//a[contains(@class, "accordion-switch")]'
+ '/ancestor::div[@class="panel panel-leg"]')
for committee_node in committee_nodes:
committee_name = self.get_node(
@@ -81,8 +83,8 @@
members = self.get_nodes(
committee_node,
- './div[@class="list-group"]/a[@class="list-group-item"]/'
- 'text()[normalize-space()]')
+ './/a[@class="list-group-item"]'
+ '/text()[normalize-space()]')
for member in members:
member_name = re.sub(r'\Sen\.\s+', '', member)
| {"golden_diff": "diff --git a/openstates/ne/committees.py b/openstates/ne/committees.py\n--- a/openstates/ne/committees.py\n+++ b/openstates/ne/committees.py\n@@ -14,8 +14,9 @@\n \n committee_nodes = self.get_nodes(\n page,\n- '//div[@class=\"main-content\"]/div[@class=\"panel panel-leg\"][1]/'\n- 'div[@class=\"list-group\"]/a[@class=\"list-group-item\"]')\n+ '//a[@class=\"accordion-switch\"][contains(text(), \"Standing Committees\")]'\n+ '/ancestor::div[@class=\"panel panel-leg\"]//div[@class=\"list-group\"]'\n+ '/a[@class=\"list-group-item\"]')\n \n for committee_node in committee_nodes:\n committee_page_url = committee_node.attrib['href']\n@@ -63,7 +64,8 @@\n \n committee_nodes = self.get_nodes(\n page,\n- '//div[@class=\"main-content\"]/div[@class=\"panel panel-leg\"]')\n+ '//a[contains(@class, \"accordion-switch\")]'\n+ '/ancestor::div[@class=\"panel panel-leg\"]')\n \n for committee_node in committee_nodes:\n committee_name = self.get_node(\n@@ -81,8 +83,8 @@\n \n members = self.get_nodes(\n committee_node,\n- './div[@class=\"list-group\"]/a[@class=\"list-group-item\"]/'\n- 'text()[normalize-space()]')\n+ './/a[@class=\"list-group-item\"]'\n+ '/text()[normalize-space()]')\n \n for member in members:\n member_name = re.sub(r'\\Sen\\.\\s+', '', member)\n", "issue": "NE failing since at least 2018-06-28\nNE has been failing since 2018-06-28\n\nBased on automated runs it appears that NE has not run successfully in 2 days (2018-06-28).\n\n\n```\n 02:30:54 INFO pupa: save membership 5b4c2cb2-7b6e-11e8-9e19-02e29baaa692 membership in ~{\"classification\": \"legislature\"} as membership_5b4c30ea-7b6e-11e8-9e19-02e29baaa692.json\n02:30:54 INFO pupa: save membership 5b4c2cb2-7b6e-11e8-9e19-02e29baaa692 membership in ~{\"classification\": \"party\", \"name\": \"Nonpartisan\"} as membership_5b4c3400-7b6e-11e8-9e19-02e29baaa692.json\n02:30:54 INFO scrapelib: GET - http://news.legislature.ne.gov/dist44\n02:30:55 INFO pupa: save person Dan Hughes as person_5bea39ca-7b6e-11e8-9e19-02e29baaa692.json\n02:30:55 INFO pupa: save membership 5bea39ca-7b6e-11e8-9e19-02e29baaa692 membership in ~{\"classification\": \"legislature\"} as membership_5bea3dda-7b6e-11e8-9e19-02e29baaa692.json\n02:30:55 INFO pupa: save membership 5bea39ca-7b6e-11e8-9e19-02e29baaa692 membership in ~{\"classification\": \"party\", \"name\": \"Nonpartisan\"} as membership_5bea4028-7b6e-11e8-9e19-02e29baaa692.json\n02:30:55 INFO scrapelib: GET - http://news.legislature.ne.gov/dist45\n02:30:56 INFO pupa: save person Sue Crawford as person_5c938c46-7b6e-11e8-9e19-02e29baaa692.json\n02:30:56 INFO pupa: save membership 5c938c46-7b6e-11e8-9e19-02e29baaa692 membership in ~{\"classification\": \"legislature\"} as membership_5c93909c-7b6e-11e8-9e19-02e29baaa692.json\n02:30:56 INFO pupa: save membership 5c938c46-7b6e-11e8-9e19-02e29baaa692 membership in ~{\"classification\": \"party\", \"name\": \"Nonpartisan\"} as membership_5c93939e-7b6e-11e8-9e19-02e29baaa692.json\n02:30:56 INFO scrapelib: GET - http://news.legislature.ne.gov/dist46\n02:30:57 INFO pupa: save person Adam Morfeld as person_5d0a16ea-7b6e-11e8-9e19-02e29baaa692.json\n02:30:57 INFO pupa: save membership 5d0a16ea-7b6e-11e8-9e19-02e29baaa692 membership in ~{\"classification\": \"legislature\"} as membership_5d0a1ae6-7b6e-11e8-9e19-02e29baaa692.json\n02:30:57 INFO pupa: save membership 5d0a16ea-7b6e-11e8-9e19-02e29baaa692 membership in ~{\"classification\": \"party\", \"name\": \"Nonpartisan\"} as membership_5d0a1d2a-7b6e-11e8-9e19-02e29baaa692.json\n02:30:57 INFO scrapelib: GET - http://news.legislature.ne.gov/dist47\n02:30:58 INFO pupa: save person Steve Erdman as person_5dc322ac-7b6e-11e8-9e19-02e29baaa692.json\n02:30:58 INFO pupa: save membership 5dc322ac-7b6e-11e8-9e19-02e29baaa692 membership in ~{\"classification\": \"legislature\"} as membership_5dc32694-7b6e-11e8-9e19-02e29baaa692.json\n02:30:58 INFO pupa: save membership 5dc322ac-7b6e-11e8-9e19-02e29baaa692 membership in ~{\"classification\": \"party\", \"name\": \"Nonpartisan\"} as membership_5dc32afe-7b6e-11e8-9e19-02e29baaa692.json\n02:30:58 INFO scrapelib: GET - http://news.legislature.ne.gov/dist48\n02:30:59 INFO pupa: save person John Stinner as person_5e89a800-7b6e-11e8-9e19-02e29baaa692.json\n02:30:59 INFO pupa: save membership 5e89a800-7b6e-11e8-9e19-02e29baaa692 membership in ~{\"classification\": \"legislature\"} as membership_5e89ac4c-7b6e-11e8-9e19-02e29baaa692.json\n02:30:59 INFO pupa: save membership 5e89a800-7b6e-11e8-9e19-02e29baaa692 membership in ~{\"classification\": \"party\", \"name\": \"Nonpartisan\"} as membership_5e89aed6-7b6e-11e8-9e19-02e29baaa692.json\n02:30:59 INFO scrapelib: GET - http://news.legislature.ne.gov/dist49\n02:31:00 INFO pupa: save person John Murante as person_5ed7988a-7b6e-11e8-9e19-02e29baaa692.json\n02:31:00 INFO pupa: save membership 5ed7988a-7b6e-11e8-9e19-02e29baaa692 membership in ~{\"classification\": \"legislature\"} as membership_5ed79c90-7b6e-11e8-9e19-02e29baaa692.json\n02:31:00 INFO pupa: save membership 5ed7988a-7b6e-11e8-9e19-02e29baaa692 membership in ~{\"classification\": \"party\", \"name\": \"Nonpartisan\"} as membership_5ed79ede-7b6e-11e8-9e19-02e29baaa692.json\n02:31:00 INFO scrapelib: GET - http://www.nebraskalegislature.gov/committees/standing-committees.php\n02:31:00 INFO scrapelib: GET - http://www.nebraskalegislature.gov/committees/select-committees.php\n02:31:01 WARNING pupa: No members found in Select committees committee.\n02:31:01 WARNING pupa: No members found in Special committees committee.\nTraceback (most recent call last):\n File \"/opt/openstates/venv-pupa//bin/pupa\", line 11, in <module>\n load_entry_point('pupa', 'console_scripts', 'pupa')()\n File \"/opt/openstates/venv-pupa/src/pupa/pupa/cli/__main__.py\", line 68, in main\n subcommands[args.subcommand].handle(args, other)\n File \"/opt/openstates/venv-pupa/src/pupa/pupa/cli/commands/update.py\", line 260, in handle\n return self.do_handle(args, other, juris)\n File \"/opt/openstates/venv-pupa/src/pupa/pupa/cli/commands/update.py\", line 305, in do_handle\n report['scrape'] = self.do_scrape(juris, args, scrapers)\n File \"/opt/openstates/venv-pupa/src/pupa/pupa/cli/commands/update.py\", line 173, in do_scrape\n report[scraper_name] = scraper.do_scrape(**scrape_args)\n File \"/opt/openstates/venv-pupa/src/pupa/pupa/scrape/base.py\", line 120, in do_scrape\n raise ScrapeError('no objects returned from {} scrape'.format(self.__class__.__name__))\npupa.exceptions.ScrapeError: no objects returned from NECommitteeScraper scrape\nloaded Open States pupa settings...\nne (scrape, import)\n bills: {}\n votes: {}\n people: {}\n committees: {}\n```\n\nVisit http://bobsled.openstates.org for more info.\n\n", "before_files": [{"content": "import re\n\nfrom pupa.scrape import Scraper, Organization\n\nfrom openstates.utils import LXMLMixin\n\n\nclass NECommitteeScraper(Scraper, LXMLMixin):\n def _scrape_standing_committees(self):\n \"\"\"Scrapes the Standing Committees page of the Nebraska state\n legislature.\"\"\"\n main_url = 'http://www.nebraskalegislature.gov/committees/standing-committees.php'\n page = self.lxmlize(main_url)\n\n committee_nodes = self.get_nodes(\n page,\n '//div[@class=\"main-content\"]/div[@class=\"panel panel-leg\"][1]/'\n 'div[@class=\"list-group\"]/a[@class=\"list-group-item\"]')\n\n for committee_node in committee_nodes:\n committee_page_url = committee_node.attrib['href']\n committee_page = self.lxmlize(committee_page_url)\n\n name_text = self.get_node(\n committee_page,\n '//div[@class=\"container view-front\"]/div[@class=\"row\"]/'\n 'div[@class=\"col-sm-6 col-md-7\"]/h1/text()[normalize-space()]')\n name = name_text.split()[0:-1]\n\n committee_name = ''\n for x in range(len(name)):\n committee_name += name[x] + ' '\n committee_name = committee_name[0: -1]\n\n org = Organization(name=committee_name, chamber='legislature',\n classification='committee')\n\n members = self.get_nodes(\n committee_page,\n '//div[@class=\"col-sm-4 col-md-3 ltc-col-right\"][1]/'\n 'div[@class=\"block-box\"][1]/ul[@class=\"list-unstyled '\n 'feature-content\"]/li/a/text()[normalize-space()]')\n\n for member in members:\n member_name = re.sub(r'\\Sen\\.\\s+', '', member)\n member_name = re.sub(r', Chairperson', '', member_name).strip()\n if 'Chairperson' in member:\n member_role = 'Chairperson'\n else:\n member_role = 'member'\n org.add_member(member_name, member_role)\n\n org.add_source(main_url)\n org.add_source(committee_page_url)\n\n yield org\n\n def _scrape_select_special_committees(self):\n \"\"\"Scrapes the Select and Special Committees page of the\n Nebraska state legislature.\"\"\"\n main_url = 'http://www.nebraskalegislature.gov/committees/select-committees.php'\n page = self.lxmlize(main_url)\n\n committee_nodes = self.get_nodes(\n page,\n '//div[@class=\"main-content\"]/div[@class=\"panel panel-leg\"]')\n\n for committee_node in committee_nodes:\n committee_name = self.get_node(\n committee_node,\n './/h2[@class=\"panel-title\"]/text()[normalize-space()]')\n\n if committee_name is None:\n committee_name = self.get_node(\n committee_node,\n './/h2[@class=\"panel-title\"]/a/text()[normalize-space()]')\n\n org = Organization(name=committee_name, chamber='legislature',\n classification='committee')\n org.add_source(main_url)\n\n members = self.get_nodes(\n committee_node,\n './div[@class=\"list-group\"]/a[@class=\"list-group-item\"]/'\n 'text()[normalize-space()]')\n\n for member in members:\n member_name = re.sub(r'\\Sen\\.\\s+', '', member)\n member_name = re.sub(r', Chairperson', '', member_name).strip()\n if 'Chairperson' in member:\n member_role = 'Chairperson'\n else:\n member_role = 'member'\n org.add_member(member_name, member_role)\n\n if not org._related:\n self.warning('No members found in {} committee.'.format(\n org.name))\n else:\n yield org\n\n def scrape(self):\n yield from self._scrape_standing_committees()\n yield from self._scrape_select_special_committees()\n", "path": "openstates/ne/committees.py"}], "after_files": [{"content": "import re\n\nfrom pupa.scrape import Scraper, Organization\n\nfrom openstates.utils import LXMLMixin\n\n\nclass NECommitteeScraper(Scraper, LXMLMixin):\n def _scrape_standing_committees(self):\n \"\"\"Scrapes the Standing Committees page of the Nebraska state\n legislature.\"\"\"\n main_url = 'http://www.nebraskalegislature.gov/committees/standing-committees.php'\n page = self.lxmlize(main_url)\n\n committee_nodes = self.get_nodes(\n page,\n '//a[@class=\"accordion-switch\"][contains(text(), \"Standing Committees\")]'\n '/ancestor::div[@class=\"panel panel-leg\"]//div[@class=\"list-group\"]'\n '/a[@class=\"list-group-item\"]')\n\n for committee_node in committee_nodes:\n committee_page_url = committee_node.attrib['href']\n committee_page = self.lxmlize(committee_page_url)\n\n name_text = self.get_node(\n committee_page,\n '//div[@class=\"container view-front\"]/div[@class=\"row\"]/'\n 'div[@class=\"col-sm-6 col-md-7\"]/h1/text()[normalize-space()]')\n name = name_text.split()[0:-1]\n\n committee_name = ''\n for x in range(len(name)):\n committee_name += name[x] + ' '\n committee_name = committee_name[0: -1]\n\n org = Organization(name=committee_name, chamber='legislature',\n classification='committee')\n\n members = self.get_nodes(\n committee_page,\n '//div[@class=\"col-sm-4 col-md-3 ltc-col-right\"][1]/'\n 'div[@class=\"block-box\"][1]/ul[@class=\"list-unstyled '\n 'feature-content\"]/li/a/text()[normalize-space()]')\n\n for member in members:\n member_name = re.sub(r'\\Sen\\.\\s+', '', member)\n member_name = re.sub(r', Chairperson', '', member_name).strip()\n if 'Chairperson' in member:\n member_role = 'Chairperson'\n else:\n member_role = 'member'\n org.add_member(member_name, member_role)\n\n org.add_source(main_url)\n org.add_source(committee_page_url)\n\n yield org\n\n def _scrape_select_special_committees(self):\n \"\"\"Scrapes the Select and Special Committees page of the\n Nebraska state legislature.\"\"\"\n main_url = 'http://www.nebraskalegislature.gov/committees/select-committees.php'\n page = self.lxmlize(main_url)\n\n committee_nodes = self.get_nodes(\n page,\n '//a[contains(@class, \"accordion-switch\")]'\n '/ancestor::div[@class=\"panel panel-leg\"]')\n\n for committee_node in committee_nodes:\n committee_name = self.get_node(\n committee_node,\n './/h2[@class=\"panel-title\"]/text()[normalize-space()]')\n\n if committee_name is None:\n committee_name = self.get_node(\n committee_node,\n './/h2[@class=\"panel-title\"]/a/text()[normalize-space()]')\n\n org = Organization(name=committee_name, chamber='legislature',\n classification='committee')\n org.add_source(main_url)\n\n members = self.get_nodes(\n committee_node,\n './/a[@class=\"list-group-item\"]'\n '/text()[normalize-space()]')\n\n for member in members:\n member_name = re.sub(r'\\Sen\\.\\s+', '', member)\n member_name = re.sub(r', Chairperson', '', member_name).strip()\n if 'Chairperson' in member:\n member_role = 'Chairperson'\n else:\n member_role = 'member'\n org.add_member(member_name, member_role)\n\n if not org._related:\n self.warning('No members found in {} committee.'.format(\n org.name))\n else:\n yield org\n\n def scrape(self):\n yield from self._scrape_standing_committees()\n yield from self._scrape_select_special_committees()\n", "path": "openstates/ne/committees.py"}]} | 3,727 | 360 |
gh_patches_debug_10614 | rasdani/github-patches | git_diff | getredash__redash-2134 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Use can't download dataset before saving query
Because the query results url contains the query id, before saving the user can't download the dataset.
We need to allow addressing query results without query id.
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `redash/handlers/api.py`
Content:
```
1 from flask_restful import Api
2 from werkzeug.wrappers import Response
3 from flask import make_response
4
5 from redash.utils import json_dumps
6 from redash.handlers.base import org_scoped_rule
7 from redash.handlers.permissions import ObjectPermissionsListResource, CheckPermissionResource
8 from redash.handlers.alerts import AlertResource, AlertListResource, AlertSubscriptionListResource, AlertSubscriptionResource
9 from redash.handlers.dashboards import DashboardListResource, RecentDashboardsResource, DashboardResource, DashboardShareResource, PublicDashboardResource
10 from redash.handlers.data_sources import DataSourceTypeListResource, DataSourceListResource, DataSourceSchemaResource, DataSourceResource, DataSourcePauseResource, DataSourceTestResource
11 from redash.handlers.events import EventResource
12 from redash.handlers.queries import QueryForkResource, QueryRefreshResource, QueryListResource, QueryRecentResource, QuerySearchResource, QueryResource, MyQueriesResource
13 from redash.handlers.query_results import QueryResultListResource, QueryResultResource, JobResource
14 from redash.handlers.users import UserResource, UserListResource, UserInviteResource, UserResetPasswordResource
15 from redash.handlers.visualizations import VisualizationListResource
16 from redash.handlers.visualizations import VisualizationResource
17 from redash.handlers.widgets import WidgetResource, WidgetListResource
18 from redash.handlers.groups import GroupListResource, GroupResource, GroupMemberListResource, GroupMemberResource, \
19 GroupDataSourceListResource, GroupDataSourceResource
20 from redash.handlers.destinations import DestinationTypeListResource, DestinationResource, DestinationListResource
21 from redash.handlers.query_snippets import QuerySnippetListResource, QuerySnippetResource
22
23
24 class ApiExt(Api):
25 def add_org_resource(self, resource, *urls, **kwargs):
26 urls = [org_scoped_rule(url) for url in urls]
27 return self.add_resource(resource, *urls, **kwargs)
28
29 api = ApiExt()
30
31
32 @api.representation('application/json')
33 def json_representation(data, code, headers=None):
34 # Flask-Restful checks only for flask.Response but flask-login uses werkzeug.wrappers.Response
35 if isinstance(data, Response):
36 return data
37 resp = make_response(json_dumps(data), code)
38 resp.headers.extend(headers or {})
39 return resp
40
41
42 api.add_org_resource(AlertResource, '/api/alerts/<alert_id>', endpoint='alert')
43 api.add_org_resource(AlertSubscriptionListResource, '/api/alerts/<alert_id>/subscriptions', endpoint='alert_subscriptions')
44 api.add_org_resource(AlertSubscriptionResource, '/api/alerts/<alert_id>/subscriptions/<subscriber_id>', endpoint='alert_subscription')
45 api.add_org_resource(AlertListResource, '/api/alerts', endpoint='alerts')
46
47 api.add_org_resource(DashboardListResource, '/api/dashboards', endpoint='dashboards')
48 api.add_org_resource(RecentDashboardsResource, '/api/dashboards/recent', endpoint='recent_dashboards')
49 api.add_org_resource(DashboardResource, '/api/dashboards/<dashboard_slug>', endpoint='dashboard')
50 api.add_org_resource(PublicDashboardResource, '/api/dashboards/public/<token>', endpoint='public_dashboard')
51 api.add_org_resource(DashboardShareResource, '/api/dashboards/<dashboard_id>/share', endpoint='dashboard_share')
52
53 api.add_org_resource(DataSourceTypeListResource, '/api/data_sources/types', endpoint='data_source_types')
54 api.add_org_resource(DataSourceListResource, '/api/data_sources', endpoint='data_sources')
55 api.add_org_resource(DataSourceSchemaResource, '/api/data_sources/<data_source_id>/schema')
56 api.add_org_resource(DataSourcePauseResource, '/api/data_sources/<data_source_id>/pause')
57 api.add_org_resource(DataSourceTestResource, '/api/data_sources/<data_source_id>/test')
58 api.add_org_resource(DataSourceResource, '/api/data_sources/<data_source_id>', endpoint='data_source')
59
60 api.add_org_resource(GroupListResource, '/api/groups', endpoint='groups')
61 api.add_org_resource(GroupResource, '/api/groups/<group_id>', endpoint='group')
62 api.add_org_resource(GroupMemberListResource, '/api/groups/<group_id>/members', endpoint='group_members')
63 api.add_org_resource(GroupMemberResource, '/api/groups/<group_id>/members/<user_id>', endpoint='group_member')
64 api.add_org_resource(GroupDataSourceListResource, '/api/groups/<group_id>/data_sources', endpoint='group_data_sources')
65 api.add_org_resource(GroupDataSourceResource, '/api/groups/<group_id>/data_sources/<data_source_id>', endpoint='group_data_source')
66
67 api.add_org_resource(EventResource, '/api/events', endpoint='events')
68
69 api.add_org_resource(QuerySearchResource, '/api/queries/search', endpoint='queries_search')
70 api.add_org_resource(QueryRecentResource, '/api/queries/recent', endpoint='recent_queries')
71 api.add_org_resource(QueryListResource, '/api/queries', endpoint='queries')
72 api.add_org_resource(MyQueriesResource, '/api/queries/my', endpoint='my_queries')
73 api.add_org_resource(QueryRefreshResource, '/api/queries/<query_id>/refresh', endpoint='query_refresh')
74 api.add_org_resource(QueryResource, '/api/queries/<query_id>', endpoint='query')
75 api.add_org_resource(QueryForkResource, '/api/queries/<query_id>/fork', endpoint='query_fork')
76
77 api.add_org_resource(ObjectPermissionsListResource, '/api/<object_type>/<object_id>/acl', endpoint='object_permissions')
78 api.add_org_resource(CheckPermissionResource, '/api/<object_type>/<object_id>/acl/<access_type>', endpoint='check_permissions')
79
80 api.add_org_resource(QueryResultListResource, '/api/query_results', endpoint='query_results')
81 api.add_org_resource(QueryResultResource,
82 '/api/query_results/<query_result_id>',
83 '/api/queries/<query_id>/results.<filetype>',
84 '/api/queries/<query_id>/results/<query_result_id>.<filetype>',
85 endpoint='query_result')
86 api.add_org_resource(JobResource, '/api/jobs/<job_id>', endpoint='job')
87
88 api.add_org_resource(UserListResource, '/api/users', endpoint='users')
89 api.add_org_resource(UserResource, '/api/users/<user_id>', endpoint='user')
90 api.add_org_resource(UserInviteResource, '/api/users/<user_id>/invite', endpoint='user_invite')
91 api.add_org_resource(UserResetPasswordResource, '/api/users/<user_id>/reset_password', endpoint='user_reset_password')
92
93 api.add_org_resource(VisualizationListResource, '/api/visualizations', endpoint='visualizations')
94 api.add_org_resource(VisualizationResource, '/api/visualizations/<visualization_id>', endpoint='visualization')
95
96 api.add_org_resource(WidgetListResource, '/api/widgets', endpoint='widgets')
97 api.add_org_resource(WidgetResource, '/api/widgets/<int:widget_id>', endpoint='widget')
98
99 api.add_org_resource(DestinationTypeListResource, '/api/destinations/types', endpoint='destination_types')
100 api.add_org_resource(DestinationResource, '/api/destinations/<destination_id>', endpoint='destination')
101 api.add_org_resource(DestinationListResource, '/api/destinations', endpoint='destinations')
102
103 api.add_org_resource(QuerySnippetResource, '/api/query_snippets/<snippet_id>', endpoint='query_snippet')
104 api.add_org_resource(QuerySnippetListResource, '/api/query_snippets', endpoint='query_snippets')
105
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/redash/handlers/api.py b/redash/handlers/api.py
--- a/redash/handlers/api.py
+++ b/redash/handlers/api.py
@@ -79,6 +79,7 @@
api.add_org_resource(QueryResultListResource, '/api/query_results', endpoint='query_results')
api.add_org_resource(QueryResultResource,
+ '/api/query_results/<query_result_id>.<filetype>',
'/api/query_results/<query_result_id>',
'/api/queries/<query_id>/results.<filetype>',
'/api/queries/<query_id>/results/<query_result_id>.<filetype>',
| {"golden_diff": "diff --git a/redash/handlers/api.py b/redash/handlers/api.py\n--- a/redash/handlers/api.py\n+++ b/redash/handlers/api.py\n@@ -79,6 +79,7 @@\n \n api.add_org_resource(QueryResultListResource, '/api/query_results', endpoint='query_results')\n api.add_org_resource(QueryResultResource,\n+ '/api/query_results/<query_result_id>.<filetype>',\n '/api/query_results/<query_result_id>',\n '/api/queries/<query_id>/results.<filetype>',\n '/api/queries/<query_id>/results/<query_result_id>.<filetype>',\n", "issue": "Use can't download dataset before saving query\nBecause the query results url contains the query id, before saving the user can't download the dataset. \n\nWe need to allow addressing query results without query id.\n\n", "before_files": [{"content": "from flask_restful import Api\nfrom werkzeug.wrappers import Response\nfrom flask import make_response\n\nfrom redash.utils import json_dumps\nfrom redash.handlers.base import org_scoped_rule\nfrom redash.handlers.permissions import ObjectPermissionsListResource, CheckPermissionResource\nfrom redash.handlers.alerts import AlertResource, AlertListResource, AlertSubscriptionListResource, AlertSubscriptionResource\nfrom redash.handlers.dashboards import DashboardListResource, RecentDashboardsResource, DashboardResource, DashboardShareResource, PublicDashboardResource \nfrom redash.handlers.data_sources import DataSourceTypeListResource, DataSourceListResource, DataSourceSchemaResource, DataSourceResource, DataSourcePauseResource, DataSourceTestResource\nfrom redash.handlers.events import EventResource\nfrom redash.handlers.queries import QueryForkResource, QueryRefreshResource, QueryListResource, QueryRecentResource, QuerySearchResource, QueryResource, MyQueriesResource\nfrom redash.handlers.query_results import QueryResultListResource, QueryResultResource, JobResource\nfrom redash.handlers.users import UserResource, UserListResource, UserInviteResource, UserResetPasswordResource\nfrom redash.handlers.visualizations import VisualizationListResource\nfrom redash.handlers.visualizations import VisualizationResource\nfrom redash.handlers.widgets import WidgetResource, WidgetListResource\nfrom redash.handlers.groups import GroupListResource, GroupResource, GroupMemberListResource, GroupMemberResource, \\\n GroupDataSourceListResource, GroupDataSourceResource\nfrom redash.handlers.destinations import DestinationTypeListResource, DestinationResource, DestinationListResource\nfrom redash.handlers.query_snippets import QuerySnippetListResource, QuerySnippetResource\n\n\nclass ApiExt(Api):\n def add_org_resource(self, resource, *urls, **kwargs):\n urls = [org_scoped_rule(url) for url in urls]\n return self.add_resource(resource, *urls, **kwargs)\n\napi = ApiExt()\n\n\[email protected]('application/json')\ndef json_representation(data, code, headers=None):\n # Flask-Restful checks only for flask.Response but flask-login uses werkzeug.wrappers.Response\n if isinstance(data, Response):\n return data\n resp = make_response(json_dumps(data), code)\n resp.headers.extend(headers or {})\n return resp\n\n\napi.add_org_resource(AlertResource, '/api/alerts/<alert_id>', endpoint='alert')\napi.add_org_resource(AlertSubscriptionListResource, '/api/alerts/<alert_id>/subscriptions', endpoint='alert_subscriptions')\napi.add_org_resource(AlertSubscriptionResource, '/api/alerts/<alert_id>/subscriptions/<subscriber_id>', endpoint='alert_subscription')\napi.add_org_resource(AlertListResource, '/api/alerts', endpoint='alerts')\n\napi.add_org_resource(DashboardListResource, '/api/dashboards', endpoint='dashboards')\napi.add_org_resource(RecentDashboardsResource, '/api/dashboards/recent', endpoint='recent_dashboards')\napi.add_org_resource(DashboardResource, '/api/dashboards/<dashboard_slug>', endpoint='dashboard')\napi.add_org_resource(PublicDashboardResource, '/api/dashboards/public/<token>', endpoint='public_dashboard')\napi.add_org_resource(DashboardShareResource, '/api/dashboards/<dashboard_id>/share', endpoint='dashboard_share')\n\napi.add_org_resource(DataSourceTypeListResource, '/api/data_sources/types', endpoint='data_source_types')\napi.add_org_resource(DataSourceListResource, '/api/data_sources', endpoint='data_sources')\napi.add_org_resource(DataSourceSchemaResource, '/api/data_sources/<data_source_id>/schema')\napi.add_org_resource(DataSourcePauseResource, '/api/data_sources/<data_source_id>/pause')\napi.add_org_resource(DataSourceTestResource, '/api/data_sources/<data_source_id>/test')\napi.add_org_resource(DataSourceResource, '/api/data_sources/<data_source_id>', endpoint='data_source')\n\napi.add_org_resource(GroupListResource, '/api/groups', endpoint='groups')\napi.add_org_resource(GroupResource, '/api/groups/<group_id>', endpoint='group')\napi.add_org_resource(GroupMemberListResource, '/api/groups/<group_id>/members', endpoint='group_members')\napi.add_org_resource(GroupMemberResource, '/api/groups/<group_id>/members/<user_id>', endpoint='group_member')\napi.add_org_resource(GroupDataSourceListResource, '/api/groups/<group_id>/data_sources', endpoint='group_data_sources')\napi.add_org_resource(GroupDataSourceResource, '/api/groups/<group_id>/data_sources/<data_source_id>', endpoint='group_data_source')\n\napi.add_org_resource(EventResource, '/api/events', endpoint='events')\n\napi.add_org_resource(QuerySearchResource, '/api/queries/search', endpoint='queries_search')\napi.add_org_resource(QueryRecentResource, '/api/queries/recent', endpoint='recent_queries')\napi.add_org_resource(QueryListResource, '/api/queries', endpoint='queries')\napi.add_org_resource(MyQueriesResource, '/api/queries/my', endpoint='my_queries')\napi.add_org_resource(QueryRefreshResource, '/api/queries/<query_id>/refresh', endpoint='query_refresh')\napi.add_org_resource(QueryResource, '/api/queries/<query_id>', endpoint='query')\napi.add_org_resource(QueryForkResource, '/api/queries/<query_id>/fork', endpoint='query_fork')\n\napi.add_org_resource(ObjectPermissionsListResource, '/api/<object_type>/<object_id>/acl', endpoint='object_permissions')\napi.add_org_resource(CheckPermissionResource, '/api/<object_type>/<object_id>/acl/<access_type>', endpoint='check_permissions')\n\napi.add_org_resource(QueryResultListResource, '/api/query_results', endpoint='query_results')\napi.add_org_resource(QueryResultResource,\n '/api/query_results/<query_result_id>',\n '/api/queries/<query_id>/results.<filetype>',\n '/api/queries/<query_id>/results/<query_result_id>.<filetype>',\n endpoint='query_result')\napi.add_org_resource(JobResource, '/api/jobs/<job_id>', endpoint='job')\n\napi.add_org_resource(UserListResource, '/api/users', endpoint='users')\napi.add_org_resource(UserResource, '/api/users/<user_id>', endpoint='user')\napi.add_org_resource(UserInviteResource, '/api/users/<user_id>/invite', endpoint='user_invite')\napi.add_org_resource(UserResetPasswordResource, '/api/users/<user_id>/reset_password', endpoint='user_reset_password')\n\napi.add_org_resource(VisualizationListResource, '/api/visualizations', endpoint='visualizations')\napi.add_org_resource(VisualizationResource, '/api/visualizations/<visualization_id>', endpoint='visualization')\n\napi.add_org_resource(WidgetListResource, '/api/widgets', endpoint='widgets')\napi.add_org_resource(WidgetResource, '/api/widgets/<int:widget_id>', endpoint='widget')\n\napi.add_org_resource(DestinationTypeListResource, '/api/destinations/types', endpoint='destination_types')\napi.add_org_resource(DestinationResource, '/api/destinations/<destination_id>', endpoint='destination')\napi.add_org_resource(DestinationListResource, '/api/destinations', endpoint='destinations')\n\napi.add_org_resource(QuerySnippetResource, '/api/query_snippets/<snippet_id>', endpoint='query_snippet')\napi.add_org_resource(QuerySnippetListResource, '/api/query_snippets', endpoint='query_snippets')\n", "path": "redash/handlers/api.py"}], "after_files": [{"content": "from flask_restful import Api\nfrom werkzeug.wrappers import Response\nfrom flask import make_response\n\nfrom redash.utils import json_dumps\nfrom redash.handlers.base import org_scoped_rule\nfrom redash.handlers.permissions import ObjectPermissionsListResource, CheckPermissionResource\nfrom redash.handlers.alerts import AlertResource, AlertListResource, AlertSubscriptionListResource, AlertSubscriptionResource\nfrom redash.handlers.dashboards import DashboardListResource, RecentDashboardsResource, DashboardResource, DashboardShareResource, PublicDashboardResource \nfrom redash.handlers.data_sources import DataSourceTypeListResource, DataSourceListResource, DataSourceSchemaResource, DataSourceResource, DataSourcePauseResource, DataSourceTestResource\nfrom redash.handlers.events import EventResource\nfrom redash.handlers.queries import QueryForkResource, QueryRefreshResource, QueryListResource, QueryRecentResource, QuerySearchResource, QueryResource, MyQueriesResource\nfrom redash.handlers.query_results import QueryResultListResource, QueryResultResource, JobResource\nfrom redash.handlers.users import UserResource, UserListResource, UserInviteResource, UserResetPasswordResource\nfrom redash.handlers.visualizations import VisualizationListResource\nfrom redash.handlers.visualizations import VisualizationResource\nfrom redash.handlers.widgets import WidgetResource, WidgetListResource\nfrom redash.handlers.groups import GroupListResource, GroupResource, GroupMemberListResource, GroupMemberResource, \\\n GroupDataSourceListResource, GroupDataSourceResource\nfrom redash.handlers.destinations import DestinationTypeListResource, DestinationResource, DestinationListResource\nfrom redash.handlers.query_snippets import QuerySnippetListResource, QuerySnippetResource\n\n\nclass ApiExt(Api):\n def add_org_resource(self, resource, *urls, **kwargs):\n urls = [org_scoped_rule(url) for url in urls]\n return self.add_resource(resource, *urls, **kwargs)\n\napi = ApiExt()\n\n\[email protected]('application/json')\ndef json_representation(data, code, headers=None):\n # Flask-Restful checks only for flask.Response but flask-login uses werkzeug.wrappers.Response\n if isinstance(data, Response):\n return data\n resp = make_response(json_dumps(data), code)\n resp.headers.extend(headers or {})\n return resp\n\n\napi.add_org_resource(AlertResource, '/api/alerts/<alert_id>', endpoint='alert')\napi.add_org_resource(AlertSubscriptionListResource, '/api/alerts/<alert_id>/subscriptions', endpoint='alert_subscriptions')\napi.add_org_resource(AlertSubscriptionResource, '/api/alerts/<alert_id>/subscriptions/<subscriber_id>', endpoint='alert_subscription')\napi.add_org_resource(AlertListResource, '/api/alerts', endpoint='alerts')\n\napi.add_org_resource(DashboardListResource, '/api/dashboards', endpoint='dashboards')\napi.add_org_resource(RecentDashboardsResource, '/api/dashboards/recent', endpoint='recent_dashboards')\napi.add_org_resource(DashboardResource, '/api/dashboards/<dashboard_slug>', endpoint='dashboard')\napi.add_org_resource(PublicDashboardResource, '/api/dashboards/public/<token>', endpoint='public_dashboard')\napi.add_org_resource(DashboardShareResource, '/api/dashboards/<dashboard_id>/share', endpoint='dashboard_share')\n\napi.add_org_resource(DataSourceTypeListResource, '/api/data_sources/types', endpoint='data_source_types')\napi.add_org_resource(DataSourceListResource, '/api/data_sources', endpoint='data_sources')\napi.add_org_resource(DataSourceSchemaResource, '/api/data_sources/<data_source_id>/schema')\napi.add_org_resource(DataSourcePauseResource, '/api/data_sources/<data_source_id>/pause')\napi.add_org_resource(DataSourceTestResource, '/api/data_sources/<data_source_id>/test')\napi.add_org_resource(DataSourceResource, '/api/data_sources/<data_source_id>', endpoint='data_source')\n\napi.add_org_resource(GroupListResource, '/api/groups', endpoint='groups')\napi.add_org_resource(GroupResource, '/api/groups/<group_id>', endpoint='group')\napi.add_org_resource(GroupMemberListResource, '/api/groups/<group_id>/members', endpoint='group_members')\napi.add_org_resource(GroupMemberResource, '/api/groups/<group_id>/members/<user_id>', endpoint='group_member')\napi.add_org_resource(GroupDataSourceListResource, '/api/groups/<group_id>/data_sources', endpoint='group_data_sources')\napi.add_org_resource(GroupDataSourceResource, '/api/groups/<group_id>/data_sources/<data_source_id>', endpoint='group_data_source')\n\napi.add_org_resource(EventResource, '/api/events', endpoint='events')\n\napi.add_org_resource(QuerySearchResource, '/api/queries/search', endpoint='queries_search')\napi.add_org_resource(QueryRecentResource, '/api/queries/recent', endpoint='recent_queries')\napi.add_org_resource(QueryListResource, '/api/queries', endpoint='queries')\napi.add_org_resource(MyQueriesResource, '/api/queries/my', endpoint='my_queries')\napi.add_org_resource(QueryRefreshResource, '/api/queries/<query_id>/refresh', endpoint='query_refresh')\napi.add_org_resource(QueryResource, '/api/queries/<query_id>', endpoint='query')\napi.add_org_resource(QueryForkResource, '/api/queries/<query_id>/fork', endpoint='query_fork')\n\napi.add_org_resource(ObjectPermissionsListResource, '/api/<object_type>/<object_id>/acl', endpoint='object_permissions')\napi.add_org_resource(CheckPermissionResource, '/api/<object_type>/<object_id>/acl/<access_type>', endpoint='check_permissions')\n\napi.add_org_resource(QueryResultListResource, '/api/query_results', endpoint='query_results')\napi.add_org_resource(QueryResultResource,\n '/api/query_results/<query_result_id>.<filetype>',\n '/api/query_results/<query_result_id>',\n '/api/queries/<query_id>/results.<filetype>',\n '/api/queries/<query_id>/results/<query_result_id>.<filetype>',\n endpoint='query_result')\napi.add_org_resource(JobResource, '/api/jobs/<job_id>', endpoint='job')\n\napi.add_org_resource(UserListResource, '/api/users', endpoint='users')\napi.add_org_resource(UserResource, '/api/users/<user_id>', endpoint='user')\napi.add_org_resource(UserInviteResource, '/api/users/<user_id>/invite', endpoint='user_invite')\napi.add_org_resource(UserResetPasswordResource, '/api/users/<user_id>/reset_password', endpoint='user_reset_password')\n\napi.add_org_resource(VisualizationListResource, '/api/visualizations', endpoint='visualizations')\napi.add_org_resource(VisualizationResource, '/api/visualizations/<visualization_id>', endpoint='visualization')\n\napi.add_org_resource(WidgetListResource, '/api/widgets', endpoint='widgets')\napi.add_org_resource(WidgetResource, '/api/widgets/<int:widget_id>', endpoint='widget')\n\napi.add_org_resource(DestinationTypeListResource, '/api/destinations/types', endpoint='destination_types')\napi.add_org_resource(DestinationResource, '/api/destinations/<destination_id>', endpoint='destination')\napi.add_org_resource(DestinationListResource, '/api/destinations', endpoint='destinations')\n\napi.add_org_resource(QuerySnippetResource, '/api/query_snippets/<snippet_id>', endpoint='query_snippet')\napi.add_org_resource(QuerySnippetListResource, '/api/query_snippets', endpoint='query_snippets')\n", "path": "redash/handlers/api.py"}]} | 2,015 | 138 |
gh_patches_debug_30965 | rasdani/github-patches | git_diff | napari__napari-5085 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Bug: values of zero in 3D MIPs render as transparent, rather than black, even in opaque blending mode
## 🐛 Bug
noticed by @maweigert:
if a volume with a bunch of zeros is in front of another volume, we would have expected that it would essentially occluded any layers behind it, assuming, of course, that the blending mode is opaque. (in additive blending, naturally, any nonzero values in other layers would result in a non-black MIP pixel).
However, MIP values of zero appear to render as transparent, even in opaque blending mode, showing layers behind it:
```python
import numpy as np
import napari
x = np.zeros((20,20,20), np.float32)
x[10,10,10] = 100
y = np.zeros((20,20,20), np.float32)
y[1:-1,1:-1,1:-1] = 100
v = napari.Viewer()
v.add_image(y)
v.add_image(x)
v.dims.ndisplay=3
```
<img width="1311" alt="Screenshot 2022-09-01 at 09 19 26" src="https://user-images.githubusercontent.com/1609449/187933993-1454960c-f9fc-4ab4-b23e-a6603e093281.png">
changing `viewer.layers['x'].rendering = 'additive'` _does_ then occlude all of `y`
not 100% sure it's a bug, but it did strike us as unexpected. @brisvag?
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `napari/_vispy/layers/image.py`
Content:
```
1 import warnings
2
3 import numpy as np
4 from vispy.color import Colormap as VispyColormap
5 from vispy.scene.node import Node
6
7 from ...utils.translations import trans
8 from ..utils.gl import fix_data_dtype, get_gl_extensions
9 from ..visuals.image import Image as ImageNode
10 from ..visuals.volume import Volume as VolumeNode
11 from .base import VispyBaseLayer
12
13
14 class ImageLayerNode:
15 def __init__(self, custom_node: Node = None, texture_format=None):
16 if (
17 texture_format == 'auto'
18 and 'texture_float' not in get_gl_extensions()
19 ):
20 # if the GPU doesn't support float textures, texture_format auto
21 # WILL fail on float dtypes
22 # https://github.com/napari/napari/issues/3988
23 texture_format = None
24
25 self._custom_node = custom_node
26 self._image_node = ImageNode(
27 None,
28 method='auto',
29 texture_format=texture_format,
30 )
31 self._volume_node = VolumeNode(
32 np.zeros((1, 1, 1), dtype=np.float32),
33 clim=[0, 1],
34 texture_format=texture_format,
35 )
36
37 def get_node(self, ndisplay: int) -> Node:
38
39 # Return custom node if we have one.
40 if self._custom_node is not None:
41 return self._custom_node
42
43 # Return Image or Volume node based on 2D or 3D.
44 if ndisplay == 2:
45 return self._image_node
46 return self._volume_node
47
48
49 class VispyImageLayer(VispyBaseLayer):
50 def __init__(self, layer, node=None, texture_format='auto'):
51
52 # Use custom node from caller, or our standard image/volume nodes.
53 self._layer_node = ImageLayerNode(node, texture_format=texture_format)
54
55 # Default to 2D (image) node.
56 super().__init__(layer, self._layer_node.get_node(2))
57
58 self._array_like = True
59
60 self.layer.events.rendering.connect(self._on_rendering_change)
61 self.layer.events.depiction.connect(self._on_depiction_change)
62 self.layer.events.interpolation2d.connect(
63 self._on_interpolation_change
64 )
65 self.layer.events.interpolation3d.connect(
66 self._on_interpolation_change
67 )
68 self.layer.events.colormap.connect(self._on_colormap_change)
69 self.layer.events.contrast_limits.connect(
70 self._on_contrast_limits_change
71 )
72 self.layer.events.gamma.connect(self._on_gamma_change)
73 self.layer.events.iso_threshold.connect(self._on_iso_threshold_change)
74 self.layer.events.attenuation.connect(self._on_attenuation_change)
75 self.layer.plane.events.position.connect(
76 self._on_plane_position_change
77 )
78 self.layer.plane.events.thickness.connect(
79 self._on_plane_thickness_change
80 )
81 self.layer.plane.events.normal.connect(self._on_plane_normal_change)
82
83 # display_change is special (like data_change) because it requires a self.reset()
84 # this means that we have to call it manually. Also, it must be called before reset
85 # in order to set the appropriate node first
86 self._on_display_change()
87 self.reset()
88 self._on_data_change()
89
90 def _on_display_change(self, data=None):
91 parent = self.node.parent
92 self.node.parent = None
93
94 self.node = self._layer_node.get_node(self.layer._ndisplay)
95
96 if data is None:
97 data = np.zeros((1,) * self.layer._ndisplay, dtype=np.float32)
98
99 if self.layer._empty:
100 self.node.visible = False
101 else:
102 self.node.visible = self.layer.visible
103
104 if self.layer.loaded:
105 self.node.set_data(data)
106
107 self.node.parent = parent
108 self.node.order = self.order
109 self.reset()
110
111 def _on_data_change(self):
112 if not self.layer.loaded:
113 # Do nothing if we are not yet loaded. Calling astype below could
114 # be very expensive. Lets not do it until our data has been loaded.
115 return
116
117 self._set_node_data(self.node, self.layer._data_view)
118
119 def _set_node_data(self, node, data):
120 """Our self.layer._data_view has been updated, update our node."""
121
122 data = fix_data_dtype(data)
123
124 if self.layer._ndisplay == 3 and self.layer.ndim == 2:
125 data = np.expand_dims(data, axis=0)
126
127 # Check if data exceeds MAX_TEXTURE_SIZE and downsample
128 if self.MAX_TEXTURE_SIZE_2D is not None and self.layer._ndisplay == 2:
129 data = self.downsample_texture(data, self.MAX_TEXTURE_SIZE_2D)
130 elif (
131 self.MAX_TEXTURE_SIZE_3D is not None and self.layer._ndisplay == 3
132 ):
133 data = self.downsample_texture(data, self.MAX_TEXTURE_SIZE_3D)
134
135 # Check if ndisplay has changed current node type needs updating
136 if (
137 self.layer._ndisplay == 3 and not isinstance(node, VolumeNode)
138 ) or (self.layer._ndisplay == 2 and not isinstance(node, ImageNode)):
139 self._on_display_change(data)
140 else:
141 node.set_data(data)
142
143 if self.layer._empty:
144 node.visible = False
145 else:
146 node.visible = self.layer.visible
147
148 # Call to update order of translation values with new dims:
149 self._on_matrix_change()
150 node.update()
151
152 def _on_interpolation_change(self):
153 self.node.interpolation = (
154 self.layer.interpolation2d
155 if self.layer._ndisplay == 2
156 else self.layer.interpolation3d
157 )
158
159 def _on_rendering_change(self):
160 if isinstance(self.node, VolumeNode):
161 self.node.method = self.layer.rendering
162 self._on_attenuation_change()
163 self._on_iso_threshold_change()
164
165 def _on_depiction_change(self):
166 if isinstance(self.node, VolumeNode):
167 self.node.raycasting_mode = str(self.layer.depiction)
168
169 def _on_colormap_change(self):
170 self.node.cmap = VispyColormap(*self.layer.colormap)
171
172 def _on_contrast_limits_change(self):
173 self.node.clim = self.layer.contrast_limits
174 if isinstance(self.node, VolumeNode):
175 self.node.mip_cutoff = self.node._texture.clim_normalized[0]
176 self.node.minip_cutoff = self.node._texture.clim_normalized[1]
177
178 def _on_gamma_change(self):
179 if len(self.node.shared_program.frag._set_items) > 0:
180 self.node.gamma = self.layer.gamma
181
182 def _on_iso_threshold_change(self):
183 if isinstance(self.node, VolumeNode):
184 self.node.threshold = self.layer.iso_threshold
185
186 def _on_attenuation_change(self):
187 if isinstance(self.node, VolumeNode):
188 self.node.attenuation = self.layer.attenuation
189
190 def _on_plane_thickness_change(self):
191 if isinstance(self.node, VolumeNode):
192 self.node.plane_thickness = self.layer.plane.thickness
193
194 def _on_plane_position_change(self):
195 if isinstance(self.node, VolumeNode):
196 self.node.plane_position = self.layer.plane.position
197
198 def _on_plane_normal_change(self):
199 if isinstance(self.node, VolumeNode):
200 self.node.plane_normal = self.layer.plane.normal
201
202 def reset(self, event=None):
203 super().reset()
204 self._on_interpolation_change()
205 self._on_colormap_change()
206 self._on_contrast_limits_change()
207 self._on_gamma_change()
208 self._on_rendering_change()
209 self._on_depiction_change()
210 self._on_plane_position_change()
211 self._on_plane_normal_change()
212 self._on_plane_thickness_change()
213
214 def downsample_texture(self, data, MAX_TEXTURE_SIZE):
215 """Downsample data based on maximum allowed texture size.
216
217 Parameters
218 ----------
219 data : array
220 Data to be downsampled if needed.
221 MAX_TEXTURE_SIZE : int
222 Maximum allowed texture size.
223
224 Returns
225 -------
226 data : array
227 Data that now fits inside texture.
228 """
229 if np.any(np.greater(data.shape, MAX_TEXTURE_SIZE)):
230 if self.layer.multiscale:
231 raise ValueError(
232 trans._(
233 "Shape of in dividual tiles in multiscale {shape} cannot exceed GL_MAX_TEXTURE_SIZE {texture_size}. Rendering is currently in {ndisplay}D mode.",
234 deferred=True,
235 shape=data.shape,
236 texture_size=MAX_TEXTURE_SIZE,
237 ndisplay=self.layer._ndisplay,
238 )
239 )
240 warnings.warn(
241 trans._(
242 "data shape {shape} exceeds GL_MAX_TEXTURE_SIZE {texture_size} in at least one axis and will be downsampled. Rendering is currently in {ndisplay}D mode.",
243 deferred=True,
244 shape=data.shape,
245 texture_size=MAX_TEXTURE_SIZE,
246 ndisplay=self.layer._ndisplay,
247 )
248 )
249 downsample = np.ceil(
250 np.divide(data.shape, MAX_TEXTURE_SIZE)
251 ).astype(int)
252 scale = np.ones(self.layer.ndim)
253 for i, d in enumerate(self.layer._dims_displayed):
254 scale[d] = downsample[i]
255 self.layer._transforms['tile2data'].scale = scale
256 self._on_matrix_change()
257 slices = tuple(slice(None, None, ds) for ds in downsample)
258 data = data[slices]
259 return data
260
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/napari/_vispy/layers/image.py b/napari/_vispy/layers/image.py
--- a/napari/_vispy/layers/image.py
+++ b/napari/_vispy/layers/image.py
@@ -4,6 +4,7 @@
from vispy.color import Colormap as VispyColormap
from vispy.scene.node import Node
+from ...layers.base._base_constants import Blending
from ...utils.translations import trans
from ..utils.gl import fix_data_dtype, get_gl_extensions
from ..visuals.image import Image as ImageNode
@@ -169,11 +170,29 @@
def _on_colormap_change(self):
self.node.cmap = VispyColormap(*self.layer.colormap)
+ def _update_mip_minip_cutoff(self):
+ # discard fragments beyond contrast limits, but only with translucent blending
+ if isinstance(self.node, VolumeNode):
+ if self.layer.blending in {
+ Blending.TRANSLUCENT,
+ Blending.TRANSLUCENT_NO_DEPTH,
+ }:
+ self.node.mip_cutoff = self.node._texture.clim_normalized[0]
+ self.node.minip_cutoff = self.node._texture.clim_normalized[1]
+ else:
+ self.node.mip_cutoff = None
+ self.node.minip_cutoff = None
+
def _on_contrast_limits_change(self):
self.node.clim = self.layer.contrast_limits
- if isinstance(self.node, VolumeNode):
- self.node.mip_cutoff = self.node._texture.clim_normalized[0]
- self.node.minip_cutoff = self.node._texture.clim_normalized[1]
+ # cutoffs must be updated after clims, so we can set them to the new values
+ self._update_mip_minip_cutoff()
+
+ def _on_blending_change(self):
+ super()._on_blending_change()
+ # cutoffs must be updated after blending, so we can know if
+ # the new blending is a translucent one
+ self._update_mip_minip_cutoff()
def _on_gamma_change(self):
if len(self.node.shared_program.frag._set_items) > 0:
| {"golden_diff": "diff --git a/napari/_vispy/layers/image.py b/napari/_vispy/layers/image.py\n--- a/napari/_vispy/layers/image.py\n+++ b/napari/_vispy/layers/image.py\n@@ -4,6 +4,7 @@\n from vispy.color import Colormap as VispyColormap\n from vispy.scene.node import Node\n \n+from ...layers.base._base_constants import Blending\n from ...utils.translations import trans\n from ..utils.gl import fix_data_dtype, get_gl_extensions\n from ..visuals.image import Image as ImageNode\n@@ -169,11 +170,29 @@\n def _on_colormap_change(self):\n self.node.cmap = VispyColormap(*self.layer.colormap)\n \n+ def _update_mip_minip_cutoff(self):\n+ # discard fragments beyond contrast limits, but only with translucent blending\n+ if isinstance(self.node, VolumeNode):\n+ if self.layer.blending in {\n+ Blending.TRANSLUCENT,\n+ Blending.TRANSLUCENT_NO_DEPTH,\n+ }:\n+ self.node.mip_cutoff = self.node._texture.clim_normalized[0]\n+ self.node.minip_cutoff = self.node._texture.clim_normalized[1]\n+ else:\n+ self.node.mip_cutoff = None\n+ self.node.minip_cutoff = None\n+\n def _on_contrast_limits_change(self):\n self.node.clim = self.layer.contrast_limits\n- if isinstance(self.node, VolumeNode):\n- self.node.mip_cutoff = self.node._texture.clim_normalized[0]\n- self.node.minip_cutoff = self.node._texture.clim_normalized[1]\n+ # cutoffs must be updated after clims, so we can set them to the new values\n+ self._update_mip_minip_cutoff()\n+\n+ def _on_blending_change(self):\n+ super()._on_blending_change()\n+ # cutoffs must be updated after blending, so we can know if\n+ # the new blending is a translucent one\n+ self._update_mip_minip_cutoff()\n \n def _on_gamma_change(self):\n if len(self.node.shared_program.frag._set_items) > 0:\n", "issue": "Bug: values of zero in 3D MIPs render as transparent, rather than black, even in opaque blending mode\n## \ud83d\udc1b Bug\r\nnoticed by @maweigert:\r\n\r\nif a volume with a bunch of zeros is in front of another volume, we would have expected that it would essentially occluded any layers behind it, assuming, of course, that the blending mode is opaque. (in additive blending, naturally, any nonzero values in other layers would result in a non-black MIP pixel).\r\n\r\nHowever, MIP values of zero appear to render as transparent, even in opaque blending mode, showing layers behind it:\r\n\r\n```python\r\nimport numpy as np \r\nimport napari \r\n\r\nx = np.zeros((20,20,20), np.float32)\r\nx[10,10,10] = 100\r\n\r\ny = np.zeros((20,20,20), np.float32)\r\ny[1:-1,1:-1,1:-1] = 100\r\n\r\nv = napari.Viewer()\r\nv.add_image(y)\r\nv.add_image(x)\r\n\r\nv.dims.ndisplay=3\r\n```\r\n\r\n<img width=\"1311\" alt=\"Screenshot 2022-09-01 at 09 19 26\" src=\"https://user-images.githubusercontent.com/1609449/187933993-1454960c-f9fc-4ab4-b23e-a6603e093281.png\">\r\n\r\nchanging `viewer.layers['x'].rendering = 'additive'` _does_ then occlude all of `y`\r\nnot 100% sure it's a bug, but it did strike us as unexpected. @brisvag?\r\n\n", "before_files": [{"content": "import warnings\n\nimport numpy as np\nfrom vispy.color import Colormap as VispyColormap\nfrom vispy.scene.node import Node\n\nfrom ...utils.translations import trans\nfrom ..utils.gl import fix_data_dtype, get_gl_extensions\nfrom ..visuals.image import Image as ImageNode\nfrom ..visuals.volume import Volume as VolumeNode\nfrom .base import VispyBaseLayer\n\n\nclass ImageLayerNode:\n def __init__(self, custom_node: Node = None, texture_format=None):\n if (\n texture_format == 'auto'\n and 'texture_float' not in get_gl_extensions()\n ):\n # if the GPU doesn't support float textures, texture_format auto\n # WILL fail on float dtypes\n # https://github.com/napari/napari/issues/3988\n texture_format = None\n\n self._custom_node = custom_node\n self._image_node = ImageNode(\n None,\n method='auto',\n texture_format=texture_format,\n )\n self._volume_node = VolumeNode(\n np.zeros((1, 1, 1), dtype=np.float32),\n clim=[0, 1],\n texture_format=texture_format,\n )\n\n def get_node(self, ndisplay: int) -> Node:\n\n # Return custom node if we have one.\n if self._custom_node is not None:\n return self._custom_node\n\n # Return Image or Volume node based on 2D or 3D.\n if ndisplay == 2:\n return self._image_node\n return self._volume_node\n\n\nclass VispyImageLayer(VispyBaseLayer):\n def __init__(self, layer, node=None, texture_format='auto'):\n\n # Use custom node from caller, or our standard image/volume nodes.\n self._layer_node = ImageLayerNode(node, texture_format=texture_format)\n\n # Default to 2D (image) node.\n super().__init__(layer, self._layer_node.get_node(2))\n\n self._array_like = True\n\n self.layer.events.rendering.connect(self._on_rendering_change)\n self.layer.events.depiction.connect(self._on_depiction_change)\n self.layer.events.interpolation2d.connect(\n self._on_interpolation_change\n )\n self.layer.events.interpolation3d.connect(\n self._on_interpolation_change\n )\n self.layer.events.colormap.connect(self._on_colormap_change)\n self.layer.events.contrast_limits.connect(\n self._on_contrast_limits_change\n )\n self.layer.events.gamma.connect(self._on_gamma_change)\n self.layer.events.iso_threshold.connect(self._on_iso_threshold_change)\n self.layer.events.attenuation.connect(self._on_attenuation_change)\n self.layer.plane.events.position.connect(\n self._on_plane_position_change\n )\n self.layer.plane.events.thickness.connect(\n self._on_plane_thickness_change\n )\n self.layer.plane.events.normal.connect(self._on_plane_normal_change)\n\n # display_change is special (like data_change) because it requires a self.reset()\n # this means that we have to call it manually. Also, it must be called before reset\n # in order to set the appropriate node first\n self._on_display_change()\n self.reset()\n self._on_data_change()\n\n def _on_display_change(self, data=None):\n parent = self.node.parent\n self.node.parent = None\n\n self.node = self._layer_node.get_node(self.layer._ndisplay)\n\n if data is None:\n data = np.zeros((1,) * self.layer._ndisplay, dtype=np.float32)\n\n if self.layer._empty:\n self.node.visible = False\n else:\n self.node.visible = self.layer.visible\n\n if self.layer.loaded:\n self.node.set_data(data)\n\n self.node.parent = parent\n self.node.order = self.order\n self.reset()\n\n def _on_data_change(self):\n if not self.layer.loaded:\n # Do nothing if we are not yet loaded. Calling astype below could\n # be very expensive. Lets not do it until our data has been loaded.\n return\n\n self._set_node_data(self.node, self.layer._data_view)\n\n def _set_node_data(self, node, data):\n \"\"\"Our self.layer._data_view has been updated, update our node.\"\"\"\n\n data = fix_data_dtype(data)\n\n if self.layer._ndisplay == 3 and self.layer.ndim == 2:\n data = np.expand_dims(data, axis=0)\n\n # Check if data exceeds MAX_TEXTURE_SIZE and downsample\n if self.MAX_TEXTURE_SIZE_2D is not None and self.layer._ndisplay == 2:\n data = self.downsample_texture(data, self.MAX_TEXTURE_SIZE_2D)\n elif (\n self.MAX_TEXTURE_SIZE_3D is not None and self.layer._ndisplay == 3\n ):\n data = self.downsample_texture(data, self.MAX_TEXTURE_SIZE_3D)\n\n # Check if ndisplay has changed current node type needs updating\n if (\n self.layer._ndisplay == 3 and not isinstance(node, VolumeNode)\n ) or (self.layer._ndisplay == 2 and not isinstance(node, ImageNode)):\n self._on_display_change(data)\n else:\n node.set_data(data)\n\n if self.layer._empty:\n node.visible = False\n else:\n node.visible = self.layer.visible\n\n # Call to update order of translation values with new dims:\n self._on_matrix_change()\n node.update()\n\n def _on_interpolation_change(self):\n self.node.interpolation = (\n self.layer.interpolation2d\n if self.layer._ndisplay == 2\n else self.layer.interpolation3d\n )\n\n def _on_rendering_change(self):\n if isinstance(self.node, VolumeNode):\n self.node.method = self.layer.rendering\n self._on_attenuation_change()\n self._on_iso_threshold_change()\n\n def _on_depiction_change(self):\n if isinstance(self.node, VolumeNode):\n self.node.raycasting_mode = str(self.layer.depiction)\n\n def _on_colormap_change(self):\n self.node.cmap = VispyColormap(*self.layer.colormap)\n\n def _on_contrast_limits_change(self):\n self.node.clim = self.layer.contrast_limits\n if isinstance(self.node, VolumeNode):\n self.node.mip_cutoff = self.node._texture.clim_normalized[0]\n self.node.minip_cutoff = self.node._texture.clim_normalized[1]\n\n def _on_gamma_change(self):\n if len(self.node.shared_program.frag._set_items) > 0:\n self.node.gamma = self.layer.gamma\n\n def _on_iso_threshold_change(self):\n if isinstance(self.node, VolumeNode):\n self.node.threshold = self.layer.iso_threshold\n\n def _on_attenuation_change(self):\n if isinstance(self.node, VolumeNode):\n self.node.attenuation = self.layer.attenuation\n\n def _on_plane_thickness_change(self):\n if isinstance(self.node, VolumeNode):\n self.node.plane_thickness = self.layer.plane.thickness\n\n def _on_plane_position_change(self):\n if isinstance(self.node, VolumeNode):\n self.node.plane_position = self.layer.plane.position\n\n def _on_plane_normal_change(self):\n if isinstance(self.node, VolumeNode):\n self.node.plane_normal = self.layer.plane.normal\n\n def reset(self, event=None):\n super().reset()\n self._on_interpolation_change()\n self._on_colormap_change()\n self._on_contrast_limits_change()\n self._on_gamma_change()\n self._on_rendering_change()\n self._on_depiction_change()\n self._on_plane_position_change()\n self._on_plane_normal_change()\n self._on_plane_thickness_change()\n\n def downsample_texture(self, data, MAX_TEXTURE_SIZE):\n \"\"\"Downsample data based on maximum allowed texture size.\n\n Parameters\n ----------\n data : array\n Data to be downsampled if needed.\n MAX_TEXTURE_SIZE : int\n Maximum allowed texture size.\n\n Returns\n -------\n data : array\n Data that now fits inside texture.\n \"\"\"\n if np.any(np.greater(data.shape, MAX_TEXTURE_SIZE)):\n if self.layer.multiscale:\n raise ValueError(\n trans._(\n \"Shape of in dividual tiles in multiscale {shape} cannot exceed GL_MAX_TEXTURE_SIZE {texture_size}. Rendering is currently in {ndisplay}D mode.\",\n deferred=True,\n shape=data.shape,\n texture_size=MAX_TEXTURE_SIZE,\n ndisplay=self.layer._ndisplay,\n )\n )\n warnings.warn(\n trans._(\n \"data shape {shape} exceeds GL_MAX_TEXTURE_SIZE {texture_size} in at least one axis and will be downsampled. Rendering is currently in {ndisplay}D mode.\",\n deferred=True,\n shape=data.shape,\n texture_size=MAX_TEXTURE_SIZE,\n ndisplay=self.layer._ndisplay,\n )\n )\n downsample = np.ceil(\n np.divide(data.shape, MAX_TEXTURE_SIZE)\n ).astype(int)\n scale = np.ones(self.layer.ndim)\n for i, d in enumerate(self.layer._dims_displayed):\n scale[d] = downsample[i]\n self.layer._transforms['tile2data'].scale = scale\n self._on_matrix_change()\n slices = tuple(slice(None, None, ds) for ds in downsample)\n data = data[slices]\n return data\n", "path": "napari/_vispy/layers/image.py"}], "after_files": [{"content": "import warnings\n\nimport numpy as np\nfrom vispy.color import Colormap as VispyColormap\nfrom vispy.scene.node import Node\n\nfrom ...layers.base._base_constants import Blending\nfrom ...utils.translations import trans\nfrom ..utils.gl import fix_data_dtype, get_gl_extensions\nfrom ..visuals.image import Image as ImageNode\nfrom ..visuals.volume import Volume as VolumeNode\nfrom .base import VispyBaseLayer\n\n\nclass ImageLayerNode:\n def __init__(self, custom_node: Node = None, texture_format=None):\n if (\n texture_format == 'auto'\n and 'texture_float' not in get_gl_extensions()\n ):\n # if the GPU doesn't support float textures, texture_format auto\n # WILL fail on float dtypes\n # https://github.com/napari/napari/issues/3988\n texture_format = None\n\n self._custom_node = custom_node\n self._image_node = ImageNode(\n None,\n method='auto',\n texture_format=texture_format,\n )\n self._volume_node = VolumeNode(\n np.zeros((1, 1, 1), dtype=np.float32),\n clim=[0, 1],\n texture_format=texture_format,\n )\n\n def get_node(self, ndisplay: int) -> Node:\n\n # Return custom node if we have one.\n if self._custom_node is not None:\n return self._custom_node\n\n # Return Image or Volume node based on 2D or 3D.\n if ndisplay == 2:\n return self._image_node\n return self._volume_node\n\n\nclass VispyImageLayer(VispyBaseLayer):\n def __init__(self, layer, node=None, texture_format='auto'):\n\n # Use custom node from caller, or our standard image/volume nodes.\n self._layer_node = ImageLayerNode(node, texture_format=texture_format)\n\n # Default to 2D (image) node.\n super().__init__(layer, self._layer_node.get_node(2))\n\n self._array_like = True\n\n self.layer.events.rendering.connect(self._on_rendering_change)\n self.layer.events.depiction.connect(self._on_depiction_change)\n self.layer.events.interpolation2d.connect(\n self._on_interpolation_change\n )\n self.layer.events.interpolation3d.connect(\n self._on_interpolation_change\n )\n self.layer.events.colormap.connect(self._on_colormap_change)\n self.layer.events.contrast_limits.connect(\n self._on_contrast_limits_change\n )\n self.layer.events.gamma.connect(self._on_gamma_change)\n self.layer.events.iso_threshold.connect(self._on_iso_threshold_change)\n self.layer.events.attenuation.connect(self._on_attenuation_change)\n self.layer.plane.events.position.connect(\n self._on_plane_position_change\n )\n self.layer.plane.events.thickness.connect(\n self._on_plane_thickness_change\n )\n self.layer.plane.events.normal.connect(self._on_plane_normal_change)\n\n # display_change is special (like data_change) because it requires a self.reset()\n # this means that we have to call it manually. Also, it must be called before reset\n # in order to set the appropriate node first\n self._on_display_change()\n self.reset()\n self._on_data_change()\n\n def _on_display_change(self, data=None):\n parent = self.node.parent\n self.node.parent = None\n\n self.node = self._layer_node.get_node(self.layer._ndisplay)\n\n if data is None:\n data = np.zeros((1,) * self.layer._ndisplay, dtype=np.float32)\n\n if self.layer._empty:\n self.node.visible = False\n else:\n self.node.visible = self.layer.visible\n\n if self.layer.loaded:\n self.node.set_data(data)\n\n self.node.parent = parent\n self.node.order = self.order\n self.reset()\n\n def _on_data_change(self):\n if not self.layer.loaded:\n # Do nothing if we are not yet loaded. Calling astype below could\n # be very expensive. Lets not do it until our data has been loaded.\n return\n\n self._set_node_data(self.node, self.layer._data_view)\n\n def _set_node_data(self, node, data):\n \"\"\"Our self.layer._data_view has been updated, update our node.\"\"\"\n\n data = fix_data_dtype(data)\n\n if self.layer._ndisplay == 3 and self.layer.ndim == 2:\n data = np.expand_dims(data, axis=0)\n\n # Check if data exceeds MAX_TEXTURE_SIZE and downsample\n if self.MAX_TEXTURE_SIZE_2D is not None and self.layer._ndisplay == 2:\n data = self.downsample_texture(data, self.MAX_TEXTURE_SIZE_2D)\n elif (\n self.MAX_TEXTURE_SIZE_3D is not None and self.layer._ndisplay == 3\n ):\n data = self.downsample_texture(data, self.MAX_TEXTURE_SIZE_3D)\n\n # Check if ndisplay has changed current node type needs updating\n if (\n self.layer._ndisplay == 3 and not isinstance(node, VolumeNode)\n ) or (self.layer._ndisplay == 2 and not isinstance(node, ImageNode)):\n self._on_display_change(data)\n else:\n node.set_data(data)\n\n if self.layer._empty:\n node.visible = False\n else:\n node.visible = self.layer.visible\n\n # Call to update order of translation values with new dims:\n self._on_matrix_change()\n node.update()\n\n def _on_interpolation_change(self):\n self.node.interpolation = (\n self.layer.interpolation2d\n if self.layer._ndisplay == 2\n else self.layer.interpolation3d\n )\n\n def _on_rendering_change(self):\n if isinstance(self.node, VolumeNode):\n self.node.method = self.layer.rendering\n self._on_attenuation_change()\n self._on_iso_threshold_change()\n\n def _on_depiction_change(self):\n if isinstance(self.node, VolumeNode):\n self.node.raycasting_mode = str(self.layer.depiction)\n\n def _on_colormap_change(self):\n self.node.cmap = VispyColormap(*self.layer.colormap)\n\n def _update_mip_minip_cutoff(self):\n # discard fragments beyond contrast limits, but only with translucent blending\n if isinstance(self.node, VolumeNode):\n if self.layer.blending in {\n Blending.TRANSLUCENT,\n Blending.TRANSLUCENT_NO_DEPTH,\n }:\n self.node.mip_cutoff = self.node._texture.clim_normalized[0]\n self.node.minip_cutoff = self.node._texture.clim_normalized[1]\n else:\n self.node.mip_cutoff = None\n self.node.minip_cutoff = None\n\n def _on_contrast_limits_change(self):\n self.node.clim = self.layer.contrast_limits\n # cutoffs must be updated after clims, so we can set them to the new values\n self._update_mip_minip_cutoff()\n\n def _on_blending_change(self):\n super()._on_blending_change()\n # cutoffs must be updated after blending, so we can know if\n # the new blending is a translucent one\n self._update_mip_minip_cutoff()\n\n def _on_gamma_change(self):\n if len(self.node.shared_program.frag._set_items) > 0:\n self.node.gamma = self.layer.gamma\n\n def _on_iso_threshold_change(self):\n if isinstance(self.node, VolumeNode):\n self.node.threshold = self.layer.iso_threshold\n\n def _on_attenuation_change(self):\n if isinstance(self.node, VolumeNode):\n self.node.attenuation = self.layer.attenuation\n\n def _on_plane_thickness_change(self):\n if isinstance(self.node, VolumeNode):\n self.node.plane_thickness = self.layer.plane.thickness\n\n def _on_plane_position_change(self):\n if isinstance(self.node, VolumeNode):\n self.node.plane_position = self.layer.plane.position\n\n def _on_plane_normal_change(self):\n if isinstance(self.node, VolumeNode):\n self.node.plane_normal = self.layer.plane.normal\n\n def reset(self, event=None):\n super().reset()\n self._on_interpolation_change()\n self._on_colormap_change()\n self._on_contrast_limits_change()\n self._on_gamma_change()\n self._on_rendering_change()\n self._on_depiction_change()\n self._on_plane_position_change()\n self._on_plane_normal_change()\n self._on_plane_thickness_change()\n\n def downsample_texture(self, data, MAX_TEXTURE_SIZE):\n \"\"\"Downsample data based on maximum allowed texture size.\n\n Parameters\n ----------\n data : array\n Data to be downsampled if needed.\n MAX_TEXTURE_SIZE : int\n Maximum allowed texture size.\n\n Returns\n -------\n data : array\n Data that now fits inside texture.\n \"\"\"\n if np.any(np.greater(data.shape, MAX_TEXTURE_SIZE)):\n if self.layer.multiscale:\n raise ValueError(\n trans._(\n \"Shape of in dividual tiles in multiscale {shape} cannot exceed GL_MAX_TEXTURE_SIZE {texture_size}. Rendering is currently in {ndisplay}D mode.\",\n deferred=True,\n shape=data.shape,\n texture_size=MAX_TEXTURE_SIZE,\n ndisplay=self.layer._ndisplay,\n )\n )\n warnings.warn(\n trans._(\n \"data shape {shape} exceeds GL_MAX_TEXTURE_SIZE {texture_size} in at least one axis and will be downsampled. Rendering is currently in {ndisplay}D mode.\",\n deferred=True,\n shape=data.shape,\n texture_size=MAX_TEXTURE_SIZE,\n ndisplay=self.layer._ndisplay,\n )\n )\n downsample = np.ceil(\n np.divide(data.shape, MAX_TEXTURE_SIZE)\n ).astype(int)\n scale = np.ones(self.layer.ndim)\n for i, d in enumerate(self.layer._dims_displayed):\n scale[d] = downsample[i]\n self.layer._transforms['tile2data'].scale = scale\n self._on_matrix_change()\n slices = tuple(slice(None, None, ds) for ds in downsample)\n data = data[slices]\n return data\n", "path": "napari/_vispy/layers/image.py"}]} | 3,386 | 487 |
gh_patches_debug_3675 | rasdani/github-patches | git_diff | conan-io__conan-center-index-8132 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
[request] perfetto/v21.0
### Package Details
* Package Name/Version: **perfetto/v21.0**
* Changelog: **https://github.com/google/perfetto/releases/tag/v21.0**
The above mentioned version is newly released by the upstream project and not yet available as a recipe. PR follows
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `recipes/perfetto/all/conanfile.py`
Content:
```
1 from conans import ConanFile, CMake, tools
2 from conans.errors import ConanInvalidConfiguration
3
4 import os
5
6 required_conan_version = ">=1.33.0"
7
8
9 class PerfettoConan(ConanFile):
10 name = "perfetto"
11 license = "Apache-2.0"
12 homepage = "https://perfetto.dev"
13 url = "https://github.com/conan-io/conan-center-index"
14 description = "Performance instrumentation and tracing for Android, Linux and Chrome"
15 topics = ("linux", "profiling", "tracing")
16 settings = "os", "compiler", "build_type", "arch"
17 options = {
18 "shared": [True, False],
19 "fPIC": [True, False]
20 }
21 default_options = {
22 "shared": False,
23 "fPIC": True
24 }
25
26 exports_sources = ["CMakeLists.txt"]
27 generators = "cmake"
28
29 _cmake = None
30
31 @property
32 def _source_subfolder(self):
33 return "source_subfolder"
34
35 def config_options(self):
36 if self.settings.os == "Windows":
37 del self.options.fPIC
38
39 def configure(self):
40 if self.options.shared:
41 del self.options.fPIC
42
43 def validate(self):
44 if self.settings.compiler == "gcc" and tools.Version(self.settings.compiler.version) < 7:
45 raise ConanInvalidConfiguration ("perfetto requires gcc >= 7")
46 if self.settings.compiler.cppstd:
47 tools.check_min_cppstd(self, 11)
48
49 def source(self):
50 tools.get(**self.conan_data["sources"][self.version],
51 strip_root=True, destination=self._source_subfolder)
52
53 def _configure_cmake(self):
54 if self._cmake:
55 return self._cmake
56 self._cmake = CMake(self)
57 self._cmake.configure()
58 return self._cmake
59
60 def build(self):
61 cmake = self._configure_cmake()
62 cmake.build()
63
64 def package(self):
65 self.copy("LICENSE", src=self._source_subfolder, dst="licenses")
66 cmake = self._configure_cmake()
67 cmake.install()
68
69 def package_info(self):
70 self.cpp_info.libs = ["perfetto"]
71 self.cpp_info.names["pkgconfig"] = "perfetto"
72 if self.settings.os == "Linux":
73 self.cpp_info.system_libs.append("pthread")
74 if self.settings.os == "Windows":
75 self.cpp_info.system_libs.append("ws2_32")
76
77
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/recipes/perfetto/all/conanfile.py b/recipes/perfetto/all/conanfile.py
--- a/recipes/perfetto/all/conanfile.py
+++ b/recipes/perfetto/all/conanfile.py
@@ -68,7 +68,6 @@
def package_info(self):
self.cpp_info.libs = ["perfetto"]
- self.cpp_info.names["pkgconfig"] = "perfetto"
if self.settings.os == "Linux":
self.cpp_info.system_libs.append("pthread")
if self.settings.os == "Windows":
| {"golden_diff": "diff --git a/recipes/perfetto/all/conanfile.py b/recipes/perfetto/all/conanfile.py\n--- a/recipes/perfetto/all/conanfile.py\n+++ b/recipes/perfetto/all/conanfile.py\n@@ -68,7 +68,6 @@\n \n def package_info(self):\n self.cpp_info.libs = [\"perfetto\"]\n- self.cpp_info.names[\"pkgconfig\"] = \"perfetto\"\n if self.settings.os == \"Linux\":\n self.cpp_info.system_libs.append(\"pthread\")\n if self.settings.os == \"Windows\":\n", "issue": "[request] perfetto/v21.0\n### Package Details\r\n * Package Name/Version: **perfetto/v21.0**\r\n * Changelog: **https://github.com/google/perfetto/releases/tag/v21.0**\r\n\r\n\r\nThe above mentioned version is newly released by the upstream project and not yet available as a recipe. PR follows\r\n\n", "before_files": [{"content": "from conans import ConanFile, CMake, tools\nfrom conans.errors import ConanInvalidConfiguration\n\nimport os\n\nrequired_conan_version = \">=1.33.0\"\n\n\nclass PerfettoConan(ConanFile):\n name = \"perfetto\"\n license = \"Apache-2.0\"\n homepage = \"https://perfetto.dev\"\n url = \"https://github.com/conan-io/conan-center-index\"\n description = \"Performance instrumentation and tracing for Android, Linux and Chrome\"\n topics = (\"linux\", \"profiling\", \"tracing\")\n settings = \"os\", \"compiler\", \"build_type\", \"arch\"\n options = {\n \"shared\": [True, False],\n \"fPIC\": [True, False]\n }\n default_options = {\n \"shared\": False,\n \"fPIC\": True\n }\n\n exports_sources = [\"CMakeLists.txt\"]\n generators = \"cmake\"\n\n _cmake = None\n\n @property\n def _source_subfolder(self):\n return \"source_subfolder\"\n\n def config_options(self):\n if self.settings.os == \"Windows\":\n del self.options.fPIC\n\n def configure(self):\n if self.options.shared:\n del self.options.fPIC\n\n def validate(self):\n if self.settings.compiler == \"gcc\" and tools.Version(self.settings.compiler.version) < 7:\n raise ConanInvalidConfiguration (\"perfetto requires gcc >= 7\")\n if self.settings.compiler.cppstd:\n tools.check_min_cppstd(self, 11)\n\n def source(self):\n tools.get(**self.conan_data[\"sources\"][self.version],\n strip_root=True, destination=self._source_subfolder)\n\n def _configure_cmake(self):\n if self._cmake:\n return self._cmake\n self._cmake = CMake(self)\n self._cmake.configure()\n return self._cmake\n\n def build(self):\n cmake = self._configure_cmake()\n cmake.build()\n\n def package(self):\n self.copy(\"LICENSE\", src=self._source_subfolder, dst=\"licenses\")\n cmake = self._configure_cmake()\n cmake.install()\n\n def package_info(self):\n self.cpp_info.libs = [\"perfetto\"]\n self.cpp_info.names[\"pkgconfig\"] = \"perfetto\"\n if self.settings.os == \"Linux\":\n self.cpp_info.system_libs.append(\"pthread\")\n if self.settings.os == \"Windows\":\n self.cpp_info.system_libs.append(\"ws2_32\")\n\n", "path": "recipes/perfetto/all/conanfile.py"}], "after_files": [{"content": "from conans import ConanFile, CMake, tools\nfrom conans.errors import ConanInvalidConfiguration\n\nimport os\n\nrequired_conan_version = \">=1.33.0\"\n\n\nclass PerfettoConan(ConanFile):\n name = \"perfetto\"\n license = \"Apache-2.0\"\n homepage = \"https://perfetto.dev\"\n url = \"https://github.com/conan-io/conan-center-index\"\n description = \"Performance instrumentation and tracing for Android, Linux and Chrome\"\n topics = (\"linux\", \"profiling\", \"tracing\")\n settings = \"os\", \"compiler\", \"build_type\", \"arch\"\n options = {\n \"shared\": [True, False],\n \"fPIC\": [True, False]\n }\n default_options = {\n \"shared\": False,\n \"fPIC\": True\n }\n\n exports_sources = [\"CMakeLists.txt\"]\n generators = \"cmake\"\n\n _cmake = None\n\n @property\n def _source_subfolder(self):\n return \"source_subfolder\"\n\n def config_options(self):\n if self.settings.os == \"Windows\":\n del self.options.fPIC\n\n def configure(self):\n if self.options.shared:\n del self.options.fPIC\n\n def validate(self):\n if self.settings.compiler == \"gcc\" and tools.Version(self.settings.compiler.version) < 7:\n raise ConanInvalidConfiguration (\"perfetto requires gcc >= 7\")\n if self.settings.compiler.cppstd:\n tools.check_min_cppstd(self, 11)\n\n def source(self):\n tools.get(**self.conan_data[\"sources\"][self.version],\n strip_root=True, destination=self._source_subfolder)\n\n def _configure_cmake(self):\n if self._cmake:\n return self._cmake\n self._cmake = CMake(self)\n self._cmake.configure()\n return self._cmake\n\n def build(self):\n cmake = self._configure_cmake()\n cmake.build()\n\n def package(self):\n self.copy(\"LICENSE\", src=self._source_subfolder, dst=\"licenses\")\n cmake = self._configure_cmake()\n cmake.install()\n\n def package_info(self):\n self.cpp_info.libs = [\"perfetto\"]\n if self.settings.os == \"Linux\":\n self.cpp_info.system_libs.append(\"pthread\")\n if self.settings.os == \"Windows\":\n self.cpp_info.system_libs.append(\"ws2_32\")\n\n", "path": "recipes/perfetto/all/conanfile.py"}]} | 1,029 | 125 |
gh_patches_debug_15591 | rasdani/github-patches | git_diff | plone__Products.CMFPlone-2714 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Bug: `_smtp` is missing at SMTPMailer's __init__ patching what is introduced in zope.sendmail from version 4.1.0
### What I did:
I am working on internal Addon development (adding support for Plone 5.2) , sending mail notification one of the part functionalities. FYI: some how I forget to active MockMailhost.
When I run all my existing unittests and got unexpected errors :
```
File "/home/nazrul/.cache/buildout/eggs/plone.testing-7.0.0-py2.7.egg/plone/testing/zope.py", line 859, in testTearDown
transaction.abort()
File "/home/nazrul/.cache/buildout/eggs/transaction-2.4.0-py2.7.egg/transaction/_manager.py", line 255, in abort
return self.manager.abort()
File "/home/nazrul/.cache/buildout/eggs/transaction-2.4.0-py2.7.egg/transaction/_manager.py", line 136, in abort
return self.get().abort()
File "/home/nazrul/.cache/buildout/eggs/transaction-2.4.0-py2.7.egg/transaction/_transaction.py", line 529, in abort
reraise(t, v, tb)
File "/home/nazrul/.cache/buildout/eggs/transaction-2.4.0-py2.7.egg/transaction/_transaction.py", line 515, in abort
rm.abort(self)
File "/home/nazrul/.cache/buildout/eggs/zope.sendmail-4.2-py2.7.egg/zope/sendmail/delivery.py", line 57, in abort
self.onAbort()
File "/home/nazrul/.cache/buildout/eggs/zope.sendmail-4.2-py2.7.egg/zope/sendmail/mailer.py", line 78, in abort
if self.connection is None:
File "/home/nazrul/.cache/buildout/eggs/zope.sendmail-4.2-py2.7.egg/zope/sendmail/mailer.py", line 48, in <lambda>
return property(lambda self: getattr(self._smtp, name),
AttributeError: 'SMTPMailer' object has no attribute '_smtp'
```
All tests are passing for earlier version of Plone.
### What is my prediction:
After day long investigation, I found [SMTPMailer __init__ method is patched here](https://github.com/plone/Products.CMFPlone/blob/master/Products/CMFPlone/patches/sendmail.py#L39) , beside Also found that [zope.sendmail from version 4.1.0 the SMTPMailer's __init__](https://github.com/zopefoundation/zope.sendmail/blob/4.1.0/src/zope/sendmail/mailer.py#L45)
has been introduced a new attribute `_smtp` what is ignored during patching.
### How to reproduce:
This is only for Plone 5.2.x
1. disable Mock Mail
2. Try to send email from your tests code
3. Or try send mail from your production/testing server.
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `Products/CMFPlone/patches/sendmail.py`
Content:
```
1 # -*- coding: utf-8 -*-
2 from plone.registry.interfaces import IRegistry
3 from Products.CMFPlone.interfaces import IMailSchema
4 from transaction._transaction import Status
5 from zope.component import getUtility
6 from zope.sendmail.mailer import SMTPMailer
7
8 import logging
9 import transaction
10
11 log = logging.getLogger("MailDataManager")
12
13
14 # BBB remove when zope.sendmail 3.8.0 is released.
15 def catchAllExceptions(func):
16 def _catch(*args, **kwargs):
17 try:
18 return func(*args, **kwargs)
19 except Exception as e:
20 txn = transaction.get()
21 if txn.status == Status.ACTIVE:
22 # sent with immediate=True
23 raise
24 else:
25 # Avoid raising errors during tpc_finish as these could lead to
26 # inconsistent state
27 log.exception(e)
28
29 return _catch
30
31
32 def applyPatches():
33 from zope.sendmail.mailer import SMTPMailer
34 old_mailer = getattr(SMTPMailer, 'vote', None) is None
35 if old_mailer:
36 SMTPMailer.send = catchAllExceptions(SMTPMailer.send)
37
38
39 def new_init(
40 self,
41 hostname='localhost',
42 port=25,
43 username=None,
44 password=None,
45 no_tls=False,
46 force_tls=False):
47
48 registry = getUtility(IRegistry)
49 mail_settings = registry.forInterface(IMailSchema, prefix='plone')
50 self.hostname = mail_settings.smtp_host
51 self.port = mail_settings.smtp_port
52 self.username = mail_settings.smtp_userid
53 self.password = mail_settings.smtp_pass
54 self.force_tls = force_tls
55 self.no_tls = no_tls
56
57 SMTPMailer.__init__ = new_init
58
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/Products/CMFPlone/patches/sendmail.py b/Products/CMFPlone/patches/sendmail.py
--- a/Products/CMFPlone/patches/sendmail.py
+++ b/Products/CMFPlone/patches/sendmail.py
@@ -3,11 +3,13 @@
from Products.CMFPlone.interfaces import IMailSchema
from transaction._transaction import Status
from zope.component import getUtility
+from zope.sendmail.mailer import _SMTPState
from zope.sendmail.mailer import SMTPMailer
import logging
import transaction
+
log = logging.getLogger("MailDataManager")
@@ -53,5 +55,7 @@
self.password = mail_settings.smtp_pass
self.force_tls = force_tls
self.no_tls = no_tls
+ self._smtp = _SMTPState()
+
SMTPMailer.__init__ = new_init
| {"golden_diff": "diff --git a/Products/CMFPlone/patches/sendmail.py b/Products/CMFPlone/patches/sendmail.py\n--- a/Products/CMFPlone/patches/sendmail.py\n+++ b/Products/CMFPlone/patches/sendmail.py\n@@ -3,11 +3,13 @@\n from Products.CMFPlone.interfaces import IMailSchema\n from transaction._transaction import Status\n from zope.component import getUtility\n+from zope.sendmail.mailer import _SMTPState\n from zope.sendmail.mailer import SMTPMailer\n \n import logging\n import transaction\n \n+\n log = logging.getLogger(\"MailDataManager\")\n \n \n@@ -53,5 +55,7 @@\n self.password = mail_settings.smtp_pass\n self.force_tls = force_tls\n self.no_tls = no_tls\n+ self._smtp = _SMTPState()\n+\n \n SMTPMailer.__init__ = new_init\n", "issue": "Bug: `_smtp` is missing at SMTPMailer's __init__ patching what is introduced in zope.sendmail from version 4.1.0\n### What I did:\r\nI am working on internal Addon development (adding support for Plone 5.2) , sending mail notification one of the part functionalities. FYI: some how I forget to active MockMailhost.\r\nWhen I run all my existing unittests and got unexpected errors : \r\n\r\n```\r\nFile \"/home/nazrul/.cache/buildout/eggs/plone.testing-7.0.0-py2.7.egg/plone/testing/zope.py\", line 859, in testTearDown\r\n transaction.abort()\r\n File \"/home/nazrul/.cache/buildout/eggs/transaction-2.4.0-py2.7.egg/transaction/_manager.py\", line 255, in abort\r\n return self.manager.abort()\r\n File \"/home/nazrul/.cache/buildout/eggs/transaction-2.4.0-py2.7.egg/transaction/_manager.py\", line 136, in abort\r\n return self.get().abort()\r\n File \"/home/nazrul/.cache/buildout/eggs/transaction-2.4.0-py2.7.egg/transaction/_transaction.py\", line 529, in abort\r\n reraise(t, v, tb)\r\n File \"/home/nazrul/.cache/buildout/eggs/transaction-2.4.0-py2.7.egg/transaction/_transaction.py\", line 515, in abort\r\n rm.abort(self)\r\n File \"/home/nazrul/.cache/buildout/eggs/zope.sendmail-4.2-py2.7.egg/zope/sendmail/delivery.py\", line 57, in abort\r\n self.onAbort()\r\n File \"/home/nazrul/.cache/buildout/eggs/zope.sendmail-4.2-py2.7.egg/zope/sendmail/mailer.py\", line 78, in abort\r\n if self.connection is None:\r\n File \"/home/nazrul/.cache/buildout/eggs/zope.sendmail-4.2-py2.7.egg/zope/sendmail/mailer.py\", line 48, in <lambda>\r\n return property(lambda self: getattr(self._smtp, name),\r\nAttributeError: 'SMTPMailer' object has no attribute '_smtp'\r\n```\r\nAll tests are passing for earlier version of Plone.\r\n\r\n### What is my prediction: \r\nAfter day long investigation, I found [SMTPMailer __init__ method is patched here](https://github.com/plone/Products.CMFPlone/blob/master/Products/CMFPlone/patches/sendmail.py#L39) , beside Also found that [zope.sendmail from version 4.1.0 the SMTPMailer's __init__](https://github.com/zopefoundation/zope.sendmail/blob/4.1.0/src/zope/sendmail/mailer.py#L45)\r\nhas been introduced a new attribute `_smtp` what is ignored during patching.\r\n\r\n### How to reproduce:\r\n\r\nThis is only for Plone 5.2.x\r\n\r\n1. disable Mock Mail\r\n2. Try to send email from your tests code \r\n3. Or try send mail from your production/testing server.\r\n\r\n\r\n\n", "before_files": [{"content": "# -*- coding: utf-8 -*-\nfrom plone.registry.interfaces import IRegistry\nfrom Products.CMFPlone.interfaces import IMailSchema\nfrom transaction._transaction import Status\nfrom zope.component import getUtility\nfrom zope.sendmail.mailer import SMTPMailer\n\nimport logging\nimport transaction\n\nlog = logging.getLogger(\"MailDataManager\")\n\n\n# BBB remove when zope.sendmail 3.8.0 is released.\ndef catchAllExceptions(func):\n def _catch(*args, **kwargs):\n try:\n return func(*args, **kwargs)\n except Exception as e:\n txn = transaction.get()\n if txn.status == Status.ACTIVE:\n # sent with immediate=True\n raise\n else:\n # Avoid raising errors during tpc_finish as these could lead to\n # inconsistent state\n log.exception(e)\n\n return _catch\n\n\ndef applyPatches():\n from zope.sendmail.mailer import SMTPMailer\n old_mailer = getattr(SMTPMailer, 'vote', None) is None\n if old_mailer:\n SMTPMailer.send = catchAllExceptions(SMTPMailer.send)\n\n\ndef new_init(\n self,\n hostname='localhost',\n port=25,\n username=None,\n password=None,\n no_tls=False,\n force_tls=False):\n\n registry = getUtility(IRegistry)\n mail_settings = registry.forInterface(IMailSchema, prefix='plone')\n self.hostname = mail_settings.smtp_host\n self.port = mail_settings.smtp_port\n self.username = mail_settings.smtp_userid\n self.password = mail_settings.smtp_pass\n self.force_tls = force_tls\n self.no_tls = no_tls\n\nSMTPMailer.__init__ = new_init\n", "path": "Products/CMFPlone/patches/sendmail.py"}], "after_files": [{"content": "# -*- coding: utf-8 -*-\nfrom plone.registry.interfaces import IRegistry\nfrom Products.CMFPlone.interfaces import IMailSchema\nfrom transaction._transaction import Status\nfrom zope.component import getUtility\nfrom zope.sendmail.mailer import _SMTPState\nfrom zope.sendmail.mailer import SMTPMailer\n\nimport logging\nimport transaction\n\n\nlog = logging.getLogger(\"MailDataManager\")\n\n\n# BBB remove when zope.sendmail 3.8.0 is released.\ndef catchAllExceptions(func):\n def _catch(*args, **kwargs):\n try:\n return func(*args, **kwargs)\n except Exception as e:\n txn = transaction.get()\n if txn.status == Status.ACTIVE:\n # sent with immediate=True\n raise\n else:\n # Avoid raising errors during tpc_finish as these could lead to\n # inconsistent state\n log.exception(e)\n\n return _catch\n\n\ndef applyPatches():\n from zope.sendmail.mailer import SMTPMailer\n old_mailer = getattr(SMTPMailer, 'vote', None) is None\n if old_mailer:\n SMTPMailer.send = catchAllExceptions(SMTPMailer.send)\n\n\ndef new_init(\n self,\n hostname='localhost',\n port=25,\n username=None,\n password=None,\n no_tls=False,\n force_tls=False):\n\n registry = getUtility(IRegistry)\n mail_settings = registry.forInterface(IMailSchema, prefix='plone')\n self.hostname = mail_settings.smtp_host\n self.port = mail_settings.smtp_port\n self.username = mail_settings.smtp_userid\n self.password = mail_settings.smtp_pass\n self.force_tls = force_tls\n self.no_tls = no_tls\n self._smtp = _SMTPState()\n\n\nSMTPMailer.__init__ = new_init\n", "path": "Products/CMFPlone/patches/sendmail.py"}]} | 1,465 | 197 |
gh_patches_debug_28476 | rasdani/github-patches | git_diff | pantsbuild__pants-13669 | We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Docker environment not passed to docker publish command
**Describe the bug**
The configured `[docker].env_vars` are not passed to `docker publish`.
**Pants version**
2.9.0.dev1
**OS**
Any
**Additional info**
As reported by chenkai036 on [Slack](https://pantsbuild.slack.com/archives/C046T6T9U/p1637248172462800?thread_ts=1637136003.393600&cid=C046T6T9U)
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `src/python/pants/backend/docker/util_rules/docker_binary.py`
Content:
```
1 # Copyright 2021 Pants project contributors (see CONTRIBUTORS.md).
2 # Licensed under the Apache License, Version 2.0 (see LICENSE).
3
4 from __future__ import annotations
5
6 from dataclasses import dataclass
7 from typing import Mapping
8
9 from pants.backend.docker.util_rules.docker_build_args import DockerBuildArgs
10 from pants.engine.fs import Digest
11 from pants.engine.process import (
12 BinaryNotFoundError,
13 BinaryPath,
14 BinaryPathRequest,
15 BinaryPaths,
16 BinaryPathTest,
17 Process,
18 ProcessCacheScope,
19 SearchPath,
20 )
21 from pants.engine.rules import Get, collect_rules, rule
22 from pants.util.logging import LogLevel
23 from pants.util.strutil import pluralize
24
25
26 class DockerBinary(BinaryPath):
27 """The `docker` binary."""
28
29 DEFAULT_SEARCH_PATH = SearchPath(("/usr/bin", "/bin", "/usr/local/bin"))
30
31 def build_image(
32 self,
33 tags: tuple[str, ...],
34 digest: Digest,
35 dockerfile: str | None = None,
36 build_args: DockerBuildArgs | None = None,
37 env: Mapping[str, str] | None = None,
38 ) -> Process:
39 args = [self.path, "build"]
40
41 for tag in tags:
42 args.extend(["-t", tag])
43
44 if build_args:
45 for build_arg in build_args:
46 args.extend(["--build-arg", build_arg])
47
48 if dockerfile:
49 args.extend(["-f", dockerfile])
50
51 # Add build context root.
52 args.append(".")
53
54 return Process(
55 argv=tuple(args),
56 description=(
57 f"Building docker image {tags[0]}"
58 + (f" +{pluralize(len(tags)-1, 'additional tag')}." if len(tags) > 1 else ".")
59 ),
60 env=env,
61 input_digest=digest,
62 cache_scope=ProcessCacheScope.PER_SESSION,
63 )
64
65 def push_image(self, tags: tuple[str, ...]) -> Process | None:
66 if not tags:
67 return None
68
69 return Process(
70 argv=(self.path, "push", *tags),
71 cache_scope=ProcessCacheScope.PER_SESSION,
72 description=f"Pushing docker image {tags[0]}",
73 )
74
75
76 @dataclass(frozen=True)
77 class DockerBinaryRequest:
78 search_path: SearchPath = DockerBinary.DEFAULT_SEARCH_PATH
79
80
81 @rule(desc="Finding the `docker` binary", level=LogLevel.DEBUG)
82 async def find_docker(docker_request: DockerBinaryRequest) -> DockerBinary:
83 request = BinaryPathRequest(
84 binary_name="docker",
85 search_path=docker_request.search_path,
86 test=BinaryPathTest(args=["-v"]),
87 )
88 paths = await Get(BinaryPaths, BinaryPathRequest, request)
89 first_path = paths.first_path
90 if not first_path:
91 raise BinaryNotFoundError.from_request(request, rationale="interact with the docker daemon")
92 return DockerBinary(first_path.path, first_path.fingerprint)
93
94
95 @rule
96 async def get_docker() -> DockerBinary:
97 return await Get(DockerBinary, DockerBinaryRequest())
98
99
100 def rules():
101 return collect_rules()
102
```
Path: `src/python/pants/backend/docker/goals/publish.py`
Content:
```
1 # Copyright 2021 Pants project contributors (see CONTRIBUTORS.md).
2 # Licensed under the Apache License, Version 2.0 (see LICENSE).
3
4 from __future__ import annotations
5
6 import logging
7 from dataclasses import dataclass
8 from itertools import chain
9 from typing import cast
10
11 from pants.backend.docker.goals.package_image import BuiltDockerImage
12 from pants.backend.docker.subsystems.docker_options import DockerOptions
13 from pants.backend.docker.target_types import DockerRegistriesField, DockerSkipPushField
14 from pants.backend.docker.util_rules.docker_binary import DockerBinary
15 from pants.core.goals.publish import (
16 PublishFieldSet,
17 PublishOutputData,
18 PublishPackages,
19 PublishProcesses,
20 PublishRequest,
21 )
22 from pants.engine.process import InteractiveProcess
23 from pants.engine.rules import collect_rules, rule
24
25 logger = logging.getLogger(__name__)
26
27
28 class PublishDockerImageRequest(PublishRequest):
29 pass
30
31
32 @dataclass(frozen=True)
33 class PublishDockerImageFieldSet(PublishFieldSet):
34 publish_request_type = PublishDockerImageRequest
35 required_fields = (DockerRegistriesField,)
36
37 registries: DockerRegistriesField
38 skip_push: DockerSkipPushField
39
40 def get_output_data(self) -> PublishOutputData:
41 return PublishOutputData(
42 {
43 "publisher": "docker",
44 "registries": self.registries.value or (),
45 **super().get_output_data(),
46 }
47 )
48
49
50 @rule
51 async def push_docker_images(
52 request: PublishDockerImageRequest, docker: DockerBinary, options: DockerOptions
53 ) -> PublishProcesses:
54 tags = tuple(
55 chain.from_iterable(
56 cast(BuiltDockerImage, image).tags
57 for pkg in request.packages
58 for image in pkg.artifacts
59 )
60 )
61
62 if request.field_set.skip_push.value:
63 return PublishProcesses(
64 [
65 PublishPackages(
66 names=tags,
67 description=f"(by `{request.field_set.skip_push.alias}` on {request.field_set.address})",
68 ),
69 ]
70 )
71
72 process = docker.push_image(tags)
73 return PublishProcesses(
74 [
75 PublishPackages(
76 names=tags,
77 process=InteractiveProcess.from_process(process) if process else None,
78 ),
79 ]
80 )
81
82
83 def rules():
84 return (
85 *collect_rules(),
86 *PublishDockerImageFieldSet.rules(),
87 )
88
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
| diff --git a/src/python/pants/backend/docker/goals/publish.py b/src/python/pants/backend/docker/goals/publish.py
--- a/src/python/pants/backend/docker/goals/publish.py
+++ b/src/python/pants/backend/docker/goals/publish.py
@@ -19,8 +19,9 @@
PublishProcesses,
PublishRequest,
)
+from pants.engine.environment import Environment, EnvironmentRequest
from pants.engine.process import InteractiveProcess
-from pants.engine.rules import collect_rules, rule
+from pants.engine.rules import Get, collect_rules, rule
logger = logging.getLogger(__name__)
@@ -69,7 +70,8 @@
]
)
- process = docker.push_image(tags)
+ env = await Get(Environment, EnvironmentRequest(options.env_vars))
+ process = docker.push_image(tags, env)
return PublishProcesses(
[
PublishPackages(
diff --git a/src/python/pants/backend/docker/util_rules/docker_binary.py b/src/python/pants/backend/docker/util_rules/docker_binary.py
--- a/src/python/pants/backend/docker/util_rules/docker_binary.py
+++ b/src/python/pants/backend/docker/util_rules/docker_binary.py
@@ -62,7 +62,9 @@
cache_scope=ProcessCacheScope.PER_SESSION,
)
- def push_image(self, tags: tuple[str, ...]) -> Process | None:
+ def push_image(
+ self, tags: tuple[str, ...], env: Mapping[str, str] | None = None
+ ) -> Process | None:
if not tags:
return None
@@ -70,6 +72,7 @@
argv=(self.path, "push", *tags),
cache_scope=ProcessCacheScope.PER_SESSION,
description=f"Pushing docker image {tags[0]}",
+ env=env,
)
| {"golden_diff": "diff --git a/src/python/pants/backend/docker/goals/publish.py b/src/python/pants/backend/docker/goals/publish.py\n--- a/src/python/pants/backend/docker/goals/publish.py\n+++ b/src/python/pants/backend/docker/goals/publish.py\n@@ -19,8 +19,9 @@\n PublishProcesses,\n PublishRequest,\n )\n+from pants.engine.environment import Environment, EnvironmentRequest\n from pants.engine.process import InteractiveProcess\n-from pants.engine.rules import collect_rules, rule\n+from pants.engine.rules import Get, collect_rules, rule\n \n logger = logging.getLogger(__name__)\n \n@@ -69,7 +70,8 @@\n ]\n )\n \n- process = docker.push_image(tags)\n+ env = await Get(Environment, EnvironmentRequest(options.env_vars))\n+ process = docker.push_image(tags, env)\n return PublishProcesses(\n [\n PublishPackages(\ndiff --git a/src/python/pants/backend/docker/util_rules/docker_binary.py b/src/python/pants/backend/docker/util_rules/docker_binary.py\n--- a/src/python/pants/backend/docker/util_rules/docker_binary.py\n+++ b/src/python/pants/backend/docker/util_rules/docker_binary.py\n@@ -62,7 +62,9 @@\n cache_scope=ProcessCacheScope.PER_SESSION,\n )\n \n- def push_image(self, tags: tuple[str, ...]) -> Process | None:\n+ def push_image(\n+ self, tags: tuple[str, ...], env: Mapping[str, str] | None = None\n+ ) -> Process | None:\n if not tags:\n return None\n \n@@ -70,6 +72,7 @@\n argv=(self.path, \"push\", *tags),\n cache_scope=ProcessCacheScope.PER_SESSION,\n description=f\"Pushing docker image {tags[0]}\",\n+ env=env,\n )\n", "issue": "Docker environment not passed to docker publish command\n**Describe the bug**\r\nThe configured `[docker].env_vars` are not passed to `docker publish`.\r\n\r\n**Pants version**\r\n2.9.0.dev1\r\n\r\n**OS**\r\nAny\r\n\r\n**Additional info**\r\nAs reported by chenkai036 on [Slack](https://pantsbuild.slack.com/archives/C046T6T9U/p1637248172462800?thread_ts=1637136003.393600&cid=C046T6T9U)\r\n\n", "before_files": [{"content": "# Copyright 2021 Pants project contributors (see CONTRIBUTORS.md).\n# Licensed under the Apache License, Version 2.0 (see LICENSE).\n\nfrom __future__ import annotations\n\nfrom dataclasses import dataclass\nfrom typing import Mapping\n\nfrom pants.backend.docker.util_rules.docker_build_args import DockerBuildArgs\nfrom pants.engine.fs import Digest\nfrom pants.engine.process import (\n BinaryNotFoundError,\n BinaryPath,\n BinaryPathRequest,\n BinaryPaths,\n BinaryPathTest,\n Process,\n ProcessCacheScope,\n SearchPath,\n)\nfrom pants.engine.rules import Get, collect_rules, rule\nfrom pants.util.logging import LogLevel\nfrom pants.util.strutil import pluralize\n\n\nclass DockerBinary(BinaryPath):\n \"\"\"The `docker` binary.\"\"\"\n\n DEFAULT_SEARCH_PATH = SearchPath((\"/usr/bin\", \"/bin\", \"/usr/local/bin\"))\n\n def build_image(\n self,\n tags: tuple[str, ...],\n digest: Digest,\n dockerfile: str | None = None,\n build_args: DockerBuildArgs | None = None,\n env: Mapping[str, str] | None = None,\n ) -> Process:\n args = [self.path, \"build\"]\n\n for tag in tags:\n args.extend([\"-t\", tag])\n\n if build_args:\n for build_arg in build_args:\n args.extend([\"--build-arg\", build_arg])\n\n if dockerfile:\n args.extend([\"-f\", dockerfile])\n\n # Add build context root.\n args.append(\".\")\n\n return Process(\n argv=tuple(args),\n description=(\n f\"Building docker image {tags[0]}\"\n + (f\" +{pluralize(len(tags)-1, 'additional tag')}.\" if len(tags) > 1 else \".\")\n ),\n env=env,\n input_digest=digest,\n cache_scope=ProcessCacheScope.PER_SESSION,\n )\n\n def push_image(self, tags: tuple[str, ...]) -> Process | None:\n if not tags:\n return None\n\n return Process(\n argv=(self.path, \"push\", *tags),\n cache_scope=ProcessCacheScope.PER_SESSION,\n description=f\"Pushing docker image {tags[0]}\",\n )\n\n\n@dataclass(frozen=True)\nclass DockerBinaryRequest:\n search_path: SearchPath = DockerBinary.DEFAULT_SEARCH_PATH\n\n\n@rule(desc=\"Finding the `docker` binary\", level=LogLevel.DEBUG)\nasync def find_docker(docker_request: DockerBinaryRequest) -> DockerBinary:\n request = BinaryPathRequest(\n binary_name=\"docker\",\n search_path=docker_request.search_path,\n test=BinaryPathTest(args=[\"-v\"]),\n )\n paths = await Get(BinaryPaths, BinaryPathRequest, request)\n first_path = paths.first_path\n if not first_path:\n raise BinaryNotFoundError.from_request(request, rationale=\"interact with the docker daemon\")\n return DockerBinary(first_path.path, first_path.fingerprint)\n\n\n@rule\nasync def get_docker() -> DockerBinary:\n return await Get(DockerBinary, DockerBinaryRequest())\n\n\ndef rules():\n return collect_rules()\n", "path": "src/python/pants/backend/docker/util_rules/docker_binary.py"}, {"content": "# Copyright 2021 Pants project contributors (see CONTRIBUTORS.md).\n# Licensed under the Apache License, Version 2.0 (see LICENSE).\n\nfrom __future__ import annotations\n\nimport logging\nfrom dataclasses import dataclass\nfrom itertools import chain\nfrom typing import cast\n\nfrom pants.backend.docker.goals.package_image import BuiltDockerImage\nfrom pants.backend.docker.subsystems.docker_options import DockerOptions\nfrom pants.backend.docker.target_types import DockerRegistriesField, DockerSkipPushField\nfrom pants.backend.docker.util_rules.docker_binary import DockerBinary\nfrom pants.core.goals.publish import (\n PublishFieldSet,\n PublishOutputData,\n PublishPackages,\n PublishProcesses,\n PublishRequest,\n)\nfrom pants.engine.process import InteractiveProcess\nfrom pants.engine.rules import collect_rules, rule\n\nlogger = logging.getLogger(__name__)\n\n\nclass PublishDockerImageRequest(PublishRequest):\n pass\n\n\n@dataclass(frozen=True)\nclass PublishDockerImageFieldSet(PublishFieldSet):\n publish_request_type = PublishDockerImageRequest\n required_fields = (DockerRegistriesField,)\n\n registries: DockerRegistriesField\n skip_push: DockerSkipPushField\n\n def get_output_data(self) -> PublishOutputData:\n return PublishOutputData(\n {\n \"publisher\": \"docker\",\n \"registries\": self.registries.value or (),\n **super().get_output_data(),\n }\n )\n\n\n@rule\nasync def push_docker_images(\n request: PublishDockerImageRequest, docker: DockerBinary, options: DockerOptions\n) -> PublishProcesses:\n tags = tuple(\n chain.from_iterable(\n cast(BuiltDockerImage, image).tags\n for pkg in request.packages\n for image in pkg.artifacts\n )\n )\n\n if request.field_set.skip_push.value:\n return PublishProcesses(\n [\n PublishPackages(\n names=tags,\n description=f\"(by `{request.field_set.skip_push.alias}` on {request.field_set.address})\",\n ),\n ]\n )\n\n process = docker.push_image(tags)\n return PublishProcesses(\n [\n PublishPackages(\n names=tags,\n process=InteractiveProcess.from_process(process) if process else None,\n ),\n ]\n )\n\n\ndef rules():\n return (\n *collect_rules(),\n *PublishDockerImageFieldSet.rules(),\n )\n", "path": "src/python/pants/backend/docker/goals/publish.py"}], "after_files": [{"content": "# Copyright 2021 Pants project contributors (see CONTRIBUTORS.md).\n# Licensed under the Apache License, Version 2.0 (see LICENSE).\n\nfrom __future__ import annotations\n\nfrom dataclasses import dataclass\nfrom typing import Mapping\n\nfrom pants.backend.docker.util_rules.docker_build_args import DockerBuildArgs\nfrom pants.engine.fs import Digest\nfrom pants.engine.process import (\n BinaryNotFoundError,\n BinaryPath,\n BinaryPathRequest,\n BinaryPaths,\n BinaryPathTest,\n Process,\n ProcessCacheScope,\n SearchPath,\n)\nfrom pants.engine.rules import Get, collect_rules, rule\nfrom pants.util.logging import LogLevel\nfrom pants.util.strutil import pluralize\n\n\nclass DockerBinary(BinaryPath):\n \"\"\"The `docker` binary.\"\"\"\n\n DEFAULT_SEARCH_PATH = SearchPath((\"/usr/bin\", \"/bin\", \"/usr/local/bin\"))\n\n def build_image(\n self,\n tags: tuple[str, ...],\n digest: Digest,\n dockerfile: str | None = None,\n build_args: DockerBuildArgs | None = None,\n env: Mapping[str, str] | None = None,\n ) -> Process:\n args = [self.path, \"build\"]\n\n for tag in tags:\n args.extend([\"-t\", tag])\n\n if build_args:\n for build_arg in build_args:\n args.extend([\"--build-arg\", build_arg])\n\n if dockerfile:\n args.extend([\"-f\", dockerfile])\n\n # Add build context root.\n args.append(\".\")\n\n return Process(\n argv=tuple(args),\n description=(\n f\"Building docker image {tags[0]}\"\n + (f\" +{pluralize(len(tags)-1, 'additional tag')}.\" if len(tags) > 1 else \".\")\n ),\n env=env,\n input_digest=digest,\n cache_scope=ProcessCacheScope.PER_SESSION,\n )\n\n def push_image(\n self, tags: tuple[str, ...], env: Mapping[str, str] | None = None\n ) -> Process | None:\n if not tags:\n return None\n\n return Process(\n argv=(self.path, \"push\", *tags),\n cache_scope=ProcessCacheScope.PER_SESSION,\n description=f\"Pushing docker image {tags[0]}\",\n env=env,\n )\n\n\n@dataclass(frozen=True)\nclass DockerBinaryRequest:\n search_path: SearchPath = DockerBinary.DEFAULT_SEARCH_PATH\n\n\n@rule(desc=\"Finding the `docker` binary\", level=LogLevel.DEBUG)\nasync def find_docker(docker_request: DockerBinaryRequest) -> DockerBinary:\n request = BinaryPathRequest(\n binary_name=\"docker\",\n search_path=docker_request.search_path,\n test=BinaryPathTest(args=[\"-v\"]),\n )\n paths = await Get(BinaryPaths, BinaryPathRequest, request)\n first_path = paths.first_path\n if not first_path:\n raise BinaryNotFoundError.from_request(request, rationale=\"interact with the docker daemon\")\n return DockerBinary(first_path.path, first_path.fingerprint)\n\n\n@rule\nasync def get_docker() -> DockerBinary:\n return await Get(DockerBinary, DockerBinaryRequest())\n\n\ndef rules():\n return collect_rules()\n", "path": "src/python/pants/backend/docker/util_rules/docker_binary.py"}, {"content": "# Copyright 2021 Pants project contributors (see CONTRIBUTORS.md).\n# Licensed under the Apache License, Version 2.0 (see LICENSE).\n\nfrom __future__ import annotations\n\nimport logging\nfrom dataclasses import dataclass\nfrom itertools import chain\nfrom typing import cast\n\nfrom pants.backend.docker.goals.package_image import BuiltDockerImage\nfrom pants.backend.docker.subsystems.docker_options import DockerOptions\nfrom pants.backend.docker.target_types import DockerRegistriesField, DockerSkipPushField\nfrom pants.backend.docker.util_rules.docker_binary import DockerBinary\nfrom pants.core.goals.publish import (\n PublishFieldSet,\n PublishOutputData,\n PublishPackages,\n PublishProcesses,\n PublishRequest,\n)\nfrom pants.engine.environment import Environment, EnvironmentRequest\nfrom pants.engine.process import InteractiveProcess\nfrom pants.engine.rules import Get, collect_rules, rule\n\nlogger = logging.getLogger(__name__)\n\n\nclass PublishDockerImageRequest(PublishRequest):\n pass\n\n\n@dataclass(frozen=True)\nclass PublishDockerImageFieldSet(PublishFieldSet):\n publish_request_type = PublishDockerImageRequest\n required_fields = (DockerRegistriesField,)\n\n registries: DockerRegistriesField\n skip_push: DockerSkipPushField\n\n def get_output_data(self) -> PublishOutputData:\n return PublishOutputData(\n {\n \"publisher\": \"docker\",\n \"registries\": self.registries.value or (),\n **super().get_output_data(),\n }\n )\n\n\n@rule\nasync def push_docker_images(\n request: PublishDockerImageRequest, docker: DockerBinary, options: DockerOptions\n) -> PublishProcesses:\n tags = tuple(\n chain.from_iterable(\n cast(BuiltDockerImage, image).tags\n for pkg in request.packages\n for image in pkg.artifacts\n )\n )\n\n if request.field_set.skip_push.value:\n return PublishProcesses(\n [\n PublishPackages(\n names=tags,\n description=f\"(by `{request.field_set.skip_push.alias}` on {request.field_set.address})\",\n ),\n ]\n )\n\n env = await Get(Environment, EnvironmentRequest(options.env_vars))\n process = docker.push_image(tags, env)\n return PublishProcesses(\n [\n PublishPackages(\n names=tags,\n process=InteractiveProcess.from_process(process) if process else None,\n ),\n ]\n )\n\n\ndef rules():\n return (\n *collect_rules(),\n *PublishDockerImageFieldSet.rules(),\n )\n", "path": "src/python/pants/backend/docker/goals/publish.py"}]} | 1,962 | 391 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.