message
stringlengths
13
484
diff
stringlengths
38
4.63k
Handle non-scalar tensors Fixes
@@ -109,7 +109,13 @@ class ScalarReader(object): except ImportError as e: log.debug("error importing make_ndarray: %s", e) raise util.TryFailed() - return val.tag, make_ndarray(val.tensor).item(), event.step + ndarray = make_ndarray(val.tensor) + try: + scalar_val = ndarray.item() + except ValueError: + return None + else: + return val.tag, scalar_val, event.step @staticmethod def _try_tfevent_v1(event, val):
Move Save button in hamburger menu This cleans up the header bar a little.
@@ -50,6 +50,7 @@ def create_hamburger_model(export_menu, tools_menu): model.append_section(None, part) part = Gio.Menu.new() + part.append(gettext("Save"), "win.file-save") part.append(gettext("Save As..."), "win.file-save-as") part.append_submenu(gettext("Export"), export_menu) model.append_section(None, part) @@ -179,7 +180,6 @@ class MainWindow(Service, ActionProvider): create_hamburger_model(self.export_menu.menu, self.tools_menu.menu) ) ) - header.pack_end(button(gettext("Save"), "win.file-save")) b = Gtk.MenuButton.new() image = Gtk.Image.new_from_icon_name(
Add delete_after to Interaction.edit_message Closes
@@ -806,6 +806,7 @@ class InteractionResponse: attachments: Sequence[Union[Attachment, File]] = MISSING, view: Optional[View] = MISSING, allowed_mentions: Optional[AllowedMentions] = MISSING, + delete_after: Optional[float] = None, ) -> None: """|coro| @@ -835,6 +836,12 @@ class InteractionResponse: allowed_mentions: Optional[:class:`~discord.AllowedMentions`] Controls the mentions being processed in this message. See :meth:`.Message.edit` for more information. + delete_after: :class:`float` + If provided, the number of seconds to wait in the background + before deleting the message we just edited. If the deletion fails, + then it is silently ignored. + + .. versionadded:: 2.2 Raises ------- @@ -885,6 +892,17 @@ class InteractionResponse: self._response_type = InteractionResponseType.message_update + if delete_after is not None: + + async def inner_call(delay: float = delete_after): + await asyncio.sleep(delay) + try: + await self._parent.delete_original_response() + except HTTPException: + pass + + asyncio.create_task(inner_call()) + async def send_modal(self, modal: Modal, /) -> None: """|coro|
limit queue size for async richlog close
@@ -13,7 +13,7 @@ from util import cvimage class _richlog_worker(threading.Thread): def __init__(self, *args, **kwargs): super().__init__(*args, **kwargs) - self.queue = Queue() + self.queue = Queue(32) self.lock = threading.Lock() self.files = {} self.daemon = True
add Jun 8 Seattle incident, woman stops breathing The account from reddit was posted by one of the medics involved.
@@ -149,6 +149,14 @@ id: wa-seattle-12 * https://www.reddit.com/r/2020PoliceBrutality/comments/gycscp/cant_go_1_day_without_teargaslighting_us/?utm_source=share&utm_medium=web2x * https://www.forbes.com/sites/jemimamcevoy/2020/06/08/seattle-police-use-tear-gas-against-protestors-despite-city-ban/#7e98a1d5b4bc +### Woman hit with police projectile resumes breathing after CPR, delivered to hospital | June 7th + +Medics attempt to treat "a young woman in her 20s who had taken a police projectile to her chest" as police throw flash bang grenades directly into the first aid center. There were several cycles of the victim no longer responding or breathing but becoming responsive again after CPR. 911 could not send an ambulance. Eventually a civilian transported her by car to the ED where she was delivered with a pulse. + +**Links** + +* [Firsthand account from the medic](https://www.reddit.com/r/Seattle/comments/gywxhz/folks_i_need_your_help/) + ### Additional Seattle OPA Case Numbers * 2020OPA-0323 - May 29 - "Punching a person on the ground who was being arrested"
Fix rebase issue with CHANGES.txt Rebasing off of master put my CHANGES.txt addition in an old release. This commit moves it up to the correct section.
@@ -15,6 +15,10 @@ RELEASE VERSION/DATE TO BE FILLED IN LATER From Adam Gross: - Fix minor bug affecting SCons.Node.FS.File.get_csig()'s usage of the MD5 chunksize. User-facing behavior does not change with this fix (GH Issue #3726). + - Added support for a new command-line parameter "--hash-format" to override the default + hash format that SCons uses. It can also be set via SetOption('hash_format'). Supported + values include md5, sha1, and sha256, but you can also use any other algorithm that is + offered by your Python interpreter's hashlib package. From Mats Wichmann: - Complete tests for Dictionary, env.keys() and env.values() for @@ -134,10 +138,6 @@ RELEASE 4.0.0 - Sat, 04 Jul 2020 12:00:27 +0000 when the value can't be converted to a string or if having a name is otherwise desirable. - Fixed usage of abspath and path for RootDir objects on Windows. Previously env.fs.Dir("T:").abspath would return "T:\T:" and now it correctly returns "T:". - - Added support for a new command-line parameter "--hash-format" to override the default - hash format that SCons uses. It can also be set via SetOption('hash_format'). Supported - values include md5, sha1, and sha256, but you can also use any other algorithm that is - offered by your Python interpreter's hashlib package. From Ivan Kravets, PlatformIO - New conditional C Scanner (`SCons.Scanner.C.CConditionalScanner()`)
Validate.py Modified validate to share a bit more info on the score that was incorrect.
@@ -176,7 +176,7 @@ def validate_standard_fields(object, uuids): if 'impact' in object['tags'] and 'confidence' in object['tags']: calculated_risk_score = int(((object['tags']['impact'])*(object['tags']['confidence']))/100) if calculated_risk_score != object['tags']['risk_score']: - errors.append("ERROR: risk_score not calulated correctly and it should be set as: %s" % calculated_risk_score) + errors.append("ERROR: risk_score not calulated correctly and it should be set to %s for " % calculated_risk_score + object['name']) return errors, uuids
Changes to the time.strftime to make it more accurate Since some host files might need to be updated more frequently it's more appropriate to show timezone and the exact time for last build, hince changed the order of month and date to %d %B
@@ -925,7 +925,7 @@ def write_opening_header(final_file, **header_params): write_data(final_file, "# This hosts file is a merged collection " "of hosts from reputable sources,\n") write_data(final_file, "# with a dash of crowd sourcing via Github\n#\n") - write_data(final_file, "# Date: " + time.strftime("%B %d %Y", time.gmtime()) + "\n") + write_data(final_file, "# Date: " + time.strftime("%d %B %Y %H:%M:%S (UTC)", time.gmtime()) + "\n") if header_params["extensions"]: write_data(final_file, "# Extensions added to this file: " + ", ".join(
FIX: Use genfromtxt to read channel names from .locs Channel names read from .loc, .locs, and .eloc files were truncated to 4 character strings ("S4" numpy dtype). Using genfromtxt with "str" as dtype lets numpy determine the required length and allows for reading unicode characters as well.
@@ -313,8 +313,7 @@ def read_montage(kind, ch_names=None, path=None, unit='m', transform=False): ch_names_ = data['name'].astype(str).tolist() pos = np.vstack((data['x'], data['y'], data['z'])).T elif ext in ('.loc', '.locs', '.eloc'): - ch_names_ = np.loadtxt(fname, dtype='S4', - usecols=[3]).astype(str).tolist() + ch_names_ = np.genfromtxt(fname, dtype=str, usecols=3).tolist() topo = np.loadtxt(fname, dtype=float, usecols=[1, 2]) sph = _topo_to_sph(topo) pos = _sph_to_cart(sph)
Fix some invalid French phone numbers starting with "+33 8x" The full range of phone number starting with 08 isn't valid (eg. 083xx is not valid) so this patch limits the range to valid numbers only. French validator used by the phonenumbers project:
@@ -11,7 +11,7 @@ class Provider(PhoneNumberProvider): '+33 (0)4 ## ## ## ##', '+33 (0)5 ## ## ## ##', '+33 (0)6 ## ## ## ##', - '+33 (0)8 ## ## ## ##', + '+33 (0)8 0# ## ## ##', '+33 1 ## ## ## ##', '+33 1 ## ## ## ##', '+33 2 ## ## ## ##', @@ -19,7 +19,7 @@ class Provider(PhoneNumberProvider): '+33 4 ## ## ## ##', '+33 5 ## ## ## ##', '+33 6 ## ## ## ##', - '+33 8 ## ## ## ##', + '+33 8 0# ## ## ##', '01########', '01########', '02########', @@ -27,7 +27,7 @@ class Provider(PhoneNumberProvider): '04########', '05########', '06########', - '08########', + '080#######', '01 ## ## ## ##', '01 ## ## ## ##', '02 ## ## ## ##', @@ -35,5 +35,5 @@ class Provider(PhoneNumberProvider): '04 ## ## ## ##', '05 ## ## ## ##', '06 ## ## ## ##', - '08 ## ## ## ##', + '08 0# ## ## ##', )
add max_keep_size Summary: Set max_keep_size to filter long utterances. Needed when trained on labeled datasets with long utterances. Pull Request resolved:
@@ -76,6 +76,10 @@ class HubertPretrainingConfig(FairseqDataclass): default=False, metadata={"help": "pad shorter samples instead of cropping"}, ) + max_keep_size: Optional[int] = field( + default=None, + metadata={"help": "exclude sample longer than this"}, + ) max_sample_size: Optional[int] = field( default=None, metadata={"help": "max sample size to crop to for batching"}, @@ -123,13 +127,11 @@ class HubertPretrainingTask(FairseqTask): else: self.state.add_factory("dictionaries", self.load_dictionaries) - self._source_dictionary = None - self.blank_symbol = "<s>" @property def source_dictionary(self) -> Optional[Dictionary]: - return self._source_dictionary + return None @property def target_dictionary(self) -> Optional[Dictionary]: @@ -174,7 +176,7 @@ class HubertPretrainingTask(FairseqTask): pad_list=pad_list, eos_list=eos_list, label_processors=procs, - max_keep_sample_size=None, + max_keep_sample_size=self.cfg.max_keep_size, min_keep_sample_size=self.cfg.min_sample_size, max_sample_size=self.cfg.max_sample_size, pad_audio=self.cfg.pad_audio,
boot: Fix ill-advised change from config component When the config component was added, a change was made in boot that causes problems. And I think was just wrong?
@@ -151,7 +151,6 @@ def _do_imports (components): return done -_inst = {} def _do_launch (argv, skip_startup=False): component_order = [] @@ -194,7 +193,7 @@ def _do_launch (argv, skip_startup=False): if modules is False: return False - inst = _inst + inst = {} for name in component_order: cname = name inst[name] = inst.get(name, -1) + 1
Add checking for 404 errors that aborts create project My apologies. I fixed my version, then added the code in your __init__.py to loop and wait for the not found error to disappear. By then, however, the error stopped occurring and I didn't notice that errors.py wasn't recognizing the error.
@@ -328,6 +328,10 @@ def get_gapi_error_detail(e, message = error['error']['errors'][0]['message'] except KeyError: message = error['error']['message'] + if http_status == 404: + if 'Requested entity was not found' in message or 'does not exist' in message: + error = _create_http_error_dict(404, ErrorReason.NOT_FOUND.value, + message) else: if 'error_description' in error: if error['error_description'] == 'Invalid Value':
Update validate-credit-card-number-with-luhn-algorithm.yml Adding digital lookup via negative numbers
@@ -12,6 +12,7 @@ rule: - match: contain loop description: Iterate over CC digits - basic block: + - or: - and: - number: 0x0 - number: 0x2 @@ -24,6 +25,17 @@ rule: - number: 0x7 - number: 0x9 description: Digital root lookup table + - and: + - number: 0x0 + - number: 0x1 + - number: 0x2 + - number: 0x3 + - number: 0x4 + - number: 0xfffffffc + - number: 0xfffffffd + - number: 0xfffffffe + - number: 0xffffffff + description: Digital root lookup table via neg numbers - basic block: - or: - and:
Update summer_2021.yaml Added anione playlist urls
@@ -59,7 +59,7 @@ info: streams: crunchyroll: '' museasia: '' - anione: '' + anione: 'https://www.youtube.com/watch?v=hasJei2WwLk&list=PLxSscENEp7Jjad7wjjOwOoFgbHC5ZXhtG' funimation|Funimation: '' wakanim|Wakanim: '' hidive: '' @@ -178,7 +178,7 @@ info: streams: crunchyroll: '' museasia: '' - anione: '' + anione: 'https://www.youtube.com/watch?v=aymf48DAXqo&list=PLxSscENEp7JhHilZTy4u4rr_s8-nEM1-8' funimation|Funimation: '' wakanim|Wakanim: '' hidive: 'https://www.hidive.com/tv/getter-robo-arc' @@ -250,7 +250,7 @@ info: streams: crunchyroll: '' museasia: '' - anione: '' + anione: 'https://www.youtube.com/watch?v=MP_5XFigDbE&list=PLxSscENEp7Jh7yn1GT-SfcTUqt5B_A7wf' funimation|Funimation: 'https://www.funimation.com/shows/higurashi-when-they-cry-gou/' wakanim|Wakanim: 'https://www.wakanim.tv/sc/v2/catalogue/show/950' hidive: '' @@ -700,7 +700,7 @@ info: streams: crunchyroll: '' museasia: '' - anione: '' + anione: 'https://www.youtube.com/watch?v=GgqqA3J80eE&list=PLxSscENEp7JhnfTQGx_aB4V2B6mc6a3gx' funimation|Funimation: 'https://www.funimation.com/shows/re-main/' wakanim|Wakanim: 'https://www.wakanim.tv/sc/v2/catalogue/show/1393' hidive: '' @@ -723,7 +723,7 @@ info: streams: crunchyroll: '' museasia: '' - anione: '' + anione: 'https://www.youtube.com/watch?v=IiuZpFogHe0&list=PLxSscENEp7Jg9olNs9aYTSFjc6PM5vrbU' funimation|Funimation: 'https://www.funimation.com/shows/scarlet-nexus/' wakanim|Wakanim: 'https://www.wakanim.tv/sc/v2/catalogue/show/1349' hidive: ''
change: relax urllib3 and requests restrictions. Relax urllib3 and requests restrictions based on and
@@ -39,7 +39,7 @@ required_packages = [ "protobuf>=3.1", "scipy>=0.19.0", "protobuf3-to-dict>=0.1.5", - "requests>=2.20.0, <2.21", + "requests>=2.20.0, <3", "smdebug-rulesconfig==0.1.2", ] @@ -47,7 +47,7 @@ required_packages = [ extras = { "analytics": ["pandas"], "local": [ - "urllib3>=1.21, <1.25", + "urllib3>=1.21.1,<1.26,!=1.25.0,!=1.25.1", "docker-compose>=1.23.0", "PyYAML>=3.10, <5", # PyYAML version has to match docker-compose requirements ],
Update ug015_storm_ref_pivot.rst Updated / fixed refs() to reflect current state.
@@ -117,11 +117,14 @@ Optional parameters: * **in:** return all nodes that have a secondary property *<type> (<ptype>) = <valu>* that is the same as (**references**) any primary *<prop> = <valu>* in the working set of nodes. * **out:** return all the nodes whose primary *<prop> = <valu>* is the same as (is **referenced by**) any secondary property *<type> (<ptype>) = <valu>* in the working set of nodes. * If no parameters are specified, ``refs()`` will return the combined results of both ``refs(in)`` and ``refs(out)`` (e.g., execute all pivots to / from the working set of nodes). +* **Return limit:** specify the maximum number of nodes returned by the ``pivot()`` query. + + * ``limit=`` (operator syntax) **Operator syntax:** .. parsed-literal:: - **refs(** [ **in** | **out** ] **)** + **refs(** [ **in** | **out , limit=** *<num>* ] **)** **Macro syntax:** @@ -150,6 +153,9 @@ N/A **Usage notes:** * ``refs()`` / ``refs(in)`` / ``refs(out)`` can be useful in an "exploratory" manner to identify what other nodes / forms are "reachable from" (can be pivoted to or from) the working set of nodes. However, because ``refs()`` essentially carries out all possible pivots, the set of nodes returned may be quite large. In such cases a more focused ``pivot()`` or ``join()`` operation may be more useful. +* ``refs()`` does not consume nodes, so the results of a ``refs()`` operation will include both the original working set as well as the resulting set of nodes. +* The ``limit=`` parameter can be provided as input to the ``pivot()`` operator itself when using Operator syntax. Alternately the ``limit()`` operator_ can be used after the ``pivot()`` operator (in either Operator or Macro syntax) to specify a limit on the number of nodes returned. +* Because ``refs()`` does not consume nodes, this impacts the results returned by the ``limit=`` parameter or the ``limit()`` operator. The ``limit=`` parameter will return **all** of the original nodes, **plus** the specified number of results. The ``limit()`` operator will return a **total** number of nodes equal to the specified limit, first including the original working nodes and then including resulting nodes (if possible). fromtags() ----------
Updating the logic to create a new spreadsheet if necessary. Creates the spreadsheet, then performs a batch update with the new data.
@@ -141,20 +141,20 @@ class GoogleSheetsTranslator(base.Translator): self._perform_batch_update(spreadsheet_id, requests) else: - sheets = [] - for catalog in catalogs: - sheets.append(self._create_sheet_from_catalog(catalog, source_lang, - prune=prune)) + # Create a new spreadsheet and use the id. + service = self._create_service() resp = service.spreadsheets().create(body={ - 'sheets': sheets, 'properties': { 'title': project_title, }, }).execute() spreadsheet_id = resp['spreadsheetId'] - for sheet in resp['sheets']: - locale = sheet['properties']['title'] - locales_to_sheet_ids[locale] = sheet['properties']['sheetId'] + + requests = [] + requests += self._generate_create_sheets_requests( + catalogs, source_lang) + self._perform_batch_update(spreadsheet_id, requests) + if 'acl' in self.config: self._do_update_acl(spreadsheet_id, self.config['acl'])
Remove type hints as part of They broke the warning message that Python 2 is no longer supported.
-import typing - import sys try: @@ -11,9 +9,7 @@ except (ImportError, SyntaxError): exit(1) -def display_image(image, - position: typing.Tuple[int, int] = None, - align: str = 'topleft'): +def display_image(image, position=None, align='topleft'): """ Display an image on the mock turtle's canvas. :param image: either a b64 encoded string of PNG image bytes or a PIL Image
[ci] Upgraded CircleCI using 13.1.0-browsers docker image. fixed
@@ -3,7 +3,7 @@ jobs: build: working_directory: ~ docker: - - image: circleci/node:12.9.1-browsers + - image: circleci/node:13.1.0-browsers steps: - checkout - run: npm install
Check if SSLContext supports loading default certs This changes behavior. When creating a TLS connection, not specifying cacert/capath and while on a Python version without load_default_certs, creating the socket will not fail as before, but verifying the connection will fail instead.
@@ -145,7 +145,7 @@ def _wrap_sni_socket(sock, sslopt, hostname, check_hostname): capath = sslopt.get('ca_cert_path', None) if cafile or capath: context.load_verify_locations(cafile=cafile, capath=capath) - else: + elif hasattr(context, 'load_default_certs'): context.load_default_certs(ssl.Purpose.SERVER_AUTH) if sslopt.get('certfile', None): context.load_cert_chain(
Fixed a syntax error on DDL statement COMPUPDATE EMPTYASNULL ON -> COMPUPDATE ON EMPTYASNULL
@@ -582,7 +582,7 @@ copy inventory from 's3://redshift-downloads/TPC-DS/30TB/inventory/' credentials copy item from 's3://redshift-downloads/TPC-DS/30TB/item/' credentials 'aws_access_key_id=<USER_ACCESS_KEY_ID> ;aws_secret_access_key=<USER_SECRET_ACCESS_KEY>' gzip delimiter '|' COMPUPDATE ON EMPTYASNULL region 'us-east-1' ; copy promotion from 's3://redshift-downloads/TPC-DS/30TB/promotion/' credentials 'aws_access_key_id=<USER_ACCESS_KEY_ID> ;aws_secret_access_key=<USER_SECRET_ACCESS_KEY>' gzip delimiter '|' COMPUPDATE ON EMPTYASNULL region 'us-east-1' ; copy reason from 's3://redshift-downloads/TPC-DS/30TB/reason/' credentials 'aws_access_key_id=<USER_ACCESS_KEY_ID> ;aws_secret_access_key=<USER_SECRET_ACCESS_KEY>' gzip delimiter '|' COMPUPDATE ON EMPTYASNULL region 'us-east-1' ; -copy ship_mode from 's3://redshift-downloads/TPC-DS/30TB/ship_mode/' credentials 'aws_access_key_id=<USER_ACCESS_KEY_ID> ;aws_secret_access_key=<USER_SECRET_ACCESS_KEY>' gzip delimiter '|' COMPUPDATE EMPTYASNULL ON region 'us-east-1' ; +copy ship_mode from 's3://redshift-downloads/TPC-DS/30TB/ship_mode/' credentials 'aws_access_key_id=<USER_ACCESS_KEY_ID> ;aws_secret_access_key=<USER_SECRET_ACCESS_KEY>' gzip delimiter '|' COMPUPDATE ON EMPTYASNULL region 'us-east-1' ; copy store from 's3://redshift-downloads/TPC-DS/30TB/store/' credentials 'aws_access_key_id=<USER_ACCESS_KEY_ID> ;aws_secret_access_key=<USER_SECRET_ACCESS_KEY>' gzip delimiter '|' COMPUPDATE ON EMPTYASNULL region 'us-east-1' ; copy time_dim from 's3://redshift-downloads/TPC-DS/30TB/time_dim/' credentials 'aws_access_key_id=<USER_ACCESS_KEY_ID> ;aws_secret_access_key=<USER_SECRET_ACCESS_KEY>' gzip delimiter '|' COMPUPDATE ON EMPTYASNULL region 'us-east-1' ; copy warehouse from 's3://redshift-downloads/TPC-DS/30TB/warehouse/' credentials 'aws_access_key_id=<USER_ACCESS_KEY_ID> ;aws_secret_access_key=<USER_SECRET_ACCESS_KEY>' gzip delimiter '|' COMPUPDATE ON EMPTYASNULL region 'us-east-1' ;
Pontoon: Update Hindi (hi-IN) localization of AMO Localization authors: Mahtab Alam heemananyatanti Shivam Singhal
@@ -5,8 +5,8 @@ msgstr "" "Project-Id-Version: PACKAGE VERSION\n" "Report-Msgid-Bugs-To: EMAIL@ADDRESS\n" "POT-Creation-Date: 2019-10-22 12:35+0000\n" -"PO-Revision-Date: 2019-11-21 10:48+0000\n" -"Last-Translator: ravi <[email protected]>\n" +"PO-Revision-Date: 2019-03-30 07:09+0000\n" +"Last-Translator: heemananyatanti <[email protected]>\n" "Language-Team: LANGUAGE <[email protected]>\n" "Language: hi_IN\n" "MIME-Version: 1.0\n"
[utils.path] added `to_existing_cased_path` works like `to_cased_path` but only returns existing paths, raises FileNotFoundError otherwise
@@ -155,10 +155,10 @@ def cased_path_candidates(path: str, root: str = osp.sep, def to_cased_path(path: str, root: str = osp.sep, is_fs_case_sensitive: bool = True) -> str: """ - Returns a cased version of the given path as far as corresponding nodes exist in the - given root directory. If multiple matches are found, only one is returned. If ``path`` - does not exist in root ``root`` or ``root`` does not exist, the return value will be - ``os.path.join(root, path)``. + Returns a cased version of the given path as far as corresponding nodes (with + arbitrary casing) exist in the given root directory. If multiple matches are found, + only one is returned. If ``path`` does not exist in root ``root`` or ``root`` does not + exist, the return value will be ``os.path.join(root, path)``. :param str path: Original path relative to ``root``. :param str root: Parent directory to search in. There are significant performance @@ -173,6 +173,33 @@ def to_cased_path(path: str, root: str = osp.sep, return candidates[0] +def to_existing_cased_path(path: str, root: str = osp.sep, + is_fs_case_sensitive: bool = True) -> str: + """ + Returns a cased version of the given path if corresponding nodes (with arbitrary + casing) exist in the given root directory. If multiple matches are found, only one is + returned. + + :param str path: Original path relative to ``root``. + :param str root: Parent directory to search in. There are significant performance + improvements if a root directory with a small tree is given. + :param bool is_fs_case_sensitive: Bool indicating if the file system is case + sensitive. If ``False``, we know that there can be at most one match and choose a + faster algorithm. + :returns: Absolute and cased version of given path. + :raises: :class:`FileNotFoundError` if ``path`` does not exist in root ``root`` or + ``root`` itself does not exist. + """ + + candidates = cased_path_candidates(path, root, is_fs_case_sensitive) + + for candidate in candidates: + if osp.exists(candidate): + return candidate + + raise FileNotFoundError(f'No matches with different casing found in "{root}"') + + def path_exists_case_insensitive(path: str, root: str = osp.sep, is_fs_case_sensitive: bool = True) -> bool: """
llvm, port: Do not create copy of port parameters for OVERRIDE modulation It's not needed as the port function is not executed in OVERRIDE case.
@@ -2289,10 +2289,10 @@ class Port_Base(Port): base_params = pnlvm.helpers.get_param_ptr(builder, self, params, "function") - if len(self.mod_afferents) > 0: - # Create a local copy of the function parameters - # only if there are modulating projections - # LLVM is not eliminating the redundant copy + if any(a.sender.modulation != OVERRIDE for a in self.mod_afferents): + # Create a local copy of the function parameters only if + # there are modulating projections of type other than OVERRIDE. + # LLVM is not eliminating the redundant copy. f_params = builder.alloca(port_f.args[0].type.pointee, name="modulated_port_params") builder.store(builder.load(base_params), f_params)
set default status for s3 posts and add support for success_action_redirect.
@@ -776,8 +776,9 @@ class ResponseObject(_TemplateEnvironmentMixin, ActionAuthenticatorMixin): template = self.response_template(S3_DELETE_BUCKET_WITH_ITEMS_ERROR) return 409, {}, template.render(bucket=removed_bucket) - def _bucket_response_post(self, request, body, bucket_name): - if not request.headers.get("Content-Length"): + def _bucket_response_post(self, request, body, bucket_name, headers): + response_headers = {} + if not request.headers.get('Content-Length'): return 411, {}, "Content-Length required" path = self._get_path(request) @@ -810,13 +811,21 @@ class ResponseObject(_TemplateEnvironmentMixin, ActionAuthenticatorMixin): else: f = request.files["file"].stream.read() + if 'success_action_redirect' in form: + response_headers['Location'] = form['success_action_redirect'] + + if 'success_action_status' in form: + status_code = form['success_action_status'] + else: + status_code = 204 + new_key = self.backend.set_key(bucket_name, key, f) # Metadata metadata = metadata_from_headers(form) new_key.set_metadata(metadata) - return 200, {}, "" + return status_code, response_headers, "" @staticmethod def _get_path(request):
Use py_compile instead of distutils to compile model file distutils is considered deprecated in python 3.10. To be removed in 3.12.
@@ -6,8 +6,8 @@ Provides the CLI for the code generator which transforms a Gaphor models """ import argparse -from distutils.util import byte_compile from pathlib import Path +from py_compile import compile from gaphor.codegen import profile_coder, uml_coder @@ -34,7 +34,7 @@ def main() -> None: ) else: uml_coder.generate(args.modelfile, args.outfile, args.overrides) - byte_compile([str(args.outfile)]) + compile(args.outfile) if __name__ == "__main__":
Fix for exec_ceph_cmd For some commands, like "ceph fs ls", the returned output is a list
@@ -277,4 +277,9 @@ def exec_ceph_cmd(ceph_cmd): ocp_pod_obj = OCP(kind='pods', namespace=defaults.ROOK_CLUSTER_NAMESPACE) ct_pod = get_ceph_tools_pod() ceph_cmd += " --format json-pretty" - return ocp_pod_obj.exec_cmd_on_pod(ct_pod, ceph_cmd).toDict() + out = ocp_pod_obj.exec_cmd_on_pod(ct_pod, ceph_cmd) + + # For some commands, like "ceph fs ls", the returned output is a list + if isinstance(out, list): + return [item.toDict() for item in out if item] + return out.toDict()
fix return multi-valued attributes from LDAP So the source attribute value can now be: * `None` if there was no value * a string (if there was one value) * a list of strings (if there was more than one value)
@@ -357,7 +357,7 @@ class LDAPValueFormatter(object): if self.string_format is not None: values = {} for attribute_name in self.attribute_names: - value = self.get_attribute_value(record, attribute_name) + value = self.get_attribute_value(record, attribute_name, first_only=True) if value is None: values = None break @@ -367,17 +367,20 @@ class LDAPValueFormatter(object): return result, attribute_name @classmethod - def get_attribute_value(cls, attributes, attribute_name): + def get_attribute_value(cls, attributes, attribute_name, first_only=False): """ The attribute value type must be decodable (str in py2, bytes in py3) :type attributes: dict :type attribute_name: unicode + :type first_only: bool """ - if attribute_name in attributes: - attribute_value = attributes[attribute_name] - if len(attribute_value) > 0: + attribute_values = attributes.get(attribute_name) + if attribute_values: try: - return attribute_value[0].decode(cls.encoding) + if first_only or len(attribute_values) == 1: + return attribute_values[0].decode(cls.encoding) + else: + return [val.decode(cls.encoding) for val in attribute_values] except UnicodeError as e: raise AssertionException("Encoding error in value of attribute '%s': %s" % (attribute_name, e)) return None
Adding import from Networkx and graph tool lib change tempoary implemented in and still not tested. Futhermore some optimization could be done especialy on the import from graph tool
@@ -131,7 +131,7 @@ class Graph(fourier.GraphFourier, difference.GraphDifference): r"""Doc TODO""" ##from graph_tool.all import * import graph_tool - g = graph_tool.Graph(directed=directed) + g = graph_tool.Graph(directed=directed) #TODO check for undirected graph nonzero = self.W.nonzero() g.add_edge_list(np.transpose(nonzero)) edge_weight = g.new_edge_property("double") @@ -139,6 +139,31 @@ class Graph(fourier.GraphFourier, difference.GraphDifference): g.edge_properties["weight"] = edge_weight return g + @classmethod + def from_networkx(cls, graph_nx): + r"""Doc TODO""" + import networkx as nx + A = nx.to_scipy_sparse_matrix(graph_nx) + G = cls(A) + return G + + @classmethod + def from_graphtool(cls, graph_gt): + r"""Doc TODO""" + nb_vertex = len(graph_gt.get_vertices()) + edge_weight = np.ones(nb_vertex) + W = np.zeros(shape=(nb_vertex, nb_vertex)) + + props_names = graph_gt.edge_properties.keys() + if "weight" in props_names: + prop = graph_gt.edge_properties["weight"] + edge_weight = prop.get_array() + + for e in graph_gt.get_edges(): + W[e[0], e[1]] = edge_weight[e[2]] + return cls(W) + + def check_weights(self): r"""Check the characteristics of the weights matrix.
removed explicit lower from repre ext to not shadow upper case issue Using lower here would hide possibly broken representation, as we would expect both repre["ext"] and repre["name"] be lowercased. In case the aren't review won't get created >> someone will notice and fix issues on source representation.
@@ -152,7 +152,7 @@ class ExtractReview(pyblish.api.InstancePlugin): if input_ext.startswith("."): input_ext = input_ext[1:] - if input_ext.lower() not in self.supported_exts: + if input_ext not in self.supported_exts: self.log.info( "Representation has unsupported extension \"{}\"".format( input_ext
wer_eligible Fix nutrition_status_weight Fix nutrition_status_weighed
@@ -966,6 +966,7 @@ class ChildHealthMonthlyAggregationHelper(BaseICDSAggregationHelper): #fully_immunized_eligible = "{} AND {} > 365".format(valid_in_month, age_in_days) pse_eligible = "({} AND {} > 36)".format(valid_in_month, age_in_months_end) ebf_eligible = "({} AND {} <= 6)".format(valid_in_month, age_in_months) + wer_eligible = "({} AND {} <= 60)".format(valid_in_month, age_in_months) columns = ( ("awc_id", "child_health.awc_id"), @@ -1034,31 +1035,31 @@ class ChildHealthMonthlyAggregationHelper(BaseICDSAggregationHelper): ("counsel_skin_to_skin", "CASE WHEN ucr.pnc_eligible = 1 THEN COALESCE(pnc.skin_to_skin, 0) ELSE 0 END"), # GM Indicators - ("wer_eligible", "ucr.wer_eligible"), + ("wer_eligible", "CASE WHEN {} THEN 1 ELSE 0 END".format(wer_eligible)), ("nutrition_status_last_recorded", "CASE " - "WHEN ucr.wer_eligible = 0 THEN NULL " + "WHEN NOT {} THEN NULL " "WHEN gm.zscore_grading_wfa = 1 THEN 'severely_underweight' " "WHEN gm.zscore_grading_wfa = 2 THEN 'moderately_underweight' " "WHEN gm.zscore_grading_wfa IN (2, 3) THEN 'normal' " - "ELSE 'unknown' END"), + "ELSE 'unknown' END".format(wer_eligible)), ("current_month_nutrition_status", "CASE " - "WHEN ucr.wer_eligible = 0 THEN NULL " + "WHEN NOT {} THEN NULL " "WHEN date_trunc('MONTH', gm.zscore_grading_wfa_last_recorded) != %(start_date)s THEN 'unweighed' " "WHEN gm.zscore_grading_wfa = 1 THEN 'severely_underweight' " "WHEN gm.zscore_grading_wfa = 2 THEN 'moderately_underweight' " "WHEN gm.zscore_grading_wfa IN (3, 4) THEN 'normal' " - "ELSE 'unweighed' END"), + "ELSE 'unweighed' END".format(wer_eligible)), ("nutrition_status_weighed", "CASE " - "WHEN ucr.wer_eligible = 1 AND current_month_nutrition_status != 'unweighed' THEN 1 " - "ELSE 0 END"), + "WHEN {} AND date_trunc('MONTH', gm.zscore_grading_wfa_last_recorded) = %(start_date)s THEN 1 " + "ELSE 0 END".format(wer_eligible)), ("recorded_weight", "CASE " - "WHEN wer_eligible = 0 THEN NULL " + "WHEN NOT {} THEN NULL " "WHEN date_trunc('MONTH', gm.weight_child_last_recorded) = %(start_date)s THEN gm.weight_child " - "ELSE NULL END"), + "ELSE NULL END".format(wer_eligible)), ("recorded_height", "CASE " "WHEN date_trunc('MONTH', gm.height_child_last_recorded) = %(start_date)s THEN gm.height_child "
Update environment variable syntax in extras.yml Update environment variable syntax in extras.yml for the windows workflow to match the powershell syntax.
@@ -59,13 +59,22 @@ jobs: # Installing with -e to keep installation local (for NOSE_NOPATH) # but still compile Cython extensions python -m pip install -e .[testing] - - name: Run test_packages ${{ matrix.nose-tests }} + - name: Run test_packages Ubuntu ${{ matrix.nose-tests }} + if: ${{matrix.os == 'ubuntu-18.04'}} env: NOSETESTS: ${{ matrix.nose-tests }} run: | python -Ic "import pygsti; print(pygsti.__version__); print(pygsti.__path__)" echo "nosetests: $NOSETESTS" nosetests $NOSETESTS + - name: Run test_packages Windows ${{ matrix.nose-tests }} + if: ${{matrix.os == ''windows-2019''}} + env: + NOSETESTS: ${{ matrix.nose-tests }} + run: | + python -Ic "import pygsti; print(pygsti.__version__); print(pygsti.__path__)" + echo "nosetests: $Env:NOSETESTS" + nosetests $Env:NOSETESTS
[profiler] Skip i386 skip condition See for some context, this test wasn't actually being skipped on i386
@@ -284,7 +284,7 @@ def test_estimate_peak_bandwidth(target, dev): ), f"Bandwidth should be between 10^9 and 10^12, but it is {bandwidth}" [email protected](platform.machine() == "i386", reason="Cannot allocate enough memory on i386") [email protected]_if_32bit(reason="Cannot allocate enough memory on i386") @tvm.testing.parametrize_targets("llvm") def test_roofline_analysis(target, dev): a = relay.var("a", relay.TensorType((512, 512), "float32"))
Integ tests: use node.name in place of node.id for test_name I had an OSError, Invalid argument, because node.nodeid was: tests_outputs/.../test_slurm.py::test_slurm[c5.xlarge-us-west-1-alinux-slurm].config instead node.name is: tests_outputs/.../test_slurm[c5.xlarge-us-west-1-alinux-slurm].config Doc:
@@ -176,11 +176,11 @@ def clusters_factory(request): def _write_cluster_config_to_outdir(request, cluster_config): out_dir = request.config.getoption("output_dir") os.makedirs( - "{out_dir}/clusters_configs/{test_dir}".format(out_dir=out_dir, test_dir=os.path.dirname(request.node.nodeid)), + "{out_dir}/clusters_configs/{test_dir}".format(out_dir=out_dir, test_dir=os.path.dirname(request.node.name)), exist_ok=True, ) cluster_config_dst = "{out_dir}/clusters_configs/{test_name}.config".format( - out_dir=out_dir, test_name=request.node.nodeid + out_dir=out_dir, test_name=request.node.name ) copyfile(cluster_config, cluster_config_dst) return cluster_config_dst
store tasks used to initiate peer connections to ensure they are kept alive and to be able to wait for them if we hit the upper limit on number of pending outgoing connections
@@ -64,7 +64,8 @@ class FullNodeDiscovery: self.cleanup_task: Optional[asyncio.Task] = None self.initial_wait: int = 0 self.resolver = dns.asyncresolver.Resolver() - self.pending_outbound_connections: Set = set() + self.pending_outbound_connections: Set[str] = set() + self.pending_tasks: Set[asyncio.Task] = set() async def initialize_address_manager(self) -> None: mkdir(self.peer_db_path.parent) @@ -88,6 +89,10 @@ class FullNodeDiscovery: self.cancel_task_safe(self.connect_peers_task) self.cancel_task_safe(self.serialize_task) self.cancel_task_safe(self.cleanup_task) + for t in self.pending_tasks: + self.cancel_task_safe(t) + if len(self.pending_tasks) > 0: + await asyncio.wait(self.pending_tasks) await self.connection.close() def cancel_task_safe(self, task: Optional[asyncio.Task]): @@ -194,8 +199,6 @@ class FullNodeDiscovery: try: if self.address_manager is None: return - if addr.host in self.pending_outbound_connections: - return self.pending_outbound_connections.add(addr.host) client_connected = await self.server.start_client( addr, @@ -367,15 +370,19 @@ class FullNodeDiscovery: if not initiate_connection: connect_peer_interval += 15 connect_peer_interval = min(connect_peer_interval, self.peer_connect_interval) - if addr is not None and initiate_connection: - while len(self.pending_outbound_connections) >= MAX_CONCURRENT_OUTBOUND_CONNECTIONS: - self.log.debug( - f"Max concurrent outbound connections reached. Retrying in {connect_peer_interval}s." - ) - await asyncio.sleep(connect_peer_interval) - self.log.debug(f"Creating connection task with {addr}.") + if addr is not None and initiate_connection and addr.host not in self.pending_outbound_connections: + if len(self.pending_outbound_connections) >= MAX_CONCURRENT_OUTBOUND_CONNECTIONS: + self.log.debug("Max concurrent outbound connections reached. waiting") + await asyncio.wait(self.pending_tasks, return_when=asyncio.FIRST_COMPLETED) + self.pending_tasks.add( asyncio.create_task(self.start_client_async(addr, disconnect_after_handshake)) + ) + await asyncio.sleep(connect_peer_interval) + + # prune completed connect tasks + self.pending_task = set(filter(lambda t: not t.done(), self.pending_tasks)) + except Exception as e: self.log.error(f"Exception in create outbound connections: {e}") self.log.error(f"Traceback: {traceback.format_exc()}") @@ -487,8 +494,8 @@ class FullNodePeers(FullNodeDiscovery): async def close(self): await self._close_common() - self.self_advertise_task.cancel() - self.address_relay_task.cancel() + self.cancel_task_safe(self.self_advertise_task) + self.cancel_task_safe(self.address_relay_task) async def _periodically_self_advertise_and_clean_data(self): while not self.is_closed:
feat(online_value_artist): add online_value_artist interface add online_value_artist interface
@@ -124,5 +124,5 @@ def news_cctv(date: str = "20130308") -> pd.DataFrame: if __name__ == "__main__": - news_cctv_df = news_cctv(date="20150208") + news_cctv_df = news_cctv(date="20211115") print(news_cctv_df)
Github workflow bot fix It seems LABEL is not a recognized argument.
@@ -52,6 +52,5 @@ jobs: steps: - uses: actions-automation/pull-request-responsibility@main with: - LABEL: "keylime-bot for keylime" reviewers: "core" num_to_request: 3
fix: remove attr helpers and type these can potentially lead to security issues, avoiding for now
@@ -196,12 +196,9 @@ def get_python_builtins(): 'all': all, 'any': any, 'bool': bool, - 'delattr': delattr, 'dict': dict, 'enumerate': enumerate, 'filter': filter, - 'getattr': getattr, - 'hasattr': hasattr, 'isinstance': isinstance, 'issubclass': issubclass, 'list': list, @@ -210,11 +207,9 @@ def get_python_builtins(): 'min': min, 'range': range, 'set': set, - 'setattr': setattr, 'sorted': sorted, 'sum': sum, 'tuple': tuple, - 'type': type, } def get_hooks(hook=None, default=None, app_name=None):
doc: don't mention legacy tools in main README Legacy tools will be removed in subsequent commits.
@@ -30,12 +30,6 @@ High-level support for implementing [repository operations](https://theupdateframework.github.io/specification/latest/#repository-operations) is planned but not yet provided: see [1.0.0 plans](https://github.com/theupdateframework/python-tuf/blob/develop/docs/1.0.0-ANNOUNCEMENT.md). -In addition to these APIs the project also provides a *legacy -implementation* with `tuf.client` implementing the client workflow and -`tuf.repository_tool` providing a high-level interface for repository -operations. The legacy implementation is going to be -[deprecated](https://github.com/theupdateframework/python-tuf/blob/develop/docs/1.0.0-ANNOUNCEMENT.md) in the near future. - The reference implementation strives to be a readable guide and demonstration for those working on implementing TUF in their own languages, environments, or update systems.
zos: don't use universal-new-lines mode PR-URL:
@@ -233,6 +233,11 @@ def LoadOneBuildFile(build_file_path, data, aux_data, includes, # Open the build file for read ('r') with universal-newlines mode ('U') # to make sure platform specific newlines ('\r\n' or '\r') are converted to '\n' # which otherwise will fail eval() + if sys.platform == 'zos': + # On z/OS, universal-newlines mode treats the file as an ascii file. But since + # node-gyp produces ebcdic files, do not use that mode. + build_file_contents = open(build_file_path, 'r').read() + else: build_file_contents = open(build_file_path, 'rU').read() else: raise GypError("%s not found (cwd: %s)" % (build_file_path, os.getcwd()))
[subset] Add --layout-scripts Fixes
@@ -168,6 +168,10 @@ Glyph set expansion: * Keep all features. --layout-features+=aalt --layout-features-=vrt2 * Keep default set of features plus 'aalt', but drop 'vrt2'. + --layout-scripts[+|-]=<script>[,<script>...] + Specify (=), add to (+=) or exclude from (-=) the comma-separated + set of OpenType layout script tags that will be preserved. By + default all scripts are retained ('*'). Hinting options: --hinting @@ -1492,6 +1496,15 @@ def subset_feature_tags(self, feature_tags): if self.table.ScriptList: self.table.ScriptList.subset_features(feature_indices, self.retain_empty_scripts()) +@_add_method(ttLib.getTableClass('GSUB'), + ttLib.getTableClass('GPOS')) +def subset_script_tags(self, script_tags): + if self.table.ScriptList: + self.table.ScriptList.ScriptRecord = \ + [s for s in self.table.ScriptList.ScriptRecord + if s.ScriptTag in script_tags] + self.table.ScriptList.ScriptCount = len(self.table.ScriptList.ScriptRecord) + @_add_method(ttLib.getTableClass('GSUB'), ttLib.getTableClass('GPOS')) def prune_features(self): @@ -1511,6 +1524,8 @@ def prune_features(self): ttLib.getTableClass('GPOS')) def prune_pre_subset(self, font, options): # Drop undesired features + if '*' not in options.layout_scripts: + self.subset_script_tags(options.layout_scripts) if '*' not in options.layout_features: self.subset_feature_tags(options.layout_features) # Neuter unreferenced lookups @@ -2755,6 +2770,7 @@ class Options(object): self.hinting_tables = self._hinting_tables_default[:] self.legacy_kern = False # drop 'kern' table if GPOS available self.layout_features = self._layout_features_default[:] + self.layout_scripts = ['*'] self.ignore_missing_glyphs = False self.ignore_missing_unicodes = True self.hinting = True
operations.repo: regen: return number of actual errors Instead of just an error status.
@@ -234,7 +234,7 @@ class operations(sync_operations): try: if sync_rate is not None: cache.set_sync_rate(1000000) - ret = 0 + errors = 0 # Force usage of unfiltered repo to include pkgs with metadata issues. # Matches are collapsed directly to a list to avoid threading issues such @@ -245,19 +245,19 @@ class operations(sync_operations): for pkg, e in regen.regen_repository( self.repo, pkgs, observer=observer, threads=threads, **kwargs): observer.error(f'caught exception {e} while processing {pkg.cpvstr}') - ret = 1 + errors += 1 # report pkgs with bad metadata -- relies on iterating over the # unfiltered repo to populate the masked repo pkgs = frozenset(pkg.cpvstr for pkg in self.repo) for pkg in sorted(self.repo._bad_masked): observer.error(f'{pkg.cpvstr}: {pkg.data.msg(verbosity=observer.verbosity)}') - ret = 1 + errors += 1 # remove old/invalid cache entries self._cmd_implementation_clean_cache(pkgs) - return ret + return errors finally: if sync_rate is not None: cache.set_sync_rate(sync_rate)
Update `numeric_match_with_0_1_relative_error_fn` Allow `numeric_match_with_0_1_relative_error_fn` to run on non numeric targets. Currently if target is non-numeric, the function will fail while converting to float. Return 0. score by return False on `check_if_numeric_value_is_within_relative_error` instead of failing.
@@ -449,7 +449,10 @@ def check_if_numeric_value_is_within_relative_error( # intentional: the task should supply valid targets. assert isinstance(target, list) assert len(target) == 1 + try: ftarget = float(target[0]) + except ValueError: + return False # We try to convert the model's response, a string, to a floating-point # number. If we fail, the model has not returned a numeric value and
doc: Fix the snippet in "The Script Validator" section In the snippet of code, store_nfs_version should point to an object instead of an array.
@@ -186,8 +186,8 @@ as well: output: OUTPUT_VAR - script: store_nfs_version: # Because inline is set, this is just a friendly name - - inline: rpm -q nfs-utils # Runs this text directly, rather than reading a file - - output: nfs-version # Places the stdout of this script into an argument + inline: rpm -q nfs-utils # Runs this text directly, rather than reading a file + output: nfs-version # Places the stdout of this script into an argument # for future scripts to consume; if none exists, the # argument is created
update rays.display : change xlabels Better names, like it was already written in the docstrings of each histogram methods. The main problem was with the word 'distance' which is ambiguous (z vs y axis).
@@ -319,7 +319,7 @@ class Rays: # axis1.set_title('Intensity profile') axis1.plot(x, y, 'k-', label="Intensity") axis1.set_ylim([0, max(y) * 1.1]) - axis1.set_xlabel("Distance") + axis1.set_xlabel("Height of ray") axis1.set_ylabel("Ray count") axis1.legend(["Intensity"]) @@ -328,7 +328,7 @@ class Rays: axis2.plot(x, y, 'k--', label="Orientation profile") axis2.set_ylim([0, max(y) * 1.1]) axis2.set_xlim([-pi / 2, pi / 2]) - axis2.set_xlabel("Angles [rad]") + axis2.set_xlabel("Angle of ray [rad]") axis2.set_ylabel("Ray count") axis2.legend(["Angle"])
SetQueryTest : Avoid `six` We don't need now we only support Python 3.
########################################################################## import inspect -import six import unittest import imath @@ -200,7 +199,7 @@ class SetQueryTest( GafferSceneTest.SceneTestCase ) : """ ) ) - with six.assertRaisesRegex( self, Gaffer.ProcessException, 'Context has no variable named "scene:path"' ) : + with self.assertRaisesRegex( Gaffer.ProcessException, 'Context has no variable named "scene:path"' ) : script["query"]["firstMatch"].getValue() script["expression"] = Gaffer.Expression() @@ -211,7 +210,7 @@ class SetQueryTest( GafferSceneTest.SceneTestCase ) : """ ) ) - with six.assertRaisesRegex( self, Gaffer.ProcessException, 'Context has no variable named "scene:path"' ) : + with self.assertRaisesRegex( Gaffer.ProcessException, 'Context has no variable named "scene:path"' ) : script["query"]["firstMatch"].getValue() @GafferTest.TestRunner.PerformanceTestMethod( repeat = 1 )
fix CreateCommCareUserModal looks like django 1.10 (deployed Mar. 21) changed something affecting the render_to_string call
@@ -12,7 +12,6 @@ from django.http import HttpResponseRedirect, HttpResponse,\ HttpResponseForbidden, HttpResponseBadRequest, Http404 from django.http.response import HttpResponseServerError from django.shortcuts import render, redirect -from django.template import RequestContext from django.template.loader import render_to_string from django.utils.decorators import method_decorator from django.utils.translation import ugettext as _, ugettext_noop @@ -872,13 +871,12 @@ class CreateCommCareUserModal(JsonRequestResponseMixin, DomainViewMixin, View): return super(CreateCommCareUserModal, self).dispatch(request, *args, **kwargs) def render_form(self, status): - context = RequestContext(self.request, { - 'form': self.new_commcare_user_form, - 'data_fields_form': self.custom_data.form, - }) return self.render_json_response({ "status": status, - "form_html": render_to_string(self.template_name, context) + "form_html": render_to_string(self.template_name, { + 'form': self.new_commcare_user_form, + 'data_fields_form': self.custom_data.form, + }, request=self.request) }) def get(self, request, *args, **kwargs):
TST: Add a test with an almost singular design matrix for lsq_linear The problem and the test was reported in issue
@@ -108,6 +108,19 @@ class BaseMixin(object): assert_(isinstance(res.message, str)) assert_(res.success) + # This is a test for issue #9982. + def test_almost_singular(self): + A = np.array( + [[0.8854232310355122, 0.0365312146937765, 0.0365312146836789], + [0.3742460132129041, 0.0130523214078376, 0.0130523214077873], + [0.9680633871281361, 0.0319366128718639, 0.0319366128718388]]) + + b = np.array( + [0.0055029366538097, 0.0026677442422208, 0.0066612514782381]) + + result = lsq_linear(A, b, method=self.method) + assert_(result.cost < 1.1e-8) + class SparseMixin(object): def test_sparse_and_LinearOperator(self):
Deepcopy the object, then run again check_trigger_amount This solves the issue of the timeseries going negative!
#!/usr/bin/env python3 # -*- coding: utf-8 -*- -import numpy as np import logging +from copy import deepcopy + +import numpy as np LGR = logging.getLogger(__name__) @@ -51,7 +53,8 @@ def find_runs(phys_in, ntp_list, tr_list, thr=None, padding=9): for run_idx, run_tps in enumerate(ntp_list): # correct time offset for this iteration's object - phys_in.check_trigger_amount(thr=thr, num_timepoints_expected=run_tps, tr=tr_list[run_idx]) + phys_in.check_trigger_amount(thr=thr, num_timepoints_expected=run_tps, + tr=tr_list[run_idx]) # If it's the very first run, start the run at sample 0, # otherwise "add" the padding @@ -98,7 +101,7 @@ def find_runs(phys_in, ntp_list, tr_list, thr=None, padding=9): # update the object so that next iteration will look for the first trigger # after previous run's last trigger. maybe padding extends to next run - phys_in = phys_in[update:-1] + phys_in = deepcopy(phys_in[update:-1]) return run_timestamps @@ -137,17 +140,16 @@ def slice4phys(phys_in, ntp_list, tr_list, thr, padding=9): # Find the timestamps run_timestamps = find_runs(phys_in, ntp_list, tr_list, thr, padding) - for run in run_timestamps.keys(): + for n, run in enumerate(run_timestamps.keys()): # tmp variable to collect run's info run_attributes = run_timestamps[run] - phys_in_slices[run] = phys_in[run_attributes[0]:run_attributes[1]] + phys_in_slices[run] = deepcopy(phys_in[run_attributes[0]:run_attributes[1]]) - # Overwrite current run phys_in attributes - # 3rd item of run_attributes is adjusted time offset - phys_in_slices[run].time_offset = run_attributes[2] - # 4th item of run_attributes is the nb of tp found by check_trigger_amount - phys_in_slices[run].num_timepoints_found = run_attributes[3] + # Run check_trigger amount + phys_in_slices[run].check_trigger_amount(thr=thr, + num_timepoints_expected=ntp_list[n], + tr=tr_list[n]) return phys_in_slices
Fix refrigeration schedules Now Qcre_sys gets calculated. Now the final energy demands are missing.
@@ -118,9 +118,9 @@ def calc_deterministic_schedules(archetype_schedules, archetype_values, bpr, lis # define schedules and codes occupant_schedules = ['ve', 'Qs', 'X'] - electricity_schedules = ['Ea', 'El', 'Ed'] + electricity_schedules = ['Ea', 'El', 'Ed', 'Qcre'] water_schedules = ['Vww', 'Vw'] - process_schedules = ['Epro', 'Qhpro', 'Qcre'] + process_schedules = ['Epro', 'Qhpro'] # schedule_codes define which archetypal schedule should be used for the given schedule schedule_codes = {'people': 0, 'electricity': 1, 'water': 2, 'processes': 3}
Connected embed Literally just changes the "Connected!" message to an embed so it looks a bit nicer, and the little added bonus that you can click on the Python Bot and it will send you to the bot's repo.
# coding=utf-8 +from discord import Embed from discord.ext.commands import AutoShardedBot from bot.constants import DEVLOG_CHANNEL @@ -14,7 +15,15 @@ class Logging: async def on_ready(self): print("Connected!") - await self.bot.get_channel(DEVLOG_CHANNEL).send("Connected!") + + embed = Embed(description="Connected!") + embed.set_author( + name="Python Bot", + url="https://github.com/discord-python/bot", + icon_url="https://raw.githubusercontent.com/discord-python/branding/master/logos/logo_circle.png" + ) + + await self.bot.get_channel(DEVLOG_CHANNEL).send(embed=embed) def setup(bot):
Add pointer to checked-in tutorial code Summary: Fixes Test Plan: Manual Reviewers: #ft, schrockn, sashank
@@ -20,3 +20,7 @@ Tutorial reusing_solids composite_solids composition_functions + + +You can find all of the tutorial code checked into the dagster repository at +``dagster/examples/dagster_examples/intro_tutorial``.
defaults for stdout and stderr should be bytes only print stdout if no OSError exception was raised
@@ -109,7 +109,8 @@ def compile(source, stderr=subprocess.PIPE) except OSError: # preserve historic status code used by exec_command() - cp = subprocess.CompletedProcess(c, 127, stdout='', stderr='') + cp = subprocess.CompletedProcess(c, 127, stdout=b'', stderr=b'') + else: if verbose: print(cp.stdout.decode()) finally:
Fix latex error summary * fix latex error summary * [pre-commit.ci] auto fixes from pre-commit.com hooks for more information, see * Move error to inside loop * change error_pos[0] to log_index
@@ -195,8 +195,10 @@ def compile_tex(tex_file, tex_compiler, output_format): if error_pos: with open(tex_file, "r") as g: tex = g.readlines() - logger.error("LaTeX compilation error: {log[error_pos][2:])}") for log_index in error_pos: + logger.error( + f"LaTeX compilation error: {log[log_index][2:]}" + ) index_line = log_index context = "Context for error:\n\n"
Remove sudo for pip install Sudo installs the packages during pip install into the systems package distribution dir, and not in the one from the virtual environment
@@ -15,8 +15,8 @@ install: - sudo ln -s /tmp/nextflow/nextflow /usr/local/bin/nextflow # Set up nf-core and test/coverage modules - cd ${TRAVIS_BUILD_DIR} - - sudo pip install . - - sudo pip install codecov nose pytest pytest-cov + - pip install . + - pip install codecov nose pytest pytest-cov script: python -m pytest --cov=nf_core .
Add 'contactable()' function to filter out non-accessible speakers Addresses requirement discussed in
@@ -512,6 +512,69 @@ def scan_network_any_soco(household_id=None, **network_scan_kwargs): return any_zone +def contactable(speakers): + """Find only contactable players in a set of `SoCo` objects. + + This function checks a set of `SoCo` objects to ensure that each + associated Sonos player is currently contactable. A new set + is returned containing only contactable players. + + If there are non-contactable players, the function return will + be delayed until at least one network timeout has expired (several + seconds). Contact attempts run in parallel threads to minimise + delays. + + Args: + speakers(set): A set of `SoCo` objects. + + Returns: + set: A set of `SoCo` objects, all of which have been + confirmed to be currently contactable. An empty set + is returned if no speakers are contactable. + """ + + def contactable_worker(): + """Worker thread helper function to check whether + speakers are contactable and, if so, to add them to + the set of contactable speakers. + """ + while True: + try: + speaker = speakers.pop() + except KeyError: + break + try: + # Try getting a device property + _ = speaker.is_visible + _LOG.info("%s is contactable", speaker.ip_address) + contactable_speakers.add(speaker) + # The exception is unimportant + # pylint: disable=bare-except + except: # noqa: E722 + _LOG.info("%s is not contactable", speaker.ip_address) + + contactable_speakers = set() + if speakers is None: + return contactable_speakers + + # Attempt to create one thread per speaker + thread_list = [] + for _ in range(len(speakers)): + thread = threading.Thread(target=contactable_worker) + try: + thread.start() + thread_list.append(thread) + except RuntimeError: + # Can't create any more threads + _LOG.info("Failed to create a new thread") + break + + for thread in thread_list: + thread.join() + + return contactable_speakers + + def _find_ipv4_networks(min_netmask): """Discover attached IP networks.
Fix wording for scheduler in Helm values Summary: The field `scheduler` is not deprecated - only the the option for the K8sScheduler. Test Plan: N/A Reviewers: dgibson, nate, sashank
@@ -229,14 +229,18 @@ pipelineRun: YOUR_ENV_VAR: "" #################################################################################################### -# Scheduler: Configuration for the K8sScheduler. (deprecated) +# Scheduler: Configuration for the scheduler #################################################################################################### scheduler: - # Type can be one of [DagsterDaemonScheduler, K8sScheduler, CustomScheduler] + # Type can be one of [ + # DagsterDaemonScheduler, + # K8sScheduler (deprecated), + # CustomScheduler, + # ] type: DagsterDaemonScheduler config: - # This configuration will only be used if the K8sScheduler is selected + # (deprecated) This configuration will only be used if the K8sScheduler is selected k8sScheduler: # Image for K8s CronJob(s) to launch K8s Job(s). This image needs access to dagster code # (to call `dagster api launch_scheduled_execution`) and instance yaml
Improve performance of ilen function. The benchmarks in and showed that this approach is faster than the current (and the one before the current) implementation. An inline comment to refer to the benchmarks has been added to the code so that any future changes can refer to the old benchmarks.
@@ -441,10 +441,12 @@ def ilen(iterable): This consumes the iterable, so handle with care. """ - length = 0 - for length, _ in enumerate(iterable, 1): - pass - return length + # This approach was selected because benchmarks showed it's likely the + # fastest of the known implementations at the time of writing. + # See GitHub tracker: #236, #230. + counter = count() + deque(zip(iterable, counter), maxlen=0) + return next(counter) def iterate(func, start):
[doc] Added update regarding URL difference based on deployment This is from [1] as networking guide content has been imported from openstack-manual [2]. [1] [2] Closes-Bug:
@@ -419,6 +419,13 @@ segment contains one IPv4 subnet and one IPv6 subnet. As of the writing of this guide, there is not placement API CLI client, so the :command:`curl` command is used for this example. + .. note:: + + Service points URLs differ depending on your OpenStack deployment. You + can discover the Placement service URL by executing the + :command:`openstack endpoint list | grep placement` command. This + command has to be executed as admin. + #. Verify that host aggregates were created for each segment in the Compute service (for the sake of brevity, only one of the segments is shown in this example).
Azure: add deploy and delete transformers They are used to deploy a VM for other transformers like export VHD, build kernel and so on.
@@ -11,6 +11,7 @@ from dataclasses_json import dataclass_json from retry import retry from lisa import schema +from lisa.environment import Environments, EnvironmentSpace from lisa.feature import Features from lisa.features import StartStop from lisa.node import RemoteNode @@ -24,6 +25,7 @@ from .common import ( AZURE_SHARED_RG_NAME, check_or_create_storage_account, get_compute_client, + get_environment_context, get_network_client, get_node_context, get_or_create_storage_container, @@ -76,6 +78,21 @@ class VhdTransformerSchema(schema.Transformer): restore: bool = False +@dataclass_json +@dataclass +class DeployTransformerSchema(schema.Transformer): + requirement: schema.Capability = field(default_factory=schema.Capability) + resource_group_name: str = "" + + +@dataclass_json +@dataclass +class DeleteTransformerSchema(schema.Transformer): + resource_group_name: str = field( + default="", metadata=schema.metadata(required=True) + ) + + class VhdTransformer(Transformer): """ convert an azure VM to VHD, which is ready to deploy. @@ -269,6 +286,80 @@ class VhdTransformer(Transformer): return public_ip_address +class DeployTransformer(Transformer): + """ + deploy a node in transformer phase for further operations + """ + + __resource_group_name = "resource_group_name" + + @classmethod + def type_name(cls) -> str: + return "azure_deploy" + + @classmethod + def type_schema(cls) -> Type[schema.TypedSchema]: + return DeployTransformerSchema + + @property + def _output_names(self) -> List[str]: + return [self.__resource_group_name] + + def _internal_run(self) -> Dict[str, Any]: + platform = _load_platform(self._runbook_builder, self.type_name()) + runbook: DeployTransformerSchema = self.runbook + + envs = Environments() + environment_requirement = EnvironmentSpace() + environment_requirement.nodes.append(runbook.requirement) + environment = envs.from_requirement(environment_requirement) + assert environment + + platform.prepare_environment(environment=environment) + + platform.deploy_environment(environment) + + resource_group_name = get_environment_context(environment).resource_group_name + + return {self.__resource_group_name: resource_group_name} + + +class DeleteTransformer(Transformer): + """ + delete an environment + """ + + @classmethod + def type_name(cls) -> str: + return "azure_delete" + + @classmethod + def type_schema(cls) -> Type[schema.TypedSchema]: + return DeleteTransformerSchema + + @property + def _output_names(self) -> List[str]: + return [] + + def _internal_run(self) -> Dict[str, Any]: + platform = _load_platform(self._runbook_builder, self.type_name()) + runbook: DeleteTransformerSchema = self.runbook + + # mock up environment for deletion + envs = Environments() + environment_requirement = EnvironmentSpace() + environment_requirement.nodes.append(schema.NodeSpace()) + environment = envs.from_requirement(environment_requirement) + assert environment + environment_context = get_environment_context(environment) + environment_context.resource_group_name = runbook.resource_group_name + environment_context.resource_group_is_created = True + + platform.delete_environment(environment) + + return {} + + def _load_platform( runbook_builder: RunbookBuilder, transformer_name: str ) -> AzurePlatform:
Add shape mismatch tests * Add shape mismatch tests Added a couple of tests in which we check tensors of different shapes for inequality. The test should return the correct result and not throw ValueError from np.allclose. * Remove whitespaces
@@ -235,11 +235,21 @@ class EqualTests(unittest.TestCase): t2 = TensorBase(np.array([1, 4, 3])) self.assertFalse(syft.equal(t1, t2)) + def test_shape_not_equal(self): + t1 = TensorBase(np.array([1, 2])) + t2 = TensorBase(np.array([1, 4, 3])) + self.assertFalse(syft.equal(t1, t2)) + def test_inequality_operation(self): t1 = TensorBase(np.array([1, 2, 3])) t2 = TensorBase(np.array([1, 4, 5])) self.assertTrue(t1 != t2) + def test_shape_inequality_operation(self): + t1 = TensorBase(np.array([1, 2])) + t2 = TensorBase(np.array([1, 4, 5])) + self.assertTrue(t1 != t2) + class sigmoidTests(unittest.TestCase): def test_sigmoid(self):
Run CI tests on Windows, too Involves adding caching on `windows-latest` too, and enforcing
@@ -5,7 +5,10 @@ on: [push, pull_request, workflow_dispatch] jobs: cache_nltk_data: name: cache nltk_data - runs-on: ubuntu-latest + strategy: + matrix: + os: [ubuntu-latest, windows-latest] + runs-on: ${{ matrix.os }} steps: - name: Checkout code uses: actions/checkout@v2 @@ -15,12 +18,12 @@ jobs: id: restore-cache with: path: ~/nltk_data - key: nltk_data + key: nltk_data_${{ secrets.CACHE_VERSION }} - name: Download nltk data packages on cache miss run: | pip install regex # dependencies needed to download nltk data - python -c "import nltk; nltk.download('all')" + python -c "import nltk; from pathlib import Path; path = Path('~/nltk_data').expanduser(); path.mkdir(exist_ok=True); nltk.download('all', download_dir=path)" shell: bash if: steps.restore-cache.outputs.cache-hit != 'true' @@ -30,7 +33,7 @@ jobs: strategy: matrix: python-version: [3.6, 3.7, 3.8, 3.9] - os: [ubuntu-latest, macos-latest] + os: [ubuntu-latest, macos-latest, windows-latest] fail-fast: false runs-on: ${{ matrix.os }} steps: @@ -57,7 +60,7 @@ jobs: uses: actions/cache@v2 with: path: ~/nltk_data - key: nltk_data + key: nltk_data_${{ secrets.CACHE_VERSION }} - name: Run pytest run: pytest --numprocesses auto -rsx nltk/test
Update ffmpeg.py get fix
@@ -130,7 +130,7 @@ class MediaStreamInfo(object): self.metadata = {} def toJson(self): - language = self.metadata.get("language", default="und").lower().strip() + language = self.metadata.get("language", "und").lower().strip() out = {'index': self.index, 'codec': self.codec, 'type': self.type}
Typo fix: example of BIF example of method BIFWriter.write_bif()
@@ -652,7 +652,7 @@ $properties}\n""" >>> from pgmpy.readwrite import BIFReader, BIFWriter >>> model = BIFReader('dog-problem.bif').get_model() >>> writer = BIFWriter(model) - >>> writer.write_bif(filname='test_file.bif') + >>> writer.write_bif(filename='test_file.bif') """ writer = self.__str__() with open(filename, "w") as fout:
macOS is now supported Remove warning about macOS support from tutorial
@@ -72,13 +72,6 @@ capabilities, and set the stage for understanding how TVM works. # -################################################################################ -# .. note:: Supported operating systems -# -# TVMC is only supported on Linux. Currently macOS and Windows default -# threading models do not support the model used for tuning by TVMC. - - ################################################################################ # Obtaining the Model # -------------------
Update environmental_inhalers.json tweaked wording to reflect to emphasise respimat [see issue 2144](https://github.com/ebmdatalab/openprescribing/issues/2144) is not an MDI as previously classified by dm+d
"why_it_matters": [ "The NHS has <a href='https://www.longtermplan.nhs.uk/wp-content/uploads/2019/08/nhs-long-term-plan-version-1.2.pdf'>committed", "to reducing its carbon footprint by 51% by 2025</a> to meet the target in the Climate Change Act, including a shift to dry powdered", - "inhalers (DPI) to deliver a reduction of 4%. DPIs are less harmful to the enviroment than traditional metered", - "dose inhalers (MDIs) and the NHS long term plan supports the use of DPI where it is clinically appropriate.", + "inhalers (DPI) to deliver a reduction of 4%. DPIs and other newer types of inhalers like soft mist inhalers are less harmful to the enviroment than traditional metered", + "dose inhalers (MDIs) and the NHS long term plan supports the use of these inahlers where it is clinically appropriate.", "<a href='https://www.nice.org.uk/guidance/ng80/resources/inhalers-for-asthma-patient-decision-aid-pdf-6727144573'>NICE has ", "produced a inhaler decision aid</a> to faciltiate discussion about inhaler options. ", "<p>You can read more about this measure ",
Removed extra memory multiplication (resolves
@@ -203,7 +203,7 @@ def parse_memory_limit(mem: float) -> str: def parse_memory(mem: float, resource: bool) -> str: """Parse memory parameter.""" lsf_unit = get_lsf_units(resource=resource) - megabytes_of_mem = convert_units(float(mem) * 1024, src_unit=lsf_unit, dst_unit='MB') + megabytes_of_mem = convert_units(float(mem), src_unit=lsf_unit, dst_unit='MB') if megabytes_of_mem < 1: megabytes_of_mem = 1.0 # round as a string here to avoid returning something like 1.231e+12
Check pre-processing output falls in valid range Summary: Check that pre-processed batch data is within range of expected magnitudes
@@ -11,6 +11,9 @@ from ml.rl.preprocessing.normalization import MISSING_VALUE, NormalizationParame from torch.nn import Module, Parameter +MAX_FEATURE_VALUE = 6 +MIN_FEATURE_VALUE = MAX_FEATURE_VALUE * -1 + logger = logging.getLogger(__name__) @@ -118,6 +121,8 @@ class Preprocessor(Module): j, input[:, j : j + 1], norm_params ) new_output *= not_missing_input[:, j : j + 1] + # TODO: Renable when T34648262 is fixed + # self._check_preprocessing_output(new_output, [norm_params]) outputs.append(new_output) else: norm_params = [] @@ -127,11 +132,17 @@ class Preprocessor(Module): begin_index, input[:, begin_index:end_index], norm_params ) new_output *= not_missing_input[:, begin_index:end_index] + # TODO: Renable when T34648262 is fixed + # self._check_preprocessing_output(new_output, norm_params) outputs.append(new_output) + # TODO: Remove clamp values when preprocessing issues are solved (T34648262) if len(outputs) == 1: - return outputs[0] - return torch.cat(outputs, dim=1) + return torch.clamp(outputs[0], MIN_FEATURE_VALUE, MAX_FEATURE_VALUE) + + return torch.clamp( + torch.cat(outputs, dim=1), MIN_FEATURE_VALUE, MAX_FEATURE_VALUE + ) def _preprocess_feature_single_column( self, @@ -467,6 +478,29 @@ class Preprocessor(Module): return fn(t1_mask, t2).float() + def _check_preprocessing_output(self, batch, norm_params): + """ + Check that preprocessed features fall within range of valid output. + :param batch: torch tensor + :param norm_params: list of normalization parameters + """ + feature_type = norm_params[0].feature_type + max_value = batch.max() + if bool(max_value > MAX_FEATURE_VALUE): + raise Exception( + "A {} feature type has max value {} which is > than accepted post pre-processing max of {}".format( + feature_type, max_value, MAX_FEATURE_VALUE + ) + ) + + min_value = batch.min() + if bool(min_value < MIN_FEATURE_VALUE): + raise Exception( + "A {} feature type has min value {} which is < accepted post pre-processing min of {}".format( + feature_type, min_value, MIN_FEATURE_VALUE + ) + ) + def hack_ge(self, t1: torch.Tensor, t2: torch.Tensor) -> torch.Tensor: return t1 > (t2 - self.epsilon_tensor)
stats: Don't set display to none in alert class. The alert class is used only by one element so this style can be applied directly to the element instead.
@@ -108,10 +108,6 @@ p { float: right; } -.alert { - display: none; -} - .button { font-family: 'Source Sans Pro', 'Helvetica Neue', sans-serif !important; border: none; @@ -185,6 +181,10 @@ p { position: relative; } +#id_stats_errors { + display: none; +} + #users_hover_info, #hoverinfo { display: none;
Update getting_started.ipynb Replaced jax.value_and_grad with jax.grad in section 7 since the outputted loss value wasn't being used
" [Module.apply](https://flax.readthedocs.io/en/latest/flax.linen.html#flax.linen.Module.apply)\n", " method.\n", "- Computes the `cross_entropy_loss` loss function.\n", - "- Evaluates the loss function and its gradient using\n", - " [jax.value_and_grad](https://jax.readthedocs.io/en/latest/jax.html#jax.value_and_grad).\n", + "- Evaluates the gradient of the loss function using\n", + " [jax.grad](https://jax.readthedocs.io/en/latest/jax.html#jax.grad).\n", "- Applies a\n", " [pytree](https://jax.readthedocs.io/en/latest/pytrees.html#pytrees-and-jax-functions)\n", " of gradients to the optimizer to update the model's parameters.\n", " logits = CNN().apply({'params': params}, batch['image'])\n", " loss = cross_entropy_loss(logits=logits, labels=batch['label'])\n", " return loss, logits\n", - " grad_fn = jax.value_and_grad(loss_fn, has_aux=True)\n", - " (_, logits), grads = grad_fn(state.params)\n", + " grad_fn = jax.grad(loss_fn, has_aux=True)\n", + " grads, logits = grad_fn(state.params)\n", " state = state.apply_gradients(grads=grads)\n", " metrics = compute_metrics(logits=logits, labels=batch['label'])\n", " return state, metrics"
[LLVM/CPU] Terminate basic block after "ret" instruction * [LLVM/CPU] Terminate basic block after "ret" instruction "Ret" is a terminator in LLVM IR and there should be no instructions in the basic block following it. When generating a "ret", end the current block and start a new one.
@@ -781,6 +781,9 @@ llvm::Value* CodeGenCPU::CreateIntrinsic(const CallNode* op) { return CreateStaticHandle(); } else if (op->op.same_as(builtin::tvm_throw_last_error())) { builder_->CreateRet(ConstInt32(-1)); + auto next_block = std::next(builder_->GetInsertBlock()->getIterator()); + llvm::BasicBlock* new_bb = llvm::BasicBlock::Create(*ctx_, "cont", function_, &*next_block); + builder_->SetInsertPoint(new_bb); return ConstInt32(-1); } else if (op->op.same_as(builtin::tvm_struct_get())) { CHECK_EQ(op->args.size(), 3U);
numpy based 16 based conversion Uses numpy to save image as 16 bit image.
This code creates a dataframe of dicom headers based on dicom files in a filepath. This code also extracts the images within those dicoms if requested. see section 'print images' pip3 install image numpy pandas pydicom pillow pypng +Make sure to have empty extracted-images, failed-dicom/1, failed-dicom/2, failed-dicom/3 folders +ready in the root folder. """ import numpy as np import pandas as pd @@ -192,9 +194,15 @@ def extract_images(i): image_path = png_destination+folderName+'/' + hashlib.sha224(imName.encode('utf-8')).hexdigest() + '.png' if is16Bit: # write the PNG file as a 16-bit greyscale - # change the input to use different transformations - http://support.dcmtk.org/docs/dcmj2pnm.html - # Popen(['dcmj2pnm', '+on2', dicom_path, image_path]).wait() - Popen(['dcmj2pnm', '+on2', '--min-max-window', dicom_path, image_path]).wait() + image_2d = ds.pixel_array.astype(np.double) + # # Rescaling grey scale between 0-255 + image_2d_scaled = (np.maximum(image_2d,0) / image_2d.max()) * 65535.0 + # # Convert to uint + shape = ds.pixel_array.shape + image_2d_scaled = np.uint16(image_2d_scaled) + with open(pngfile , 'wb') as png_file: + w = png.Writer(shape[1], shape[0], greyscale=True,bitdepth=16) + w.write(png_file, image_2d_scaled) else: shape = ds.pixel_array.shape # # Convert to float to avoid overflow or underflow losses.
TF on some platforms never landed as 2.0.0 Using any available version between 2.0 and 2.1 for our TF2 tests.
@@ -7,7 +7,7 @@ These tests use `guild.tfevent`, which requires that we call Install TensorFlow 2: - >>> quiet("pip install tensorflow==2.0.0 --upgrade", + >>> quiet("pip install tensorflow>=2.0.0,<2.1 --upgrade", ... ignore="DEPRECATION") NOTE: We install 2.0.0 here because `tensorflow` 2.1 appears to be @@ -25,7 +25,7 @@ Check: >>> run("guild check --offline --tensorflow") guild_version: ... tensorboard_version: 2.0.2 - tensorflow_version: 2.0.0 + tensorflow_version: 2.0... tensorflow_cuda_support: no tensorflow_gpu_available: no ...
Cleaning up some types Summary: We don't need metaclass here.
@@ -21,14 +21,9 @@ from typing import ( # noqa FEATURES = Dict[int, float] ACTION = Union[str, FEATURES] -ACTION_TYPEVAR = TypeVar("ACTION_TYPEVAR", str, FEATURES) -class NamedTupleGenericMeta(NamedTupleMeta, GenericMeta): - pass - - -class Samples(NamedTuple, Generic[ACTION_TYPEVAR], metaclass=NamedTupleGenericMeta): +class Samples(NamedTuple): mdp_ids: List[str] sequence_numbers: List[int] sequence_number_ordinals: List[int] @@ -43,9 +38,7 @@ class Samples(NamedTuple, Generic[ACTION_TYPEVAR], metaclass=NamedTupleGenericMe possible_next_actions: List[List[ACTION]] -class MultiStepSamples( - NamedTuple, Generic[ACTION_TYPEVAR], metaclass=NamedTupleGenericMeta -): +class MultiStepSamples(NamedTuple): mdp_ids: List[str] sequence_numbers: List[int] sequence_number_ordinals: List[int]
Update lhs.py added a "approximate time left" print for console output
@@ -110,10 +110,14 @@ class lhs(_algorithm): self.datawriter.save(like, randompar, simulations=simulations) # Progress bar acttime = time.time() + + #create string of the aproximate time left to complete all runs + timestr = time.strftime("%H:%M:%S", time.gmtime(round(((acttime - starttime) / + (rep + 1)) * (repetitions - (rep + 1))))) # Refresh progressbar every second if acttime - intervaltime >= 2: - text = '%i of %i (best like=%g)' % ( - rep, repetitions, self.status.objectivefunction) + text = '%i of %i (best like=%g) est. time remaining: %s' % (rep, repetitions, + self.status.objectivefunction, timestr) print(text) intervaltime = time.time() self.repeat.terminate()
added originalBasename and originalDirname to instance Will be used to fill 'source' (and 'online') template. Source template is used to publish in-situ, eg. without copying possibly massive files (as pointcaches etc.)
+import os.path + import pyblish.api @@ -22,3 +24,9 @@ class CollectSource(pyblish.api.ContextPlugin): self.log.info(( "Source of instance \"{}\" was already set to \"{}\"" ).format(instance.data["name"], source)) + + if not instance.data.get("originalBasename"): + instance.data["originalBasename"] = os.path.basename(source) + + if not instance.data.get("originalDirname"): + instance.data["originalDirname"] = os.path.dirname(source)
Removing radiation check Since it wasn't working and isn't really necessary
@@ -1304,6 +1304,8 @@ class SensitivityDemandSimulateTool(object): # num_simulations, sample_index samples_folder = parameters[2].valueAsText + num_simulations = parameters[4] + sample_index = parameters[5] if samples_folder is None: return elif not os.path.exists(samples_folder): @@ -1315,11 +1317,11 @@ class SensitivityDemandSimulateTool(object): parameters[2].setErrorMessage('Samples file not found in %s' % samples_folder) return else: - # num_simulations - parameters[4].enabled = True - parameters[4].value = np.load(os.path.join(samples_folder,'samples.npy')).shape[0] - # sample_index - parameters[5].enabled = True + if not num_simulations.enabled: + # only overwrite on first try + num_simulations.enabled = True + num_simulations.value = np.load(os.path.join(samples_folder,'samples.npy')).shape[0] + sample_index.enabled = True return def updateMessages(self, parameters): @@ -1330,10 +1332,6 @@ class SensitivityDemandSimulateTool(object): elif not os.path.exists(scenario_path): parameters[0].setErrorMessage('Scenario folder not found: %s' % scenario_path) return - elif not os.path.exists(get_radiation(scenario_path)): - parameters[0].setErrorMessage("No radiation data found for scenario. Run radiation script first.") - elif not os.path.exists(get_surface_properties(scenario_path)): - parameters[0].setErrorMessage("No radiation data found for scenario. Run radiation script first.") # samples_folder samples_folder = parameters[2].valueAsText
Fix: typo in link to Papermill library I found this issue when I clicked the link and got a 404
@@ -91,7 +91,7 @@ Dagster works with the tools and systems that you're already using with your dat &nbsp;/&nbsp; <img style="vertical-align:middle" src="https://user-images.githubusercontent.com/609349/57987827-fa268b80-7a3b-11e9-8a18-b675d76c19aa.png"> </td> <td style="border-left: 0px"> <b>Jupyter / Papermill</b></td> - <td><a href="https://github.com/dagster-io/dagster/tree/master/python_modules/dagstermill" />dagstermill</a><br />Built on the <a href=https://github.com/nteract/papermill">papermill library</a>, dagstermill is meant for integrating productionized Jupyter notebooks into dagster pipelines.</td> + <td><a href="https://github.com/dagster-io/dagster/tree/master/python_modules/dagstermill" />dagstermill</a><br />Built on the <a href="https://github.com/nteract/papermill">papermill library</a>, dagstermill is meant for integrating productionized Jupyter notebooks into dagster pipelines.</td> </tr> <tr> <td align="center" style="border-right: 0px"><img style="vertical-align:middle" src="https://user-images.githubusercontent.com/609349/57988016-f431aa00-7a3d-11e9-8cb6-1309d4246b27.png"></td>
Fix cleaning documents Field <node> is missing in CLI example, this patch will make it right.
@@ -169,20 +169,20 @@ higher. Examples of doing this with a JSON string:: - ironic --ironic-api-version 1.15 node-set-provision-state \ + ironic --ironic-api-version 1.15 node-set-provision-state <node> \ clean --clean-steps '[{"interface": "deploy", "step": "erase_devices_metadata"}]' - ironic --ironic-api-version 1.15 node-set-provision-state \ + ironic --ironic-api-version 1.15 node-set-provision-state <node> \ clean --clean-steps '[{"interface": "deploy", "step": "erase_devices"}]' Or with a file:: - ironic --ironic-api-version 1.15 node-set-provision-state \ + ironic --ironic-api-version 1.15 node-set-provision-state <node> \ clean --clean-steps my-clean-steps.txt Or with stdin:: - cat my-clean-steps.txt | ironic --ironic-api-version 1.15 \ + cat my-clean-steps.txt | ironic --ironic-api-version 1.15 <node> \ node-set-provision-state clean --clean-steps - Cleaning Network
[nightly] Move scheduling tests into one suite For future convenience, we are moving scheduling-related tests into one suite for easier monitoring and benchmarking.
@@ -61,8 +61,6 @@ CORE_NIGHTLY_TESTS = { "non_streaming_shuffle_100gb", "non_streaming_shuffle_50gb_large_partition", "non_streaming_shuffle_50gb", - "dask_on_ray_10gb_sort", - "dask_on_ray_100gb_sort", SmokeTest("dask_on_ray_large_scale_test_no_spilling"), SmokeTest("dask_on_ray_large_scale_test_spilling"), "stress_test_placement_group", @@ -78,9 +76,7 @@ CORE_NIGHTLY_TESTS = { SmokeTest("stress_test_many_tasks"), SmokeTest("stress_test_dead_actors"), "shuffle_data_loader", - "dask_on_ray_1tb_sort", SmokeTest("threaded_actors_stress_test"), - "placement_group_performance_test", "pg_long_running_performance_test", ], "~/ray/benchmarks/benchmark_tests.yaml": [ @@ -135,7 +131,6 @@ CORE_DAILY_TESTS = { "k8s_threaded_actors_stress_test", "stress_test_many_tasks", "stress_test_dead_actors", - "many_nodes_actor_test", ], "~/ray/release/nightly_tests/chaos_test.yaml": [ "chaos_dask_on_ray_large_scale_test_no_spilling", @@ -149,11 +144,6 @@ CORE_SCALABILITY_TESTS_DAILY = { "many_tasks", "many_pgs", "many_nodes", - "scheduling_test_many_0s_tasks_single_node", - "scheduling_test_many_0s_tasks_many_nodes", - # Reenable these two once we got right setup - # "scheduling_test_many_5s_tasks_single_node", - # "scheduling_test_many_5s_tasks_many_nodes", ], } @@ -166,6 +156,23 @@ CORE_REDIS_HA_TESTS_DAILY = { ], } +CORE_SCHEDULING_DAILY = { + "~/ray/benchmarks/benchmark_tests.yaml": [ + "scheduling_test_many_0s_tasks_single_node", + "scheduling_test_many_0s_tasks_many_nodes", + # Reenable these two once we got right setup + # "scheduling_test_many_5s_tasks_single_node", + # "scheduling_test_many_5s_tasks_many_nodes", + ], + "~/ray/release/nightly_tests/nightly_tests.yaml": [ + "many_nodes_actor_test", + "dask_on_ray_10gb_sort", + "dask_on_ray_100gb_sort", + "dask_on_ray_1tb_sort", + "placement_group_performance_test", + ], +} + NIGHTLY_TESTS = { # "~/ray/release/horovod_tests/horovod_tests.yaml": [ # SmokeTest("horovod_test"), @@ -293,6 +300,7 @@ SUITES = { **NIGHTLY_TESTS, **USER_TESTS }, + "core-scheduling-daily": CORE_SCHEDULING_DAILY, "weekly": WEEKLY_TESTS, }
Fix issue with column delimitor in CHECKSUMS Changed the way of parsing CHECKSUMS results
@@ -53,10 +53,10 @@ recipe: wget -c $vcf_url.tbi sum_obs=`sum $vcf_local` - sum_exp=`grep $vcf_local CHECKSUMS | awk '{print $1,$2}'` + sum_exp=sum_exp=`awk 'BEGIN{OFS=" "} $3==FN{gsub("\t"FN,""); print $0}' "FN=$vcf_local" CHECKSUMS` sum_obs_index=`sum $vcf_local.tbi` - sum_exp_index=`grep $vcf_local.tbi CHECKSUMS | awk '{print $1,$2}'` + sum_exp_index=sum_exp=`awk 'BEGIN{OFS=" "} $3==FN{gsub("\t"FN,""); print $0}' "FN=$vcf_local.tbi" CHECKSUMS` if [[ $sum_obs != $sum_exp ]] then
Standalone: Fix, wasn't detecting the standard library extension modules as standard library. * On Windows these live in a dedicated directory, which meant they were not discovered by standard library freezing and could be missing therefore.
@@ -28,6 +28,8 @@ module. import os +from nuitka.utils.Utils import getOS + def getStandardLibraryPaths(): """ Get the standard library paths. @@ -108,6 +110,12 @@ def getStandardLibraryPaths(): if os.path.isdir(candidate): stdlib_paths.add(candidate) + if getOS() == "Windows": + import _ctypes + stdlib_paths.add( + os.path.dirname(_ctypes.__file__) + ) + getStandardLibraryPaths.result = [ os.path.normcase(stdlib_path) for stdlib_path in
Solve_captcha option added to config Added solve_captcha to config. Setting to true will allow captcha solving, either auto (2Captcha is token found) or manual.
@@ -6,6 +6,7 @@ import requests import os from pokemongo_bot.event_manager import EventHandler +from pokemongo_bot.base_task import BaseTask from sys import platform as _platform from selenium import webdriver from selenium.webdriver.support.ui import WebDriverWait @@ -14,9 +15,11 @@ SITE_KEY = '6LeeTScTAAAAADqvhqVMhPpr_vB9D364Ia-1dSgK' class CaptchaHandler(EventHandler): - def __init__(self, bot): + def __init__(self, bot, captcha_solving): super(CaptchaHandler, self).__init__() self.bot = bot + self.enabled = captcha_solving + def get_token(self, url): token = '' @@ -61,6 +64,7 @@ class CaptchaHandler(EventHandler): def handle_event(self, event, sender, level, formatted_msg, data): if event in ('pokestop_searching_too_often', 'login_successful'): self.bot.logger.info('Checking for captcha challenge.') + # test manual url = 'http://www.google.com/xhtml' response_dict = self.bot.api.check_challenge() @@ -75,6 +79,16 @@ class CaptchaHandler(EventHandler): return url = challenge['challenge_url'] + if self.enabled == False: + self.bot.event_manager.emit( + 'captcha', + sender=self, + level='info', + formatted="Captcha encountered but solving diabled, exiting..." + ) + sys.exit(1) + return + if not self.bot.config.twocaptcha_token: self.bot.logger.warn('No 2captcha token set, executing manual solving') token = self.get_token(url)
GCE: return a clear error message on deletes I'm working on another PR to actually handle VM deletes the way we discussed, but this at least allows GCE to return a clear message.
@@ -1130,14 +1130,20 @@ class GceCmds(CommonCloudFunctions) : self.take_action_if_requested("VM", obj_attr_list, "deprovision_finished") except CldOpsException, obj : + for line in traceback.format_exc().splitlines() : + cbwarn(line, True) _status = obj.status _fmsg = str(obj.msg) except GCEException, obj : + for line in traceback.format_exc().splitlines() : + cbwarn(line, True) _status = int(obj.error_code) _fmsg = str(obj.error_message) except Exception, msg : + for line in traceback.format_exc().splitlines() : + cbwarn(line, True) _fmsg = str(msg) _status = 23
Removed AddonIsEnabled check from UpNext menu not working anymore
<setting id="ProgressManager_enabled" type="bool" label="30235" default="false"/> <setting label="30238" type="lsep"/> <setting id="upnext_install" type="action" label="30234" action="InstallAddon(service.upnext)" visible="!System.HasAddon(service.upnext)" option="close"/> - <setting id="UpNextNotifier_enabled" type="bool" label="30129" default="false" enable="System.AddonIsEnabled(service.upnext)"/> + <setting id="UpNextNotifier_enabled" type="bool" label="30129" default="false" visible="System.HasAddon(service.upnext)"/> <setting id="upnext_settings" type="action" label="30178" action="Addon.OpenSettings(service.upnext)" visible="System.HasAddon(service.upnext)" option="close"/> </category> <category label="30023"><!--Expert-->
Update docker.py The repository:tag should be like seqrise.com:5000/clabee/bwa:latest, which is accepted by docker. If sp[1] doesn't match valid tag names, it maybe a part of repository. And if len(sp) == 3, sp[0] and sp[1] should merged as a repository name
@@ -27,6 +27,17 @@ def get_image(dockerRequirement, pull_image, dry_run=False): sp = dockerRequirement["dockerImageId"].split(":") if len(sp) == 1: sp.append("latest") + elif len(sp) == 2: + # if sp[1] doesn't match valid tag names, it is a part of repository + if not re.match(r'[\w][\w.-]{0,127}', sp[1]): + sp[0] = sp[0] + ":" + sp[1] + sp[1] = "latest" + elif len(sp) == 3: + if re.match(r'[\w][\w.-]{0,127}', sp[2]): + sp[0] = sp[0] + ":" + sp[1] + sp[1] = sp[2] + del sp[2] + # check for repository:tag match or image id match if ((sp[0] == m.group(1) and sp[1] == m.group(2)) or dockerRequirement["dockerImageId"] == m.group(3)): found = True
DontSkip.traverse_dontskip: add rules without processing locations These grammar rules are artificial, i.e. are implementation/lowering details, so manually propagate original location to them instead of trying to guess the location from the traceback. TN:
@@ -669,16 +669,15 @@ class Parser(object): self.dontskip_parser = _pick_impl( [Null(get_context().root_grammar_class)] + list(self.dontskip_parsers), - True + True, + location=self.location ) self.dontskip_parser.is_dont_skip_parser = True # Add a named rule for the the DontSkip parsers. Don't forget to # compile it (compute their types). - grammar.add_rules(**{ - gen_name('dontskip_{}'.format(self.name)).lower: - self.dontskip_parser - }) + grammar._add_rule(gen_name('dontskip_{}'.format(self.name)).lower, + self.dontskip_parser) self.dontskip_parser.compute_types() self.dontskip_parser.freeze_types()
4.3.1 Changelog Adding 4.3.1 changelog,
@@ -4,8 +4,15 @@ This changelog summarizes updates to [Mattermost Team Edition](http://www.matter Also see [changelog in progress](http://bit.ly/2nK3cVf) for the next release. -## Release v4.3.0 -Release date: 2017-10-16 +## Release v4.3.1 + + - **v4.3.1, released 2017-10-20** + - Fixed an upgrade issue where the database schema would appear to be out of date and throw a log warning. + - Fixed the Idle Timeout setting in `config.json` by changing the setting title from `SessionIdleTimeout` to `SessionIdleTimeoutInMinutes`. + - Fixed a regression where slash commands were not functional in Direct or Group Messages. + - Mattermost v4.3.1 contains a low severity security fix. [Upgrading](http://docs.mattermost.com/administration/upgrade.html) is highly recommended. Details will be posted on our [security updates page](https://about.mattermost.com/security-updates/) 14 days after release as per the [Mattermost Responsible Disclosure Policy](https://www.mattermost.org/responsible-disclosure-policy/). + - **v4.3.0, released 2017-10-16** + - Original 4.3.0 release ### Security Update @@ -79,7 +86,7 @@ For a list of past and upcoming deprecated features, [see our website](https://a #### config.json -Multiple setting options were added to `default.json`. Below is a list of the additions and their default values on install. The settings can be modified in `default.json`, or the System Console when available. +Multiple setting options were added to `config.json`. Below is a list of the additions and their default values on install. The settings can be modified in `config.json`, or the System Console when available. **Changes to Enterprise Edition**:
SplineWidget : Fix drawing width Qt 5 seems to be defaulting to a non-cosmetic pen, whereas Qt 4 presumably defaulted to a cosmetic one.
@@ -207,6 +207,7 @@ class SplineWidget( GafferUI.Widget ) : painter.setTransform( transform ) for s in self.__splinesToDraw : pen = QtGui.QPen( self._qtColor( s.color ) ) + pen.setCosmetic( True ) painter.setPen( pen ) painter.drawPath( s.path )
Start saving comment history again Restore a line of code which was inadvertently deleted in Set the time of the comment history record to be the time its content was entered, whether it is the first edit or a later one.
@@ -30,7 +30,7 @@ from ..badges import badges from ..misc import cache, send_email, allowedNames, get_errors, engine, ensure_locale_loaded from ..misc import ratelimit, POSTING_LIMIT, AUTH_LIMIT, is_domain_banned from ..models import SubPost, SubPostComment, Sub, Message, User, UserIgnores, SubMetadata, UserSaved -from ..models import SubMod, SubBan, InviteCode, Notification, SubPostContentHistory, SubPostTitleHistory +from ..models import SubMod, SubBan, SubPostCommentHistory, InviteCode, Notification, SubPostContentHistory, SubPostTitleHistory from ..models import SubStylesheet, SubSubscriber, SubUploads, UserUploads, SiteMetadata, SubPostMetadata, SubPostReport from ..models import SubPostVote, SubPostCommentVote, UserMetadata, SubFlair, SubPostPollOption, SubPostPollVote, SubPostCommentReport, SubRule from ..models import rconn, UserStatus @@ -1837,6 +1837,8 @@ def edit_comment(): return jsonify(status='error', error=_("Post is archived")) dt = datetime.datetime.utcnow() + spm = SubPostCommentHistory.create(cid=comment.cid, content=comment.content, + datetime=(comment.lastedit or comment.time)) comment.content = form.text.data comment.lastedit = dt comment.save()
Re-add Python 3.10 to the CI tests As scikit-learn has been updated, allowing Python 3.10 versions to build on Windows
@@ -76,7 +76,7 @@ jobs: needs: [cache_nltk_data, cache_third_party] strategy: matrix: - python-version: ['3.7', '3.8', '3.9'] + python-version: ['3.7', '3.8', '3.9', '3.10'] os: [ubuntu-latest, macos-latest, windows-latest] fail-fast: false runs-on: ${{ matrix.os }}
Adds Warning For Permission Error In Ping Mostly affects linux machines not running as root.
@@ -42,8 +42,12 @@ class Latency(commands.Cog): try: url = urllib.parse.urlparse(URLs.site_schema + URLs.site).hostname + try: delay = await aioping.ping(url, family=socket.AddressFamily.AF_INET) * 1000 site_ping = f"{delay:.{ROUND_LATENCY}f} ms" + except OSError: + # Some machines do not have permission to run ping + site_ping = "Permission denied, could not ping." except TimeoutError: site_ping = f"{Emojis.cross_mark} Connection timed out."
MonadWide: docstring correction. Closes
@@ -819,7 +819,7 @@ class MonadWide(MonadTall): | | --------------------- - Using the `cmd_flip' method will switch which vertical side the + Using the ``cmd_flip`` method will switch which vertical side the main pane will occupy. The main pane is considered the "top" of the stack. @@ -838,8 +838,8 @@ class MonadWide(MonadTall): Occupying the rest of the screen are one or more secondary panes. The secondary panes will share the horizontal space of the screen - however they can be resized at will with the `cmd_grow' and - `cmd_shrink' methods. The other secondary panes will adjust their + however they can be resized at will with the ``cmd_grow`` and + ``cmd_shrink`` methods. The other secondary panes will adjust their sizes to smoothly fill all of the space. :: @@ -853,7 +853,7 @@ class MonadWide(MonadTall): | | | | | | | | --------------------- --------------------- - Panes can be moved with the `cmd_shuffle_up' and `cmd_shuffle_down' + Panes can be moved with the ``cmd_shuffle_up`` and ``cmd_shuffle_down`` methods. As mentioned the main pane is considered the top of the stack; moving up is counter-clockwise and moving down is clockwise. @@ -873,13 +873,13 @@ class MonadWide(MonadTall): Normalizing: To restore all client windows to their default size ratios simply - use the `cmd_normalize' method. + use the ``cmd_normalize`` method. Maximizing: To toggle a client window between its minimum and maximum sizes - simply use the `cmd_maximize' on a focused client. + simply use the ``cmd_maximize`` on a focused client. Suggested Bindings::
enc2gen.py: fix create_legacy_two_gpr_one_scalable_one_fixed * a few typos crept in to mess up the logic for selecting the widths of the register encoder functions. * Fix from
@@ -1310,7 +1310,7 @@ def create_legacy_two_gpr_one_scalable_one_fixed(env,ii): opsz_codes =[] for op in _gen_opnds(ii): opnds.append(op) - opsz_codes.append( get_gpr_opsz_code(opnds[0])) + opsz_codes.append( get_gpr_opsz_code(op) ) for osz in osz_list: opsig = make_opnd_signature(ii,osz) fname = "{}_{}_{}".format(enc_fn_prefix, @@ -1340,13 +1340,13 @@ def create_legacy_two_gpr_one_scalable_one_fixed(env,ii): if opsz_codes[0] in ['rv','ry']: op0_bits = osz else: - op0_bits = opsz_to_bits(opsz_codes[0]) + op0_bits = opsz_to_bits[opsz_codes[0]] fo.add_code_eol('enc_modrm_{}_gpr{}(r,{})'.format(f1,osz,var_reg0)) if opsz_codes[1] in ['rv','ry']: op1_bits = osz else: - op1_bits = opsz_to_bits(opsz_codes[1]) + op1_bits = opsz_to_bits[opsz_codes[1]] fo.add_code_eol('enc_modrm_{}_gpr{}(r,{})'.format(f2,op1_bits,var_reg1)) emit_rex(env,fo,rexw_forced)
Update workflow status badges in readme Our two workflows, Lint and Build, now have separate status badges that link to the latest results from that workflow.
# SeasonalBot -[![Build Status](https://github.com/python-discord/seasonalbot/workflows/Lint,%20Build%20&%20Deploy/badge.svg?branch=master) +[![Lint Badge][1]][2] +[![Build Badge][3]][4] [![Discord](https://img.shields.io/static/v1?label=Python%20Discord&logo=discord&message=%3E100k%20members&color=%237289DA&logoColor=white)](https://discord.gg/2B963hn) A Discord bot for the Python Discord community which changes with the seasons, and provides useful event features. @@ -17,3 +18,8 @@ This later evolved into a bot that runs all year, providing season-appropriate f Before you start, please take some time to read through our [contributing guidelines](CONTRIBUTING.md). See [Seasonalbot's Wiki](https://pythondiscord.com/pages/contributing/seasonalbot/) for in-depth guides on getting started with the project! + +[1]:https://github.com/python-discord/seasonalbot/workflows/Lint/badge.svg?branch=master +[2]:https://github.com/python-discord/seasonalbot/actions?query=workflow%3ALint+branch%3Amaster +[3]:https://github.com/python-discord/seasonalbot/workflows/Build/badge.svg?branch=master +[4]:https://github.com/python-discord/seasonalbot/actions?query=workflow%3ABuild+branch%3Amaster
settings: Display "Never" over "Invalid Date" for inactive users. This shows the text "Never" for users who are part of a realm but have never been active, rather than a more vague JavaScript output of "Invalid Date" due to the fact that their last presence evaluates to NaN.
@@ -145,8 +145,13 @@ function populate_users(realm_people_data) { } else if (presence.presence_info[item.user_id]) { // XDate takes number of milliseconds since UTC epoch. var last_active = presence.presence_info[item.user_id].last_active * 1000; + + if (!isNaN(last_active)) { var last_active_date = new XDate(last_active); activity_rendered = timerender.render_date(last_active_date, undefined, today); + } else { + activity_rendered = $("<span></span>").text(i18n.t("Never")); + } } else { activity_rendered = $("<span></span>").text(i18n.t("Unknown")); }
Fix error in shutdown_watcher Fixes the following: ``` 2017-08-08 17:04:48,799 [ERROR] conjure-up/kubernetes-core - events.py:206 - Error in cleanup code: 'generator' object has no attribute 'cr_code' Traceback (most recent call last): File "/snap/conjure-up/589/lib/python3.6/site-packages/conjureup/events.py", line 201, in shutdown_watcher if coro and coro.cr_code is not shutdown_watcher.__code__: AttributeError: 'generator' object has no attribute 'cr_code' ```
@@ -198,7 +198,8 @@ async def shutdown_watcher(): for task in asyncio.Task.all_tasks(app.loop): # cancel all other tasks coro = getattr(task, '_coro', None) - if coro and coro.cr_code is not shutdown_watcher.__code__: + cr_code = getattr(coro, 'cr_code', None) + if cr_code is not shutdown_watcher.__code__: app.log.debug('Cancelling pending task: {}'.format(task)) task.cancel() await asyncio.sleep(0.1) # give tasks a chance to see the cancel
minor: Refactor `userMentionHandler` to avoid duplicating escaping logic. We avoid duplicating escaping logic for the same variable, which helps to reduce the risk of any future bugs.
@@ -165,21 +165,24 @@ export function apply_markdown(message) { // flags on the message itself that get used by the message // view code and possibly our filtering code. - if (helpers.my_user_id() === user_id && !silently) { + // If I mention "@aLiCe sMITH", I still want "Alice Smith" to + // show in the pill. + let display_text = helpers.get_actual_name_from_user_id(user_id); + let classes; + if (silently) { + classes = "user-mention silent"; + } else { + if (helpers.my_user_id() === user_id) { message.mentioned = true; message.mentioned_me_directly = true; } - let str = ""; - if (silently) { - str += `<span class="user-mention silent" data-user-id="${_.escape(user_id)}">`; - } else { - str += `<span class="user-mention" data-user-id="${_.escape(user_id)}">@`; + classes = "user-mention"; + display_text = "@" + display_text; } - // If I mention "@aLiCe sMITH", I still want "Alice Smith" to - // show in the pill. - const actual_full_name = helpers.get_actual_name_from_user_id(user_id); - return `${str}${_.escape(actual_full_name)}</span>`; + return `<span class="${classes}" data-user-id="${_.escape(user_id)}">${_.escape( + display_text, + )}</span>`; }, groupMentionHandler(name) { const group = helpers.get_user_group_from_name(name);
DOC: fixed format and updated CHANGELOG entries Fixed format in 3.0.0 block and updated entries to avoid duplicate entries, place updates in correct categories, and describe the changes made in this PR.
@@ -9,14 +9,16 @@ This project adheres to [Semantic Versioning](http://semver.org/). - custom.attach replaces custom.add - Unit tests are now pytest compatible and use parametrize - Added altitudes to test instruments - - New flags added to instruments to streamline unit testing: `_test_download`, `_test_download_travis`, `_password_req` - - Madrigal instruments migrated to pysatMadrigal + - New flags added to instruments to streamline unit testing: + `_test_download`, `_test_download_travis`, `_password_req` - methods.nasa_cdaweb.list_files moved to methods.general - `strict_time_flag` now defaults to True - Use of start / stop notation in remote_file_list - Added variable rename method to Instrument object (#91) - Migrated file methods to pysat.utils.files (#336) - Deprecations + - Migraged instruments to pysatMadrigal, pysatNASA, pysatSpaceWeather, + pysatIncubator, pysatModels, pysatCDAAC, and pysatMissions - Removed ssnl - Removed utils.stats - Removed model_utils @@ -25,9 +27,11 @@ This project adheres to [Semantic Versioning](http://semver.org/). - import DataFrame and Series directly from pandas, not pysat - Removed coords.scale_units - Removed time.season_date_range - - DeprecationWarning for strict_time_flag only triggered if sloppy data is found + - DeprecationWarning for strict_time_flag only triggered if sloppy data is + found - Remove convenience methods imported from pandas - - Changed the default `custom.attatch` input to allow keyword arguement use when additional function input is required + - Changed the default `custom.attatch` input to allow keyword arguement use + when additional function input is required - Removed python 2.7 syntax - Removed utils.coords.geodetic_to_geocentric - Removed utils.coords.geodetic_to_geocentric_horizontal @@ -35,6 +39,7 @@ This project adheres to [Semantic Versioning](http://semver.org/). - Removed utils.coords.global_to_local_cartesian - Removed utils.coords.local_horizontal_to_global_geo - Deprecation Warnings for methods in `pysat._files` + - Addressed several Warnings raised by incorrect use of dependent packages - Documentation - Added info on how to register new instruments - Fixed description of tag and sat_id behaviour in testing instruments @@ -47,7 +52,8 @@ This project adheres to [Semantic Versioning](http://semver.org/). - Fixed custom instrument attribute persistence upon load - Improved string handling robustness when writing netCDF4 files in Python 3 - Improved pandas 1.1.0 compatibility in tests - - Fixed coupling of two_digit_year_break keyword to underlying method in methods.general.list_files + - Fixed coupling of two_digit_year_break keyword to underlying method in + methods.general.list_files - Fixed additional file date range for monthly data with gaps - Maintenance - nose dependency removed from unit tests @@ -56,8 +62,8 @@ This project adheres to [Semantic Versioning](http://semver.org/). - Include flake8 check as part of testing suites - Improve unit testing coverage of instrument functions and instrument object - Add tests for acknowledgements and references - - Removed implicit conversion to integers in methods.general.convert_timestamp_to_datetime - - Removed instruments (migrated to new libraries) + - Removed implicit conversion to integers in + methods.general.convert_timestamp_to_datetime ## [2.2.1] - 2020-07-29 - Documentation