repo
stringlengths 7
55
| path
stringlengths 4
127
| func_name
stringlengths 1
88
| original_string
stringlengths 75
19.8k
| language
stringclasses 1
value | code
stringlengths 75
19.8k
| code_tokens
list | docstring
stringlengths 3
17.3k
| docstring_tokens
list | sha
stringlengths 40
40
| url
stringlengths 87
242
| partition
stringclasses 1
value |
---|---|---|---|---|---|---|---|---|---|---|---|
googlefonts/fontbakery | Lib/fontbakery/profiles/googlefonts.py | com_google_fonts_check_family_control_chars | def com_google_fonts_check_family_control_chars(ttFonts):
"""Does font file include unacceptable control character glyphs?"""
# list of unacceptable control character glyph names
# definition includes the entire control character Unicode block except:
# - .null (U+0000)
# - CR (U+000D)
unacceptable_cc_list = [
"uni0001",
"uni0002",
"uni0003",
"uni0004",
"uni0005",
"uni0006",
"uni0007",
"uni0008",
"uni0009",
"uni000A",
"uni000B",
"uni000C",
"uni000E",
"uni000F",
"uni0010",
"uni0011",
"uni0012",
"uni0013",
"uni0014",
"uni0015",
"uni0016",
"uni0017",
"uni0018",
"uni0019",
"uni001A",
"uni001B",
"uni001C",
"uni001D",
"uni001E",
"uni001F"
]
# a dict with key:value of font path that failed check : list of unacceptable glyph names
failed_font_dict = {}
for ttFont in ttFonts:
font_failed = False
unacceptable_glyphs_in_set = [] # a list of unacceptable glyph names identified
glyph_name_set = set(ttFont["glyf"].glyphs.keys())
fontname = ttFont.reader.file.name
for unacceptable_glyph_name in unacceptable_cc_list:
if unacceptable_glyph_name in glyph_name_set:
font_failed = True
unacceptable_glyphs_in_set.append(unacceptable_glyph_name)
if font_failed:
failed_font_dict[fontname] = unacceptable_glyphs_in_set
if len(failed_font_dict) > 0:
unacceptable_cc_report_string = "The following unacceptable control characters were identified:\n"
for fnt in failed_font_dict.keys():
unacceptable_cc_report_string += " {}: {}\n".format(
fnt, ", ".join(failed_font_dict[fnt])
)
yield FAIL, ("{}".format(unacceptable_cc_report_string))
else:
yield PASS, ("Unacceptable control characters were not identified.") | python | def com_google_fonts_check_family_control_chars(ttFonts):
"""Does font file include unacceptable control character glyphs?"""
# list of unacceptable control character glyph names
# definition includes the entire control character Unicode block except:
# - .null (U+0000)
# - CR (U+000D)
unacceptable_cc_list = [
"uni0001",
"uni0002",
"uni0003",
"uni0004",
"uni0005",
"uni0006",
"uni0007",
"uni0008",
"uni0009",
"uni000A",
"uni000B",
"uni000C",
"uni000E",
"uni000F",
"uni0010",
"uni0011",
"uni0012",
"uni0013",
"uni0014",
"uni0015",
"uni0016",
"uni0017",
"uni0018",
"uni0019",
"uni001A",
"uni001B",
"uni001C",
"uni001D",
"uni001E",
"uni001F"
]
# a dict with key:value of font path that failed check : list of unacceptable glyph names
failed_font_dict = {}
for ttFont in ttFonts:
font_failed = False
unacceptable_glyphs_in_set = [] # a list of unacceptable glyph names identified
glyph_name_set = set(ttFont["glyf"].glyphs.keys())
fontname = ttFont.reader.file.name
for unacceptable_glyph_name in unacceptable_cc_list:
if unacceptable_glyph_name in glyph_name_set:
font_failed = True
unacceptable_glyphs_in_set.append(unacceptable_glyph_name)
if font_failed:
failed_font_dict[fontname] = unacceptable_glyphs_in_set
if len(failed_font_dict) > 0:
unacceptable_cc_report_string = "The following unacceptable control characters were identified:\n"
for fnt in failed_font_dict.keys():
unacceptable_cc_report_string += " {}: {}\n".format(
fnt, ", ".join(failed_font_dict[fnt])
)
yield FAIL, ("{}".format(unacceptable_cc_report_string))
else:
yield PASS, ("Unacceptable control characters were not identified.") | [
"def",
"com_google_fonts_check_family_control_chars",
"(",
"ttFonts",
")",
":",
"# list of unacceptable control character glyph names",
"# definition includes the entire control character Unicode block except:",
"# - .null (U+0000)",
"# - CR (U+000D)",
"unacceptable_cc_list",
"=",
"[",
"\"uni0001\"",
",",
"\"uni0002\"",
",",
"\"uni0003\"",
",",
"\"uni0004\"",
",",
"\"uni0005\"",
",",
"\"uni0006\"",
",",
"\"uni0007\"",
",",
"\"uni0008\"",
",",
"\"uni0009\"",
",",
"\"uni000A\"",
",",
"\"uni000B\"",
",",
"\"uni000C\"",
",",
"\"uni000E\"",
",",
"\"uni000F\"",
",",
"\"uni0010\"",
",",
"\"uni0011\"",
",",
"\"uni0012\"",
",",
"\"uni0013\"",
",",
"\"uni0014\"",
",",
"\"uni0015\"",
",",
"\"uni0016\"",
",",
"\"uni0017\"",
",",
"\"uni0018\"",
",",
"\"uni0019\"",
",",
"\"uni001A\"",
",",
"\"uni001B\"",
",",
"\"uni001C\"",
",",
"\"uni001D\"",
",",
"\"uni001E\"",
",",
"\"uni001F\"",
"]",
"# a dict with key:value of font path that failed check : list of unacceptable glyph names",
"failed_font_dict",
"=",
"{",
"}",
"for",
"ttFont",
"in",
"ttFonts",
":",
"font_failed",
"=",
"False",
"unacceptable_glyphs_in_set",
"=",
"[",
"]",
"# a list of unacceptable glyph names identified",
"glyph_name_set",
"=",
"set",
"(",
"ttFont",
"[",
"\"glyf\"",
"]",
".",
"glyphs",
".",
"keys",
"(",
")",
")",
"fontname",
"=",
"ttFont",
".",
"reader",
".",
"file",
".",
"name",
"for",
"unacceptable_glyph_name",
"in",
"unacceptable_cc_list",
":",
"if",
"unacceptable_glyph_name",
"in",
"glyph_name_set",
":",
"font_failed",
"=",
"True",
"unacceptable_glyphs_in_set",
".",
"append",
"(",
"unacceptable_glyph_name",
")",
"if",
"font_failed",
":",
"failed_font_dict",
"[",
"fontname",
"]",
"=",
"unacceptable_glyphs_in_set",
"if",
"len",
"(",
"failed_font_dict",
")",
">",
"0",
":",
"unacceptable_cc_report_string",
"=",
"\"The following unacceptable control characters were identified:\\n\"",
"for",
"fnt",
"in",
"failed_font_dict",
".",
"keys",
"(",
")",
":",
"unacceptable_cc_report_string",
"+=",
"\" {}: {}\\n\"",
".",
"format",
"(",
"fnt",
",",
"\", \"",
".",
"join",
"(",
"failed_font_dict",
"[",
"fnt",
"]",
")",
")",
"yield",
"FAIL",
",",
"(",
"\"{}\"",
".",
"format",
"(",
"unacceptable_cc_report_string",
")",
")",
"else",
":",
"yield",
"PASS",
",",
"(",
"\"Unacceptable control characters were not identified.\"",
")"
] | Does font file include unacceptable control character glyphs? | [
"Does",
"font",
"file",
"include",
"unacceptable",
"control",
"character",
"glyphs?"
] | b355aea2e619a4477769e060d24c32448aa65399 | https://github.com/googlefonts/fontbakery/blob/b355aea2e619a4477769e060d24c32448aa65399/Lib/fontbakery/profiles/googlefonts.py#L3818-L3882 | train |
googlefonts/fontbakery | Lib/fontbakery/profiles/googlefonts.py | gfonts_repo_structure | def gfonts_repo_structure(fonts):
""" The family at the given font path
follows the files and directory structure
typical of a font project hosted on
the Google Fonts repo on GitHub ? """
from fontbakery.utils import get_absolute_path
# FIXME: Improve this with more details
# about the expected structure.
abspath = get_absolute_path(fonts[0])
return abspath.split(os.path.sep)[-3] in ["ufl", "ofl", "apache"] | python | def gfonts_repo_structure(fonts):
""" The family at the given font path
follows the files and directory structure
typical of a font project hosted on
the Google Fonts repo on GitHub ? """
from fontbakery.utils import get_absolute_path
# FIXME: Improve this with more details
# about the expected structure.
abspath = get_absolute_path(fonts[0])
return abspath.split(os.path.sep)[-3] in ["ufl", "ofl", "apache"] | [
"def",
"gfonts_repo_structure",
"(",
"fonts",
")",
":",
"from",
"fontbakery",
".",
"utils",
"import",
"get_absolute_path",
"# FIXME: Improve this with more details",
"# about the expected structure.",
"abspath",
"=",
"get_absolute_path",
"(",
"fonts",
"[",
"0",
"]",
")",
"return",
"abspath",
".",
"split",
"(",
"os",
".",
"path",
".",
"sep",
")",
"[",
"-",
"3",
"]",
"in",
"[",
"\"ufl\"",
",",
"\"ofl\"",
",",
"\"apache\"",
"]"
] | The family at the given font path
follows the files and directory structure
typical of a font project hosted on
the Google Fonts repo on GitHub ? | [
"The",
"family",
"at",
"the",
"given",
"font",
"path",
"follows",
"the",
"files",
"and",
"directory",
"structure",
"typical",
"of",
"a",
"font",
"project",
"hosted",
"on",
"the",
"Google",
"Fonts",
"repo",
"on",
"GitHub",
"?"
] | b355aea2e619a4477769e060d24c32448aa65399 | https://github.com/googlefonts/fontbakery/blob/b355aea2e619a4477769e060d24c32448aa65399/Lib/fontbakery/profiles/googlefonts.py#L3886-L3896 | train |
googlefonts/fontbakery | Lib/fontbakery/profiles/googlefonts.py | com_google_fonts_check_repo_dirname_match_nameid_1 | def com_google_fonts_check_repo_dirname_match_nameid_1(fonts,
gfonts_repo_structure):
"""Directory name in GFonts repo structure must
match NameID 1 of the regular."""
from fontTools.ttLib import TTFont
from fontbakery.utils import (get_name_entry_strings,
get_absolute_path,
get_regular)
regular = get_regular(fonts)
if not regular:
yield FAIL, "The font seems to lack a regular."
entry = get_name_entry_strings(TTFont(regular), NameID.FONT_FAMILY_NAME)[0]
expected = entry.lower()
expected = "".join(expected.split(' '))
expected = "".join(expected.split('-'))
license, familypath, filename = get_absolute_path(regular).split(os.path.sep)[-3:]
if familypath == expected:
yield PASS, "OK"
else:
yield FAIL, (f"Family name on the name table ('{entry}') does not match"
f" directory name in the repo structure ('{familypath}')."
f" Expected '{expected}'.") | python | def com_google_fonts_check_repo_dirname_match_nameid_1(fonts,
gfonts_repo_structure):
"""Directory name in GFonts repo structure must
match NameID 1 of the regular."""
from fontTools.ttLib import TTFont
from fontbakery.utils import (get_name_entry_strings,
get_absolute_path,
get_regular)
regular = get_regular(fonts)
if not regular:
yield FAIL, "The font seems to lack a regular."
entry = get_name_entry_strings(TTFont(regular), NameID.FONT_FAMILY_NAME)[0]
expected = entry.lower()
expected = "".join(expected.split(' '))
expected = "".join(expected.split('-'))
license, familypath, filename = get_absolute_path(regular).split(os.path.sep)[-3:]
if familypath == expected:
yield PASS, "OK"
else:
yield FAIL, (f"Family name on the name table ('{entry}') does not match"
f" directory name in the repo structure ('{familypath}')."
f" Expected '{expected}'.") | [
"def",
"com_google_fonts_check_repo_dirname_match_nameid_1",
"(",
"fonts",
",",
"gfonts_repo_structure",
")",
":",
"from",
"fontTools",
".",
"ttLib",
"import",
"TTFont",
"from",
"fontbakery",
".",
"utils",
"import",
"(",
"get_name_entry_strings",
",",
"get_absolute_path",
",",
"get_regular",
")",
"regular",
"=",
"get_regular",
"(",
"fonts",
")",
"if",
"not",
"regular",
":",
"yield",
"FAIL",
",",
"\"The font seems to lack a regular.\"",
"entry",
"=",
"get_name_entry_strings",
"(",
"TTFont",
"(",
"regular",
")",
",",
"NameID",
".",
"FONT_FAMILY_NAME",
")",
"[",
"0",
"]",
"expected",
"=",
"entry",
".",
"lower",
"(",
")",
"expected",
"=",
"\"\"",
".",
"join",
"(",
"expected",
".",
"split",
"(",
"' '",
")",
")",
"expected",
"=",
"\"\"",
".",
"join",
"(",
"expected",
".",
"split",
"(",
"'-'",
")",
")",
"license",
",",
"familypath",
",",
"filename",
"=",
"get_absolute_path",
"(",
"regular",
")",
".",
"split",
"(",
"os",
".",
"path",
".",
"sep",
")",
"[",
"-",
"3",
":",
"]",
"if",
"familypath",
"==",
"expected",
":",
"yield",
"PASS",
",",
"\"OK\"",
"else",
":",
"yield",
"FAIL",
",",
"(",
"f\"Family name on the name table ('{entry}') does not match\"",
"f\" directory name in the repo structure ('{familypath}').\"",
"f\" Expected '{expected}'.\"",
")"
] | Directory name in GFonts repo structure must
match NameID 1 of the regular. | [
"Directory",
"name",
"in",
"GFonts",
"repo",
"structure",
"must",
"match",
"NameID",
"1",
"of",
"the",
"regular",
"."
] | b355aea2e619a4477769e060d24c32448aa65399 | https://github.com/googlefonts/fontbakery/blob/b355aea2e619a4477769e060d24c32448aa65399/Lib/fontbakery/profiles/googlefonts.py#L3906-L3929 | train |
googlefonts/fontbakery | Lib/fontbakery/profiles/os2.py | com_google_fonts_check_family_panose_proportion | def com_google_fonts_check_family_panose_proportion(ttFonts):
"""Fonts have consistent PANOSE proportion?"""
failed = False
proportion = None
for ttFont in ttFonts:
if proportion is None:
proportion = ttFont['OS/2'].panose.bProportion
if proportion != ttFont['OS/2'].panose.bProportion:
failed = True
if failed:
yield FAIL, ("PANOSE proportion is not"
" the same accross this family."
" In order to fix this,"
" please make sure that the panose.bProportion value"
" is the same in the OS/2 table of all of this family"
" font files.")
else:
yield PASS, "Fonts have consistent PANOSE proportion." | python | def com_google_fonts_check_family_panose_proportion(ttFonts):
"""Fonts have consistent PANOSE proportion?"""
failed = False
proportion = None
for ttFont in ttFonts:
if proportion is None:
proportion = ttFont['OS/2'].panose.bProportion
if proportion != ttFont['OS/2'].panose.bProportion:
failed = True
if failed:
yield FAIL, ("PANOSE proportion is not"
" the same accross this family."
" In order to fix this,"
" please make sure that the panose.bProportion value"
" is the same in the OS/2 table of all of this family"
" font files.")
else:
yield PASS, "Fonts have consistent PANOSE proportion." | [
"def",
"com_google_fonts_check_family_panose_proportion",
"(",
"ttFonts",
")",
":",
"failed",
"=",
"False",
"proportion",
"=",
"None",
"for",
"ttFont",
"in",
"ttFonts",
":",
"if",
"proportion",
"is",
"None",
":",
"proportion",
"=",
"ttFont",
"[",
"'OS/2'",
"]",
".",
"panose",
".",
"bProportion",
"if",
"proportion",
"!=",
"ttFont",
"[",
"'OS/2'",
"]",
".",
"panose",
".",
"bProportion",
":",
"failed",
"=",
"True",
"if",
"failed",
":",
"yield",
"FAIL",
",",
"(",
"\"PANOSE proportion is not\"",
"\" the same accross this family.\"",
"\" In order to fix this,\"",
"\" please make sure that the panose.bProportion value\"",
"\" is the same in the OS/2 table of all of this family\"",
"\" font files.\"",
")",
"else",
":",
"yield",
"PASS",
",",
"\"Fonts have consistent PANOSE proportion.\""
] | Fonts have consistent PANOSE proportion? | [
"Fonts",
"have",
"consistent",
"PANOSE",
"proportion?"
] | b355aea2e619a4477769e060d24c32448aa65399 | https://github.com/googlefonts/fontbakery/blob/b355aea2e619a4477769e060d24c32448aa65399/Lib/fontbakery/profiles/os2.py#L14-L32 | train |
googlefonts/fontbakery | Lib/fontbakery/profiles/os2.py | com_google_fonts_check_family_panose_familytype | def com_google_fonts_check_family_panose_familytype(ttFonts):
"""Fonts have consistent PANOSE family type?"""
failed = False
familytype = None
for ttfont in ttFonts:
if familytype is None:
familytype = ttfont['OS/2'].panose.bFamilyType
if familytype != ttfont['OS/2'].panose.bFamilyType:
failed = True
if failed:
yield FAIL, ("PANOSE family type is not"
" the same accross this family."
" In order to fix this,"
" please make sure that the panose.bFamilyType value"
" is the same in the OS/2 table of all of this family"
" font files.")
else:
yield PASS, "Fonts have consistent PANOSE family type." | python | def com_google_fonts_check_family_panose_familytype(ttFonts):
"""Fonts have consistent PANOSE family type?"""
failed = False
familytype = None
for ttfont in ttFonts:
if familytype is None:
familytype = ttfont['OS/2'].panose.bFamilyType
if familytype != ttfont['OS/2'].panose.bFamilyType:
failed = True
if failed:
yield FAIL, ("PANOSE family type is not"
" the same accross this family."
" In order to fix this,"
" please make sure that the panose.bFamilyType value"
" is the same in the OS/2 table of all of this family"
" font files.")
else:
yield PASS, "Fonts have consistent PANOSE family type." | [
"def",
"com_google_fonts_check_family_panose_familytype",
"(",
"ttFonts",
")",
":",
"failed",
"=",
"False",
"familytype",
"=",
"None",
"for",
"ttfont",
"in",
"ttFonts",
":",
"if",
"familytype",
"is",
"None",
":",
"familytype",
"=",
"ttfont",
"[",
"'OS/2'",
"]",
".",
"panose",
".",
"bFamilyType",
"if",
"familytype",
"!=",
"ttfont",
"[",
"'OS/2'",
"]",
".",
"panose",
".",
"bFamilyType",
":",
"failed",
"=",
"True",
"if",
"failed",
":",
"yield",
"FAIL",
",",
"(",
"\"PANOSE family type is not\"",
"\" the same accross this family.\"",
"\" In order to fix this,\"",
"\" please make sure that the panose.bFamilyType value\"",
"\" is the same in the OS/2 table of all of this family\"",
"\" font files.\"",
")",
"else",
":",
"yield",
"PASS",
",",
"\"Fonts have consistent PANOSE family type.\""
] | Fonts have consistent PANOSE family type? | [
"Fonts",
"have",
"consistent",
"PANOSE",
"family",
"type?"
] | b355aea2e619a4477769e060d24c32448aa65399 | https://github.com/googlefonts/fontbakery/blob/b355aea2e619a4477769e060d24c32448aa65399/Lib/fontbakery/profiles/os2.py#L38-L56 | train |
googlefonts/fontbakery | Lib/fontbakery/profiles/os2.py | com_google_fonts_check_code_pages | def com_google_fonts_check_code_pages(ttFont):
"""Check code page character ranges"""
if not hasattr(ttFont['OS/2'], "ulCodePageRange1") or \
not hasattr(ttFont['OS/2'], "ulCodePageRange2") or \
(ttFont['OS/2'].ulCodePageRange1 == 0 and \
ttFont['OS/2'].ulCodePageRange2 == 0):
yield FAIL, ("No code pages defined in the OS/2 table"
" ulCodePageRage1 and CodePageRage2 fields.")
else:
yield PASS, "At least one code page is defined." | python | def com_google_fonts_check_code_pages(ttFont):
"""Check code page character ranges"""
if not hasattr(ttFont['OS/2'], "ulCodePageRange1") or \
not hasattr(ttFont['OS/2'], "ulCodePageRange2") or \
(ttFont['OS/2'].ulCodePageRange1 == 0 and \
ttFont['OS/2'].ulCodePageRange2 == 0):
yield FAIL, ("No code pages defined in the OS/2 table"
" ulCodePageRage1 and CodePageRage2 fields.")
else:
yield PASS, "At least one code page is defined." | [
"def",
"com_google_fonts_check_code_pages",
"(",
"ttFont",
")",
":",
"if",
"not",
"hasattr",
"(",
"ttFont",
"[",
"'OS/2'",
"]",
",",
"\"ulCodePageRange1\"",
")",
"or",
"not",
"hasattr",
"(",
"ttFont",
"[",
"'OS/2'",
"]",
",",
"\"ulCodePageRange2\"",
")",
"or",
"(",
"ttFont",
"[",
"'OS/2'",
"]",
".",
"ulCodePageRange1",
"==",
"0",
"and",
"ttFont",
"[",
"'OS/2'",
"]",
".",
"ulCodePageRange2",
"==",
"0",
")",
":",
"yield",
"FAIL",
",",
"(",
"\"No code pages defined in the OS/2 table\"",
"\" ulCodePageRage1 and CodePageRage2 fields.\"",
")",
"else",
":",
"yield",
"PASS",
",",
"\"At least one code page is defined.\""
] | Check code page character ranges | [
"Check",
"code",
"page",
"character",
"ranges"
] | b355aea2e619a4477769e060d24c32448aa65399 | https://github.com/googlefonts/fontbakery/blob/b355aea2e619a4477769e060d24c32448aa65399/Lib/fontbakery/profiles/os2.py#L245-L255 | train |
googlefonts/fontbakery | Lib/fontbakery/profiles/glyf.py | com_google_fonts_check_glyf_unused_data | def com_google_fonts_check_glyf_unused_data(ttFont):
"""Is there any unused data at the end of the glyf table?"""
try:
expected_glyphs = len(ttFont.getGlyphOrder())
actual_glyphs = len(ttFont['glyf'].glyphs)
diff = actual_glyphs - expected_glyphs
if diff < 0:
yield FAIL, Message("unreachable-data",
("Glyf table has unreachable data at the end of "
" the table. Expected glyf table length {}"
" (from loca table), got length"
" {} (difference: {})").format(
expected_glyphs, actual_glyphs, diff))
elif not diff: # negative diff -> exception below
yield PASS, "There is no unused data at the end of the glyf table."
else:
raise Exception("Bug: fontTools did not raise an expected exception.")
except fontTools.ttLib.TTLibError as error:
if "not enough 'glyf' table data" in format(error):
yield FAIL, Message("missing-data",
("Loca table references data beyond"
" the end of the glyf table."
" Expected glyf table length {}"
" (from loca table).").format(expected_glyphs))
else:
raise Exception("Bug: Unexpected fontTools exception.") | python | def com_google_fonts_check_glyf_unused_data(ttFont):
"""Is there any unused data at the end of the glyf table?"""
try:
expected_glyphs = len(ttFont.getGlyphOrder())
actual_glyphs = len(ttFont['glyf'].glyphs)
diff = actual_glyphs - expected_glyphs
if diff < 0:
yield FAIL, Message("unreachable-data",
("Glyf table has unreachable data at the end of "
" the table. Expected glyf table length {}"
" (from loca table), got length"
" {} (difference: {})").format(
expected_glyphs, actual_glyphs, diff))
elif not diff: # negative diff -> exception below
yield PASS, "There is no unused data at the end of the glyf table."
else:
raise Exception("Bug: fontTools did not raise an expected exception.")
except fontTools.ttLib.TTLibError as error:
if "not enough 'glyf' table data" in format(error):
yield FAIL, Message("missing-data",
("Loca table references data beyond"
" the end of the glyf table."
" Expected glyf table length {}"
" (from loca table).").format(expected_glyphs))
else:
raise Exception("Bug: Unexpected fontTools exception.") | [
"def",
"com_google_fonts_check_glyf_unused_data",
"(",
"ttFont",
")",
":",
"try",
":",
"expected_glyphs",
"=",
"len",
"(",
"ttFont",
".",
"getGlyphOrder",
"(",
")",
")",
"actual_glyphs",
"=",
"len",
"(",
"ttFont",
"[",
"'glyf'",
"]",
".",
"glyphs",
")",
"diff",
"=",
"actual_glyphs",
"-",
"expected_glyphs",
"if",
"diff",
"<",
"0",
":",
"yield",
"FAIL",
",",
"Message",
"(",
"\"unreachable-data\"",
",",
"(",
"\"Glyf table has unreachable data at the end of \"",
"\" the table. Expected glyf table length {}\"",
"\" (from loca table), got length\"",
"\" {} (difference: {})\"",
")",
".",
"format",
"(",
"expected_glyphs",
",",
"actual_glyphs",
",",
"diff",
")",
")",
"elif",
"not",
"diff",
":",
"# negative diff -> exception below",
"yield",
"PASS",
",",
"\"There is no unused data at the end of the glyf table.\"",
"else",
":",
"raise",
"Exception",
"(",
"\"Bug: fontTools did not raise an expected exception.\"",
")",
"except",
"fontTools",
".",
"ttLib",
".",
"TTLibError",
"as",
"error",
":",
"if",
"\"not enough 'glyf' table data\"",
"in",
"format",
"(",
"error",
")",
":",
"yield",
"FAIL",
",",
"Message",
"(",
"\"missing-data\"",
",",
"(",
"\"Loca table references data beyond\"",
"\" the end of the glyf table.\"",
"\" Expected glyf table length {}\"",
"\" (from loca table).\"",
")",
".",
"format",
"(",
"expected_glyphs",
")",
")",
"else",
":",
"raise",
"Exception",
"(",
"\"Bug: Unexpected fontTools exception.\"",
")"
] | Is there any unused data at the end of the glyf table? | [
"Is",
"there",
"any",
"unused",
"data",
"at",
"the",
"end",
"of",
"the",
"glyf",
"table?"
] | b355aea2e619a4477769e060d24c32448aa65399 | https://github.com/googlefonts/fontbakery/blob/b355aea2e619a4477769e060d24c32448aa65399/Lib/fontbakery/profiles/glyf.py#L13-L39 | train |
googlefonts/fontbakery | Lib/fontbakery/profiles/glyf.py | com_google_fonts_check_points_out_of_bounds | def com_google_fonts_check_points_out_of_bounds(ttFont):
"""Check for points out of bounds."""
failed = False
out_of_bounds = []
for glyphName in ttFont['glyf'].keys():
glyph = ttFont['glyf'][glyphName]
coords = glyph.getCoordinates(ttFont['glyf'])[0]
for x, y in coords:
if x < glyph.xMin or x > glyph.xMax or \
y < glyph.yMin or y > glyph.yMax or \
abs(x) > 32766 or abs(y) > 32766:
failed = True
out_of_bounds.append((glyphName, x, y))
if failed:
yield WARN, ("The following glyphs have coordinates which are"
" out of bounds:\n{}\nThis happens a lot when points"
" are not extremes, which is usually bad. However,"
" fixing this alert by adding points on extremes may"
" do more harm than good, especially with italics,"
" calligraphic-script, handwriting, rounded and"
" other fonts. So it is common to"
" ignore this message".format(out_of_bounds))
else:
yield PASS, "All glyph paths have coordinates within bounds!" | python | def com_google_fonts_check_points_out_of_bounds(ttFont):
"""Check for points out of bounds."""
failed = False
out_of_bounds = []
for glyphName in ttFont['glyf'].keys():
glyph = ttFont['glyf'][glyphName]
coords = glyph.getCoordinates(ttFont['glyf'])[0]
for x, y in coords:
if x < glyph.xMin or x > glyph.xMax or \
y < glyph.yMin or y > glyph.yMax or \
abs(x) > 32766 or abs(y) > 32766:
failed = True
out_of_bounds.append((glyphName, x, y))
if failed:
yield WARN, ("The following glyphs have coordinates which are"
" out of bounds:\n{}\nThis happens a lot when points"
" are not extremes, which is usually bad. However,"
" fixing this alert by adding points on extremes may"
" do more harm than good, especially with italics,"
" calligraphic-script, handwriting, rounded and"
" other fonts. So it is common to"
" ignore this message".format(out_of_bounds))
else:
yield PASS, "All glyph paths have coordinates within bounds!" | [
"def",
"com_google_fonts_check_points_out_of_bounds",
"(",
"ttFont",
")",
":",
"failed",
"=",
"False",
"out_of_bounds",
"=",
"[",
"]",
"for",
"glyphName",
"in",
"ttFont",
"[",
"'glyf'",
"]",
".",
"keys",
"(",
")",
":",
"glyph",
"=",
"ttFont",
"[",
"'glyf'",
"]",
"[",
"glyphName",
"]",
"coords",
"=",
"glyph",
".",
"getCoordinates",
"(",
"ttFont",
"[",
"'glyf'",
"]",
")",
"[",
"0",
"]",
"for",
"x",
",",
"y",
"in",
"coords",
":",
"if",
"x",
"<",
"glyph",
".",
"xMin",
"or",
"x",
">",
"glyph",
".",
"xMax",
"or",
"y",
"<",
"glyph",
".",
"yMin",
"or",
"y",
">",
"glyph",
".",
"yMax",
"or",
"abs",
"(",
"x",
")",
">",
"32766",
"or",
"abs",
"(",
"y",
")",
">",
"32766",
":",
"failed",
"=",
"True",
"out_of_bounds",
".",
"append",
"(",
"(",
"glyphName",
",",
"x",
",",
"y",
")",
")",
"if",
"failed",
":",
"yield",
"WARN",
",",
"(",
"\"The following glyphs have coordinates which are\"",
"\" out of bounds:\\n{}\\nThis happens a lot when points\"",
"\" are not extremes, which is usually bad. However,\"",
"\" fixing this alert by adding points on extremes may\"",
"\" do more harm than good, especially with italics,\"",
"\" calligraphic-script, handwriting, rounded and\"",
"\" other fonts. So it is common to\"",
"\" ignore this message\"",
".",
"format",
"(",
"out_of_bounds",
")",
")",
"else",
":",
"yield",
"PASS",
",",
"\"All glyph paths have coordinates within bounds!\""
] | Check for points out of bounds. | [
"Check",
"for",
"points",
"out",
"of",
"bounds",
"."
] | b355aea2e619a4477769e060d24c32448aa65399 | https://github.com/googlefonts/fontbakery/blob/b355aea2e619a4477769e060d24c32448aa65399/Lib/fontbakery/profiles/glyf.py#L51-L75 | train |
googlefonts/fontbakery | Lib/fontbakery/profiles/ufo_sources.py | com_daltonmaag_check_ufolint | def com_daltonmaag_check_ufolint(font):
"""Run ufolint on UFO source directory."""
import subprocess
ufolint_cmd = ["ufolint", font]
try:
subprocess.check_output(ufolint_cmd, stderr=subprocess.STDOUT)
except subprocess.CalledProcessError as e:
yield FAIL, ("ufolint failed the UFO source. Output follows :"
"\n\n{}\n").format(e.output.decode())
except OSError:
yield ERROR, "ufolint is not available!"
else:
yield PASS, "ufolint passed the UFO source." | python | def com_daltonmaag_check_ufolint(font):
"""Run ufolint on UFO source directory."""
import subprocess
ufolint_cmd = ["ufolint", font]
try:
subprocess.check_output(ufolint_cmd, stderr=subprocess.STDOUT)
except subprocess.CalledProcessError as e:
yield FAIL, ("ufolint failed the UFO source. Output follows :"
"\n\n{}\n").format(e.output.decode())
except OSError:
yield ERROR, "ufolint is not available!"
else:
yield PASS, "ufolint passed the UFO source." | [
"def",
"com_daltonmaag_check_ufolint",
"(",
"font",
")",
":",
"import",
"subprocess",
"ufolint_cmd",
"=",
"[",
"\"ufolint\"",
",",
"font",
"]",
"try",
":",
"subprocess",
".",
"check_output",
"(",
"ufolint_cmd",
",",
"stderr",
"=",
"subprocess",
".",
"STDOUT",
")",
"except",
"subprocess",
".",
"CalledProcessError",
"as",
"e",
":",
"yield",
"FAIL",
",",
"(",
"\"ufolint failed the UFO source. Output follows :\"",
"\"\\n\\n{}\\n\"",
")",
".",
"format",
"(",
"e",
".",
"output",
".",
"decode",
"(",
")",
")",
"except",
"OSError",
":",
"yield",
"ERROR",
",",
"\"ufolint is not available!\"",
"else",
":",
"yield",
"PASS",
",",
"\"ufolint passed the UFO source.\""
] | Run ufolint on UFO source directory. | [
"Run",
"ufolint",
"on",
"UFO",
"source",
"directory",
"."
] | b355aea2e619a4477769e060d24c32448aa65399 | https://github.com/googlefonts/fontbakery/blob/b355aea2e619a4477769e060d24c32448aa65399/Lib/fontbakery/profiles/ufo_sources.py#L91-L104 | train |
googlefonts/fontbakery | Lib/fontbakery/profiles/ufo_sources.py | com_daltonmaag_check_required_fields | def com_daltonmaag_check_required_fields(ufo_font):
"""Check that required fields are present in the UFO fontinfo.
ufo2ft requires these info fields to compile a font binary:
unitsPerEm, ascender, descender, xHeight, capHeight and familyName.
"""
recommended_fields = []
for field in [
"unitsPerEm", "ascender", "descender", "xHeight", "capHeight",
"familyName"
]:
if ufo_font.info.__dict__.get("_" + field) is None:
recommended_fields.append(field)
if recommended_fields:
yield FAIL, f"Required field(s) missing: {recommended_fields}"
else:
yield PASS, "Required fields present." | python | def com_daltonmaag_check_required_fields(ufo_font):
"""Check that required fields are present in the UFO fontinfo.
ufo2ft requires these info fields to compile a font binary:
unitsPerEm, ascender, descender, xHeight, capHeight and familyName.
"""
recommended_fields = []
for field in [
"unitsPerEm", "ascender", "descender", "xHeight", "capHeight",
"familyName"
]:
if ufo_font.info.__dict__.get("_" + field) is None:
recommended_fields.append(field)
if recommended_fields:
yield FAIL, f"Required field(s) missing: {recommended_fields}"
else:
yield PASS, "Required fields present." | [
"def",
"com_daltonmaag_check_required_fields",
"(",
"ufo_font",
")",
":",
"recommended_fields",
"=",
"[",
"]",
"for",
"field",
"in",
"[",
"\"unitsPerEm\"",
",",
"\"ascender\"",
",",
"\"descender\"",
",",
"\"xHeight\"",
",",
"\"capHeight\"",
",",
"\"familyName\"",
"]",
":",
"if",
"ufo_font",
".",
"info",
".",
"__dict__",
".",
"get",
"(",
"\"_\"",
"+",
"field",
")",
"is",
"None",
":",
"recommended_fields",
".",
"append",
"(",
"field",
")",
"if",
"recommended_fields",
":",
"yield",
"FAIL",
",",
"f\"Required field(s) missing: {recommended_fields}\"",
"else",
":",
"yield",
"PASS",
",",
"\"Required fields present.\""
] | Check that required fields are present in the UFO fontinfo.
ufo2ft requires these info fields to compile a font binary:
unitsPerEm, ascender, descender, xHeight, capHeight and familyName. | [
"Check",
"that",
"required",
"fields",
"are",
"present",
"in",
"the",
"UFO",
"fontinfo",
"."
] | b355aea2e619a4477769e060d24c32448aa65399 | https://github.com/googlefonts/fontbakery/blob/b355aea2e619a4477769e060d24c32448aa65399/Lib/fontbakery/profiles/ufo_sources.py#L111-L129 | train |
googlefonts/fontbakery | Lib/fontbakery/profiles/ufo_sources.py | com_daltonmaag_check_recommended_fields | def com_daltonmaag_check_recommended_fields(ufo_font):
"""Check that recommended fields are present in the UFO fontinfo.
This includes fields that should be in any production font.
"""
recommended_fields = []
for field in [
"postscriptUnderlineThickness", "postscriptUnderlinePosition",
"versionMajor", "versionMinor", "styleName", "copyright",
"openTypeOS2Panose"
]:
if ufo_font.info.__dict__.get("_" + field) is None:
recommended_fields.append(field)
if recommended_fields:
yield WARN, f"Recommended field(s) missing: {recommended_fields}"
else:
yield PASS, "Recommended fields present." | python | def com_daltonmaag_check_recommended_fields(ufo_font):
"""Check that recommended fields are present in the UFO fontinfo.
This includes fields that should be in any production font.
"""
recommended_fields = []
for field in [
"postscriptUnderlineThickness", "postscriptUnderlinePosition",
"versionMajor", "versionMinor", "styleName", "copyright",
"openTypeOS2Panose"
]:
if ufo_font.info.__dict__.get("_" + field) is None:
recommended_fields.append(field)
if recommended_fields:
yield WARN, f"Recommended field(s) missing: {recommended_fields}"
else:
yield PASS, "Recommended fields present." | [
"def",
"com_daltonmaag_check_recommended_fields",
"(",
"ufo_font",
")",
":",
"recommended_fields",
"=",
"[",
"]",
"for",
"field",
"in",
"[",
"\"postscriptUnderlineThickness\"",
",",
"\"postscriptUnderlinePosition\"",
",",
"\"versionMajor\"",
",",
"\"versionMinor\"",
",",
"\"styleName\"",
",",
"\"copyright\"",
",",
"\"openTypeOS2Panose\"",
"]",
":",
"if",
"ufo_font",
".",
"info",
".",
"__dict__",
".",
"get",
"(",
"\"_\"",
"+",
"field",
")",
"is",
"None",
":",
"recommended_fields",
".",
"append",
"(",
"field",
")",
"if",
"recommended_fields",
":",
"yield",
"WARN",
",",
"f\"Recommended field(s) missing: {recommended_fields}\"",
"else",
":",
"yield",
"PASS",
",",
"\"Recommended fields present.\""
] | Check that recommended fields are present in the UFO fontinfo.
This includes fields that should be in any production font. | [
"Check",
"that",
"recommended",
"fields",
"are",
"present",
"in",
"the",
"UFO",
"fontinfo",
"."
] | b355aea2e619a4477769e060d24c32448aa65399 | https://github.com/googlefonts/fontbakery/blob/b355aea2e619a4477769e060d24c32448aa65399/Lib/fontbakery/profiles/ufo_sources.py#L136-L154 | train |
googlefonts/fontbakery | Lib/fontbakery/profiles/ufo_sources.py | com_daltonmaag_check_unnecessary_fields | def com_daltonmaag_check_unnecessary_fields(ufo_font):
"""Check that no unnecessary fields are present in the UFO fontinfo.
ufo2ft will generate these.
openTypeOS2UnicodeRanges and openTypeOS2CodePageRanges are exempted
because it is useful to toggle a range when not _all_ the glyphs in that
region are present.
year is deprecated since UFO v2.
"""
unnecessary_fields = []
for field in [
"openTypeNameUniqueID", "openTypeNameVersion", "postscriptUniqueID",
"year"
]:
if ufo_font.info.__dict__.get("_" + field) is not None:
unnecessary_fields.append(field)
if unnecessary_fields:
yield WARN, f"Unnecessary field(s) present: {unnecessary_fields}"
else:
yield PASS, "Unnecessary fields omitted." | python | def com_daltonmaag_check_unnecessary_fields(ufo_font):
"""Check that no unnecessary fields are present in the UFO fontinfo.
ufo2ft will generate these.
openTypeOS2UnicodeRanges and openTypeOS2CodePageRanges are exempted
because it is useful to toggle a range when not _all_ the glyphs in that
region are present.
year is deprecated since UFO v2.
"""
unnecessary_fields = []
for field in [
"openTypeNameUniqueID", "openTypeNameVersion", "postscriptUniqueID",
"year"
]:
if ufo_font.info.__dict__.get("_" + field) is not None:
unnecessary_fields.append(field)
if unnecessary_fields:
yield WARN, f"Unnecessary field(s) present: {unnecessary_fields}"
else:
yield PASS, "Unnecessary fields omitted." | [
"def",
"com_daltonmaag_check_unnecessary_fields",
"(",
"ufo_font",
")",
":",
"unnecessary_fields",
"=",
"[",
"]",
"for",
"field",
"in",
"[",
"\"openTypeNameUniqueID\"",
",",
"\"openTypeNameVersion\"",
",",
"\"postscriptUniqueID\"",
",",
"\"year\"",
"]",
":",
"if",
"ufo_font",
".",
"info",
".",
"__dict__",
".",
"get",
"(",
"\"_\"",
"+",
"field",
")",
"is",
"not",
"None",
":",
"unnecessary_fields",
".",
"append",
"(",
"field",
")",
"if",
"unnecessary_fields",
":",
"yield",
"WARN",
",",
"f\"Unnecessary field(s) present: {unnecessary_fields}\"",
"else",
":",
"yield",
"PASS",
",",
"\"Unnecessary fields omitted.\""
] | Check that no unnecessary fields are present in the UFO fontinfo.
ufo2ft will generate these.
openTypeOS2UnicodeRanges and openTypeOS2CodePageRanges are exempted
because it is useful to toggle a range when not _all_ the glyphs in that
region are present.
year is deprecated since UFO v2. | [
"Check",
"that",
"no",
"unnecessary",
"fields",
"are",
"present",
"in",
"the",
"UFO",
"fontinfo",
"."
] | b355aea2e619a4477769e060d24c32448aa65399 | https://github.com/googlefonts/fontbakery/blob/b355aea2e619a4477769e060d24c32448aa65399/Lib/fontbakery/profiles/ufo_sources.py#L161-L184 | train |
googlefonts/fontbakery | Lib/fontbakery/profiles/ufo_sources.py | UFOProfile.setup_argparse | def setup_argparse(self, argument_parser):
"""Set up custom arguments needed for this profile."""
import glob
import logging
import argparse
def get_fonts(pattern):
fonts_to_check = []
# use glob.glob to accept *.ufo
for fullpath in glob.glob(pattern):
fullpath_absolute = os.path.abspath(fullpath)
if fullpath_absolute.lower().endswith(".ufo") and os.path.isdir(
fullpath_absolute):
fonts_to_check.append(fullpath)
else:
logging.warning(
("Skipping '{}' as it does not seem "
"to be valid UFO source directory.").format(fullpath))
return fonts_to_check
class MergeAction(argparse.Action):
def __call__(self, parser, namespace, values, option_string=None):
target = [item for l in values for item in l]
setattr(namespace, self.dest, target)
argument_parser.add_argument(
'fonts',
# To allow optional commands like "-L" to work without other input
# files:
nargs='*',
type=get_fonts,
action=MergeAction,
help='font file path(s) to check.'
' Wildcards like *.ufo are allowed.')
return ('fonts',) | python | def setup_argparse(self, argument_parser):
"""Set up custom arguments needed for this profile."""
import glob
import logging
import argparse
def get_fonts(pattern):
fonts_to_check = []
# use glob.glob to accept *.ufo
for fullpath in glob.glob(pattern):
fullpath_absolute = os.path.abspath(fullpath)
if fullpath_absolute.lower().endswith(".ufo") and os.path.isdir(
fullpath_absolute):
fonts_to_check.append(fullpath)
else:
logging.warning(
("Skipping '{}' as it does not seem "
"to be valid UFO source directory.").format(fullpath))
return fonts_to_check
class MergeAction(argparse.Action):
def __call__(self, parser, namespace, values, option_string=None):
target = [item for l in values for item in l]
setattr(namespace, self.dest, target)
argument_parser.add_argument(
'fonts',
# To allow optional commands like "-L" to work without other input
# files:
nargs='*',
type=get_fonts,
action=MergeAction,
help='font file path(s) to check.'
' Wildcards like *.ufo are allowed.')
return ('fonts',) | [
"def",
"setup_argparse",
"(",
"self",
",",
"argument_parser",
")",
":",
"import",
"glob",
"import",
"logging",
"import",
"argparse",
"def",
"get_fonts",
"(",
"pattern",
")",
":",
"fonts_to_check",
"=",
"[",
"]",
"# use glob.glob to accept *.ufo",
"for",
"fullpath",
"in",
"glob",
".",
"glob",
"(",
"pattern",
")",
":",
"fullpath_absolute",
"=",
"os",
".",
"path",
".",
"abspath",
"(",
"fullpath",
")",
"if",
"fullpath_absolute",
".",
"lower",
"(",
")",
".",
"endswith",
"(",
"\".ufo\"",
")",
"and",
"os",
".",
"path",
".",
"isdir",
"(",
"fullpath_absolute",
")",
":",
"fonts_to_check",
".",
"append",
"(",
"fullpath",
")",
"else",
":",
"logging",
".",
"warning",
"(",
"(",
"\"Skipping '{}' as it does not seem \"",
"\"to be valid UFO source directory.\"",
")",
".",
"format",
"(",
"fullpath",
")",
")",
"return",
"fonts_to_check",
"class",
"MergeAction",
"(",
"argparse",
".",
"Action",
")",
":",
"def",
"__call__",
"(",
"self",
",",
"parser",
",",
"namespace",
",",
"values",
",",
"option_string",
"=",
"None",
")",
":",
"target",
"=",
"[",
"item",
"for",
"l",
"in",
"values",
"for",
"item",
"in",
"l",
"]",
"setattr",
"(",
"namespace",
",",
"self",
".",
"dest",
",",
"target",
")",
"argument_parser",
".",
"add_argument",
"(",
"'fonts'",
",",
"# To allow optional commands like \"-L\" to work without other input",
"# files:",
"nargs",
"=",
"'*'",
",",
"type",
"=",
"get_fonts",
",",
"action",
"=",
"MergeAction",
",",
"help",
"=",
"'font file path(s) to check.'",
"' Wildcards like *.ufo are allowed.'",
")",
"return",
"(",
"'fonts'",
",",
")"
] | Set up custom arguments needed for this profile. | [
"Set",
"up",
"custom",
"arguments",
"needed",
"for",
"this",
"profile",
"."
] | b355aea2e619a4477769e060d24c32448aa65399 | https://github.com/googlefonts/fontbakery/blob/b355aea2e619a4477769e060d24c32448aa65399/Lib/fontbakery/profiles/ufo_sources.py#L11-L49 | train |
googlefonts/fontbakery | Lib/fontbakery/profiles/hmtx.py | com_google_fonts_check_whitespace_widths | def com_google_fonts_check_whitespace_widths(ttFont):
"""Whitespace and non-breaking space have the same width?"""
from fontbakery.utils import get_glyph_name
space_name = get_glyph_name(ttFont, 0x0020)
nbsp_name = get_glyph_name(ttFont, 0x00A0)
space_width = ttFont['hmtx'][space_name][0]
nbsp_width = ttFont['hmtx'][nbsp_name][0]
if space_width > 0 and space_width == nbsp_width:
yield PASS, "Whitespace and non-breaking space have the same width."
else:
yield FAIL, ("Whitespace and non-breaking space have differing width:"
" Whitespace ({}) is {} font units wide, non-breaking space"
" ({}) is {} font units wide. Both should be positive and the"
" same.").format(space_name, space_width, nbsp_name,
nbsp_width) | python | def com_google_fonts_check_whitespace_widths(ttFont):
"""Whitespace and non-breaking space have the same width?"""
from fontbakery.utils import get_glyph_name
space_name = get_glyph_name(ttFont, 0x0020)
nbsp_name = get_glyph_name(ttFont, 0x00A0)
space_width = ttFont['hmtx'][space_name][0]
nbsp_width = ttFont['hmtx'][nbsp_name][0]
if space_width > 0 and space_width == nbsp_width:
yield PASS, "Whitespace and non-breaking space have the same width."
else:
yield FAIL, ("Whitespace and non-breaking space have differing width:"
" Whitespace ({}) is {} font units wide, non-breaking space"
" ({}) is {} font units wide. Both should be positive and the"
" same.").format(space_name, space_width, nbsp_name,
nbsp_width) | [
"def",
"com_google_fonts_check_whitespace_widths",
"(",
"ttFont",
")",
":",
"from",
"fontbakery",
".",
"utils",
"import",
"get_glyph_name",
"space_name",
"=",
"get_glyph_name",
"(",
"ttFont",
",",
"0x0020",
")",
"nbsp_name",
"=",
"get_glyph_name",
"(",
"ttFont",
",",
"0x00A0",
")",
"space_width",
"=",
"ttFont",
"[",
"'hmtx'",
"]",
"[",
"space_name",
"]",
"[",
"0",
"]",
"nbsp_width",
"=",
"ttFont",
"[",
"'hmtx'",
"]",
"[",
"nbsp_name",
"]",
"[",
"0",
"]",
"if",
"space_width",
">",
"0",
"and",
"space_width",
"==",
"nbsp_width",
":",
"yield",
"PASS",
",",
"\"Whitespace and non-breaking space have the same width.\"",
"else",
":",
"yield",
"FAIL",
",",
"(",
"\"Whitespace and non-breaking space have differing width:\"",
"\" Whitespace ({}) is {} font units wide, non-breaking space\"",
"\" ({}) is {} font units wide. Both should be positive and the\"",
"\" same.\"",
")",
".",
"format",
"(",
"space_name",
",",
"space_width",
",",
"nbsp_name",
",",
"nbsp_width",
")"
] | Whitespace and non-breaking space have the same width? | [
"Whitespace",
"and",
"non",
"-",
"breaking",
"space",
"have",
"the",
"same",
"width?"
] | b355aea2e619a4477769e060d24c32448aa65399 | https://github.com/googlefonts/fontbakery/blob/b355aea2e619a4477769e060d24c32448aa65399/Lib/fontbakery/profiles/hmtx.py#L13-L30 | train |
Kuniwak/vint | vint/linting/policy_set.py | PolicySet.update_by_config | def update_by_config(self, config_dict):
""" Update policies set by the config dictionary.
Expect the policy_enabling_map structure to be (represented by YAML):
- PolicyFoo:
enabled: True
- PolicyBar:
enabled: False
additional_field: 'is_ok'
"""
policy_enabling_map = self._get_enabling_map(config_dict)
self.enabled_policies = []
for policy_name, is_policy_enabled in policy_enabling_map.items():
if not self._is_policy_exists(policy_name):
self._warn_unexistent_policy(policy_name)
continue
if is_policy_enabled:
enabled_policy = self._get_policy(policy_name)
self.enabled_policies.append(enabled_policy) | python | def update_by_config(self, config_dict):
""" Update policies set by the config dictionary.
Expect the policy_enabling_map structure to be (represented by YAML):
- PolicyFoo:
enabled: True
- PolicyBar:
enabled: False
additional_field: 'is_ok'
"""
policy_enabling_map = self._get_enabling_map(config_dict)
self.enabled_policies = []
for policy_name, is_policy_enabled in policy_enabling_map.items():
if not self._is_policy_exists(policy_name):
self._warn_unexistent_policy(policy_name)
continue
if is_policy_enabled:
enabled_policy = self._get_policy(policy_name)
self.enabled_policies.append(enabled_policy) | [
"def",
"update_by_config",
"(",
"self",
",",
"config_dict",
")",
":",
"policy_enabling_map",
"=",
"self",
".",
"_get_enabling_map",
"(",
"config_dict",
")",
"self",
".",
"enabled_policies",
"=",
"[",
"]",
"for",
"policy_name",
",",
"is_policy_enabled",
"in",
"policy_enabling_map",
".",
"items",
"(",
")",
":",
"if",
"not",
"self",
".",
"_is_policy_exists",
"(",
"policy_name",
")",
":",
"self",
".",
"_warn_unexistent_policy",
"(",
"policy_name",
")",
"continue",
"if",
"is_policy_enabled",
":",
"enabled_policy",
"=",
"self",
".",
"_get_policy",
"(",
"policy_name",
")",
"self",
".",
"enabled_policies",
".",
"append",
"(",
"enabled_policy",
")"
] | Update policies set by the config dictionary.
Expect the policy_enabling_map structure to be (represented by YAML):
- PolicyFoo:
enabled: True
- PolicyBar:
enabled: False
additional_field: 'is_ok' | [
"Update",
"policies",
"set",
"by",
"the",
"config",
"dictionary",
"."
] | db29337d859d88239c282c2e9d84c858f23a4a09 | https://github.com/Kuniwak/vint/blob/db29337d859d88239c282c2e9d84c858f23a4a09/vint/linting/policy_set.py#L48-L68 | train |
Kuniwak/vint | vint/linting/cli.py | _build_cmdargs | def _build_cmdargs(argv):
""" Build command line arguments dict to use;
- displaying usages
- vint.linting.env.build_environment
This method take an argv parameter to make function pure.
"""
parser = _build_arg_parser()
namespace = parser.parse_args(argv[1:])
cmdargs = vars(namespace)
return cmdargs | python | def _build_cmdargs(argv):
""" Build command line arguments dict to use;
- displaying usages
- vint.linting.env.build_environment
This method take an argv parameter to make function pure.
"""
parser = _build_arg_parser()
namespace = parser.parse_args(argv[1:])
cmdargs = vars(namespace)
return cmdargs | [
"def",
"_build_cmdargs",
"(",
"argv",
")",
":",
"parser",
"=",
"_build_arg_parser",
"(",
")",
"namespace",
"=",
"parser",
".",
"parse_args",
"(",
"argv",
"[",
"1",
":",
"]",
")",
"cmdargs",
"=",
"vars",
"(",
"namespace",
")",
"return",
"cmdargs"
] | Build command line arguments dict to use;
- displaying usages
- vint.linting.env.build_environment
This method take an argv parameter to make function pure. | [
"Build",
"command",
"line",
"arguments",
"dict",
"to",
"use",
";",
"-",
"displaying",
"usages",
"-",
"vint",
".",
"linting",
".",
"env",
".",
"build_environment"
] | db29337d859d88239c282c2e9d84c858f23a4a09 | https://github.com/Kuniwak/vint/blob/db29337d859d88239c282c2e9d84c858f23a4a09/vint/linting/cli.py#L79-L90 | train |
Kuniwak/vint | vint/ast/parsing.py | Parser.parse | def parse(self, lint_target): # type: (AbstractLintTarget) -> Dict[str, Any]
""" Parse vim script file and return the AST. """
decoder = Decoder(default_decoding_strategy)
decoded = decoder.decode(lint_target.read())
decoded_and_lf_normalized = decoded.replace('\r\n', '\n')
return self.parse_string(decoded_and_lf_normalized) | python | def parse(self, lint_target): # type: (AbstractLintTarget) -> Dict[str, Any]
""" Parse vim script file and return the AST. """
decoder = Decoder(default_decoding_strategy)
decoded = decoder.decode(lint_target.read())
decoded_and_lf_normalized = decoded.replace('\r\n', '\n')
return self.parse_string(decoded_and_lf_normalized) | [
"def",
"parse",
"(",
"self",
",",
"lint_target",
")",
":",
"# type: (AbstractLintTarget) -> Dict[str, Any]",
"decoder",
"=",
"Decoder",
"(",
"default_decoding_strategy",
")",
"decoded",
"=",
"decoder",
".",
"decode",
"(",
"lint_target",
".",
"read",
"(",
")",
")",
"decoded_and_lf_normalized",
"=",
"decoded",
".",
"replace",
"(",
"'\\r\\n'",
",",
"'\\n'",
")",
"return",
"self",
".",
"parse_string",
"(",
"decoded_and_lf_normalized",
")"
] | Parse vim script file and return the AST. | [
"Parse",
"vim",
"script",
"file",
"and",
"return",
"the",
"AST",
"."
] | db29337d859d88239c282c2e9d84c858f23a4a09 | https://github.com/Kuniwak/vint/blob/db29337d859d88239c282c2e9d84c858f23a4a09/vint/ast/parsing.py#L19-L25 | train |
Kuniwak/vint | vint/ast/parsing.py | Parser.parse_string | def parse_string(self, string): # type: (str) -> Dict[str, Any]
""" Parse vim script string and return the AST. """
lines = string.split('\n')
reader = vimlparser.StringReader(lines)
parser = vimlparser.VimLParser(self._enable_neovim)
ast = parser.parse(reader)
# TOPLEVEL does not have a pos, but we need pos for all nodes
ast['pos'] = {'col': 1, 'i': 0, 'lnum': 1}
for plugin in self.plugins:
plugin.process(ast)
return ast | python | def parse_string(self, string): # type: (str) -> Dict[str, Any]
""" Parse vim script string and return the AST. """
lines = string.split('\n')
reader = vimlparser.StringReader(lines)
parser = vimlparser.VimLParser(self._enable_neovim)
ast = parser.parse(reader)
# TOPLEVEL does not have a pos, but we need pos for all nodes
ast['pos'] = {'col': 1, 'i': 0, 'lnum': 1}
for plugin in self.plugins:
plugin.process(ast)
return ast | [
"def",
"parse_string",
"(",
"self",
",",
"string",
")",
":",
"# type: (str) -> Dict[str, Any]",
"lines",
"=",
"string",
".",
"split",
"(",
"'\\n'",
")",
"reader",
"=",
"vimlparser",
".",
"StringReader",
"(",
"lines",
")",
"parser",
"=",
"vimlparser",
".",
"VimLParser",
"(",
"self",
".",
"_enable_neovim",
")",
"ast",
"=",
"parser",
".",
"parse",
"(",
"reader",
")",
"# TOPLEVEL does not have a pos, but we need pos for all nodes",
"ast",
"[",
"'pos'",
"]",
"=",
"{",
"'col'",
":",
"1",
",",
"'i'",
":",
"0",
",",
"'lnum'",
":",
"1",
"}",
"for",
"plugin",
"in",
"self",
".",
"plugins",
":",
"plugin",
".",
"process",
"(",
"ast",
")",
"return",
"ast"
] | Parse vim script string and return the AST. | [
"Parse",
"vim",
"script",
"string",
"and",
"return",
"the",
"AST",
"."
] | db29337d859d88239c282c2e9d84c858f23a4a09 | https://github.com/Kuniwak/vint/blob/db29337d859d88239c282c2e9d84c858f23a4a09/vint/ast/parsing.py#L28-L42 | train |
Kuniwak/vint | vint/ast/parsing.py | Parser.parse_string_expr | def parse_string_expr(self, string_expr_node):
""" Parse a string node content. """
string_expr_node_value = string_expr_node['value']
string_expr_str = string_expr_node_value[1:-1]
# Care escaped string literals
if string_expr_node_value[0] == "'":
string_expr_str = string_expr_str.replace("''", "'")
else:
string_expr_str = string_expr_str.replace('\\"', '"')
# NOTE: This is a hack to parse expr1. See :help expr1
raw_ast = self.parse_string('echo ' + string_expr_str)
# We need the left node of ECHO node
parsed_string_expr_nodes = raw_ast['body'][0]['list']
start_pos = string_expr_node['pos']
def adjust_position(node):
pos = node['pos']
# Care 1-based index and the length of "echo ".
pos['col'] += start_pos['col'] - 1 - 5
# Care the length of "echo ".
pos['i'] += start_pos['i'] - 5
# Care 1-based index
pos['lnum'] += start_pos['lnum'] - 1
for parsed_string_expr_node in parsed_string_expr_nodes:
traverse(parsed_string_expr_node, on_enter=adjust_position)
return parsed_string_expr_nodes | python | def parse_string_expr(self, string_expr_node):
""" Parse a string node content. """
string_expr_node_value = string_expr_node['value']
string_expr_str = string_expr_node_value[1:-1]
# Care escaped string literals
if string_expr_node_value[0] == "'":
string_expr_str = string_expr_str.replace("''", "'")
else:
string_expr_str = string_expr_str.replace('\\"', '"')
# NOTE: This is a hack to parse expr1. See :help expr1
raw_ast = self.parse_string('echo ' + string_expr_str)
# We need the left node of ECHO node
parsed_string_expr_nodes = raw_ast['body'][0]['list']
start_pos = string_expr_node['pos']
def adjust_position(node):
pos = node['pos']
# Care 1-based index and the length of "echo ".
pos['col'] += start_pos['col'] - 1 - 5
# Care the length of "echo ".
pos['i'] += start_pos['i'] - 5
# Care 1-based index
pos['lnum'] += start_pos['lnum'] - 1
for parsed_string_expr_node in parsed_string_expr_nodes:
traverse(parsed_string_expr_node, on_enter=adjust_position)
return parsed_string_expr_nodes | [
"def",
"parse_string_expr",
"(",
"self",
",",
"string_expr_node",
")",
":",
"string_expr_node_value",
"=",
"string_expr_node",
"[",
"'value'",
"]",
"string_expr_str",
"=",
"string_expr_node_value",
"[",
"1",
":",
"-",
"1",
"]",
"# Care escaped string literals",
"if",
"string_expr_node_value",
"[",
"0",
"]",
"==",
"\"'\"",
":",
"string_expr_str",
"=",
"string_expr_str",
".",
"replace",
"(",
"\"''\"",
",",
"\"'\"",
")",
"else",
":",
"string_expr_str",
"=",
"string_expr_str",
".",
"replace",
"(",
"'\\\\\"'",
",",
"'\"'",
")",
"# NOTE: This is a hack to parse expr1. See :help expr1",
"raw_ast",
"=",
"self",
".",
"parse_string",
"(",
"'echo '",
"+",
"string_expr_str",
")",
"# We need the left node of ECHO node",
"parsed_string_expr_nodes",
"=",
"raw_ast",
"[",
"'body'",
"]",
"[",
"0",
"]",
"[",
"'list'",
"]",
"start_pos",
"=",
"string_expr_node",
"[",
"'pos'",
"]",
"def",
"adjust_position",
"(",
"node",
")",
":",
"pos",
"=",
"node",
"[",
"'pos'",
"]",
"# Care 1-based index and the length of \"echo \".",
"pos",
"[",
"'col'",
"]",
"+=",
"start_pos",
"[",
"'col'",
"]",
"-",
"1",
"-",
"5",
"# Care the length of \"echo \".",
"pos",
"[",
"'i'",
"]",
"+=",
"start_pos",
"[",
"'i'",
"]",
"-",
"5",
"# Care 1-based index",
"pos",
"[",
"'lnum'",
"]",
"+=",
"start_pos",
"[",
"'lnum'",
"]",
"-",
"1",
"for",
"parsed_string_expr_node",
"in",
"parsed_string_expr_nodes",
":",
"traverse",
"(",
"parsed_string_expr_node",
",",
"on_enter",
"=",
"adjust_position",
")",
"return",
"parsed_string_expr_nodes"
] | Parse a string node content. | [
"Parse",
"a",
"string",
"node",
"content",
"."
] | db29337d859d88239c282c2e9d84c858f23a4a09 | https://github.com/Kuniwak/vint/blob/db29337d859d88239c282c2e9d84c858f23a4a09/vint/ast/parsing.py#L87-L121 | train |
Kuniwak/vint | vint/ast/plugin/scope_plugin/scope_detector.py | is_builtin_variable | def is_builtin_variable(id_node): # type: (Dict[str, Any]) -> bool
""" Whether the specified node is a builtin identifier. """
# Builtin variables are always IDENTIFIER.
if NodeType(id_node['type']) is not NodeType.IDENTIFIER:
return False
id_value = id_node['value']
if id_value.startswith('v:'):
# It is an explicit builtin variable such as: "v:count", "v:char"
# TODO: Add unknown builtin flag
return True
if is_builtin_function(id_node):
return True
if id_value in ['key', 'val']:
# These builtin variable names are available on only map() or filter().
return is_on_lambda_string_context(id_node)
# It is an implicit builtin variable such as: "count", "char"
return id_value in BuiltinVariablesCanHaveImplicitScope | python | def is_builtin_variable(id_node): # type: (Dict[str, Any]) -> bool
""" Whether the specified node is a builtin identifier. """
# Builtin variables are always IDENTIFIER.
if NodeType(id_node['type']) is not NodeType.IDENTIFIER:
return False
id_value = id_node['value']
if id_value.startswith('v:'):
# It is an explicit builtin variable such as: "v:count", "v:char"
# TODO: Add unknown builtin flag
return True
if is_builtin_function(id_node):
return True
if id_value in ['key', 'val']:
# These builtin variable names are available on only map() or filter().
return is_on_lambda_string_context(id_node)
# It is an implicit builtin variable such as: "count", "char"
return id_value in BuiltinVariablesCanHaveImplicitScope | [
"def",
"is_builtin_variable",
"(",
"id_node",
")",
":",
"# type: (Dict[str, Any]) -> bool",
"# Builtin variables are always IDENTIFIER.",
"if",
"NodeType",
"(",
"id_node",
"[",
"'type'",
"]",
")",
"is",
"not",
"NodeType",
".",
"IDENTIFIER",
":",
"return",
"False",
"id_value",
"=",
"id_node",
"[",
"'value'",
"]",
"if",
"id_value",
".",
"startswith",
"(",
"'v:'",
")",
":",
"# It is an explicit builtin variable such as: \"v:count\", \"v:char\"",
"# TODO: Add unknown builtin flag",
"return",
"True",
"if",
"is_builtin_function",
"(",
"id_node",
")",
":",
"return",
"True",
"if",
"id_value",
"in",
"[",
"'key'",
",",
"'val'",
"]",
":",
"# These builtin variable names are available on only map() or filter().",
"return",
"is_on_lambda_string_context",
"(",
"id_node",
")",
"# It is an implicit builtin variable such as: \"count\", \"char\"",
"return",
"id_value",
"in",
"BuiltinVariablesCanHaveImplicitScope"
] | Whether the specified node is a builtin identifier. | [
"Whether",
"the",
"specified",
"node",
"is",
"a",
"builtin",
"identifier",
"."
] | db29337d859d88239c282c2e9d84c858f23a4a09 | https://github.com/Kuniwak/vint/blob/db29337d859d88239c282c2e9d84c858f23a4a09/vint/ast/plugin/scope_plugin/scope_detector.py#L69-L90 | train |
Kuniwak/vint | vint/ast/plugin/scope_plugin/scope_detector.py | is_builtin_function | def is_builtin_function(id_node): # type: (Dict[str, Any]) -> bool
""" Whether the specified node is a builtin function name identifier.
The given identifier should be a child node of NodeType.CALL.
"""
# Builtin functions are always IDENTIFIER.
if NodeType(id_node['type']) is not NodeType.IDENTIFIER:
return False
id_value = id_node['value']
if not is_function_identifier(id_node):
return False
# There are difference between a function identifier and variable
# identifier:
#
# let localtime = 0
# echo localtime " => 0
# echo localtime() " => 1420011455
return id_value in BuiltinFunctions | python | def is_builtin_function(id_node): # type: (Dict[str, Any]) -> bool
""" Whether the specified node is a builtin function name identifier.
The given identifier should be a child node of NodeType.CALL.
"""
# Builtin functions are always IDENTIFIER.
if NodeType(id_node['type']) is not NodeType.IDENTIFIER:
return False
id_value = id_node['value']
if not is_function_identifier(id_node):
return False
# There are difference between a function identifier and variable
# identifier:
#
# let localtime = 0
# echo localtime " => 0
# echo localtime() " => 1420011455
return id_value in BuiltinFunctions | [
"def",
"is_builtin_function",
"(",
"id_node",
")",
":",
"# type: (Dict[str, Any]) -> bool",
"# Builtin functions are always IDENTIFIER.",
"if",
"NodeType",
"(",
"id_node",
"[",
"'type'",
"]",
")",
"is",
"not",
"NodeType",
".",
"IDENTIFIER",
":",
"return",
"False",
"id_value",
"=",
"id_node",
"[",
"'value'",
"]",
"if",
"not",
"is_function_identifier",
"(",
"id_node",
")",
":",
"return",
"False",
"# There are difference between a function identifier and variable",
"# identifier:",
"#",
"# let localtime = 0",
"# echo localtime \" => 0",
"# echo localtime() \" => 1420011455",
"return",
"id_value",
"in",
"BuiltinFunctions"
] | Whether the specified node is a builtin function name identifier.
The given identifier should be a child node of NodeType.CALL. | [
"Whether",
"the",
"specified",
"node",
"is",
"a",
"builtin",
"function",
"name",
"identifier",
".",
"The",
"given",
"identifier",
"should",
"be",
"a",
"child",
"node",
"of",
"NodeType",
".",
"CALL",
"."
] | db29337d859d88239c282c2e9d84c858f23a4a09 | https://github.com/Kuniwak/vint/blob/db29337d859d88239c282c2e9d84c858f23a4a09/vint/ast/plugin/scope_plugin/scope_detector.py#L93-L112 | train |
Kuniwak/vint | vint/ast/plugin/scope_plugin/identifier_classifier.py | IdentifierClassifier.attach_identifier_attributes | def attach_identifier_attributes(self, ast): # type: (Dict[str, Any]) -> Dict[str, Any]
""" Attach 5 flags to the AST.
- is dynamic: True if the identifier name can be determined by static analysis.
- is member: True if the identifier is a member of a subscription/dot/slice node.
- is declaring: True if the identifier is used to declare.
- is autoload: True if the identifier is declared with autoload.
- is function: True if the identifier is a function. Vim distinguish
between function identifiers and variable identifiers.
- is declarative parameter: True if the identifier is a declarative
parameter. For example, the identifier "param" in Func(param) is a
declarative parameter.
- is on string expression context: True if the variable is on the
string expression context. The string expression context is the
string content on the 2nd argument of the map or filter function.
- is lambda argument: True if the identifier is a lambda argument.
"""
redir_assignment_parser = RedirAssignmentParser()
ast_with_parsed_redir = redir_assignment_parser.process(ast)
map_and_filter_parser = CallNodeParser()
ast_with_parse_map_and_filter_and_redir = \
map_and_filter_parser.process(ast_with_parsed_redir)
traverse(
ast_with_parse_map_and_filter_and_redir,
on_enter=lambda node: self._enter_handler(
node,
is_on_lambda_str=None,
is_on_lambda_body=None,
)
)
return ast | python | def attach_identifier_attributes(self, ast): # type: (Dict[str, Any]) -> Dict[str, Any]
""" Attach 5 flags to the AST.
- is dynamic: True if the identifier name can be determined by static analysis.
- is member: True if the identifier is a member of a subscription/dot/slice node.
- is declaring: True if the identifier is used to declare.
- is autoload: True if the identifier is declared with autoload.
- is function: True if the identifier is a function. Vim distinguish
between function identifiers and variable identifiers.
- is declarative parameter: True if the identifier is a declarative
parameter. For example, the identifier "param" in Func(param) is a
declarative parameter.
- is on string expression context: True if the variable is on the
string expression context. The string expression context is the
string content on the 2nd argument of the map or filter function.
- is lambda argument: True if the identifier is a lambda argument.
"""
redir_assignment_parser = RedirAssignmentParser()
ast_with_parsed_redir = redir_assignment_parser.process(ast)
map_and_filter_parser = CallNodeParser()
ast_with_parse_map_and_filter_and_redir = \
map_and_filter_parser.process(ast_with_parsed_redir)
traverse(
ast_with_parse_map_and_filter_and_redir,
on_enter=lambda node: self._enter_handler(
node,
is_on_lambda_str=None,
is_on_lambda_body=None,
)
)
return ast | [
"def",
"attach_identifier_attributes",
"(",
"self",
",",
"ast",
")",
":",
"# type: (Dict[str, Any]) -> Dict[str, Any]",
"redir_assignment_parser",
"=",
"RedirAssignmentParser",
"(",
")",
"ast_with_parsed_redir",
"=",
"redir_assignment_parser",
".",
"process",
"(",
"ast",
")",
"map_and_filter_parser",
"=",
"CallNodeParser",
"(",
")",
"ast_with_parse_map_and_filter_and_redir",
"=",
"map_and_filter_parser",
".",
"process",
"(",
"ast_with_parsed_redir",
")",
"traverse",
"(",
"ast_with_parse_map_and_filter_and_redir",
",",
"on_enter",
"=",
"lambda",
"node",
":",
"self",
".",
"_enter_handler",
"(",
"node",
",",
"is_on_lambda_str",
"=",
"None",
",",
"is_on_lambda_body",
"=",
"None",
",",
")",
")",
"return",
"ast"
] | Attach 5 flags to the AST.
- is dynamic: True if the identifier name can be determined by static analysis.
- is member: True if the identifier is a member of a subscription/dot/slice node.
- is declaring: True if the identifier is used to declare.
- is autoload: True if the identifier is declared with autoload.
- is function: True if the identifier is a function. Vim distinguish
between function identifiers and variable identifiers.
- is declarative parameter: True if the identifier is a declarative
parameter. For example, the identifier "param" in Func(param) is a
declarative parameter.
- is on string expression context: True if the variable is on the
string expression context. The string expression context is the
string content on the 2nd argument of the map or filter function.
- is lambda argument: True if the identifier is a lambda argument. | [
"Attach",
"5",
"flags",
"to",
"the",
"AST",
"."
] | db29337d859d88239c282c2e9d84c858f23a4a09 | https://github.com/Kuniwak/vint/blob/db29337d859d88239c282c2e9d84c858f23a4a09/vint/ast/plugin/scope_plugin/identifier_classifier.py#L118-L150 | train |
Kuniwak/vint | vint/linting/policy/abstract_policy.py | AbstractPolicy.create_violation_report | def create_violation_report(self, node, lint_context):
""" Returns a violation report for the node. """
return {
'name': self.name,
'level': self.level,
'description': self.description,
'reference': self.reference,
'position': {
'line': node['pos']['lnum'],
'column': node['pos']['col'],
'path': lint_context['lint_target'].path,
},
} | python | def create_violation_report(self, node, lint_context):
""" Returns a violation report for the node. """
return {
'name': self.name,
'level': self.level,
'description': self.description,
'reference': self.reference,
'position': {
'line': node['pos']['lnum'],
'column': node['pos']['col'],
'path': lint_context['lint_target'].path,
},
} | [
"def",
"create_violation_report",
"(",
"self",
",",
"node",
",",
"lint_context",
")",
":",
"return",
"{",
"'name'",
":",
"self",
".",
"name",
",",
"'level'",
":",
"self",
".",
"level",
",",
"'description'",
":",
"self",
".",
"description",
",",
"'reference'",
":",
"self",
".",
"reference",
",",
"'position'",
":",
"{",
"'line'",
":",
"node",
"[",
"'pos'",
"]",
"[",
"'lnum'",
"]",
",",
"'column'",
":",
"node",
"[",
"'pos'",
"]",
"[",
"'col'",
"]",
",",
"'path'",
":",
"lint_context",
"[",
"'lint_target'",
"]",
".",
"path",
",",
"}",
",",
"}"
] | Returns a violation report for the node. | [
"Returns",
"a",
"violation",
"report",
"for",
"the",
"node",
"."
] | db29337d859d88239c282c2e9d84c858f23a4a09 | https://github.com/Kuniwak/vint/blob/db29337d859d88239c282c2e9d84c858f23a4a09/vint/linting/policy/abstract_policy.py#L22-L34 | train |
Kuniwak/vint | vint/linting/policy/abstract_policy.py | AbstractPolicy.get_policy_config | def get_policy_config(self, lint_context):
"""
Returns a config of the concrete policy. For example, a config of ProhibitSomethingEvil is located on
config.policies.ProhibitSomethingEvil.
"""
policy_config = lint_context['config']\
.get('policies', {})\
.get(self.__class__.__name__, {})
return policy_config | python | def get_policy_config(self, lint_context):
"""
Returns a config of the concrete policy. For example, a config of ProhibitSomethingEvil is located on
config.policies.ProhibitSomethingEvil.
"""
policy_config = lint_context['config']\
.get('policies', {})\
.get(self.__class__.__name__, {})
return policy_config | [
"def",
"get_policy_config",
"(",
"self",
",",
"lint_context",
")",
":",
"policy_config",
"=",
"lint_context",
"[",
"'config'",
"]",
".",
"get",
"(",
"'policies'",
",",
"{",
"}",
")",
".",
"get",
"(",
"self",
".",
"__class__",
".",
"__name__",
",",
"{",
"}",
")",
"return",
"policy_config"
] | Returns a config of the concrete policy. For example, a config of ProhibitSomethingEvil is located on
config.policies.ProhibitSomethingEvil. | [
"Returns",
"a",
"config",
"of",
"the",
"concrete",
"policy",
".",
"For",
"example",
"a",
"config",
"of",
"ProhibitSomethingEvil",
"is",
"located",
"on",
"config",
".",
"policies",
".",
"ProhibitSomethingEvil",
"."
] | db29337d859d88239c282c2e9d84c858f23a4a09 | https://github.com/Kuniwak/vint/blob/db29337d859d88239c282c2e9d84c858f23a4a09/vint/linting/policy/abstract_policy.py#L37-L46 | train |
Kuniwak/vint | vint/linting/policy/abstract_policy.py | AbstractPolicy.get_violation_if_found | def get_violation_if_found(self, node, lint_context):
""" Returns a violation if the node is invalid. """
if self.is_valid(node, lint_context):
return None
return self.create_violation_report(node, lint_context) | python | def get_violation_if_found(self, node, lint_context):
""" Returns a violation if the node is invalid. """
if self.is_valid(node, lint_context):
return None
return self.create_violation_report(node, lint_context) | [
"def",
"get_violation_if_found",
"(",
"self",
",",
"node",
",",
"lint_context",
")",
":",
"if",
"self",
".",
"is_valid",
"(",
"node",
",",
"lint_context",
")",
":",
"return",
"None",
"return",
"self",
".",
"create_violation_report",
"(",
"node",
",",
"lint_context",
")"
] | Returns a violation if the node is invalid. | [
"Returns",
"a",
"violation",
"if",
"the",
"node",
"is",
"invalid",
"."
] | db29337d859d88239c282c2e9d84c858f23a4a09 | https://github.com/Kuniwak/vint/blob/db29337d859d88239c282c2e9d84c858f23a4a09/vint/linting/policy/abstract_policy.py#L49-L54 | train |
Kuniwak/vint | vint/bootstrap.py | import_all_policies | def import_all_policies():
""" Import all policies that were registered by vint.linting.policy_registry.
Dynamic policy importing is comprised of the 3 steps
1. Try to import all policy modules (then we can't know what policies exist)
2. In policy module, register itself by using vint.linting.policy_registry
3. After all policies registered by itself, we can get policy classes
"""
pkg_name = _get_policy_package_name_for_test()
pkg_path_list = pkg_name.split('.')
pkg_path = str(Path(_get_vint_root(), *pkg_path_list).resolve())
for _, module_name, is_pkg in pkgutil.iter_modules([pkg_path]):
if not is_pkg:
module_fqn = pkg_name + '.' + module_name
logging.debug('Loading the policy module: `{fqn}`'.format(fqn=module_fqn))
importlib.import_module(module_fqn) | python | def import_all_policies():
""" Import all policies that were registered by vint.linting.policy_registry.
Dynamic policy importing is comprised of the 3 steps
1. Try to import all policy modules (then we can't know what policies exist)
2. In policy module, register itself by using vint.linting.policy_registry
3. After all policies registered by itself, we can get policy classes
"""
pkg_name = _get_policy_package_name_for_test()
pkg_path_list = pkg_name.split('.')
pkg_path = str(Path(_get_vint_root(), *pkg_path_list).resolve())
for _, module_name, is_pkg in pkgutil.iter_modules([pkg_path]):
if not is_pkg:
module_fqn = pkg_name + '.' + module_name
logging.debug('Loading the policy module: `{fqn}`'.format(fqn=module_fqn))
importlib.import_module(module_fqn) | [
"def",
"import_all_policies",
"(",
")",
":",
"pkg_name",
"=",
"_get_policy_package_name_for_test",
"(",
")",
"pkg_path_list",
"=",
"pkg_name",
".",
"split",
"(",
"'.'",
")",
"pkg_path",
"=",
"str",
"(",
"Path",
"(",
"_get_vint_root",
"(",
")",
",",
"*",
"pkg_path_list",
")",
".",
"resolve",
"(",
")",
")",
"for",
"_",
",",
"module_name",
",",
"is_pkg",
"in",
"pkgutil",
".",
"iter_modules",
"(",
"[",
"pkg_path",
"]",
")",
":",
"if",
"not",
"is_pkg",
":",
"module_fqn",
"=",
"pkg_name",
"+",
"'.'",
"+",
"module_name",
"logging",
".",
"debug",
"(",
"'Loading the policy module: `{fqn}`'",
".",
"format",
"(",
"fqn",
"=",
"module_fqn",
")",
")",
"importlib",
".",
"import_module",
"(",
"module_fqn",
")"
] | Import all policies that were registered by vint.linting.policy_registry.
Dynamic policy importing is comprised of the 3 steps
1. Try to import all policy modules (then we can't know what policies exist)
2. In policy module, register itself by using vint.linting.policy_registry
3. After all policies registered by itself, we can get policy classes | [
"Import",
"all",
"policies",
"that",
"were",
"registered",
"by",
"vint",
".",
"linting",
".",
"policy_registry",
"."
] | db29337d859d88239c282c2e9d84c858f23a4a09 | https://github.com/Kuniwak/vint/blob/db29337d859d88239c282c2e9d84c858f23a4a09/vint/bootstrap.py#L24-L41 | train |
Kuniwak/vint | vint/ast/plugin/scope_plugin/scope_linker.py | ScopeLinker.process | def process(self, ast): # type: (Dict[str, Any]) -> None
""" Build a scope tree and links between scopes and identifiers by the
specified ast. You can access the built scope tree and the built links
by .scope_tree and .link_registry.
"""
id_classifier = IdentifierClassifier()
attached_ast = id_classifier.attach_identifier_attributes(ast)
# We are already in script local scope.
self._scope_tree_builder.enter_new_scope(ScopeVisibility.SCRIPT_LOCAL)
traverse(attached_ast,
on_enter=self._enter_handler,
on_leave=self._leave_handler)
self.scope_tree = self._scope_tree_builder.get_global_scope()
self.link_registry = self._scope_tree_builder.link_registry | python | def process(self, ast): # type: (Dict[str, Any]) -> None
""" Build a scope tree and links between scopes and identifiers by the
specified ast. You can access the built scope tree and the built links
by .scope_tree and .link_registry.
"""
id_classifier = IdentifierClassifier()
attached_ast = id_classifier.attach_identifier_attributes(ast)
# We are already in script local scope.
self._scope_tree_builder.enter_new_scope(ScopeVisibility.SCRIPT_LOCAL)
traverse(attached_ast,
on_enter=self._enter_handler,
on_leave=self._leave_handler)
self.scope_tree = self._scope_tree_builder.get_global_scope()
self.link_registry = self._scope_tree_builder.link_registry | [
"def",
"process",
"(",
"self",
",",
"ast",
")",
":",
"# type: (Dict[str, Any]) -> None",
"id_classifier",
"=",
"IdentifierClassifier",
"(",
")",
"attached_ast",
"=",
"id_classifier",
".",
"attach_identifier_attributes",
"(",
"ast",
")",
"# We are already in script local scope.",
"self",
".",
"_scope_tree_builder",
".",
"enter_new_scope",
"(",
"ScopeVisibility",
".",
"SCRIPT_LOCAL",
")",
"traverse",
"(",
"attached_ast",
",",
"on_enter",
"=",
"self",
".",
"_enter_handler",
",",
"on_leave",
"=",
"self",
".",
"_leave_handler",
")",
"self",
".",
"scope_tree",
"=",
"self",
".",
"_scope_tree_builder",
".",
"get_global_scope",
"(",
")",
"self",
".",
"link_registry",
"=",
"self",
".",
"_scope_tree_builder",
".",
"link_registry"
] | Build a scope tree and links between scopes and identifiers by the
specified ast. You can access the built scope tree and the built links
by .scope_tree and .link_registry. | [
"Build",
"a",
"scope",
"tree",
"and",
"links",
"between",
"scopes",
"and",
"identifiers",
"by",
"the",
"specified",
"ast",
".",
"You",
"can",
"access",
"the",
"built",
"scope",
"tree",
"and",
"the",
"built",
"links",
"by",
".",
"scope_tree",
"and",
".",
"link_registry",
"."
] | db29337d859d88239c282c2e9d84c858f23a4a09 | https://github.com/Kuniwak/vint/blob/db29337d859d88239c282c2e9d84c858f23a4a09/vint/ast/plugin/scope_plugin/scope_linker.py#L326-L342 | train |
mozilla/mozdownload | mozdownload/cli.py | cli | def cli(argv=None):
"""CLI entry point for mozdownload."""
kwargs = parse_arguments(argv or sys.argv[1:])
log_level = kwargs.pop('log_level')
logging.basicConfig(format='%(levelname)s | %(message)s', level=log_level)
logger = logging.getLogger(__name__)
# Configure logging levels for sub modules. Set to ERROR by default.
sub_log_level = logging.ERROR
if log_level == logging.getLevelName(logging.DEBUG):
sub_log_level = logging.DEBUG
logging.getLogger('redo').setLevel(sub_log_level)
logging.getLogger('requests').setLevel(sub_log_level)
logging.getLogger('thclient').setLevel(sub_log_level)
try:
scraper_type = kwargs.pop('scraper_type')
# If a URL has been specified use the direct scraper
if kwargs.get('url'):
scraper_type = 'direct'
build = factory.FactoryScraper(scraper_type, **kwargs)
if kwargs.get('print_url'):
logger.info(build.url)
else:
build.download()
except KeyboardInterrupt:
logger.error('Download interrupted by the user') | python | def cli(argv=None):
"""CLI entry point for mozdownload."""
kwargs = parse_arguments(argv or sys.argv[1:])
log_level = kwargs.pop('log_level')
logging.basicConfig(format='%(levelname)s | %(message)s', level=log_level)
logger = logging.getLogger(__name__)
# Configure logging levels for sub modules. Set to ERROR by default.
sub_log_level = logging.ERROR
if log_level == logging.getLevelName(logging.DEBUG):
sub_log_level = logging.DEBUG
logging.getLogger('redo').setLevel(sub_log_level)
logging.getLogger('requests').setLevel(sub_log_level)
logging.getLogger('thclient').setLevel(sub_log_level)
try:
scraper_type = kwargs.pop('scraper_type')
# If a URL has been specified use the direct scraper
if kwargs.get('url'):
scraper_type = 'direct'
build = factory.FactoryScraper(scraper_type, **kwargs)
if kwargs.get('print_url'):
logger.info(build.url)
else:
build.download()
except KeyboardInterrupt:
logger.error('Download interrupted by the user') | [
"def",
"cli",
"(",
"argv",
"=",
"None",
")",
":",
"kwargs",
"=",
"parse_arguments",
"(",
"argv",
"or",
"sys",
".",
"argv",
"[",
"1",
":",
"]",
")",
"log_level",
"=",
"kwargs",
".",
"pop",
"(",
"'log_level'",
")",
"logging",
".",
"basicConfig",
"(",
"format",
"=",
"'%(levelname)s | %(message)s'",
",",
"level",
"=",
"log_level",
")",
"logger",
"=",
"logging",
".",
"getLogger",
"(",
"__name__",
")",
"# Configure logging levels for sub modules. Set to ERROR by default.",
"sub_log_level",
"=",
"logging",
".",
"ERROR",
"if",
"log_level",
"==",
"logging",
".",
"getLevelName",
"(",
"logging",
".",
"DEBUG",
")",
":",
"sub_log_level",
"=",
"logging",
".",
"DEBUG",
"logging",
".",
"getLogger",
"(",
"'redo'",
")",
".",
"setLevel",
"(",
"sub_log_level",
")",
"logging",
".",
"getLogger",
"(",
"'requests'",
")",
".",
"setLevel",
"(",
"sub_log_level",
")",
"logging",
".",
"getLogger",
"(",
"'thclient'",
")",
".",
"setLevel",
"(",
"sub_log_level",
")",
"try",
":",
"scraper_type",
"=",
"kwargs",
".",
"pop",
"(",
"'scraper_type'",
")",
"# If a URL has been specified use the direct scraper",
"if",
"kwargs",
".",
"get",
"(",
"'url'",
")",
":",
"scraper_type",
"=",
"'direct'",
"build",
"=",
"factory",
".",
"FactoryScraper",
"(",
"scraper_type",
",",
"*",
"*",
"kwargs",
")",
"if",
"kwargs",
".",
"get",
"(",
"'print_url'",
")",
":",
"logger",
".",
"info",
"(",
"build",
".",
"url",
")",
"else",
":",
"build",
".",
"download",
"(",
")",
"except",
"KeyboardInterrupt",
":",
"logger",
".",
"error",
"(",
"'Download interrupted by the user'",
")"
] | CLI entry point for mozdownload. | [
"CLI",
"entry",
"point",
"for",
"mozdownload",
"."
] | 97796a028455bb5200434562d23b66d5a5eb537b | https://github.com/mozilla/mozdownload/blob/97796a028455bb5200434562d23b66d5a5eb537b/mozdownload/cli.py#L146-L175 | train |
mozilla/mozdownload | mozdownload/treeherder.py | Treeherder.query_builds_by_revision | def query_builds_by_revision(self, revision, job_type_name='Build', debug_build=False):
"""Retrieve build folders for a given revision with the help of Treeherder.
:param revision: Revision of the build to download.
:param job_type_name: Name of the job to look for. For builds it should be
'Build', 'Nightly', and 'L10n Nightly'. Defaults to `Build`.
:param debug_build: Download a debug build.
"""
builds = set()
try:
self.logger.info('Querying {url} for list of builds for revision: {revision}'.format(
url=self.client.server_url, revision=revision))
# Retrieve the option hash to filter for type of build (opt, and debug for now)
option_hash = None
for key, values in self.client.get_option_collection_hash().iteritems():
for value in values:
if value['name'] == ('debug' if debug_build else 'opt'):
option_hash = key
break
if option_hash:
break
resultsets = self.client.get_pushes(self.branch, revision=revision)
# Set filters to speed-up querying jobs
kwargs = {
'option_collection_hash': option_hash,
'job_type_name': job_type_name,
'exclusion_profile': False,
}
kwargs.update(self.get_treeherder_platform(self.platform))
for resultset in resultsets:
kwargs.update({'result_set_id': resultset['id']})
jobs = self.client.get_jobs(self.branch, **kwargs)
for job in jobs:
log_urls = self.client.get_job_log_url(self.branch, job_id=job['id'])
for log_url in log_urls:
if self.application in log_url['url']:
self.logger.debug('Found build folder: {}'.format(log_url['url']))
builds.update([log_url['url']])
except Exception:
self.logger.exception('Failure occurred when querying Treeherder for builds')
return list(builds) | python | def query_builds_by_revision(self, revision, job_type_name='Build', debug_build=False):
"""Retrieve build folders for a given revision with the help of Treeherder.
:param revision: Revision of the build to download.
:param job_type_name: Name of the job to look for. For builds it should be
'Build', 'Nightly', and 'L10n Nightly'. Defaults to `Build`.
:param debug_build: Download a debug build.
"""
builds = set()
try:
self.logger.info('Querying {url} for list of builds for revision: {revision}'.format(
url=self.client.server_url, revision=revision))
# Retrieve the option hash to filter for type of build (opt, and debug for now)
option_hash = None
for key, values in self.client.get_option_collection_hash().iteritems():
for value in values:
if value['name'] == ('debug' if debug_build else 'opt'):
option_hash = key
break
if option_hash:
break
resultsets = self.client.get_pushes(self.branch, revision=revision)
# Set filters to speed-up querying jobs
kwargs = {
'option_collection_hash': option_hash,
'job_type_name': job_type_name,
'exclusion_profile': False,
}
kwargs.update(self.get_treeherder_platform(self.platform))
for resultset in resultsets:
kwargs.update({'result_set_id': resultset['id']})
jobs = self.client.get_jobs(self.branch, **kwargs)
for job in jobs:
log_urls = self.client.get_job_log_url(self.branch, job_id=job['id'])
for log_url in log_urls:
if self.application in log_url['url']:
self.logger.debug('Found build folder: {}'.format(log_url['url']))
builds.update([log_url['url']])
except Exception:
self.logger.exception('Failure occurred when querying Treeherder for builds')
return list(builds) | [
"def",
"query_builds_by_revision",
"(",
"self",
",",
"revision",
",",
"job_type_name",
"=",
"'Build'",
",",
"debug_build",
"=",
"False",
")",
":",
"builds",
"=",
"set",
"(",
")",
"try",
":",
"self",
".",
"logger",
".",
"info",
"(",
"'Querying {url} for list of builds for revision: {revision}'",
".",
"format",
"(",
"url",
"=",
"self",
".",
"client",
".",
"server_url",
",",
"revision",
"=",
"revision",
")",
")",
"# Retrieve the option hash to filter for type of build (opt, and debug for now)",
"option_hash",
"=",
"None",
"for",
"key",
",",
"values",
"in",
"self",
".",
"client",
".",
"get_option_collection_hash",
"(",
")",
".",
"iteritems",
"(",
")",
":",
"for",
"value",
"in",
"values",
":",
"if",
"value",
"[",
"'name'",
"]",
"==",
"(",
"'debug'",
"if",
"debug_build",
"else",
"'opt'",
")",
":",
"option_hash",
"=",
"key",
"break",
"if",
"option_hash",
":",
"break",
"resultsets",
"=",
"self",
".",
"client",
".",
"get_pushes",
"(",
"self",
".",
"branch",
",",
"revision",
"=",
"revision",
")",
"# Set filters to speed-up querying jobs",
"kwargs",
"=",
"{",
"'option_collection_hash'",
":",
"option_hash",
",",
"'job_type_name'",
":",
"job_type_name",
",",
"'exclusion_profile'",
":",
"False",
",",
"}",
"kwargs",
".",
"update",
"(",
"self",
".",
"get_treeherder_platform",
"(",
"self",
".",
"platform",
")",
")",
"for",
"resultset",
"in",
"resultsets",
":",
"kwargs",
".",
"update",
"(",
"{",
"'result_set_id'",
":",
"resultset",
"[",
"'id'",
"]",
"}",
")",
"jobs",
"=",
"self",
".",
"client",
".",
"get_jobs",
"(",
"self",
".",
"branch",
",",
"*",
"*",
"kwargs",
")",
"for",
"job",
"in",
"jobs",
":",
"log_urls",
"=",
"self",
".",
"client",
".",
"get_job_log_url",
"(",
"self",
".",
"branch",
",",
"job_id",
"=",
"job",
"[",
"'id'",
"]",
")",
"for",
"log_url",
"in",
"log_urls",
":",
"if",
"self",
".",
"application",
"in",
"log_url",
"[",
"'url'",
"]",
":",
"self",
".",
"logger",
".",
"debug",
"(",
"'Found build folder: {}'",
".",
"format",
"(",
"log_url",
"[",
"'url'",
"]",
")",
")",
"builds",
".",
"update",
"(",
"[",
"log_url",
"[",
"'url'",
"]",
"]",
")",
"except",
"Exception",
":",
"self",
".",
"logger",
".",
"exception",
"(",
"'Failure occurred when querying Treeherder for builds'",
")",
"return",
"list",
"(",
"builds",
")"
] | Retrieve build folders for a given revision with the help of Treeherder.
:param revision: Revision of the build to download.
:param job_type_name: Name of the job to look for. For builds it should be
'Build', 'Nightly', and 'L10n Nightly'. Defaults to `Build`.
:param debug_build: Download a debug build. | [
"Retrieve",
"build",
"folders",
"for",
"a",
"given",
"revision",
"with",
"the",
"help",
"of",
"Treeherder",
"."
] | 97796a028455bb5200434562d23b66d5a5eb537b | https://github.com/mozilla/mozdownload/blob/97796a028455bb5200434562d23b66d5a5eb537b/mozdownload/treeherder.py#L60-L107 | train |
mozilla/mozdownload | mozdownload/utils.py | urljoin | def urljoin(*fragments):
"""Concatenate multi part strings into urls."""
# Strip possible already existent final slashes of fragments except for the last one
parts = [fragment.rstrip('/') for fragment in fragments[:len(fragments) - 1]]
parts.append(fragments[-1])
return '/'.join(parts) | python | def urljoin(*fragments):
"""Concatenate multi part strings into urls."""
# Strip possible already existent final slashes of fragments except for the last one
parts = [fragment.rstrip('/') for fragment in fragments[:len(fragments) - 1]]
parts.append(fragments[-1])
return '/'.join(parts) | [
"def",
"urljoin",
"(",
"*",
"fragments",
")",
":",
"# Strip possible already existent final slashes of fragments except for the last one",
"parts",
"=",
"[",
"fragment",
".",
"rstrip",
"(",
"'/'",
")",
"for",
"fragment",
"in",
"fragments",
"[",
":",
"len",
"(",
"fragments",
")",
"-",
"1",
"]",
"]",
"parts",
".",
"append",
"(",
"fragments",
"[",
"-",
"1",
"]",
")",
"return",
"'/'",
".",
"join",
"(",
"parts",
")"
] | Concatenate multi part strings into urls. | [
"Concatenate",
"multi",
"part",
"strings",
"into",
"urls",
"."
] | 97796a028455bb5200434562d23b66d5a5eb537b | https://github.com/mozilla/mozdownload/blob/97796a028455bb5200434562d23b66d5a5eb537b/mozdownload/utils.py#L12-L18 | train |
mozilla/mozdownload | mozdownload/utils.py | create_md5 | def create_md5(path):
"""Create the md5 hash of a file using the hashlib library."""
m = hashlib.md5()
# rb necessary to run correctly in windows.
with open(path, "rb") as f:
while True:
data = f.read(8192)
if not data:
break
m.update(data)
return m.hexdigest() | python | def create_md5(path):
"""Create the md5 hash of a file using the hashlib library."""
m = hashlib.md5()
# rb necessary to run correctly in windows.
with open(path, "rb") as f:
while True:
data = f.read(8192)
if not data:
break
m.update(data)
return m.hexdigest() | [
"def",
"create_md5",
"(",
"path",
")",
":",
"m",
"=",
"hashlib",
".",
"md5",
"(",
")",
"# rb necessary to run correctly in windows.",
"with",
"open",
"(",
"path",
",",
"\"rb\"",
")",
"as",
"f",
":",
"while",
"True",
":",
"data",
"=",
"f",
".",
"read",
"(",
"8192",
")",
"if",
"not",
"data",
":",
"break",
"m",
".",
"update",
"(",
"data",
")",
"return",
"m",
".",
"hexdigest",
"(",
")"
] | Create the md5 hash of a file using the hashlib library. | [
"Create",
"the",
"md5",
"hash",
"of",
"a",
"file",
"using",
"the",
"hashlib",
"library",
"."
] | 97796a028455bb5200434562d23b66d5a5eb537b | https://github.com/mozilla/mozdownload/blob/97796a028455bb5200434562d23b66d5a5eb537b/mozdownload/utils.py#L21-L32 | train |
mozilla/mozdownload | mozdownload/parser.py | DirectoryParser.filter | def filter(self, filter):
"""Filter entries by calling function or applying regex."""
if hasattr(filter, '__call__'):
return [entry for entry in self.entries if filter(entry)]
else:
pattern = re.compile(filter, re.IGNORECASE)
return [entry for entry in self.entries if pattern.match(entry)] | python | def filter(self, filter):
"""Filter entries by calling function or applying regex."""
if hasattr(filter, '__call__'):
return [entry for entry in self.entries if filter(entry)]
else:
pattern = re.compile(filter, re.IGNORECASE)
return [entry for entry in self.entries if pattern.match(entry)] | [
"def",
"filter",
"(",
"self",
",",
"filter",
")",
":",
"if",
"hasattr",
"(",
"filter",
",",
"'__call__'",
")",
":",
"return",
"[",
"entry",
"for",
"entry",
"in",
"self",
".",
"entries",
"if",
"filter",
"(",
"entry",
")",
"]",
"else",
":",
"pattern",
"=",
"re",
".",
"compile",
"(",
"filter",
",",
"re",
".",
"IGNORECASE",
")",
"return",
"[",
"entry",
"for",
"entry",
"in",
"self",
".",
"entries",
"if",
"pattern",
".",
"match",
"(",
"entry",
")",
"]"
] | Filter entries by calling function or applying regex. | [
"Filter",
"entries",
"by",
"calling",
"function",
"or",
"applying",
"regex",
"."
] | 97796a028455bb5200434562d23b66d5a5eb537b | https://github.com/mozilla/mozdownload/blob/97796a028455bb5200434562d23b66d5a5eb537b/mozdownload/parser.py#L53-L59 | train |
mozilla/mozdownload | mozdownload/parser.py | DirectoryParser.handle_starttag | def handle_starttag(self, tag, attrs):
"""Callback for when a tag gets opened."""
if not tag == 'a':
return
for attr in attrs:
if attr[0] == 'href':
# Links look like: /pub/firefox/nightly/2015/
# We have to trim the fragment down to the last item. Also to ensure we
# always get it, we remove a possible final slash first
url = urllib.unquote(attr[1])
self.active_url = url.rstrip('/').split('/')[-1]
return | python | def handle_starttag(self, tag, attrs):
"""Callback for when a tag gets opened."""
if not tag == 'a':
return
for attr in attrs:
if attr[0] == 'href':
# Links look like: /pub/firefox/nightly/2015/
# We have to trim the fragment down to the last item. Also to ensure we
# always get it, we remove a possible final slash first
url = urllib.unquote(attr[1])
self.active_url = url.rstrip('/').split('/')[-1]
return | [
"def",
"handle_starttag",
"(",
"self",
",",
"tag",
",",
"attrs",
")",
":",
"if",
"not",
"tag",
"==",
"'a'",
":",
"return",
"for",
"attr",
"in",
"attrs",
":",
"if",
"attr",
"[",
"0",
"]",
"==",
"'href'",
":",
"# Links look like: /pub/firefox/nightly/2015/",
"# We have to trim the fragment down to the last item. Also to ensure we",
"# always get it, we remove a possible final slash first",
"url",
"=",
"urllib",
".",
"unquote",
"(",
"attr",
"[",
"1",
"]",
")",
"self",
".",
"active_url",
"=",
"url",
".",
"rstrip",
"(",
"'/'",
")",
".",
"split",
"(",
"'/'",
")",
"[",
"-",
"1",
"]",
"return"
] | Callback for when a tag gets opened. | [
"Callback",
"for",
"when",
"a",
"tag",
"gets",
"opened",
"."
] | 97796a028455bb5200434562d23b66d5a5eb537b | https://github.com/mozilla/mozdownload/blob/97796a028455bb5200434562d23b66d5a5eb537b/mozdownload/parser.py#L61-L74 | train |
mozilla/mozdownload | mozdownload/parser.py | DirectoryParser.handle_data | def handle_data(self, data):
"""Callback when the data of a tag has been collected."""
# Only process the data when we are in an active a tag and have an URL.
if not self.active_url:
return
# The visible text can have a final slash so strip it off
if data.strip('/') == self.active_url:
self.entries.append(self.active_url) | python | def handle_data(self, data):
"""Callback when the data of a tag has been collected."""
# Only process the data when we are in an active a tag and have an URL.
if not self.active_url:
return
# The visible text can have a final slash so strip it off
if data.strip('/') == self.active_url:
self.entries.append(self.active_url) | [
"def",
"handle_data",
"(",
"self",
",",
"data",
")",
":",
"# Only process the data when we are in an active a tag and have an URL.",
"if",
"not",
"self",
".",
"active_url",
":",
"return",
"# The visible text can have a final slash so strip it off",
"if",
"data",
".",
"strip",
"(",
"'/'",
")",
"==",
"self",
".",
"active_url",
":",
"self",
".",
"entries",
".",
"append",
"(",
"self",
".",
"active_url",
")"
] | Callback when the data of a tag has been collected. | [
"Callback",
"when",
"the",
"data",
"of",
"a",
"tag",
"has",
"been",
"collected",
"."
] | 97796a028455bb5200434562d23b66d5a5eb537b | https://github.com/mozilla/mozdownload/blob/97796a028455bb5200434562d23b66d5a5eb537b/mozdownload/parser.py#L81-L89 | train |
mozilla/mozdownload | mozdownload/timezones.py | PacificTimezone.dst | def dst(self, dt):
"""Calculate delta for daylight saving."""
# Daylight saving starts on the second Sunday of March at 2AM standard
dst_start_date = self.first_sunday(dt.year, 3) + timedelta(days=7) \
+ timedelta(hours=2)
# Daylight saving ends on the first Sunday of November at 2AM standard
dst_end_date = self.first_sunday(dt.year, 11) + timedelta(hours=2)
if dst_start_date <= dt.replace(tzinfo=None) < dst_end_date:
return timedelta(hours=1)
else:
return timedelta(0) | python | def dst(self, dt):
"""Calculate delta for daylight saving."""
# Daylight saving starts on the second Sunday of March at 2AM standard
dst_start_date = self.first_sunday(dt.year, 3) + timedelta(days=7) \
+ timedelta(hours=2)
# Daylight saving ends on the first Sunday of November at 2AM standard
dst_end_date = self.first_sunday(dt.year, 11) + timedelta(hours=2)
if dst_start_date <= dt.replace(tzinfo=None) < dst_end_date:
return timedelta(hours=1)
else:
return timedelta(0) | [
"def",
"dst",
"(",
"self",
",",
"dt",
")",
":",
"# Daylight saving starts on the second Sunday of March at 2AM standard",
"dst_start_date",
"=",
"self",
".",
"first_sunday",
"(",
"dt",
".",
"year",
",",
"3",
")",
"+",
"timedelta",
"(",
"days",
"=",
"7",
")",
"+",
"timedelta",
"(",
"hours",
"=",
"2",
")",
"# Daylight saving ends on the first Sunday of November at 2AM standard",
"dst_end_date",
"=",
"self",
".",
"first_sunday",
"(",
"dt",
".",
"year",
",",
"11",
")",
"+",
"timedelta",
"(",
"hours",
"=",
"2",
")",
"if",
"dst_start_date",
"<=",
"dt",
".",
"replace",
"(",
"tzinfo",
"=",
"None",
")",
"<",
"dst_end_date",
":",
"return",
"timedelta",
"(",
"hours",
"=",
"1",
")",
"else",
":",
"return",
"timedelta",
"(",
"0",
")"
] | Calculate delta for daylight saving. | [
"Calculate",
"delta",
"for",
"daylight",
"saving",
"."
] | 97796a028455bb5200434562d23b66d5a5eb537b | https://github.com/mozilla/mozdownload/blob/97796a028455bb5200434562d23b66d5a5eb537b/mozdownload/timezones.py#L23-L34 | train |
mozilla/mozdownload | mozdownload/timezones.py | PacificTimezone.first_sunday | def first_sunday(self, year, month):
"""Get the first sunday of a month."""
date = datetime(year, month, 1, 0)
days_until_sunday = 6 - date.weekday()
return date + timedelta(days=days_until_sunday) | python | def first_sunday(self, year, month):
"""Get the first sunday of a month."""
date = datetime(year, month, 1, 0)
days_until_sunday = 6 - date.weekday()
return date + timedelta(days=days_until_sunday) | [
"def",
"first_sunday",
"(",
"self",
",",
"year",
",",
"month",
")",
":",
"date",
"=",
"datetime",
"(",
"year",
",",
"month",
",",
"1",
",",
"0",
")",
"days_until_sunday",
"=",
"6",
"-",
"date",
".",
"weekday",
"(",
")",
"return",
"date",
"+",
"timedelta",
"(",
"days",
"=",
"days_until_sunday",
")"
] | Get the first sunday of a month. | [
"Get",
"the",
"first",
"sunday",
"of",
"a",
"month",
"."
] | 97796a028455bb5200434562d23b66d5a5eb537b | https://github.com/mozilla/mozdownload/blob/97796a028455bb5200434562d23b66d5a5eb537b/mozdownload/timezones.py#L36-L41 | train |
mozilla/mozdownload | mozdownload/scraper.py | Scraper.binary | def binary(self):
"""Return the name of the build."""
def _get_binary():
# Retrieve all entries from the remote virtual folder
parser = self._create_directory_parser(self.path)
if not parser.entries:
raise errors.NotFoundError('No entries found', self.path)
# Download the first matched directory entry
pattern = re.compile(self.binary_regex, re.IGNORECASE)
for entry in parser.entries:
try:
self._binary = pattern.match(entry).group()
break
except Exception:
# No match, continue with next entry
continue
else:
raise errors.NotFoundError("Binary not found in folder",
self.path)
self._retry_check_404(_get_binary)
return self._binary | python | def binary(self):
"""Return the name of the build."""
def _get_binary():
# Retrieve all entries from the remote virtual folder
parser = self._create_directory_parser(self.path)
if not parser.entries:
raise errors.NotFoundError('No entries found', self.path)
# Download the first matched directory entry
pattern = re.compile(self.binary_regex, re.IGNORECASE)
for entry in parser.entries:
try:
self._binary = pattern.match(entry).group()
break
except Exception:
# No match, continue with next entry
continue
else:
raise errors.NotFoundError("Binary not found in folder",
self.path)
self._retry_check_404(_get_binary)
return self._binary | [
"def",
"binary",
"(",
"self",
")",
":",
"def",
"_get_binary",
"(",
")",
":",
"# Retrieve all entries from the remote virtual folder",
"parser",
"=",
"self",
".",
"_create_directory_parser",
"(",
"self",
".",
"path",
")",
"if",
"not",
"parser",
".",
"entries",
":",
"raise",
"errors",
".",
"NotFoundError",
"(",
"'No entries found'",
",",
"self",
".",
"path",
")",
"# Download the first matched directory entry",
"pattern",
"=",
"re",
".",
"compile",
"(",
"self",
".",
"binary_regex",
",",
"re",
".",
"IGNORECASE",
")",
"for",
"entry",
"in",
"parser",
".",
"entries",
":",
"try",
":",
"self",
".",
"_binary",
"=",
"pattern",
".",
"match",
"(",
"entry",
")",
".",
"group",
"(",
")",
"break",
"except",
"Exception",
":",
"# No match, continue with next entry",
"continue",
"else",
":",
"raise",
"errors",
".",
"NotFoundError",
"(",
"\"Binary not found in folder\"",
",",
"self",
".",
"path",
")",
"self",
".",
"_retry_check_404",
"(",
"_get_binary",
")",
"return",
"self",
".",
"_binary"
] | Return the name of the build. | [
"Return",
"the",
"name",
"of",
"the",
"build",
"."
] | 97796a028455bb5200434562d23b66d5a5eb537b | https://github.com/mozilla/mozdownload/blob/97796a028455bb5200434562d23b66d5a5eb537b/mozdownload/scraper.py#L165-L188 | train |
mozilla/mozdownload | mozdownload/scraper.py | Scraper.url | def url(self):
"""Return the URL of the build."""
return urllib.quote(urljoin(self.path, self.binary),
safe='%/:=&?~#+!$,;\'@()*[]|') | python | def url(self):
"""Return the URL of the build."""
return urllib.quote(urljoin(self.path, self.binary),
safe='%/:=&?~#+!$,;\'@()*[]|') | [
"def",
"url",
"(",
"self",
")",
":",
"return",
"urllib",
".",
"quote",
"(",
"urljoin",
"(",
"self",
".",
"path",
",",
"self",
".",
"binary",
")",
",",
"safe",
"=",
"'%/:=&?~#+!$,;\\'@()*[]|'",
")"
] | Return the URL of the build. | [
"Return",
"the",
"URL",
"of",
"the",
"build",
"."
] | 97796a028455bb5200434562d23b66d5a5eb537b | https://github.com/mozilla/mozdownload/blob/97796a028455bb5200434562d23b66d5a5eb537b/mozdownload/scraper.py#L196-L199 | train |
mozilla/mozdownload | mozdownload/scraper.py | Scraper.filename | def filename(self):
"""Return the local filename of the build."""
if self._filename is None:
if os.path.splitext(self.destination)[1]:
# If the filename has been given make use of it
target_file = self.destination
else:
# Otherwise create it from the build details
target_file = os.path.join(self.destination,
self.build_filename(self.binary))
self._filename = os.path.abspath(target_file)
return self._filename | python | def filename(self):
"""Return the local filename of the build."""
if self._filename is None:
if os.path.splitext(self.destination)[1]:
# If the filename has been given make use of it
target_file = self.destination
else:
# Otherwise create it from the build details
target_file = os.path.join(self.destination,
self.build_filename(self.binary))
self._filename = os.path.abspath(target_file)
return self._filename | [
"def",
"filename",
"(",
"self",
")",
":",
"if",
"self",
".",
"_filename",
"is",
"None",
":",
"if",
"os",
".",
"path",
".",
"splitext",
"(",
"self",
".",
"destination",
")",
"[",
"1",
"]",
":",
"# If the filename has been given make use of it",
"target_file",
"=",
"self",
".",
"destination",
"else",
":",
"# Otherwise create it from the build details",
"target_file",
"=",
"os",
".",
"path",
".",
"join",
"(",
"self",
".",
"destination",
",",
"self",
".",
"build_filename",
"(",
"self",
".",
"binary",
")",
")",
"self",
".",
"_filename",
"=",
"os",
".",
"path",
".",
"abspath",
"(",
"target_file",
")",
"return",
"self",
".",
"_filename"
] | Return the local filename of the build. | [
"Return",
"the",
"local",
"filename",
"of",
"the",
"build",
"."
] | 97796a028455bb5200434562d23b66d5a5eb537b | https://github.com/mozilla/mozdownload/blob/97796a028455bb5200434562d23b66d5a5eb537b/mozdownload/scraper.py#L217-L230 | train |
mozilla/mozdownload | mozdownload/scraper.py | Scraper.download | def download(self):
"""Download the specified file."""
def total_seconds(td):
# Keep backward compatibility with Python 2.6 which doesn't have
# this method
if hasattr(td, 'total_seconds'):
return td.total_seconds()
else:
return (td.microseconds +
(td.seconds + td.days * 24 * 3600) * 10 ** 6) / 10 ** 6
# Don't re-download the file
if os.path.isfile(os.path.abspath(self.filename)):
self.logger.info("File has already been downloaded: %s" %
(self.filename))
return self.filename
directory = os.path.dirname(self.filename)
if not os.path.isdir(directory):
os.makedirs(directory)
self.logger.info('Downloading from: %s' % self.url)
self.logger.info('Saving as: %s' % self.filename)
tmp_file = self.filename + ".part"
def _download():
try:
start_time = datetime.now()
# Enable streaming mode so we can download content in chunks
r = self.session.get(self.url, stream=True)
r.raise_for_status()
content_length = r.headers.get('Content-length')
# ValueError: Value out of range if only total_size given
if content_length:
total_size = int(content_length.strip())
max_value = ((total_size / CHUNK_SIZE) + 1) * CHUNK_SIZE
bytes_downloaded = 0
log_level = self.logger.getEffectiveLevel()
if log_level <= logging.INFO and content_length:
widgets = [pb.Percentage(), ' ', pb.Bar(), ' ', pb.ETA(),
' ', pb.FileTransferSpeed()]
pbar = pb.ProgressBar(widgets=widgets,
maxval=max_value).start()
with open(tmp_file, 'wb') as f:
for chunk in r.iter_content(CHUNK_SIZE):
f.write(chunk)
bytes_downloaded += CHUNK_SIZE
if log_level <= logging.INFO and content_length:
pbar.update(bytes_downloaded)
t1 = total_seconds(datetime.now() - start_time)
if self.timeout_download and \
t1 >= self.timeout_download:
raise errors.TimeoutError
if log_level <= logging.INFO and content_length:
pbar.finish()
except Exception:
if os.path.isfile(tmp_file):
os.remove(tmp_file)
raise
self._retry(_download,
retry_exceptions=(requests.exceptions.RequestException,
errors.TimeoutError))
os.rename(tmp_file, self.filename)
return self.filename | python | def download(self):
"""Download the specified file."""
def total_seconds(td):
# Keep backward compatibility with Python 2.6 which doesn't have
# this method
if hasattr(td, 'total_seconds'):
return td.total_seconds()
else:
return (td.microseconds +
(td.seconds + td.days * 24 * 3600) * 10 ** 6) / 10 ** 6
# Don't re-download the file
if os.path.isfile(os.path.abspath(self.filename)):
self.logger.info("File has already been downloaded: %s" %
(self.filename))
return self.filename
directory = os.path.dirname(self.filename)
if not os.path.isdir(directory):
os.makedirs(directory)
self.logger.info('Downloading from: %s' % self.url)
self.logger.info('Saving as: %s' % self.filename)
tmp_file = self.filename + ".part"
def _download():
try:
start_time = datetime.now()
# Enable streaming mode so we can download content in chunks
r = self.session.get(self.url, stream=True)
r.raise_for_status()
content_length = r.headers.get('Content-length')
# ValueError: Value out of range if only total_size given
if content_length:
total_size = int(content_length.strip())
max_value = ((total_size / CHUNK_SIZE) + 1) * CHUNK_SIZE
bytes_downloaded = 0
log_level = self.logger.getEffectiveLevel()
if log_level <= logging.INFO and content_length:
widgets = [pb.Percentage(), ' ', pb.Bar(), ' ', pb.ETA(),
' ', pb.FileTransferSpeed()]
pbar = pb.ProgressBar(widgets=widgets,
maxval=max_value).start()
with open(tmp_file, 'wb') as f:
for chunk in r.iter_content(CHUNK_SIZE):
f.write(chunk)
bytes_downloaded += CHUNK_SIZE
if log_level <= logging.INFO and content_length:
pbar.update(bytes_downloaded)
t1 = total_seconds(datetime.now() - start_time)
if self.timeout_download and \
t1 >= self.timeout_download:
raise errors.TimeoutError
if log_level <= logging.INFO and content_length:
pbar.finish()
except Exception:
if os.path.isfile(tmp_file):
os.remove(tmp_file)
raise
self._retry(_download,
retry_exceptions=(requests.exceptions.RequestException,
errors.TimeoutError))
os.rename(tmp_file, self.filename)
return self.filename | [
"def",
"download",
"(",
"self",
")",
":",
"def",
"total_seconds",
"(",
"td",
")",
":",
"# Keep backward compatibility with Python 2.6 which doesn't have",
"# this method",
"if",
"hasattr",
"(",
"td",
",",
"'total_seconds'",
")",
":",
"return",
"td",
".",
"total_seconds",
"(",
")",
"else",
":",
"return",
"(",
"td",
".",
"microseconds",
"+",
"(",
"td",
".",
"seconds",
"+",
"td",
".",
"days",
"*",
"24",
"*",
"3600",
")",
"*",
"10",
"**",
"6",
")",
"/",
"10",
"**",
"6",
"# Don't re-download the file",
"if",
"os",
".",
"path",
".",
"isfile",
"(",
"os",
".",
"path",
".",
"abspath",
"(",
"self",
".",
"filename",
")",
")",
":",
"self",
".",
"logger",
".",
"info",
"(",
"\"File has already been downloaded: %s\"",
"%",
"(",
"self",
".",
"filename",
")",
")",
"return",
"self",
".",
"filename",
"directory",
"=",
"os",
".",
"path",
".",
"dirname",
"(",
"self",
".",
"filename",
")",
"if",
"not",
"os",
".",
"path",
".",
"isdir",
"(",
"directory",
")",
":",
"os",
".",
"makedirs",
"(",
"directory",
")",
"self",
".",
"logger",
".",
"info",
"(",
"'Downloading from: %s'",
"%",
"self",
".",
"url",
")",
"self",
".",
"logger",
".",
"info",
"(",
"'Saving as: %s'",
"%",
"self",
".",
"filename",
")",
"tmp_file",
"=",
"self",
".",
"filename",
"+",
"\".part\"",
"def",
"_download",
"(",
")",
":",
"try",
":",
"start_time",
"=",
"datetime",
".",
"now",
"(",
")",
"# Enable streaming mode so we can download content in chunks",
"r",
"=",
"self",
".",
"session",
".",
"get",
"(",
"self",
".",
"url",
",",
"stream",
"=",
"True",
")",
"r",
".",
"raise_for_status",
"(",
")",
"content_length",
"=",
"r",
".",
"headers",
".",
"get",
"(",
"'Content-length'",
")",
"# ValueError: Value out of range if only total_size given",
"if",
"content_length",
":",
"total_size",
"=",
"int",
"(",
"content_length",
".",
"strip",
"(",
")",
")",
"max_value",
"=",
"(",
"(",
"total_size",
"/",
"CHUNK_SIZE",
")",
"+",
"1",
")",
"*",
"CHUNK_SIZE",
"bytes_downloaded",
"=",
"0",
"log_level",
"=",
"self",
".",
"logger",
".",
"getEffectiveLevel",
"(",
")",
"if",
"log_level",
"<=",
"logging",
".",
"INFO",
"and",
"content_length",
":",
"widgets",
"=",
"[",
"pb",
".",
"Percentage",
"(",
")",
",",
"' '",
",",
"pb",
".",
"Bar",
"(",
")",
",",
"' '",
",",
"pb",
".",
"ETA",
"(",
")",
",",
"' '",
",",
"pb",
".",
"FileTransferSpeed",
"(",
")",
"]",
"pbar",
"=",
"pb",
".",
"ProgressBar",
"(",
"widgets",
"=",
"widgets",
",",
"maxval",
"=",
"max_value",
")",
".",
"start",
"(",
")",
"with",
"open",
"(",
"tmp_file",
",",
"'wb'",
")",
"as",
"f",
":",
"for",
"chunk",
"in",
"r",
".",
"iter_content",
"(",
"CHUNK_SIZE",
")",
":",
"f",
".",
"write",
"(",
"chunk",
")",
"bytes_downloaded",
"+=",
"CHUNK_SIZE",
"if",
"log_level",
"<=",
"logging",
".",
"INFO",
"and",
"content_length",
":",
"pbar",
".",
"update",
"(",
"bytes_downloaded",
")",
"t1",
"=",
"total_seconds",
"(",
"datetime",
".",
"now",
"(",
")",
"-",
"start_time",
")",
"if",
"self",
".",
"timeout_download",
"and",
"t1",
">=",
"self",
".",
"timeout_download",
":",
"raise",
"errors",
".",
"TimeoutError",
"if",
"log_level",
"<=",
"logging",
".",
"INFO",
"and",
"content_length",
":",
"pbar",
".",
"finish",
"(",
")",
"except",
"Exception",
":",
"if",
"os",
".",
"path",
".",
"isfile",
"(",
"tmp_file",
")",
":",
"os",
".",
"remove",
"(",
"tmp_file",
")",
"raise",
"self",
".",
"_retry",
"(",
"_download",
",",
"retry_exceptions",
"=",
"(",
"requests",
".",
"exceptions",
".",
"RequestException",
",",
"errors",
".",
"TimeoutError",
")",
")",
"os",
".",
"rename",
"(",
"tmp_file",
",",
"self",
".",
"filename",
")",
"return",
"self",
".",
"filename"
] | Download the specified file. | [
"Download",
"the",
"specified",
"file",
"."
] | 97796a028455bb5200434562d23b66d5a5eb537b | https://github.com/mozilla/mozdownload/blob/97796a028455bb5200434562d23b66d5a5eb537b/mozdownload/scraper.py#L249-L324 | train |
mozilla/mozdownload | mozdownload/scraper.py | Scraper.show_matching_builds | def show_matching_builds(self, builds):
"""Output the matching builds."""
self.logger.info('Found %s build%s: %s' % (
len(builds),
len(builds) > 1 and 's' or '',
len(builds) > 10 and
' ... '.join([', '.join(builds[:5]), ', '.join(builds[-5:])]) or
', '.join(builds))) | python | def show_matching_builds(self, builds):
"""Output the matching builds."""
self.logger.info('Found %s build%s: %s' % (
len(builds),
len(builds) > 1 and 's' or '',
len(builds) > 10 and
' ... '.join([', '.join(builds[:5]), ', '.join(builds[-5:])]) or
', '.join(builds))) | [
"def",
"show_matching_builds",
"(",
"self",
",",
"builds",
")",
":",
"self",
".",
"logger",
".",
"info",
"(",
"'Found %s build%s: %s'",
"%",
"(",
"len",
"(",
"builds",
")",
",",
"len",
"(",
"builds",
")",
">",
"1",
"and",
"'s'",
"or",
"''",
",",
"len",
"(",
"builds",
")",
">",
"10",
"and",
"' ... '",
".",
"join",
"(",
"[",
"', '",
".",
"join",
"(",
"builds",
"[",
":",
"5",
"]",
")",
",",
"', '",
".",
"join",
"(",
"builds",
"[",
"-",
"5",
":",
"]",
")",
"]",
")",
"or",
"', '",
".",
"join",
"(",
"builds",
")",
")",
")"
] | Output the matching builds. | [
"Output",
"the",
"matching",
"builds",
"."
] | 97796a028455bb5200434562d23b66d5a5eb537b | https://github.com/mozilla/mozdownload/blob/97796a028455bb5200434562d23b66d5a5eb537b/mozdownload/scraper.py#L326-L333 | train |
mozilla/mozdownload | mozdownload/scraper.py | DailyScraper.is_build_dir | def is_build_dir(self, folder_name):
"""Return whether or not the given dir contains a build."""
# Cannot move up to base scraper due to parser.entries call in
# get_build_info_for_date (see below)
url = '%s/' % urljoin(self.base_url, self.monthly_build_list_regex, folder_name)
if self.application in APPLICATIONS_MULTI_LOCALE \
and self.locale != 'multi':
url = '%s/' % urljoin(url, self.locale)
parser = self._create_directory_parser(url)
pattern = re.compile(self.binary_regex, re.IGNORECASE)
for entry in parser.entries:
try:
pattern.match(entry).group()
return True
except Exception:
# No match, continue with next entry
continue
return False | python | def is_build_dir(self, folder_name):
"""Return whether or not the given dir contains a build."""
# Cannot move up to base scraper due to parser.entries call in
# get_build_info_for_date (see below)
url = '%s/' % urljoin(self.base_url, self.monthly_build_list_regex, folder_name)
if self.application in APPLICATIONS_MULTI_LOCALE \
and self.locale != 'multi':
url = '%s/' % urljoin(url, self.locale)
parser = self._create_directory_parser(url)
pattern = re.compile(self.binary_regex, re.IGNORECASE)
for entry in parser.entries:
try:
pattern.match(entry).group()
return True
except Exception:
# No match, continue with next entry
continue
return False | [
"def",
"is_build_dir",
"(",
"self",
",",
"folder_name",
")",
":",
"# Cannot move up to base scraper due to parser.entries call in",
"# get_build_info_for_date (see below)",
"url",
"=",
"'%s/'",
"%",
"urljoin",
"(",
"self",
".",
"base_url",
",",
"self",
".",
"monthly_build_list_regex",
",",
"folder_name",
")",
"if",
"self",
".",
"application",
"in",
"APPLICATIONS_MULTI_LOCALE",
"and",
"self",
".",
"locale",
"!=",
"'multi'",
":",
"url",
"=",
"'%s/'",
"%",
"urljoin",
"(",
"url",
",",
"self",
".",
"locale",
")",
"parser",
"=",
"self",
".",
"_create_directory_parser",
"(",
"url",
")",
"pattern",
"=",
"re",
".",
"compile",
"(",
"self",
".",
"binary_regex",
",",
"re",
".",
"IGNORECASE",
")",
"for",
"entry",
"in",
"parser",
".",
"entries",
":",
"try",
":",
"pattern",
".",
"match",
"(",
"entry",
")",
".",
"group",
"(",
")",
"return",
"True",
"except",
"Exception",
":",
"# No match, continue with next entry",
"continue",
"return",
"False"
] | Return whether or not the given dir contains a build. | [
"Return",
"whether",
"or",
"not",
"the",
"given",
"dir",
"contains",
"a",
"build",
"."
] | 97796a028455bb5200434562d23b66d5a5eb537b | https://github.com/mozilla/mozdownload/blob/97796a028455bb5200434562d23b66d5a5eb537b/mozdownload/scraper.py#L435-L455 | train |
mozilla/mozdownload | mozdownload/scraper.py | DailyScraper.get_build_info_for_date | def get_build_info_for_date(self, date, build_index=None):
"""Return the build information for a given date."""
url = urljoin(self.base_url, self.monthly_build_list_regex)
has_time = date and date.time()
self.logger.info('Retrieving list of builds from %s' % url)
parser = self._create_directory_parser(url)
regex = r'%(DATE)s-(\d+-)+%(BRANCH)s%(L10N)s%(PLATFORM)s$' % {
'DATE': date.strftime('%Y-%m-%d'),
'BRANCH': self.branch,
# ensure to select the correct subfolder for localized builds
'L10N': '(-l10n)?' if self.locale_build else '',
'PLATFORM': '' if self.application not in (
'fennec') else '-' + self.platform
}
parser.entries = parser.filter(regex)
parser.entries = parser.filter(self.is_build_dir)
if has_time:
# If a time is included in the date, use it to determine the
# build's index
regex = r'.*%s.*' % date.strftime('%H-%M-%S')
parser.entries = parser.filter(regex)
if not parser.entries:
date_format = '%Y-%m-%d-%H-%M-%S' if has_time else '%Y-%m-%d'
message = 'Folder for builds on %s has not been found' % \
self.date.strftime(date_format)
raise errors.NotFoundError(message, url)
# If no index has been given, set it to the last build of the day.
self.show_matching_builds(parser.entries)
# If no index has been given, set it to the last build of the day.
if build_index is None:
# Find the most recent non-empty entry.
build_index = len(parser.entries)
for build in reversed(parser.entries):
build_index -= 1
if not build_index or self.is_build_dir(build):
break
self.logger.info('Selected build: %s' % parser.entries[build_index])
return (parser.entries, build_index) | python | def get_build_info_for_date(self, date, build_index=None):
"""Return the build information for a given date."""
url = urljoin(self.base_url, self.monthly_build_list_regex)
has_time = date and date.time()
self.logger.info('Retrieving list of builds from %s' % url)
parser = self._create_directory_parser(url)
regex = r'%(DATE)s-(\d+-)+%(BRANCH)s%(L10N)s%(PLATFORM)s$' % {
'DATE': date.strftime('%Y-%m-%d'),
'BRANCH': self.branch,
# ensure to select the correct subfolder for localized builds
'L10N': '(-l10n)?' if self.locale_build else '',
'PLATFORM': '' if self.application not in (
'fennec') else '-' + self.platform
}
parser.entries = parser.filter(regex)
parser.entries = parser.filter(self.is_build_dir)
if has_time:
# If a time is included in the date, use it to determine the
# build's index
regex = r'.*%s.*' % date.strftime('%H-%M-%S')
parser.entries = parser.filter(regex)
if not parser.entries:
date_format = '%Y-%m-%d-%H-%M-%S' if has_time else '%Y-%m-%d'
message = 'Folder for builds on %s has not been found' % \
self.date.strftime(date_format)
raise errors.NotFoundError(message, url)
# If no index has been given, set it to the last build of the day.
self.show_matching_builds(parser.entries)
# If no index has been given, set it to the last build of the day.
if build_index is None:
# Find the most recent non-empty entry.
build_index = len(parser.entries)
for build in reversed(parser.entries):
build_index -= 1
if not build_index or self.is_build_dir(build):
break
self.logger.info('Selected build: %s' % parser.entries[build_index])
return (parser.entries, build_index) | [
"def",
"get_build_info_for_date",
"(",
"self",
",",
"date",
",",
"build_index",
"=",
"None",
")",
":",
"url",
"=",
"urljoin",
"(",
"self",
".",
"base_url",
",",
"self",
".",
"monthly_build_list_regex",
")",
"has_time",
"=",
"date",
"and",
"date",
".",
"time",
"(",
")",
"self",
".",
"logger",
".",
"info",
"(",
"'Retrieving list of builds from %s'",
"%",
"url",
")",
"parser",
"=",
"self",
".",
"_create_directory_parser",
"(",
"url",
")",
"regex",
"=",
"r'%(DATE)s-(\\d+-)+%(BRANCH)s%(L10N)s%(PLATFORM)s$'",
"%",
"{",
"'DATE'",
":",
"date",
".",
"strftime",
"(",
"'%Y-%m-%d'",
")",
",",
"'BRANCH'",
":",
"self",
".",
"branch",
",",
"# ensure to select the correct subfolder for localized builds",
"'L10N'",
":",
"'(-l10n)?'",
"if",
"self",
".",
"locale_build",
"else",
"''",
",",
"'PLATFORM'",
":",
"''",
"if",
"self",
".",
"application",
"not",
"in",
"(",
"'fennec'",
")",
"else",
"'-'",
"+",
"self",
".",
"platform",
"}",
"parser",
".",
"entries",
"=",
"parser",
".",
"filter",
"(",
"regex",
")",
"parser",
".",
"entries",
"=",
"parser",
".",
"filter",
"(",
"self",
".",
"is_build_dir",
")",
"if",
"has_time",
":",
"# If a time is included in the date, use it to determine the",
"# build's index",
"regex",
"=",
"r'.*%s.*'",
"%",
"date",
".",
"strftime",
"(",
"'%H-%M-%S'",
")",
"parser",
".",
"entries",
"=",
"parser",
".",
"filter",
"(",
"regex",
")",
"if",
"not",
"parser",
".",
"entries",
":",
"date_format",
"=",
"'%Y-%m-%d-%H-%M-%S'",
"if",
"has_time",
"else",
"'%Y-%m-%d'",
"message",
"=",
"'Folder for builds on %s has not been found'",
"%",
"self",
".",
"date",
".",
"strftime",
"(",
"date_format",
")",
"raise",
"errors",
".",
"NotFoundError",
"(",
"message",
",",
"url",
")",
"# If no index has been given, set it to the last build of the day.",
"self",
".",
"show_matching_builds",
"(",
"parser",
".",
"entries",
")",
"# If no index has been given, set it to the last build of the day.",
"if",
"build_index",
"is",
"None",
":",
"# Find the most recent non-empty entry.",
"build_index",
"=",
"len",
"(",
"parser",
".",
"entries",
")",
"for",
"build",
"in",
"reversed",
"(",
"parser",
".",
"entries",
")",
":",
"build_index",
"-=",
"1",
"if",
"not",
"build_index",
"or",
"self",
".",
"is_build_dir",
"(",
"build",
")",
":",
"break",
"self",
".",
"logger",
".",
"info",
"(",
"'Selected build: %s'",
"%",
"parser",
".",
"entries",
"[",
"build_index",
"]",
")",
"return",
"(",
"parser",
".",
"entries",
",",
"build_index",
")"
] | Return the build information for a given date. | [
"Return",
"the",
"build",
"information",
"for",
"a",
"given",
"date",
"."
] | 97796a028455bb5200434562d23b66d5a5eb537b | https://github.com/mozilla/mozdownload/blob/97796a028455bb5200434562d23b66d5a5eb537b/mozdownload/scraper.py#L457-L500 | train |
mozilla/mozdownload | mozdownload/scraper.py | DailyScraper.monthly_build_list_regex | def monthly_build_list_regex(self):
"""Return the regex for the folder containing builds of a month."""
# Regex for possible builds for the given date
return r'nightly/%(YEAR)s/%(MONTH)s/' % {
'YEAR': self.date.year,
'MONTH': str(self.date.month).zfill(2)} | python | def monthly_build_list_regex(self):
"""Return the regex for the folder containing builds of a month."""
# Regex for possible builds for the given date
return r'nightly/%(YEAR)s/%(MONTH)s/' % {
'YEAR': self.date.year,
'MONTH': str(self.date.month).zfill(2)} | [
"def",
"monthly_build_list_regex",
"(",
"self",
")",
":",
"# Regex for possible builds for the given date",
"return",
"r'nightly/%(YEAR)s/%(MONTH)s/'",
"%",
"{",
"'YEAR'",
":",
"self",
".",
"date",
".",
"year",
",",
"'MONTH'",
":",
"str",
"(",
"self",
".",
"date",
".",
"month",
")",
".",
"zfill",
"(",
"2",
")",
"}"
] | Return the regex for the folder containing builds of a month. | [
"Return",
"the",
"regex",
"for",
"the",
"folder",
"containing",
"builds",
"of",
"a",
"month",
"."
] | 97796a028455bb5200434562d23b66d5a5eb537b | https://github.com/mozilla/mozdownload/blob/97796a028455bb5200434562d23b66d5a5eb537b/mozdownload/scraper.py#L544-L549 | train |
mozilla/mozdownload | mozdownload/scraper.py | DirectScraper.filename | def filename(self):
"""File name of the downloaded file."""
if os.path.splitext(self.destination)[1]:
# If the filename has been given make use of it
target_file = self.destination
else:
# Otherwise determine it from the url.
parsed_url = urlparse(self.url)
source_filename = (parsed_url.path.rpartition('/')[-1] or
parsed_url.hostname)
target_file = os.path.join(self.destination, source_filename)
return os.path.abspath(target_file) | python | def filename(self):
"""File name of the downloaded file."""
if os.path.splitext(self.destination)[1]:
# If the filename has been given make use of it
target_file = self.destination
else:
# Otherwise determine it from the url.
parsed_url = urlparse(self.url)
source_filename = (parsed_url.path.rpartition('/')[-1] or
parsed_url.hostname)
target_file = os.path.join(self.destination, source_filename)
return os.path.abspath(target_file) | [
"def",
"filename",
"(",
"self",
")",
":",
"if",
"os",
".",
"path",
".",
"splitext",
"(",
"self",
".",
"destination",
")",
"[",
"1",
"]",
":",
"# If the filename has been given make use of it",
"target_file",
"=",
"self",
".",
"destination",
"else",
":",
"# Otherwise determine it from the url.",
"parsed_url",
"=",
"urlparse",
"(",
"self",
".",
"url",
")",
"source_filename",
"=",
"(",
"parsed_url",
".",
"path",
".",
"rpartition",
"(",
"'/'",
")",
"[",
"-",
"1",
"]",
"or",
"parsed_url",
".",
"hostname",
")",
"target_file",
"=",
"os",
".",
"path",
".",
"join",
"(",
"self",
".",
"destination",
",",
"source_filename",
")",
"return",
"os",
".",
"path",
".",
"abspath",
"(",
"target_file",
")"
] | File name of the downloaded file. | [
"File",
"name",
"of",
"the",
"downloaded",
"file",
"."
] | 97796a028455bb5200434562d23b66d5a5eb537b | https://github.com/mozilla/mozdownload/blob/97796a028455bb5200434562d23b66d5a5eb537b/mozdownload/scraper.py#L577-L589 | train |
mozilla/mozdownload | mozdownload/scraper.py | ReleaseScraper.query_versions | def query_versions(self, version=None):
"""Check specified version and resolve special values."""
if version not in RELEASE_AND_CANDIDATE_LATEST_VERSIONS:
return [version]
url = urljoin(self.base_url, 'releases/')
parser = self._create_directory_parser(url)
if version:
versions = parser.filter(RELEASE_AND_CANDIDATE_LATEST_VERSIONS[version])
from distutils.version import LooseVersion
versions.sort(key=LooseVersion)
return [versions[-1]]
else:
return parser.entries | python | def query_versions(self, version=None):
"""Check specified version and resolve special values."""
if version not in RELEASE_AND_CANDIDATE_LATEST_VERSIONS:
return [version]
url = urljoin(self.base_url, 'releases/')
parser = self._create_directory_parser(url)
if version:
versions = parser.filter(RELEASE_AND_CANDIDATE_LATEST_VERSIONS[version])
from distutils.version import LooseVersion
versions.sort(key=LooseVersion)
return [versions[-1]]
else:
return parser.entries | [
"def",
"query_versions",
"(",
"self",
",",
"version",
"=",
"None",
")",
":",
"if",
"version",
"not",
"in",
"RELEASE_AND_CANDIDATE_LATEST_VERSIONS",
":",
"return",
"[",
"version",
"]",
"url",
"=",
"urljoin",
"(",
"self",
".",
"base_url",
",",
"'releases/'",
")",
"parser",
"=",
"self",
".",
"_create_directory_parser",
"(",
"url",
")",
"if",
"version",
":",
"versions",
"=",
"parser",
".",
"filter",
"(",
"RELEASE_AND_CANDIDATE_LATEST_VERSIONS",
"[",
"version",
"]",
")",
"from",
"distutils",
".",
"version",
"import",
"LooseVersion",
"versions",
".",
"sort",
"(",
"key",
"=",
"LooseVersion",
")",
"return",
"[",
"versions",
"[",
"-",
"1",
"]",
"]",
"else",
":",
"return",
"parser",
".",
"entries"
] | Check specified version and resolve special values. | [
"Check",
"specified",
"version",
"and",
"resolve",
"special",
"values",
"."
] | 97796a028455bb5200434562d23b66d5a5eb537b | https://github.com/mozilla/mozdownload/blob/97796a028455bb5200434562d23b66d5a5eb537b/mozdownload/scraper.py#L657-L670 | train |
mozilla/mozdownload | mozdownload/scraper.py | TinderboxScraper.build_list_regex | def build_list_regex(self):
"""Return the regex for the folder which contains the list of builds."""
regex = 'tinderbox-builds/%(BRANCH)s-%(PLATFORM)s%(L10N)s%(DEBUG)s/'
return regex % {
'BRANCH': self.branch,
'PLATFORM': '' if self.locale_build else self.platform_regex,
'L10N': 'l10n' if self.locale_build else '',
'DEBUG': '-debug' if self.debug_build else ''} | python | def build_list_regex(self):
"""Return the regex for the folder which contains the list of builds."""
regex = 'tinderbox-builds/%(BRANCH)s-%(PLATFORM)s%(L10N)s%(DEBUG)s/'
return regex % {
'BRANCH': self.branch,
'PLATFORM': '' if self.locale_build else self.platform_regex,
'L10N': 'l10n' if self.locale_build else '',
'DEBUG': '-debug' if self.debug_build else ''} | [
"def",
"build_list_regex",
"(",
"self",
")",
":",
"regex",
"=",
"'tinderbox-builds/%(BRANCH)s-%(PLATFORM)s%(L10N)s%(DEBUG)s/'",
"return",
"regex",
"%",
"{",
"'BRANCH'",
":",
"self",
".",
"branch",
",",
"'PLATFORM'",
":",
"''",
"if",
"self",
".",
"locale_build",
"else",
"self",
".",
"platform_regex",
",",
"'L10N'",
":",
"'l10n'",
"if",
"self",
".",
"locale_build",
"else",
"''",
",",
"'DEBUG'",
":",
"'-debug'",
"if",
"self",
".",
"debug_build",
"else",
"''",
"}"
] | Return the regex for the folder which contains the list of builds. | [
"Return",
"the",
"regex",
"for",
"the",
"folder",
"which",
"contains",
"the",
"list",
"of",
"builds",
"."
] | 97796a028455bb5200434562d23b66d5a5eb537b | https://github.com/mozilla/mozdownload/blob/97796a028455bb5200434562d23b66d5a5eb537b/mozdownload/scraper.py#L845-L853 | train |
mozilla/mozdownload | mozdownload/scraper.py | TinderboxScraper.date_matches | def date_matches(self, timestamp):
"""Determine whether the timestamp date is equal to the argument date."""
if self.date is None:
return False
timestamp = datetime.fromtimestamp(float(timestamp), self.timezone)
if self.date.date() == timestamp.date():
return True
return False | python | def date_matches(self, timestamp):
"""Determine whether the timestamp date is equal to the argument date."""
if self.date is None:
return False
timestamp = datetime.fromtimestamp(float(timestamp), self.timezone)
if self.date.date() == timestamp.date():
return True
return False | [
"def",
"date_matches",
"(",
"self",
",",
"timestamp",
")",
":",
"if",
"self",
".",
"date",
"is",
"None",
":",
"return",
"False",
"timestamp",
"=",
"datetime",
".",
"fromtimestamp",
"(",
"float",
"(",
"timestamp",
")",
",",
"self",
".",
"timezone",
")",
"if",
"self",
".",
"date",
".",
"date",
"(",
")",
"==",
"timestamp",
".",
"date",
"(",
")",
":",
"return",
"True",
"return",
"False"
] | Determine whether the timestamp date is equal to the argument date. | [
"Determine",
"whether",
"the",
"timestamp",
"date",
"is",
"equal",
"to",
"the",
"argument",
"date",
"."
] | 97796a028455bb5200434562d23b66d5a5eb537b | https://github.com/mozilla/mozdownload/blob/97796a028455bb5200434562d23b66d5a5eb537b/mozdownload/scraper.py#L855-L864 | train |
mozilla/mozdownload | mozdownload/scraper.py | TinderboxScraper.get_build_info_for_index | def get_build_info_for_index(self, build_index=None):
"""Get additional information for the build at the given index."""
url = urljoin(self.base_url, self.build_list_regex)
self.logger.info('Retrieving list of builds from %s' % url)
parser = self._create_directory_parser(url)
parser.entries = parser.filter(r'^\d+$')
if self.timestamp:
# If a timestamp is given, retrieve the folder with the timestamp
# as name
parser.entries = self.timestamp in parser.entries and \
[self.timestamp]
elif self.date:
# If date is given, retrieve the subset of builds on that date
parser.entries = filter(self.date_matches, parser.entries)
if not parser.entries:
message = 'No builds have been found'
raise errors.NotFoundError(message, url)
self.show_matching_builds(parser.entries)
# If no index has been given, set it to the last build of the day.
if build_index is None:
# Find the most recent non-empty entry.
build_index = len(parser.entries)
for build in reversed(parser.entries):
build_index -= 1
if not build_index or self.is_build_dir(build):
break
self.logger.info('Selected build: %s' % parser.entries[build_index])
return (parser.entries, build_index) | python | def get_build_info_for_index(self, build_index=None):
"""Get additional information for the build at the given index."""
url = urljoin(self.base_url, self.build_list_regex)
self.logger.info('Retrieving list of builds from %s' % url)
parser = self._create_directory_parser(url)
parser.entries = parser.filter(r'^\d+$')
if self.timestamp:
# If a timestamp is given, retrieve the folder with the timestamp
# as name
parser.entries = self.timestamp in parser.entries and \
[self.timestamp]
elif self.date:
# If date is given, retrieve the subset of builds on that date
parser.entries = filter(self.date_matches, parser.entries)
if not parser.entries:
message = 'No builds have been found'
raise errors.NotFoundError(message, url)
self.show_matching_builds(parser.entries)
# If no index has been given, set it to the last build of the day.
if build_index is None:
# Find the most recent non-empty entry.
build_index = len(parser.entries)
for build in reversed(parser.entries):
build_index -= 1
if not build_index or self.is_build_dir(build):
break
self.logger.info('Selected build: %s' % parser.entries[build_index])
return (parser.entries, build_index) | [
"def",
"get_build_info_for_index",
"(",
"self",
",",
"build_index",
"=",
"None",
")",
":",
"url",
"=",
"urljoin",
"(",
"self",
".",
"base_url",
",",
"self",
".",
"build_list_regex",
")",
"self",
".",
"logger",
".",
"info",
"(",
"'Retrieving list of builds from %s'",
"%",
"url",
")",
"parser",
"=",
"self",
".",
"_create_directory_parser",
"(",
"url",
")",
"parser",
".",
"entries",
"=",
"parser",
".",
"filter",
"(",
"r'^\\d+$'",
")",
"if",
"self",
".",
"timestamp",
":",
"# If a timestamp is given, retrieve the folder with the timestamp",
"# as name",
"parser",
".",
"entries",
"=",
"self",
".",
"timestamp",
"in",
"parser",
".",
"entries",
"and",
"[",
"self",
".",
"timestamp",
"]",
"elif",
"self",
".",
"date",
":",
"# If date is given, retrieve the subset of builds on that date",
"parser",
".",
"entries",
"=",
"filter",
"(",
"self",
".",
"date_matches",
",",
"parser",
".",
"entries",
")",
"if",
"not",
"parser",
".",
"entries",
":",
"message",
"=",
"'No builds have been found'",
"raise",
"errors",
".",
"NotFoundError",
"(",
"message",
",",
"url",
")",
"self",
".",
"show_matching_builds",
"(",
"parser",
".",
"entries",
")",
"# If no index has been given, set it to the last build of the day.",
"if",
"build_index",
"is",
"None",
":",
"# Find the most recent non-empty entry.",
"build_index",
"=",
"len",
"(",
"parser",
".",
"entries",
")",
"for",
"build",
"in",
"reversed",
"(",
"parser",
".",
"entries",
")",
":",
"build_index",
"-=",
"1",
"if",
"not",
"build_index",
"or",
"self",
".",
"is_build_dir",
"(",
"build",
")",
":",
"break",
"self",
".",
"logger",
".",
"info",
"(",
"'Selected build: %s'",
"%",
"parser",
".",
"entries",
"[",
"build_index",
"]",
")",
"return",
"(",
"parser",
".",
"entries",
",",
"build_index",
")"
] | Get additional information for the build at the given index. | [
"Get",
"additional",
"information",
"for",
"the",
"build",
"at",
"the",
"given",
"index",
"."
] | 97796a028455bb5200434562d23b66d5a5eb537b | https://github.com/mozilla/mozdownload/blob/97796a028455bb5200434562d23b66d5a5eb537b/mozdownload/scraper.py#L900-L935 | train |
romanz/trezor-agent | libagent/device/ui.py | create_default_options_getter | def create_default_options_getter():
"""Return current TTY and DISPLAY settings for GnuPG pinentry."""
options = []
try:
ttyname = subprocess.check_output(args=['tty']).strip()
options.append(b'ttyname=' + ttyname)
except subprocess.CalledProcessError as e:
log.warning('no TTY found: %s', e)
display = os.environ.get('DISPLAY')
if display is not None:
options.append('display={}'.format(display).encode('ascii'))
else:
log.warning('DISPLAY not defined')
log.info('using %s for pinentry options', options)
return lambda: options | python | def create_default_options_getter():
"""Return current TTY and DISPLAY settings for GnuPG pinentry."""
options = []
try:
ttyname = subprocess.check_output(args=['tty']).strip()
options.append(b'ttyname=' + ttyname)
except subprocess.CalledProcessError as e:
log.warning('no TTY found: %s', e)
display = os.environ.get('DISPLAY')
if display is not None:
options.append('display={}'.format(display).encode('ascii'))
else:
log.warning('DISPLAY not defined')
log.info('using %s for pinentry options', options)
return lambda: options | [
"def",
"create_default_options_getter",
"(",
")",
":",
"options",
"=",
"[",
"]",
"try",
":",
"ttyname",
"=",
"subprocess",
".",
"check_output",
"(",
"args",
"=",
"[",
"'tty'",
"]",
")",
".",
"strip",
"(",
")",
"options",
".",
"append",
"(",
"b'ttyname='",
"+",
"ttyname",
")",
"except",
"subprocess",
".",
"CalledProcessError",
"as",
"e",
":",
"log",
".",
"warning",
"(",
"'no TTY found: %s'",
",",
"e",
")",
"display",
"=",
"os",
".",
"environ",
".",
"get",
"(",
"'DISPLAY'",
")",
"if",
"display",
"is",
"not",
"None",
":",
"options",
".",
"append",
"(",
"'display={}'",
".",
"format",
"(",
"display",
")",
".",
"encode",
"(",
"'ascii'",
")",
")",
"else",
":",
"log",
".",
"warning",
"(",
"'DISPLAY not defined'",
")",
"log",
".",
"info",
"(",
"'using %s for pinentry options'",
",",
"options",
")",
"return",
"lambda",
":",
"options"
] | Return current TTY and DISPLAY settings for GnuPG pinentry. | [
"Return",
"current",
"TTY",
"and",
"DISPLAY",
"settings",
"for",
"GnuPG",
"pinentry",
"."
] | 513b1259c4d7aca5f88cd958edc11828d0712f1b | https://github.com/romanz/trezor-agent/blob/513b1259c4d7aca5f88cd958edc11828d0712f1b/libagent/device/ui.py#L64-L80 | train |
romanz/trezor-agent | libagent/device/ui.py | write | def write(p, line):
"""Send and flush a single line to the subprocess' stdin."""
log.debug('%s <- %r', p.args, line)
p.stdin.write(line)
p.stdin.flush() | python | def write(p, line):
"""Send and flush a single line to the subprocess' stdin."""
log.debug('%s <- %r', p.args, line)
p.stdin.write(line)
p.stdin.flush() | [
"def",
"write",
"(",
"p",
",",
"line",
")",
":",
"log",
".",
"debug",
"(",
"'%s <- %r'",
",",
"p",
".",
"args",
",",
"line",
")",
"p",
".",
"stdin",
".",
"write",
"(",
"line",
")",
"p",
".",
"stdin",
".",
"flush",
"(",
")"
] | Send and flush a single line to the subprocess' stdin. | [
"Send",
"and",
"flush",
"a",
"single",
"line",
"to",
"the",
"subprocess",
"stdin",
"."
] | 513b1259c4d7aca5f88cd958edc11828d0712f1b | https://github.com/romanz/trezor-agent/blob/513b1259c4d7aca5f88cd958edc11828d0712f1b/libagent/device/ui.py#L83-L87 | train |
romanz/trezor-agent | libagent/device/ui.py | expect | def expect(p, prefixes, confidential=False):
"""Read a line and return it without required prefix."""
resp = p.stdout.readline()
log.debug('%s -> %r', p.args, resp if not confidential else '********')
for prefix in prefixes:
if resp.startswith(prefix):
return resp[len(prefix):]
raise UnexpectedError(resp) | python | def expect(p, prefixes, confidential=False):
"""Read a line and return it without required prefix."""
resp = p.stdout.readline()
log.debug('%s -> %r', p.args, resp if not confidential else '********')
for prefix in prefixes:
if resp.startswith(prefix):
return resp[len(prefix):]
raise UnexpectedError(resp) | [
"def",
"expect",
"(",
"p",
",",
"prefixes",
",",
"confidential",
"=",
"False",
")",
":",
"resp",
"=",
"p",
".",
"stdout",
".",
"readline",
"(",
")",
"log",
".",
"debug",
"(",
"'%s -> %r'",
",",
"p",
".",
"args",
",",
"resp",
"if",
"not",
"confidential",
"else",
"'********'",
")",
"for",
"prefix",
"in",
"prefixes",
":",
"if",
"resp",
".",
"startswith",
"(",
"prefix",
")",
":",
"return",
"resp",
"[",
"len",
"(",
"prefix",
")",
":",
"]",
"raise",
"UnexpectedError",
"(",
"resp",
")"
] | Read a line and return it without required prefix. | [
"Read",
"a",
"line",
"and",
"return",
"it",
"without",
"required",
"prefix",
"."
] | 513b1259c4d7aca5f88cd958edc11828d0712f1b | https://github.com/romanz/trezor-agent/blob/513b1259c4d7aca5f88cd958edc11828d0712f1b/libagent/device/ui.py#L94-L101 | train |
romanz/trezor-agent | libagent/device/ui.py | interact | def interact(title, description, prompt, binary, options):
"""Use GPG pinentry program to interact with the user."""
args = [binary]
p = subprocess.Popen(args=args,
stdin=subprocess.PIPE,
stdout=subprocess.PIPE,
env=os.environ)
p.args = args # TODO: remove after Python 2 deprecation.
expect(p, [b'OK'])
title = util.assuan_serialize(title.encode('ascii'))
write(p, b'SETTITLE ' + title + b'\n')
expect(p, [b'OK'])
if description:
description = util.assuan_serialize(description.encode('ascii'))
write(p, b'SETDESC ' + description + b'\n')
expect(p, [b'OK'])
if prompt:
prompt = util.assuan_serialize(prompt.encode('ascii'))
write(p, b'SETPROMPT ' + prompt + b'\n')
expect(p, [b'OK'])
log.debug('setting %d options', len(options))
for opt in options:
write(p, b'OPTION ' + opt + b'\n')
expect(p, [b'OK', b'ERR'])
write(p, b'GETPIN\n')
pin = expect(p, [b'OK', b'D '], confidential=True)
p.communicate() # close stdin and wait for the process to exit
exit_code = p.wait()
if exit_code:
raise subprocess.CalledProcessError(exit_code, binary)
return pin.decode('ascii').strip() | python | def interact(title, description, prompt, binary, options):
"""Use GPG pinentry program to interact with the user."""
args = [binary]
p = subprocess.Popen(args=args,
stdin=subprocess.PIPE,
stdout=subprocess.PIPE,
env=os.environ)
p.args = args # TODO: remove after Python 2 deprecation.
expect(p, [b'OK'])
title = util.assuan_serialize(title.encode('ascii'))
write(p, b'SETTITLE ' + title + b'\n')
expect(p, [b'OK'])
if description:
description = util.assuan_serialize(description.encode('ascii'))
write(p, b'SETDESC ' + description + b'\n')
expect(p, [b'OK'])
if prompt:
prompt = util.assuan_serialize(prompt.encode('ascii'))
write(p, b'SETPROMPT ' + prompt + b'\n')
expect(p, [b'OK'])
log.debug('setting %d options', len(options))
for opt in options:
write(p, b'OPTION ' + opt + b'\n')
expect(p, [b'OK', b'ERR'])
write(p, b'GETPIN\n')
pin = expect(p, [b'OK', b'D '], confidential=True)
p.communicate() # close stdin and wait for the process to exit
exit_code = p.wait()
if exit_code:
raise subprocess.CalledProcessError(exit_code, binary)
return pin.decode('ascii').strip() | [
"def",
"interact",
"(",
"title",
",",
"description",
",",
"prompt",
",",
"binary",
",",
"options",
")",
":",
"args",
"=",
"[",
"binary",
"]",
"p",
"=",
"subprocess",
".",
"Popen",
"(",
"args",
"=",
"args",
",",
"stdin",
"=",
"subprocess",
".",
"PIPE",
",",
"stdout",
"=",
"subprocess",
".",
"PIPE",
",",
"env",
"=",
"os",
".",
"environ",
")",
"p",
".",
"args",
"=",
"args",
"# TODO: remove after Python 2 deprecation.",
"expect",
"(",
"p",
",",
"[",
"b'OK'",
"]",
")",
"title",
"=",
"util",
".",
"assuan_serialize",
"(",
"title",
".",
"encode",
"(",
"'ascii'",
")",
")",
"write",
"(",
"p",
",",
"b'SETTITLE '",
"+",
"title",
"+",
"b'\\n'",
")",
"expect",
"(",
"p",
",",
"[",
"b'OK'",
"]",
")",
"if",
"description",
":",
"description",
"=",
"util",
".",
"assuan_serialize",
"(",
"description",
".",
"encode",
"(",
"'ascii'",
")",
")",
"write",
"(",
"p",
",",
"b'SETDESC '",
"+",
"description",
"+",
"b'\\n'",
")",
"expect",
"(",
"p",
",",
"[",
"b'OK'",
"]",
")",
"if",
"prompt",
":",
"prompt",
"=",
"util",
".",
"assuan_serialize",
"(",
"prompt",
".",
"encode",
"(",
"'ascii'",
")",
")",
"write",
"(",
"p",
",",
"b'SETPROMPT '",
"+",
"prompt",
"+",
"b'\\n'",
")",
"expect",
"(",
"p",
",",
"[",
"b'OK'",
"]",
")",
"log",
".",
"debug",
"(",
"'setting %d options'",
",",
"len",
"(",
"options",
")",
")",
"for",
"opt",
"in",
"options",
":",
"write",
"(",
"p",
",",
"b'OPTION '",
"+",
"opt",
"+",
"b'\\n'",
")",
"expect",
"(",
"p",
",",
"[",
"b'OK'",
",",
"b'ERR'",
"]",
")",
"write",
"(",
"p",
",",
"b'GETPIN\\n'",
")",
"pin",
"=",
"expect",
"(",
"p",
",",
"[",
"b'OK'",
",",
"b'D '",
"]",
",",
"confidential",
"=",
"True",
")",
"p",
".",
"communicate",
"(",
")",
"# close stdin and wait for the process to exit",
"exit_code",
"=",
"p",
".",
"wait",
"(",
")",
"if",
"exit_code",
":",
"raise",
"subprocess",
".",
"CalledProcessError",
"(",
"exit_code",
",",
"binary",
")",
"return",
"pin",
".",
"decode",
"(",
"'ascii'",
")",
".",
"strip",
"(",
")"
] | Use GPG pinentry program to interact with the user. | [
"Use",
"GPG",
"pinentry",
"program",
"to",
"interact",
"with",
"the",
"user",
"."
] | 513b1259c4d7aca5f88cd958edc11828d0712f1b | https://github.com/romanz/trezor-agent/blob/513b1259c4d7aca5f88cd958edc11828d0712f1b/libagent/device/ui.py#L104-L141 | train |
romanz/trezor-agent | libagent/device/ui.py | UI.get_passphrase | def get_passphrase(self, prompt='Passphrase:'):
"""Ask the user for passphrase."""
passphrase = None
if self.cached_passphrase_ack:
passphrase = self.cached_passphrase_ack.get()
if passphrase is None:
passphrase = interact(
title='{} passphrase'.format(self.device_name),
prompt=prompt,
description=None,
binary=self.passphrase_entry_binary,
options=self.options_getter())
if self.cached_passphrase_ack:
self.cached_passphrase_ack.set(passphrase)
return passphrase | python | def get_passphrase(self, prompt='Passphrase:'):
"""Ask the user for passphrase."""
passphrase = None
if self.cached_passphrase_ack:
passphrase = self.cached_passphrase_ack.get()
if passphrase is None:
passphrase = interact(
title='{} passphrase'.format(self.device_name),
prompt=prompt,
description=None,
binary=self.passphrase_entry_binary,
options=self.options_getter())
if self.cached_passphrase_ack:
self.cached_passphrase_ack.set(passphrase)
return passphrase | [
"def",
"get_passphrase",
"(",
"self",
",",
"prompt",
"=",
"'Passphrase:'",
")",
":",
"passphrase",
"=",
"None",
"if",
"self",
".",
"cached_passphrase_ack",
":",
"passphrase",
"=",
"self",
".",
"cached_passphrase_ack",
".",
"get",
"(",
")",
"if",
"passphrase",
"is",
"None",
":",
"passphrase",
"=",
"interact",
"(",
"title",
"=",
"'{} passphrase'",
".",
"format",
"(",
"self",
".",
"device_name",
")",
",",
"prompt",
"=",
"prompt",
",",
"description",
"=",
"None",
",",
"binary",
"=",
"self",
".",
"passphrase_entry_binary",
",",
"options",
"=",
"self",
".",
"options_getter",
"(",
")",
")",
"if",
"self",
".",
"cached_passphrase_ack",
":",
"self",
".",
"cached_passphrase_ack",
".",
"set",
"(",
"passphrase",
")",
"return",
"passphrase"
] | Ask the user for passphrase. | [
"Ask",
"the",
"user",
"for",
"passphrase",
"."
] | 513b1259c4d7aca5f88cd958edc11828d0712f1b | https://github.com/romanz/trezor-agent/blob/513b1259c4d7aca5f88cd958edc11828d0712f1b/libagent/device/ui.py#L43-L57 | train |
romanz/trezor-agent | libagent/ssh/client.py | Client.export_public_keys | def export_public_keys(self, identities):
"""Export SSH public keys from the device."""
public_keys = []
with self.device:
for i in identities:
pubkey = self.device.pubkey(identity=i)
vk = formats.decompress_pubkey(pubkey=pubkey,
curve_name=i.curve_name)
public_key = formats.export_public_key(vk=vk,
label=i.to_string())
public_keys.append(public_key)
return public_keys | python | def export_public_keys(self, identities):
"""Export SSH public keys from the device."""
public_keys = []
with self.device:
for i in identities:
pubkey = self.device.pubkey(identity=i)
vk = formats.decompress_pubkey(pubkey=pubkey,
curve_name=i.curve_name)
public_key = formats.export_public_key(vk=vk,
label=i.to_string())
public_keys.append(public_key)
return public_keys | [
"def",
"export_public_keys",
"(",
"self",
",",
"identities",
")",
":",
"public_keys",
"=",
"[",
"]",
"with",
"self",
".",
"device",
":",
"for",
"i",
"in",
"identities",
":",
"pubkey",
"=",
"self",
".",
"device",
".",
"pubkey",
"(",
"identity",
"=",
"i",
")",
"vk",
"=",
"formats",
".",
"decompress_pubkey",
"(",
"pubkey",
"=",
"pubkey",
",",
"curve_name",
"=",
"i",
".",
"curve_name",
")",
"public_key",
"=",
"formats",
".",
"export_public_key",
"(",
"vk",
"=",
"vk",
",",
"label",
"=",
"i",
".",
"to_string",
"(",
")",
")",
"public_keys",
".",
"append",
"(",
"public_key",
")",
"return",
"public_keys"
] | Export SSH public keys from the device. | [
"Export",
"SSH",
"public",
"keys",
"from",
"the",
"device",
"."
] | 513b1259c4d7aca5f88cd958edc11828d0712f1b | https://github.com/romanz/trezor-agent/blob/513b1259c4d7aca5f88cd958edc11828d0712f1b/libagent/ssh/client.py#L21-L32 | train |
romanz/trezor-agent | libagent/ssh/client.py | Client.sign_ssh_challenge | def sign_ssh_challenge(self, blob, identity):
"""Sign given blob using a private key on the device."""
msg = _parse_ssh_blob(blob)
log.debug('%s: user %r via %r (%r)',
msg['conn'], msg['user'], msg['auth'], msg['key_type'])
log.debug('nonce: %r', msg['nonce'])
fp = msg['public_key']['fingerprint']
log.debug('fingerprint: %s', fp)
log.debug('hidden challenge size: %d bytes', len(blob))
log.info('please confirm user "%s" login to "%s" using %s...',
msg['user'].decode('ascii'), identity.to_string(),
self.device)
with self.device:
return self.device.sign(blob=blob, identity=identity) | python | def sign_ssh_challenge(self, blob, identity):
"""Sign given blob using a private key on the device."""
msg = _parse_ssh_blob(blob)
log.debug('%s: user %r via %r (%r)',
msg['conn'], msg['user'], msg['auth'], msg['key_type'])
log.debug('nonce: %r', msg['nonce'])
fp = msg['public_key']['fingerprint']
log.debug('fingerprint: %s', fp)
log.debug('hidden challenge size: %d bytes', len(blob))
log.info('please confirm user "%s" login to "%s" using %s...',
msg['user'].decode('ascii'), identity.to_string(),
self.device)
with self.device:
return self.device.sign(blob=blob, identity=identity) | [
"def",
"sign_ssh_challenge",
"(",
"self",
",",
"blob",
",",
"identity",
")",
":",
"msg",
"=",
"_parse_ssh_blob",
"(",
"blob",
")",
"log",
".",
"debug",
"(",
"'%s: user %r via %r (%r)'",
",",
"msg",
"[",
"'conn'",
"]",
",",
"msg",
"[",
"'user'",
"]",
",",
"msg",
"[",
"'auth'",
"]",
",",
"msg",
"[",
"'key_type'",
"]",
")",
"log",
".",
"debug",
"(",
"'nonce: %r'",
",",
"msg",
"[",
"'nonce'",
"]",
")",
"fp",
"=",
"msg",
"[",
"'public_key'",
"]",
"[",
"'fingerprint'",
"]",
"log",
".",
"debug",
"(",
"'fingerprint: %s'",
",",
"fp",
")",
"log",
".",
"debug",
"(",
"'hidden challenge size: %d bytes'",
",",
"len",
"(",
"blob",
")",
")",
"log",
".",
"info",
"(",
"'please confirm user \"%s\" login to \"%s\" using %s...'",
",",
"msg",
"[",
"'user'",
"]",
".",
"decode",
"(",
"'ascii'",
")",
",",
"identity",
".",
"to_string",
"(",
")",
",",
"self",
".",
"device",
")",
"with",
"self",
".",
"device",
":",
"return",
"self",
".",
"device",
".",
"sign",
"(",
"blob",
"=",
"blob",
",",
"identity",
"=",
"identity",
")"
] | Sign given blob using a private key on the device. | [
"Sign",
"given",
"blob",
"using",
"a",
"private",
"key",
"on",
"the",
"device",
"."
] | 513b1259c4d7aca5f88cd958edc11828d0712f1b | https://github.com/romanz/trezor-agent/blob/513b1259c4d7aca5f88cd958edc11828d0712f1b/libagent/ssh/client.py#L34-L49 | train |
romanz/trezor-agent | libagent/formats.py | fingerprint | def fingerprint(blob):
"""
Compute SSH fingerprint for specified blob.
See https://en.wikipedia.org/wiki/Public_key_fingerprint for details.
"""
digest = hashlib.md5(blob).digest()
return ':'.join('{:02x}'.format(c) for c in bytearray(digest)) | python | def fingerprint(blob):
"""
Compute SSH fingerprint for specified blob.
See https://en.wikipedia.org/wiki/Public_key_fingerprint for details.
"""
digest = hashlib.md5(blob).digest()
return ':'.join('{:02x}'.format(c) for c in bytearray(digest)) | [
"def",
"fingerprint",
"(",
"blob",
")",
":",
"digest",
"=",
"hashlib",
".",
"md5",
"(",
"blob",
")",
".",
"digest",
"(",
")",
"return",
"':'",
".",
"join",
"(",
"'{:02x}'",
".",
"format",
"(",
"c",
")",
"for",
"c",
"in",
"bytearray",
"(",
"digest",
")",
")"
] | Compute SSH fingerprint for specified blob.
See https://en.wikipedia.org/wiki/Public_key_fingerprint for details. | [
"Compute",
"SSH",
"fingerprint",
"for",
"specified",
"blob",
"."
] | 513b1259c4d7aca5f88cd958edc11828d0712f1b | https://github.com/romanz/trezor-agent/blob/513b1259c4d7aca5f88cd958edc11828d0712f1b/libagent/formats.py#L34-L41 | train |
romanz/trezor-agent | libagent/formats.py | parse_pubkey | def parse_pubkey(blob):
"""
Parse SSH public key from given blob.
Construct a verifier for ECDSA signatures.
The verifier returns the signatures in the required SSH format.
Currently, NIST256P1 and ED25519 elliptic curves are supported.
"""
fp = fingerprint(blob)
s = io.BytesIO(blob)
key_type = util.read_frame(s)
log.debug('key type: %s', key_type)
assert key_type in SUPPORTED_KEY_TYPES, key_type
result = {'blob': blob, 'type': key_type, 'fingerprint': fp}
if key_type == SSH_NIST256_KEY_TYPE:
curve_name = util.read_frame(s)
log.debug('curve name: %s', curve_name)
point = util.read_frame(s)
assert s.read() == b''
_type, point = point[:1], point[1:]
assert _type == SSH_NIST256_DER_OCTET
size = len(point) // 2
assert len(point) == 2 * size
coords = (util.bytes2num(point[:size]), util.bytes2num(point[size:]))
curve = ecdsa.NIST256p
point = ecdsa.ellipticcurve.Point(curve.curve, *coords)
def ecdsa_verifier(sig, msg):
assert len(sig) == 2 * size
sig_decode = ecdsa.util.sigdecode_string
vk = ecdsa.VerifyingKey.from_public_point(point, curve, hashfunc)
vk.verify(signature=sig, data=msg, sigdecode=sig_decode)
parts = [sig[:size], sig[size:]]
return b''.join([util.frame(b'\x00' + p) for p in parts])
result.update(point=coords, curve=CURVE_NIST256,
verifier=ecdsa_verifier)
if key_type == SSH_ED25519_KEY_TYPE:
pubkey = util.read_frame(s)
assert s.read() == b''
def ed25519_verify(sig, msg):
assert len(sig) == 64
vk = ed25519.VerifyingKey(pubkey)
vk.verify(sig, msg)
return sig
result.update(curve=CURVE_ED25519, verifier=ed25519_verify)
return result | python | def parse_pubkey(blob):
"""
Parse SSH public key from given blob.
Construct a verifier for ECDSA signatures.
The verifier returns the signatures in the required SSH format.
Currently, NIST256P1 and ED25519 elliptic curves are supported.
"""
fp = fingerprint(blob)
s = io.BytesIO(blob)
key_type = util.read_frame(s)
log.debug('key type: %s', key_type)
assert key_type in SUPPORTED_KEY_TYPES, key_type
result = {'blob': blob, 'type': key_type, 'fingerprint': fp}
if key_type == SSH_NIST256_KEY_TYPE:
curve_name = util.read_frame(s)
log.debug('curve name: %s', curve_name)
point = util.read_frame(s)
assert s.read() == b''
_type, point = point[:1], point[1:]
assert _type == SSH_NIST256_DER_OCTET
size = len(point) // 2
assert len(point) == 2 * size
coords = (util.bytes2num(point[:size]), util.bytes2num(point[size:]))
curve = ecdsa.NIST256p
point = ecdsa.ellipticcurve.Point(curve.curve, *coords)
def ecdsa_verifier(sig, msg):
assert len(sig) == 2 * size
sig_decode = ecdsa.util.sigdecode_string
vk = ecdsa.VerifyingKey.from_public_point(point, curve, hashfunc)
vk.verify(signature=sig, data=msg, sigdecode=sig_decode)
parts = [sig[:size], sig[size:]]
return b''.join([util.frame(b'\x00' + p) for p in parts])
result.update(point=coords, curve=CURVE_NIST256,
verifier=ecdsa_verifier)
if key_type == SSH_ED25519_KEY_TYPE:
pubkey = util.read_frame(s)
assert s.read() == b''
def ed25519_verify(sig, msg):
assert len(sig) == 64
vk = ed25519.VerifyingKey(pubkey)
vk.verify(sig, msg)
return sig
result.update(curve=CURVE_ED25519, verifier=ed25519_verify)
return result | [
"def",
"parse_pubkey",
"(",
"blob",
")",
":",
"fp",
"=",
"fingerprint",
"(",
"blob",
")",
"s",
"=",
"io",
".",
"BytesIO",
"(",
"blob",
")",
"key_type",
"=",
"util",
".",
"read_frame",
"(",
"s",
")",
"log",
".",
"debug",
"(",
"'key type: %s'",
",",
"key_type",
")",
"assert",
"key_type",
"in",
"SUPPORTED_KEY_TYPES",
",",
"key_type",
"result",
"=",
"{",
"'blob'",
":",
"blob",
",",
"'type'",
":",
"key_type",
",",
"'fingerprint'",
":",
"fp",
"}",
"if",
"key_type",
"==",
"SSH_NIST256_KEY_TYPE",
":",
"curve_name",
"=",
"util",
".",
"read_frame",
"(",
"s",
")",
"log",
".",
"debug",
"(",
"'curve name: %s'",
",",
"curve_name",
")",
"point",
"=",
"util",
".",
"read_frame",
"(",
"s",
")",
"assert",
"s",
".",
"read",
"(",
")",
"==",
"b''",
"_type",
",",
"point",
"=",
"point",
"[",
":",
"1",
"]",
",",
"point",
"[",
"1",
":",
"]",
"assert",
"_type",
"==",
"SSH_NIST256_DER_OCTET",
"size",
"=",
"len",
"(",
"point",
")",
"//",
"2",
"assert",
"len",
"(",
"point",
")",
"==",
"2",
"*",
"size",
"coords",
"=",
"(",
"util",
".",
"bytes2num",
"(",
"point",
"[",
":",
"size",
"]",
")",
",",
"util",
".",
"bytes2num",
"(",
"point",
"[",
"size",
":",
"]",
")",
")",
"curve",
"=",
"ecdsa",
".",
"NIST256p",
"point",
"=",
"ecdsa",
".",
"ellipticcurve",
".",
"Point",
"(",
"curve",
".",
"curve",
",",
"*",
"coords",
")",
"def",
"ecdsa_verifier",
"(",
"sig",
",",
"msg",
")",
":",
"assert",
"len",
"(",
"sig",
")",
"==",
"2",
"*",
"size",
"sig_decode",
"=",
"ecdsa",
".",
"util",
".",
"sigdecode_string",
"vk",
"=",
"ecdsa",
".",
"VerifyingKey",
".",
"from_public_point",
"(",
"point",
",",
"curve",
",",
"hashfunc",
")",
"vk",
".",
"verify",
"(",
"signature",
"=",
"sig",
",",
"data",
"=",
"msg",
",",
"sigdecode",
"=",
"sig_decode",
")",
"parts",
"=",
"[",
"sig",
"[",
":",
"size",
"]",
",",
"sig",
"[",
"size",
":",
"]",
"]",
"return",
"b''",
".",
"join",
"(",
"[",
"util",
".",
"frame",
"(",
"b'\\x00'",
"+",
"p",
")",
"for",
"p",
"in",
"parts",
"]",
")",
"result",
".",
"update",
"(",
"point",
"=",
"coords",
",",
"curve",
"=",
"CURVE_NIST256",
",",
"verifier",
"=",
"ecdsa_verifier",
")",
"if",
"key_type",
"==",
"SSH_ED25519_KEY_TYPE",
":",
"pubkey",
"=",
"util",
".",
"read_frame",
"(",
"s",
")",
"assert",
"s",
".",
"read",
"(",
")",
"==",
"b''",
"def",
"ed25519_verify",
"(",
"sig",
",",
"msg",
")",
":",
"assert",
"len",
"(",
"sig",
")",
"==",
"64",
"vk",
"=",
"ed25519",
".",
"VerifyingKey",
"(",
"pubkey",
")",
"vk",
".",
"verify",
"(",
"sig",
",",
"msg",
")",
"return",
"sig",
"result",
".",
"update",
"(",
"curve",
"=",
"CURVE_ED25519",
",",
"verifier",
"=",
"ed25519_verify",
")",
"return",
"result"
] | Parse SSH public key from given blob.
Construct a verifier for ECDSA signatures.
The verifier returns the signatures in the required SSH format.
Currently, NIST256P1 and ED25519 elliptic curves are supported. | [
"Parse",
"SSH",
"public",
"key",
"from",
"given",
"blob",
"."
] | 513b1259c4d7aca5f88cd958edc11828d0712f1b | https://github.com/romanz/trezor-agent/blob/513b1259c4d7aca5f88cd958edc11828d0712f1b/libagent/formats.py#L44-L97 | train |
romanz/trezor-agent | libagent/formats.py | export_public_key | def export_public_key(vk, label):
"""
Export public key to text format.
The resulting string can be written into a .pub file or
appended to the ~/.ssh/authorized_keys file.
"""
key_type, blob = serialize_verifying_key(vk)
log.debug('fingerprint: %s', fingerprint(blob))
b64 = base64.b64encode(blob).decode('ascii')
return u'{} {} {}\n'.format(key_type.decode('ascii'), b64, label) | python | def export_public_key(vk, label):
"""
Export public key to text format.
The resulting string can be written into a .pub file or
appended to the ~/.ssh/authorized_keys file.
"""
key_type, blob = serialize_verifying_key(vk)
log.debug('fingerprint: %s', fingerprint(blob))
b64 = base64.b64encode(blob).decode('ascii')
return u'{} {} {}\n'.format(key_type.decode('ascii'), b64, label) | [
"def",
"export_public_key",
"(",
"vk",
",",
"label",
")",
":",
"key_type",
",",
"blob",
"=",
"serialize_verifying_key",
"(",
"vk",
")",
"log",
".",
"debug",
"(",
"'fingerprint: %s'",
",",
"fingerprint",
"(",
"blob",
")",
")",
"b64",
"=",
"base64",
".",
"b64encode",
"(",
"blob",
")",
".",
"decode",
"(",
"'ascii'",
")",
"return",
"u'{} {} {}\\n'",
".",
"format",
"(",
"key_type",
".",
"decode",
"(",
"'ascii'",
")",
",",
"b64",
",",
"label",
")"
] | Export public key to text format.
The resulting string can be written into a .pub file or
appended to the ~/.ssh/authorized_keys file. | [
"Export",
"public",
"key",
"to",
"text",
"format",
"."
] | 513b1259c4d7aca5f88cd958edc11828d0712f1b | https://github.com/romanz/trezor-agent/blob/513b1259c4d7aca5f88cd958edc11828d0712f1b/libagent/formats.py#L181-L191 | train |
romanz/trezor-agent | libagent/formats.py | import_public_key | def import_public_key(line):
"""Parse public key textual format, as saved at a .pub file."""
log.debug('loading SSH public key: %r', line)
file_type, base64blob, name = line.split()
blob = base64.b64decode(base64blob)
result = parse_pubkey(blob)
result['name'] = name.encode('utf-8')
assert result['type'] == file_type.encode('ascii')
log.debug('loaded %s public key: %s', file_type, result['fingerprint'])
return result | python | def import_public_key(line):
"""Parse public key textual format, as saved at a .pub file."""
log.debug('loading SSH public key: %r', line)
file_type, base64blob, name = line.split()
blob = base64.b64decode(base64blob)
result = parse_pubkey(blob)
result['name'] = name.encode('utf-8')
assert result['type'] == file_type.encode('ascii')
log.debug('loaded %s public key: %s', file_type, result['fingerprint'])
return result | [
"def",
"import_public_key",
"(",
"line",
")",
":",
"log",
".",
"debug",
"(",
"'loading SSH public key: %r'",
",",
"line",
")",
"file_type",
",",
"base64blob",
",",
"name",
"=",
"line",
".",
"split",
"(",
")",
"blob",
"=",
"base64",
".",
"b64decode",
"(",
"base64blob",
")",
"result",
"=",
"parse_pubkey",
"(",
"blob",
")",
"result",
"[",
"'name'",
"]",
"=",
"name",
".",
"encode",
"(",
"'utf-8'",
")",
"assert",
"result",
"[",
"'type'",
"]",
"==",
"file_type",
".",
"encode",
"(",
"'ascii'",
")",
"log",
".",
"debug",
"(",
"'loaded %s public key: %s'",
",",
"file_type",
",",
"result",
"[",
"'fingerprint'",
"]",
")",
"return",
"result"
] | Parse public key textual format, as saved at a .pub file. | [
"Parse",
"public",
"key",
"textual",
"format",
"as",
"saved",
"at",
"a",
".",
"pub",
"file",
"."
] | 513b1259c4d7aca5f88cd958edc11828d0712f1b | https://github.com/romanz/trezor-agent/blob/513b1259c4d7aca5f88cd958edc11828d0712f1b/libagent/formats.py#L194-L203 | train |
romanz/trezor-agent | libagent/gpg/decode.py | parse_packets | def parse_packets(stream):
"""
Support iterative parsing of available GPG packets.
See https://tools.ietf.org/html/rfc4880#section-4.2 for details.
"""
reader = util.Reader(stream)
while True:
try:
value = reader.readfmt('B')
except EOFError:
return
log.debug('prefix byte: %s', bin(value))
assert util.bit(value, 7) == 1
tag = util.low_bits(value, 6)
if util.bit(value, 6) == 0:
length_type = util.low_bits(tag, 2)
tag = tag >> 2
fmt = {0: '>B', 1: '>H', 2: '>L'}[length_type]
packet_size = reader.readfmt(fmt)
else:
first = reader.readfmt('B')
if first < 192:
packet_size = first
elif first < 224:
packet_size = ((first - 192) << 8) + reader.readfmt('B') + 192
elif first == 255:
packet_size = reader.readfmt('>L')
else:
log.error('Partial Body Lengths unsupported')
log.debug('packet length: %d', packet_size)
packet_data = reader.read(packet_size)
packet_type = PACKET_TYPES.get(tag)
p = {'type': 'unknown', 'tag': tag, 'raw': packet_data}
if packet_type is not None:
try:
p = packet_type(util.Reader(io.BytesIO(packet_data)))
p['tag'] = tag
except ValueError:
log.exception('Skipping packet: %s', util.hexlify(packet_data))
log.debug('packet "%s": %s', p['type'], p)
yield p | python | def parse_packets(stream):
"""
Support iterative parsing of available GPG packets.
See https://tools.ietf.org/html/rfc4880#section-4.2 for details.
"""
reader = util.Reader(stream)
while True:
try:
value = reader.readfmt('B')
except EOFError:
return
log.debug('prefix byte: %s', bin(value))
assert util.bit(value, 7) == 1
tag = util.low_bits(value, 6)
if util.bit(value, 6) == 0:
length_type = util.low_bits(tag, 2)
tag = tag >> 2
fmt = {0: '>B', 1: '>H', 2: '>L'}[length_type]
packet_size = reader.readfmt(fmt)
else:
first = reader.readfmt('B')
if first < 192:
packet_size = first
elif first < 224:
packet_size = ((first - 192) << 8) + reader.readfmt('B') + 192
elif first == 255:
packet_size = reader.readfmt('>L')
else:
log.error('Partial Body Lengths unsupported')
log.debug('packet length: %d', packet_size)
packet_data = reader.read(packet_size)
packet_type = PACKET_TYPES.get(tag)
p = {'type': 'unknown', 'tag': tag, 'raw': packet_data}
if packet_type is not None:
try:
p = packet_type(util.Reader(io.BytesIO(packet_data)))
p['tag'] = tag
except ValueError:
log.exception('Skipping packet: %s', util.hexlify(packet_data))
log.debug('packet "%s": %s', p['type'], p)
yield p | [
"def",
"parse_packets",
"(",
"stream",
")",
":",
"reader",
"=",
"util",
".",
"Reader",
"(",
"stream",
")",
"while",
"True",
":",
"try",
":",
"value",
"=",
"reader",
".",
"readfmt",
"(",
"'B'",
")",
"except",
"EOFError",
":",
"return",
"log",
".",
"debug",
"(",
"'prefix byte: %s'",
",",
"bin",
"(",
"value",
")",
")",
"assert",
"util",
".",
"bit",
"(",
"value",
",",
"7",
")",
"==",
"1",
"tag",
"=",
"util",
".",
"low_bits",
"(",
"value",
",",
"6",
")",
"if",
"util",
".",
"bit",
"(",
"value",
",",
"6",
")",
"==",
"0",
":",
"length_type",
"=",
"util",
".",
"low_bits",
"(",
"tag",
",",
"2",
")",
"tag",
"=",
"tag",
">>",
"2",
"fmt",
"=",
"{",
"0",
":",
"'>B'",
",",
"1",
":",
"'>H'",
",",
"2",
":",
"'>L'",
"}",
"[",
"length_type",
"]",
"packet_size",
"=",
"reader",
".",
"readfmt",
"(",
"fmt",
")",
"else",
":",
"first",
"=",
"reader",
".",
"readfmt",
"(",
"'B'",
")",
"if",
"first",
"<",
"192",
":",
"packet_size",
"=",
"first",
"elif",
"first",
"<",
"224",
":",
"packet_size",
"=",
"(",
"(",
"first",
"-",
"192",
")",
"<<",
"8",
")",
"+",
"reader",
".",
"readfmt",
"(",
"'B'",
")",
"+",
"192",
"elif",
"first",
"==",
"255",
":",
"packet_size",
"=",
"reader",
".",
"readfmt",
"(",
"'>L'",
")",
"else",
":",
"log",
".",
"error",
"(",
"'Partial Body Lengths unsupported'",
")",
"log",
".",
"debug",
"(",
"'packet length: %d'",
",",
"packet_size",
")",
"packet_data",
"=",
"reader",
".",
"read",
"(",
"packet_size",
")",
"packet_type",
"=",
"PACKET_TYPES",
".",
"get",
"(",
"tag",
")",
"p",
"=",
"{",
"'type'",
":",
"'unknown'",
",",
"'tag'",
":",
"tag",
",",
"'raw'",
":",
"packet_data",
"}",
"if",
"packet_type",
"is",
"not",
"None",
":",
"try",
":",
"p",
"=",
"packet_type",
"(",
"util",
".",
"Reader",
"(",
"io",
".",
"BytesIO",
"(",
"packet_data",
")",
")",
")",
"p",
"[",
"'tag'",
"]",
"=",
"tag",
"except",
"ValueError",
":",
"log",
".",
"exception",
"(",
"'Skipping packet: %s'",
",",
"util",
".",
"hexlify",
"(",
"packet_data",
")",
")",
"log",
".",
"debug",
"(",
"'packet \"%s\": %s'",
",",
"p",
"[",
"'type'",
"]",
",",
"p",
")",
"yield",
"p"
] | Support iterative parsing of available GPG packets.
See https://tools.ietf.org/html/rfc4880#section-4.2 for details. | [
"Support",
"iterative",
"parsing",
"of",
"available",
"GPG",
"packets",
"."
] | 513b1259c4d7aca5f88cd958edc11828d0712f1b | https://github.com/romanz/trezor-agent/blob/513b1259c4d7aca5f88cd958edc11828d0712f1b/libagent/gpg/decode.py#L215-L261 | train |
romanz/trezor-agent | libagent/gpg/decode.py | digest_packets | def digest_packets(packets, hasher):
"""Compute digest on specified packets, according to '_to_hash' field."""
data_to_hash = io.BytesIO()
for p in packets:
data_to_hash.write(p['_to_hash'])
hasher.update(data_to_hash.getvalue())
return hasher.digest() | python | def digest_packets(packets, hasher):
"""Compute digest on specified packets, according to '_to_hash' field."""
data_to_hash = io.BytesIO()
for p in packets:
data_to_hash.write(p['_to_hash'])
hasher.update(data_to_hash.getvalue())
return hasher.digest() | [
"def",
"digest_packets",
"(",
"packets",
",",
"hasher",
")",
":",
"data_to_hash",
"=",
"io",
".",
"BytesIO",
"(",
")",
"for",
"p",
"in",
"packets",
":",
"data_to_hash",
".",
"write",
"(",
"p",
"[",
"'_to_hash'",
"]",
")",
"hasher",
".",
"update",
"(",
"data_to_hash",
".",
"getvalue",
"(",
")",
")",
"return",
"hasher",
".",
"digest",
"(",
")"
] | Compute digest on specified packets, according to '_to_hash' field. | [
"Compute",
"digest",
"on",
"specified",
"packets",
"according",
"to",
"_to_hash",
"field",
"."
] | 513b1259c4d7aca5f88cd958edc11828d0712f1b | https://github.com/romanz/trezor-agent/blob/513b1259c4d7aca5f88cd958edc11828d0712f1b/libagent/gpg/decode.py#L264-L270 | train |
romanz/trezor-agent | libagent/gpg/decode.py | load_by_keygrip | def load_by_keygrip(pubkey_bytes, keygrip):
"""Return public key and first user ID for specified keygrip."""
stream = io.BytesIO(pubkey_bytes)
packets = list(parse_packets(stream))
packets_per_pubkey = []
for p in packets:
if p['type'] == 'pubkey':
# Add a new packet list for each pubkey.
packets_per_pubkey.append([])
packets_per_pubkey[-1].append(p)
for packets in packets_per_pubkey:
user_ids = [p for p in packets if p['type'] == 'user_id']
for p in packets:
if p.get('keygrip') == keygrip:
return p, user_ids
raise KeyError('{} keygrip not found'.format(util.hexlify(keygrip))) | python | def load_by_keygrip(pubkey_bytes, keygrip):
"""Return public key and first user ID for specified keygrip."""
stream = io.BytesIO(pubkey_bytes)
packets = list(parse_packets(stream))
packets_per_pubkey = []
for p in packets:
if p['type'] == 'pubkey':
# Add a new packet list for each pubkey.
packets_per_pubkey.append([])
packets_per_pubkey[-1].append(p)
for packets in packets_per_pubkey:
user_ids = [p for p in packets if p['type'] == 'user_id']
for p in packets:
if p.get('keygrip') == keygrip:
return p, user_ids
raise KeyError('{} keygrip not found'.format(util.hexlify(keygrip))) | [
"def",
"load_by_keygrip",
"(",
"pubkey_bytes",
",",
"keygrip",
")",
":",
"stream",
"=",
"io",
".",
"BytesIO",
"(",
"pubkey_bytes",
")",
"packets",
"=",
"list",
"(",
"parse_packets",
"(",
"stream",
")",
")",
"packets_per_pubkey",
"=",
"[",
"]",
"for",
"p",
"in",
"packets",
":",
"if",
"p",
"[",
"'type'",
"]",
"==",
"'pubkey'",
":",
"# Add a new packet list for each pubkey.",
"packets_per_pubkey",
".",
"append",
"(",
"[",
"]",
")",
"packets_per_pubkey",
"[",
"-",
"1",
"]",
".",
"append",
"(",
"p",
")",
"for",
"packets",
"in",
"packets_per_pubkey",
":",
"user_ids",
"=",
"[",
"p",
"for",
"p",
"in",
"packets",
"if",
"p",
"[",
"'type'",
"]",
"==",
"'user_id'",
"]",
"for",
"p",
"in",
"packets",
":",
"if",
"p",
".",
"get",
"(",
"'keygrip'",
")",
"==",
"keygrip",
":",
"return",
"p",
",",
"user_ids",
"raise",
"KeyError",
"(",
"'{} keygrip not found'",
".",
"format",
"(",
"util",
".",
"hexlify",
"(",
"keygrip",
")",
")",
")"
] | Return public key and first user ID for specified keygrip. | [
"Return",
"public",
"key",
"and",
"first",
"user",
"ID",
"for",
"specified",
"keygrip",
"."
] | 513b1259c4d7aca5f88cd958edc11828d0712f1b | https://github.com/romanz/trezor-agent/blob/513b1259c4d7aca5f88cd958edc11828d0712f1b/libagent/gpg/decode.py#L284-L300 | train |
romanz/trezor-agent | libagent/gpg/decode.py | load_signature | def load_signature(stream, original_data):
"""Load signature from stream, and compute GPG digest for verification."""
signature, = list(parse_packets((stream)))
hash_alg = HASH_ALGORITHMS[signature['hash_alg']]
digest = digest_packets([{'_to_hash': original_data}, signature],
hasher=hashlib.new(hash_alg))
assert signature['hash_prefix'] == digest[:2]
return signature, digest | python | def load_signature(stream, original_data):
"""Load signature from stream, and compute GPG digest for verification."""
signature, = list(parse_packets((stream)))
hash_alg = HASH_ALGORITHMS[signature['hash_alg']]
digest = digest_packets([{'_to_hash': original_data}, signature],
hasher=hashlib.new(hash_alg))
assert signature['hash_prefix'] == digest[:2]
return signature, digest | [
"def",
"load_signature",
"(",
"stream",
",",
"original_data",
")",
":",
"signature",
",",
"=",
"list",
"(",
"parse_packets",
"(",
"(",
"stream",
")",
")",
")",
"hash_alg",
"=",
"HASH_ALGORITHMS",
"[",
"signature",
"[",
"'hash_alg'",
"]",
"]",
"digest",
"=",
"digest_packets",
"(",
"[",
"{",
"'_to_hash'",
":",
"original_data",
"}",
",",
"signature",
"]",
",",
"hasher",
"=",
"hashlib",
".",
"new",
"(",
"hash_alg",
")",
")",
"assert",
"signature",
"[",
"'hash_prefix'",
"]",
"==",
"digest",
"[",
":",
"2",
"]",
"return",
"signature",
",",
"digest"
] | Load signature from stream, and compute GPG digest for verification. | [
"Load",
"signature",
"from",
"stream",
"and",
"compute",
"GPG",
"digest",
"for",
"verification",
"."
] | 513b1259c4d7aca5f88cd958edc11828d0712f1b | https://github.com/romanz/trezor-agent/blob/513b1259c4d7aca5f88cd958edc11828d0712f1b/libagent/gpg/decode.py#L303-L310 | train |
romanz/trezor-agent | libagent/gpg/decode.py | remove_armor | def remove_armor(armored_data):
"""Decode armored data into its binary form."""
stream = io.BytesIO(armored_data)
lines = stream.readlines()[3:-1]
data = base64.b64decode(b''.join(lines))
payload, checksum = data[:-3], data[-3:]
assert util.crc24(payload) == checksum
return payload | python | def remove_armor(armored_data):
"""Decode armored data into its binary form."""
stream = io.BytesIO(armored_data)
lines = stream.readlines()[3:-1]
data = base64.b64decode(b''.join(lines))
payload, checksum = data[:-3], data[-3:]
assert util.crc24(payload) == checksum
return payload | [
"def",
"remove_armor",
"(",
"armored_data",
")",
":",
"stream",
"=",
"io",
".",
"BytesIO",
"(",
"armored_data",
")",
"lines",
"=",
"stream",
".",
"readlines",
"(",
")",
"[",
"3",
":",
"-",
"1",
"]",
"data",
"=",
"base64",
".",
"b64decode",
"(",
"b''",
".",
"join",
"(",
"lines",
")",
")",
"payload",
",",
"checksum",
"=",
"data",
"[",
":",
"-",
"3",
"]",
",",
"data",
"[",
"-",
"3",
":",
"]",
"assert",
"util",
".",
"crc24",
"(",
"payload",
")",
"==",
"checksum",
"return",
"payload"
] | Decode armored data into its binary form. | [
"Decode",
"armored",
"data",
"into",
"its",
"binary",
"form",
"."
] | 513b1259c4d7aca5f88cd958edc11828d0712f1b | https://github.com/romanz/trezor-agent/blob/513b1259c4d7aca5f88cd958edc11828d0712f1b/libagent/gpg/decode.py#L313-L320 | train |
romanz/trezor-agent | libagent/server.py | remove_file | def remove_file(path, remove=os.remove, exists=os.path.exists):
"""Remove file, and raise OSError if still exists."""
try:
remove(path)
except OSError:
if exists(path):
raise | python | def remove_file(path, remove=os.remove, exists=os.path.exists):
"""Remove file, and raise OSError if still exists."""
try:
remove(path)
except OSError:
if exists(path):
raise | [
"def",
"remove_file",
"(",
"path",
",",
"remove",
"=",
"os",
".",
"remove",
",",
"exists",
"=",
"os",
".",
"path",
".",
"exists",
")",
":",
"try",
":",
"remove",
"(",
"path",
")",
"except",
"OSError",
":",
"if",
"exists",
"(",
"path",
")",
":",
"raise"
] | Remove file, and raise OSError if still exists. | [
"Remove",
"file",
"and",
"raise",
"OSError",
"if",
"still",
"exists",
"."
] | 513b1259c4d7aca5f88cd958edc11828d0712f1b | https://github.com/romanz/trezor-agent/blob/513b1259c4d7aca5f88cd958edc11828d0712f1b/libagent/server.py#L14-L20 | train |
romanz/trezor-agent | libagent/server.py | unix_domain_socket_server | def unix_domain_socket_server(sock_path):
"""
Create UNIX-domain socket on specified path.
Listen on it, and delete it after the generated context is over.
"""
log.debug('serving on %s', sock_path)
remove_file(sock_path)
server = socket.socket(socket.AF_UNIX, socket.SOCK_STREAM)
server.bind(sock_path)
server.listen(1)
try:
yield server
finally:
remove_file(sock_path) | python | def unix_domain_socket_server(sock_path):
"""
Create UNIX-domain socket on specified path.
Listen on it, and delete it after the generated context is over.
"""
log.debug('serving on %s', sock_path)
remove_file(sock_path)
server = socket.socket(socket.AF_UNIX, socket.SOCK_STREAM)
server.bind(sock_path)
server.listen(1)
try:
yield server
finally:
remove_file(sock_path) | [
"def",
"unix_domain_socket_server",
"(",
"sock_path",
")",
":",
"log",
".",
"debug",
"(",
"'serving on %s'",
",",
"sock_path",
")",
"remove_file",
"(",
"sock_path",
")",
"server",
"=",
"socket",
".",
"socket",
"(",
"socket",
".",
"AF_UNIX",
",",
"socket",
".",
"SOCK_STREAM",
")",
"server",
".",
"bind",
"(",
"sock_path",
")",
"server",
".",
"listen",
"(",
"1",
")",
"try",
":",
"yield",
"server",
"finally",
":",
"remove_file",
"(",
"sock_path",
")"
] | Create UNIX-domain socket on specified path.
Listen on it, and delete it after the generated context is over. | [
"Create",
"UNIX",
"-",
"domain",
"socket",
"on",
"specified",
"path",
"."
] | 513b1259c4d7aca5f88cd958edc11828d0712f1b | https://github.com/romanz/trezor-agent/blob/513b1259c4d7aca5f88cd958edc11828d0712f1b/libagent/server.py#L24-L39 | train |
romanz/trezor-agent | libagent/server.py | handle_connection | def handle_connection(conn, handler, mutex):
"""
Handle a single connection using the specified protocol handler in a loop.
Since this function may be called concurrently from server_thread,
the specified mutex is used to synchronize the device handling.
Exit when EOFError is raised.
All other exceptions are logged as warnings.
"""
try:
log.debug('welcome agent')
with contextlib.closing(conn):
while True:
msg = util.read_frame(conn)
with mutex:
reply = handler.handle(msg=msg)
util.send(conn, reply)
except EOFError:
log.debug('goodbye agent')
except Exception as e: # pylint: disable=broad-except
log.warning('error: %s', e, exc_info=True) | python | def handle_connection(conn, handler, mutex):
"""
Handle a single connection using the specified protocol handler in a loop.
Since this function may be called concurrently from server_thread,
the specified mutex is used to synchronize the device handling.
Exit when EOFError is raised.
All other exceptions are logged as warnings.
"""
try:
log.debug('welcome agent')
with contextlib.closing(conn):
while True:
msg = util.read_frame(conn)
with mutex:
reply = handler.handle(msg=msg)
util.send(conn, reply)
except EOFError:
log.debug('goodbye agent')
except Exception as e: # pylint: disable=broad-except
log.warning('error: %s', e, exc_info=True) | [
"def",
"handle_connection",
"(",
"conn",
",",
"handler",
",",
"mutex",
")",
":",
"try",
":",
"log",
".",
"debug",
"(",
"'welcome agent'",
")",
"with",
"contextlib",
".",
"closing",
"(",
"conn",
")",
":",
"while",
"True",
":",
"msg",
"=",
"util",
".",
"read_frame",
"(",
"conn",
")",
"with",
"mutex",
":",
"reply",
"=",
"handler",
".",
"handle",
"(",
"msg",
"=",
"msg",
")",
"util",
".",
"send",
"(",
"conn",
",",
"reply",
")",
"except",
"EOFError",
":",
"log",
".",
"debug",
"(",
"'goodbye agent'",
")",
"except",
"Exception",
"as",
"e",
":",
"# pylint: disable=broad-except",
"log",
".",
"warning",
"(",
"'error: %s'",
",",
"e",
",",
"exc_info",
"=",
"True",
")"
] | Handle a single connection using the specified protocol handler in a loop.
Since this function may be called concurrently from server_thread,
the specified mutex is used to synchronize the device handling.
Exit when EOFError is raised.
All other exceptions are logged as warnings. | [
"Handle",
"a",
"single",
"connection",
"using",
"the",
"specified",
"protocol",
"handler",
"in",
"a",
"loop",
"."
] | 513b1259c4d7aca5f88cd958edc11828d0712f1b | https://github.com/romanz/trezor-agent/blob/513b1259c4d7aca5f88cd958edc11828d0712f1b/libagent/server.py#L79-L100 | train |
romanz/trezor-agent | libagent/server.py | retry | def retry(func, exception_type, quit_event):
"""
Run the function, retrying when the specified exception_type occurs.
Poll quit_event on each iteration, to be responsive to an external
exit request.
"""
while True:
if quit_event.is_set():
raise StopIteration
try:
return func()
except exception_type:
pass | python | def retry(func, exception_type, quit_event):
"""
Run the function, retrying when the specified exception_type occurs.
Poll quit_event on each iteration, to be responsive to an external
exit request.
"""
while True:
if quit_event.is_set():
raise StopIteration
try:
return func()
except exception_type:
pass | [
"def",
"retry",
"(",
"func",
",",
"exception_type",
",",
"quit_event",
")",
":",
"while",
"True",
":",
"if",
"quit_event",
".",
"is_set",
"(",
")",
":",
"raise",
"StopIteration",
"try",
":",
"return",
"func",
"(",
")",
"except",
"exception_type",
":",
"pass"
] | Run the function, retrying when the specified exception_type occurs.
Poll quit_event on each iteration, to be responsive to an external
exit request. | [
"Run",
"the",
"function",
"retrying",
"when",
"the",
"specified",
"exception_type",
"occurs",
"."
] | 513b1259c4d7aca5f88cd958edc11828d0712f1b | https://github.com/romanz/trezor-agent/blob/513b1259c4d7aca5f88cd958edc11828d0712f1b/libagent/server.py#L103-L116 | train |
romanz/trezor-agent | libagent/server.py | spawn | def spawn(func, kwargs):
"""Spawn a thread, and join it after the context is over."""
t = threading.Thread(target=func, kwargs=kwargs)
t.start()
yield
t.join() | python | def spawn(func, kwargs):
"""Spawn a thread, and join it after the context is over."""
t = threading.Thread(target=func, kwargs=kwargs)
t.start()
yield
t.join() | [
"def",
"spawn",
"(",
"func",
",",
"kwargs",
")",
":",
"t",
"=",
"threading",
".",
"Thread",
"(",
"target",
"=",
"func",
",",
"kwargs",
"=",
"kwargs",
")",
"t",
".",
"start",
"(",
")",
"yield",
"t",
".",
"join",
"(",
")"
] | Spawn a thread, and join it after the context is over. | [
"Spawn",
"a",
"thread",
"and",
"join",
"it",
"after",
"the",
"context",
"is",
"over",
"."
] | 513b1259c4d7aca5f88cd958edc11828d0712f1b | https://github.com/romanz/trezor-agent/blob/513b1259c4d7aca5f88cd958edc11828d0712f1b/libagent/server.py#L142-L147 | train |
romanz/trezor-agent | libagent/server.py | run_process | def run_process(command, environ):
"""
Run the specified process and wait until it finishes.
Use environ dict for environment variables.
"""
log.info('running %r with %r', command, environ)
env = dict(os.environ)
env.update(environ)
try:
p = subprocess.Popen(args=command, env=env)
except OSError as e:
raise OSError('cannot run %r: %s' % (command, e))
log.debug('subprocess %d is running', p.pid)
ret = p.wait()
log.debug('subprocess %d exited: %d', p.pid, ret)
return ret | python | def run_process(command, environ):
"""
Run the specified process and wait until it finishes.
Use environ dict for environment variables.
"""
log.info('running %r with %r', command, environ)
env = dict(os.environ)
env.update(environ)
try:
p = subprocess.Popen(args=command, env=env)
except OSError as e:
raise OSError('cannot run %r: %s' % (command, e))
log.debug('subprocess %d is running', p.pid)
ret = p.wait()
log.debug('subprocess %d exited: %d', p.pid, ret)
return ret | [
"def",
"run_process",
"(",
"command",
",",
"environ",
")",
":",
"log",
".",
"info",
"(",
"'running %r with %r'",
",",
"command",
",",
"environ",
")",
"env",
"=",
"dict",
"(",
"os",
".",
"environ",
")",
"env",
".",
"update",
"(",
"environ",
")",
"try",
":",
"p",
"=",
"subprocess",
".",
"Popen",
"(",
"args",
"=",
"command",
",",
"env",
"=",
"env",
")",
"except",
"OSError",
"as",
"e",
":",
"raise",
"OSError",
"(",
"'cannot run %r: %s'",
"%",
"(",
"command",
",",
"e",
")",
")",
"log",
".",
"debug",
"(",
"'subprocess %d is running'",
",",
"p",
".",
"pid",
")",
"ret",
"=",
"p",
".",
"wait",
"(",
")",
"log",
".",
"debug",
"(",
"'subprocess %d exited: %d'",
",",
"p",
".",
"pid",
",",
"ret",
")",
"return",
"ret"
] | Run the specified process and wait until it finishes.
Use environ dict for environment variables. | [
"Run",
"the",
"specified",
"process",
"and",
"wait",
"until",
"it",
"finishes",
"."
] | 513b1259c4d7aca5f88cd958edc11828d0712f1b | https://github.com/romanz/trezor-agent/blob/513b1259c4d7aca5f88cd958edc11828d0712f1b/libagent/server.py#L150-L166 | train |
romanz/trezor-agent | libagent/gpg/keyring.py | check_output | def check_output(args, env=None, sp=subprocess):
"""Call an external binary and return its stdout."""
log.debug('calling %s with env %s', args, env)
output = sp.check_output(args=args, env=env)
log.debug('output: %r', output)
return output | python | def check_output(args, env=None, sp=subprocess):
"""Call an external binary and return its stdout."""
log.debug('calling %s with env %s', args, env)
output = sp.check_output(args=args, env=env)
log.debug('output: %r', output)
return output | [
"def",
"check_output",
"(",
"args",
",",
"env",
"=",
"None",
",",
"sp",
"=",
"subprocess",
")",
":",
"log",
".",
"debug",
"(",
"'calling %s with env %s'",
",",
"args",
",",
"env",
")",
"output",
"=",
"sp",
".",
"check_output",
"(",
"args",
"=",
"args",
",",
"env",
"=",
"env",
")",
"log",
".",
"debug",
"(",
"'output: %r'",
",",
"output",
")",
"return",
"output"
] | Call an external binary and return its stdout. | [
"Call",
"an",
"external",
"binary",
"and",
"return",
"its",
"stdout",
"."
] | 513b1259c4d7aca5f88cd958edc11828d0712f1b | https://github.com/romanz/trezor-agent/blob/513b1259c4d7aca5f88cd958edc11828d0712f1b/libagent/gpg/keyring.py#L17-L22 | train |
romanz/trezor-agent | libagent/gpg/keyring.py | get_agent_sock_path | def get_agent_sock_path(env=None, sp=subprocess):
"""Parse gpgconf output to find out GPG agent UNIX socket path."""
args = [util.which('gpgconf'), '--list-dirs']
output = check_output(args=args, env=env, sp=sp)
lines = output.strip().split(b'\n')
dirs = dict(line.split(b':', 1) for line in lines)
log.debug('%s: %s', args, dirs)
return dirs[b'agent-socket'] | python | def get_agent_sock_path(env=None, sp=subprocess):
"""Parse gpgconf output to find out GPG agent UNIX socket path."""
args = [util.which('gpgconf'), '--list-dirs']
output = check_output(args=args, env=env, sp=sp)
lines = output.strip().split(b'\n')
dirs = dict(line.split(b':', 1) for line in lines)
log.debug('%s: %s', args, dirs)
return dirs[b'agent-socket'] | [
"def",
"get_agent_sock_path",
"(",
"env",
"=",
"None",
",",
"sp",
"=",
"subprocess",
")",
":",
"args",
"=",
"[",
"util",
".",
"which",
"(",
"'gpgconf'",
")",
",",
"'--list-dirs'",
"]",
"output",
"=",
"check_output",
"(",
"args",
"=",
"args",
",",
"env",
"=",
"env",
",",
"sp",
"=",
"sp",
")",
"lines",
"=",
"output",
".",
"strip",
"(",
")",
".",
"split",
"(",
"b'\\n'",
")",
"dirs",
"=",
"dict",
"(",
"line",
".",
"split",
"(",
"b':'",
",",
"1",
")",
"for",
"line",
"in",
"lines",
")",
"log",
".",
"debug",
"(",
"'%s: %s'",
",",
"args",
",",
"dirs",
")",
"return",
"dirs",
"[",
"b'agent-socket'",
"]"
] | Parse gpgconf output to find out GPG agent UNIX socket path. | [
"Parse",
"gpgconf",
"output",
"to",
"find",
"out",
"GPG",
"agent",
"UNIX",
"socket",
"path",
"."
] | 513b1259c4d7aca5f88cd958edc11828d0712f1b | https://github.com/romanz/trezor-agent/blob/513b1259c4d7aca5f88cd958edc11828d0712f1b/libagent/gpg/keyring.py#L25-L32 | train |
romanz/trezor-agent | libagent/gpg/keyring.py | connect_to_agent | def connect_to_agent(env=None, sp=subprocess):
"""Connect to GPG agent's UNIX socket."""
sock_path = get_agent_sock_path(sp=sp, env=env)
# Make sure the original gpg-agent is running.
check_output(args=['gpg-connect-agent', '/bye'], sp=sp)
sock = socket.socket(socket.AF_UNIX, socket.SOCK_STREAM)
sock.connect(sock_path)
return sock | python | def connect_to_agent(env=None, sp=subprocess):
"""Connect to GPG agent's UNIX socket."""
sock_path = get_agent_sock_path(sp=sp, env=env)
# Make sure the original gpg-agent is running.
check_output(args=['gpg-connect-agent', '/bye'], sp=sp)
sock = socket.socket(socket.AF_UNIX, socket.SOCK_STREAM)
sock.connect(sock_path)
return sock | [
"def",
"connect_to_agent",
"(",
"env",
"=",
"None",
",",
"sp",
"=",
"subprocess",
")",
":",
"sock_path",
"=",
"get_agent_sock_path",
"(",
"sp",
"=",
"sp",
",",
"env",
"=",
"env",
")",
"# Make sure the original gpg-agent is running.",
"check_output",
"(",
"args",
"=",
"[",
"'gpg-connect-agent'",
",",
"'/bye'",
"]",
",",
"sp",
"=",
"sp",
")",
"sock",
"=",
"socket",
".",
"socket",
"(",
"socket",
".",
"AF_UNIX",
",",
"socket",
".",
"SOCK_STREAM",
")",
"sock",
".",
"connect",
"(",
"sock_path",
")",
"return",
"sock"
] | Connect to GPG agent's UNIX socket. | [
"Connect",
"to",
"GPG",
"agent",
"s",
"UNIX",
"socket",
"."
] | 513b1259c4d7aca5f88cd958edc11828d0712f1b | https://github.com/romanz/trezor-agent/blob/513b1259c4d7aca5f88cd958edc11828d0712f1b/libagent/gpg/keyring.py#L35-L42 | train |
romanz/trezor-agent | libagent/gpg/keyring.py | sendline | def sendline(sock, msg, confidential=False):
"""Send a binary message, followed by EOL."""
log.debug('<- %r', ('<snip>' if confidential else msg))
sock.sendall(msg + b'\n') | python | def sendline(sock, msg, confidential=False):
"""Send a binary message, followed by EOL."""
log.debug('<- %r', ('<snip>' if confidential else msg))
sock.sendall(msg + b'\n') | [
"def",
"sendline",
"(",
"sock",
",",
"msg",
",",
"confidential",
"=",
"False",
")",
":",
"log",
".",
"debug",
"(",
"'<- %r'",
",",
"(",
"'<snip>'",
"if",
"confidential",
"else",
"msg",
")",
")",
"sock",
".",
"sendall",
"(",
"msg",
"+",
"b'\\n'",
")"
] | Send a binary message, followed by EOL. | [
"Send",
"a",
"binary",
"message",
"followed",
"by",
"EOL",
"."
] | 513b1259c4d7aca5f88cd958edc11828d0712f1b | https://github.com/romanz/trezor-agent/blob/513b1259c4d7aca5f88cd958edc11828d0712f1b/libagent/gpg/keyring.py#L51-L54 | train |
romanz/trezor-agent | libagent/gpg/keyring.py | recvline | def recvline(sock):
"""Receive a single line from the socket."""
reply = io.BytesIO()
while True:
c = sock.recv(1)
if not c:
return None # socket is closed
if c == b'\n':
break
reply.write(c)
result = reply.getvalue()
log.debug('-> %r', result)
return result | python | def recvline(sock):
"""Receive a single line from the socket."""
reply = io.BytesIO()
while True:
c = sock.recv(1)
if not c:
return None # socket is closed
if c == b'\n':
break
reply.write(c)
result = reply.getvalue()
log.debug('-> %r', result)
return result | [
"def",
"recvline",
"(",
"sock",
")",
":",
"reply",
"=",
"io",
".",
"BytesIO",
"(",
")",
"while",
"True",
":",
"c",
"=",
"sock",
".",
"recv",
"(",
"1",
")",
"if",
"not",
"c",
":",
"return",
"None",
"# socket is closed",
"if",
"c",
"==",
"b'\\n'",
":",
"break",
"reply",
".",
"write",
"(",
"c",
")",
"result",
"=",
"reply",
".",
"getvalue",
"(",
")",
"log",
".",
"debug",
"(",
"'-> %r'",
",",
"result",
")",
"return",
"result"
] | Receive a single line from the socket. | [
"Receive",
"a",
"single",
"line",
"from",
"the",
"socket",
"."
] | 513b1259c4d7aca5f88cd958edc11828d0712f1b | https://github.com/romanz/trezor-agent/blob/513b1259c4d7aca5f88cd958edc11828d0712f1b/libagent/gpg/keyring.py#L57-L72 | train |
romanz/trezor-agent | libagent/gpg/keyring.py | parse_term | def parse_term(s):
"""Parse single s-expr term from bytes."""
size, s = s.split(b':', 1)
size = int(size)
return s[:size], s[size:] | python | def parse_term(s):
"""Parse single s-expr term from bytes."""
size, s = s.split(b':', 1)
size = int(size)
return s[:size], s[size:] | [
"def",
"parse_term",
"(",
"s",
")",
":",
"size",
",",
"s",
"=",
"s",
".",
"split",
"(",
"b':'",
",",
"1",
")",
"size",
"=",
"int",
"(",
"size",
")",
"return",
"s",
"[",
":",
"size",
"]",
",",
"s",
"[",
"size",
":",
"]"
] | Parse single s-expr term from bytes. | [
"Parse",
"single",
"s",
"-",
"expr",
"term",
"from",
"bytes",
"."
] | 513b1259c4d7aca5f88cd958edc11828d0712f1b | https://github.com/romanz/trezor-agent/blob/513b1259c4d7aca5f88cd958edc11828d0712f1b/libagent/gpg/keyring.py#L97-L101 | train |
romanz/trezor-agent | libagent/gpg/keyring.py | parse | def parse(s):
"""Parse full s-expr from bytes."""
if s.startswith(b'('):
s = s[1:]
name, s = parse_term(s)
values = [name]
while not s.startswith(b')'):
value, s = parse(s)
values.append(value)
return values, s[1:]
return parse_term(s) | python | def parse(s):
"""Parse full s-expr from bytes."""
if s.startswith(b'('):
s = s[1:]
name, s = parse_term(s)
values = [name]
while not s.startswith(b')'):
value, s = parse(s)
values.append(value)
return values, s[1:]
return parse_term(s) | [
"def",
"parse",
"(",
"s",
")",
":",
"if",
"s",
".",
"startswith",
"(",
"b'('",
")",
":",
"s",
"=",
"s",
"[",
"1",
":",
"]",
"name",
",",
"s",
"=",
"parse_term",
"(",
"s",
")",
"values",
"=",
"[",
"name",
"]",
"while",
"not",
"s",
".",
"startswith",
"(",
"b')'",
")",
":",
"value",
",",
"s",
"=",
"parse",
"(",
"s",
")",
"values",
".",
"append",
"(",
"value",
")",
"return",
"values",
",",
"s",
"[",
"1",
":",
"]",
"return",
"parse_term",
"(",
"s",
")"
] | Parse full s-expr from bytes. | [
"Parse",
"full",
"s",
"-",
"expr",
"from",
"bytes",
"."
] | 513b1259c4d7aca5f88cd958edc11828d0712f1b | https://github.com/romanz/trezor-agent/blob/513b1259c4d7aca5f88cd958edc11828d0712f1b/libagent/gpg/keyring.py#L104-L115 | train |
romanz/trezor-agent | libagent/gpg/keyring.py | parse_sig | def parse_sig(sig):
"""Parse signature integer values from s-expr."""
label, sig = sig
assert label == b'sig-val'
algo_name = sig[0]
parser = {b'rsa': _parse_rsa_sig,
b'ecdsa': _parse_ecdsa_sig,
b'eddsa': _parse_eddsa_sig,
b'dsa': _parse_dsa_sig}[algo_name]
return parser(args=sig[1:]) | python | def parse_sig(sig):
"""Parse signature integer values from s-expr."""
label, sig = sig
assert label == b'sig-val'
algo_name = sig[0]
parser = {b'rsa': _parse_rsa_sig,
b'ecdsa': _parse_ecdsa_sig,
b'eddsa': _parse_eddsa_sig,
b'dsa': _parse_dsa_sig}[algo_name]
return parser(args=sig[1:]) | [
"def",
"parse_sig",
"(",
"sig",
")",
":",
"label",
",",
"sig",
"=",
"sig",
"assert",
"label",
"==",
"b'sig-val'",
"algo_name",
"=",
"sig",
"[",
"0",
"]",
"parser",
"=",
"{",
"b'rsa'",
":",
"_parse_rsa_sig",
",",
"b'ecdsa'",
":",
"_parse_ecdsa_sig",
",",
"b'eddsa'",
":",
"_parse_eddsa_sig",
",",
"b'dsa'",
":",
"_parse_dsa_sig",
"}",
"[",
"algo_name",
"]",
"return",
"parser",
"(",
"args",
"=",
"sig",
"[",
"1",
":",
"]",
")"
] | Parse signature integer values from s-expr. | [
"Parse",
"signature",
"integer",
"values",
"from",
"s",
"-",
"expr",
"."
] | 513b1259c4d7aca5f88cd958edc11828d0712f1b | https://github.com/romanz/trezor-agent/blob/513b1259c4d7aca5f88cd958edc11828d0712f1b/libagent/gpg/keyring.py#L137-L146 | train |
romanz/trezor-agent | libagent/gpg/keyring.py | sign_digest | def sign_digest(sock, keygrip, digest, sp=subprocess, environ=None):
"""Sign a digest using specified key using GPG agent."""
hash_algo = 8 # SHA256
assert len(digest) == 32
assert communicate(sock, 'RESET').startswith(b'OK')
ttyname = check_output(args=['tty'], sp=sp).strip()
options = ['ttyname={}'.format(ttyname)] # set TTY for passphrase entry
display = (environ or os.environ).get('DISPLAY')
if display is not None:
options.append('display={}'.format(display))
for opt in options:
assert communicate(sock, 'OPTION {}'.format(opt)) == b'OK'
assert communicate(sock, 'SIGKEY {}'.format(keygrip)) == b'OK'
hex_digest = binascii.hexlify(digest).upper().decode('ascii')
assert communicate(sock, 'SETHASH {} {}'.format(hash_algo,
hex_digest)) == b'OK'
assert communicate(sock, 'SETKEYDESC '
'Sign+a+new+TREZOR-based+subkey') == b'OK'
assert communicate(sock, 'PKSIGN') == b'OK'
while True:
line = recvline(sock).strip()
if line.startswith(b'S PROGRESS'):
continue
else:
break
line = unescape(line)
log.debug('unescaped: %r', line)
prefix, sig = line.split(b' ', 1)
if prefix != b'D':
raise ValueError(prefix)
sig, leftover = parse(sig)
assert not leftover, leftover
return parse_sig(sig) | python | def sign_digest(sock, keygrip, digest, sp=subprocess, environ=None):
"""Sign a digest using specified key using GPG agent."""
hash_algo = 8 # SHA256
assert len(digest) == 32
assert communicate(sock, 'RESET').startswith(b'OK')
ttyname = check_output(args=['tty'], sp=sp).strip()
options = ['ttyname={}'.format(ttyname)] # set TTY for passphrase entry
display = (environ or os.environ).get('DISPLAY')
if display is not None:
options.append('display={}'.format(display))
for opt in options:
assert communicate(sock, 'OPTION {}'.format(opt)) == b'OK'
assert communicate(sock, 'SIGKEY {}'.format(keygrip)) == b'OK'
hex_digest = binascii.hexlify(digest).upper().decode('ascii')
assert communicate(sock, 'SETHASH {} {}'.format(hash_algo,
hex_digest)) == b'OK'
assert communicate(sock, 'SETKEYDESC '
'Sign+a+new+TREZOR-based+subkey') == b'OK'
assert communicate(sock, 'PKSIGN') == b'OK'
while True:
line = recvline(sock).strip()
if line.startswith(b'S PROGRESS'):
continue
else:
break
line = unescape(line)
log.debug('unescaped: %r', line)
prefix, sig = line.split(b' ', 1)
if prefix != b'D':
raise ValueError(prefix)
sig, leftover = parse(sig)
assert not leftover, leftover
return parse_sig(sig) | [
"def",
"sign_digest",
"(",
"sock",
",",
"keygrip",
",",
"digest",
",",
"sp",
"=",
"subprocess",
",",
"environ",
"=",
"None",
")",
":",
"hash_algo",
"=",
"8",
"# SHA256",
"assert",
"len",
"(",
"digest",
")",
"==",
"32",
"assert",
"communicate",
"(",
"sock",
",",
"'RESET'",
")",
".",
"startswith",
"(",
"b'OK'",
")",
"ttyname",
"=",
"check_output",
"(",
"args",
"=",
"[",
"'tty'",
"]",
",",
"sp",
"=",
"sp",
")",
".",
"strip",
"(",
")",
"options",
"=",
"[",
"'ttyname={}'",
".",
"format",
"(",
"ttyname",
")",
"]",
"# set TTY for passphrase entry",
"display",
"=",
"(",
"environ",
"or",
"os",
".",
"environ",
")",
".",
"get",
"(",
"'DISPLAY'",
")",
"if",
"display",
"is",
"not",
"None",
":",
"options",
".",
"append",
"(",
"'display={}'",
".",
"format",
"(",
"display",
")",
")",
"for",
"opt",
"in",
"options",
":",
"assert",
"communicate",
"(",
"sock",
",",
"'OPTION {}'",
".",
"format",
"(",
"opt",
")",
")",
"==",
"b'OK'",
"assert",
"communicate",
"(",
"sock",
",",
"'SIGKEY {}'",
".",
"format",
"(",
"keygrip",
")",
")",
"==",
"b'OK'",
"hex_digest",
"=",
"binascii",
".",
"hexlify",
"(",
"digest",
")",
".",
"upper",
"(",
")",
".",
"decode",
"(",
"'ascii'",
")",
"assert",
"communicate",
"(",
"sock",
",",
"'SETHASH {} {}'",
".",
"format",
"(",
"hash_algo",
",",
"hex_digest",
")",
")",
"==",
"b'OK'",
"assert",
"communicate",
"(",
"sock",
",",
"'SETKEYDESC '",
"'Sign+a+new+TREZOR-based+subkey'",
")",
"==",
"b'OK'",
"assert",
"communicate",
"(",
"sock",
",",
"'PKSIGN'",
")",
"==",
"b'OK'",
"while",
"True",
":",
"line",
"=",
"recvline",
"(",
"sock",
")",
".",
"strip",
"(",
")",
"if",
"line",
".",
"startswith",
"(",
"b'S PROGRESS'",
")",
":",
"continue",
"else",
":",
"break",
"line",
"=",
"unescape",
"(",
"line",
")",
"log",
".",
"debug",
"(",
"'unescaped: %r'",
",",
"line",
")",
"prefix",
",",
"sig",
"=",
"line",
".",
"split",
"(",
"b' '",
",",
"1",
")",
"if",
"prefix",
"!=",
"b'D'",
":",
"raise",
"ValueError",
"(",
"prefix",
")",
"sig",
",",
"leftover",
"=",
"parse",
"(",
"sig",
")",
"assert",
"not",
"leftover",
",",
"leftover",
"return",
"parse_sig",
"(",
"sig",
")"
] | Sign a digest using specified key using GPG agent. | [
"Sign",
"a",
"digest",
"using",
"specified",
"key",
"using",
"GPG",
"agent",
"."
] | 513b1259c4d7aca5f88cd958edc11828d0712f1b | https://github.com/romanz/trezor-agent/blob/513b1259c4d7aca5f88cd958edc11828d0712f1b/libagent/gpg/keyring.py#L149-L188 | train |
romanz/trezor-agent | libagent/gpg/keyring.py | get_gnupg_components | def get_gnupg_components(sp=subprocess):
"""Parse GnuPG components' paths."""
args = [util.which('gpgconf'), '--list-components']
output = check_output(args=args, sp=sp)
components = dict(re.findall('(.*):.*:(.*)', output.decode('utf-8')))
log.debug('gpgconf --list-components: %s', components)
return components | python | def get_gnupg_components(sp=subprocess):
"""Parse GnuPG components' paths."""
args = [util.which('gpgconf'), '--list-components']
output = check_output(args=args, sp=sp)
components = dict(re.findall('(.*):.*:(.*)', output.decode('utf-8')))
log.debug('gpgconf --list-components: %s', components)
return components | [
"def",
"get_gnupg_components",
"(",
"sp",
"=",
"subprocess",
")",
":",
"args",
"=",
"[",
"util",
".",
"which",
"(",
"'gpgconf'",
")",
",",
"'--list-components'",
"]",
"output",
"=",
"check_output",
"(",
"args",
"=",
"args",
",",
"sp",
"=",
"sp",
")",
"components",
"=",
"dict",
"(",
"re",
".",
"findall",
"(",
"'(.*):.*:(.*)'",
",",
"output",
".",
"decode",
"(",
"'utf-8'",
")",
")",
")",
"log",
".",
"debug",
"(",
"'gpgconf --list-components: %s'",
",",
"components",
")",
"return",
"components"
] | Parse GnuPG components' paths. | [
"Parse",
"GnuPG",
"components",
"paths",
"."
] | 513b1259c4d7aca5f88cd958edc11828d0712f1b | https://github.com/romanz/trezor-agent/blob/513b1259c4d7aca5f88cd958edc11828d0712f1b/libagent/gpg/keyring.py#L191-L197 | train |
romanz/trezor-agent | libagent/gpg/keyring.py | gpg_command | def gpg_command(args, env=None):
"""Prepare common GPG command line arguments."""
if env is None:
env = os.environ
cmd = get_gnupg_binary(neopg_binary=env.get('NEOPG_BINARY'))
return [cmd] + args | python | def gpg_command(args, env=None):
"""Prepare common GPG command line arguments."""
if env is None:
env = os.environ
cmd = get_gnupg_binary(neopg_binary=env.get('NEOPG_BINARY'))
return [cmd] + args | [
"def",
"gpg_command",
"(",
"args",
",",
"env",
"=",
"None",
")",
":",
"if",
"env",
"is",
"None",
":",
"env",
"=",
"os",
".",
"environ",
"cmd",
"=",
"get_gnupg_binary",
"(",
"neopg_binary",
"=",
"env",
".",
"get",
"(",
"'NEOPG_BINARY'",
")",
")",
"return",
"[",
"cmd",
"]",
"+",
"args"
] | Prepare common GPG command line arguments. | [
"Prepare",
"common",
"GPG",
"command",
"line",
"arguments",
"."
] | 513b1259c4d7aca5f88cd958edc11828d0712f1b | https://github.com/romanz/trezor-agent/blob/513b1259c4d7aca5f88cd958edc11828d0712f1b/libagent/gpg/keyring.py#L208-L213 | train |
romanz/trezor-agent | libagent/gpg/keyring.py | export_public_key | def export_public_key(user_id, env=None, sp=subprocess):
"""Export GPG public key for specified `user_id`."""
args = gpg_command(['--export', user_id])
result = check_output(args=args, env=env, sp=sp)
if not result:
log.error('could not find public key %r in local GPG keyring', user_id)
raise KeyError(user_id)
return result | python | def export_public_key(user_id, env=None, sp=subprocess):
"""Export GPG public key for specified `user_id`."""
args = gpg_command(['--export', user_id])
result = check_output(args=args, env=env, sp=sp)
if not result:
log.error('could not find public key %r in local GPG keyring', user_id)
raise KeyError(user_id)
return result | [
"def",
"export_public_key",
"(",
"user_id",
",",
"env",
"=",
"None",
",",
"sp",
"=",
"subprocess",
")",
":",
"args",
"=",
"gpg_command",
"(",
"[",
"'--export'",
",",
"user_id",
"]",
")",
"result",
"=",
"check_output",
"(",
"args",
"=",
"args",
",",
"env",
"=",
"env",
",",
"sp",
"=",
"sp",
")",
"if",
"not",
"result",
":",
"log",
".",
"error",
"(",
"'could not find public key %r in local GPG keyring'",
",",
"user_id",
")",
"raise",
"KeyError",
"(",
"user_id",
")",
"return",
"result"
] | Export GPG public key for specified `user_id`. | [
"Export",
"GPG",
"public",
"key",
"for",
"specified",
"user_id",
"."
] | 513b1259c4d7aca5f88cd958edc11828d0712f1b | https://github.com/romanz/trezor-agent/blob/513b1259c4d7aca5f88cd958edc11828d0712f1b/libagent/gpg/keyring.py#L233-L240 | train |
romanz/trezor-agent | libagent/gpg/keyring.py | export_public_keys | def export_public_keys(env=None, sp=subprocess):
"""Export all GPG public keys."""
args = gpg_command(['--export'])
result = check_output(args=args, env=env, sp=sp)
if not result:
raise KeyError('No GPG public keys found at env: {!r}'.format(env))
return result | python | def export_public_keys(env=None, sp=subprocess):
"""Export all GPG public keys."""
args = gpg_command(['--export'])
result = check_output(args=args, env=env, sp=sp)
if not result:
raise KeyError('No GPG public keys found at env: {!r}'.format(env))
return result | [
"def",
"export_public_keys",
"(",
"env",
"=",
"None",
",",
"sp",
"=",
"subprocess",
")",
":",
"args",
"=",
"gpg_command",
"(",
"[",
"'--export'",
"]",
")",
"result",
"=",
"check_output",
"(",
"args",
"=",
"args",
",",
"env",
"=",
"env",
",",
"sp",
"=",
"sp",
")",
"if",
"not",
"result",
":",
"raise",
"KeyError",
"(",
"'No GPG public keys found at env: {!r}'",
".",
"format",
"(",
"env",
")",
")",
"return",
"result"
] | Export all GPG public keys. | [
"Export",
"all",
"GPG",
"public",
"keys",
"."
] | 513b1259c4d7aca5f88cd958edc11828d0712f1b | https://github.com/romanz/trezor-agent/blob/513b1259c4d7aca5f88cd958edc11828d0712f1b/libagent/gpg/keyring.py#L243-L249 | train |
romanz/trezor-agent | libagent/gpg/keyring.py | create_agent_signer | def create_agent_signer(user_id):
"""Sign digest with existing GPG keys using gpg-agent tool."""
sock = connect_to_agent(env=os.environ)
keygrip = get_keygrip(user_id)
def sign(digest):
"""Sign the digest and return an ECDSA/RSA/DSA signature."""
return sign_digest(sock=sock, keygrip=keygrip, digest=digest)
return sign | python | def create_agent_signer(user_id):
"""Sign digest with existing GPG keys using gpg-agent tool."""
sock = connect_to_agent(env=os.environ)
keygrip = get_keygrip(user_id)
def sign(digest):
"""Sign the digest and return an ECDSA/RSA/DSA signature."""
return sign_digest(sock=sock, keygrip=keygrip, digest=digest)
return sign | [
"def",
"create_agent_signer",
"(",
"user_id",
")",
":",
"sock",
"=",
"connect_to_agent",
"(",
"env",
"=",
"os",
".",
"environ",
")",
"keygrip",
"=",
"get_keygrip",
"(",
"user_id",
")",
"def",
"sign",
"(",
"digest",
")",
":",
"\"\"\"Sign the digest and return an ECDSA/RSA/DSA signature.\"\"\"",
"return",
"sign_digest",
"(",
"sock",
"=",
"sock",
",",
"keygrip",
"=",
"keygrip",
",",
"digest",
"=",
"digest",
")",
"return",
"sign"
] | Sign digest with existing GPG keys using gpg-agent tool. | [
"Sign",
"digest",
"with",
"existing",
"GPG",
"keys",
"using",
"gpg",
"-",
"agent",
"tool",
"."
] | 513b1259c4d7aca5f88cd958edc11828d0712f1b | https://github.com/romanz/trezor-agent/blob/513b1259c4d7aca5f88cd958edc11828d0712f1b/libagent/gpg/keyring.py#L252-L261 | train |
romanz/trezor-agent | libagent/ssh/protocol.py | msg_name | def msg_name(code):
"""Convert integer message code into a string name."""
ids = {v: k for k, v in COMMANDS.items()}
return ids[code] | python | def msg_name(code):
"""Convert integer message code into a string name."""
ids = {v: k for k, v in COMMANDS.items()}
return ids[code] | [
"def",
"msg_name",
"(",
"code",
")",
":",
"ids",
"=",
"{",
"v",
":",
"k",
"for",
"k",
",",
"v",
"in",
"COMMANDS",
".",
"items",
"(",
")",
"}",
"return",
"ids",
"[",
"code",
"]"
] | Convert integer message code into a string name. | [
"Convert",
"integer",
"message",
"code",
"into",
"a",
"string",
"name",
"."
] | 513b1259c4d7aca5f88cd958edc11828d0712f1b | https://github.com/romanz/trezor-agent/blob/513b1259c4d7aca5f88cd958edc11828d0712f1b/libagent/ssh/protocol.py#L51-L54 | train |
romanz/trezor-agent | libagent/ssh/protocol.py | _legacy_pubs | def _legacy_pubs(buf):
"""SSH v1 public keys are not supported."""
leftover = buf.read()
if leftover:
log.warning('skipping leftover: %r', leftover)
code = util.pack('B', msg_code('SSH_AGENT_RSA_IDENTITIES_ANSWER'))
num = util.pack('L', 0) # no SSH v1 keys
return util.frame(code, num) | python | def _legacy_pubs(buf):
"""SSH v1 public keys are not supported."""
leftover = buf.read()
if leftover:
log.warning('skipping leftover: %r', leftover)
code = util.pack('B', msg_code('SSH_AGENT_RSA_IDENTITIES_ANSWER'))
num = util.pack('L', 0) # no SSH v1 keys
return util.frame(code, num) | [
"def",
"_legacy_pubs",
"(",
"buf",
")",
":",
"leftover",
"=",
"buf",
".",
"read",
"(",
")",
"if",
"leftover",
":",
"log",
".",
"warning",
"(",
"'skipping leftover: %r'",
",",
"leftover",
")",
"code",
"=",
"util",
".",
"pack",
"(",
"'B'",
",",
"msg_code",
"(",
"'SSH_AGENT_RSA_IDENTITIES_ANSWER'",
")",
")",
"num",
"=",
"util",
".",
"pack",
"(",
"'L'",
",",
"0",
")",
"# no SSH v1 keys",
"return",
"util",
".",
"frame",
"(",
"code",
",",
"num",
")"
] | SSH v1 public keys are not supported. | [
"SSH",
"v1",
"public",
"keys",
"are",
"not",
"supported",
"."
] | 513b1259c4d7aca5f88cd958edc11828d0712f1b | https://github.com/romanz/trezor-agent/blob/513b1259c4d7aca5f88cd958edc11828d0712f1b/libagent/ssh/protocol.py#L63-L70 | train |
romanz/trezor-agent | libagent/ssh/protocol.py | Handler.handle | def handle(self, msg):
"""Handle SSH message from the SSH client and return the response."""
debug_msg = ': {!r}'.format(msg) if self.debug else ''
log.debug('request: %d bytes%s', len(msg), debug_msg)
buf = io.BytesIO(msg)
code, = util.recv(buf, '>B')
if code not in self.methods:
log.warning('Unsupported command: %s (%d)', msg_name(code), code)
return failure()
method = self.methods[code]
log.debug('calling %s()', method.__name__)
reply = method(buf=buf)
debug_reply = ': {!r}'.format(reply) if self.debug else ''
log.debug('reply: %d bytes%s', len(reply), debug_reply)
return reply | python | def handle(self, msg):
"""Handle SSH message from the SSH client and return the response."""
debug_msg = ': {!r}'.format(msg) if self.debug else ''
log.debug('request: %d bytes%s', len(msg), debug_msg)
buf = io.BytesIO(msg)
code, = util.recv(buf, '>B')
if code not in self.methods:
log.warning('Unsupported command: %s (%d)', msg_name(code), code)
return failure()
method = self.methods[code]
log.debug('calling %s()', method.__name__)
reply = method(buf=buf)
debug_reply = ': {!r}'.format(reply) if self.debug else ''
log.debug('reply: %d bytes%s', len(reply), debug_reply)
return reply | [
"def",
"handle",
"(",
"self",
",",
"msg",
")",
":",
"debug_msg",
"=",
"': {!r}'",
".",
"format",
"(",
"msg",
")",
"if",
"self",
".",
"debug",
"else",
"''",
"log",
".",
"debug",
"(",
"'request: %d bytes%s'",
",",
"len",
"(",
"msg",
")",
",",
"debug_msg",
")",
"buf",
"=",
"io",
".",
"BytesIO",
"(",
"msg",
")",
"code",
",",
"=",
"util",
".",
"recv",
"(",
"buf",
",",
"'>B'",
")",
"if",
"code",
"not",
"in",
"self",
".",
"methods",
":",
"log",
".",
"warning",
"(",
"'Unsupported command: %s (%d)'",
",",
"msg_name",
"(",
"code",
")",
",",
"code",
")",
"return",
"failure",
"(",
")",
"method",
"=",
"self",
".",
"methods",
"[",
"code",
"]",
"log",
".",
"debug",
"(",
"'calling %s()'",
",",
"method",
".",
"__name__",
")",
"reply",
"=",
"method",
"(",
"buf",
"=",
"buf",
")",
"debug_reply",
"=",
"': {!r}'",
".",
"format",
"(",
"reply",
")",
"if",
"self",
".",
"debug",
"else",
"''",
"log",
".",
"debug",
"(",
"'reply: %d bytes%s'",
",",
"len",
"(",
"reply",
")",
",",
"debug_reply",
")",
"return",
"reply"
] | Handle SSH message from the SSH client and return the response. | [
"Handle",
"SSH",
"message",
"from",
"the",
"SSH",
"client",
"and",
"return",
"the",
"response",
"."
] | 513b1259c4d7aca5f88cd958edc11828d0712f1b | https://github.com/romanz/trezor-agent/blob/513b1259c4d7aca5f88cd958edc11828d0712f1b/libagent/ssh/protocol.py#L91-L106 | train |
romanz/trezor-agent | libagent/ssh/protocol.py | Handler.list_pubs | def list_pubs(self, buf):
"""SSH v2 public keys are serialized and returned."""
assert not buf.read()
keys = self.conn.parse_public_keys()
code = util.pack('B', msg_code('SSH2_AGENT_IDENTITIES_ANSWER'))
num = util.pack('L', len(keys))
log.debug('available keys: %s', [k['name'] for k in keys])
for i, k in enumerate(keys):
log.debug('%2d) %s', i+1, k['fingerprint'])
pubs = [util.frame(k['blob']) + util.frame(k['name']) for k in keys]
return util.frame(code, num, *pubs) | python | def list_pubs(self, buf):
"""SSH v2 public keys are serialized and returned."""
assert not buf.read()
keys = self.conn.parse_public_keys()
code = util.pack('B', msg_code('SSH2_AGENT_IDENTITIES_ANSWER'))
num = util.pack('L', len(keys))
log.debug('available keys: %s', [k['name'] for k in keys])
for i, k in enumerate(keys):
log.debug('%2d) %s', i+1, k['fingerprint'])
pubs = [util.frame(k['blob']) + util.frame(k['name']) for k in keys]
return util.frame(code, num, *pubs) | [
"def",
"list_pubs",
"(",
"self",
",",
"buf",
")",
":",
"assert",
"not",
"buf",
".",
"read",
"(",
")",
"keys",
"=",
"self",
".",
"conn",
".",
"parse_public_keys",
"(",
")",
"code",
"=",
"util",
".",
"pack",
"(",
"'B'",
",",
"msg_code",
"(",
"'SSH2_AGENT_IDENTITIES_ANSWER'",
")",
")",
"num",
"=",
"util",
".",
"pack",
"(",
"'L'",
",",
"len",
"(",
"keys",
")",
")",
"log",
".",
"debug",
"(",
"'available keys: %s'",
",",
"[",
"k",
"[",
"'name'",
"]",
"for",
"k",
"in",
"keys",
"]",
")",
"for",
"i",
",",
"k",
"in",
"enumerate",
"(",
"keys",
")",
":",
"log",
".",
"debug",
"(",
"'%2d) %s'",
",",
"i",
"+",
"1",
",",
"k",
"[",
"'fingerprint'",
"]",
")",
"pubs",
"=",
"[",
"util",
".",
"frame",
"(",
"k",
"[",
"'blob'",
"]",
")",
"+",
"util",
".",
"frame",
"(",
"k",
"[",
"'name'",
"]",
")",
"for",
"k",
"in",
"keys",
"]",
"return",
"util",
".",
"frame",
"(",
"code",
",",
"num",
",",
"*",
"pubs",
")"
] | SSH v2 public keys are serialized and returned. | [
"SSH",
"v2",
"public",
"keys",
"are",
"serialized",
"and",
"returned",
"."
] | 513b1259c4d7aca5f88cd958edc11828d0712f1b | https://github.com/romanz/trezor-agent/blob/513b1259c4d7aca5f88cd958edc11828d0712f1b/libagent/ssh/protocol.py#L108-L118 | train |
romanz/trezor-agent | libagent/ssh/protocol.py | Handler.sign_message | def sign_message(self, buf):
"""
SSH v2 public key authentication is performed.
If the required key is not supported, raise KeyError
If the signature is invalid, raise ValueError
"""
key = formats.parse_pubkey(util.read_frame(buf))
log.debug('looking for %s', key['fingerprint'])
blob = util.read_frame(buf)
assert util.read_frame(buf) == b''
assert not buf.read()
for k in self.conn.parse_public_keys():
if (k['fingerprint']) == (key['fingerprint']):
log.debug('using key %r (%s)', k['name'], k['fingerprint'])
key = k
break
else:
raise KeyError('key not found')
label = key['name'].decode('utf-8')
log.debug('signing %d-byte blob with "%s" key', len(blob), label)
try:
signature = self.conn.sign(blob=blob, identity=key['identity'])
except IOError:
return failure()
log.debug('signature: %r', signature)
try:
sig_bytes = key['verifier'](sig=signature, msg=blob)
log.info('signature status: OK')
except formats.ecdsa.BadSignatureError:
log.exception('signature status: ERROR')
raise ValueError('invalid ECDSA signature')
log.debug('signature size: %d bytes', len(sig_bytes))
data = util.frame(util.frame(key['type']), util.frame(sig_bytes))
code = util.pack('B', msg_code('SSH2_AGENT_SIGN_RESPONSE'))
return util.frame(code, data) | python | def sign_message(self, buf):
"""
SSH v2 public key authentication is performed.
If the required key is not supported, raise KeyError
If the signature is invalid, raise ValueError
"""
key = formats.parse_pubkey(util.read_frame(buf))
log.debug('looking for %s', key['fingerprint'])
blob = util.read_frame(buf)
assert util.read_frame(buf) == b''
assert not buf.read()
for k in self.conn.parse_public_keys():
if (k['fingerprint']) == (key['fingerprint']):
log.debug('using key %r (%s)', k['name'], k['fingerprint'])
key = k
break
else:
raise KeyError('key not found')
label = key['name'].decode('utf-8')
log.debug('signing %d-byte blob with "%s" key', len(blob), label)
try:
signature = self.conn.sign(blob=blob, identity=key['identity'])
except IOError:
return failure()
log.debug('signature: %r', signature)
try:
sig_bytes = key['verifier'](sig=signature, msg=blob)
log.info('signature status: OK')
except formats.ecdsa.BadSignatureError:
log.exception('signature status: ERROR')
raise ValueError('invalid ECDSA signature')
log.debug('signature size: %d bytes', len(sig_bytes))
data = util.frame(util.frame(key['type']), util.frame(sig_bytes))
code = util.pack('B', msg_code('SSH2_AGENT_SIGN_RESPONSE'))
return util.frame(code, data) | [
"def",
"sign_message",
"(",
"self",
",",
"buf",
")",
":",
"key",
"=",
"formats",
".",
"parse_pubkey",
"(",
"util",
".",
"read_frame",
"(",
"buf",
")",
")",
"log",
".",
"debug",
"(",
"'looking for %s'",
",",
"key",
"[",
"'fingerprint'",
"]",
")",
"blob",
"=",
"util",
".",
"read_frame",
"(",
"buf",
")",
"assert",
"util",
".",
"read_frame",
"(",
"buf",
")",
"==",
"b''",
"assert",
"not",
"buf",
".",
"read",
"(",
")",
"for",
"k",
"in",
"self",
".",
"conn",
".",
"parse_public_keys",
"(",
")",
":",
"if",
"(",
"k",
"[",
"'fingerprint'",
"]",
")",
"==",
"(",
"key",
"[",
"'fingerprint'",
"]",
")",
":",
"log",
".",
"debug",
"(",
"'using key %r (%s)'",
",",
"k",
"[",
"'name'",
"]",
",",
"k",
"[",
"'fingerprint'",
"]",
")",
"key",
"=",
"k",
"break",
"else",
":",
"raise",
"KeyError",
"(",
"'key not found'",
")",
"label",
"=",
"key",
"[",
"'name'",
"]",
".",
"decode",
"(",
"'utf-8'",
")",
"log",
".",
"debug",
"(",
"'signing %d-byte blob with \"%s\" key'",
",",
"len",
"(",
"blob",
")",
",",
"label",
")",
"try",
":",
"signature",
"=",
"self",
".",
"conn",
".",
"sign",
"(",
"blob",
"=",
"blob",
",",
"identity",
"=",
"key",
"[",
"'identity'",
"]",
")",
"except",
"IOError",
":",
"return",
"failure",
"(",
")",
"log",
".",
"debug",
"(",
"'signature: %r'",
",",
"signature",
")",
"try",
":",
"sig_bytes",
"=",
"key",
"[",
"'verifier'",
"]",
"(",
"sig",
"=",
"signature",
",",
"msg",
"=",
"blob",
")",
"log",
".",
"info",
"(",
"'signature status: OK'",
")",
"except",
"formats",
".",
"ecdsa",
".",
"BadSignatureError",
":",
"log",
".",
"exception",
"(",
"'signature status: ERROR'",
")",
"raise",
"ValueError",
"(",
"'invalid ECDSA signature'",
")",
"log",
".",
"debug",
"(",
"'signature size: %d bytes'",
",",
"len",
"(",
"sig_bytes",
")",
")",
"data",
"=",
"util",
".",
"frame",
"(",
"util",
".",
"frame",
"(",
"key",
"[",
"'type'",
"]",
")",
",",
"util",
".",
"frame",
"(",
"sig_bytes",
")",
")",
"code",
"=",
"util",
".",
"pack",
"(",
"'B'",
",",
"msg_code",
"(",
"'SSH2_AGENT_SIGN_RESPONSE'",
")",
")",
"return",
"util",
".",
"frame",
"(",
"code",
",",
"data",
")"
] | SSH v2 public key authentication is performed.
If the required key is not supported, raise KeyError
If the signature is invalid, raise ValueError | [
"SSH",
"v2",
"public",
"key",
"authentication",
"is",
"performed",
"."
] | 513b1259c4d7aca5f88cd958edc11828d0712f1b | https://github.com/romanz/trezor-agent/blob/513b1259c4d7aca5f88cd958edc11828d0712f1b/libagent/ssh/protocol.py#L120-L160 | train |
romanz/trezor-agent | libagent/util.py | recv | def recv(conn, size):
"""
Receive bytes from connection socket or stream.
If size is struct.calcsize()-compatible format, use it to unpack the data.
Otherwise, return the plain blob as bytes.
"""
try:
fmt = size
size = struct.calcsize(fmt)
except TypeError:
fmt = None
try:
_read = conn.recv
except AttributeError:
_read = conn.read
res = io.BytesIO()
while size > 0:
buf = _read(size)
if not buf:
raise EOFError
size = size - len(buf)
res.write(buf)
res = res.getvalue()
if fmt:
return struct.unpack(fmt, res)
else:
return res | python | def recv(conn, size):
"""
Receive bytes from connection socket or stream.
If size is struct.calcsize()-compatible format, use it to unpack the data.
Otherwise, return the plain blob as bytes.
"""
try:
fmt = size
size = struct.calcsize(fmt)
except TypeError:
fmt = None
try:
_read = conn.recv
except AttributeError:
_read = conn.read
res = io.BytesIO()
while size > 0:
buf = _read(size)
if not buf:
raise EOFError
size = size - len(buf)
res.write(buf)
res = res.getvalue()
if fmt:
return struct.unpack(fmt, res)
else:
return res | [
"def",
"recv",
"(",
"conn",
",",
"size",
")",
":",
"try",
":",
"fmt",
"=",
"size",
"size",
"=",
"struct",
".",
"calcsize",
"(",
"fmt",
")",
"except",
"TypeError",
":",
"fmt",
"=",
"None",
"try",
":",
"_read",
"=",
"conn",
".",
"recv",
"except",
"AttributeError",
":",
"_read",
"=",
"conn",
".",
"read",
"res",
"=",
"io",
".",
"BytesIO",
"(",
")",
"while",
"size",
">",
"0",
":",
"buf",
"=",
"_read",
"(",
"size",
")",
"if",
"not",
"buf",
":",
"raise",
"EOFError",
"size",
"=",
"size",
"-",
"len",
"(",
"buf",
")",
"res",
".",
"write",
"(",
"buf",
")",
"res",
"=",
"res",
".",
"getvalue",
"(",
")",
"if",
"fmt",
":",
"return",
"struct",
".",
"unpack",
"(",
"fmt",
",",
"res",
")",
"else",
":",
"return",
"res"
] | Receive bytes from connection socket or stream.
If size is struct.calcsize()-compatible format, use it to unpack the data.
Otherwise, return the plain blob as bytes. | [
"Receive",
"bytes",
"from",
"connection",
"socket",
"or",
"stream",
"."
] | 513b1259c4d7aca5f88cd958edc11828d0712f1b | https://github.com/romanz/trezor-agent/blob/513b1259c4d7aca5f88cd958edc11828d0712f1b/libagent/util.py#L18-L46 | train |
romanz/trezor-agent | libagent/util.py | bytes2num | def bytes2num(s):
"""Convert MSB-first bytes to an unsigned integer."""
res = 0
for i, c in enumerate(reversed(bytearray(s))):
res += c << (i * 8)
return res | python | def bytes2num(s):
"""Convert MSB-first bytes to an unsigned integer."""
res = 0
for i, c in enumerate(reversed(bytearray(s))):
res += c << (i * 8)
return res | [
"def",
"bytes2num",
"(",
"s",
")",
":",
"res",
"=",
"0",
"for",
"i",
",",
"c",
"in",
"enumerate",
"(",
"reversed",
"(",
"bytearray",
"(",
"s",
")",
")",
")",
":",
"res",
"+=",
"c",
"<<",
"(",
"i",
"*",
"8",
")",
"return",
"res"
] | Convert MSB-first bytes to an unsigned integer. | [
"Convert",
"MSB",
"-",
"first",
"bytes",
"to",
"an",
"unsigned",
"integer",
"."
] | 513b1259c4d7aca5f88cd958edc11828d0712f1b | https://github.com/romanz/trezor-agent/blob/513b1259c4d7aca5f88cd958edc11828d0712f1b/libagent/util.py#L55-L60 | train |
romanz/trezor-agent | libagent/util.py | num2bytes | def num2bytes(value, size):
"""Convert an unsigned integer to MSB-first bytes with specified size."""
res = []
for _ in range(size):
res.append(value & 0xFF)
value = value >> 8
assert value == 0
return bytes(bytearray(list(reversed(res)))) | python | def num2bytes(value, size):
"""Convert an unsigned integer to MSB-first bytes with specified size."""
res = []
for _ in range(size):
res.append(value & 0xFF)
value = value >> 8
assert value == 0
return bytes(bytearray(list(reversed(res)))) | [
"def",
"num2bytes",
"(",
"value",
",",
"size",
")",
":",
"res",
"=",
"[",
"]",
"for",
"_",
"in",
"range",
"(",
"size",
")",
":",
"res",
".",
"append",
"(",
"value",
"&",
"0xFF",
")",
"value",
"=",
"value",
">>",
"8",
"assert",
"value",
"==",
"0",
"return",
"bytes",
"(",
"bytearray",
"(",
"list",
"(",
"reversed",
"(",
"res",
")",
")",
")",
")"
] | Convert an unsigned integer to MSB-first bytes with specified size. | [
"Convert",
"an",
"unsigned",
"integer",
"to",
"MSB",
"-",
"first",
"bytes",
"with",
"specified",
"size",
"."
] | 513b1259c4d7aca5f88cd958edc11828d0712f1b | https://github.com/romanz/trezor-agent/blob/513b1259c4d7aca5f88cd958edc11828d0712f1b/libagent/util.py#L63-L70 | train |
romanz/trezor-agent | libagent/util.py | frame | def frame(*msgs):
"""Serialize MSB-first length-prefixed frame."""
res = io.BytesIO()
for msg in msgs:
res.write(msg)
msg = res.getvalue()
return pack('L', len(msg)) + msg | python | def frame(*msgs):
"""Serialize MSB-first length-prefixed frame."""
res = io.BytesIO()
for msg in msgs:
res.write(msg)
msg = res.getvalue()
return pack('L', len(msg)) + msg | [
"def",
"frame",
"(",
"*",
"msgs",
")",
":",
"res",
"=",
"io",
".",
"BytesIO",
"(",
")",
"for",
"msg",
"in",
"msgs",
":",
"res",
".",
"write",
"(",
"msg",
")",
"msg",
"=",
"res",
".",
"getvalue",
"(",
")",
"return",
"pack",
"(",
"'L'",
",",
"len",
"(",
"msg",
")",
")",
"+",
"msg"
] | Serialize MSB-first length-prefixed frame. | [
"Serialize",
"MSB",
"-",
"first",
"length",
"-",
"prefixed",
"frame",
"."
] | 513b1259c4d7aca5f88cd958edc11828d0712f1b | https://github.com/romanz/trezor-agent/blob/513b1259c4d7aca5f88cd958edc11828d0712f1b/libagent/util.py#L78-L84 | train |
romanz/trezor-agent | libagent/util.py | split_bits | def split_bits(value, *bits):
"""
Split integer value into list of ints, according to `bits` list.
For example, split_bits(0x1234, 4, 8, 4) == [0x1, 0x23, 0x4]
"""
result = []
for b in reversed(bits):
mask = (1 << b) - 1
result.append(value & mask)
value = value >> b
assert value == 0
result.reverse()
return result | python | def split_bits(value, *bits):
"""
Split integer value into list of ints, according to `bits` list.
For example, split_bits(0x1234, 4, 8, 4) == [0x1, 0x23, 0x4]
"""
result = []
for b in reversed(bits):
mask = (1 << b) - 1
result.append(value & mask)
value = value >> b
assert value == 0
result.reverse()
return result | [
"def",
"split_bits",
"(",
"value",
",",
"*",
"bits",
")",
":",
"result",
"=",
"[",
"]",
"for",
"b",
"in",
"reversed",
"(",
"bits",
")",
":",
"mask",
"=",
"(",
"1",
"<<",
"b",
")",
"-",
"1",
"result",
".",
"append",
"(",
"value",
"&",
"mask",
")",
"value",
"=",
"value",
">>",
"b",
"assert",
"value",
"==",
"0",
"result",
".",
"reverse",
"(",
")",
"return",
"result"
] | Split integer value into list of ints, according to `bits` list.
For example, split_bits(0x1234, 4, 8, 4) == [0x1, 0x23, 0x4] | [
"Split",
"integer",
"value",
"into",
"list",
"of",
"ints",
"according",
"to",
"bits",
"list",
"."
] | 513b1259c4d7aca5f88cd958edc11828d0712f1b | https://github.com/romanz/trezor-agent/blob/513b1259c4d7aca5f88cd958edc11828d0712f1b/libagent/util.py#L115-L129 | train |
romanz/trezor-agent | libagent/util.py | readfmt | def readfmt(stream, fmt):
"""Read and unpack an object from stream, using a struct format string."""
size = struct.calcsize(fmt)
blob = stream.read(size)
return struct.unpack(fmt, blob) | python | def readfmt(stream, fmt):
"""Read and unpack an object from stream, using a struct format string."""
size = struct.calcsize(fmt)
blob = stream.read(size)
return struct.unpack(fmt, blob) | [
"def",
"readfmt",
"(",
"stream",
",",
"fmt",
")",
":",
"size",
"=",
"struct",
".",
"calcsize",
"(",
"fmt",
")",
"blob",
"=",
"stream",
".",
"read",
"(",
"size",
")",
"return",
"struct",
".",
"unpack",
"(",
"fmt",
",",
"blob",
")"
] | Read and unpack an object from stream, using a struct format string. | [
"Read",
"and",
"unpack",
"an",
"object",
"from",
"stream",
"using",
"a",
"struct",
"format",
"string",
"."
] | 513b1259c4d7aca5f88cd958edc11828d0712f1b | https://github.com/romanz/trezor-agent/blob/513b1259c4d7aca5f88cd958edc11828d0712f1b/libagent/util.py#L132-L136 | train |
romanz/trezor-agent | libagent/util.py | setup_logging | def setup_logging(verbosity, filename=None):
"""Configure logging for this tool."""
levels = [logging.WARNING, logging.INFO, logging.DEBUG]
level = levels[min(verbosity, len(levels) - 1)]
logging.root.setLevel(level)
fmt = logging.Formatter('%(asctime)s %(levelname)-12s %(message)-100s '
'[%(filename)s:%(lineno)d]')
hdlr = logging.StreamHandler() # stderr
hdlr.setFormatter(fmt)
logging.root.addHandler(hdlr)
if filename:
hdlr = logging.FileHandler(filename, 'a')
hdlr.setFormatter(fmt)
logging.root.addHandler(hdlr) | python | def setup_logging(verbosity, filename=None):
"""Configure logging for this tool."""
levels = [logging.WARNING, logging.INFO, logging.DEBUG]
level = levels[min(verbosity, len(levels) - 1)]
logging.root.setLevel(level)
fmt = logging.Formatter('%(asctime)s %(levelname)-12s %(message)-100s '
'[%(filename)s:%(lineno)d]')
hdlr = logging.StreamHandler() # stderr
hdlr.setFormatter(fmt)
logging.root.addHandler(hdlr)
if filename:
hdlr = logging.FileHandler(filename, 'a')
hdlr.setFormatter(fmt)
logging.root.addHandler(hdlr) | [
"def",
"setup_logging",
"(",
"verbosity",
",",
"filename",
"=",
"None",
")",
":",
"levels",
"=",
"[",
"logging",
".",
"WARNING",
",",
"logging",
".",
"INFO",
",",
"logging",
".",
"DEBUG",
"]",
"level",
"=",
"levels",
"[",
"min",
"(",
"verbosity",
",",
"len",
"(",
"levels",
")",
"-",
"1",
")",
"]",
"logging",
".",
"root",
".",
"setLevel",
"(",
"level",
")",
"fmt",
"=",
"logging",
".",
"Formatter",
"(",
"'%(asctime)s %(levelname)-12s %(message)-100s '",
"'[%(filename)s:%(lineno)d]'",
")",
"hdlr",
"=",
"logging",
".",
"StreamHandler",
"(",
")",
"# stderr",
"hdlr",
".",
"setFormatter",
"(",
"fmt",
")",
"logging",
".",
"root",
".",
"addHandler",
"(",
"hdlr",
")",
"if",
"filename",
":",
"hdlr",
"=",
"logging",
".",
"FileHandler",
"(",
"filename",
",",
"'a'",
")",
"hdlr",
".",
"setFormatter",
"(",
"fmt",
")",
"logging",
".",
"root",
".",
"addHandler",
"(",
"hdlr",
")"
] | Configure logging for this tool. | [
"Configure",
"logging",
"for",
"this",
"tool",
"."
] | 513b1259c4d7aca5f88cd958edc11828d0712f1b | https://github.com/romanz/trezor-agent/blob/513b1259c4d7aca5f88cd958edc11828d0712f1b/libagent/util.py#L183-L198 | train |
romanz/trezor-agent | libagent/util.py | which | def which(cmd):
"""Return full path to specified command, or raise OSError if missing."""
try:
# For Python 3
from shutil import which as _which
except ImportError:
# For Python 2
from backports.shutil_which import which as _which # pylint: disable=relative-import
full_path = _which(cmd)
if full_path is None:
raise OSError('Cannot find {!r} in $PATH'.format(cmd))
log.debug('which %r => %r', cmd, full_path)
return full_path | python | def which(cmd):
"""Return full path to specified command, or raise OSError if missing."""
try:
# For Python 3
from shutil import which as _which
except ImportError:
# For Python 2
from backports.shutil_which import which as _which # pylint: disable=relative-import
full_path = _which(cmd)
if full_path is None:
raise OSError('Cannot find {!r} in $PATH'.format(cmd))
log.debug('which %r => %r', cmd, full_path)
return full_path | [
"def",
"which",
"(",
"cmd",
")",
":",
"try",
":",
"# For Python 3",
"from",
"shutil",
"import",
"which",
"as",
"_which",
"except",
"ImportError",
":",
"# For Python 2",
"from",
"backports",
".",
"shutil_which",
"import",
"which",
"as",
"_which",
"# pylint: disable=relative-import",
"full_path",
"=",
"_which",
"(",
"cmd",
")",
"if",
"full_path",
"is",
"None",
":",
"raise",
"OSError",
"(",
"'Cannot find {!r} in $PATH'",
".",
"format",
"(",
"cmd",
")",
")",
"log",
".",
"debug",
"(",
"'which %r => %r'",
",",
"cmd",
",",
"full_path",
")",
"return",
"full_path"
] | Return full path to specified command, or raise OSError if missing. | [
"Return",
"full",
"path",
"to",
"specified",
"command",
"or",
"raise",
"OSError",
"if",
"missing",
"."
] | 513b1259c4d7aca5f88cd958edc11828d0712f1b | https://github.com/romanz/trezor-agent/blob/513b1259c4d7aca5f88cd958edc11828d0712f1b/libagent/util.py#L238-L250 | train |
romanz/trezor-agent | libagent/util.py | Reader.readfmt | def readfmt(self, fmt):
"""Read a specified object, using a struct format string."""
size = struct.calcsize(fmt)
blob = self.read(size)
obj, = struct.unpack(fmt, blob)
return obj | python | def readfmt(self, fmt):
"""Read a specified object, using a struct format string."""
size = struct.calcsize(fmt)
blob = self.read(size)
obj, = struct.unpack(fmt, blob)
return obj | [
"def",
"readfmt",
"(",
"self",
",",
"fmt",
")",
":",
"size",
"=",
"struct",
".",
"calcsize",
"(",
"fmt",
")",
"blob",
"=",
"self",
".",
"read",
"(",
"size",
")",
"obj",
",",
"=",
"struct",
".",
"unpack",
"(",
"fmt",
",",
"blob",
")",
"return",
"obj"
] | Read a specified object, using a struct format string. | [
"Read",
"a",
"specified",
"object",
"using",
"a",
"struct",
"format",
"string",
"."
] | 513b1259c4d7aca5f88cd958edc11828d0712f1b | https://github.com/romanz/trezor-agent/blob/513b1259c4d7aca5f88cd958edc11828d0712f1b/libagent/util.py#L157-L162 | train |
romanz/trezor-agent | libagent/util.py | Reader.read | def read(self, size=None):
"""Read `size` bytes from stream."""
blob = self.s.read(size)
if size is not None and len(blob) < size:
raise EOFError
if self._captured:
self._captured.write(blob)
return blob | python | def read(self, size=None):
"""Read `size` bytes from stream."""
blob = self.s.read(size)
if size is not None and len(blob) < size:
raise EOFError
if self._captured:
self._captured.write(blob)
return blob | [
"def",
"read",
"(",
"self",
",",
"size",
"=",
"None",
")",
":",
"blob",
"=",
"self",
".",
"s",
".",
"read",
"(",
"size",
")",
"if",
"size",
"is",
"not",
"None",
"and",
"len",
"(",
"blob",
")",
"<",
"size",
":",
"raise",
"EOFError",
"if",
"self",
".",
"_captured",
":",
"self",
".",
"_captured",
".",
"write",
"(",
"blob",
")",
"return",
"blob"
] | Read `size` bytes from stream. | [
"Read",
"size",
"bytes",
"from",
"stream",
"."
] | 513b1259c4d7aca5f88cd958edc11828d0712f1b | https://github.com/romanz/trezor-agent/blob/513b1259c4d7aca5f88cd958edc11828d0712f1b/libagent/util.py#L164-L171 | train |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.